[ 536.386355] env[65726]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'linux_bridge' {{(pid=65726) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 536.386582] env[65726]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'noop' {{(pid=65726) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 536.386934] env[65726]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'ovs' {{(pid=65726) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 536.387350] env[65726]: INFO os_vif [-] Loaded VIF plugins: linux_bridge, noop, ovs [ 536.509105] env[65726]: DEBUG oslo_concurrency.processutils [-] Running cmd (subprocess): grep -F node.session.scan /sbin/iscsiadm {{(pid=65726) execute /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/processutils.py:349}} [ 536.519328] env[65726]: DEBUG oslo_concurrency.processutils [-] CMD "grep -F node.session.scan /sbin/iscsiadm" returned: 0 in 0.010s {{(pid=65726) execute /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/processutils.py:372}} [ 536.568830] env[65726]: INFO oslo_service.periodic_task [-] Skipping periodic task _heal_instance_info_cache because its interval is negative [ 537.129233] env[65726]: INFO nova.virt.driver [None req-6fa113db-a7e2-4e32-b048-8407386fb266 None None] Loading compute driver 'vmwareapi.VMwareVCDriver' [ 537.201574] env[65726]: DEBUG oslo_concurrency.lockutils [-] Acquiring lock "oslo_vmware_api_lock" by "oslo_vmware.api.VMwareAPISession._create_session" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 537.201739] env[65726]: DEBUG oslo_concurrency.lockutils [-] Lock "oslo_vmware_api_lock" acquired by "oslo_vmware.api.VMwareAPISession._create_session" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 537.201848] env[65726]: DEBUG oslo_vmware.service [-] Creating suds client with soap_url='https://vc1.osci.c.eu-de-1.cloud.sap:443/sdk' and wsdl_url='https://vc1.osci.c.eu-de-1.cloud.sap:443/sdk/vimService.wsdl' {{(pid=65726) __init__ /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:242}} [ 540.420813] env[65726]: DEBUG oslo_vmware.service [-] Invoking ServiceInstance.RetrieveServiceContent with opID=oslo.vmware-626d729c-8c45-401a-a7f1-bf777a248a78 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.437078] env[65726]: DEBUG oslo_vmware.api [-] Logging into host: vc1.osci.c.eu-de-1.cloud.sap. {{(pid=65726) _create_session /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:242}} [ 540.437238] env[65726]: DEBUG oslo_vmware.service [-] Invoking SessionManager.Login with opID=oslo.vmware-985e7c77-5554-4f10-b928-65d32fc05d4e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.463671] env[65726]: INFO oslo_vmware.api [-] Successfully established new session; session ID is 80dcc. [ 540.463854] env[65726]: DEBUG oslo_concurrency.lockutils [-] Lock "oslo_vmware_api_lock" "released" by "oslo_vmware.api.VMwareAPISession._create_session" :: held 3.262s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 540.464390] env[65726]: INFO nova.virt.vmwareapi.driver [None req-6fa113db-a7e2-4e32-b048-8407386fb266 None None] VMware vCenter version: 7.0.3 [ 540.467851] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39d4721e-cf1e-4f78-a01f-cb3d478b2d8f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.485987] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8acc826-0b34-4a53-acec-42ee7700a990 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.492441] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd055305-ca0e-47a2-8099-da93d9fc04a1 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.499345] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4ec16e7-e4f7-459c-8829-0b9a4461a355 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.512866] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc765d98-1c9e-4eb3-b30b-1b37ce0e0386 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.519353] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8722482-7eca-4a9e-874d-c1ef70a7ff5e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.550576] env[65726]: DEBUG oslo_vmware.service [-] Invoking ExtensionManager.FindExtension with opID=oslo.vmware-8a969185-f624-4448-aac4-4672700810a4 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.556757] env[65726]: DEBUG nova.virt.vmwareapi.driver [None req-6fa113db-a7e2-4e32-b048-8407386fb266 None None] Extension org.openstack.compute already exists. {{(pid=65726) _register_openstack_extension /opt/stack/nova/nova/virt/vmwareapi/driver.py:228}} [ 540.559588] env[65726]: INFO nova.compute.provider_config [None req-6fa113db-a7e2-4e32-b048-8407386fb266 None None] No provider configs found in /etc/nova/provider_config/. If files are present, ensure the Nova process has access. [ 541.063050] env[65726]: DEBUG nova.context [None req-6fa113db-a7e2-4e32-b048-8407386fb266 None None] Found 2 cells: 00000000-0000-0000-0000-000000000000(cell0),c97bc470-e578-4d7f-aa1c-378316e2293c(cell1) {{(pid=65726) load_cells /opt/stack/nova/nova/context.py:472}} [ 541.063331] env[65726]: INFO nova.utils [None req-6fa113db-a7e2-4e32-b048-8407386fb266 None None] The cell worker thread pool MainProcess.cell_worker is initialized [ 541.065590] env[65726]: DEBUG oslo_concurrency.lockutils [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 541.065799] env[65726]: DEBUG oslo_concurrency.lockutils [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 541.066501] env[65726]: DEBUG oslo_concurrency.lockutils [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.001s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 541.066963] env[65726]: DEBUG oslo_concurrency.lockutils [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] Acquiring lock "c97bc470-e578-4d7f-aa1c-378316e2293c" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 541.067156] env[65726]: DEBUG oslo_concurrency.lockutils [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] Lock "c97bc470-e578-4d7f-aa1c-378316e2293c" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 541.068227] env[65726]: DEBUG oslo_concurrency.lockutils [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] Lock "c97bc470-e578-4d7f-aa1c-378316e2293c" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.001s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 541.089385] env[65726]: INFO dbcounter [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] Registered counter for database nova_cell0 [ 541.098156] env[65726]: INFO dbcounter [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] Registered counter for database nova_cell1 [ 541.101645] env[65726]: DEBUG oslo_db.sqlalchemy.engines [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] MySQL server mode set to STRICT_TRANS_TABLES,STRICT_ALL_TABLES,NO_ZERO_IN_DATE,NO_ZERO_DATE,ERROR_FOR_DIVISION_BY_ZERO,TRADITIONAL,NO_AUTO_CREATE_USER,NO_ENGINE_SUBSTITUTION {{(pid=65726) _check_effective_sql_mode /opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/sqlalchemy/engines.py:397}} [ 541.102296] env[65726]: DEBUG oslo_db.sqlalchemy.engines [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] MySQL server mode set to STRICT_TRANS_TABLES,STRICT_ALL_TABLES,NO_ZERO_IN_DATE,NO_ZERO_DATE,ERROR_FOR_DIVISION_BY_ZERO,TRADITIONAL,NO_AUTO_CREATE_USER,NO_ENGINE_SUBSTITUTION {{(pid=65726) _check_effective_sql_mode /opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/sqlalchemy/engines.py:397}} [ 541.106611] env[65726]: ERROR nova.db.main.api [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] No DB access allowed in nova-compute: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenpool.py", line 87, in _spawn_n_impl [ 541.106611] env[65726]: func(*args, **kwargs) [ 541.106611] env[65726]: File "/opt/stack/data/venv/lib/python3.10/site-packages/futurist/_green.py", line 69, in __call__ [ 541.106611] env[65726]: self.work.run() [ 541.106611] env[65726]: File "/opt/stack/data/venv/lib/python3.10/site-packages/futurist/_utils.py", line 45, in run [ 541.106611] env[65726]: result = self.fn(*self.args, **self.kwargs) [ 541.106611] env[65726]: File "/opt/stack/nova/nova/utils.py", line 585, in context_wrapper [ 541.106611] env[65726]: return func(*args, **kwargs) [ 541.106611] env[65726]: File "/opt/stack/nova/nova/context.py", line 420, in gather_result [ 541.106611] env[65726]: result = fn(*args, **kwargs) [ 541.106611] env[65726]: File "/opt/stack/nova/nova/db/main/api.py", line 179, in wrapper [ 541.106611] env[65726]: return f(*args, **kwargs) [ 541.106611] env[65726]: File "/opt/stack/nova/nova/objects/service.py", line 568, in _db_service_get_minimum_version [ 541.106611] env[65726]: return db.service_get_minimum_version(context, binaries) [ 541.106611] env[65726]: File "/opt/stack/nova/nova/db/main/api.py", line 238, in wrapper [ 541.106611] env[65726]: _check_db_access() [ 541.106611] env[65726]: File "/opt/stack/nova/nova/db/main/api.py", line 188, in _check_db_access [ 541.106611] env[65726]: stacktrace = ''.join(traceback.format_stack()) [ 541.106611] env[65726]: [ 541.108332] env[65726]: ERROR nova.db.main.api [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] No DB access allowed in nova-compute: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenpool.py", line 87, in _spawn_n_impl [ 541.108332] env[65726]: func(*args, **kwargs) [ 541.108332] env[65726]: File "/opt/stack/data/venv/lib/python3.10/site-packages/futurist/_green.py", line 69, in __call__ [ 541.108332] env[65726]: self.work.run() [ 541.108332] env[65726]: File "/opt/stack/data/venv/lib/python3.10/site-packages/futurist/_utils.py", line 45, in run [ 541.108332] env[65726]: result = self.fn(*self.args, **self.kwargs) [ 541.108332] env[65726]: File "/opt/stack/nova/nova/utils.py", line 585, in context_wrapper [ 541.108332] env[65726]: return func(*args, **kwargs) [ 541.108332] env[65726]: File "/opt/stack/nova/nova/context.py", line 420, in gather_result [ 541.108332] env[65726]: result = fn(*args, **kwargs) [ 541.108332] env[65726]: File "/opt/stack/nova/nova/db/main/api.py", line 179, in wrapper [ 541.108332] env[65726]: return f(*args, **kwargs) [ 541.108332] env[65726]: File "/opt/stack/nova/nova/objects/service.py", line 568, in _db_service_get_minimum_version [ 541.108332] env[65726]: return db.service_get_minimum_version(context, binaries) [ 541.108332] env[65726]: File "/opt/stack/nova/nova/db/main/api.py", line 238, in wrapper [ 541.108332] env[65726]: _check_db_access() [ 541.108332] env[65726]: File "/opt/stack/nova/nova/db/main/api.py", line 188, in _check_db_access [ 541.108332] env[65726]: stacktrace = ''.join(traceback.format_stack()) [ 541.108332] env[65726]: [ 541.108880] env[65726]: WARNING nova.objects.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] Failed to get minimum service version for cell 00000000-0000-0000-0000-000000000000 [ 541.109016] env[65726]: WARNING nova.objects.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] Failed to get minimum service version for cell c97bc470-e578-4d7f-aa1c-378316e2293c [ 541.109429] env[65726]: DEBUG oslo_concurrency.lockutils [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] Acquiring lock "singleton_lock" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 541.109611] env[65726]: DEBUG oslo_concurrency.lockutils [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] Acquired lock "singleton_lock" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 541.109856] env[65726]: DEBUG oslo_concurrency.lockutils [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] Releasing lock "singleton_lock" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 541.110191] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] Full set of CONF: {{(pid=65726) _wait_for_exit_or_signal /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/service.py:275}} [ 541.110346] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] ******************************************************************************** {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2804}} [ 541.110468] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] Configuration options gathered from: {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2805}} [ 541.110599] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] command line args: ['--config-file', '/etc/nova/nova.conf', '--config-file', '/etc/nova/nova-cpu-common.conf', '--config-file', '/etc/nova/nova-cpu-1.conf'] {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2806}} [ 541.110784] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] config files: ['/etc/nova/nova.conf', '/etc/nova/nova-cpu-common.conf', '/etc/nova/nova-cpu-1.conf'] {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2807}} [ 541.110899] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] ================================================================================ {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2809}} [ 541.111108] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] allow_resize_to_same_host = True {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 541.111277] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] arq_binding_timeout = 300 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 541.111393] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] backdoor_port = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 541.111508] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] backdoor_socket = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 541.111661] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] block_device_allocate_retries = 60 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 541.111808] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] block_device_allocate_retries_interval = 3 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 541.111955] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] cell_worker_thread_pool_size = 5 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 541.112127] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] cert = self.pem {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 541.112284] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] compute_driver = vmwareapi.VMwareVCDriver {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 541.112440] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] compute_monitors = [] {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 541.112592] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] config_dir = [] {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 541.112749] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] config_drive_format = iso9660 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 541.112870] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] config_file = ['/etc/nova/nova.conf', '/etc/nova/nova-cpu-common.conf', '/etc/nova/nova-cpu-1.conf'] {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 541.113032] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] config_source = [] {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 541.113193] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] console_host = devstack {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 541.113346] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] control_exchange = nova {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 541.113572] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] cpu_allocation_ratio = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 541.113826] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] daemon = False {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 541.114103] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] debug = True {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 541.114360] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] default_access_ip_network_name = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 541.114631] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] default_availability_zone = nova {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 541.114882] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] default_ephemeral_format = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 541.115187] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] default_green_pool_size = 1000 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 541.115426] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] default_log_levels = ['amqp=WARN', 'amqplib=WARN', 'boto=WARN', 'qpid=WARN', 'sqlalchemy=WARN', 'suds=INFO', 'oslo.messaging=INFO', 'oslo_messaging=INFO', 'iso8601=WARN', 'requests.packages.urllib3.connectionpool=WARN', 'urllib3.connectionpool=WARN', 'websocket=WARN', 'requests.packages.urllib3.util.retry=WARN', 'urllib3.util.retry=WARN', 'keystonemiddleware=WARN', 'routes.middleware=WARN', 'stevedore=WARN', 'taskflow=WARN', 'keystoneauth=WARN', 'oslo.cache=INFO', 'oslo_policy=INFO', 'dogpile.core.dogpile=INFO', 'glanceclient=WARN', 'oslo.privsep.daemon=INFO'] {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 541.115587] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] default_schedule_zone = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 541.115744] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] default_thread_pool_size = 10 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 541.115932] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] disk_allocation_ratio = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 541.116046] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] enable_new_services = True {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 541.116203] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] flat_injected = False {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 541.116349] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] force_config_drive = False {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 541.116493] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] force_raw_images = True {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 541.116651] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] graceful_shutdown_timeout = 5 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 541.116798] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] heal_instance_info_cache_interval = -1 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 541.117027] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] host = cpu-1 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 541.117230] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] initial_cpu_allocation_ratio = 4.0 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 541.117401] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] initial_disk_allocation_ratio = 1.0 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 541.117556] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] initial_ram_allocation_ratio = 1.0 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 541.117769] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] injected_network_template = /opt/stack/nova/nova/virt/interfaces.template {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 541.117919] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] instance_build_timeout = 0 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 541.118081] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] instance_delete_interval = 300 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 541.118257] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] instance_format = [instance: %(uuid)s] {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 541.118424] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] instance_name_template = instance-%08x {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 541.118576] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] instance_usage_audit = False {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 541.118734] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] instance_usage_audit_period = month {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 541.118887] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] instance_uuid_format = [instance: %(uuid)s] {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 541.119055] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] instances_path = /opt/stack/data/nova/instances {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 541.119214] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] internal_service_availability_zone = internal {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 541.119359] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] key = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 541.119510] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] live_migration_retry_count = 30 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 541.119666] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] log_color = False {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 541.119816] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] log_config_append = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 541.119968] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] log_date_format = %Y-%m-%d %H:%M:%S {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 541.120128] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] log_dir = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 541.120274] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] log_file = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 541.120391] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] log_options = True {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 541.120538] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] log_rotate_interval = 1 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 541.120693] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] log_rotate_interval_type = days {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 541.120846] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] log_rotation_type = none {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 541.120964] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] logging_context_format_string = %(color)s%(levelname)s %(name)s [%(global_request_id)s %(request_id)s %(project_name)s %(user_name)s%(color)s] %(instance)s%(color)s%(message)s {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 541.121088] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] logging_debug_format_suffix = {{(pid=%(process)d) %(funcName)s %(pathname)s:%(lineno)d}} {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 541.121246] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] logging_default_format_string = %(color)s%(levelname)s %(name)s [-%(color)s] %(instance)s%(color)s%(message)s {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 541.121400] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] logging_exception_prefix = ERROR %(name)s %(instance)s {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 541.121518] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] logging_user_identity_format = %(user)s %(project)s %(domain)s %(system_scope)s %(user_domain)s %(project_domain)s {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 541.121666] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] long_rpc_timeout = 1800 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 541.121816] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] max_concurrent_builds = 10 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 541.121962] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] max_concurrent_live_migrations = 1 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 541.122119] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] max_concurrent_snapshots = 5 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 541.122266] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] max_local_block_devices = 3 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 541.122415] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] max_logfile_count = 30 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 541.122558] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] max_logfile_size_mb = 200 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 541.122706] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] maximum_instance_delete_attempts = 5 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 541.122850] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] migrate_max_retries = -1 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 541.123011] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] mkisofs_cmd = genisoimage {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 541.123210] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] my_block_storage_ip = 10.180.1.21 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 541.123333] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] my_ip = 10.180.1.21 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 541.123520] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] my_shared_fs_storage_ip = 10.180.1.21 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 541.123670] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] network_allocate_retries = 0 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 541.123835] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] non_inheritable_image_properties = ['cache_in_nova', 'bittorrent'] {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 541.123988] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] osapi_compute_unique_server_name_scope = {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 541.124151] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] password_length = 12 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 541.124301] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] periodic_enable = True {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 541.124448] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] periodic_fuzzy_delay = 60 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 541.124598] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] pointer_model = usbtablet {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 541.124750] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] preallocate_images = none {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 541.124941] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] publish_errors = False {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 541.125102] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] pybasedir = /opt/stack/nova {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 541.125262] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] ram_allocation_ratio = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 541.125415] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] rate_limit_burst = 0 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 541.125573] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] rate_limit_except_level = CRITICAL {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 541.125719] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] rate_limit_interval = 0 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 541.125869] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] reboot_timeout = 0 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 541.126029] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] reclaim_instance_interval = 0 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 541.126182] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] record = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 541.126342] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] reimage_timeout_per_gb = 60 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 541.126493] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] report_interval = 120 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 541.126641] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] rescue_timeout = 0 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 541.126786] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] reserved_host_cpus = 0 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 541.126936] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] reserved_host_disk_mb = 0 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 541.127095] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] reserved_host_memory_mb = 512 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 541.127271] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] reserved_huge_pages = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 541.127430] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] resize_confirm_window = 0 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 541.127578] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] resize_fs_using_block_device = False {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 541.127724] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] resume_guests_state_on_host_boot = False {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 541.127879] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] rootwrap_config = /etc/nova/rootwrap.conf {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 541.128039] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] rpc_response_timeout = 60 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 541.128215] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] run_external_periodic_tasks = True {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 541.128388] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] running_deleted_instance_action = reap {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 541.128539] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] running_deleted_instance_poll_interval = 1800 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 541.128689] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] running_deleted_instance_timeout = 0 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 541.128835] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] scheduler_instance_sync_interval = 120 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 541.128991] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] service_down_time = 720 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 541.129162] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] servicegroup_driver = db {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 541.129307] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] shell_completion = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 541.129455] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] shelved_offload_time = 0 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 541.129604] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] shelved_poll_interval = 3600 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 541.129758] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] shutdown_timeout = 0 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 541.129909] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] source_is_ipv6 = False {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 541.130065] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] ssl_only = False {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 541.130318] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] state_path = /opt/stack/data/n-cpu-1 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 541.130478] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] sync_power_state_interval = 600 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 541.130627] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] sync_power_state_pool_size = 1000 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 541.130780] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] syslog_log_facility = LOG_USER {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 541.130923] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] tempdir = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 541.131085] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] thread_pool_statistic_period = -1 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 541.131236] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] timeout_nbd = 10 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 541.131396] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] transport_url = **** {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 541.131543] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] update_resources_interval = 0 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 541.131686] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] use_cow_images = True {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 541.131831] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] use_journal = False {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 541.131974] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] use_json = False {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 541.132132] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] use_rootwrap_daemon = False {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 541.132279] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] use_stderr = False {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 541.132425] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] use_syslog = False {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 541.132566] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] vcpu_pin_set = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 541.132714] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] vif_plugging_is_fatal = True {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 541.132865] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] vif_plugging_timeout = 300 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 541.133026] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] virt_mkfs = [] {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 541.133178] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] volume_usage_poll_interval = 0 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 541.133327] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] watch_log_file = False {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 541.133479] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] web = /usr/share/spice-html5 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 541.133648] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] oslo_concurrency.disable_process_locking = False {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.134293] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] oslo_concurrency.lock_path = /opt/stack/data/n-cpu-1 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.134479] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] os_brick.lock_path = /opt/stack/data/os_brick {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.134644] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] os_brick.wait_mpath_device_attempts = 4 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.134800] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] os_brick.wait_mpath_device_interval = 1 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.134964] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] oslo_messaging_metrics.metrics_buffer_size = 1000 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.135133] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] oslo_messaging_metrics.metrics_enabled = False {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.135297] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] oslo_messaging_metrics.metrics_process_name = {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.135459] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] oslo_messaging_metrics.metrics_socket_file = /var/tmp/metrics_collector.sock {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.135612] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] oslo_messaging_metrics.metrics_thread_stop_timeout = 10 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.135770] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] api.compute_link_prefix = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.135927] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] api.config_drive_skip_versions = 1.0 2007-01-19 2007-03-01 2007-08-29 2007-10-10 2007-12-15 2008-02-01 2008-09-01 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.136092] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] api.dhcp_domain = novalocal {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.136248] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] api.enable_instance_password = True {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.136397] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] api.glance_link_prefix = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.136548] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] api.instance_list_cells_batch_fixed_size = 100 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.136701] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] api.instance_list_cells_batch_strategy = distributed {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.136851] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] api.instance_list_per_project_cells = False {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.137006] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] api.list_records_by_skipping_down_cells = True {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.137167] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] api.local_metadata_per_cell = False {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.137350] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] api.max_limit = 1000 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.137511] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] api.metadata_cache_expiration = 15 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.137672] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] api.neutron_default_project_id = default {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.137829] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] api.response_validation = warn {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.137977] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] api.use_neutron_default_nets = False {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.138388] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] api.vendordata_dynamic_connect_timeout = 5 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.138388] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] api.vendordata_dynamic_failure_fatal = False {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.138462] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] api.vendordata_dynamic_read_timeout = 5 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.138607] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] api.vendordata_dynamic_ssl_certfile = {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.138764] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] api.vendordata_dynamic_targets = [] {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.138915] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] api.vendordata_jsonfile_path = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.139090] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] api.vendordata_providers = ['StaticJSON'] {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.139287] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] cache.backend = dogpile.cache.memcached {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.139512] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] cache.backend_argument = **** {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.139684] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] cache.backend_expiration_time = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.139845] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] cache.config_prefix = cache.oslo {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.139994] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] cache.debug_cache_backend = False {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.140160] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] cache.enable_retry_client = False {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.140310] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] cache.enable_socket_keepalive = False {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.140467] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] cache.enabled = True {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.140618] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] cache.enforce_fips_mode = False {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.140770] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] cache.expiration_time = 600 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.140930] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] cache.hashclient_dead_timeout = 60.0 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.141096] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] cache.hashclient_retry_attempts = 2 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.141260] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] cache.hashclient_retry_timeout = 1.0 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.141415] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] cache.memcache_dead_retry = 300 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.141561] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] cache.memcache_password = **** {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.141712] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] cache.memcache_pool_connection_get_timeout = 10 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.141863] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] cache.memcache_pool_flush_on_reconnect = False {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.142023] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] cache.memcache_pool_maxsize = 10 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.142179] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] cache.memcache_pool_unused_timeout = 60 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.142328] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] cache.memcache_sasl_enabled = False {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.142493] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] cache.memcache_servers = ['localhost:11211'] {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.142651] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] cache.memcache_socket_timeout = 1.0 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.142801] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] cache.memcache_username = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.142956] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] cache.proxies = [] {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.143124] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] cache.redis_db = 0 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.143275] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] cache.redis_password = **** {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.143435] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] cache.redis_sentinel_service_name = mymaster {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.143620] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] cache.redis_sentinels = ['localhost:26379'] {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.143791] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] cache.redis_server = localhost:6379 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.143950] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] cache.redis_socket_timeout = 1.0 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.144115] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] cache.redis_username = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.144270] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] cache.retry_attempts = 2 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.144426] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] cache.retry_delay = 0.0 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.144578] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] cache.socket_keepalive_count = 1 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.144725] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] cache.socket_keepalive_idle = 1 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.144870] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] cache.socket_keepalive_interval = 1 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.145031] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] cache.tls_allowed_ciphers = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.145183] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] cache.tls_cafile = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.145326] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] cache.tls_certfile = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.145472] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] cache.tls_enabled = False {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.145613] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] cache.tls_keyfile = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.145769] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] cinder.auth_section = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.145930] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] cinder.auth_type = password {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.146090] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] cinder.cafile = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.146256] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] cinder.catalog_info = volumev3::publicURL {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.146403] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] cinder.certfile = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.146556] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] cinder.collect_timing = False {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.146704] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] cinder.cross_az_attach = True {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.146851] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] cinder.debug = False {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.146997] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] cinder.endpoint_template = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.147170] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] cinder.http_retries = 3 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.147353] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] cinder.insecure = False {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.147508] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] cinder.keyfile = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.147667] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] cinder.os_region_name = RegionOne {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.147819] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] cinder.split_loggers = False {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.147966] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] cinder.timeout = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.148144] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] compute.consecutive_build_service_disable_threshold = 10 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.148348] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] compute.cpu_dedicated_set = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.148562] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] compute.cpu_shared_set = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.148733] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] compute.image_type_exclude_list = [] {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.148887] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] compute.live_migration_wait_for_vif_plug = True {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.149054] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] compute.max_concurrent_disk_ops = 0 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.149211] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] compute.max_disk_devices_to_attach = -1 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.149364] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] compute.packing_host_numa_cells_allocation_strategy = False {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.149542] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] compute.provider_config_location = /etc/nova/provider_config/ {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.149728] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] compute.resource_provider_association_refresh = 300 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.149885] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] compute.sharing_providers_max_uuids_per_request = 200 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.150046] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] compute.shutdown_retry_interval = 10 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.150217] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] compute.vmdk_allowed_types = ['streamOptimized', 'monolithicSparse'] {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.150386] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] conductor.workers = 2 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.150547] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] console.allowed_origins = [] {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.150736] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] console.ssl_ciphers = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.150907] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] console.ssl_minimum_version = default {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.151076] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] consoleauth.enforce_session_timeout = False {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.151238] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] consoleauth.token_ttl = 600 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.151395] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] cyborg.cafile = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.151544] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] cyborg.certfile = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.151694] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] cyborg.collect_timing = False {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.151840] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] cyborg.connect_retries = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.151987] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] cyborg.connect_retry_delay = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.152148] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] cyborg.endpoint_override = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.152298] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] cyborg.insecure = False {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.152441] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] cyborg.keyfile = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.152590] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] cyborg.max_version = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.152733] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] cyborg.min_version = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.152877] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] cyborg.region_name = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.153034] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] cyborg.retriable_status_codes = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.153183] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] cyborg.service_name = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.153340] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] cyborg.service_type = accelerator {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.153486] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] cyborg.split_loggers = False {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.153631] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] cyborg.status_code_retries = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.153774] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] cyborg.status_code_retry_delay = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.153917] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] cyborg.timeout = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.154092] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] cyborg.valid_interfaces = ['internal', 'public'] {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.154242] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] cyborg.version = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.154399] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] database.asyncio_connection = **** {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.154547] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] database.asyncio_slave_connection = **** {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.154702] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] database.backend = sqlalchemy {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.154857] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] database.connection = **** {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.155025] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] database.connection_debug = 0 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.155183] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] database.connection_parameters = {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.155340] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] database.connection_recycle_time = 3600 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.155492] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] database.connection_trace = False {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.155665] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] database.db_inc_retry_interval = True {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.155832] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] database.db_max_retries = 20 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.155986] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] database.db_max_retry_interval = 10 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.156153] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] database.db_retry_interval = 1 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.156305] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] database.max_overflow = 50 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.156459] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] database.max_pool_size = 5 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.156606] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] database.max_retries = 10 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.156760] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] database.mysql_sql_mode = TRADITIONAL {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.156910] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] database.mysql_wsrep_sync_wait = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.157066] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] database.pool_timeout = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.157246] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] database.retry_interval = 10 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.157408] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] database.slave_connection = **** {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.157559] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] database.sqlite_synchronous = True {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.157707] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] database.use_db_reconnect = False {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.157861] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] api_database.asyncio_connection = **** {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.158013] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] api_database.asyncio_slave_connection = **** {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.158192] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] api_database.backend = sqlalchemy {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.158381] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] api_database.connection = **** {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.158543] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] api_database.connection_debug = 0 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.158699] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] api_database.connection_parameters = {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.158851] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] api_database.connection_recycle_time = 3600 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.159013] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] api_database.connection_trace = False {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.159167] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] api_database.db_inc_retry_interval = True {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.159320] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] api_database.db_max_retries = 20 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.159471] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] api_database.db_max_retry_interval = 10 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.159658] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] api_database.db_retry_interval = 1 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.159826] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] api_database.max_overflow = 50 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.159976] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] api_database.max_pool_size = 5 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.160141] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] api_database.max_retries = 10 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.160300] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] api_database.mysql_sql_mode = TRADITIONAL {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.160448] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] api_database.mysql_wsrep_sync_wait = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.160594] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] api_database.pool_timeout = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.160743] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] api_database.retry_interval = 10 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.160887] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] api_database.slave_connection = **** {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.161046] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] api_database.sqlite_synchronous = True {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.161216] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] devices.enabled_mdev_types = [] {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.161380] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] ephemeral_storage_encryption.cipher = aes-xts-plain64 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.161537] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] ephemeral_storage_encryption.default_format = luks {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.161755] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] ephemeral_storage_encryption.enabled = False {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.161935] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] ephemeral_storage_encryption.key_size = 512 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.162123] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] glance.api_servers = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.162283] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] glance.cafile = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.162439] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] glance.certfile = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.162595] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] glance.collect_timing = False {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.162742] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] glance.connect_retries = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.162888] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] glance.connect_retry_delay = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.163048] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] glance.debug = False {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.163207] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] glance.default_trusted_certificate_ids = [] {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.163359] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] glance.enable_certificate_validation = False {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.163505] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] glance.enable_rbd_download = False {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.163650] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] glance.endpoint_override = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.163804] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] glance.insecure = False {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.163948] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] glance.keyfile = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.164109] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] glance.max_version = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.164255] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] glance.min_version = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.164406] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] glance.num_retries = 3 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.164561] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] glance.rbd_ceph_conf = {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.164710] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] glance.rbd_connect_timeout = 5 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.164862] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] glance.rbd_pool = {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.165032] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] glance.rbd_user = {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.165211] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] glance.region_name = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.165367] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] glance.retriable_status_codes = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.165515] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] glance.service_name = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.165675] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] glance.service_type = image {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.165822] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] glance.split_loggers = False {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.165968] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] glance.status_code_retries = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.166155] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] glance.status_code_retry_delay = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.166305] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] glance.timeout = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.166476] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] glance.valid_interfaces = ['internal', 'public'] {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.166628] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] glance.verify_glance_signatures = False {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.166774] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] glance.version = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.166930] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] guestfs.debug = False {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.167100] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] manila.auth_section = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.167277] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] manila.auth_type = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.167434] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] manila.cafile = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.167579] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] manila.certfile = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.167731] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] manila.collect_timing = False {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.167880] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] manila.connect_retries = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.168039] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] manila.connect_retry_delay = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.168190] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] manila.endpoint_override = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.168341] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] manila.insecure = False {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.168486] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] manila.keyfile = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.168631] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] manila.max_version = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.168773] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] manila.min_version = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.168917] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] manila.region_name = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.169072] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] manila.retriable_status_codes = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.169220] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] manila.service_name = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.169377] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] manila.service_type = shared-file-system {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.169528] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] manila.share_apply_policy_timeout = 10 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.169675] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] manila.split_loggers = False {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.169821] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] manila.status_code_retries = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.169967] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] manila.status_code_retry_delay = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.170127] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] manila.timeout = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.170295] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] manila.valid_interfaces = ['internal', 'public'] {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.170444] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] manila.version = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.170599] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] mks.enabled = False {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.170948] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] mks.mksproxy_base_url = http://127.0.0.1:6090/ {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.171139] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] image_cache.manager_interval = 2400 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.171299] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] image_cache.precache_concurrency = 1 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.171469] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] image_cache.remove_unused_base_images = True {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.171631] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] image_cache.remove_unused_original_minimum_age_seconds = 86400 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.171786] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] image_cache.remove_unused_resized_minimum_age_seconds = 3600 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.171952] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] image_cache.subdirectory_name = _base {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.172130] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] ironic.api_max_retries = 60 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.172286] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] ironic.api_retry_interval = 2 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.172436] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] ironic.auth_section = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.172584] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] ironic.auth_type = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.172731] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] ironic.cafile = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.172875] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] ironic.certfile = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.173037] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] ironic.collect_timing = False {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.173194] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] ironic.conductor_group = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.173342] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] ironic.connect_retries = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.173490] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] ironic.connect_retry_delay = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.173635] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] ironic.endpoint_override = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.173782] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] ironic.insecure = False {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.173924] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] ironic.keyfile = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.174087] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] ironic.max_version = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.174239] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] ironic.min_version = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.174392] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] ironic.peer_list = [] {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.174539] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] ironic.region_name = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.174684] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] ironic.retriable_status_codes = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.174836] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] ironic.serial_console_state_timeout = 10 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.174982] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] ironic.service_name = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.175183] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] ironic.service_type = baremetal {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.175345] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] ironic.shard = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.175493] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] ironic.split_loggers = False {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.175641] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] ironic.status_code_retries = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.175789] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] ironic.status_code_retry_delay = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.175935] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] ironic.timeout = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.176116] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] ironic.valid_interfaces = ['internal', 'public'] {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.176266] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] ironic.version = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.176440] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] key_manager.backend = nova.keymgr.conf_key_mgr.ConfKeyManager {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.176602] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] key_manager.fixed_key = **** {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.176771] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] barbican.auth_endpoint = http://localhost/identity/v3 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.176922] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] barbican.barbican_api_version = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.177081] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] barbican.barbican_endpoint = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.177272] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] barbican.barbican_endpoint_type = public {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.177427] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] barbican.barbican_region_name = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.177605] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] barbican.cafile = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.177840] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] barbican.certfile = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.178027] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] barbican.collect_timing = False {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.178188] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] barbican.insecure = False {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.178376] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] barbican.keyfile = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.178565] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] barbican.number_of_retries = 60 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.178756] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] barbican.retry_delay = 1 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.178933] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] barbican.send_service_user_token = False {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.179117] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] barbican.split_loggers = False {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.179268] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] barbican.timeout = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.179422] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] barbican.verify_ssl = True {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.179568] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] barbican.verify_ssl_path = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.179722] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] barbican_service_user.auth_section = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.179873] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] barbican_service_user.auth_type = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.180030] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] barbican_service_user.cafile = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.180181] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] barbican_service_user.certfile = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.180335] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] barbican_service_user.collect_timing = False {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.180481] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] barbican_service_user.insecure = False {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.180627] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] barbican_service_user.keyfile = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.180776] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] barbican_service_user.split_loggers = False {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.180920] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] barbican_service_user.timeout = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.181099] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] vault.approle_role_id = **** {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.181345] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] vault.approle_secret_id = **** {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.181620] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] vault.kv_mountpoint = secret {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.181881] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] vault.kv_path = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.182157] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] vault.kv_version = 2 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.182420] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] vault.namespace = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.182680] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] vault.root_token_id = **** {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.182931] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] vault.ssl_ca_crt_file = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.183214] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] vault.timeout = 60.0 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.183481] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] vault.use_ssl = False {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.183732] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] vault.vault_url = http://127.0.0.1:8200 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.183914] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] keystone.cafile = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.184081] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] keystone.certfile = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.184289] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] keystone.collect_timing = False {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.184541] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] keystone.connect_retries = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.184747] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] keystone.connect_retry_delay = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.184917] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] keystone.endpoint_override = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.185124] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] keystone.insecure = False {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.185299] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] keystone.keyfile = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.185453] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] keystone.max_version = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.185599] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] keystone.min_version = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.185812] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] keystone.region_name = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.186062] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] keystone.retriable_status_codes = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.186268] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] keystone.service_name = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.186445] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] keystone.service_type = identity {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.186605] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] keystone.split_loggers = False {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.186770] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] keystone.status_code_retries = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.186926] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] keystone.status_code_retry_delay = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.187087] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] keystone.timeout = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.187287] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] keystone.valid_interfaces = ['internal', 'public'] {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.187440] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] keystone.version = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.187621] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] libvirt.ceph_mount_options = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.187954] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] libvirt.ceph_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.188146] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] libvirt.connection_uri = {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.188330] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] libvirt.cpu_mode = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.188533] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] libvirt.cpu_model_extra_flags = [] {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.188700] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] libvirt.cpu_models = [] {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.188862] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] libvirt.cpu_power_governor_high = performance {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.189054] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] libvirt.cpu_power_governor_low = powersave {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.189230] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] libvirt.cpu_power_management = False {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.189401] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] libvirt.cpu_power_management_strategy = cpu_state {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.189555] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] libvirt.device_detach_attempts = 8 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.189705] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] libvirt.device_detach_timeout = 20 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.189857] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] libvirt.disk_cachemodes = [] {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.190009] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] libvirt.disk_prefix = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.190179] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] libvirt.enabled_perf_events = [] {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.190328] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] libvirt.file_backed_memory = 0 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.190479] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] libvirt.gid_maps = [] {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.190646] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] libvirt.hw_disk_discard = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.190805] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] libvirt.hw_machine_type = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.190991] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] libvirt.images_rbd_ceph_conf = {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.191198] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] libvirt.images_rbd_glance_copy_poll_interval = 15 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.191366] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] libvirt.images_rbd_glance_copy_timeout = 600 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.191551] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] libvirt.images_rbd_glance_store_name = {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.191726] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] libvirt.images_rbd_pool = rbd {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.191887] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] libvirt.images_type = default {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.192213] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] libvirt.images_volume_group = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.192213] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] libvirt.inject_key = False {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.192415] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] libvirt.inject_partition = -2 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.192478] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] libvirt.inject_password = False {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.192632] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] libvirt.iscsi_iface = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.192791] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] libvirt.iser_use_multipath = False {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.193046] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] libvirt.live_migration_bandwidth = 0 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.193205] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] libvirt.live_migration_completion_timeout = 800 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.193426] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] libvirt.live_migration_downtime = 500 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.193717] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] libvirt.live_migration_downtime_delay = 75 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.193888] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] libvirt.live_migration_downtime_steps = 10 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.194158] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] libvirt.live_migration_inbound_addr = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.194491] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] libvirt.live_migration_parallel_connections = 1 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.194616] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] libvirt.live_migration_permit_auto_converge = False {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.194791] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] libvirt.live_migration_permit_post_copy = False {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.194967] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] libvirt.live_migration_scheme = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.195121] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] libvirt.live_migration_timeout_action = abort {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.195278] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] libvirt.live_migration_tunnelled = False {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.195432] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] libvirt.live_migration_uri = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.195581] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] libvirt.live_migration_with_native_tls = False {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.195730] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] libvirt.max_queues = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.195882] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] libvirt.mem_stats_period_seconds = 10 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.196113] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] libvirt.migration_inbound_addr = 10.180.1.21 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.196268] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] libvirt.nfs_mount_options = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.196579] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] libvirt.nfs_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.196753] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] libvirt.num_aoe_discover_tries = 3 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.196910] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] libvirt.num_iser_scan_tries = 5 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.197074] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] libvirt.num_memory_encrypted_guests = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.197258] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] libvirt.num_nvme_discover_tries = 5 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.197426] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] libvirt.num_pcie_ports = 0 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.197584] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] libvirt.num_volume_scan_tries = 5 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.197741] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] libvirt.pmem_namespaces = [] {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.197891] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] libvirt.quobyte_client_cfg = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.198207] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] libvirt.quobyte_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.198384] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] libvirt.rbd_connect_timeout = 5 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.198533] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] libvirt.rbd_destroy_volume_retries = 12 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.198688] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] libvirt.rbd_destroy_volume_retry_interval = 5 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.198836] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] libvirt.rbd_secret_uuid = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.198982] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] libvirt.rbd_user = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.199147] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] libvirt.realtime_scheduler_priority = 1 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.199305] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] libvirt.remote_filesystem_transport = ssh {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.199454] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] libvirt.rescue_image_id = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.199625] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] libvirt.rescue_kernel_id = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.199779] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] libvirt.rescue_ramdisk_id = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.199934] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] libvirt.rng_dev_path = /dev/urandom {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.200092] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] libvirt.rx_queue_size = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.200249] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] libvirt.smbfs_mount_options = {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.200532] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] libvirt.smbfs_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.200700] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] libvirt.snapshot_compression = False {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.200850] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] libvirt.snapshot_image_format = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.201077] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] libvirt.snapshots_directory = /opt/stack/data/nova/instances/snapshots {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.201234] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] libvirt.sparse_logical_volumes = False {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.201406] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] libvirt.supported_tpm_secret_security = ['user', 'host'] {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.201559] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] libvirt.swtpm_enabled = False {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.201716] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] libvirt.swtpm_group = tss {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.201874] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] libvirt.swtpm_user = tss {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.202044] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] libvirt.sysinfo_serial = unique {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.202198] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] libvirt.tb_cache_size = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.202350] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] libvirt.tx_queue_size = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.202500] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] libvirt.uid_maps = [] {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.202649] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] libvirt.use_default_aio_mode_for_volumes = False {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.202796] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] libvirt.use_virtio_for_bridges = True {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.202952] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] libvirt.virt_type = kvm {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.203119] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] libvirt.volume_clear = zero {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.203271] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] libvirt.volume_clear_size = 0 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.203422] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] libvirt.volume_enforce_multipath = False {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.203572] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] libvirt.volume_use_multipath = False {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.203716] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] libvirt.vzstorage_cache_path = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.203870] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] libvirt.vzstorage_log_path = /var/log/vstorage/%(cluster_name)s/nova.log.gz {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.204335] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] libvirt.vzstorage_mount_group = qemu {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.204551] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] libvirt.vzstorage_mount_opts = [] {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.204717] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] libvirt.vzstorage_mount_perms = 0770 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.204996] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] libvirt.vzstorage_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.205185] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] libvirt.vzstorage_mount_user = stack {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.205345] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] libvirt.wait_soft_reboot_seconds = 120 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.205510] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] neutron.auth_section = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.205681] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] neutron.auth_type = password {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.205833] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] neutron.cafile = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.205980] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] neutron.certfile = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.206144] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] neutron.collect_timing = False {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.206295] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] neutron.connect_retries = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.206446] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] neutron.connect_retry_delay = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.206606] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] neutron.default_floating_pool = public {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.206755] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] neutron.endpoint_override = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.206907] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] neutron.extension_sync_interval = 600 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.207069] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] neutron.http_retries = 3 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.207255] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] neutron.insecure = False {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.207412] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] neutron.keyfile = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.207566] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] neutron.max_version = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.207723] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] neutron.metadata_proxy_shared_secret = **** {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.207871] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] neutron.min_version = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.208041] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] neutron.ovs_bridge = br-int {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.208221] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] neutron.physnets = [] {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.208399] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] neutron.region_name = RegionOne {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.208550] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] neutron.retriable_status_codes = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.208707] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] neutron.service_metadata_proxy = True {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.208856] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] neutron.service_name = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.209024] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] neutron.service_type = network {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.209181] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] neutron.split_loggers = False {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.209338] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] neutron.status_code_retries = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.209586] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] neutron.status_code_retry_delay = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.209664] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] neutron.timeout = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.209841] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] neutron.valid_interfaces = ['internal', 'public'] {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.209995] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] neutron.version = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.210172] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] notifications.bdms_in_notifications = False {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.210341] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] notifications.default_level = INFO {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.210496] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] notifications.include_share_mapping = False {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.210659] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] notifications.notification_format = unversioned {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.210809] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] notifications.notify_on_state_change = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.210971] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] notifications.versioned_notifications_topics = ['versioned_notifications'] {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.211153] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] pci.alias = [] {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.211311] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] pci.device_spec = [] {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.211464] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] pci.report_in_placement = False {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.211637] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] placement.auth_section = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.211851] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] placement.auth_type = password {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.212028] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] placement.auth_url = http://10.180.1.21/identity {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.212186] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] placement.cafile = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.212382] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] placement.certfile = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.212489] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] placement.collect_timing = False {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.212636] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] placement.connect_retries = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.212782] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] placement.connect_retry_delay = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.212931] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] placement.default_domain_id = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.213089] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] placement.default_domain_name = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.213238] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] placement.domain_id = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.213383] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] placement.domain_name = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.213528] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] placement.endpoint_override = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.213675] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] placement.insecure = False {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.213817] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] placement.keyfile = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.213959] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] placement.max_version = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.214116] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] placement.min_version = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.214272] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] placement.password = **** {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.214419] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] placement.project_domain_id = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.214571] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] placement.project_domain_name = Default {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.214726] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] placement.project_id = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.214886] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] placement.project_name = service {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.215050] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] placement.region_name = RegionOne {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.215215] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] placement.retriable_status_codes = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.215365] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] placement.service_name = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.215520] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] placement.service_type = placement {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.215670] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] placement.split_loggers = False {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.215812] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] placement.status_code_retries = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.215960] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] placement.status_code_retry_delay = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.216118] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] placement.system_scope = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.216265] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] placement.timeout = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.216412] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] placement.trust_id = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.216574] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] placement.user_domain_id = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.216708] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] placement.user_domain_name = Default {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.216853] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] placement.user_id = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.217019] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] placement.username = nova {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.217206] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] placement.valid_interfaces = ['internal', 'public'] {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.217376] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] placement.version = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.217577] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] quota.cores = 20 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.217735] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] quota.count_usage_from_placement = False {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.217897] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] quota.driver = nova.quota.DbQuotaDriver {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.218066] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] quota.injected_file_content_bytes = 10240 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.218229] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] quota.injected_file_path_length = 255 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.218385] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] quota.injected_files = 5 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.218538] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] quota.instances = 10 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.218689] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] quota.key_pairs = 100 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.218841] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] quota.metadata_items = 128 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.218993] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] quota.ram = 51200 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.219157] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] quota.recheck_quota = True {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.219312] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] quota.server_group_members = 10 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.219467] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] quota.server_groups = 10 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.219688] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] quota.unified_limits_resource_list = ['servers'] {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.219852] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] quota.unified_limits_resource_strategy = require {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.220285] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] scheduler.discover_hosts_in_cells_interval = -1 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.220285] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] scheduler.enable_isolated_aggregate_filtering = False {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.220370] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] scheduler.image_metadata_prefilter = False {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.220452] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] scheduler.limit_tenants_to_placement_aggregate = False {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.220606] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] scheduler.max_attempts = 3 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.220757] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] scheduler.max_placement_results = 1000 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.220905] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] scheduler.placement_aggregate_required_for_tenants = False {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.221065] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] scheduler.query_placement_for_image_type_support = False {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.221217] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] scheduler.query_placement_for_routed_network_aggregates = False {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.221379] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] scheduler.workers = 2 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.221536] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] filter_scheduler.aggregate_image_properties_isolation_namespace = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.221691] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] filter_scheduler.aggregate_image_properties_isolation_separator = . {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.221856] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] filter_scheduler.available_filters = ['nova.scheduler.filters.all_filters'] {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.222023] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] filter_scheduler.build_failure_weight_multiplier = 1000000.0 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.222179] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] filter_scheduler.cpu_weight_multiplier = 1.0 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.222333] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] filter_scheduler.cross_cell_move_weight_multiplier = 1000000.0 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.222482] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] filter_scheduler.disk_weight_multiplier = 1.0 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.222656] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] filter_scheduler.enabled_filters = ['ComputeFilter', 'ComputeCapabilitiesFilter', 'ImagePropertiesFilter', 'ServerGroupAntiAffinityFilter', 'ServerGroupAffinityFilter', 'SameHostFilter', 'DifferentHostFilter'] {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.222809] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] filter_scheduler.host_subset_size = 1 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.222960] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] filter_scheduler.hypervisor_version_weight_multiplier = 1.0 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.223121] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] filter_scheduler.image_properties_default_architecture = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.223272] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] filter_scheduler.image_props_weight_multiplier = 0.0 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.223424] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] filter_scheduler.image_props_weight_setting = [] {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.223575] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] filter_scheduler.io_ops_weight_multiplier = -1.0 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.223725] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] filter_scheduler.isolated_hosts = [] {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.223875] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] filter_scheduler.isolated_images = [] {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.224048] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] filter_scheduler.max_instances_per_host = 50 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.224201] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] filter_scheduler.max_io_ops_per_host = 8 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.224350] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] filter_scheduler.num_instances_weight_multiplier = 0.0 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.224499] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] filter_scheduler.pci_in_placement = False {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.224650] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] filter_scheduler.pci_weight_multiplier = 1.0 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.224797] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] filter_scheduler.ram_weight_multiplier = 1.0 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.224943] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] filter_scheduler.restrict_isolated_hosts_to_isolated_images = True {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.225103] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] filter_scheduler.shuffle_best_same_weighed_hosts = False {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.225256] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] filter_scheduler.soft_affinity_weight_multiplier = 1.0 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.225405] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] filter_scheduler.soft_anti_affinity_weight_multiplier = 1.0 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.225551] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] filter_scheduler.track_instance_changes = True {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.225711] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] filter_scheduler.weight_classes = ['nova.scheduler.weights.all_weighers'] {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.225865] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] metrics.required = True {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.226022] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] metrics.weight_multiplier = 1.0 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.226177] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] metrics.weight_of_unavailable = -10000.0 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.226330] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] metrics.weight_setting = [] {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.226629] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] serial_console.base_url = ws://127.0.0.1:6083/ {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.226789] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] serial_console.enabled = False {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.226950] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] serial_console.port_range = 10000:20000 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.227120] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] serial_console.proxyclient_address = 127.0.0.1 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.227302] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] serial_console.serialproxy_host = 0.0.0.0 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.227467] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] serial_console.serialproxy_port = 6083 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.227625] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] service_user.auth_section = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.227785] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] service_user.auth_type = password {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.227935] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] service_user.cafile = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.228094] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] service_user.certfile = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.228271] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] service_user.collect_timing = False {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.228432] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] service_user.insecure = False {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.228581] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] service_user.keyfile = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.228740] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] service_user.send_service_user_token = True {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.228892] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] service_user.split_loggers = False {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.229053] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] service_user.timeout = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.229218] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] spice.agent_enabled = True {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.229371] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] spice.enabled = False {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.229703] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] spice.html5proxy_base_url = http://127.0.0.1:6082/spice_auto.html {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.229905] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] spice.html5proxy_host = 0.0.0.0 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.230078] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] spice.html5proxy_port = 6082 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.230233] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] spice.image_compression = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.230383] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] spice.jpeg_compression = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.230528] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] spice.playback_compression = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.230676] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] spice.require_secure = False {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.230832] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] spice.server_listen = 127.0.0.1 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.230988] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] spice.server_proxyclient_address = 127.0.0.1 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.231279] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] spice.spice_direct_proxy_base_url = http://127.0.0.1:13002/nova {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.231440] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] spice.streaming_mode = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.231588] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] spice.zlib_compression = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.231744] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] upgrade_levels.baseapi = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.231901] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] upgrade_levels.compute = auto {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.232060] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] upgrade_levels.conductor = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.232288] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] upgrade_levels.scheduler = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.232531] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] vendordata_dynamic_auth.auth_section = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.232766] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] vendordata_dynamic_auth.auth_type = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.232960] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] vendordata_dynamic_auth.cafile = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.233128] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] vendordata_dynamic_auth.certfile = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.233282] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] vendordata_dynamic_auth.collect_timing = False {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.233432] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] vendordata_dynamic_auth.insecure = False {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.233578] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] vendordata_dynamic_auth.keyfile = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.233727] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] vendordata_dynamic_auth.split_loggers = False {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.233872] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] vendordata_dynamic_auth.timeout = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.234043] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] vmware.api_retry_count = 10 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.234199] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] vmware.ca_file = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.234359] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] vmware.cache_prefix = devstack-image-cache {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.234513] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] vmware.cluster_name = testcl1 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.234665] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] vmware.connection_pool_size = 10 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.234809] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] vmware.console_delay_seconds = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.234965] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] vmware.datastore_regex = ^datastore.* {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.235191] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] vmware.host_ip = vc1.osci.c.eu-de-1.cloud.sap {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.235356] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] vmware.host_password = **** {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.235511] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] vmware.host_port = 443 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.235666] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] vmware.host_username = administrator@vsphere.local {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.235820] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] vmware.insecure = True {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.235968] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] vmware.integration_bridge = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.236134] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] vmware.maximum_objects = 100 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.236284] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] vmware.pbm_default_policy = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.236435] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] vmware.pbm_enabled = False {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.236580] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] vmware.pbm_wsdl_location = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.236733] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] vmware.serial_log_dir = /opt/vmware/vspc {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.236877] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] vmware.serial_port_proxy_uri = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.237031] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] vmware.serial_port_service_uri = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.237213] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] vmware.task_poll_interval = 0.5 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.237396] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] vmware.use_linked_clone = False {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.237557] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] vmware.vnc_keymap = en-us {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.237738] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] vmware.vnc_port = 5900 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.237962] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] vmware.vnc_port_total = 10000 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.238220] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] vnc.auth_schemes = ['none'] {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.238409] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] vnc.enabled = False {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.238700] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] vnc.novncproxy_base_url = http://127.0.0.1:6080/vnc_auto.html {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.238874] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] vnc.novncproxy_host = 0.0.0.0 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.239047] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] vnc.novncproxy_port = 6080 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.239231] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] vnc.server_listen = 127.0.0.1 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.239403] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] vnc.server_proxyclient_address = 127.0.0.1 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.239576] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] vnc.vencrypt_ca_certs = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.239738] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] vnc.vencrypt_client_cert = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.239885] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] vnc.vencrypt_client_key = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.240056] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] workarounds.disable_compute_service_check_for_ffu = False {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.240209] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] workarounds.disable_deep_image_inspection = False {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.240359] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] workarounds.disable_fallback_pcpu_query = False {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.240506] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] workarounds.disable_group_policy_check_upcall = False {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.240651] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] workarounds.disable_libvirt_livesnapshot = False {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.240794] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] workarounds.disable_rootwrap = False {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.240938] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] workarounds.enable_numa_live_migration = False {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.241097] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] workarounds.enable_qemu_monitor_announce_self = False {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.241248] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] workarounds.ensure_libvirt_rbd_instance_dir_cleanup = False {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.241394] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] workarounds.handle_virt_lifecycle_events = True {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.241541] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] workarounds.libvirt_disable_apic = False {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.241686] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] workarounds.never_download_image_if_on_rbd = False {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.241832] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] workarounds.qemu_monitor_announce_self_count = 3 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.241977] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] workarounds.qemu_monitor_announce_self_interval = 1 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.242136] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] workarounds.reserve_disk_resource_for_image_cache = False {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.242283] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] workarounds.skip_cpu_compare_at_startup = False {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.242430] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] workarounds.skip_cpu_compare_on_dest = False {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.242574] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] workarounds.skip_hypervisor_version_check_on_lm = False {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.242718] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] workarounds.skip_reserve_in_use_ironic_nodes = False {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.242862] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] workarounds.unified_limits_count_pcpu_as_vcpu = False {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.243019] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] workarounds.wait_for_vif_plugged_event_during_hard_reboot = [] {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.243194] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] wsgi.api_paste_config = /etc/nova/api-paste.ini {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.243346] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] wsgi.secure_proxy_ssl_header = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.243497] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] zvm.ca_file = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.243687] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] zvm.cloud_connector_url = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.244008] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] zvm.image_tmp_path = /opt/stack/data/n-cpu-1/images {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.244194] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] zvm.reachable_timeout = 300 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.244359] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] oslo_versionedobjects.fatal_exception_format_errors = False {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.244523] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] oslo_middleware.http_basic_auth_user_file = /etc/htpasswd {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.244695] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] profiler.connection_string = messaging:// {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.244854] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] profiler.enabled = False {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.245036] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] profiler.es_doc_type = notification {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.245185] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] profiler.es_scroll_size = 10000 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.245339] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] profiler.es_scroll_time = 2m {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.245488] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] profiler.filter_error_trace = False {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.245641] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] profiler.hmac_keys = **** {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.245793] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] profiler.sentinel_service_name = mymaster {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.245948] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] profiler.socket_timeout = 0.1 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.246112] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] profiler.trace_requests = False {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.246259] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] profiler.trace_sqlalchemy = False {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.246426] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] profiler_jaeger.process_tags = {} {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.246573] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] profiler_jaeger.service_name_prefix = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.246724] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] profiler_otlp.service_name_prefix = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.246885] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] oslo_messaging_rabbit.amqp_auto_delete = False {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.247048] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] oslo_messaging_rabbit.amqp_durable_queues = False {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.247225] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] oslo_messaging_rabbit.conn_pool_min_size = 2 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.247392] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] oslo_messaging_rabbit.conn_pool_ttl = 1200 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.247546] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] oslo_messaging_rabbit.direct_mandatory_flag = True {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.247695] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] oslo_messaging_rabbit.enable_cancel_on_failover = False {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.247844] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] oslo_messaging_rabbit.heartbeat_in_pthread = False {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.247994] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] oslo_messaging_rabbit.heartbeat_rate = 3 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.248190] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] oslo_messaging_rabbit.heartbeat_timeout_threshold = 60 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.248343] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] oslo_messaging_rabbit.hostname = devstack {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.248495] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] oslo_messaging_rabbit.kombu_compression = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.248654] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] oslo_messaging_rabbit.kombu_failover_strategy = round-robin {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.248808] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] oslo_messaging_rabbit.kombu_missing_consumer_retry_timeout = 60 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.248960] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] oslo_messaging_rabbit.kombu_reconnect_delay = 1.0 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.249125] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] oslo_messaging_rabbit.kombu_reconnect_splay = 0.0 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.249282] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] oslo_messaging_rabbit.processname = nova-compute {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.249433] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] oslo_messaging_rabbit.rabbit_ha_queues = False {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.249593] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] oslo_messaging_rabbit.rabbit_interval_max = 30 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.249749] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] oslo_messaging_rabbit.rabbit_login_method = AMQPLAIN {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.249900] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] oslo_messaging_rabbit.rabbit_qos_prefetch_count = 0 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.250067] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] oslo_messaging_rabbit.rabbit_quorum_delivery_limit = 0 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.250222] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] oslo_messaging_rabbit.rabbit_quorum_max_memory_bytes = 0 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.250376] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] oslo_messaging_rabbit.rabbit_quorum_max_memory_length = 0 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.250525] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] oslo_messaging_rabbit.rabbit_quorum_queue = False {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.250675] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] oslo_messaging_rabbit.rabbit_retry_backoff = 2 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.250832] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] oslo_messaging_rabbit.rabbit_retry_interval = 1 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.250973] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] oslo_messaging_rabbit.rabbit_stream_fanout = False {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.251148] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] oslo_messaging_rabbit.rabbit_transient_queues_ttl = 1800 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.251300] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] oslo_messaging_rabbit.rabbit_transient_quorum_queue = False {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.251450] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] oslo_messaging_rabbit.rpc_conn_pool_size = 30 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.251599] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] oslo_messaging_rabbit.ssl = False {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.251755] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] oslo_messaging_rabbit.ssl_ca_file = {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.251911] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] oslo_messaging_rabbit.ssl_cert_file = {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.252073] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] oslo_messaging_rabbit.ssl_enforce_fips_mode = False {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.252232] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] oslo_messaging_rabbit.ssl_key_file = {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.252388] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] oslo_messaging_rabbit.ssl_version = {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.252538] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] oslo_messaging_rabbit.use_queue_manager = False {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.252709] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] oslo_messaging_notifications.driver = ['messagingv2'] {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.252863] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] oslo_messaging_notifications.retry = -1 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.253034] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] oslo_messaging_notifications.topics = ['notifications'] {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.253198] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] oslo_messaging_notifications.transport_url = **** {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.253357] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] oslo_limit.auth_section = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.253506] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] oslo_limit.auth_type = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.253650] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] oslo_limit.cafile = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.253793] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] oslo_limit.certfile = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.253940] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] oslo_limit.collect_timing = False {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.254096] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] oslo_limit.connect_retries = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.254252] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] oslo_limit.connect_retry_delay = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.254402] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] oslo_limit.endpoint_id = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.254557] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] oslo_limit.endpoint_interface = publicURL {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.254700] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] oslo_limit.endpoint_override = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.254841] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] oslo_limit.endpoint_region_name = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.254984] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] oslo_limit.endpoint_service_name = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.255143] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] oslo_limit.endpoint_service_type = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.255291] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] oslo_limit.insecure = False {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.255437] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] oslo_limit.keyfile = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.255580] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] oslo_limit.max_version = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.255720] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] oslo_limit.min_version = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.255861] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] oslo_limit.region_name = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.256011] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] oslo_limit.retriable_status_codes = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.256162] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] oslo_limit.service_name = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.256306] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] oslo_limit.service_type = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.256453] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] oslo_limit.split_loggers = False {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.256595] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] oslo_limit.status_code_retries = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.256740] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] oslo_limit.status_code_retry_delay = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.256882] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] oslo_limit.timeout = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.257033] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] oslo_limit.valid_interfaces = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.257204] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] oslo_limit.version = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.257360] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] oslo_reports.file_event_handler = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.257515] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] oslo_reports.file_event_handler_interval = 1 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.257661] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] oslo_reports.log_dir = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.257829] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] vif_plug_linux_bridge_privileged.capabilities = [12] {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.257976] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] vif_plug_linux_bridge_privileged.group = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.258137] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] vif_plug_linux_bridge_privileged.helper_command = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.258303] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] vif_plug_linux_bridge_privileged.log_daemon_traceback = False {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.258459] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] vif_plug_linux_bridge_privileged.logger_name = oslo_privsep.daemon {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.258612] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] vif_plug_linux_bridge_privileged.thread_pool_size = 8 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.258760] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] vif_plug_linux_bridge_privileged.user = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.258917] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] vif_plug_ovs_privileged.capabilities = [12, 1] {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.259079] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] vif_plug_ovs_privileged.group = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.259229] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] vif_plug_ovs_privileged.helper_command = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.259383] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] vif_plug_ovs_privileged.log_daemon_traceback = False {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.259550] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] vif_plug_ovs_privileged.logger_name = oslo_privsep.daemon {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.259713] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] vif_plug_ovs_privileged.thread_pool_size = 8 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.259858] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] vif_plug_ovs_privileged.user = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.260025] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] os_vif_linux_bridge.flat_interface = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.260199] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] os_vif_linux_bridge.forward_bridge_interface = ['all'] {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.260360] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] os_vif_linux_bridge.iptables_bottom_regex = {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.260516] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] os_vif_linux_bridge.iptables_drop_action = DROP {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.260669] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] os_vif_linux_bridge.iptables_top_regex = {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.260829] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] os_vif_linux_bridge.network_device_mtu = 1500 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.260981] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] os_vif_linux_bridge.use_ipv6 = False {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.261140] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] os_vif_linux_bridge.vlan_interface = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.261306] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] os_vif_ovs.default_qos_type = linux-noop {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.261464] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] os_vif_ovs.isolate_vif = False {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.261619] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] os_vif_ovs.network_device_mtu = 1500 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.261768] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] os_vif_ovs.ovs_vsctl_timeout = 120 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.261923] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] os_vif_ovs.ovsdb_connection = tcp:127.0.0.1:6640 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.262093] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] os_vif_ovs.ovsdb_interface = native {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.262246] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] os_vif_ovs.per_port_bridge = False {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.262406] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] privsep_osbrick.capabilities = [21, 2] {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.262554] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] privsep_osbrick.group = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.262697] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] privsep_osbrick.helper_command = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.262844] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] privsep_osbrick.log_daemon_traceback = False {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.262994] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] privsep_osbrick.logger_name = os_brick.privileged {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.263156] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] privsep_osbrick.thread_pool_size = 8 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.263302] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] privsep_osbrick.user = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.263463] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] nova_sys_admin.capabilities = [0, 1, 2, 3, 12, 21] {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.263609] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] nova_sys_admin.group = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.263753] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] nova_sys_admin.helper_command = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.263899] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] nova_sys_admin.log_daemon_traceback = False {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.264061] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] nova_sys_admin.logger_name = oslo_privsep.daemon {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.264214] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] nova_sys_admin.thread_pool_size = 8 {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.264359] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] nova_sys_admin.user = None {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 541.264472] env[65726]: DEBUG oslo_service.backend._eventlet.service [None req-4bb946d6-92ab-4ca4-aa9c-6350b5329f95 None None] ******************************************************************************** {{(pid=65726) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2828}} [ 541.264901] env[65726]: INFO nova.service [-] Starting compute node (version 0.0.1) [ 541.768888] env[65726]: WARNING nova.virt.vmwareapi.driver [None req-04e75d53-dab9-4278-afda-f5e29352c96d None None] The vmwareapi driver is not tested by the OpenStack project nor does it have clear maintainer(s) and thus its quality can not be ensured. It should be considered experimental and may be removed in a future release. If you are using the driver in production please let us know via the openstack-discuss mailing list. [ 541.769705] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-04e75d53-dab9-4278-afda-f5e29352c96d None None] Getting list of instances from cluster (obj){ [ 541.769705] env[65726]: value = "domain-c8" [ 541.769705] env[65726]: _type = "ClusterComputeResource" [ 541.769705] env[65726]: } {{(pid=65726) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 541.770846] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d7f3df1-f17d-4c1d-be4a-0851cb460032 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.780135] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-04e75d53-dab9-4278-afda-f5e29352c96d None None] Got total of 0 instances {{(pid=65726) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 541.780621] env[65726]: INFO nova.virt.node [None req-04e75d53-dab9-4278-afda-f5e29352c96d None None] Generated node identity 07c4692f-bdb4-4058-9173-ff9664830295 [ 541.780862] env[65726]: INFO nova.virt.node [None req-04e75d53-dab9-4278-afda-f5e29352c96d None None] Wrote node identity 07c4692f-bdb4-4058-9173-ff9664830295 to /opt/stack/data/n-cpu-1/compute_id [ 542.284073] env[65726]: WARNING nova.compute.manager [None req-04e75d53-dab9-4278-afda-f5e29352c96d None None] Compute nodes ['07c4692f-bdb4-4058-9173-ff9664830295'] for host cpu-1 were not found in the database. If this is the first time this service is starting on this host, then you can ignore this warning. [ 543.289350] env[65726]: INFO nova.compute.manager [None req-04e75d53-dab9-4278-afda-f5e29352c96d None None] Looking for unclaimed instances stuck in BUILDING status for nodes managed by this host [ 544.294162] env[65726]: WARNING nova.compute.manager [None req-04e75d53-dab9-4278-afda-f5e29352c96d None None] No compute node record found for host cpu-1. If this is the first time this service is starting on this host, then you can ignore this warning.: nova.exception_Remote.ComputeHostNotFound_Remote: Compute host cpu-1 could not be found. [ 544.294582] env[65726]: DEBUG oslo_concurrency.lockutils [None req-04e75d53-dab9-4278-afda-f5e29352c96d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 544.294582] env[65726]: DEBUG oslo_concurrency.lockutils [None req-04e75d53-dab9-4278-afda-f5e29352c96d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 544.294736] env[65726]: DEBUG oslo_concurrency.lockutils [None req-04e75d53-dab9-4278-afda-f5e29352c96d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 544.294834] env[65726]: DEBUG nova.compute.resource_tracker [None req-04e75d53-dab9-4278-afda-f5e29352c96d None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=65726) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:937}} [ 544.295781] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ceb2a2f5-43f2-47eb-8461-c74d891f7915 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.304148] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2da0efc6-ac8b-45ea-bafa-cf5d6605bf9d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.318300] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17684247-8c62-4463-87eb-fe00b9da451e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.325233] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6de4577f-1261-4ecb-a52b-e4f209cec8dc {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.354077] env[65726]: DEBUG nova.compute.resource_tracker [None req-04e75d53-dab9-4278-afda-f5e29352c96d None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180514MB free_disk=96GB free_vcpus=48 pci_devices=None {{(pid=65726) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1136}} [ 544.354221] env[65726]: DEBUG oslo_concurrency.lockutils [None req-04e75d53-dab9-4278-afda-f5e29352c96d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 544.354425] env[65726]: DEBUG oslo_concurrency.lockutils [None req-04e75d53-dab9-4278-afda-f5e29352c96d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 544.857252] env[65726]: WARNING nova.compute.resource_tracker [None req-04e75d53-dab9-4278-afda-f5e29352c96d None None] No compute node record for cpu-1:07c4692f-bdb4-4058-9173-ff9664830295: nova.exception_Remote.ComputeHostNotFound_Remote: Compute host 07c4692f-bdb4-4058-9173-ff9664830295 could not be found. [ 545.361888] env[65726]: INFO nova.compute.resource_tracker [None req-04e75d53-dab9-4278-afda-f5e29352c96d None None] Compute node record created for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 with uuid: 07c4692f-bdb4-4058-9173-ff9664830295 [ 546.869147] env[65726]: DEBUG nova.compute.resource_tracker [None req-04e75d53-dab9-4278-afda-f5e29352c96d None None] Total usable vcpus: 48, total allocated vcpus: 0 {{(pid=65726) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1159}} [ 546.869609] env[65726]: DEBUG nova.compute.resource_tracker [None req-04e75d53-dab9-4278-afda-f5e29352c96d None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=512MB phys_disk=100GB used_disk=0GB total_vcpus=48 used_vcpus=0 pci_stats=[] stats={'failed_builds': '0'} {{(pid=65726) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1168}} [ 547.053221] env[65726]: INFO nova.scheduler.client.report [None req-04e75d53-dab9-4278-afda-f5e29352c96d None None] [req-793930ce-9d35-4153-b276-00a0787250a6] Created resource provider record via placement API for resource provider with UUID 07c4692f-bdb4-4058-9173-ff9664830295 and name domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28. [ 547.077878] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1f540af-8a49-4e34-9010-7285b0648172 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 547.085939] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-744f3dff-acda-4ffc-ae66-5e79bd543d57 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 547.115917] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b6f511a-3c05-432e-8e21-a1d732d7ed66 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 547.124069] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-036a7432-58da-44ec-aecb-8019c9c2953b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 547.137891] env[65726]: DEBUG nova.compute.provider_tree [None req-04e75d53-dab9-4278-afda-f5e29352c96d None None] Updating inventory in ProviderTree for provider 07c4692f-bdb4-4058-9173-ff9664830295 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 547.678741] env[65726]: DEBUG nova.scheduler.client.report [None req-04e75d53-dab9-4278-afda-f5e29352c96d None None] Updated inventory for provider 07c4692f-bdb4-4058-9173-ff9664830295 with generation 0 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 547.678961] env[65726]: DEBUG nova.compute.provider_tree [None req-04e75d53-dab9-4278-afda-f5e29352c96d None None] Updating resource provider 07c4692f-bdb4-4058-9173-ff9664830295 generation from 0 to 1 during operation: update_inventory {{(pid=65726) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 547.679110] env[65726]: DEBUG nova.compute.provider_tree [None req-04e75d53-dab9-4278-afda-f5e29352c96d None None] Updating inventory in ProviderTree for provider 07c4692f-bdb4-4058-9173-ff9664830295 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 547.731175] env[65726]: DEBUG nova.compute.provider_tree [None req-04e75d53-dab9-4278-afda-f5e29352c96d None None] Updating resource provider 07c4692f-bdb4-4058-9173-ff9664830295 generation from 1 to 2 during operation: update_traits {{(pid=65726) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 548.236160] env[65726]: DEBUG nova.compute.resource_tracker [None req-04e75d53-dab9-4278-afda-f5e29352c96d None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=65726) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1097}} [ 548.236610] env[65726]: DEBUG oslo_concurrency.lockutils [None req-04e75d53-dab9-4278-afda-f5e29352c96d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 3.882s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 548.236610] env[65726]: DEBUG nova.service [None req-04e75d53-dab9-4278-afda-f5e29352c96d None None] Creating RPC server for service compute {{(pid=65726) start /opt/stack/nova/nova/service.py:177}} [ 548.250944] env[65726]: DEBUG nova.service [None req-04e75d53-dab9-4278-afda-f5e29352c96d None None] Join ServiceGroup membership for this service compute {{(pid=65726) start /opt/stack/nova/nova/service.py:194}} [ 548.251147] env[65726]: DEBUG nova.servicegroup.drivers.db [None req-04e75d53-dab9-4278-afda-f5e29352c96d None None] DB_Driver: join new ServiceGroup member cpu-1 to the compute group, service = {{(pid=65726) join /opt/stack/nova/nova/servicegroup/drivers/db.py:44}} [ 592.317558] env[65726]: INFO nova.utils [None req-9ed3c3d0-6db2-48b9-8065-95d9e297f871 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] The default thread pool MainProcess.default is initialized [ 592.318719] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9ed3c3d0-6db2-48b9-8065-95d9e297f871 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Acquiring lock "0e064341-4e4a-407b-8c26-3eb04b409029" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 592.318719] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9ed3c3d0-6db2-48b9-8065-95d9e297f871 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Lock "0e064341-4e4a-407b-8c26-3eb04b409029" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 592.355906] env[65726]: DEBUG oslo_concurrency.lockutils [None req-816382da-6d37-4962-9c02-92a18cd283e5 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Acquiring lock "bb64fc41-19d7-49dc-a17a-9dd74730130f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 592.356852] env[65726]: DEBUG oslo_concurrency.lockutils [None req-816382da-6d37-4962-9c02-92a18cd283e5 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Lock "bb64fc41-19d7-49dc-a17a-9dd74730130f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 592.822449] env[65726]: DEBUG nova.compute.manager [None req-9ed3c3d0-6db2-48b9-8065-95d9e297f871 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] [instance: 0e064341-4e4a-407b-8c26-3eb04b409029] Starting instance... {{(pid=65726) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 592.859794] env[65726]: DEBUG nova.compute.manager [None req-816382da-6d37-4962-9c02-92a18cd283e5 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] [instance: bb64fc41-19d7-49dc-a17a-9dd74730130f] Starting instance... {{(pid=65726) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 593.371925] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9ed3c3d0-6db2-48b9-8065-95d9e297f871 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 593.372262] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9ed3c3d0-6db2-48b9-8065-95d9e297f871 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 593.375280] env[65726]: INFO nova.compute.claims [None req-9ed3c3d0-6db2-48b9-8065-95d9e297f871 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] [instance: 0e064341-4e4a-407b-8c26-3eb04b409029] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 593.386802] env[65726]: DEBUG oslo_concurrency.lockutils [None req-816382da-6d37-4962-9c02-92a18cd283e5 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 594.472291] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85453d0b-ae07-4683-ac4e-375a58ec1f51 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.488257] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c3051dd-3112-41be-b61b-ed3eec4af0ac {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.540025] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87d5406a-5ac7-4270-b2f1-b79910e5adf6 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.547449] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efad86e5-e798-4db9-8850-16ce49dfdbc1 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.564538] env[65726]: DEBUG nova.compute.provider_tree [None req-9ed3c3d0-6db2-48b9-8065-95d9e297f871 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 595.073316] env[65726]: DEBUG nova.scheduler.client.report [None req-9ed3c3d0-6db2-48b9-8065-95d9e297f871 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 595.253039] env[65726]: DEBUG oslo_service.periodic_task [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Running periodic task ComputeManager._sync_power_states {{(pid=65726) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 595.371111] env[65726]: DEBUG oslo_concurrency.lockutils [None req-af5fca6d-c57c-4762-93f4-4ba84e22ca7b tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Acquiring lock "8e9c27f6-e9fa-4ed0-b6ae-3b5df52d1c44" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 595.372032] env[65726]: DEBUG oslo_concurrency.lockutils [None req-af5fca6d-c57c-4762-93f4-4ba84e22ca7b tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Lock "8e9c27f6-e9fa-4ed0-b6ae-3b5df52d1c44" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 595.580105] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9ed3c3d0-6db2-48b9-8065-95d9e297f871 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.207s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 595.580387] env[65726]: DEBUG nova.compute.manager [None req-9ed3c3d0-6db2-48b9-8065-95d9e297f871 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] [instance: 0e064341-4e4a-407b-8c26-3eb04b409029] Start building networks asynchronously for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 595.584168] env[65726]: DEBUG oslo_concurrency.lockutils [None req-816382da-6d37-4962-9c02-92a18cd283e5 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.197s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 595.585220] env[65726]: INFO nova.compute.claims [None req-816382da-6d37-4962-9c02-92a18cd283e5 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] [instance: bb64fc41-19d7-49dc-a17a-9dd74730130f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 595.760272] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Getting list of instances from cluster (obj){ [ 595.760272] env[65726]: value = "domain-c8" [ 595.760272] env[65726]: _type = "ClusterComputeResource" [ 595.760272] env[65726]: } {{(pid=65726) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 595.762219] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8f0ce31-d4f4-466a-a740-69cf5eba55da {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.779517] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Got total of 0 instances {{(pid=65726) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 595.780552] env[65726]: WARNING nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] While synchronizing instance power states, found 1 instances in the database and 0 instances on the hypervisor. [ 595.780652] env[65726]: DEBUG nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Triggering sync for uuid 0e064341-4e4a-407b-8c26-3eb04b409029 {{(pid=65726) _sync_power_states /opt/stack/nova/nova/compute/manager.py:11043}} [ 595.781254] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Acquiring lock "0e064341-4e4a-407b-8c26-3eb04b409029" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 595.781460] env[65726]: DEBUG oslo_service.periodic_task [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Running periodic task ComputeManager._cleanup_running_deleted_instances {{(pid=65726) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 595.781857] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Getting list of instances from cluster (obj){ [ 595.781857] env[65726]: value = "domain-c8" [ 595.781857] env[65726]: _type = "ClusterComputeResource" [ 595.781857] env[65726]: } {{(pid=65726) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 595.783064] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d4717b4-4b7a-43d1-9e80-c1fbd86b48e2 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.795451] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Got total of 0 instances {{(pid=65726) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 595.875440] env[65726]: DEBUG nova.compute.manager [None req-af5fca6d-c57c-4762-93f4-4ba84e22ca7b tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 8e9c27f6-e9fa-4ed0-b6ae-3b5df52d1c44] Starting instance... {{(pid=65726) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 596.085844] env[65726]: DEBUG nova.compute.utils [None req-9ed3c3d0-6db2-48b9-8065-95d9e297f871 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Using /dev/sd instead of None {{(pid=65726) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 596.089169] env[65726]: DEBUG nova.compute.manager [None req-9ed3c3d0-6db2-48b9-8065-95d9e297f871 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] [instance: 0e064341-4e4a-407b-8c26-3eb04b409029] Allocating IP information in the background. {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 596.089169] env[65726]: DEBUG nova.network.neutron [None req-9ed3c3d0-6db2-48b9-8065-95d9e297f871 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] [instance: 0e064341-4e4a-407b-8c26-3eb04b409029] allocate_for_instance() {{(pid=65726) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 596.089866] env[65726]: WARNING neutronclient.v2_0.client [None req-9ed3c3d0-6db2-48b9-8065-95d9e297f871 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 596.091666] env[65726]: WARNING neutronclient.v2_0.client [None req-9ed3c3d0-6db2-48b9-8065-95d9e297f871 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 596.094506] env[65726]: WARNING openstack [None req-9ed3c3d0-6db2-48b9-8065-95d9e297f871 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 596.094506] env[65726]: WARNING openstack [None req-9ed3c3d0-6db2-48b9-8065-95d9e297f871 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 596.415857] env[65726]: DEBUG oslo_concurrency.lockutils [None req-af5fca6d-c57c-4762-93f4-4ba84e22ca7b tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 596.584326] env[65726]: DEBUG oslo_service.periodic_task [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=65726) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 596.588121] env[65726]: DEBUG oslo_service.periodic_task [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=65726) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 596.588121] env[65726]: DEBUG oslo_service.periodic_task [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=65726) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 596.588121] env[65726]: DEBUG oslo_service.periodic_task [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=65726) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 596.588121] env[65726]: DEBUG oslo_service.periodic_task [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=65726) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 596.588121] env[65726]: DEBUG oslo_service.periodic_task [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=65726) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 596.588121] env[65726]: DEBUG oslo_service.periodic_task [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=65726) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 596.588323] env[65726]: DEBUG nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=65726) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11251}} [ 596.588323] env[65726]: DEBUG oslo_service.periodic_task [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Running periodic task ComputeManager.update_available_resource {{(pid=65726) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 596.606734] env[65726]: DEBUG nova.compute.manager [None req-9ed3c3d0-6db2-48b9-8065-95d9e297f871 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] [instance: 0e064341-4e4a-407b-8c26-3eb04b409029] Start building block device mappings for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 596.695270] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b828369-4742-4ded-8852-365db3083138 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.704183] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acf34800-c155-4097-a58f-312d9aa412b4 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.753773] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ca651f7-1208-4670-80e6-bb7e1f38a78c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.763057] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-850a22ce-484b-449d-8397-2c97df748faf {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.780923] env[65726]: DEBUG nova.compute.provider_tree [None req-816382da-6d37-4962-9c02-92a18cd283e5 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 597.091582] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 597.289404] env[65726]: DEBUG nova.scheduler.client.report [None req-816382da-6d37-4962-9c02-92a18cd283e5 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 597.318321] env[65726]: DEBUG nova.policy [None req-9ed3c3d0-6db2-48b9-8065-95d9e297f871 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '58e5e1469d2241d3ad3178f085a7bee6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'fae9ef0198a2468a9793374b76378abf', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65726) authorize /opt/stack/nova/nova/policy.py:192}} [ 597.326650] env[65726]: DEBUG oslo_concurrency.lockutils [None req-816382da-6d37-4962-9c02-92a18cd283e5 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 1.743s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 597.327775] env[65726]: DEBUG nova.compute.manager [None req-816382da-6d37-4962-9c02-92a18cd283e5 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] [instance: bb64fc41-19d7-49dc-a17a-9dd74730130f] Start building networks asynchronously for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 597.333256] env[65726]: DEBUG oslo_concurrency.lockutils [None req-af5fca6d-c57c-4762-93f4-4ba84e22ca7b tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.917s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 597.337019] env[65726]: INFO nova.compute.claims [None req-af5fca6d-c57c-4762-93f4-4ba84e22ca7b tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 8e9c27f6-e9fa-4ed0-b6ae-3b5df52d1c44] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 597.621676] env[65726]: DEBUG nova.compute.manager [None req-9ed3c3d0-6db2-48b9-8065-95d9e297f871 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] [instance: 0e064341-4e4a-407b-8c26-3eb04b409029] Start spawning the instance on the hypervisor. {{(pid=65726) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 597.680242] env[65726]: DEBUG nova.virt.hardware [None req-9ed3c3d0-6db2-48b9-8065-95d9e297f871 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T19:28:34Z,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1539f06fa3534e90acbc3a60c4b8bd3f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T19:28:35Z,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 597.680606] env[65726]: DEBUG nova.virt.hardware [None req-9ed3c3d0-6db2-48b9-8065-95d9e297f871 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 597.680816] env[65726]: DEBUG nova.virt.hardware [None req-9ed3c3d0-6db2-48b9-8065-95d9e297f871 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 597.681523] env[65726]: DEBUG nova.virt.hardware [None req-9ed3c3d0-6db2-48b9-8065-95d9e297f871 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 597.681523] env[65726]: DEBUG nova.virt.hardware [None req-9ed3c3d0-6db2-48b9-8065-95d9e297f871 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 597.681903] env[65726]: DEBUG nova.virt.hardware [None req-9ed3c3d0-6db2-48b9-8065-95d9e297f871 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 597.681903] env[65726]: DEBUG nova.virt.hardware [None req-9ed3c3d0-6db2-48b9-8065-95d9e297f871 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 597.682025] env[65726]: DEBUG nova.virt.hardware [None req-9ed3c3d0-6db2-48b9-8065-95d9e297f871 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 597.682785] env[65726]: DEBUG nova.virt.hardware [None req-9ed3c3d0-6db2-48b9-8065-95d9e297f871 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 597.682858] env[65726]: DEBUG nova.virt.hardware [None req-9ed3c3d0-6db2-48b9-8065-95d9e297f871 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 597.683033] env[65726]: DEBUG nova.virt.hardware [None req-9ed3c3d0-6db2-48b9-8065-95d9e297f871 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 597.684222] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb1e59ab-6657-49fc-a126-a47be7661ab4 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.693325] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44949cb5-731a-4cec-a870-5a7ac84ced2b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.719455] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb2bcb43-192b-4831-9b9a-d81e30463e80 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.834883] env[65726]: DEBUG nova.compute.utils [None req-816382da-6d37-4962-9c02-92a18cd283e5 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Using /dev/sd instead of None {{(pid=65726) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 597.836493] env[65726]: DEBUG nova.compute.manager [None req-816382da-6d37-4962-9c02-92a18cd283e5 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] [instance: bb64fc41-19d7-49dc-a17a-9dd74730130f] Allocating IP information in the background. {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 597.837093] env[65726]: DEBUG nova.network.neutron [None req-816382da-6d37-4962-9c02-92a18cd283e5 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] [instance: bb64fc41-19d7-49dc-a17a-9dd74730130f] allocate_for_instance() {{(pid=65726) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 597.840860] env[65726]: WARNING neutronclient.v2_0.client [None req-816382da-6d37-4962-9c02-92a18cd283e5 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 597.840860] env[65726]: WARNING neutronclient.v2_0.client [None req-816382da-6d37-4962-9c02-92a18cd283e5 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 597.841558] env[65726]: WARNING openstack [None req-816382da-6d37-4962-9c02-92a18cd283e5 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 597.842152] env[65726]: WARNING openstack [None req-816382da-6d37-4962-9c02-92a18cd283e5 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 598.008516] env[65726]: WARNING openstack [None req-9ed3c3d0-6db2-48b9-8065-95d9e297f871 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 598.009197] env[65726]: WARNING openstack [None req-9ed3c3d0-6db2-48b9-8065-95d9e297f871 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 598.138973] env[65726]: DEBUG nova.network.neutron [None req-9ed3c3d0-6db2-48b9-8065-95d9e297f871 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] [instance: 0e064341-4e4a-407b-8c26-3eb04b409029] Successfully created port: 9d3c5bc9-aa25-4013-b8dc-f565887d2736 {{(pid=65726) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 598.146358] env[65726]: DEBUG nova.policy [None req-816382da-6d37-4962-9c02-92a18cd283e5 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6ce8616097d940ffa52ef44c14b84622', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '06ad412f524c40538c835d9d664cd85e', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65726) authorize /opt/stack/nova/nova/policy.py:192}} [ 598.232284] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5c1fab1f-7856-43d0-84bd-904452799dc9 tempest-ServersTestJSON-1793637688 tempest-ServersTestJSON-1793637688-project-member] Acquiring lock "88fb7235-cd0e-49dc-9d78-3583d5d3f528" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 598.232518] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5c1fab1f-7856-43d0-84bd-904452799dc9 tempest-ServersTestJSON-1793637688 tempest-ServersTestJSON-1793637688-project-member] Lock "88fb7235-cd0e-49dc-9d78-3583d5d3f528" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 598.354824] env[65726]: DEBUG nova.compute.manager [None req-816382da-6d37-4962-9c02-92a18cd283e5 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] [instance: bb64fc41-19d7-49dc-a17a-9dd74730130f] Start building block device mappings for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 598.451162] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ff529aa-092a-48a6-baec-f7e1d2669ff9 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.460253] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb58e7d1-0030-4ecd-ac59-04e3e9029935 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.495859] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a4dd559-8d33-4968-a935-8a377ef5c1f3 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.506071] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21932f55-2dca-4420-af82-7ffabcbeed3e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.522072] env[65726]: DEBUG nova.compute.provider_tree [None req-af5fca6d-c57c-4762-93f4-4ba84e22ca7b tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 598.645257] env[65726]: DEBUG nova.network.neutron [None req-816382da-6d37-4962-9c02-92a18cd283e5 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] [instance: bb64fc41-19d7-49dc-a17a-9dd74730130f] Successfully created port: b6351d8e-b77c-4b29-b18b-02fc5e9e0c72 {{(pid=65726) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 598.736840] env[65726]: DEBUG nova.compute.manager [None req-5c1fab1f-7856-43d0-84bd-904452799dc9 tempest-ServersTestJSON-1793637688 tempest-ServersTestJSON-1793637688-project-member] [instance: 88fb7235-cd0e-49dc-9d78-3583d5d3f528] Starting instance... {{(pid=65726) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 599.026938] env[65726]: DEBUG nova.scheduler.client.report [None req-af5fca6d-c57c-4762-93f4-4ba84e22ca7b tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 599.268721] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5c1fab1f-7856-43d0-84bd-904452799dc9 tempest-ServersTestJSON-1793637688 tempest-ServersTestJSON-1793637688-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 599.369714] env[65726]: DEBUG nova.compute.manager [None req-816382da-6d37-4962-9c02-92a18cd283e5 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] [instance: bb64fc41-19d7-49dc-a17a-9dd74730130f] Start spawning the instance on the hypervisor. {{(pid=65726) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 599.408040] env[65726]: DEBUG nova.virt.hardware [None req-816382da-6d37-4962-9c02-92a18cd283e5 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T19:28:34Z,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1539f06fa3534e90acbc3a60c4b8bd3f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T19:28:35Z,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 599.408040] env[65726]: DEBUG nova.virt.hardware [None req-816382da-6d37-4962-9c02-92a18cd283e5 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 599.408230] env[65726]: DEBUG nova.virt.hardware [None req-816382da-6d37-4962-9c02-92a18cd283e5 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 599.408352] env[65726]: DEBUG nova.virt.hardware [None req-816382da-6d37-4962-9c02-92a18cd283e5 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 599.408481] env[65726]: DEBUG nova.virt.hardware [None req-816382da-6d37-4962-9c02-92a18cd283e5 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 599.408614] env[65726]: DEBUG nova.virt.hardware [None req-816382da-6d37-4962-9c02-92a18cd283e5 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 599.408814] env[65726]: DEBUG nova.virt.hardware [None req-816382da-6d37-4962-9c02-92a18cd283e5 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 599.409111] env[65726]: DEBUG nova.virt.hardware [None req-816382da-6d37-4962-9c02-92a18cd283e5 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 599.410225] env[65726]: DEBUG nova.virt.hardware [None req-816382da-6d37-4962-9c02-92a18cd283e5 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 599.410822] env[65726]: DEBUG nova.virt.hardware [None req-816382da-6d37-4962-9c02-92a18cd283e5 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 599.410822] env[65726]: DEBUG nova.virt.hardware [None req-816382da-6d37-4962-9c02-92a18cd283e5 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 599.412425] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e029acc8-5b65-48b7-b56b-d3e35e9695be {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.424615] env[65726]: DEBUG oslo_concurrency.lockutils [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Acquiring lock "703fff2d-5aec-4ed1-b987-9e1cc3a25f67" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 599.424868] env[65726]: DEBUG oslo_concurrency.lockutils [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Lock "703fff2d-5aec-4ed1-b987-9e1cc3a25f67" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 599.430269] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-102b390e-0df2-4e13-827f-783ac1baca67 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.536300] env[65726]: DEBUG oslo_concurrency.lockutils [None req-af5fca6d-c57c-4762-93f4-4ba84e22ca7b tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.203s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 599.536680] env[65726]: DEBUG nova.compute.manager [None req-af5fca6d-c57c-4762-93f4-4ba84e22ca7b tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 8e9c27f6-e9fa-4ed0-b6ae-3b5df52d1c44] Start building networks asynchronously for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 599.539817] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 2.448s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 599.539817] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 599.540953] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=65726) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:937}} [ 599.540953] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5c1fab1f-7856-43d0-84bd-904452799dc9 tempest-ServersTestJSON-1793637688 tempest-ServersTestJSON-1793637688-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.272s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 599.542078] env[65726]: INFO nova.compute.claims [None req-5c1fab1f-7856-43d0-84bd-904452799dc9 tempest-ServersTestJSON-1793637688 tempest-ServersTestJSON-1793637688-project-member] [instance: 88fb7235-cd0e-49dc-9d78-3583d5d3f528] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 599.546190] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa4350ab-0efb-4c8d-a9f3-d31e58e83dba {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.557930] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b749db5-d74d-495b-b939-88a9722027b3 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.573021] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20109a28-5054-4227-99ba-b9da07f85790 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.582012] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0a7a528-8d02-443a-9c76-e0ee22ffdf80 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.616052] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180517MB free_disk=96GB free_vcpus=48 pci_devices=None {{(pid=65726) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1136}} [ 599.616153] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 599.930356] env[65726]: DEBUG nova.compute.manager [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] [instance: 703fff2d-5aec-4ed1-b987-9e1cc3a25f67] Starting instance... {{(pid=65726) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 599.958858] env[65726]: DEBUG nova.network.neutron [None req-9ed3c3d0-6db2-48b9-8065-95d9e297f871 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] [instance: 0e064341-4e4a-407b-8c26-3eb04b409029] Successfully updated port: 9d3c5bc9-aa25-4013-b8dc-f565887d2736 {{(pid=65726) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 600.047690] env[65726]: DEBUG nova.compute.utils [None req-af5fca6d-c57c-4762-93f4-4ba84e22ca7b tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Using /dev/sd instead of None {{(pid=65726) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 600.052194] env[65726]: DEBUG nova.compute.manager [None req-af5fca6d-c57c-4762-93f4-4ba84e22ca7b tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 8e9c27f6-e9fa-4ed0-b6ae-3b5df52d1c44] Allocating IP information in the background. {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 600.053819] env[65726]: DEBUG nova.network.neutron [None req-af5fca6d-c57c-4762-93f4-4ba84e22ca7b tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 8e9c27f6-e9fa-4ed0-b6ae-3b5df52d1c44] allocate_for_instance() {{(pid=65726) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 600.053819] env[65726]: WARNING neutronclient.v2_0.client [None req-af5fca6d-c57c-4762-93f4-4ba84e22ca7b tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 600.057020] env[65726]: WARNING neutronclient.v2_0.client [None req-af5fca6d-c57c-4762-93f4-4ba84e22ca7b tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 600.057020] env[65726]: WARNING openstack [None req-af5fca6d-c57c-4762-93f4-4ba84e22ca7b tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 600.057020] env[65726]: WARNING openstack [None req-af5fca6d-c57c-4762-93f4-4ba84e22ca7b tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 600.463284] env[65726]: DEBUG oslo_concurrency.lockutils [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 600.464797] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9ed3c3d0-6db2-48b9-8065-95d9e297f871 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Acquiring lock "refresh_cache-0e064341-4e4a-407b-8c26-3eb04b409029" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 600.464797] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9ed3c3d0-6db2-48b9-8065-95d9e297f871 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Acquired lock "refresh_cache-0e064341-4e4a-407b-8c26-3eb04b409029" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 600.464797] env[65726]: DEBUG nova.network.neutron [None req-9ed3c3d0-6db2-48b9-8065-95d9e297f871 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] [instance: 0e064341-4e4a-407b-8c26-3eb04b409029] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 600.578553] env[65726]: DEBUG nova.compute.manager [None req-af5fca6d-c57c-4762-93f4-4ba84e22ca7b tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 8e9c27f6-e9fa-4ed0-b6ae-3b5df52d1c44] Start building block device mappings for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 600.614357] env[65726]: DEBUG nova.policy [None req-af5fca6d-c57c-4762-93f4-4ba84e22ca7b tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9b6087cfa55d41a4bad9f6cee89865d5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '39067912d091464192516c08eb576696', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65726) authorize /opt/stack/nova/nova/policy.py:192}} [ 600.770577] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ad2e2ee-a71e-40a7-9e28-3d46aa439cfd {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.777716] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-583e22c8-de5c-45e1-9335-37db44c63386 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.810478] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a921d0e-2da0-421b-9b31-e9d4747e4c17 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.820185] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51836d1c-e39d-404f-8d3f-b0677072c7a2 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.837342] env[65726]: DEBUG nova.compute.provider_tree [None req-5c1fab1f-7856-43d0-84bd-904452799dc9 tempest-ServersTestJSON-1793637688 tempest-ServersTestJSON-1793637688-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 600.972175] env[65726]: WARNING openstack [None req-9ed3c3d0-6db2-48b9-8065-95d9e297f871 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 600.972175] env[65726]: WARNING openstack [None req-9ed3c3d0-6db2-48b9-8065-95d9e297f871 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 600.977473] env[65726]: DEBUG nova.network.neutron [None req-9ed3c3d0-6db2-48b9-8065-95d9e297f871 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] [instance: 0e064341-4e4a-407b-8c26-3eb04b409029] Instance cache missing network info. {{(pid=65726) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 601.226984] env[65726]: DEBUG nova.network.neutron [None req-816382da-6d37-4962-9c02-92a18cd283e5 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] [instance: bb64fc41-19d7-49dc-a17a-9dd74730130f] Successfully updated port: b6351d8e-b77c-4b29-b18b-02fc5e9e0c72 {{(pid=65726) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 601.343778] env[65726]: DEBUG nova.scheduler.client.report [None req-5c1fab1f-7856-43d0-84bd-904452799dc9 tempest-ServersTestJSON-1793637688 tempest-ServersTestJSON-1793637688-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 601.542483] env[65726]: DEBUG nova.network.neutron [None req-af5fca6d-c57c-4762-93f4-4ba84e22ca7b tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 8e9c27f6-e9fa-4ed0-b6ae-3b5df52d1c44] Successfully created port: cc1dce8f-441c-42a7-b3ba-7f3d3fc9ca66 {{(pid=65726) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 601.597462] env[65726]: DEBUG nova.compute.manager [None req-af5fca6d-c57c-4762-93f4-4ba84e22ca7b tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 8e9c27f6-e9fa-4ed0-b6ae-3b5df52d1c44] Start spawning the instance on the hypervisor. {{(pid=65726) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 601.632592] env[65726]: DEBUG nova.virt.hardware [None req-af5fca6d-c57c-4762-93f4-4ba84e22ca7b tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T19:28:34Z,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1539f06fa3534e90acbc3a60c4b8bd3f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T19:28:35Z,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 601.633310] env[65726]: DEBUG nova.virt.hardware [None req-af5fca6d-c57c-4762-93f4-4ba84e22ca7b tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 601.633480] env[65726]: DEBUG nova.virt.hardware [None req-af5fca6d-c57c-4762-93f4-4ba84e22ca7b tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 601.633667] env[65726]: DEBUG nova.virt.hardware [None req-af5fca6d-c57c-4762-93f4-4ba84e22ca7b tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 601.633804] env[65726]: DEBUG nova.virt.hardware [None req-af5fca6d-c57c-4762-93f4-4ba84e22ca7b tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 601.633941] env[65726]: DEBUG nova.virt.hardware [None req-af5fca6d-c57c-4762-93f4-4ba84e22ca7b tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 601.634161] env[65726]: DEBUG nova.virt.hardware [None req-af5fca6d-c57c-4762-93f4-4ba84e22ca7b tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 601.634310] env[65726]: DEBUG nova.virt.hardware [None req-af5fca6d-c57c-4762-93f4-4ba84e22ca7b tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 601.634467] env[65726]: DEBUG nova.virt.hardware [None req-af5fca6d-c57c-4762-93f4-4ba84e22ca7b tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 601.634614] env[65726]: DEBUG nova.virt.hardware [None req-af5fca6d-c57c-4762-93f4-4ba84e22ca7b tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 601.634776] env[65726]: DEBUG nova.virt.hardware [None req-af5fca6d-c57c-4762-93f4-4ba84e22ca7b tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 601.635711] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-794aa33f-0af0-47e8-b847-5389a249dc71 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.650031] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e1ed1bb-df19-4727-92df-5ce3d35b71c6 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.730706] env[65726]: DEBUG oslo_concurrency.lockutils [None req-816382da-6d37-4962-9c02-92a18cd283e5 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Acquiring lock "refresh_cache-bb64fc41-19d7-49dc-a17a-9dd74730130f" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 601.730900] env[65726]: DEBUG oslo_concurrency.lockutils [None req-816382da-6d37-4962-9c02-92a18cd283e5 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Acquired lock "refresh_cache-bb64fc41-19d7-49dc-a17a-9dd74730130f" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 601.731085] env[65726]: DEBUG nova.network.neutron [None req-816382da-6d37-4962-9c02-92a18cd283e5 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] [instance: bb64fc41-19d7-49dc-a17a-9dd74730130f] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 601.849363] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5c1fab1f-7856-43d0-84bd-904452799dc9 tempest-ServersTestJSON-1793637688 tempest-ServersTestJSON-1793637688-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.309s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 601.849950] env[65726]: DEBUG nova.compute.manager [None req-5c1fab1f-7856-43d0-84bd-904452799dc9 tempest-ServersTestJSON-1793637688 tempest-ServersTestJSON-1793637688-project-member] [instance: 88fb7235-cd0e-49dc-9d78-3583d5d3f528] Start building networks asynchronously for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 601.854631] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 2.238s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 602.235894] env[65726]: WARNING openstack [None req-816382da-6d37-4962-9c02-92a18cd283e5 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 602.235894] env[65726]: WARNING openstack [None req-816382da-6d37-4962-9c02-92a18cd283e5 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 602.242972] env[65726]: DEBUG nova.network.neutron [None req-816382da-6d37-4962-9c02-92a18cd283e5 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] [instance: bb64fc41-19d7-49dc-a17a-9dd74730130f] Instance cache missing network info. {{(pid=65726) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 602.360840] env[65726]: DEBUG nova.compute.utils [None req-5c1fab1f-7856-43d0-84bd-904452799dc9 tempest-ServersTestJSON-1793637688 tempest-ServersTestJSON-1793637688-project-member] Using /dev/sd instead of None {{(pid=65726) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 602.367657] env[65726]: DEBUG nova.compute.manager [None req-5c1fab1f-7856-43d0-84bd-904452799dc9 tempest-ServersTestJSON-1793637688 tempest-ServersTestJSON-1793637688-project-member] [instance: 88fb7235-cd0e-49dc-9d78-3583d5d3f528] Allocating IP information in the background. {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 602.367657] env[65726]: DEBUG nova.network.neutron [None req-5c1fab1f-7856-43d0-84bd-904452799dc9 tempest-ServersTestJSON-1793637688 tempest-ServersTestJSON-1793637688-project-member] [instance: 88fb7235-cd0e-49dc-9d78-3583d5d3f528] allocate_for_instance() {{(pid=65726) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 602.367657] env[65726]: WARNING neutronclient.v2_0.client [None req-5c1fab1f-7856-43d0-84bd-904452799dc9 tempest-ServersTestJSON-1793637688 tempest-ServersTestJSON-1793637688-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 602.367657] env[65726]: WARNING neutronclient.v2_0.client [None req-5c1fab1f-7856-43d0-84bd-904452799dc9 tempest-ServersTestJSON-1793637688 tempest-ServersTestJSON-1793637688-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 602.368496] env[65726]: WARNING openstack [None req-5c1fab1f-7856-43d0-84bd-904452799dc9 tempest-ServersTestJSON-1793637688 tempest-ServersTestJSON-1793637688-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 602.368670] env[65726]: WARNING openstack [None req-5c1fab1f-7856-43d0-84bd-904452799dc9 tempest-ServersTestJSON-1793637688 tempest-ServersTestJSON-1793637688-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 602.593200] env[65726]: WARNING openstack [None req-9ed3c3d0-6db2-48b9-8065-95d9e297f871 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 602.593200] env[65726]: WARNING openstack [None req-9ed3c3d0-6db2-48b9-8065-95d9e297f871 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 602.867168] env[65726]: DEBUG nova.compute.manager [None req-5c1fab1f-7856-43d0-84bd-904452799dc9 tempest-ServersTestJSON-1793637688 tempest-ServersTestJSON-1793637688-project-member] [instance: 88fb7235-cd0e-49dc-9d78-3583d5d3f528] Start building block device mappings for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 602.906903] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Instance bb64fc41-19d7-49dc-a17a-9dd74730130f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65726) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 602.907141] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Instance 0e064341-4e4a-407b-8c26-3eb04b409029 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65726) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 602.907352] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Instance 8e9c27f6-e9fa-4ed0-b6ae-3b5df52d1c44 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65726) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 602.908476] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Instance 88fb7235-cd0e-49dc-9d78-3583d5d3f528 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65726) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 602.915589] env[65726]: DEBUG nova.policy [None req-5c1fab1f-7856-43d0-84bd-904452799dc9 tempest-ServersTestJSON-1793637688 tempest-ServersTestJSON-1793637688-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3032cbe03e7a4174ac794478777ba30a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '53f76b7cd812445da88b09acb430da70', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65726) authorize /opt/stack/nova/nova/policy.py:192}} [ 603.006294] env[65726]: WARNING openstack [None req-816382da-6d37-4962-9c02-92a18cd283e5 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 603.006500] env[65726]: WARNING openstack [None req-816382da-6d37-4962-9c02-92a18cd283e5 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 603.108530] env[65726]: WARNING openstack [None req-9ed3c3d0-6db2-48b9-8065-95d9e297f871 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 603.108868] env[65726]: WARNING openstack [None req-9ed3c3d0-6db2-48b9-8065-95d9e297f871 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 603.413110] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Instance 703fff2d-5aec-4ed1-b987-9e1cc3a25f67 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65726) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 603.413530] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Total usable vcpus: 48, total allocated vcpus: 4 {{(pid=65726) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1159}} [ 603.413530] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1280MB phys_disk=100GB used_disk=4GB total_vcpus=48 used_vcpus=4 pci_stats=[] stats={'failed_builds': '0', 'num_instances': '4', 'num_vm_building': '4', 'num_task_spawning': '3', 'num_os_type_None': '4', 'num_proj_06ad412f524c40538c835d9d664cd85e': '1', 'io_workload': '4', 'num_proj_fae9ef0198a2468a9793374b76378abf': '1', 'num_proj_39067912d091464192516c08eb576696': '1', 'num_task_None': '1', 'num_proj_53f76b7cd812445da88b09acb430da70': '1'} {{(pid=65726) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1168}} [ 603.523601] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9290f22f-11c2-4116-a8cb-8eceee5c094c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.534217] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff5f1ee9-b05e-495a-a8b8-dd3696b8e5db {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.574224] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5d4c0e0-1505-47c0-a080-70c97b1e816f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.584066] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab1ee36c-c100-41fe-a7c5-ef08a51d17d1 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.598335] env[65726]: DEBUG nova.compute.provider_tree [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Updating inventory in ProviderTree for provider 07c4692f-bdb4-4058-9173-ff9664830295 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 603.828425] env[65726]: DEBUG nova.network.neutron [None req-af5fca6d-c57c-4762-93f4-4ba84e22ca7b tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 8e9c27f6-e9fa-4ed0-b6ae-3b5df52d1c44] Successfully updated port: cc1dce8f-441c-42a7-b3ba-7f3d3fc9ca66 {{(pid=65726) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 603.888455] env[65726]: DEBUG nova.compute.manager [None req-5c1fab1f-7856-43d0-84bd-904452799dc9 tempest-ServersTestJSON-1793637688 tempest-ServersTestJSON-1793637688-project-member] [instance: 88fb7235-cd0e-49dc-9d78-3583d5d3f528] Start spawning the instance on the hypervisor. {{(pid=65726) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 603.929610] env[65726]: DEBUG nova.virt.hardware [None req-5c1fab1f-7856-43d0-84bd-904452799dc9 tempest-ServersTestJSON-1793637688 tempest-ServersTestJSON-1793637688-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T19:28:34Z,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1539f06fa3534e90acbc3a60c4b8bd3f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T19:28:35Z,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 603.933237] env[65726]: DEBUG nova.virt.hardware [None req-5c1fab1f-7856-43d0-84bd-904452799dc9 tempest-ServersTestJSON-1793637688 tempest-ServersTestJSON-1793637688-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 603.933538] env[65726]: DEBUG nova.virt.hardware [None req-5c1fab1f-7856-43d0-84bd-904452799dc9 tempest-ServersTestJSON-1793637688 tempest-ServersTestJSON-1793637688-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 603.933821] env[65726]: DEBUG nova.virt.hardware [None req-5c1fab1f-7856-43d0-84bd-904452799dc9 tempest-ServersTestJSON-1793637688 tempest-ServersTestJSON-1793637688-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 603.933980] env[65726]: DEBUG nova.virt.hardware [None req-5c1fab1f-7856-43d0-84bd-904452799dc9 tempest-ServersTestJSON-1793637688 tempest-ServersTestJSON-1793637688-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 603.935609] env[65726]: DEBUG nova.virt.hardware [None req-5c1fab1f-7856-43d0-84bd-904452799dc9 tempest-ServersTestJSON-1793637688 tempest-ServersTestJSON-1793637688-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 603.935609] env[65726]: DEBUG nova.virt.hardware [None req-5c1fab1f-7856-43d0-84bd-904452799dc9 tempest-ServersTestJSON-1793637688 tempest-ServersTestJSON-1793637688-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 603.935609] env[65726]: DEBUG nova.virt.hardware [None req-5c1fab1f-7856-43d0-84bd-904452799dc9 tempest-ServersTestJSON-1793637688 tempest-ServersTestJSON-1793637688-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 603.936300] env[65726]: DEBUG nova.virt.hardware [None req-5c1fab1f-7856-43d0-84bd-904452799dc9 tempest-ServersTestJSON-1793637688 tempest-ServersTestJSON-1793637688-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 603.936378] env[65726]: DEBUG nova.virt.hardware [None req-5c1fab1f-7856-43d0-84bd-904452799dc9 tempest-ServersTestJSON-1793637688 tempest-ServersTestJSON-1793637688-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 603.936491] env[65726]: DEBUG nova.virt.hardware [None req-5c1fab1f-7856-43d0-84bd-904452799dc9 tempest-ServersTestJSON-1793637688 tempest-ServersTestJSON-1793637688-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 603.937400] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d41195c-704d-4623-91cd-27b959e171d1 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.946983] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f50021d-dcb6-4e7b-8bea-1ef54e67714d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.000047] env[65726]: DEBUG nova.network.neutron [None req-5c1fab1f-7856-43d0-84bd-904452799dc9 tempest-ServersTestJSON-1793637688 tempest-ServersTestJSON-1793637688-project-member] [instance: 88fb7235-cd0e-49dc-9d78-3583d5d3f528] Successfully created port: b915a591-ee09-4390-ae01-09778bd3f2c3 {{(pid=65726) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 604.050274] env[65726]: WARNING openstack [None req-816382da-6d37-4962-9c02-92a18cd283e5 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 604.050773] env[65726]: WARNING openstack [None req-816382da-6d37-4962-9c02-92a18cd283e5 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 604.106207] env[65726]: DEBUG nova.network.neutron [None req-9ed3c3d0-6db2-48b9-8065-95d9e297f871 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] [instance: 0e064341-4e4a-407b-8c26-3eb04b409029] Updating instance_info_cache with network_info: [{"id": "9d3c5bc9-aa25-4013-b8dc-f565887d2736", "address": "fa:16:3e:f9:95:de", "network": {"id": "55d629b8-a541-4720-9289-927bbc6e6f22", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.93", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "1539f06fa3534e90acbc3a60c4b8bd3f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "35342bcb-8b06-472e-b3c0-43fd3d6c4b30", "external-id": "nsx-vlan-transportzone-524", "segmentation_id": 524, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9d3c5bc9-aa", "ovs_interfaceid": "9d3c5bc9-aa25-4013-b8dc-f565887d2736", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 604.136400] env[65726]: ERROR nova.scheduler.client.report [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] [req-64bbca29-605f-47ed-b110-2eff6c0f4672] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 07c4692f-bdb4-4058-9173-ff9664830295. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-64bbca29-605f-47ed-b110-2eff6c0f4672"}]} [ 604.165255] env[65726]: DEBUG nova.scheduler.client.report [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Refreshing inventories for resource provider 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 604.187734] env[65726]: DEBUG nova.scheduler.client.report [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Updating ProviderTree inventory for provider 07c4692f-bdb4-4058-9173-ff9664830295 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 604.188288] env[65726]: DEBUG nova.compute.provider_tree [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Updating inventory in ProviderTree for provider 07c4692f-bdb4-4058-9173-ff9664830295 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 604.209990] env[65726]: DEBUG nova.scheduler.client.report [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Refreshing aggregate associations for resource provider 07c4692f-bdb4-4058-9173-ff9664830295, aggregates: aa903671-5612-4351-baf5-6ce80b30b4cc {{(pid=65726) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 604.237215] env[65726]: DEBUG nova.scheduler.client.report [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Refreshing trait associations for resource provider 07c4692f-bdb4-4058-9173-ff9664830295, traits: COMPUTE_IMAGE_TYPE_ISO,HW_ARCH_X86_64,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=65726) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 604.335865] env[65726]: DEBUG oslo_concurrency.lockutils [None req-af5fca6d-c57c-4762-93f4-4ba84e22ca7b tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Acquiring lock "refresh_cache-8e9c27f6-e9fa-4ed0-b6ae-3b5df52d1c44" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 604.337385] env[65726]: DEBUG oslo_concurrency.lockutils [None req-af5fca6d-c57c-4762-93f4-4ba84e22ca7b tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Acquired lock "refresh_cache-8e9c27f6-e9fa-4ed0-b6ae-3b5df52d1c44" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 604.337606] env[65726]: DEBUG nova.network.neutron [None req-af5fca6d-c57c-4762-93f4-4ba84e22ca7b tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 8e9c27f6-e9fa-4ed0-b6ae-3b5df52d1c44] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 604.381411] env[65726]: DEBUG nova.network.neutron [None req-816382da-6d37-4962-9c02-92a18cd283e5 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] [instance: bb64fc41-19d7-49dc-a17a-9dd74730130f] Updating instance_info_cache with network_info: [{"id": "b6351d8e-b77c-4b29-b18b-02fc5e9e0c72", "address": "fa:16:3e:7e:3e:65", "network": {"id": "55d629b8-a541-4720-9289-927bbc6e6f22", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.234", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "1539f06fa3534e90acbc3a60c4b8bd3f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "35342bcb-8b06-472e-b3c0-43fd3d6c4b30", "external-id": "nsx-vlan-transportzone-524", "segmentation_id": 524, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb6351d8e-b7", "ovs_interfaceid": "b6351d8e-b77c-4b29-b18b-02fc5e9e0c72", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 604.399253] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-213439f7-0aee-47ad-8c8c-bb8f6c295f31 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.414882] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98c15be7-2db6-4187-920a-cfdb83934191 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.426297] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9da01a84-79f6-4bc1-8b36-334b89c8de59 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Acquiring lock "ee676e3b-3326-46a3-940d-ebbb8b108991" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 604.426297] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9da01a84-79f6-4bc1-8b36-334b89c8de59 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Lock "ee676e3b-3326-46a3-940d-ebbb8b108991" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 604.458887] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08b70ba8-396d-4412-b25c-6209341853c9 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.470051] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4eae981d-ac5a-4fb4-ad38-1be4a4259339 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.491860] env[65726]: DEBUG nova.compute.provider_tree [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Updating inventory in ProviderTree for provider 07c4692f-bdb4-4058-9173-ff9664830295 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 604.615850] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9ed3c3d0-6db2-48b9-8065-95d9e297f871 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Releasing lock "refresh_cache-0e064341-4e4a-407b-8c26-3eb04b409029" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 604.616202] env[65726]: DEBUG nova.compute.manager [None req-9ed3c3d0-6db2-48b9-8065-95d9e297f871 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] [instance: 0e064341-4e4a-407b-8c26-3eb04b409029] Instance network_info: |[{"id": "9d3c5bc9-aa25-4013-b8dc-f565887d2736", "address": "fa:16:3e:f9:95:de", "network": {"id": "55d629b8-a541-4720-9289-927bbc6e6f22", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.93", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "1539f06fa3534e90acbc3a60c4b8bd3f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "35342bcb-8b06-472e-b3c0-43fd3d6c4b30", "external-id": "nsx-vlan-transportzone-524", "segmentation_id": 524, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9d3c5bc9-aa", "ovs_interfaceid": "9d3c5bc9-aa25-4013-b8dc-f565887d2736", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 604.616641] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-9ed3c3d0-6db2-48b9-8065-95d9e297f871 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] [instance: 0e064341-4e4a-407b-8c26-3eb04b409029] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f9:95:de', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '35342bcb-8b06-472e-b3c0-43fd3d6c4b30', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9d3c5bc9-aa25-4013-b8dc-f565887d2736', 'vif_model': 'vmxnet3'}] {{(pid=65726) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 604.631711] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-9ed3c3d0-6db2-48b9-8065-95d9e297f871 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Creating folder: OpenStack. Parent ref: group-v4. {{(pid=65726) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 604.632493] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-33e679ba-f37c-4d30-a771-346c042d4c5e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.647736] env[65726]: WARNING suds.client [-] Web service reported a SOAP processing fault using an unexpected HTTP status code 200. Reporting as an internal server error. [ 604.647736] env[65726]: DEBUG oslo_vmware.api [-] Fault list: [DuplicateName] {{(pid=65726) _invoke_api /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:337}} [ 604.647736] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-9ed3c3d0-6db2-48b9-8065-95d9e297f871 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Folder already exists: OpenStack. Parent ref: group-v4. {{(pid=65726) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1609}} [ 604.647968] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-9ed3c3d0-6db2-48b9-8065-95d9e297f871 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Creating folder: Project (fae9ef0198a2468a9793374b76378abf). Parent ref: group-v995008. {{(pid=65726) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 604.649038] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-592193b6-2e01-460e-9dd2-ee84d61bac43 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.660629] env[65726]: INFO nova.virt.vmwareapi.vm_util [None req-9ed3c3d0-6db2-48b9-8065-95d9e297f871 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Created folder: Project (fae9ef0198a2468a9793374b76378abf) in parent group-v995008. [ 604.660725] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-9ed3c3d0-6db2-48b9-8065-95d9e297f871 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Creating folder: Instances. Parent ref: group-v995012. {{(pid=65726) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 604.661096] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-50879a6e-b4a5-4dac-8d8a-b57052267256 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.672409] env[65726]: INFO nova.virt.vmwareapi.vm_util [None req-9ed3c3d0-6db2-48b9-8065-95d9e297f871 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Created folder: Instances in parent group-v995012. [ 604.672684] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-9ed3c3d0-6db2-48b9-8065-95d9e297f871 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 604.672896] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0e064341-4e4a-407b-8c26-3eb04b409029] Creating VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 604.673135] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2a0b9a5b-225b-4724-a841-9d57b3728767 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.699362] env[65726]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 604.699362] env[65726]: value = "task-5115361" [ 604.699362] env[65726]: _type = "Task" [ 604.699362] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 604.709073] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115361, 'name': CreateVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 604.841877] env[65726]: WARNING openstack [None req-af5fca6d-c57c-4762-93f4-4ba84e22ca7b tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 604.844161] env[65726]: WARNING openstack [None req-af5fca6d-c57c-4762-93f4-4ba84e22ca7b tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 604.850401] env[65726]: DEBUG nova.network.neutron [None req-af5fca6d-c57c-4762-93f4-4ba84e22ca7b tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 8e9c27f6-e9fa-4ed0-b6ae-3b5df52d1c44] Instance cache missing network info. {{(pid=65726) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 604.888436] env[65726]: DEBUG oslo_concurrency.lockutils [None req-816382da-6d37-4962-9c02-92a18cd283e5 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Releasing lock "refresh_cache-bb64fc41-19d7-49dc-a17a-9dd74730130f" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 604.888918] env[65726]: DEBUG nova.compute.manager [None req-816382da-6d37-4962-9c02-92a18cd283e5 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] [instance: bb64fc41-19d7-49dc-a17a-9dd74730130f] Instance network_info: |[{"id": "b6351d8e-b77c-4b29-b18b-02fc5e9e0c72", "address": "fa:16:3e:7e:3e:65", "network": {"id": "55d629b8-a541-4720-9289-927bbc6e6f22", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.234", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "1539f06fa3534e90acbc3a60c4b8bd3f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "35342bcb-8b06-472e-b3c0-43fd3d6c4b30", "external-id": "nsx-vlan-transportzone-524", "segmentation_id": 524, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb6351d8e-b7", "ovs_interfaceid": "b6351d8e-b77c-4b29-b18b-02fc5e9e0c72", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 604.889571] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-816382da-6d37-4962-9c02-92a18cd283e5 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] [instance: bb64fc41-19d7-49dc-a17a-9dd74730130f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:7e:3e:65', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '35342bcb-8b06-472e-b3c0-43fd3d6c4b30', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b6351d8e-b77c-4b29-b18b-02fc5e9e0c72', 'vif_model': 'vmxnet3'}] {{(pid=65726) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 604.899991] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-816382da-6d37-4962-9c02-92a18cd283e5 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Creating folder: Project (06ad412f524c40538c835d9d664cd85e). Parent ref: group-v995008. {{(pid=65726) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 604.900332] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9afeea07-0e85-4ff2-b143-6b86b6dc3eea {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.914985] env[65726]: INFO nova.virt.vmwareapi.vm_util [None req-816382da-6d37-4962-9c02-92a18cd283e5 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Created folder: Project (06ad412f524c40538c835d9d664cd85e) in parent group-v995008. [ 604.917340] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-816382da-6d37-4962-9c02-92a18cd283e5 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Creating folder: Instances. Parent ref: group-v995015. {{(pid=65726) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 604.917340] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ec060a8c-d1ff-4e91-bbf9-8ed0e34df15c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.926739] env[65726]: INFO nova.virt.vmwareapi.vm_util [None req-816382da-6d37-4962-9c02-92a18cd283e5 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Created folder: Instances in parent group-v995015. [ 604.927130] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-816382da-6d37-4962-9c02-92a18cd283e5 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 604.927707] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bb64fc41-19d7-49dc-a17a-9dd74730130f] Creating VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 604.927707] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-af3dc94a-6fc5-4005-b9ca-508af6a32291 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.946113] env[65726]: DEBUG nova.compute.manager [None req-9da01a84-79f6-4bc1-8b36-334b89c8de59 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] [instance: ee676e3b-3326-46a3-940d-ebbb8b108991] Starting instance... {{(pid=65726) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 604.957338] env[65726]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 604.957338] env[65726]: value = "task-5115365" [ 604.957338] env[65726]: _type = "Task" [ 604.957338] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 604.967422] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115365, 'name': CreateVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 605.022260] env[65726]: ERROR nova.scheduler.client.report [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] [req-c7885005-cf1e-41a2-a937-6f344fe5e308] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 07c4692f-bdb4-4058-9173-ff9664830295. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-c7885005-cf1e-41a2-a937-6f344fe5e308"}]} [ 605.043647] env[65726]: DEBUG nova.scheduler.client.report [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Refreshing inventories for resource provider 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 605.067256] env[65726]: DEBUG nova.scheduler.client.report [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Updating ProviderTree inventory for provider 07c4692f-bdb4-4058-9173-ff9664830295 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 605.067922] env[65726]: DEBUG nova.compute.provider_tree [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Updating inventory in ProviderTree for provider 07c4692f-bdb4-4058-9173-ff9664830295 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 605.083034] env[65726]: DEBUG nova.scheduler.client.report [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Refreshing aggregate associations for resource provider 07c4692f-bdb4-4058-9173-ff9664830295, aggregates: None {{(pid=65726) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 605.107414] env[65726]: DEBUG nova.scheduler.client.report [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Refreshing trait associations for resource provider 07c4692f-bdb4-4058-9173-ff9664830295, traits: COMPUTE_IMAGE_TYPE_ISO,HW_ARCH_X86_64,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=65726) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 605.207472] env[65726]: DEBUG nova.compute.manager [req-3d31aac9-2dc0-4643-aa4e-81b13e178d48 req-b58e95ec-3a10-4aae-bba0-9796391b14c6 service nova] [instance: 0e064341-4e4a-407b-8c26-3eb04b409029] Received event network-vif-plugged-9d3c5bc9-aa25-4013-b8dc-f565887d2736 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 605.207472] env[65726]: DEBUG oslo_concurrency.lockutils [req-3d31aac9-2dc0-4643-aa4e-81b13e178d48 req-b58e95ec-3a10-4aae-bba0-9796391b14c6 service nova] Acquiring lock "0e064341-4e4a-407b-8c26-3eb04b409029-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 605.207638] env[65726]: DEBUG oslo_concurrency.lockutils [req-3d31aac9-2dc0-4643-aa4e-81b13e178d48 req-b58e95ec-3a10-4aae-bba0-9796391b14c6 service nova] Lock "0e064341-4e4a-407b-8c26-3eb04b409029-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 605.207789] env[65726]: DEBUG oslo_concurrency.lockutils [req-3d31aac9-2dc0-4643-aa4e-81b13e178d48 req-b58e95ec-3a10-4aae-bba0-9796391b14c6 service nova] Lock "0e064341-4e4a-407b-8c26-3eb04b409029-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 605.207871] env[65726]: DEBUG nova.compute.manager [req-3d31aac9-2dc0-4643-aa4e-81b13e178d48 req-b58e95ec-3a10-4aae-bba0-9796391b14c6 service nova] [instance: 0e064341-4e4a-407b-8c26-3eb04b409029] No waiting events found dispatching network-vif-plugged-9d3c5bc9-aa25-4013-b8dc-f565887d2736 {{(pid=65726) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 605.210183] env[65726]: WARNING nova.compute.manager [req-3d31aac9-2dc0-4643-aa4e-81b13e178d48 req-b58e95ec-3a10-4aae-bba0-9796391b14c6 service nova] [instance: 0e064341-4e4a-407b-8c26-3eb04b409029] Received unexpected event network-vif-plugged-9d3c5bc9-aa25-4013-b8dc-f565887d2736 for instance with vm_state building and task_state spawning. [ 605.215738] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115361, 'name': CreateVM_Task, 'duration_secs': 0.39305} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 605.220978] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0e064341-4e4a-407b-8c26-3eb04b409029] Created VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 605.221627] env[65726]: WARNING openstack [None req-9ed3c3d0-6db2-48b9-8065-95d9e297f871 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 605.221972] env[65726]: WARNING openstack [None req-9ed3c3d0-6db2-48b9-8065-95d9e297f871 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 605.227280] env[65726]: WARNING openstack [None req-9ed3c3d0-6db2-48b9-8065-95d9e297f871 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 605.227614] env[65726]: WARNING openstack [None req-9ed3c3d0-6db2-48b9-8065-95d9e297f871 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 605.262919] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1cce59cf-2f2b-457f-9243-a9443d147274 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.273052] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-924056e8-bcf7-47a1-bed0-a223a38adec4 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.323589] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43e3fec6-ce8a-4ea0-a0f7-65502cf72200 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.336954] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-017184e1-0759-4a8c-8711-5774de385e48 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.350794] env[65726]: DEBUG nova.compute.provider_tree [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Updating inventory in ProviderTree for provider 07c4692f-bdb4-4058-9173-ff9664830295 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 605.381783] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Acquiring lock "8493ba52-b950-4727-970a-19b1797b2ca1" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 605.382028] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Lock "8493ba52-b950-4727-970a-19b1797b2ca1" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 605.429897] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Acquiring lock "21730277-cef6-4bfe-9f67-de71f6f615ff" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 605.430224] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Lock "21730277-cef6-4bfe-9f67-de71f6f615ff" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 605.479272] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115365, 'name': CreateVM_Task, 'duration_secs': 0.373944} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 605.479272] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bb64fc41-19d7-49dc-a17a-9dd74730130f] Created VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 605.479272] env[65726]: WARNING openstack [None req-816382da-6d37-4962-9c02-92a18cd283e5 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 605.479272] env[65726]: WARNING openstack [None req-816382da-6d37-4962-9c02-92a18cd283e5 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 605.479272] env[65726]: DEBUG oslo_vmware.service [None req-816382da-6d37-4962-9c02-92a18cd283e5 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3049589c-ce37-4927-aa48-df86a4deb7d9 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.482653] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9da01a84-79f6-4bc1-8b36-334b89c8de59 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 605.491733] env[65726]: DEBUG oslo_concurrency.lockutils [None req-816382da-6d37-4962-9c02-92a18cd283e5 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 605.491958] env[65726]: DEBUG oslo_concurrency.lockutils [None req-816382da-6d37-4962-9c02-92a18cd283e5 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 605.492727] env[65726]: DEBUG oslo_concurrency.lockutils [None req-816382da-6d37-4962-9c02-92a18cd283e5 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 605.493030] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1ab77f5d-6d53-4b60-87c0-9c910527fcd7 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.499013] env[65726]: DEBUG oslo_vmware.api [None req-816382da-6d37-4962-9c02-92a18cd283e5 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Waiting for the task: (returnval){ [ 605.499013] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]528c33c4-f8b0-1898-5cc6-36c18988eadd" [ 605.499013] env[65726]: _type = "Task" [ 605.499013] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 605.507439] env[65726]: DEBUG oslo_vmware.api [None req-816382da-6d37-4962-9c02-92a18cd283e5 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]528c33c4-f8b0-1898-5cc6-36c18988eadd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 605.628595] env[65726]: WARNING openstack [None req-af5fca6d-c57c-4762-93f4-4ba84e22ca7b tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 605.628946] env[65726]: WARNING openstack [None req-af5fca6d-c57c-4762-93f4-4ba84e22ca7b tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 605.698585] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9ed3c3d0-6db2-48b9-8065-95d9e297f871 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 605.703967] env[65726]: DEBUG oslo_concurrency.lockutils [None req-d2d75a44-1ded-480a-a8b4-89859e465e33 tempest-ServerDiagnosticsV248Test-347810772 tempest-ServerDiagnosticsV248Test-347810772-project-member] Acquiring lock "4d4dc88b-0d6c-430c-ab4f-c81a56e4657b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 605.703967] env[65726]: DEBUG oslo_concurrency.lockutils [None req-d2d75a44-1ded-480a-a8b4-89859e465e33 tempest-ServerDiagnosticsV248Test-347810772 tempest-ServerDiagnosticsV248Test-347810772-project-member] Lock "4d4dc88b-0d6c-430c-ab4f-c81a56e4657b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 605.781989] env[65726]: DEBUG nova.network.neutron [None req-5c1fab1f-7856-43d0-84bd-904452799dc9 tempest-ServersTestJSON-1793637688 tempest-ServersTestJSON-1793637688-project-member] [instance: 88fb7235-cd0e-49dc-9d78-3583d5d3f528] Successfully updated port: b915a591-ee09-4390-ae01-09778bd3f2c3 {{(pid=65726) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 605.876781] env[65726]: ERROR nova.scheduler.client.report [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] [req-aec937bd-3521-4b43-b378-b3857ee781a6] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 07c4692f-bdb4-4058-9173-ff9664830295. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-aec937bd-3521-4b43-b378-b3857ee781a6"}]} [ 605.884841] env[65726]: DEBUG nova.compute.manager [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] [instance: 8493ba52-b950-4727-970a-19b1797b2ca1] Starting instance... {{(pid=65726) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 605.904031] env[65726]: DEBUG nova.scheduler.client.report [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Refreshing inventories for resource provider 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 605.927645] env[65726]: DEBUG nova.scheduler.client.report [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Updating ProviderTree inventory for provider 07c4692f-bdb4-4058-9173-ff9664830295 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 605.927645] env[65726]: DEBUG nova.compute.provider_tree [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Updating inventory in ProviderTree for provider 07c4692f-bdb4-4058-9173-ff9664830295 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 605.933039] env[65726]: DEBUG nova.compute.manager [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] [instance: 21730277-cef6-4bfe-9f67-de71f6f615ff] Starting instance... {{(pid=65726) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 605.949471] env[65726]: DEBUG nova.scheduler.client.report [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Refreshing aggregate associations for resource provider 07c4692f-bdb4-4058-9173-ff9664830295, aggregates: None {{(pid=65726) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 605.984989] env[65726]: DEBUG nova.scheduler.client.report [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Refreshing trait associations for resource provider 07c4692f-bdb4-4058-9173-ff9664830295, traits: COMPUTE_IMAGE_TYPE_ISO,HW_ARCH_X86_64,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=65726) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 606.016666] env[65726]: DEBUG oslo_concurrency.lockutils [None req-816382da-6d37-4962-9c02-92a18cd283e5 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 606.016934] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-816382da-6d37-4962-9c02-92a18cd283e5 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] [instance: bb64fc41-19d7-49dc-a17a-9dd74730130f] Processing image b52362a3-ee8a-4cbf-b06f-513b0cc8f95c {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 606.017180] env[65726]: DEBUG oslo_concurrency.lockutils [None req-816382da-6d37-4962-9c02-92a18cd283e5 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 606.017320] env[65726]: DEBUG oslo_concurrency.lockutils [None req-816382da-6d37-4962-9c02-92a18cd283e5 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 606.019611] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-816382da-6d37-4962-9c02-92a18cd283e5 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 606.020134] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9ed3c3d0-6db2-48b9-8065-95d9e297f871 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 606.020620] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9ed3c3d0-6db2-48b9-8065-95d9e297f871 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 606.020929] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-75f5aa68-7469-4407-9009-182ad22b978d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.023194] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a7da73aa-f8b5-43ac-923d-94eb8af2ad1b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.038864] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-816382da-6d37-4962-9c02-92a18cd283e5 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 606.038948] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-816382da-6d37-4962-9c02-92a18cd283e5 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65726) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 606.040189] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f367d58-4f94-4885-a9a8-af26d9a5e60b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.046408] env[65726]: WARNING openstack [None req-af5fca6d-c57c-4762-93f4-4ba84e22ca7b tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 606.046785] env[65726]: WARNING openstack [None req-af5fca6d-c57c-4762-93f4-4ba84e22ca7b tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 606.057129] env[65726]: DEBUG oslo_vmware.api [None req-9ed3c3d0-6db2-48b9-8065-95d9e297f871 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Waiting for the task: (returnval){ [ 606.057129] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]526457d7-d119-76a2-c7f3-5bc2413796e7" [ 606.057129] env[65726]: _type = "Task" [ 606.057129] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 606.067293] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-195e3836-7d49-4ffc-ba5c-b072b468cf34 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.081697] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9ed3c3d0-6db2-48b9-8065-95d9e297f871 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 606.081973] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-9ed3c3d0-6db2-48b9-8065-95d9e297f871 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] [instance: 0e064341-4e4a-407b-8c26-3eb04b409029] Processing image b52362a3-ee8a-4cbf-b06f-513b0cc8f95c {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 606.082189] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9ed3c3d0-6db2-48b9-8065-95d9e297f871 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 606.083386] env[65726]: DEBUG nova.compute.manager [req-c62332c3-460c-48c8-a8a4-8720ba9f3f1a req-5abcbe57-4c7e-4ab5-8acd-fc5fc062e78c service nova] [instance: bb64fc41-19d7-49dc-a17a-9dd74730130f] Received event network-vif-plugged-b6351d8e-b77c-4b29-b18b-02fc5e9e0c72 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 606.083505] env[65726]: DEBUG oslo_concurrency.lockutils [req-c62332c3-460c-48c8-a8a4-8720ba9f3f1a req-5abcbe57-4c7e-4ab5-8acd-fc5fc062e78c service nova] Acquiring lock "bb64fc41-19d7-49dc-a17a-9dd74730130f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 606.085827] env[65726]: DEBUG oslo_concurrency.lockutils [req-c62332c3-460c-48c8-a8a4-8720ba9f3f1a req-5abcbe57-4c7e-4ab5-8acd-fc5fc062e78c service nova] Lock "bb64fc41-19d7-49dc-a17a-9dd74730130f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 606.085827] env[65726]: DEBUG oslo_concurrency.lockutils [req-c62332c3-460c-48c8-a8a4-8720ba9f3f1a req-5abcbe57-4c7e-4ab5-8acd-fc5fc062e78c service nova] Lock "bb64fc41-19d7-49dc-a17a-9dd74730130f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 606.085827] env[65726]: DEBUG nova.compute.manager [req-c62332c3-460c-48c8-a8a4-8720ba9f3f1a req-5abcbe57-4c7e-4ab5-8acd-fc5fc062e78c service nova] [instance: bb64fc41-19d7-49dc-a17a-9dd74730130f] No waiting events found dispatching network-vif-plugged-b6351d8e-b77c-4b29-b18b-02fc5e9e0c72 {{(pid=65726) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 606.085827] env[65726]: WARNING nova.compute.manager [req-c62332c3-460c-48c8-a8a4-8720ba9f3f1a req-5abcbe57-4c7e-4ab5-8acd-fc5fc062e78c service nova] [instance: bb64fc41-19d7-49dc-a17a-9dd74730130f] Received unexpected event network-vif-plugged-b6351d8e-b77c-4b29-b18b-02fc5e9e0c72 for instance with vm_state building and task_state spawning. [ 606.086608] env[65726]: DEBUG oslo_vmware.api [None req-816382da-6d37-4962-9c02-92a18cd283e5 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Waiting for the task: (returnval){ [ 606.086608] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52a0b7d9-6cfc-17ba-0704-406676db79cd" [ 606.086608] env[65726]: _type = "Task" [ 606.086608] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 606.096772] env[65726]: DEBUG oslo_vmware.api [None req-816382da-6d37-4962-9c02-92a18cd283e5 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52a0b7d9-6cfc-17ba-0704-406676db79cd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 606.165063] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c611d34-6db1-4d62-a75d-6db0052db0c8 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.174379] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9a4e987-73a5-42b3-ad5f-5210fb727c1e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.212433] env[65726]: DEBUG nova.compute.manager [None req-d2d75a44-1ded-480a-a8b4-89859e465e33 tempest-ServerDiagnosticsV248Test-347810772 tempest-ServerDiagnosticsV248Test-347810772-project-member] [instance: 4d4dc88b-0d6c-430c-ab4f-c81a56e4657b] Starting instance... {{(pid=65726) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 606.216455] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ba0bb24-f68f-4b63-b00a-34ef6880cb78 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.224087] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d8b61dc-69af-41e7-9d59-058a641f6487 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.240069] env[65726]: DEBUG nova.compute.provider_tree [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Updating inventory in ProviderTree for provider 07c4692f-bdb4-4058-9173-ff9664830295 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 606.286952] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5c1fab1f-7856-43d0-84bd-904452799dc9 tempest-ServersTestJSON-1793637688 tempest-ServersTestJSON-1793637688-project-member] Acquiring lock "refresh_cache-88fb7235-cd0e-49dc-9d78-3583d5d3f528" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 606.286952] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5c1fab1f-7856-43d0-84bd-904452799dc9 tempest-ServersTestJSON-1793637688 tempest-ServersTestJSON-1793637688-project-member] Acquired lock "refresh_cache-88fb7235-cd0e-49dc-9d78-3583d5d3f528" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 606.287849] env[65726]: DEBUG nova.network.neutron [None req-5c1fab1f-7856-43d0-84bd-904452799dc9 tempest-ServersTestJSON-1793637688 tempest-ServersTestJSON-1793637688-project-member] [instance: 88fb7235-cd0e-49dc-9d78-3583d5d3f528] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 606.413696] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 606.427369] env[65726]: DEBUG nova.network.neutron [None req-af5fca6d-c57c-4762-93f4-4ba84e22ca7b tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 8e9c27f6-e9fa-4ed0-b6ae-3b5df52d1c44] Updating instance_info_cache with network_info: [{"id": "cc1dce8f-441c-42a7-b3ba-7f3d3fc9ca66", "address": "fa:16:3e:8b:2a:82", "network": {"id": "e206b5b9-da36-4e6f-926f-914dec38fac2", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1053201655-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "39067912d091464192516c08eb576696", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7c80243e-93a7-4a95-bc8d-e9534bacd66e", "external-id": "nsx-vlan-transportzone-306", "segmentation_id": 306, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcc1dce8f-44", "ovs_interfaceid": "cc1dce8f-441c-42a7-b3ba-7f3d3fc9ca66", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 606.461699] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 606.588739] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b82a7f4d-3603-48cd-8e48-1d790e39dbfb tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Acquiring lock "cd539d8b-921b-4947-954c-8573380dbdc8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 606.588739] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b82a7f4d-3603-48cd-8e48-1d790e39dbfb tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Lock "cd539d8b-921b-4947-954c-8573380dbdc8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 606.602958] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-816382da-6d37-4962-9c02-92a18cd283e5 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] [instance: bb64fc41-19d7-49dc-a17a-9dd74730130f] Preparing fetch location {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 606.603314] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-816382da-6d37-4962-9c02-92a18cd283e5 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Creating directory with path [datastore1] vmware_temp/442b7d56-8f11-4536-a59e-c880610d03eb/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 606.603902] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4a0e35e9-5f67-4f3e-a229-2cc0ac71573a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.626559] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-816382da-6d37-4962-9c02-92a18cd283e5 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Created directory with path [datastore1] vmware_temp/442b7d56-8f11-4536-a59e-c880610d03eb/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 606.626742] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-816382da-6d37-4962-9c02-92a18cd283e5 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] [instance: bb64fc41-19d7-49dc-a17a-9dd74730130f] Fetch image to [datastore1] vmware_temp/442b7d56-8f11-4536-a59e-c880610d03eb/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/tmp-sparse.vmdk {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 606.626906] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-816382da-6d37-4962-9c02-92a18cd283e5 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] [instance: bb64fc41-19d7-49dc-a17a-9dd74730130f] Downloading image file data b52362a3-ee8a-4cbf-b06f-513b0cc8f95c to [datastore1] vmware_temp/442b7d56-8f11-4536-a59e-c880610d03eb/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/tmp-sparse.vmdk on the data store datastore1 {{(pid=65726) _fetch_image_as_file /opt/stack/nova/nova/virt/vmwareapi/vmops.py:399}} [ 606.628101] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df3c6449-36c5-4edd-9fd7-364faec9f55b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.636042] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01bc1614-6639-4169-902e-ca24ee21d45b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.649273] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b067f45a-24d3-45bb-aee0-ee93eff87ab6 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.691163] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5bb73f13-077c-4ff6-b57f-c8c428099347 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.698967] env[65726]: DEBUG oslo_vmware.service [-] Invoking SessionManager.AcquireGenericServiceTicket with opID=oslo.vmware-ee048a73-646c-40e3-91cb-f4b053c46d8f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.724796] env[65726]: DEBUG nova.virt.vmwareapi.images [None req-816382da-6d37-4962-9c02-92a18cd283e5 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] [instance: bb64fc41-19d7-49dc-a17a-9dd74730130f] Downloading image file data b52362a3-ee8a-4cbf-b06f-513b0cc8f95c to the data store datastore1 {{(pid=65726) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:245}} [ 606.741780] env[65726]: DEBUG oslo_concurrency.lockutils [None req-d2d75a44-1ded-480a-a8b4-89859e465e33 tempest-ServerDiagnosticsV248Test-347810772 tempest-ServerDiagnosticsV248Test-347810772-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 606.767645] env[65726]: ERROR nova.scheduler.client.report [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] [req-9b4931d4-a3dd-49ee-bc96-3820ce538c60] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 07c4692f-bdb4-4058-9173-ff9664830295. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-9b4931d4-a3dd-49ee-bc96-3820ce538c60"}]} [ 606.768106] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 4.914s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 606.771232] env[65726]: ERROR nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Error updating resources for node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28.: nova.exception.ResourceProviderUpdateConflict: A conflict was encountered attempting to update resource provider 07c4692f-bdb4-4058-9173-ff9664830295 (generation 13): {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-9b4931d4-a3dd-49ee-bc96-3820ce538c60"}]} [ 606.771232] env[65726]: ERROR nova.compute.manager Traceback (most recent call last): [ 606.771232] env[65726]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 11287, in _update_available_resource_for_node [ 606.771232] env[65726]: ERROR nova.compute.manager self.rt.update_available_resource(context, nodename, [ 606.771232] env[65726]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/resource_tracker.py", line 965, in update_available_resource [ 606.771232] env[65726]: ERROR nova.compute.manager self._update_available_resource(context, resources, startup=startup) [ 606.771232] env[65726]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 415, in inner [ 606.771232] env[65726]: ERROR nova.compute.manager return f(*args, **kwargs) [ 606.771232] env[65726]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1096, in _update_available_resource [ 606.771232] env[65726]: ERROR nova.compute.manager self._update(context, cn, startup=startup) [ 606.771232] env[65726]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1408, in _update [ 606.771232] env[65726]: ERROR nova.compute.manager self._update_to_placement(context, compute_node, startup) [ 606.771232] env[65726]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 55, in wrapped_f [ 606.771757] env[65726]: ERROR nova.compute.manager return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 606.771757] env[65726]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 289, in call [ 606.771757] env[65726]: ERROR nova.compute.manager raise attempt.get() [ 606.771757] env[65726]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 326, in get [ 606.771757] env[65726]: ERROR nova.compute.manager raise exc.with_traceback(tb) [ 606.771757] env[65726]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 273, in call [ 606.771757] env[65726]: ERROR nova.compute.manager attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 606.771757] env[65726]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1393, in _update_to_placement [ 606.771757] env[65726]: ERROR nova.compute.manager self.reportclient.update_from_provider_tree( [ 606.771757] env[65726]: ERROR nova.compute.manager File "/opt/stack/nova/nova/scheduler/client/report.py", line 1502, in update_from_provider_tree [ 606.771757] env[65726]: ERROR nova.compute.manager self.set_inventory_for_provider( [ 606.771757] env[65726]: ERROR nova.compute.manager File "/opt/stack/nova/nova/scheduler/client/report.py", line 1007, in set_inventory_for_provider [ 606.771757] env[65726]: ERROR nova.compute.manager raise exception.ResourceProviderUpdateConflict( [ 606.771757] env[65726]: ERROR nova.compute.manager nova.exception.ResourceProviderUpdateConflict: A conflict was encountered attempting to update resource provider 07c4692f-bdb4-4058-9173-ff9664830295 (generation 13): {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-9b4931d4-a3dd-49ee-bc96-3820ce538c60"}]} [ 606.771757] env[65726]: ERROR nova.compute.manager [ 606.772423] env[65726]: DEBUG oslo_concurrency.lockutils [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.308s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 606.773214] env[65726]: INFO nova.compute.claims [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] [instance: 703fff2d-5aec-4ed1-b987-9e1cc3a25f67] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 606.791854] env[65726]: WARNING openstack [None req-5c1fab1f-7856-43d0-84bd-904452799dc9 tempest-ServersTestJSON-1793637688 tempest-ServersTestJSON-1793637688-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 606.792333] env[65726]: WARNING openstack [None req-5c1fab1f-7856-43d0-84bd-904452799dc9 tempest-ServersTestJSON-1793637688 tempest-ServersTestJSON-1793637688-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 606.798450] env[65726]: DEBUG nova.network.neutron [None req-5c1fab1f-7856-43d0-84bd-904452799dc9 tempest-ServersTestJSON-1793637688 tempest-ServersTestJSON-1793637688-project-member] [instance: 88fb7235-cd0e-49dc-9d78-3583d5d3f528] Instance cache missing network info. {{(pid=65726) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 606.801562] env[65726]: DEBUG oslo_vmware.rw_handles [None req-816382da-6d37-4962-9c02-92a18cd283e5 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Creating HTTP connection to write to file with size = 21318656 and URL = https://esx7c2n1.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/442b7d56-8f11-4536-a59e-c880610d03eb/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore1. {{(pid=65726) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 606.935775] env[65726]: DEBUG oslo_concurrency.lockutils [None req-af5fca6d-c57c-4762-93f4-4ba84e22ca7b tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Releasing lock "refresh_cache-8e9c27f6-e9fa-4ed0-b6ae-3b5df52d1c44" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 606.935775] env[65726]: DEBUG nova.compute.manager [None req-af5fca6d-c57c-4762-93f4-4ba84e22ca7b tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 8e9c27f6-e9fa-4ed0-b6ae-3b5df52d1c44] Instance network_info: |[{"id": "cc1dce8f-441c-42a7-b3ba-7f3d3fc9ca66", "address": "fa:16:3e:8b:2a:82", "network": {"id": "e206b5b9-da36-4e6f-926f-914dec38fac2", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1053201655-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "39067912d091464192516c08eb576696", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7c80243e-93a7-4a95-bc8d-e9534bacd66e", "external-id": "nsx-vlan-transportzone-306", "segmentation_id": 306, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcc1dce8f-44", "ovs_interfaceid": "cc1dce8f-441c-42a7-b3ba-7f3d3fc9ca66", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 606.935915] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-af5fca6d-c57c-4762-93f4-4ba84e22ca7b tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 8e9c27f6-e9fa-4ed0-b6ae-3b5df52d1c44] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:8b:2a:82', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '7c80243e-93a7-4a95-bc8d-e9534bacd66e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'cc1dce8f-441c-42a7-b3ba-7f3d3fc9ca66', 'vif_model': 'vmxnet3'}] {{(pid=65726) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 606.941108] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-af5fca6d-c57c-4762-93f4-4ba84e22ca7b tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Creating folder: Project (39067912d091464192516c08eb576696). Parent ref: group-v995008. {{(pid=65726) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 606.943076] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-761ea607-d655-4b53-98c7-d0cad6396e5d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.961363] env[65726]: INFO nova.virt.vmwareapi.vm_util [None req-af5fca6d-c57c-4762-93f4-4ba84e22ca7b tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Created folder: Project (39067912d091464192516c08eb576696) in parent group-v995008. [ 606.961639] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-af5fca6d-c57c-4762-93f4-4ba84e22ca7b tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Creating folder: Instances. Parent ref: group-v995018. {{(pid=65726) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 606.962020] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-6ebc3618-dcb6-432a-a8c7-c3c92bc02a57 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.975594] env[65726]: INFO nova.virt.vmwareapi.vm_util [None req-af5fca6d-c57c-4762-93f4-4ba84e22ca7b tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Created folder: Instances in parent group-v995018. [ 606.976208] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-af5fca6d-c57c-4762-93f4-4ba84e22ca7b tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 606.976443] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8e9c27f6-e9fa-4ed0-b6ae-3b5df52d1c44] Creating VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 606.976726] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d620bc51-ea87-4c29-ba87-837ca0240698 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.005263] env[65726]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 607.005263] env[65726]: value = "task-5115369" [ 607.005263] env[65726]: _type = "Task" [ 607.005263] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 607.021244] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115369, 'name': CreateVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 607.094899] env[65726]: DEBUG nova.compute.manager [None req-b82a7f4d-3603-48cd-8e48-1d790e39dbfb tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] [instance: cd539d8b-921b-4947-954c-8573380dbdc8] Starting instance... {{(pid=65726) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 607.391338] env[65726]: WARNING openstack [None req-5c1fab1f-7856-43d0-84bd-904452799dc9 tempest-ServersTestJSON-1793637688 tempest-ServersTestJSON-1793637688-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 607.391777] env[65726]: WARNING openstack [None req-5c1fab1f-7856-43d0-84bd-904452799dc9 tempest-ServersTestJSON-1793637688 tempest-ServersTestJSON-1793637688-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 607.518947] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115369, 'name': CreateVM_Task, 'duration_secs': 0.460895} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 607.520129] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8e9c27f6-e9fa-4ed0-b6ae-3b5df52d1c44] Created VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 607.520617] env[65726]: WARNING openstack [None req-af5fca6d-c57c-4762-93f4-4ba84e22ca7b tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 607.521061] env[65726]: WARNING openstack [None req-af5fca6d-c57c-4762-93f4-4ba84e22ca7b tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 607.527172] env[65726]: DEBUG oslo_concurrency.lockutils [None req-af5fca6d-c57c-4762-93f4-4ba84e22ca7b tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 607.527361] env[65726]: DEBUG oslo_concurrency.lockutils [None req-af5fca6d-c57c-4762-93f4-4ba84e22ca7b tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 607.527681] env[65726]: DEBUG oslo_concurrency.lockutils [None req-af5fca6d-c57c-4762-93f4-4ba84e22ca7b tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 607.531127] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6b358378-8c10-42ae-90df-093fca4256a3 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.542311] env[65726]: DEBUG oslo_vmware.api [None req-af5fca6d-c57c-4762-93f4-4ba84e22ca7b tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Waiting for the task: (returnval){ [ 607.542311] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52d26aac-1faa-82cd-3e93-1b0378883b0e" [ 607.542311] env[65726]: _type = "Task" [ 607.542311] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 607.546635] env[65726]: DEBUG oslo_vmware.rw_handles [None req-816382da-6d37-4962-9c02-92a18cd283e5 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Completed reading data from the image iterator. {{(pid=65726) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 607.546820] env[65726]: DEBUG oslo_vmware.rw_handles [None req-816382da-6d37-4962-9c02-92a18cd283e5 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Closing write handle for https://esx7c2n1.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/442b7d56-8f11-4536-a59e-c880610d03eb/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore1. {{(pid=65726) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:281}} [ 607.553100] env[65726]: DEBUG oslo_vmware.api [None req-af5fca6d-c57c-4762-93f4-4ba84e22ca7b tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52d26aac-1faa-82cd-3e93-1b0378883b0e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 607.618158] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b82a7f4d-3603-48cd-8e48-1d790e39dbfb tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 607.689324] env[65726]: DEBUG nova.virt.vmwareapi.images [None req-816382da-6d37-4962-9c02-92a18cd283e5 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] [instance: bb64fc41-19d7-49dc-a17a-9dd74730130f] Downloaded image file data b52362a3-ee8a-4cbf-b06f-513b0cc8f95c to vmware_temp/442b7d56-8f11-4536-a59e-c880610d03eb/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/tmp-sparse.vmdk on the data store datastore1 {{(pid=65726) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:258}} [ 607.690861] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-816382da-6d37-4962-9c02-92a18cd283e5 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] [instance: bb64fc41-19d7-49dc-a17a-9dd74730130f] Caching image {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 607.691190] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-816382da-6d37-4962-9c02-92a18cd283e5 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Copying Virtual Disk [datastore1] vmware_temp/442b7d56-8f11-4536-a59e-c880610d03eb/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/tmp-sparse.vmdk to [datastore1] vmware_temp/442b7d56-8f11-4536-a59e-c880610d03eb/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 607.691459] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2c78666f-b866-45da-bd5b-92355d6f3c1f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.699390] env[65726]: DEBUG oslo_vmware.api [None req-816382da-6d37-4962-9c02-92a18cd283e5 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Waiting for the task: (returnval){ [ 607.699390] env[65726]: value = "task-5115370" [ 607.699390] env[65726]: _type = "Task" [ 607.699390] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 607.708553] env[65726]: DEBUG oslo_vmware.api [None req-816382da-6d37-4962-9c02-92a18cd283e5 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Task: {'id': task-5115370, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 607.810416] env[65726]: DEBUG nova.scheduler.client.report [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Refreshing inventories for resource provider 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 607.826671] env[65726]: DEBUG nova.scheduler.client.report [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Updating ProviderTree inventory for provider 07c4692f-bdb4-4058-9173-ff9664830295 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 607.826671] env[65726]: DEBUG nova.compute.provider_tree [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Updating inventory in ProviderTree for provider 07c4692f-bdb4-4058-9173-ff9664830295 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 607.845865] env[65726]: DEBUG nova.scheduler.client.report [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Refreshing aggregate associations for resource provider 07c4692f-bdb4-4058-9173-ff9664830295, aggregates: None {{(pid=65726) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 607.874799] env[65726]: DEBUG nova.scheduler.client.report [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Refreshing trait associations for resource provider 07c4692f-bdb4-4058-9173-ff9664830295, traits: COMPUTE_IMAGE_TYPE_ISO,HW_ARCH_X86_64,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=65726) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 607.908870] env[65726]: WARNING openstack [None req-5c1fab1f-7856-43d0-84bd-904452799dc9 tempest-ServersTestJSON-1793637688 tempest-ServersTestJSON-1793637688-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 607.909362] env[65726]: WARNING openstack [None req-5c1fab1f-7856-43d0-84bd-904452799dc9 tempest-ServersTestJSON-1793637688 tempest-ServersTestJSON-1793637688-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 608.059177] env[65726]: DEBUG oslo_concurrency.lockutils [None req-af5fca6d-c57c-4762-93f4-4ba84e22ca7b tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 608.059478] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-af5fca6d-c57c-4762-93f4-4ba84e22ca7b tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 8e9c27f6-e9fa-4ed0-b6ae-3b5df52d1c44] Processing image b52362a3-ee8a-4cbf-b06f-513b0cc8f95c {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 608.059716] env[65726]: DEBUG oslo_concurrency.lockutils [None req-af5fca6d-c57c-4762-93f4-4ba84e22ca7b tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 608.092851] env[65726]: DEBUG nova.network.neutron [None req-5c1fab1f-7856-43d0-84bd-904452799dc9 tempest-ServersTestJSON-1793637688 tempest-ServersTestJSON-1793637688-project-member] [instance: 88fb7235-cd0e-49dc-9d78-3583d5d3f528] Updating instance_info_cache with network_info: [{"id": "b915a591-ee09-4390-ae01-09778bd3f2c3", "address": "fa:16:3e:e3:6b:b7", "network": {"id": "ae04280a-5cd5-46c3-a6f2-8020d38faea4", "bridge": "br-int", "label": "tempest-ServersTestJSON-1476935398-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "53f76b7cd812445da88b09acb430da70", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "49b5df12-d801-4140-8816-2fd401608c7d", "external-id": "nsx-vlan-transportzone-326", "segmentation_id": 326, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb915a591-ee", "ovs_interfaceid": "b915a591-ee09-4390-ae01-09778bd3f2c3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 608.120810] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6cb163b2-18a8-4c85-9ab4-d6894f83b834 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.129568] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a99b7693-f335-4858-86c6-67017ebd0631 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.169904] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a457095c-536a-4ae6-8a60-ca804e064e2e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.180270] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b7782d1-9b4a-49a9-9eef-d016c0bb825c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.202670] env[65726]: DEBUG nova.compute.provider_tree [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Updating inventory in ProviderTree for provider 07c4692f-bdb4-4058-9173-ff9664830295 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 608.214810] env[65726]: DEBUG oslo_vmware.api [None req-816382da-6d37-4962-9c02-92a18cd283e5 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Task: {'id': task-5115370, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 608.600700] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5c1fab1f-7856-43d0-84bd-904452799dc9 tempest-ServersTestJSON-1793637688 tempest-ServersTestJSON-1793637688-project-member] Releasing lock "refresh_cache-88fb7235-cd0e-49dc-9d78-3583d5d3f528" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 608.601785] env[65726]: DEBUG nova.compute.manager [None req-5c1fab1f-7856-43d0-84bd-904452799dc9 tempest-ServersTestJSON-1793637688 tempest-ServersTestJSON-1793637688-project-member] [instance: 88fb7235-cd0e-49dc-9d78-3583d5d3f528] Instance network_info: |[{"id": "b915a591-ee09-4390-ae01-09778bd3f2c3", "address": "fa:16:3e:e3:6b:b7", "network": {"id": "ae04280a-5cd5-46c3-a6f2-8020d38faea4", "bridge": "br-int", "label": "tempest-ServersTestJSON-1476935398-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "53f76b7cd812445da88b09acb430da70", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "49b5df12-d801-4140-8816-2fd401608c7d", "external-id": "nsx-vlan-transportzone-326", "segmentation_id": 326, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb915a591-ee", "ovs_interfaceid": "b915a591-ee09-4390-ae01-09778bd3f2c3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 608.605928] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-5c1fab1f-7856-43d0-84bd-904452799dc9 tempest-ServersTestJSON-1793637688 tempest-ServersTestJSON-1793637688-project-member] [instance: 88fb7235-cd0e-49dc-9d78-3583d5d3f528] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e3:6b:b7', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '49b5df12-d801-4140-8816-2fd401608c7d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b915a591-ee09-4390-ae01-09778bd3f2c3', 'vif_model': 'vmxnet3'}] {{(pid=65726) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 608.615185] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-5c1fab1f-7856-43d0-84bd-904452799dc9 tempest-ServersTestJSON-1793637688 tempest-ServersTestJSON-1793637688-project-member] Creating folder: Project (53f76b7cd812445da88b09acb430da70). Parent ref: group-v995008. {{(pid=65726) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 608.615996] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b8e65ecf-6b31-4881-a7b5-2f0a252ee6d8 tempest-VolumesAssistedSnapshotsTest-1875841935 tempest-VolumesAssistedSnapshotsTest-1875841935-project-member] Acquiring lock "aa2716a9-2cf8-4e51-800d-1c8f98b2cf05" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 608.616552] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b8e65ecf-6b31-4881-a7b5-2f0a252ee6d8 tempest-VolumesAssistedSnapshotsTest-1875841935 tempest-VolumesAssistedSnapshotsTest-1875841935-project-member] Lock "aa2716a9-2cf8-4e51-800d-1c8f98b2cf05" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 608.616552] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e3f8de12-910a-44e6-be59-2abd2ccb8aac {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.630542] env[65726]: INFO nova.virt.vmwareapi.vm_util [None req-5c1fab1f-7856-43d0-84bd-904452799dc9 tempest-ServersTestJSON-1793637688 tempest-ServersTestJSON-1793637688-project-member] Created folder: Project (53f76b7cd812445da88b09acb430da70) in parent group-v995008. [ 608.630678] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-5c1fab1f-7856-43d0-84bd-904452799dc9 tempest-ServersTestJSON-1793637688 tempest-ServersTestJSON-1793637688-project-member] Creating folder: Instances. Parent ref: group-v995022. {{(pid=65726) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 608.630824] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d6e2e43c-a614-4fad-8fee-5014273acd10 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.641636] env[65726]: INFO nova.virt.vmwareapi.vm_util [None req-5c1fab1f-7856-43d0-84bd-904452799dc9 tempest-ServersTestJSON-1793637688 tempest-ServersTestJSON-1793637688-project-member] Created folder: Instances in parent group-v995022. [ 608.641859] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-5c1fab1f-7856-43d0-84bd-904452799dc9 tempest-ServersTestJSON-1793637688 tempest-ServersTestJSON-1793637688-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 608.642070] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 88fb7235-cd0e-49dc-9d78-3583d5d3f528] Creating VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 608.642349] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2d9c3829-32a6-4fc5-bcf6-6fe632907891 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.663882] env[65726]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 608.663882] env[65726]: value = "task-5115373" [ 608.663882] env[65726]: _type = "Task" [ 608.663882] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 608.674501] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115373, 'name': CreateVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 608.715922] env[65726]: DEBUG oslo_vmware.api [None req-816382da-6d37-4962-9c02-92a18cd283e5 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Task: {'id': task-5115370, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.740681} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 608.716112] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-816382da-6d37-4962-9c02-92a18cd283e5 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Copied Virtual Disk [datastore1] vmware_temp/442b7d56-8f11-4536-a59e-c880610d03eb/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/tmp-sparse.vmdk to [datastore1] vmware_temp/442b7d56-8f11-4536-a59e-c880610d03eb/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 608.716426] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-816382da-6d37-4962-9c02-92a18cd283e5 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Deleting the datastore file [datastore1] vmware_temp/442b7d56-8f11-4536-a59e-c880610d03eb/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/tmp-sparse.vmdk {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 608.716915] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6d1b2512-71d7-4944-8574-287497367295 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.727340] env[65726]: DEBUG oslo_vmware.api [None req-816382da-6d37-4962-9c02-92a18cd283e5 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Waiting for the task: (returnval){ [ 608.727340] env[65726]: value = "task-5115374" [ 608.727340] env[65726]: _type = "Task" [ 608.727340] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 608.742097] env[65726]: DEBUG oslo_vmware.api [None req-816382da-6d37-4962-9c02-92a18cd283e5 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Task: {'id': task-5115374, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 608.742097] env[65726]: ERROR nova.scheduler.client.report [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] [req-e10aad5e-07e8-4f14-8c23-c33bdf03e348] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 07c4692f-bdb4-4058-9173-ff9664830295. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-e10aad5e-07e8-4f14-8c23-c33bdf03e348"}]} [ 608.765843] env[65726]: DEBUG nova.scheduler.client.report [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Refreshing inventories for resource provider 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 608.789282] env[65726]: DEBUG nova.scheduler.client.report [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Updating ProviderTree inventory for provider 07c4692f-bdb4-4058-9173-ff9664830295 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 608.789282] env[65726]: DEBUG nova.compute.provider_tree [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Updating inventory in ProviderTree for provider 07c4692f-bdb4-4058-9173-ff9664830295 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 608.817588] env[65726]: DEBUG nova.scheduler.client.report [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Refreshing aggregate associations for resource provider 07c4692f-bdb4-4058-9173-ff9664830295, aggregates: None {{(pid=65726) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 608.846044] env[65726]: DEBUG nova.scheduler.client.report [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Refreshing trait associations for resource provider 07c4692f-bdb4-4058-9173-ff9664830295, traits: COMPUTE_IMAGE_TYPE_ISO,HW_ARCH_X86_64,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=65726) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 609.113393] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d30e247-9d99-4238-a8a0-099ba8fac258 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.127412] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17288110-84e0-44c1-9773-94934771902b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.163933] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95bd7b6c-7528-47f7-9006-3af5d5c32858 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.182890] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad6c3dd3-27b3-4e37-a6e8-41247e0c7d68 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.187108] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115373, 'name': CreateVM_Task, 'duration_secs': 0.3898} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 609.187363] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 88fb7235-cd0e-49dc-9d78-3583d5d3f528] Created VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 609.188488] env[65726]: WARNING openstack [None req-5c1fab1f-7856-43d0-84bd-904452799dc9 tempest-ServersTestJSON-1793637688 tempest-ServersTestJSON-1793637688-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 609.188953] env[65726]: WARNING openstack [None req-5c1fab1f-7856-43d0-84bd-904452799dc9 tempest-ServersTestJSON-1793637688 tempest-ServersTestJSON-1793637688-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 609.194479] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5c1fab1f-7856-43d0-84bd-904452799dc9 tempest-ServersTestJSON-1793637688 tempest-ServersTestJSON-1793637688-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 609.194626] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5c1fab1f-7856-43d0-84bd-904452799dc9 tempest-ServersTestJSON-1793637688 tempest-ServersTestJSON-1793637688-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 609.195789] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5c1fab1f-7856-43d0-84bd-904452799dc9 tempest-ServersTestJSON-1793637688 tempest-ServersTestJSON-1793637688-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 609.205062] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-78dabb22-4645-4546-bbf9-9507eb901b01 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.207211] env[65726]: DEBUG nova.compute.provider_tree [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Updating inventory in ProviderTree for provider 07c4692f-bdb4-4058-9173-ff9664830295 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 609.212632] env[65726]: DEBUG oslo_vmware.api [None req-5c1fab1f-7856-43d0-84bd-904452799dc9 tempest-ServersTestJSON-1793637688 tempest-ServersTestJSON-1793637688-project-member] Waiting for the task: (returnval){ [ 609.212632] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52d80482-4ba3-eb40-af2a-130182a5ee33" [ 609.212632] env[65726]: _type = "Task" [ 609.212632] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 609.222351] env[65726]: DEBUG oslo_vmware.api [None req-5c1fab1f-7856-43d0-84bd-904452799dc9 tempest-ServersTestJSON-1793637688 tempest-ServersTestJSON-1793637688-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52d80482-4ba3-eb40-af2a-130182a5ee33, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 609.240662] env[65726]: DEBUG oslo_vmware.api [None req-816382da-6d37-4962-9c02-92a18cd283e5 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Task: {'id': task-5115374, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.063801} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 609.240939] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-816382da-6d37-4962-9c02-92a18cd283e5 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Deleted the datastore file {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 609.241865] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-816382da-6d37-4962-9c02-92a18cd283e5 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Moving file from [datastore1] vmware_temp/442b7d56-8f11-4536-a59e-c880610d03eb/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c to [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c. {{(pid=65726) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:334}} [ 609.241865] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MoveDatastoreFile_Task with opID=oslo.vmware-4df209c7-0e4e-43e4-9d3b-3d7cc4dc157d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.249633] env[65726]: DEBUG oslo_vmware.api [None req-816382da-6d37-4962-9c02-92a18cd283e5 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Waiting for the task: (returnval){ [ 609.249633] env[65726]: value = "task-5115376" [ 609.249633] env[65726]: _type = "Task" [ 609.249633] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 609.261259] env[65726]: DEBUG oslo_vmware.api [None req-816382da-6d37-4962-9c02-92a18cd283e5 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Task: {'id': task-5115376, 'name': MoveDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 609.541594] env[65726]: DEBUG oslo_concurrency.lockutils [None req-8fd7433e-ebac-43f4-aa60-6301be278066 tempest-ServerDiagnosticsTest-2021442419 tempest-ServerDiagnosticsTest-2021442419-project-member] Acquiring lock "de5cad99-1dbc-4435-98b4-987e24ba8c3c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 609.541813] env[65726]: DEBUG oslo_concurrency.lockutils [None req-8fd7433e-ebac-43f4-aa60-6301be278066 tempest-ServerDiagnosticsTest-2021442419 tempest-ServerDiagnosticsTest-2021442419-project-member] Lock "de5cad99-1dbc-4435-98b4-987e24ba8c3c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 609.649808] env[65726]: DEBUG nova.compute.manager [req-593c5f97-a230-40c6-aec7-28e40819c5fb req-77be3017-6f29-4849-87a9-dc59653a8645 service nova] [instance: 0e064341-4e4a-407b-8c26-3eb04b409029] Received event network-changed-9d3c5bc9-aa25-4013-b8dc-f565887d2736 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 609.650136] env[65726]: DEBUG nova.compute.manager [req-593c5f97-a230-40c6-aec7-28e40819c5fb req-77be3017-6f29-4849-87a9-dc59653a8645 service nova] [instance: 0e064341-4e4a-407b-8c26-3eb04b409029] Refreshing instance network info cache due to event network-changed-9d3c5bc9-aa25-4013-b8dc-f565887d2736. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 609.653741] env[65726]: DEBUG oslo_concurrency.lockutils [req-593c5f97-a230-40c6-aec7-28e40819c5fb req-77be3017-6f29-4849-87a9-dc59653a8645 service nova] Acquiring lock "refresh_cache-0e064341-4e4a-407b-8c26-3eb04b409029" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 609.653741] env[65726]: DEBUG oslo_concurrency.lockutils [req-593c5f97-a230-40c6-aec7-28e40819c5fb req-77be3017-6f29-4849-87a9-dc59653a8645 service nova] Acquired lock "refresh_cache-0e064341-4e4a-407b-8c26-3eb04b409029" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 609.653741] env[65726]: DEBUG nova.network.neutron [req-593c5f97-a230-40c6-aec7-28e40819c5fb req-77be3017-6f29-4849-87a9-dc59653a8645 service nova] [instance: 0e064341-4e4a-407b-8c26-3eb04b409029] Refreshing network info cache for port 9d3c5bc9-aa25-4013-b8dc-f565887d2736 {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 609.731300] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5c1fab1f-7856-43d0-84bd-904452799dc9 tempest-ServersTestJSON-1793637688 tempest-ServersTestJSON-1793637688-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 609.731470] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-5c1fab1f-7856-43d0-84bd-904452799dc9 tempest-ServersTestJSON-1793637688 tempest-ServersTestJSON-1793637688-project-member] [instance: 88fb7235-cd0e-49dc-9d78-3583d5d3f528] Processing image b52362a3-ee8a-4cbf-b06f-513b0cc8f95c {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 609.732040] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5c1fab1f-7856-43d0-84bd-904452799dc9 tempest-ServersTestJSON-1793637688 tempest-ServersTestJSON-1793637688-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 609.743019] env[65726]: ERROR nova.scheduler.client.report [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] [req-e3fe7db8-39af-42b0-ba15-64ca2acc5eea] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 07c4692f-bdb4-4058-9173-ff9664830295. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-e3fe7db8-39af-42b0-ba15-64ca2acc5eea"}]} [ 609.762536] env[65726]: DEBUG oslo_vmware.api [None req-816382da-6d37-4962-9c02-92a18cd283e5 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Task: {'id': task-5115376, 'name': MoveDatastoreFile_Task, 'duration_secs': 0.034132} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 609.762999] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-816382da-6d37-4962-9c02-92a18cd283e5 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] File moved {{(pid=65726) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:346}} [ 609.763324] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-816382da-6d37-4962-9c02-92a18cd283e5 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] [instance: bb64fc41-19d7-49dc-a17a-9dd74730130f] Cleaning up location [datastore1] vmware_temp/442b7d56-8f11-4536-a59e-c880610d03eb {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 609.763691] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-816382da-6d37-4962-9c02-92a18cd283e5 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Deleting the datastore file [datastore1] vmware_temp/442b7d56-8f11-4536-a59e-c880610d03eb {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 609.764042] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-de952fa3-b8b6-428c-9fbf-d1e542ad0228 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.767269] env[65726]: DEBUG nova.scheduler.client.report [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Refreshing inventories for resource provider 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 609.776037] env[65726]: DEBUG oslo_vmware.api [None req-816382da-6d37-4962-9c02-92a18cd283e5 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Waiting for the task: (returnval){ [ 609.776037] env[65726]: value = "task-5115377" [ 609.776037] env[65726]: _type = "Task" [ 609.776037] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 609.789229] env[65726]: DEBUG oslo_vmware.api [None req-816382da-6d37-4962-9c02-92a18cd283e5 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Task: {'id': task-5115377, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 609.790312] env[65726]: DEBUG nova.scheduler.client.report [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Updating ProviderTree inventory for provider 07c4692f-bdb4-4058-9173-ff9664830295 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 609.791295] env[65726]: DEBUG nova.compute.provider_tree [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Updating inventory in ProviderTree for provider 07c4692f-bdb4-4058-9173-ff9664830295 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 609.806604] env[65726]: DEBUG nova.scheduler.client.report [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Refreshing aggregate associations for resource provider 07c4692f-bdb4-4058-9173-ff9664830295, aggregates: None {{(pid=65726) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 609.834062] env[65726]: DEBUG nova.scheduler.client.report [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Refreshing trait associations for resource provider 07c4692f-bdb4-4058-9173-ff9664830295, traits: COMPUTE_IMAGE_TYPE_ISO,HW_ARCH_X86_64,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=65726) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 610.133021] env[65726]: DEBUG nova.compute.manager [req-fccf0b77-e7bb-41aa-b6c5-117d34566265 req-34c11ed9-fede-4a19-967a-1b3c57bdf97e service nova] [instance: bb64fc41-19d7-49dc-a17a-9dd74730130f] Received event network-changed-b6351d8e-b77c-4b29-b18b-02fc5e9e0c72 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 610.133021] env[65726]: DEBUG nova.compute.manager [req-fccf0b77-e7bb-41aa-b6c5-117d34566265 req-34c11ed9-fede-4a19-967a-1b3c57bdf97e service nova] [instance: bb64fc41-19d7-49dc-a17a-9dd74730130f] Refreshing instance network info cache due to event network-changed-b6351d8e-b77c-4b29-b18b-02fc5e9e0c72. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 610.133021] env[65726]: DEBUG oslo_concurrency.lockutils [req-fccf0b77-e7bb-41aa-b6c5-117d34566265 req-34c11ed9-fede-4a19-967a-1b3c57bdf97e service nova] Acquiring lock "refresh_cache-bb64fc41-19d7-49dc-a17a-9dd74730130f" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 610.133021] env[65726]: DEBUG oslo_concurrency.lockutils [req-fccf0b77-e7bb-41aa-b6c5-117d34566265 req-34c11ed9-fede-4a19-967a-1b3c57bdf97e service nova] Acquired lock "refresh_cache-bb64fc41-19d7-49dc-a17a-9dd74730130f" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 610.133219] env[65726]: DEBUG nova.network.neutron [req-fccf0b77-e7bb-41aa-b6c5-117d34566265 req-34c11ed9-fede-4a19-967a-1b3c57bdf97e service nova] [instance: bb64fc41-19d7-49dc-a17a-9dd74730130f] Refreshing network info cache for port b6351d8e-b77c-4b29-b18b-02fc5e9e0c72 {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 610.140405] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75a90918-5028-401d-bc08-e674958e9456 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.151166] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82cf0423-9d1f-4755-9a60-3f99ed4fa22f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.159170] env[65726]: WARNING openstack [req-593c5f97-a230-40c6-aec7-28e40819c5fb req-77be3017-6f29-4849-87a9-dc59653a8645 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 610.159630] env[65726]: WARNING openstack [req-593c5f97-a230-40c6-aec7-28e40819c5fb req-77be3017-6f29-4849-87a9-dc59653a8645 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 610.207021] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61b489c6-da23-45d0-84c2-435a180b4ad2 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.215769] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b44cd33f-06d5-4d2e-a68e-1db9a923bd26 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.234013] env[65726]: DEBUG nova.compute.provider_tree [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Updating inventory in ProviderTree for provider 07c4692f-bdb4-4058-9173-ff9664830295 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 610.287285] env[65726]: DEBUG oslo_vmware.api [None req-816382da-6d37-4962-9c02-92a18cd283e5 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Task: {'id': task-5115377, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.028943} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 610.287595] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-816382da-6d37-4962-9c02-92a18cd283e5 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Deleted the datastore file {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 610.288745] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-99c0690e-f232-4395-8525-ba63f5cf0ff9 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.295730] env[65726]: DEBUG oslo_vmware.api [None req-816382da-6d37-4962-9c02-92a18cd283e5 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Waiting for the task: (returnval){ [ 610.295730] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52d45307-1a00-6119-e733-59607a44f347" [ 610.295730] env[65726]: _type = "Task" [ 610.295730] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 610.305581] env[65726]: DEBUG oslo_vmware.api [None req-816382da-6d37-4962-9c02-92a18cd283e5 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52d45307-1a00-6119-e733-59607a44f347, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 610.517586] env[65726]: DEBUG oslo_concurrency.lockutils [None req-8cc1dec8-17ac-4db3-bea9-0fb861f8b4b8 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Acquiring lock "04ffdf80-c54c-4587-9bb4-d520dc440501" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 610.518243] env[65726]: DEBUG oslo_concurrency.lockutils [None req-8cc1dec8-17ac-4db3-bea9-0fb861f8b4b8 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Lock "04ffdf80-c54c-4587-9bb4-d520dc440501" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 610.645021] env[65726]: WARNING openstack [req-fccf0b77-e7bb-41aa-b6c5-117d34566265 req-34c11ed9-fede-4a19-967a-1b3c57bdf97e service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 610.645690] env[65726]: WARNING openstack [req-fccf0b77-e7bb-41aa-b6c5-117d34566265 req-34c11ed9-fede-4a19-967a-1b3c57bdf97e service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 610.768436] env[65726]: ERROR nova.scheduler.client.report [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] [req-e56886a7-8d89-4223-a158-2dc41050cc6d] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 07c4692f-bdb4-4058-9173-ff9664830295. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-e56886a7-8d89-4223-a158-2dc41050cc6d"}]} [ 610.793099] env[65726]: DEBUG nova.scheduler.client.report [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Refreshing inventories for resource provider 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 610.809195] env[65726]: DEBUG oslo_vmware.api [None req-816382da-6d37-4962-9c02-92a18cd283e5 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52d45307-1a00-6119-e733-59607a44f347, 'name': SearchDatastore_Task, 'duration_secs': 0.009808} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 610.809720] env[65726]: DEBUG oslo_concurrency.lockutils [None req-816382da-6d37-4962-9c02-92a18cd283e5 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 610.809720] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-816382da-6d37-4962-9c02-92a18cd283e5 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] bb64fc41-19d7-49dc-a17a-9dd74730130f/bb64fc41-19d7-49dc-a17a-9dd74730130f.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 610.810012] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9ed3c3d0-6db2-48b9-8065-95d9e297f871 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 610.810376] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-9ed3c3d0-6db2-48b9-8065-95d9e297f871 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 610.810456] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-cb6c4df1-ce31-4383-933a-58beac230fa9 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.813342] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-befdbe15-c25f-4bf0-96fb-4494aecca9ca {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.819451] env[65726]: DEBUG nova.scheduler.client.report [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Updating ProviderTree inventory for provider 07c4692f-bdb4-4058-9173-ff9664830295 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 610.819769] env[65726]: DEBUG nova.compute.provider_tree [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Updating inventory in ProviderTree for provider 07c4692f-bdb4-4058-9173-ff9664830295 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 610.823280] env[65726]: DEBUG oslo_vmware.api [None req-816382da-6d37-4962-9c02-92a18cd283e5 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Waiting for the task: (returnval){ [ 610.823280] env[65726]: value = "task-5115378" [ 610.823280] env[65726]: _type = "Task" [ 610.823280] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 610.824619] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-9ed3c3d0-6db2-48b9-8065-95d9e297f871 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 610.824777] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-9ed3c3d0-6db2-48b9-8065-95d9e297f871 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65726) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 610.828574] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9090636d-cb35-4cf9-a715-918289e6e509 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.836627] env[65726]: DEBUG oslo_vmware.api [None req-816382da-6d37-4962-9c02-92a18cd283e5 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Task: {'id': task-5115378, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 610.838213] env[65726]: DEBUG oslo_vmware.api [None req-9ed3c3d0-6db2-48b9-8065-95d9e297f871 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Waiting for the task: (returnval){ [ 610.838213] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5278dbcf-77f6-21eb-5c2e-a9a095c388d1" [ 610.838213] env[65726]: _type = "Task" [ 610.838213] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 610.839607] env[65726]: DEBUG nova.scheduler.client.report [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Refreshing aggregate associations for resource provider 07c4692f-bdb4-4058-9173-ff9664830295, aggregates: None {{(pid=65726) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 610.850535] env[65726]: DEBUG oslo_vmware.api [None req-9ed3c3d0-6db2-48b9-8065-95d9e297f871 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5278dbcf-77f6-21eb-5c2e-a9a095c388d1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 610.864870] env[65726]: DEBUG nova.scheduler.client.report [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Refreshing trait associations for resource provider 07c4692f-bdb4-4058-9173-ff9664830295, traits: COMPUTE_IMAGE_TYPE_ISO,HW_ARCH_X86_64,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=65726) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 611.143893] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-778f06bb-68cc-4fce-9d0a-bec312c4cb5b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.156482] env[65726]: WARNING openstack [req-593c5f97-a230-40c6-aec7-28e40819c5fb req-77be3017-6f29-4849-87a9-dc59653a8645 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 611.156781] env[65726]: WARNING openstack [req-593c5f97-a230-40c6-aec7-28e40819c5fb req-77be3017-6f29-4849-87a9-dc59653a8645 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 611.166440] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b8f201c-f6bc-4347-aab7-801c96a9b0e4 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.221654] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ee5d3f8-713d-4ca0-990e-817acd8892d6 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.231205] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c710020-dbb2-4431-a9c7-fa044681428e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.251155] env[65726]: DEBUG nova.compute.provider_tree [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Updating inventory in ProviderTree for provider 07c4692f-bdb4-4058-9173-ff9664830295 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 611.335784] env[65726]: WARNING openstack [req-fccf0b77-e7bb-41aa-b6c5-117d34566265 req-34c11ed9-fede-4a19-967a-1b3c57bdf97e service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 611.335784] env[65726]: WARNING openstack [req-fccf0b77-e7bb-41aa-b6c5-117d34566265 req-34c11ed9-fede-4a19-967a-1b3c57bdf97e service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 611.348709] env[65726]: DEBUG oslo_vmware.api [None req-816382da-6d37-4962-9c02-92a18cd283e5 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Task: {'id': task-5115378, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.509745} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 611.349466] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-816382da-6d37-4962-9c02-92a18cd283e5 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] bb64fc41-19d7-49dc-a17a-9dd74730130f/bb64fc41-19d7-49dc-a17a-9dd74730130f.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 611.349772] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-816382da-6d37-4962-9c02-92a18cd283e5 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] [instance: bb64fc41-19d7-49dc-a17a-9dd74730130f] Extending root virtual disk to 1048576 {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 611.349956] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d158bc4b-1f95-47da-bc4f-f177194a9227 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.357416] env[65726]: DEBUG oslo_vmware.api [None req-9ed3c3d0-6db2-48b9-8065-95d9e297f871 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5278dbcf-77f6-21eb-5c2e-a9a095c388d1, 'name': SearchDatastore_Task, 'duration_secs': 0.017494} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 611.357416] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-537162a0-32a5-4e45-8cb3-f308137bd5fc {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.362233] env[65726]: DEBUG oslo_vmware.api [None req-816382da-6d37-4962-9c02-92a18cd283e5 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Waiting for the task: (returnval){ [ 611.362233] env[65726]: value = "task-5115380" [ 611.362233] env[65726]: _type = "Task" [ 611.362233] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 611.363348] env[65726]: DEBUG oslo_vmware.api [None req-9ed3c3d0-6db2-48b9-8065-95d9e297f871 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Waiting for the task: (returnval){ [ 611.363348] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52da4d85-f4c0-1ebe-df93-a32c03800bdc" [ 611.363348] env[65726]: _type = "Task" [ 611.363348] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 611.373650] env[65726]: WARNING openstack [req-593c5f97-a230-40c6-aec7-28e40819c5fb req-77be3017-6f29-4849-87a9-dc59653a8645 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 611.374061] env[65726]: WARNING openstack [req-593c5f97-a230-40c6-aec7-28e40819c5fb req-77be3017-6f29-4849-87a9-dc59653a8645 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 611.390974] env[65726]: DEBUG oslo_vmware.api [None req-816382da-6d37-4962-9c02-92a18cd283e5 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Task: {'id': task-5115380, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 611.393988] env[65726]: DEBUG oslo_vmware.api [None req-9ed3c3d0-6db2-48b9-8065-95d9e297f871 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52da4d85-f4c0-1ebe-df93-a32c03800bdc, 'name': SearchDatastore_Task, 'duration_secs': 0.017979} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 611.398448] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9ed3c3d0-6db2-48b9-8065-95d9e297f871 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 611.398744] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-9ed3c3d0-6db2-48b9-8065-95d9e297f871 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] 0e064341-4e4a-407b-8c26-3eb04b409029/0e064341-4e4a-407b-8c26-3eb04b409029.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 611.399709] env[65726]: DEBUG oslo_concurrency.lockutils [None req-af5fca6d-c57c-4762-93f4-4ba84e22ca7b tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 611.399827] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-af5fca6d-c57c-4762-93f4-4ba84e22ca7b tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 611.400544] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c16490eb-e865-45f2-b3b9-c531e8342980 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.402457] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-577c6e25-d2e8-429a-84b9-e36f653f552c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.413342] env[65726]: DEBUG oslo_vmware.api [None req-9ed3c3d0-6db2-48b9-8065-95d9e297f871 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Waiting for the task: (returnval){ [ 611.413342] env[65726]: value = "task-5115381" [ 611.413342] env[65726]: _type = "Task" [ 611.413342] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 611.416782] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-af5fca6d-c57c-4762-93f4-4ba84e22ca7b tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 611.416962] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-af5fca6d-c57c-4762-93f4-4ba84e22ca7b tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65726) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 611.418383] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-edabeed8-4b8c-4b98-b0c1-e7359ab54d7e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.425311] env[65726]: DEBUG oslo_vmware.api [None req-9ed3c3d0-6db2-48b9-8065-95d9e297f871 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Task: {'id': task-5115381, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 611.427117] env[65726]: DEBUG oslo_vmware.api [None req-af5fca6d-c57c-4762-93f4-4ba84e22ca7b tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Waiting for the task: (returnval){ [ 611.427117] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]522da169-e62d-cd9c-7fce-a096337678c6" [ 611.427117] env[65726]: _type = "Task" [ 611.427117] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 611.437462] env[65726]: DEBUG oslo_vmware.api [None req-af5fca6d-c57c-4762-93f4-4ba84e22ca7b tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]522da169-e62d-cd9c-7fce-a096337678c6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 611.479856] env[65726]: WARNING openstack [req-fccf0b77-e7bb-41aa-b6c5-117d34566265 req-34c11ed9-fede-4a19-967a-1b3c57bdf97e service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 611.480299] env[65726]: WARNING openstack [req-fccf0b77-e7bb-41aa-b6c5-117d34566265 req-34c11ed9-fede-4a19-967a-1b3c57bdf97e service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 611.524364] env[65726]: DEBUG nova.network.neutron [req-593c5f97-a230-40c6-aec7-28e40819c5fb req-77be3017-6f29-4849-87a9-dc59653a8645 service nova] [instance: 0e064341-4e4a-407b-8c26-3eb04b409029] Updated VIF entry in instance network info cache for port 9d3c5bc9-aa25-4013-b8dc-f565887d2736. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 611.524364] env[65726]: DEBUG nova.network.neutron [req-593c5f97-a230-40c6-aec7-28e40819c5fb req-77be3017-6f29-4849-87a9-dc59653a8645 service nova] [instance: 0e064341-4e4a-407b-8c26-3eb04b409029] Updating instance_info_cache with network_info: [{"id": "9d3c5bc9-aa25-4013-b8dc-f565887d2736", "address": "fa:16:3e:f9:95:de", "network": {"id": "55d629b8-a541-4720-9289-927bbc6e6f22", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.93", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "1539f06fa3534e90acbc3a60c4b8bd3f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "35342bcb-8b06-472e-b3c0-43fd3d6c4b30", "external-id": "nsx-vlan-transportzone-524", "segmentation_id": 524, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9d3c5bc9-aa", "ovs_interfaceid": "9d3c5bc9-aa25-4013-b8dc-f565887d2736", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 611.702929] env[65726]: DEBUG nova.network.neutron [req-fccf0b77-e7bb-41aa-b6c5-117d34566265 req-34c11ed9-fede-4a19-967a-1b3c57bdf97e service nova] [instance: bb64fc41-19d7-49dc-a17a-9dd74730130f] Updated VIF entry in instance network info cache for port b6351d8e-b77c-4b29-b18b-02fc5e9e0c72. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 611.705159] env[65726]: DEBUG nova.network.neutron [req-fccf0b77-e7bb-41aa-b6c5-117d34566265 req-34c11ed9-fede-4a19-967a-1b3c57bdf97e service nova] [instance: bb64fc41-19d7-49dc-a17a-9dd74730130f] Updating instance_info_cache with network_info: [{"id": "b6351d8e-b77c-4b29-b18b-02fc5e9e0c72", "address": "fa:16:3e:7e:3e:65", "network": {"id": "55d629b8-a541-4720-9289-927bbc6e6f22", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.234", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "1539f06fa3534e90acbc3a60c4b8bd3f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "35342bcb-8b06-472e-b3c0-43fd3d6c4b30", "external-id": "nsx-vlan-transportzone-524", "segmentation_id": 524, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb6351d8e-b7", "ovs_interfaceid": "b6351d8e-b77c-4b29-b18b-02fc5e9e0c72", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 611.804702] env[65726]: DEBUG nova.scheduler.client.report [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Updated inventory for provider 07c4692f-bdb4-4058-9173-ff9664830295 with generation 17 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 611.805056] env[65726]: DEBUG nova.compute.provider_tree [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Updating resource provider 07c4692f-bdb4-4058-9173-ff9664830295 generation from 17 to 18 during operation: update_inventory {{(pid=65726) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 611.805163] env[65726]: DEBUG nova.compute.provider_tree [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Updating inventory in ProviderTree for provider 07c4692f-bdb4-4058-9173-ff9664830295 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 611.875596] env[65726]: DEBUG oslo_vmware.api [None req-816382da-6d37-4962-9c02-92a18cd283e5 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Task: {'id': task-5115380, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.072184} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 611.876056] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-816382da-6d37-4962-9c02-92a18cd283e5 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] [instance: bb64fc41-19d7-49dc-a17a-9dd74730130f] Extended root virtual disk {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 611.877243] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f821de0b-53fb-4c5c-a22c-cab6f7269785 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.906086] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-816382da-6d37-4962-9c02-92a18cd283e5 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] [instance: bb64fc41-19d7-49dc-a17a-9dd74730130f] Reconfiguring VM instance instance-00000001 to attach disk [datastore1] bb64fc41-19d7-49dc-a17a-9dd74730130f/bb64fc41-19d7-49dc-a17a-9dd74730130f.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 611.906472] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6a083379-d9d7-4d64-989e-0f4629b8b416 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.935846] env[65726]: DEBUG oslo_vmware.api [None req-9ed3c3d0-6db2-48b9-8065-95d9e297f871 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Task: {'id': task-5115381, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.504752} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 611.937883] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-9ed3c3d0-6db2-48b9-8065-95d9e297f871 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] 0e064341-4e4a-407b-8c26-3eb04b409029/0e064341-4e4a-407b-8c26-3eb04b409029.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 611.938159] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-9ed3c3d0-6db2-48b9-8065-95d9e297f871 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] [instance: 0e064341-4e4a-407b-8c26-3eb04b409029] Extending root virtual disk to 1048576 {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 611.942037] env[65726]: DEBUG oslo_vmware.api [None req-816382da-6d37-4962-9c02-92a18cd283e5 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Waiting for the task: (returnval){ [ 611.942037] env[65726]: value = "task-5115382" [ 611.942037] env[65726]: _type = "Task" [ 611.942037] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 611.942037] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-439be33f-ea4f-4f77-85dd-a2f29713d016 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.952422] env[65726]: DEBUG oslo_vmware.api [None req-af5fca6d-c57c-4762-93f4-4ba84e22ca7b tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]522da169-e62d-cd9c-7fce-a096337678c6, 'name': SearchDatastore_Task, 'duration_secs': 0.010021} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 611.952422] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d055d859-f0de-4f48-b229-b94cfcc49491 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.955738] env[65726]: DEBUG oslo_vmware.api [None req-9ed3c3d0-6db2-48b9-8065-95d9e297f871 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Waiting for the task: (returnval){ [ 611.955738] env[65726]: value = "task-5115383" [ 611.955738] env[65726]: _type = "Task" [ 611.955738] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 611.963554] env[65726]: DEBUG oslo_vmware.api [None req-af5fca6d-c57c-4762-93f4-4ba84e22ca7b tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Waiting for the task: (returnval){ [ 611.963554] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]526cc49c-6205-3925-82ff-428ac7edcc2b" [ 611.963554] env[65726]: _type = "Task" [ 611.963554] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 611.974762] env[65726]: DEBUG oslo_vmware.api [None req-816382da-6d37-4962-9c02-92a18cd283e5 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Task: {'id': task-5115382, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 611.975192] env[65726]: DEBUG oslo_vmware.api [None req-9ed3c3d0-6db2-48b9-8065-95d9e297f871 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Task: {'id': task-5115383, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 611.981566] env[65726]: DEBUG oslo_vmware.api [None req-af5fca6d-c57c-4762-93f4-4ba84e22ca7b tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]526cc49c-6205-3925-82ff-428ac7edcc2b, 'name': SearchDatastore_Task, 'duration_secs': 0.010971} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 611.981819] env[65726]: DEBUG oslo_concurrency.lockutils [None req-af5fca6d-c57c-4762-93f4-4ba84e22ca7b tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 611.982077] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-af5fca6d-c57c-4762-93f4-4ba84e22ca7b tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] 8e9c27f6-e9fa-4ed0-b6ae-3b5df52d1c44/8e9c27f6-e9fa-4ed0-b6ae-3b5df52d1c44.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 611.982343] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5c1fab1f-7856-43d0-84bd-904452799dc9 tempest-ServersTestJSON-1793637688 tempest-ServersTestJSON-1793637688-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 611.982572] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-5c1fab1f-7856-43d0-84bd-904452799dc9 tempest-ServersTestJSON-1793637688 tempest-ServersTestJSON-1793637688-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 611.983051] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0b24403d-6f40-4bf1-aca9-fa2173b5c9c0 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.985757] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4b603b0e-6502-4be5-b492-7184022392f0 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.994647] env[65726]: DEBUG oslo_vmware.api [None req-af5fca6d-c57c-4762-93f4-4ba84e22ca7b tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Waiting for the task: (returnval){ [ 611.994647] env[65726]: value = "task-5115384" [ 611.994647] env[65726]: _type = "Task" [ 611.994647] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 612.000659] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-5c1fab1f-7856-43d0-84bd-904452799dc9 tempest-ServersTestJSON-1793637688 tempest-ServersTestJSON-1793637688-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 612.000913] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-5c1fab1f-7856-43d0-84bd-904452799dc9 tempest-ServersTestJSON-1793637688 tempest-ServersTestJSON-1793637688-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65726) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 612.001807] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-569b1329-456a-498c-9634-4e9729566375 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.013155] env[65726]: DEBUG oslo_vmware.api [None req-5c1fab1f-7856-43d0-84bd-904452799dc9 tempest-ServersTestJSON-1793637688 tempest-ServersTestJSON-1793637688-project-member] Waiting for the task: (returnval){ [ 612.013155] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52a27988-ec2f-ed34-23af-93294d48d9ac" [ 612.013155] env[65726]: _type = "Task" [ 612.013155] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 612.013155] env[65726]: DEBUG oslo_vmware.api [None req-af5fca6d-c57c-4762-93f4-4ba84e22ca7b tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': task-5115384, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 612.023852] env[65726]: DEBUG oslo_vmware.api [None req-5c1fab1f-7856-43d0-84bd-904452799dc9 tempest-ServersTestJSON-1793637688 tempest-ServersTestJSON-1793637688-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52a27988-ec2f-ed34-23af-93294d48d9ac, 'name': SearchDatastore_Task, 'duration_secs': 0.011051} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 612.028026] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-13aff8cb-9bab-4309-8e37-f3f4f1837e36 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.028026] env[65726]: DEBUG oslo_concurrency.lockutils [req-593c5f97-a230-40c6-aec7-28e40819c5fb req-77be3017-6f29-4849-87a9-dc59653a8645 service nova] Releasing lock "refresh_cache-0e064341-4e4a-407b-8c26-3eb04b409029" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 612.028813] env[65726]: DEBUG nova.compute.manager [req-593c5f97-a230-40c6-aec7-28e40819c5fb req-77be3017-6f29-4849-87a9-dc59653a8645 service nova] [instance: 8e9c27f6-e9fa-4ed0-b6ae-3b5df52d1c44] Received event network-vif-plugged-cc1dce8f-441c-42a7-b3ba-7f3d3fc9ca66 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 612.028813] env[65726]: DEBUG oslo_concurrency.lockutils [req-593c5f97-a230-40c6-aec7-28e40819c5fb req-77be3017-6f29-4849-87a9-dc59653a8645 service nova] Acquiring lock "8e9c27f6-e9fa-4ed0-b6ae-3b5df52d1c44-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 612.028813] env[65726]: DEBUG oslo_concurrency.lockutils [req-593c5f97-a230-40c6-aec7-28e40819c5fb req-77be3017-6f29-4849-87a9-dc59653a8645 service nova] Lock "8e9c27f6-e9fa-4ed0-b6ae-3b5df52d1c44-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 612.028964] env[65726]: DEBUG oslo_concurrency.lockutils [req-593c5f97-a230-40c6-aec7-28e40819c5fb req-77be3017-6f29-4849-87a9-dc59653a8645 service nova] Lock "8e9c27f6-e9fa-4ed0-b6ae-3b5df52d1c44-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 612.029080] env[65726]: DEBUG nova.compute.manager [req-593c5f97-a230-40c6-aec7-28e40819c5fb req-77be3017-6f29-4849-87a9-dc59653a8645 service nova] [instance: 8e9c27f6-e9fa-4ed0-b6ae-3b5df52d1c44] No waiting events found dispatching network-vif-plugged-cc1dce8f-441c-42a7-b3ba-7f3d3fc9ca66 {{(pid=65726) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 612.029294] env[65726]: WARNING nova.compute.manager [req-593c5f97-a230-40c6-aec7-28e40819c5fb req-77be3017-6f29-4849-87a9-dc59653a8645 service nova] [instance: 8e9c27f6-e9fa-4ed0-b6ae-3b5df52d1c44] Received unexpected event network-vif-plugged-cc1dce8f-441c-42a7-b3ba-7f3d3fc9ca66 for instance with vm_state building and task_state spawning. [ 612.029469] env[65726]: DEBUG nova.compute.manager [req-593c5f97-a230-40c6-aec7-28e40819c5fb req-77be3017-6f29-4849-87a9-dc59653a8645 service nova] [instance: 8e9c27f6-e9fa-4ed0-b6ae-3b5df52d1c44] Received event network-changed-cc1dce8f-441c-42a7-b3ba-7f3d3fc9ca66 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 612.029629] env[65726]: DEBUG nova.compute.manager [req-593c5f97-a230-40c6-aec7-28e40819c5fb req-77be3017-6f29-4849-87a9-dc59653a8645 service nova] [instance: 8e9c27f6-e9fa-4ed0-b6ae-3b5df52d1c44] Refreshing instance network info cache due to event network-changed-cc1dce8f-441c-42a7-b3ba-7f3d3fc9ca66. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 612.029841] env[65726]: DEBUG oslo_concurrency.lockutils [req-593c5f97-a230-40c6-aec7-28e40819c5fb req-77be3017-6f29-4849-87a9-dc59653a8645 service nova] Acquiring lock "refresh_cache-8e9c27f6-e9fa-4ed0-b6ae-3b5df52d1c44" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 612.029988] env[65726]: DEBUG oslo_concurrency.lockutils [req-593c5f97-a230-40c6-aec7-28e40819c5fb req-77be3017-6f29-4849-87a9-dc59653a8645 service nova] Acquired lock "refresh_cache-8e9c27f6-e9fa-4ed0-b6ae-3b5df52d1c44" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 612.030155] env[65726]: DEBUG nova.network.neutron [req-593c5f97-a230-40c6-aec7-28e40819c5fb req-77be3017-6f29-4849-87a9-dc59653a8645 service nova] [instance: 8e9c27f6-e9fa-4ed0-b6ae-3b5df52d1c44] Refreshing network info cache for port cc1dce8f-441c-42a7-b3ba-7f3d3fc9ca66 {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 612.035762] env[65726]: DEBUG oslo_vmware.api [None req-5c1fab1f-7856-43d0-84bd-904452799dc9 tempest-ServersTestJSON-1793637688 tempest-ServersTestJSON-1793637688-project-member] Waiting for the task: (returnval){ [ 612.035762] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]529d74ce-7f84-4d10-defa-3511af4dec8f" [ 612.035762] env[65726]: _type = "Task" [ 612.035762] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 612.046233] env[65726]: DEBUG oslo_vmware.api [None req-5c1fab1f-7856-43d0-84bd-904452799dc9 tempest-ServersTestJSON-1793637688 tempest-ServersTestJSON-1793637688-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]529d74ce-7f84-4d10-defa-3511af4dec8f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 612.208428] env[65726]: DEBUG oslo_concurrency.lockutils [req-fccf0b77-e7bb-41aa-b6c5-117d34566265 req-34c11ed9-fede-4a19-967a-1b3c57bdf97e service nova] Releasing lock "refresh_cache-bb64fc41-19d7-49dc-a17a-9dd74730130f" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 612.310747] env[65726]: DEBUG oslo_concurrency.lockutils [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 5.538s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 612.310747] env[65726]: DEBUG nova.compute.manager [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] [instance: 703fff2d-5aec-4ed1-b987-9e1cc3a25f67] Start building networks asynchronously for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 612.315326] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9da01a84-79f6-4bc1-8b36-334b89c8de59 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.833s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 612.317034] env[65726]: INFO nova.compute.claims [None req-9da01a84-79f6-4bc1-8b36-334b89c8de59 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] [instance: ee676e3b-3326-46a3-940d-ebbb8b108991] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 612.454370] env[65726]: DEBUG oslo_vmware.api [None req-816382da-6d37-4962-9c02-92a18cd283e5 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Task: {'id': task-5115382, 'name': ReconfigVM_Task, 'duration_secs': 0.34331} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 612.454500] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-816382da-6d37-4962-9c02-92a18cd283e5 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] [instance: bb64fc41-19d7-49dc-a17a-9dd74730130f] Reconfigured VM instance instance-00000001 to attach disk [datastore1] bb64fc41-19d7-49dc-a17a-9dd74730130f/bb64fc41-19d7-49dc-a17a-9dd74730130f.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 612.455266] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-09076a45-e6f5-414c-9de4-b5a64bc569bf {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.463670] env[65726]: DEBUG oslo_vmware.api [None req-816382da-6d37-4962-9c02-92a18cd283e5 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Waiting for the task: (returnval){ [ 612.463670] env[65726]: value = "task-5115385" [ 612.463670] env[65726]: _type = "Task" [ 612.463670] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 612.470200] env[65726]: DEBUG oslo_vmware.api [None req-9ed3c3d0-6db2-48b9-8065-95d9e297f871 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Task: {'id': task-5115383, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.076213} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 612.470870] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-9ed3c3d0-6db2-48b9-8065-95d9e297f871 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] [instance: 0e064341-4e4a-407b-8c26-3eb04b409029] Extended root virtual disk {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 612.471970] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac16b544-0f64-48b6-b281-1cecf7f52316 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.479623] env[65726]: DEBUG oslo_vmware.api [None req-816382da-6d37-4962-9c02-92a18cd283e5 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Task: {'id': task-5115385, 'name': Rename_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 612.503942] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-9ed3c3d0-6db2-48b9-8065-95d9e297f871 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] [instance: 0e064341-4e4a-407b-8c26-3eb04b409029] Reconfiguring VM instance instance-00000002 to attach disk [datastore1] 0e064341-4e4a-407b-8c26-3eb04b409029/0e064341-4e4a-407b-8c26-3eb04b409029.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 612.505805] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-801cd53c-a5ed-4f2b-b6c4-f7e8d651c03a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.536685] env[65726]: WARNING openstack [req-593c5f97-a230-40c6-aec7-28e40819c5fb req-77be3017-6f29-4849-87a9-dc59653a8645 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 612.537077] env[65726]: WARNING openstack [req-593c5f97-a230-40c6-aec7-28e40819c5fb req-77be3017-6f29-4849-87a9-dc59653a8645 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 612.543671] env[65726]: DEBUG oslo_vmware.api [None req-af5fca6d-c57c-4762-93f4-4ba84e22ca7b tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': task-5115384, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 612.543967] env[65726]: DEBUG oslo_vmware.api [None req-9ed3c3d0-6db2-48b9-8065-95d9e297f871 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Waiting for the task: (returnval){ [ 612.543967] env[65726]: value = "task-5115386" [ 612.543967] env[65726]: _type = "Task" [ 612.543967] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 612.567113] env[65726]: DEBUG oslo_vmware.api [None req-5c1fab1f-7856-43d0-84bd-904452799dc9 tempest-ServersTestJSON-1793637688 tempest-ServersTestJSON-1793637688-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]529d74ce-7f84-4d10-defa-3511af4dec8f, 'name': SearchDatastore_Task, 'duration_secs': 0.009769} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 612.567348] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5c1fab1f-7856-43d0-84bd-904452799dc9 tempest-ServersTestJSON-1793637688 tempest-ServersTestJSON-1793637688-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 612.567607] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-5c1fab1f-7856-43d0-84bd-904452799dc9 tempest-ServersTestJSON-1793637688 tempest-ServersTestJSON-1793637688-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] 88fb7235-cd0e-49dc-9d78-3583d5d3f528/88fb7235-cd0e-49dc-9d78-3583d5d3f528.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 612.567882] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b555aa90-e282-4dea-b706-a91ebaff2031 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.581491] env[65726]: DEBUG oslo_vmware.api [None req-5c1fab1f-7856-43d0-84bd-904452799dc9 tempest-ServersTestJSON-1793637688 tempest-ServersTestJSON-1793637688-project-member] Waiting for the task: (returnval){ [ 612.581491] env[65726]: value = "task-5115387" [ 612.581491] env[65726]: _type = "Task" [ 612.581491] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 612.594930] env[65726]: DEBUG oslo_vmware.api [None req-5c1fab1f-7856-43d0-84bd-904452799dc9 tempest-ServersTestJSON-1793637688 tempest-ServersTestJSON-1793637688-project-member] Task: {'id': task-5115387, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 612.747201] env[65726]: WARNING openstack [req-593c5f97-a230-40c6-aec7-28e40819c5fb req-77be3017-6f29-4849-87a9-dc59653a8645 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 612.748852] env[65726]: WARNING openstack [req-593c5f97-a230-40c6-aec7-28e40819c5fb req-77be3017-6f29-4849-87a9-dc59653a8645 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 612.822701] env[65726]: DEBUG nova.compute.utils [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Using /dev/sd instead of None {{(pid=65726) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 612.826110] env[65726]: DEBUG nova.compute.manager [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] [instance: 703fff2d-5aec-4ed1-b987-9e1cc3a25f67] Allocating IP information in the background. {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 612.826330] env[65726]: DEBUG nova.network.neutron [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] [instance: 703fff2d-5aec-4ed1-b987-9e1cc3a25f67] allocate_for_instance() {{(pid=65726) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 612.826646] env[65726]: WARNING neutronclient.v2_0.client [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 612.826993] env[65726]: WARNING neutronclient.v2_0.client [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 612.827918] env[65726]: WARNING openstack [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 612.828397] env[65726]: WARNING openstack [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 612.843333] env[65726]: WARNING openstack [req-593c5f97-a230-40c6-aec7-28e40819c5fb req-77be3017-6f29-4849-87a9-dc59653a8645 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 612.843333] env[65726]: WARNING openstack [req-593c5f97-a230-40c6-aec7-28e40819c5fb req-77be3017-6f29-4849-87a9-dc59653a8645 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 612.900893] env[65726]: DEBUG nova.policy [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a78a1cdb82644c1abe2b0ab2e0f621a5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3a680eae0e734ea8b2274b504d0a46d8', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65726) authorize /opt/stack/nova/nova/policy.py:192}} [ 612.941679] env[65726]: DEBUG nova.network.neutron [req-593c5f97-a230-40c6-aec7-28e40819c5fb req-77be3017-6f29-4849-87a9-dc59653a8645 service nova] [instance: 8e9c27f6-e9fa-4ed0-b6ae-3b5df52d1c44] Updated VIF entry in instance network info cache for port cc1dce8f-441c-42a7-b3ba-7f3d3fc9ca66. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 612.942805] env[65726]: DEBUG nova.network.neutron [req-593c5f97-a230-40c6-aec7-28e40819c5fb req-77be3017-6f29-4849-87a9-dc59653a8645 service nova] [instance: 8e9c27f6-e9fa-4ed0-b6ae-3b5df52d1c44] Updating instance_info_cache with network_info: [{"id": "cc1dce8f-441c-42a7-b3ba-7f3d3fc9ca66", "address": "fa:16:3e:8b:2a:82", "network": {"id": "e206b5b9-da36-4e6f-926f-914dec38fac2", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1053201655-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "39067912d091464192516c08eb576696", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7c80243e-93a7-4a95-bc8d-e9534bacd66e", "external-id": "nsx-vlan-transportzone-306", "segmentation_id": 306, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcc1dce8f-44", "ovs_interfaceid": "cc1dce8f-441c-42a7-b3ba-7f3d3fc9ca66", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 612.977526] env[65726]: DEBUG oslo_vmware.api [None req-816382da-6d37-4962-9c02-92a18cd283e5 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Task: {'id': task-5115385, 'name': Rename_Task, 'duration_secs': 0.159018} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 612.977805] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-816382da-6d37-4962-9c02-92a18cd283e5 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] [instance: bb64fc41-19d7-49dc-a17a-9dd74730130f] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 612.978543] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6d2d1f21-e150-4ca3-8209-e64790537313 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.989157] env[65726]: DEBUG oslo_vmware.api [None req-816382da-6d37-4962-9c02-92a18cd283e5 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Waiting for the task: (returnval){ [ 612.989157] env[65726]: value = "task-5115388" [ 612.989157] env[65726]: _type = "Task" [ 612.989157] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 613.000654] env[65726]: DEBUG oslo_vmware.api [None req-816382da-6d37-4962-9c02-92a18cd283e5 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Task: {'id': task-5115388, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 613.021272] env[65726]: DEBUG oslo_vmware.api [None req-af5fca6d-c57c-4762-93f4-4ba84e22ca7b tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': task-5115384, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 613.058620] env[65726]: DEBUG oslo_vmware.api [None req-9ed3c3d0-6db2-48b9-8065-95d9e297f871 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Task: {'id': task-5115386, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 613.103528] env[65726]: DEBUG oslo_vmware.api [None req-5c1fab1f-7856-43d0-84bd-904452799dc9 tempest-ServersTestJSON-1793637688 tempest-ServersTestJSON-1793637688-project-member] Task: {'id': task-5115387, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 613.281084] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Acquiring lock "811c1bbd-259c-4f5e-9a61-581bfba0ec4d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 613.281230] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Lock "811c1bbd-259c-4f5e-9a61-581bfba0ec4d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 613.308184] env[65726]: DEBUG nova.network.neutron [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] [instance: 703fff2d-5aec-4ed1-b987-9e1cc3a25f67] Successfully created port: 29330b5b-6171-4de8-bf77-6c57f6a8bd58 {{(pid=65726) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 613.329608] env[65726]: DEBUG nova.compute.manager [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] [instance: 703fff2d-5aec-4ed1-b987-9e1cc3a25f67] Start building block device mappings for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 613.335614] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Acquiring lock "3a60b719-9db7-436e-9908-25f6a233c465" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 613.335941] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Lock "3a60b719-9db7-436e-9908-25f6a233c465" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 613.418430] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Acquiring lock "f5151062-57b1-4e4e-93f4-aab0e7f504d4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 613.420056] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Lock "f5151062-57b1-4e4e-93f4-aab0e7f504d4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 613.446832] env[65726]: DEBUG oslo_concurrency.lockutils [req-593c5f97-a230-40c6-aec7-28e40819c5fb req-77be3017-6f29-4849-87a9-dc59653a8645 service nova] Releasing lock "refresh_cache-8e9c27f6-e9fa-4ed0-b6ae-3b5df52d1c44" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 613.447391] env[65726]: DEBUG nova.compute.manager [req-593c5f97-a230-40c6-aec7-28e40819c5fb req-77be3017-6f29-4849-87a9-dc59653a8645 service nova] [instance: 88fb7235-cd0e-49dc-9d78-3583d5d3f528] Received event network-vif-plugged-b915a591-ee09-4390-ae01-09778bd3f2c3 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 613.447802] env[65726]: DEBUG oslo_concurrency.lockutils [req-593c5f97-a230-40c6-aec7-28e40819c5fb req-77be3017-6f29-4849-87a9-dc59653a8645 service nova] Acquiring lock "88fb7235-cd0e-49dc-9d78-3583d5d3f528-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 613.450751] env[65726]: DEBUG oslo_concurrency.lockutils [req-593c5f97-a230-40c6-aec7-28e40819c5fb req-77be3017-6f29-4849-87a9-dc59653a8645 service nova] Lock "88fb7235-cd0e-49dc-9d78-3583d5d3f528-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 613.450751] env[65726]: DEBUG oslo_concurrency.lockutils [req-593c5f97-a230-40c6-aec7-28e40819c5fb req-77be3017-6f29-4849-87a9-dc59653a8645 service nova] Lock "88fb7235-cd0e-49dc-9d78-3583d5d3f528-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 613.450751] env[65726]: DEBUG nova.compute.manager [req-593c5f97-a230-40c6-aec7-28e40819c5fb req-77be3017-6f29-4849-87a9-dc59653a8645 service nova] [instance: 88fb7235-cd0e-49dc-9d78-3583d5d3f528] No waiting events found dispatching network-vif-plugged-b915a591-ee09-4390-ae01-09778bd3f2c3 {{(pid=65726) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 613.450751] env[65726]: WARNING nova.compute.manager [req-593c5f97-a230-40c6-aec7-28e40819c5fb req-77be3017-6f29-4849-87a9-dc59653a8645 service nova] [instance: 88fb7235-cd0e-49dc-9d78-3583d5d3f528] Received unexpected event network-vif-plugged-b915a591-ee09-4390-ae01-09778bd3f2c3 for instance with vm_state building and task_state spawning. [ 613.450751] env[65726]: DEBUG nova.compute.manager [req-593c5f97-a230-40c6-aec7-28e40819c5fb req-77be3017-6f29-4849-87a9-dc59653a8645 service nova] [instance: 88fb7235-cd0e-49dc-9d78-3583d5d3f528] Received event network-changed-b915a591-ee09-4390-ae01-09778bd3f2c3 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 613.450947] env[65726]: DEBUG nova.compute.manager [req-593c5f97-a230-40c6-aec7-28e40819c5fb req-77be3017-6f29-4849-87a9-dc59653a8645 service nova] [instance: 88fb7235-cd0e-49dc-9d78-3583d5d3f528] Refreshing instance network info cache due to event network-changed-b915a591-ee09-4390-ae01-09778bd3f2c3. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 613.450947] env[65726]: DEBUG oslo_concurrency.lockutils [req-593c5f97-a230-40c6-aec7-28e40819c5fb req-77be3017-6f29-4849-87a9-dc59653a8645 service nova] Acquiring lock "refresh_cache-88fb7235-cd0e-49dc-9d78-3583d5d3f528" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 613.450947] env[65726]: DEBUG oslo_concurrency.lockutils [req-593c5f97-a230-40c6-aec7-28e40819c5fb req-77be3017-6f29-4849-87a9-dc59653a8645 service nova] Acquired lock "refresh_cache-88fb7235-cd0e-49dc-9d78-3583d5d3f528" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 613.450947] env[65726]: DEBUG nova.network.neutron [req-593c5f97-a230-40c6-aec7-28e40819c5fb req-77be3017-6f29-4849-87a9-dc59653a8645 service nova] [instance: 88fb7235-cd0e-49dc-9d78-3583d5d3f528] Refreshing network info cache for port b915a591-ee09-4390-ae01-09778bd3f2c3 {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 613.521317] env[65726]: DEBUG oslo_vmware.api [None req-816382da-6d37-4962-9c02-92a18cd283e5 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Task: {'id': task-5115388, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 613.536235] env[65726]: DEBUG oslo_vmware.api [None req-af5fca6d-c57c-4762-93f4-4ba84e22ca7b tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': task-5115384, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 613.564141] env[65726]: DEBUG oslo_vmware.api [None req-9ed3c3d0-6db2-48b9-8065-95d9e297f871 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Task: {'id': task-5115386, 'name': ReconfigVM_Task, 'duration_secs': 0.913182} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 613.564844] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-9ed3c3d0-6db2-48b9-8065-95d9e297f871 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] [instance: 0e064341-4e4a-407b-8c26-3eb04b409029] Reconfigured VM instance instance-00000002 to attach disk [datastore1] 0e064341-4e4a-407b-8c26-3eb04b409029/0e064341-4e4a-407b-8c26-3eb04b409029.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 613.565532] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-267e7bca-7b8d-4e1a-a37a-19e6b8b191c6 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.575466] env[65726]: DEBUG oslo_vmware.api [None req-9ed3c3d0-6db2-48b9-8065-95d9e297f871 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Waiting for the task: (returnval){ [ 613.575466] env[65726]: value = "task-5115390" [ 613.575466] env[65726]: _type = "Task" [ 613.575466] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 613.595570] env[65726]: DEBUG oslo_vmware.api [None req-9ed3c3d0-6db2-48b9-8065-95d9e297f871 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Task: {'id': task-5115390, 'name': Rename_Task} progress is 6%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 613.608888] env[65726]: DEBUG oslo_vmware.api [None req-5c1fab1f-7856-43d0-84bd-904452799dc9 tempest-ServersTestJSON-1793637688 tempest-ServersTestJSON-1793637688-project-member] Task: {'id': task-5115387, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 613.727733] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9880d939-39d6-4d38-b59d-fdb561c24364 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.740231] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1eaa5c0-a47d-47ae-ba1b-842c381c79ba {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.775239] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-569cb840-b5af-42c8-848a-12d4ab33a444 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.785129] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ed8bed5-d137-4e67-9258-eba3c5cc2106 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.800184] env[65726]: DEBUG nova.compute.provider_tree [None req-9da01a84-79f6-4bc1-8b36-334b89c8de59 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 613.953436] env[65726]: WARNING openstack [req-593c5f97-a230-40c6-aec7-28e40819c5fb req-77be3017-6f29-4849-87a9-dc59653a8645 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 613.954511] env[65726]: WARNING openstack [req-593c5f97-a230-40c6-aec7-28e40819c5fb req-77be3017-6f29-4849-87a9-dc59653a8645 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 614.003990] env[65726]: DEBUG oslo_vmware.api [None req-816382da-6d37-4962-9c02-92a18cd283e5 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Task: {'id': task-5115388, 'name': PowerOnVM_Task, 'duration_secs': 0.662281} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 614.003990] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-816382da-6d37-4962-9c02-92a18cd283e5 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] [instance: bb64fc41-19d7-49dc-a17a-9dd74730130f] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 614.003990] env[65726]: INFO nova.compute.manager [None req-816382da-6d37-4962-9c02-92a18cd283e5 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] [instance: bb64fc41-19d7-49dc-a17a-9dd74730130f] Took 14.63 seconds to spawn the instance on the hypervisor. [ 614.003990] env[65726]: DEBUG nova.compute.manager [None req-816382da-6d37-4962-9c02-92a18cd283e5 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] [instance: bb64fc41-19d7-49dc-a17a-9dd74730130f] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 614.003990] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-845a793b-401a-409a-9ba0-3e39f87a196e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.033035] env[65726]: DEBUG oslo_vmware.api [None req-af5fca6d-c57c-4762-93f4-4ba84e22ca7b tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': task-5115384, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.62154} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 614.033625] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-af5fca6d-c57c-4762-93f4-4ba84e22ca7b tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] 8e9c27f6-e9fa-4ed0-b6ae-3b5df52d1c44/8e9c27f6-e9fa-4ed0-b6ae-3b5df52d1c44.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 614.033926] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-af5fca6d-c57c-4762-93f4-4ba84e22ca7b tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 8e9c27f6-e9fa-4ed0-b6ae-3b5df52d1c44] Extending root virtual disk to 1048576 {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 614.037406] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-62f60a37-1844-4e2b-980c-efb61a1757cf {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.042636] env[65726]: DEBUG oslo_vmware.api [None req-af5fca6d-c57c-4762-93f4-4ba84e22ca7b tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Waiting for the task: (returnval){ [ 614.042636] env[65726]: value = "task-5115391" [ 614.042636] env[65726]: _type = "Task" [ 614.042636] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 614.054022] env[65726]: DEBUG oslo_vmware.api [None req-af5fca6d-c57c-4762-93f4-4ba84e22ca7b tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': task-5115391, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 614.097068] env[65726]: DEBUG oslo_vmware.api [None req-9ed3c3d0-6db2-48b9-8065-95d9e297f871 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Task: {'id': task-5115390, 'name': Rename_Task, 'duration_secs': 0.164329} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 614.100047] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-9ed3c3d0-6db2-48b9-8065-95d9e297f871 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] [instance: 0e064341-4e4a-407b-8c26-3eb04b409029] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 614.100406] env[65726]: DEBUG oslo_vmware.api [None req-5c1fab1f-7856-43d0-84bd-904452799dc9 tempest-ServersTestJSON-1793637688 tempest-ServersTestJSON-1793637688-project-member] Task: {'id': task-5115387, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.032302} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 614.100722] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2d0c532e-8062-47f4-9710-2586d562194d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.102391] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-5c1fab1f-7856-43d0-84bd-904452799dc9 tempest-ServersTestJSON-1793637688 tempest-ServersTestJSON-1793637688-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] 88fb7235-cd0e-49dc-9d78-3583d5d3f528/88fb7235-cd0e-49dc-9d78-3583d5d3f528.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 614.102588] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-5c1fab1f-7856-43d0-84bd-904452799dc9 tempest-ServersTestJSON-1793637688 tempest-ServersTestJSON-1793637688-project-member] [instance: 88fb7235-cd0e-49dc-9d78-3583d5d3f528] Extending root virtual disk to 1048576 {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 614.103027] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-eb1c91fb-c1e1-4d76-8556-ae140a95b8bb {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.113973] env[65726]: DEBUG oslo_vmware.api [None req-9ed3c3d0-6db2-48b9-8065-95d9e297f871 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Waiting for the task: (returnval){ [ 614.113973] env[65726]: value = "task-5115392" [ 614.113973] env[65726]: _type = "Task" [ 614.113973] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 614.114410] env[65726]: DEBUG oslo_vmware.api [None req-5c1fab1f-7856-43d0-84bd-904452799dc9 tempest-ServersTestJSON-1793637688 tempest-ServersTestJSON-1793637688-project-member] Waiting for the task: (returnval){ [ 614.114410] env[65726]: value = "task-5115393" [ 614.114410] env[65726]: _type = "Task" [ 614.114410] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 614.133376] env[65726]: DEBUG oslo_vmware.api [None req-5c1fab1f-7856-43d0-84bd-904452799dc9 tempest-ServersTestJSON-1793637688 tempest-ServersTestJSON-1793637688-project-member] Task: {'id': task-5115393, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 614.136558] env[65726]: DEBUG oslo_vmware.api [None req-9ed3c3d0-6db2-48b9-8065-95d9e297f871 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Task: {'id': task-5115392, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 614.175800] env[65726]: WARNING openstack [req-593c5f97-a230-40c6-aec7-28e40819c5fb req-77be3017-6f29-4849-87a9-dc59653a8645 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 614.176210] env[65726]: WARNING openstack [req-593c5f97-a230-40c6-aec7-28e40819c5fb req-77be3017-6f29-4849-87a9-dc59653a8645 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 614.276169] env[65726]: WARNING openstack [req-593c5f97-a230-40c6-aec7-28e40819c5fb req-77be3017-6f29-4849-87a9-dc59653a8645 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 614.276169] env[65726]: WARNING openstack [req-593c5f97-a230-40c6-aec7-28e40819c5fb req-77be3017-6f29-4849-87a9-dc59653a8645 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 614.303810] env[65726]: DEBUG nova.scheduler.client.report [None req-9da01a84-79f6-4bc1-8b36-334b89c8de59 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 614.341456] env[65726]: DEBUG nova.compute.manager [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] [instance: 703fff2d-5aec-4ed1-b987-9e1cc3a25f67] Start spawning the instance on the hypervisor. {{(pid=65726) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 614.370367] env[65726]: DEBUG nova.virt.hardware [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T19:28:34Z,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1539f06fa3534e90acbc3a60c4b8bd3f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T19:28:35Z,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 614.370367] env[65726]: DEBUG nova.virt.hardware [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 614.370367] env[65726]: DEBUG nova.virt.hardware [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 614.370568] env[65726]: DEBUG nova.virt.hardware [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 614.370568] env[65726]: DEBUG nova.virt.hardware [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 614.370568] env[65726]: DEBUG nova.virt.hardware [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 614.370568] env[65726]: DEBUG nova.virt.hardware [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 614.370568] env[65726]: DEBUG nova.virt.hardware [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 614.370737] env[65726]: DEBUG nova.virt.hardware [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 614.370737] env[65726]: DEBUG nova.virt.hardware [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 614.370929] env[65726]: DEBUG nova.virt.hardware [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 614.371816] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12d360ac-396d-4b41-8c64-ce154391c0e4 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.381024] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82c710e2-1be2-4948-bfab-c1626442e11f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.534542] env[65726]: INFO nova.compute.manager [None req-816382da-6d37-4962-9c02-92a18cd283e5 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] [instance: bb64fc41-19d7-49dc-a17a-9dd74730130f] Took 21.17 seconds to build instance. [ 614.561638] env[65726]: DEBUG oslo_vmware.api [None req-af5fca6d-c57c-4762-93f4-4ba84e22ca7b tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': task-5115391, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.10209} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 614.562538] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-af5fca6d-c57c-4762-93f4-4ba84e22ca7b tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 8e9c27f6-e9fa-4ed0-b6ae-3b5df52d1c44] Extended root virtual disk {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 614.563680] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84ef7ef4-3bdf-46b0-9837-e7bd6280e7df {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.588949] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-af5fca6d-c57c-4762-93f4-4ba84e22ca7b tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 8e9c27f6-e9fa-4ed0-b6ae-3b5df52d1c44] Reconfiguring VM instance instance-00000003 to attach disk [datastore1] 8e9c27f6-e9fa-4ed0-b6ae-3b5df52d1c44/8e9c27f6-e9fa-4ed0-b6ae-3b5df52d1c44.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 614.589514] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-90e334ce-1d0a-41dc-9d4b-2ab85391563a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.610170] env[65726]: DEBUG oslo_vmware.api [None req-af5fca6d-c57c-4762-93f4-4ba84e22ca7b tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Waiting for the task: (returnval){ [ 614.610170] env[65726]: value = "task-5115394" [ 614.610170] env[65726]: _type = "Task" [ 614.610170] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 614.624749] env[65726]: DEBUG oslo_vmware.api [None req-af5fca6d-c57c-4762-93f4-4ba84e22ca7b tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': task-5115394, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 614.630484] env[65726]: DEBUG oslo_vmware.api [None req-5c1fab1f-7856-43d0-84bd-904452799dc9 tempest-ServersTestJSON-1793637688 tempest-ServersTestJSON-1793637688-project-member] Task: {'id': task-5115393, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.16059} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 614.633666] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-5c1fab1f-7856-43d0-84bd-904452799dc9 tempest-ServersTestJSON-1793637688 tempest-ServersTestJSON-1793637688-project-member] [instance: 88fb7235-cd0e-49dc-9d78-3583d5d3f528] Extended root virtual disk {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 614.633997] env[65726]: DEBUG oslo_vmware.api [None req-9ed3c3d0-6db2-48b9-8065-95d9e297f871 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Task: {'id': task-5115392, 'name': PowerOnVM_Task, 'duration_secs': 0.501867} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 614.635927] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7df10027-1859-4cc8-9c2d-97f93bdc0b79 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.638687] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-9ed3c3d0-6db2-48b9-8065-95d9e297f871 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] [instance: 0e064341-4e4a-407b-8c26-3eb04b409029] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 614.638951] env[65726]: INFO nova.compute.manager [None req-9ed3c3d0-6db2-48b9-8065-95d9e297f871 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] [instance: 0e064341-4e4a-407b-8c26-3eb04b409029] Took 17.02 seconds to spawn the instance on the hypervisor. [ 614.639178] env[65726]: DEBUG nova.compute.manager [None req-9ed3c3d0-6db2-48b9-8065-95d9e297f871 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] [instance: 0e064341-4e4a-407b-8c26-3eb04b409029] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 614.643579] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3bd9e5c5-9855-4e7a-b6e8-027e5334d203 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.646232] env[65726]: DEBUG oslo_concurrency.lockutils [None req-881e6c53-2eb2-4fe8-9bf1-92c962897b3f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Acquiring lock "f2d9090c-988f-43f4-9c81-7aa718a3438a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 614.646443] env[65726]: DEBUG oslo_concurrency.lockutils [None req-881e6c53-2eb2-4fe8-9bf1-92c962897b3f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Lock "f2d9090c-988f-43f4-9c81-7aa718a3438a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 614.667805] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-5c1fab1f-7856-43d0-84bd-904452799dc9 tempest-ServersTestJSON-1793637688 tempest-ServersTestJSON-1793637688-project-member] [instance: 88fb7235-cd0e-49dc-9d78-3583d5d3f528] Reconfiguring VM instance instance-00000004 to attach disk [datastore1] 88fb7235-cd0e-49dc-9d78-3583d5d3f528/88fb7235-cd0e-49dc-9d78-3583d5d3f528.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 614.670218] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6738b0a0-196e-4bb2-8cc1-6959ff054427 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.694107] env[65726]: DEBUG oslo_vmware.api [None req-5c1fab1f-7856-43d0-84bd-904452799dc9 tempest-ServersTestJSON-1793637688 tempest-ServersTestJSON-1793637688-project-member] Waiting for the task: (returnval){ [ 614.694107] env[65726]: value = "task-5115395" [ 614.694107] env[65726]: _type = "Task" [ 614.694107] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 614.703504] env[65726]: DEBUG oslo_vmware.api [None req-5c1fab1f-7856-43d0-84bd-904452799dc9 tempest-ServersTestJSON-1793637688 tempest-ServersTestJSON-1793637688-project-member] Task: {'id': task-5115395, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 614.809895] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9da01a84-79f6-4bc1-8b36-334b89c8de59 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.495s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 614.810737] env[65726]: DEBUG nova.compute.manager [None req-9da01a84-79f6-4bc1-8b36-334b89c8de59 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] [instance: ee676e3b-3326-46a3-940d-ebbb8b108991] Start building networks asynchronously for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 614.813257] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 8.400s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 614.814676] env[65726]: INFO nova.compute.claims [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] [instance: 8493ba52-b950-4727-970a-19b1797b2ca1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 614.848301] env[65726]: DEBUG nova.network.neutron [req-593c5f97-a230-40c6-aec7-28e40819c5fb req-77be3017-6f29-4849-87a9-dc59653a8645 service nova] [instance: 88fb7235-cd0e-49dc-9d78-3583d5d3f528] Updated VIF entry in instance network info cache for port b915a591-ee09-4390-ae01-09778bd3f2c3. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 614.848514] env[65726]: DEBUG nova.network.neutron [req-593c5f97-a230-40c6-aec7-28e40819c5fb req-77be3017-6f29-4849-87a9-dc59653a8645 service nova] [instance: 88fb7235-cd0e-49dc-9d78-3583d5d3f528] Updating instance_info_cache with network_info: [{"id": "b915a591-ee09-4390-ae01-09778bd3f2c3", "address": "fa:16:3e:e3:6b:b7", "network": {"id": "ae04280a-5cd5-46c3-a6f2-8020d38faea4", "bridge": "br-int", "label": "tempest-ServersTestJSON-1476935398-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "53f76b7cd812445da88b09acb430da70", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "49b5df12-d801-4140-8816-2fd401608c7d", "external-id": "nsx-vlan-transportzone-326", "segmentation_id": 326, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb915a591-ee", "ovs_interfaceid": "b915a591-ee09-4390-ae01-09778bd3f2c3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 615.013937] env[65726]: DEBUG nova.network.neutron [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] [instance: 703fff2d-5aec-4ed1-b987-9e1cc3a25f67] Successfully updated port: 29330b5b-6171-4de8-bf77-6c57f6a8bd58 {{(pid=65726) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 615.038277] env[65726]: DEBUG oslo_concurrency.lockutils [None req-816382da-6d37-4962-9c02-92a18cd283e5 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Lock "bb64fc41-19d7-49dc-a17a-9dd74730130f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 22.681s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 615.121530] env[65726]: DEBUG oslo_vmware.api [None req-af5fca6d-c57c-4762-93f4-4ba84e22ca7b tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': task-5115394, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 615.204727] env[65726]: DEBUG oslo_vmware.api [None req-5c1fab1f-7856-43d0-84bd-904452799dc9 tempest-ServersTestJSON-1793637688 tempest-ServersTestJSON-1793637688-project-member] Task: {'id': task-5115395, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 615.206934] env[65726]: INFO nova.compute.manager [None req-9ed3c3d0-6db2-48b9-8065-95d9e297f871 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] [instance: 0e064341-4e4a-407b-8c26-3eb04b409029] Took 21.88 seconds to build instance. [ 615.319407] env[65726]: DEBUG nova.compute.utils [None req-9da01a84-79f6-4bc1-8b36-334b89c8de59 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Using /dev/sd instead of None {{(pid=65726) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 615.324288] env[65726]: DEBUG nova.compute.manager [None req-9da01a84-79f6-4bc1-8b36-334b89c8de59 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] [instance: ee676e3b-3326-46a3-940d-ebbb8b108991] Allocating IP information in the background. {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 615.324633] env[65726]: DEBUG nova.network.neutron [None req-9da01a84-79f6-4bc1-8b36-334b89c8de59 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] [instance: ee676e3b-3326-46a3-940d-ebbb8b108991] allocate_for_instance() {{(pid=65726) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 615.325188] env[65726]: WARNING neutronclient.v2_0.client [None req-9da01a84-79f6-4bc1-8b36-334b89c8de59 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 615.326365] env[65726]: WARNING neutronclient.v2_0.client [None req-9da01a84-79f6-4bc1-8b36-334b89c8de59 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 615.327059] env[65726]: WARNING openstack [None req-9da01a84-79f6-4bc1-8b36-334b89c8de59 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 615.327148] env[65726]: WARNING openstack [None req-9da01a84-79f6-4bc1-8b36-334b89c8de59 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 615.352783] env[65726]: DEBUG oslo_concurrency.lockutils [req-593c5f97-a230-40c6-aec7-28e40819c5fb req-77be3017-6f29-4849-87a9-dc59653a8645 service nova] Releasing lock "refresh_cache-88fb7235-cd0e-49dc-9d78-3583d5d3f528" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 615.450880] env[65726]: DEBUG nova.policy [None req-9da01a84-79f6-4bc1-8b36-334b89c8de59 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e13d7ce17e914f49ac5f47aba8de2fc1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '27fdee694977439d88ed2b08abb28405', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65726) authorize /opt/stack/nova/nova/policy.py:192}} [ 615.517428] env[65726]: DEBUG oslo_concurrency.lockutils [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Acquiring lock "refresh_cache-703fff2d-5aec-4ed1-b987-9e1cc3a25f67" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 615.517656] env[65726]: DEBUG oslo_concurrency.lockutils [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Acquired lock "refresh_cache-703fff2d-5aec-4ed1-b987-9e1cc3a25f67" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 615.517790] env[65726]: DEBUG nova.network.neutron [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] [instance: 703fff2d-5aec-4ed1-b987-9e1cc3a25f67] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 615.542530] env[65726]: DEBUG nova.compute.manager [None req-b8e65ecf-6b31-4881-a7b5-2f0a252ee6d8 tempest-VolumesAssistedSnapshotsTest-1875841935 tempest-VolumesAssistedSnapshotsTest-1875841935-project-member] [instance: aa2716a9-2cf8-4e51-800d-1c8f98b2cf05] Starting instance... {{(pid=65726) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 615.621998] env[65726]: DEBUG oslo_vmware.api [None req-af5fca6d-c57c-4762-93f4-4ba84e22ca7b tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': task-5115394, 'name': ReconfigVM_Task, 'duration_secs': 0.693776} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 615.622390] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-af5fca6d-c57c-4762-93f4-4ba84e22ca7b tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 8e9c27f6-e9fa-4ed0-b6ae-3b5df52d1c44] Reconfigured VM instance instance-00000003 to attach disk [datastore1] 8e9c27f6-e9fa-4ed0-b6ae-3b5df52d1c44/8e9c27f6-e9fa-4ed0-b6ae-3b5df52d1c44.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 615.623128] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c66cb435-25d0-4bbf-9fd2-22890333ad8b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.630794] env[65726]: DEBUG oslo_vmware.api [None req-af5fca6d-c57c-4762-93f4-4ba84e22ca7b tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Waiting for the task: (returnval){ [ 615.630794] env[65726]: value = "task-5115396" [ 615.630794] env[65726]: _type = "Task" [ 615.630794] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 615.642333] env[65726]: DEBUG oslo_vmware.api [None req-af5fca6d-c57c-4762-93f4-4ba84e22ca7b tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': task-5115396, 'name': Rename_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 615.708902] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9ed3c3d0-6db2-48b9-8065-95d9e297f871 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Lock "0e064341-4e4a-407b-8c26-3eb04b409029" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 23.390s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 615.709269] env[65726]: DEBUG oslo_vmware.api [None req-5c1fab1f-7856-43d0-84bd-904452799dc9 tempest-ServersTestJSON-1793637688 tempest-ServersTestJSON-1793637688-project-member] Task: {'id': task-5115395, 'name': ReconfigVM_Task, 'duration_secs': 0.706219} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 615.710694] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Lock "0e064341-4e4a-407b-8c26-3eb04b409029" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 19.929s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 615.711114] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-5c1fab1f-7856-43d0-84bd-904452799dc9 tempest-ServersTestJSON-1793637688 tempest-ServersTestJSON-1793637688-project-member] [instance: 88fb7235-cd0e-49dc-9d78-3583d5d3f528] Reconfigured VM instance instance-00000004 to attach disk [datastore1] 88fb7235-cd0e-49dc-9d78-3583d5d3f528/88fb7235-cd0e-49dc-9d78-3583d5d3f528.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 615.715018] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1002a444-8dbc-41d6-9c11-8668b5dbd359 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.715472] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7c1968eb-402e-4237-a803-39b6605663bf {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.724971] env[65726]: DEBUG oslo_vmware.api [None req-5c1fab1f-7856-43d0-84bd-904452799dc9 tempest-ServersTestJSON-1793637688 tempest-ServersTestJSON-1793637688-project-member] Waiting for the task: (returnval){ [ 615.724971] env[65726]: value = "task-5115397" [ 615.724971] env[65726]: _type = "Task" [ 615.724971] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 615.748416] env[65726]: DEBUG oslo_vmware.api [None req-5c1fab1f-7856-43d0-84bd-904452799dc9 tempest-ServersTestJSON-1793637688 tempest-ServersTestJSON-1793637688-project-member] Task: {'id': task-5115397, 'name': Rename_Task} progress is 6%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 615.827321] env[65726]: DEBUG nova.compute.manager [None req-9da01a84-79f6-4bc1-8b36-334b89c8de59 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] [instance: ee676e3b-3326-46a3-940d-ebbb8b108991] Start building block device mappings for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 615.846760] env[65726]: DEBUG nova.network.neutron [None req-9da01a84-79f6-4bc1-8b36-334b89c8de59 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] [instance: ee676e3b-3326-46a3-940d-ebbb8b108991] Successfully created port: 30651f8a-abdc-487b-9044-1c8a17beab23 {{(pid=65726) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 616.022177] env[65726]: WARNING openstack [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 616.023379] env[65726]: WARNING openstack [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 616.029140] env[65726]: DEBUG nova.network.neutron [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] [instance: 703fff2d-5aec-4ed1-b987-9e1cc3a25f67] Instance cache missing network info. {{(pid=65726) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 616.060316] env[65726]: DEBUG oslo_concurrency.lockutils [None req-cc552ef7-8733-47e0-9cfd-32896dfc9132 tempest-TenantUsagesTestJSON-327211524 tempest-TenantUsagesTestJSON-327211524-project-member] Acquiring lock "7748c23f-9ea9-4172-8ab7-187678272bb3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 616.060612] env[65726]: DEBUG oslo_concurrency.lockutils [None req-cc552ef7-8733-47e0-9cfd-32896dfc9132 tempest-TenantUsagesTestJSON-327211524 tempest-TenantUsagesTestJSON-327211524-project-member] Lock "7748c23f-9ea9-4172-8ab7-187678272bb3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 616.082092] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b8e65ecf-6b31-4881-a7b5-2f0a252ee6d8 tempest-VolumesAssistedSnapshotsTest-1875841935 tempest-VolumesAssistedSnapshotsTest-1875841935-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 616.143683] env[65726]: DEBUG oslo_vmware.api [None req-af5fca6d-c57c-4762-93f4-4ba84e22ca7b tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': task-5115396, 'name': Rename_Task, 'duration_secs': 0.210629} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 616.145407] env[65726]: WARNING openstack [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 616.145900] env[65726]: WARNING openstack [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 616.154425] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-af5fca6d-c57c-4762-93f4-4ba84e22ca7b tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 8e9c27f6-e9fa-4ed0-b6ae-3b5df52d1c44] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 616.157468] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5622c049-1bb0-48b5-8e78-b8ffa44d826a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.165130] env[65726]: DEBUG oslo_vmware.api [None req-af5fca6d-c57c-4762-93f4-4ba84e22ca7b tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Waiting for the task: (returnval){ [ 616.165130] env[65726]: value = "task-5115398" [ 616.165130] env[65726]: _type = "Task" [ 616.165130] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 616.175771] env[65726]: DEBUG oslo_vmware.api [None req-af5fca6d-c57c-4762-93f4-4ba84e22ca7b tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': task-5115398, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 616.187716] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e25fd29e-3f6a-4d7f-b226-d8a101c8fa9f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.197783] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00c36492-7210-4878-ba33-69755857a4c8 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.230094] env[65726]: DEBUG nova.compute.manager [None req-8fd7433e-ebac-43f4-aa60-6301be278066 tempest-ServerDiagnosticsTest-2021442419 tempest-ServerDiagnosticsTest-2021442419-project-member] [instance: de5cad99-1dbc-4435-98b4-987e24ba8c3c] Starting instance... {{(pid=65726) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 616.236022] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Lock "0e064341-4e4a-407b-8c26-3eb04b409029" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.522s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 616.241744] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f30b6ae-470f-4df3-9b45-84581d050fbb {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.257557] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1de06b90-6ae6-42d4-8998-e5be251672a8 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.262521] env[65726]: DEBUG oslo_vmware.api [None req-5c1fab1f-7856-43d0-84bd-904452799dc9 tempest-ServersTestJSON-1793637688 tempest-ServersTestJSON-1793637688-project-member] Task: {'id': task-5115397, 'name': Rename_Task, 'duration_secs': 0.144358} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 616.262521] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-5c1fab1f-7856-43d0-84bd-904452799dc9 tempest-ServersTestJSON-1793637688 tempest-ServersTestJSON-1793637688-project-member] [instance: 88fb7235-cd0e-49dc-9d78-3583d5d3f528] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 616.263166] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f88fdd60-8922-40c2-960a-ce7ba391b9e6 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.274531] env[65726]: DEBUG nova.compute.provider_tree [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 616.277443] env[65726]: DEBUG oslo_vmware.api [None req-5c1fab1f-7856-43d0-84bd-904452799dc9 tempest-ServersTestJSON-1793637688 tempest-ServersTestJSON-1793637688-project-member] Waiting for the task: (returnval){ [ 616.277443] env[65726]: value = "task-5115399" [ 616.277443] env[65726]: _type = "Task" [ 616.277443] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 616.287591] env[65726]: DEBUG oslo_vmware.api [None req-5c1fab1f-7856-43d0-84bd-904452799dc9 tempest-ServersTestJSON-1793637688 tempest-ServersTestJSON-1793637688-project-member] Task: {'id': task-5115399, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 616.365711] env[65726]: WARNING openstack [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 616.366220] env[65726]: WARNING openstack [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 616.445994] env[65726]: DEBUG oslo_concurrency.lockutils [None req-40525054-994c-41ab-bed7-c877834cae18 tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] Acquiring lock "3b933684-ac19-44b0-a49d-6af45501e166" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 616.446331] env[65726]: DEBUG oslo_concurrency.lockutils [None req-40525054-994c-41ab-bed7-c877834cae18 tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] Lock "3b933684-ac19-44b0-a49d-6af45501e166" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 616.503346] env[65726]: DEBUG nova.network.neutron [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] [instance: 703fff2d-5aec-4ed1-b987-9e1cc3a25f67] Updating instance_info_cache with network_info: [{"id": "29330b5b-6171-4de8-bf77-6c57f6a8bd58", "address": "fa:16:3e:1d:eb:7e", "network": {"id": "f639fd12-fb47-4b3f-9895-6facebc89726", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-2112264538-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.2", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "3a680eae0e734ea8b2274b504d0a46d8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "257e5ea7-8b80-4301-9900-a754f1fe2031", "external-id": "nsx-vlan-transportzone-682", "segmentation_id": 682, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap29330b5b-61", "ovs_interfaceid": "29330b5b-6171-4de8-bf77-6c57f6a8bd58", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 616.676537] env[65726]: DEBUG oslo_vmware.api [None req-af5fca6d-c57c-4762-93f4-4ba84e22ca7b tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': task-5115398, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 616.758687] env[65726]: DEBUG oslo_concurrency.lockutils [None req-8fd7433e-ebac-43f4-aa60-6301be278066 tempest-ServerDiagnosticsTest-2021442419 tempest-ServerDiagnosticsTest-2021442419-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 616.779160] env[65726]: DEBUG nova.scheduler.client.report [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 616.796851] env[65726]: DEBUG oslo_vmware.api [None req-5c1fab1f-7856-43d0-84bd-904452799dc9 tempest-ServersTestJSON-1793637688 tempest-ServersTestJSON-1793637688-project-member] Task: {'id': task-5115399, 'name': PowerOnVM_Task, 'duration_secs': 0.500511} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 616.796944] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-5c1fab1f-7856-43d0-84bd-904452799dc9 tempest-ServersTestJSON-1793637688 tempest-ServersTestJSON-1793637688-project-member] [instance: 88fb7235-cd0e-49dc-9d78-3583d5d3f528] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 616.797147] env[65726]: INFO nova.compute.manager [None req-5c1fab1f-7856-43d0-84bd-904452799dc9 tempest-ServersTestJSON-1793637688 tempest-ServersTestJSON-1793637688-project-member] [instance: 88fb7235-cd0e-49dc-9d78-3583d5d3f528] Took 12.91 seconds to spawn the instance on the hypervisor. [ 616.797351] env[65726]: DEBUG nova.compute.manager [None req-5c1fab1f-7856-43d0-84bd-904452799dc9 tempest-ServersTestJSON-1793637688 tempest-ServersTestJSON-1793637688-project-member] [instance: 88fb7235-cd0e-49dc-9d78-3583d5d3f528] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 616.799864] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d1925ee-0cda-4765-99fd-7884eefdedb2 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.838011] env[65726]: DEBUG nova.compute.manager [None req-9da01a84-79f6-4bc1-8b36-334b89c8de59 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] [instance: ee676e3b-3326-46a3-940d-ebbb8b108991] Start spawning the instance on the hypervisor. {{(pid=65726) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 616.866164] env[65726]: DEBUG nova.virt.hardware [None req-9da01a84-79f6-4bc1-8b36-334b89c8de59 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T19:28:34Z,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1539f06fa3534e90acbc3a60c4b8bd3f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T19:28:35Z,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 616.866164] env[65726]: DEBUG nova.virt.hardware [None req-9da01a84-79f6-4bc1-8b36-334b89c8de59 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 616.866164] env[65726]: DEBUG nova.virt.hardware [None req-9da01a84-79f6-4bc1-8b36-334b89c8de59 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 616.866346] env[65726]: DEBUG nova.virt.hardware [None req-9da01a84-79f6-4bc1-8b36-334b89c8de59 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 616.866346] env[65726]: DEBUG nova.virt.hardware [None req-9da01a84-79f6-4bc1-8b36-334b89c8de59 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 616.866346] env[65726]: DEBUG nova.virt.hardware [None req-9da01a84-79f6-4bc1-8b36-334b89c8de59 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 616.866346] env[65726]: DEBUG nova.virt.hardware [None req-9da01a84-79f6-4bc1-8b36-334b89c8de59 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 616.866799] env[65726]: DEBUG nova.virt.hardware [None req-9da01a84-79f6-4bc1-8b36-334b89c8de59 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 616.867120] env[65726]: DEBUG nova.virt.hardware [None req-9da01a84-79f6-4bc1-8b36-334b89c8de59 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 616.867415] env[65726]: DEBUG nova.virt.hardware [None req-9da01a84-79f6-4bc1-8b36-334b89c8de59 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 616.867739] env[65726]: DEBUG nova.virt.hardware [None req-9da01a84-79f6-4bc1-8b36-334b89c8de59 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 616.869106] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebdb0675-17e8-493e-92fe-d9ecf0ea82fd {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.879458] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2b60187-7376-4996-89a7-5f62f76a892c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.006842] env[65726]: DEBUG oslo_concurrency.lockutils [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Releasing lock "refresh_cache-703fff2d-5aec-4ed1-b987-9e1cc3a25f67" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 617.007318] env[65726]: DEBUG nova.compute.manager [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] [instance: 703fff2d-5aec-4ed1-b987-9e1cc3a25f67] Instance network_info: |[{"id": "29330b5b-6171-4de8-bf77-6c57f6a8bd58", "address": "fa:16:3e:1d:eb:7e", "network": {"id": "f639fd12-fb47-4b3f-9895-6facebc89726", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-2112264538-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.2", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "3a680eae0e734ea8b2274b504d0a46d8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "257e5ea7-8b80-4301-9900-a754f1fe2031", "external-id": "nsx-vlan-transportzone-682", "segmentation_id": 682, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap29330b5b-61", "ovs_interfaceid": "29330b5b-6171-4de8-bf77-6c57f6a8bd58", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 617.008142] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] [instance: 703fff2d-5aec-4ed1-b987-9e1cc3a25f67] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:1d:eb:7e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '257e5ea7-8b80-4301-9900-a754f1fe2031', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '29330b5b-6171-4de8-bf77-6c57f6a8bd58', 'vif_model': 'vmxnet3'}] {{(pid=65726) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 617.016115] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Creating folder: Project (3a680eae0e734ea8b2274b504d0a46d8). Parent ref: group-v995008. {{(pid=65726) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 617.016478] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-84c546f9-13bf-4289-8ec5-9a8a68b43745 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.030851] env[65726]: INFO nova.virt.vmwareapi.vm_util [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Created folder: Project (3a680eae0e734ea8b2274b504d0a46d8) in parent group-v995008. [ 617.031260] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Creating folder: Instances. Parent ref: group-v995025. {{(pid=65726) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 617.031400] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7afbaace-2dd2-4673-86da-485e0246eddf {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.041733] env[65726]: INFO nova.virt.vmwareapi.vm_util [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Created folder: Instances in parent group-v995025. [ 617.042019] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 617.042258] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 703fff2d-5aec-4ed1-b987-9e1cc3a25f67] Creating VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 617.042507] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-dd070e95-68d6-4e42-86d8-de8c12b2961b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.063268] env[65726]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 617.063268] env[65726]: value = "task-5115402" [ 617.063268] env[65726]: _type = "Task" [ 617.063268] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 617.071752] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115402, 'name': CreateVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 617.136114] env[65726]: DEBUG nova.compute.manager [req-89abbed9-79c1-4f97-8578-1d6d6acc3b1f req-8eeca103-4440-4f90-aced-7f9c5ebad600 service nova] [instance: 703fff2d-5aec-4ed1-b987-9e1cc3a25f67] Received event network-vif-plugged-29330b5b-6171-4de8-bf77-6c57f6a8bd58 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 617.136398] env[65726]: DEBUG oslo_concurrency.lockutils [req-89abbed9-79c1-4f97-8578-1d6d6acc3b1f req-8eeca103-4440-4f90-aced-7f9c5ebad600 service nova] Acquiring lock "703fff2d-5aec-4ed1-b987-9e1cc3a25f67-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 617.136714] env[65726]: DEBUG oslo_concurrency.lockutils [req-89abbed9-79c1-4f97-8578-1d6d6acc3b1f req-8eeca103-4440-4f90-aced-7f9c5ebad600 service nova] Lock "703fff2d-5aec-4ed1-b987-9e1cc3a25f67-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 617.136964] env[65726]: DEBUG oslo_concurrency.lockutils [req-89abbed9-79c1-4f97-8578-1d6d6acc3b1f req-8eeca103-4440-4f90-aced-7f9c5ebad600 service nova] Lock "703fff2d-5aec-4ed1-b987-9e1cc3a25f67-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 617.137188] env[65726]: DEBUG nova.compute.manager [req-89abbed9-79c1-4f97-8578-1d6d6acc3b1f req-8eeca103-4440-4f90-aced-7f9c5ebad600 service nova] [instance: 703fff2d-5aec-4ed1-b987-9e1cc3a25f67] No waiting events found dispatching network-vif-plugged-29330b5b-6171-4de8-bf77-6c57f6a8bd58 {{(pid=65726) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 617.137445] env[65726]: WARNING nova.compute.manager [req-89abbed9-79c1-4f97-8578-1d6d6acc3b1f req-8eeca103-4440-4f90-aced-7f9c5ebad600 service nova] [instance: 703fff2d-5aec-4ed1-b987-9e1cc3a25f67] Received unexpected event network-vif-plugged-29330b5b-6171-4de8-bf77-6c57f6a8bd58 for instance with vm_state building and task_state spawning. [ 617.178062] env[65726]: DEBUG oslo_vmware.api [None req-af5fca6d-c57c-4762-93f4-4ba84e22ca7b tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': task-5115398, 'name': PowerOnVM_Task, 'duration_secs': 0.640363} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 617.178498] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-af5fca6d-c57c-4762-93f4-4ba84e22ca7b tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 8e9c27f6-e9fa-4ed0-b6ae-3b5df52d1c44] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 617.178733] env[65726]: INFO nova.compute.manager [None req-af5fca6d-c57c-4762-93f4-4ba84e22ca7b tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 8e9c27f6-e9fa-4ed0-b6ae-3b5df52d1c44] Took 15.58 seconds to spawn the instance on the hypervisor. [ 617.179013] env[65726]: DEBUG nova.compute.manager [None req-af5fca6d-c57c-4762-93f4-4ba84e22ca7b tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 8e9c27f6-e9fa-4ed0-b6ae-3b5df52d1c44] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 617.179972] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13a95c3b-fd9d-494a-b33d-f3228c80f6bc {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.293189] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.479s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 617.293189] env[65726]: DEBUG nova.compute.manager [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] [instance: 8493ba52-b950-4727-970a-19b1797b2ca1] Start building networks asynchronously for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 617.296256] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 10.835s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 617.297843] env[65726]: INFO nova.compute.claims [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] [instance: 21730277-cef6-4bfe-9f67-de71f6f615ff] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 617.322257] env[65726]: INFO nova.compute.manager [None req-5c1fab1f-7856-43d0-84bd-904452799dc9 tempest-ServersTestJSON-1793637688 tempest-ServersTestJSON-1793637688-project-member] [instance: 88fb7235-cd0e-49dc-9d78-3583d5d3f528] Took 18.08 seconds to build instance. [ 617.567388] env[65726]: DEBUG nova.network.neutron [None req-9da01a84-79f6-4bc1-8b36-334b89c8de59 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] [instance: ee676e3b-3326-46a3-940d-ebbb8b108991] Successfully updated port: 30651f8a-abdc-487b-9044-1c8a17beab23 {{(pid=65726) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 617.583895] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115402, 'name': CreateVM_Task, 'duration_secs': 0.415359} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 617.584361] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 703fff2d-5aec-4ed1-b987-9e1cc3a25f67] Created VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 617.585329] env[65726]: WARNING openstack [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 617.586321] env[65726]: WARNING openstack [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 617.593866] env[65726]: DEBUG oslo_vmware.service [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8006ed5-3266-4f7e-870f-be7a290ca324 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.604135] env[65726]: DEBUG oslo_concurrency.lockutils [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 617.604135] env[65726]: DEBUG oslo_concurrency.lockutils [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 617.604135] env[65726]: DEBUG oslo_concurrency.lockutils [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 617.604135] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2a5dfa0d-00fb-43d8-9e8d-7765822dddd9 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.612374] env[65726]: DEBUG oslo_vmware.api [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Waiting for the task: (returnval){ [ 617.612374] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5276fd04-43ac-0c4a-e624-4a3564a29414" [ 617.612374] env[65726]: _type = "Task" [ 617.612374] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 617.621909] env[65726]: DEBUG oslo_vmware.api [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5276fd04-43ac-0c4a-e624-4a3564a29414, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 617.699224] env[65726]: INFO nova.compute.manager [None req-af5fca6d-c57c-4762-93f4-4ba84e22ca7b tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 8e9c27f6-e9fa-4ed0-b6ae-3b5df52d1c44] Took 21.32 seconds to build instance. [ 617.802578] env[65726]: DEBUG nova.compute.utils [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Using /dev/sd instead of None {{(pid=65726) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 617.804186] env[65726]: DEBUG nova.compute.manager [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] [instance: 8493ba52-b950-4727-970a-19b1797b2ca1] Allocating IP information in the background. {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 617.807039] env[65726]: DEBUG nova.network.neutron [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] [instance: 8493ba52-b950-4727-970a-19b1797b2ca1] allocate_for_instance() {{(pid=65726) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 617.807039] env[65726]: WARNING neutronclient.v2_0.client [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 617.807039] env[65726]: WARNING neutronclient.v2_0.client [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 617.807039] env[65726]: WARNING openstack [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 617.807039] env[65726]: WARNING openstack [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 617.817381] env[65726]: DEBUG nova.compute.manager [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] [instance: 8493ba52-b950-4727-970a-19b1797b2ca1] Start building block device mappings for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 617.824302] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5c1fab1f-7856-43d0-84bd-904452799dc9 tempest-ServersTestJSON-1793637688 tempest-ServersTestJSON-1793637688-project-member] Lock "88fb7235-cd0e-49dc-9d78-3583d5d3f528" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 19.592s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 617.863898] env[65726]: DEBUG nova.policy [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '838543f31e974acf83678c7f5c27b102', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7c3641c4a34b4bf8b09403220e153e65', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65726) authorize /opt/stack/nova/nova/policy.py:192}} [ 617.981085] env[65726]: DEBUG oslo_concurrency.lockutils [None req-362ec7a6-07fd-4c6b-b5fb-9865bdcb8c1a tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Acquiring lock "274ab469-61a9-4b7e-852c-074c871e3abf" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 617.981372] env[65726]: DEBUG oslo_concurrency.lockutils [None req-362ec7a6-07fd-4c6b-b5fb-9865bdcb8c1a tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Lock "274ab469-61a9-4b7e-852c-074c871e3abf" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 618.077678] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9da01a84-79f6-4bc1-8b36-334b89c8de59 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Acquiring lock "refresh_cache-ee676e3b-3326-46a3-940d-ebbb8b108991" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 618.077935] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9da01a84-79f6-4bc1-8b36-334b89c8de59 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Acquired lock "refresh_cache-ee676e3b-3326-46a3-940d-ebbb8b108991" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 618.078111] env[65726]: DEBUG nova.network.neutron [None req-9da01a84-79f6-4bc1-8b36-334b89c8de59 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] [instance: ee676e3b-3326-46a3-940d-ebbb8b108991] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 618.125579] env[65726]: DEBUG oslo_concurrency.lockutils [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 618.125579] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] [instance: 703fff2d-5aec-4ed1-b987-9e1cc3a25f67] Processing image b52362a3-ee8a-4cbf-b06f-513b0cc8f95c {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 618.125579] env[65726]: DEBUG oslo_concurrency.lockutils [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 618.125579] env[65726]: DEBUG oslo_concurrency.lockutils [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 618.125950] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 618.125950] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9871a46b-2f35-44d4-80a1-6f1fc7bdc65b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.148053] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 618.148053] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65726) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 618.148800] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e36439f7-32e4-49f3-8560-0b8bb882e964 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.157111] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d6aa3bed-6ce8-4c11-bd2f-c49b63cf94b5 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.163607] env[65726]: DEBUG oslo_vmware.api [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Waiting for the task: (returnval){ [ 618.163607] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52947b53-11df-17c1-e369-3c19ce6f7388" [ 618.163607] env[65726]: _type = "Task" [ 618.163607] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 618.175602] env[65726]: DEBUG oslo_vmware.api [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52947b53-11df-17c1-e369-3c19ce6f7388, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 618.200853] env[65726]: DEBUG oslo_concurrency.lockutils [None req-af5fca6d-c57c-4762-93f4-4ba84e22ca7b tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Lock "8e9c27f6-e9fa-4ed0-b6ae-3b5df52d1c44" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 22.829s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 618.272430] env[65726]: DEBUG nova.network.neutron [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] [instance: 8493ba52-b950-4727-970a-19b1797b2ca1] Successfully created port: 964ef0b7-9857-4032-ac36-efb1ecd1419d {{(pid=65726) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 618.329256] env[65726]: DEBUG nova.compute.manager [None req-8cc1dec8-17ac-4db3-bea9-0fb861f8b4b8 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] [instance: 04ffdf80-c54c-4587-9bb4-d520dc440501] Starting instance... {{(pid=65726) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 618.588415] env[65726]: WARNING openstack [None req-9da01a84-79f6-4bc1-8b36-334b89c8de59 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 618.588952] env[65726]: WARNING openstack [None req-9da01a84-79f6-4bc1-8b36-334b89c8de59 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 618.597983] env[65726]: DEBUG nova.network.neutron [None req-9da01a84-79f6-4bc1-8b36-334b89c8de59 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] [instance: ee676e3b-3326-46a3-940d-ebbb8b108991] Instance cache missing network info. {{(pid=65726) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 618.660599] env[65726]: WARNING openstack [None req-9da01a84-79f6-4bc1-8b36-334b89c8de59 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 618.660979] env[65726]: WARNING openstack [None req-9da01a84-79f6-4bc1-8b36-334b89c8de59 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 618.681280] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] [instance: 703fff2d-5aec-4ed1-b987-9e1cc3a25f67] Preparing fetch location {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 618.681280] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Creating directory with path [datastore2] vmware_temp/faf7f68d-e90f-4f24-809d-008da3fe89b9/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 618.681465] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b9680fab-e98d-4aad-8288-e8d2e840031a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.703107] env[65726]: DEBUG nova.compute.manager [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] [instance: 811c1bbd-259c-4f5e-9a61-581bfba0ec4d] Starting instance... {{(pid=65726) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 618.727042] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Created directory with path [datastore2] vmware_temp/faf7f68d-e90f-4f24-809d-008da3fe89b9/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 618.727042] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] [instance: 703fff2d-5aec-4ed1-b987-9e1cc3a25f67] Fetch image to [datastore2] vmware_temp/faf7f68d-e90f-4f24-809d-008da3fe89b9/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/tmp-sparse.vmdk {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 618.727042] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] [instance: 703fff2d-5aec-4ed1-b987-9e1cc3a25f67] Downloading image file data b52362a3-ee8a-4cbf-b06f-513b0cc8f95c to [datastore2] vmware_temp/faf7f68d-e90f-4f24-809d-008da3fe89b9/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/tmp-sparse.vmdk on the data store datastore2 {{(pid=65726) _fetch_image_as_file /opt/stack/nova/nova/virt/vmwareapi/vmops.py:399}} [ 618.727042] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ce5aa1a-d931-4b1e-a462-e3da8f246bb4 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.739450] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c678807-6378-4da3-bc47-4a615f5f666a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.746132] env[65726]: WARNING openstack [None req-9da01a84-79f6-4bc1-8b36-334b89c8de59 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 618.746552] env[65726]: WARNING openstack [None req-9da01a84-79f6-4bc1-8b36-334b89c8de59 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 618.765555] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea9d1cd2-8d21-41cc-b58e-9dec5677e41b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.804748] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67e760a3-e4c0-480d-ab7e-4d037a6e1f23 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.808126] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9717052-8c26-41ba-82e8-63424ee8622c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.822156] env[65726]: DEBUG oslo_vmware.service [-] Invoking SessionManager.AcquireGenericServiceTicket with opID=oslo.vmware-5ff3dc5b-8fe2-4fd5-8a48-1b6bf7d320c9 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.825226] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb4a3e0d-62c8-4fe0-b9e8-68d2c59570de {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.857129] env[65726]: DEBUG nova.compute.manager [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] [instance: 8493ba52-b950-4727-970a-19b1797b2ca1] Start spawning the instance on the hypervisor. {{(pid=65726) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 618.865933] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55fcfb69-9afa-4fb6-845c-b9057972ace8 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.869038] env[65726]: DEBUG nova.virt.vmwareapi.images [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] [instance: 703fff2d-5aec-4ed1-b987-9e1cc3a25f67] Downloading image file data b52362a3-ee8a-4cbf-b06f-513b0cc8f95c to the data store datastore2 {{(pid=65726) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:245}} [ 618.872899] env[65726]: DEBUG nova.network.neutron [None req-9da01a84-79f6-4bc1-8b36-334b89c8de59 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] [instance: ee676e3b-3326-46a3-940d-ebbb8b108991] Updating instance_info_cache with network_info: [{"id": "30651f8a-abdc-487b-9044-1c8a17beab23", "address": "fa:16:3e:13:1f:1c", "network": {"id": "55d629b8-a541-4720-9289-927bbc6e6f22", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.53", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "1539f06fa3534e90acbc3a60c4b8bd3f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "35342bcb-8b06-472e-b3c0-43fd3d6c4b30", "external-id": "nsx-vlan-transportzone-524", "segmentation_id": 524, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap30651f8a-ab", "ovs_interfaceid": "30651f8a-abdc-487b-9044-1c8a17beab23", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 618.880831] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-073f23c4-dff3-4768-9194-ebeee56fe395 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.901527] env[65726]: DEBUG oslo_concurrency.lockutils [None req-8cc1dec8-17ac-4db3-bea9-0fb861f8b4b8 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 618.903044] env[65726]: DEBUG nova.compute.provider_tree [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 618.905096] env[65726]: DEBUG nova.virt.hardware [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T19:28:34Z,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1539f06fa3534e90acbc3a60c4b8bd3f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T19:28:35Z,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 618.905309] env[65726]: DEBUG nova.virt.hardware [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 618.905501] env[65726]: DEBUG nova.virt.hardware [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 618.905695] env[65726]: DEBUG nova.virt.hardware [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 618.905945] env[65726]: DEBUG nova.virt.hardware [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 618.905945] env[65726]: DEBUG nova.virt.hardware [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 618.907894] env[65726]: DEBUG nova.virt.hardware [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 618.907894] env[65726]: DEBUG nova.virt.hardware [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 618.907894] env[65726]: DEBUG nova.virt.hardware [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 618.907894] env[65726]: DEBUG nova.virt.hardware [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 618.907894] env[65726]: DEBUG nova.virt.hardware [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 618.909580] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31b4a196-83d8-49f9-97ab-ffb1cefd5b13 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.923244] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cefd4ff4-9401-48aa-84ac-9f6e8d74e184 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.971089] env[65726]: DEBUG oslo_vmware.rw_handles [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Creating HTTP connection to write to file with size = 21318656 and URL = https://esx7c2n1.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/faf7f68d-e90f-4f24-809d-008da3fe89b9/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore2. {{(pid=65726) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 619.062562] env[65726]: DEBUG oslo_concurrency.lockutils [None req-621af3d8-d8b6-4b45-86f5-d66b91010e2e tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Acquiring lock "fd481728-3ef0-4a96-affd-ab2dd3f596bb" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 619.062811] env[65726]: DEBUG oslo_concurrency.lockutils [None req-621af3d8-d8b6-4b45-86f5-d66b91010e2e tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Lock "fd481728-3ef0-4a96-affd-ab2dd3f596bb" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 619.063084] env[65726]: DEBUG oslo_concurrency.lockutils [None req-06f67a27-e6e0-4078-804d-cff1137abcad tempest-ServersTestBootFromVolume-762458528 tempest-ServersTestBootFromVolume-762458528-project-member] Acquiring lock "5a8d4357-4b44-4a19-b1da-42d188c38adc" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 619.063337] env[65726]: DEBUG oslo_concurrency.lockutils [None req-06f67a27-e6e0-4078-804d-cff1137abcad tempest-ServersTestBootFromVolume-762458528 tempest-ServersTestBootFromVolume-762458528-project-member] Lock "5a8d4357-4b44-4a19-b1da-42d188c38adc" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 619.233202] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 619.316019] env[65726]: DEBUG nova.compute.manager [req-58d4953d-3e66-45f1-b4b2-20300fccb186 req-c7aae91c-6959-462a-87be-3af0fe9087e0 service nova] [instance: ee676e3b-3326-46a3-940d-ebbb8b108991] Received event network-vif-plugged-30651f8a-abdc-487b-9044-1c8a17beab23 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 619.316019] env[65726]: DEBUG oslo_concurrency.lockutils [req-58d4953d-3e66-45f1-b4b2-20300fccb186 req-c7aae91c-6959-462a-87be-3af0fe9087e0 service nova] Acquiring lock "ee676e3b-3326-46a3-940d-ebbb8b108991-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 619.316269] env[65726]: DEBUG oslo_concurrency.lockutils [req-58d4953d-3e66-45f1-b4b2-20300fccb186 req-c7aae91c-6959-462a-87be-3af0fe9087e0 service nova] Lock "ee676e3b-3326-46a3-940d-ebbb8b108991-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 619.316424] env[65726]: DEBUG oslo_concurrency.lockutils [req-58d4953d-3e66-45f1-b4b2-20300fccb186 req-c7aae91c-6959-462a-87be-3af0fe9087e0 service nova] Lock "ee676e3b-3326-46a3-940d-ebbb8b108991-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 619.316615] env[65726]: DEBUG nova.compute.manager [req-58d4953d-3e66-45f1-b4b2-20300fccb186 req-c7aae91c-6959-462a-87be-3af0fe9087e0 service nova] [instance: ee676e3b-3326-46a3-940d-ebbb8b108991] No waiting events found dispatching network-vif-plugged-30651f8a-abdc-487b-9044-1c8a17beab23 {{(pid=65726) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 619.316772] env[65726]: WARNING nova.compute.manager [req-58d4953d-3e66-45f1-b4b2-20300fccb186 req-c7aae91c-6959-462a-87be-3af0fe9087e0 service nova] [instance: ee676e3b-3326-46a3-940d-ebbb8b108991] Received unexpected event network-vif-plugged-30651f8a-abdc-487b-9044-1c8a17beab23 for instance with vm_state building and task_state spawning. [ 619.380204] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9da01a84-79f6-4bc1-8b36-334b89c8de59 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Releasing lock "refresh_cache-ee676e3b-3326-46a3-940d-ebbb8b108991" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 619.380485] env[65726]: DEBUG nova.compute.manager [None req-9da01a84-79f6-4bc1-8b36-334b89c8de59 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] [instance: ee676e3b-3326-46a3-940d-ebbb8b108991] Instance network_info: |[{"id": "30651f8a-abdc-487b-9044-1c8a17beab23", "address": "fa:16:3e:13:1f:1c", "network": {"id": "55d629b8-a541-4720-9289-927bbc6e6f22", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.53", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "1539f06fa3534e90acbc3a60c4b8bd3f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "35342bcb-8b06-472e-b3c0-43fd3d6c4b30", "external-id": "nsx-vlan-transportzone-524", "segmentation_id": 524, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap30651f8a-ab", "ovs_interfaceid": "30651f8a-abdc-487b-9044-1c8a17beab23", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 619.381844] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-9da01a84-79f6-4bc1-8b36-334b89c8de59 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] [instance: ee676e3b-3326-46a3-940d-ebbb8b108991] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:13:1f:1c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '35342bcb-8b06-472e-b3c0-43fd3d6c4b30', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '30651f8a-abdc-487b-9044-1c8a17beab23', 'vif_model': 'vmxnet3'}] {{(pid=65726) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 619.395506] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-9da01a84-79f6-4bc1-8b36-334b89c8de59 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Creating folder: Project (27fdee694977439d88ed2b08abb28405). Parent ref: group-v995008. {{(pid=65726) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 619.395506] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-3aa3089b-4fad-4096-863a-2981aaa87535 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.407347] env[65726]: INFO nova.virt.vmwareapi.vm_util [None req-9da01a84-79f6-4bc1-8b36-334b89c8de59 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Created folder: Project (27fdee694977439d88ed2b08abb28405) in parent group-v995008. [ 619.407828] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-9da01a84-79f6-4bc1-8b36-334b89c8de59 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Creating folder: Instances. Parent ref: group-v995028. {{(pid=65726) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 619.407957] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-63ede940-1e6f-4c3d-8e6b-81cd7bdc0b43 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.417056] env[65726]: DEBUG nova.scheduler.client.report [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 619.427191] env[65726]: INFO nova.virt.vmwareapi.vm_util [None req-9da01a84-79f6-4bc1-8b36-334b89c8de59 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Created folder: Instances in parent group-v995028. [ 619.427451] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-9da01a84-79f6-4bc1-8b36-334b89c8de59 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 619.428535] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ee676e3b-3326-46a3-940d-ebbb8b108991] Creating VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 619.428760] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-55697a5e-e4c0-4d2b-ae8f-39155412b375 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.459282] env[65726]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 619.459282] env[65726]: value = "task-5115405" [ 619.459282] env[65726]: _type = "Task" [ 619.459282] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 619.470442] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115405, 'name': CreateVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 619.531957] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a0754b39-9b94-45fe-86fb-b66ad1bd90cb tempest-DeleteServersAdminTestJSON-1071157516 tempest-DeleteServersAdminTestJSON-1071157516-project-admin] Acquiring lock "bb64fc41-19d7-49dc-a17a-9dd74730130f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 619.532424] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a0754b39-9b94-45fe-86fb-b66ad1bd90cb tempest-DeleteServersAdminTestJSON-1071157516 tempest-DeleteServersAdminTestJSON-1071157516-project-admin] Lock "bb64fc41-19d7-49dc-a17a-9dd74730130f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 619.532605] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a0754b39-9b94-45fe-86fb-b66ad1bd90cb tempest-DeleteServersAdminTestJSON-1071157516 tempest-DeleteServersAdminTestJSON-1071157516-project-admin] Acquiring lock "bb64fc41-19d7-49dc-a17a-9dd74730130f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 619.532839] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a0754b39-9b94-45fe-86fb-b66ad1bd90cb tempest-DeleteServersAdminTestJSON-1071157516 tempest-DeleteServersAdminTestJSON-1071157516-project-admin] Lock "bb64fc41-19d7-49dc-a17a-9dd74730130f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 619.533025] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a0754b39-9b94-45fe-86fb-b66ad1bd90cb tempest-DeleteServersAdminTestJSON-1071157516 tempest-DeleteServersAdminTestJSON-1071157516-project-admin] Lock "bb64fc41-19d7-49dc-a17a-9dd74730130f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 619.536817] env[65726]: INFO nova.compute.manager [None req-a0754b39-9b94-45fe-86fb-b66ad1bd90cb tempest-DeleteServersAdminTestJSON-1071157516 tempest-DeleteServersAdminTestJSON-1071157516-project-admin] [instance: bb64fc41-19d7-49dc-a17a-9dd74730130f] Terminating instance [ 619.687501] env[65726]: DEBUG oslo_vmware.rw_handles [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Completed reading data from the image iterator. {{(pid=65726) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 619.687782] env[65726]: DEBUG oslo_vmware.rw_handles [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Closing write handle for https://esx7c2n1.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/faf7f68d-e90f-4f24-809d-008da3fe89b9/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore2. {{(pid=65726) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:281}} [ 619.851353] env[65726]: DEBUG nova.virt.vmwareapi.images [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] [instance: 703fff2d-5aec-4ed1-b987-9e1cc3a25f67] Downloaded image file data b52362a3-ee8a-4cbf-b06f-513b0cc8f95c to vmware_temp/faf7f68d-e90f-4f24-809d-008da3fe89b9/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/tmp-sparse.vmdk on the data store datastore2 {{(pid=65726) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:258}} [ 619.853267] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] [instance: 703fff2d-5aec-4ed1-b987-9e1cc3a25f67] Caching image {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 619.853587] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Copying Virtual Disk [datastore2] vmware_temp/faf7f68d-e90f-4f24-809d-008da3fe89b9/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/tmp-sparse.vmdk to [datastore2] vmware_temp/faf7f68d-e90f-4f24-809d-008da3fe89b9/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 619.853962] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-15eb6323-6a25-4e6e-9e25-3343a6b7cde9 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.862547] env[65726]: DEBUG oslo_vmware.api [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Waiting for the task: (returnval){ [ 619.862547] env[65726]: value = "task-5115406" [ 619.862547] env[65726]: _type = "Task" [ 619.862547] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 619.871368] env[65726]: DEBUG oslo_vmware.api [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Task: {'id': task-5115406, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 619.909607] env[65726]: DEBUG nova.network.neutron [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] [instance: 8493ba52-b950-4727-970a-19b1797b2ca1] Successfully updated port: 964ef0b7-9857-4032-ac36-efb1ecd1419d {{(pid=65726) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 619.927204] env[65726]: DEBUG oslo_concurrency.lockutils [None req-f6d9baae-af6b-4ded-b5be-fe592833addb tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Acquiring lock "8e9c27f6-e9fa-4ed0-b6ae-3b5df52d1c44" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 619.927493] env[65726]: DEBUG oslo_concurrency.lockutils [None req-f6d9baae-af6b-4ded-b5be-fe592833addb tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Lock "8e9c27f6-e9fa-4ed0-b6ae-3b5df52d1c44" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 619.927714] env[65726]: DEBUG nova.compute.manager [None req-f6d9baae-af6b-4ded-b5be-fe592833addb tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 8e9c27f6-e9fa-4ed0-b6ae-3b5df52d1c44] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 619.929074] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5ec1470-d2bc-44f8-926d-3675d1f2c093 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.932721] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.637s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 619.933206] env[65726]: DEBUG nova.compute.manager [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] [instance: 21730277-cef6-4bfe-9f67-de71f6f615ff] Start building networks asynchronously for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 619.935865] env[65726]: DEBUG oslo_concurrency.lockutils [None req-d2d75a44-1ded-480a-a8b4-89859e465e33 tempest-ServerDiagnosticsV248Test-347810772 tempest-ServerDiagnosticsV248Test-347810772-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.194s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 619.937661] env[65726]: INFO nova.compute.claims [None req-d2d75a44-1ded-480a-a8b4-89859e465e33 tempest-ServerDiagnosticsV248Test-347810772 tempest-ServerDiagnosticsV248Test-347810772-project-member] [instance: 4d4dc88b-0d6c-430c-ab4f-c81a56e4657b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 619.948139] env[65726]: DEBUG nova.compute.manager [None req-f6d9baae-af6b-4ded-b5be-fe592833addb tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 8e9c27f6-e9fa-4ed0-b6ae-3b5df52d1c44] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=65726) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3459}} [ 619.948139] env[65726]: DEBUG nova.objects.instance [None req-f6d9baae-af6b-4ded-b5be-fe592833addb tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Lazy-loading 'flavor' on Instance uuid 8e9c27f6-e9fa-4ed0-b6ae-3b5df52d1c44 {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 619.972227] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115405, 'name': CreateVM_Task} progress is 99%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 620.047023] env[65726]: DEBUG nova.compute.manager [None req-a0754b39-9b94-45fe-86fb-b66ad1bd90cb tempest-DeleteServersAdminTestJSON-1071157516 tempest-DeleteServersAdminTestJSON-1071157516-project-admin] [instance: bb64fc41-19d7-49dc-a17a-9dd74730130f] Start destroying the instance on the hypervisor. {{(pid=65726) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 620.047023] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-a0754b39-9b94-45fe-86fb-b66ad1bd90cb tempest-DeleteServersAdminTestJSON-1071157516 tempest-DeleteServersAdminTestJSON-1071157516-project-admin] [instance: bb64fc41-19d7-49dc-a17a-9dd74730130f] Destroying instance {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 620.047023] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d201169-1b8c-403d-91d2-dd423c80ebf3 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.054462] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-a0754b39-9b94-45fe-86fb-b66ad1bd90cb tempest-DeleteServersAdminTestJSON-1071157516 tempest-DeleteServersAdminTestJSON-1071157516-project-admin] [instance: bb64fc41-19d7-49dc-a17a-9dd74730130f] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 620.054802] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a091a7d4-2e83-4e34-9818-aa8e1569e1f4 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.063356] env[65726]: DEBUG oslo_vmware.api [None req-a0754b39-9b94-45fe-86fb-b66ad1bd90cb tempest-DeleteServersAdminTestJSON-1071157516 tempest-DeleteServersAdminTestJSON-1071157516-project-admin] Waiting for the task: (returnval){ [ 620.063356] env[65726]: value = "task-5115407" [ 620.063356] env[65726]: _type = "Task" [ 620.063356] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 620.076481] env[65726]: DEBUG oslo_vmware.api [None req-a0754b39-9b94-45fe-86fb-b66ad1bd90cb tempest-DeleteServersAdminTestJSON-1071157516 tempest-DeleteServersAdminTestJSON-1071157516-project-admin] Task: {'id': task-5115407, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 620.377691] env[65726]: DEBUG oslo_vmware.api [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Task: {'id': task-5115406, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 620.412818] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Acquiring lock "refresh_cache-8493ba52-b950-4727-970a-19b1797b2ca1" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 620.413018] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Acquired lock "refresh_cache-8493ba52-b950-4727-970a-19b1797b2ca1" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 620.413200] env[65726]: DEBUG nova.network.neutron [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] [instance: 8493ba52-b950-4727-970a-19b1797b2ca1] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 620.443044] env[65726]: DEBUG nova.compute.utils [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Using /dev/sd instead of None {{(pid=65726) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 620.451444] env[65726]: DEBUG nova.compute.manager [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] [instance: 21730277-cef6-4bfe-9f67-de71f6f615ff] Allocating IP information in the background. {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 620.451769] env[65726]: DEBUG nova.network.neutron [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] [instance: 21730277-cef6-4bfe-9f67-de71f6f615ff] allocate_for_instance() {{(pid=65726) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 620.452729] env[65726]: WARNING neutronclient.v2_0.client [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 620.452729] env[65726]: WARNING neutronclient.v2_0.client [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 620.453143] env[65726]: WARNING openstack [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 620.453482] env[65726]: WARNING openstack [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 620.477197] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115405, 'name': CreateVM_Task} progress is 99%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 620.541976] env[65726]: DEBUG nova.policy [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '838543f31e974acf83678c7f5c27b102', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7c3641c4a34b4bf8b09403220e153e65', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65726) authorize /opt/stack/nova/nova/policy.py:192}} [ 620.580154] env[65726]: DEBUG oslo_vmware.api [None req-a0754b39-9b94-45fe-86fb-b66ad1bd90cb tempest-DeleteServersAdminTestJSON-1071157516 tempest-DeleteServersAdminTestJSON-1071157516-project-admin] Task: {'id': task-5115407, 'name': PowerOffVM_Task, 'duration_secs': 0.283542} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 620.580154] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-a0754b39-9b94-45fe-86fb-b66ad1bd90cb tempest-DeleteServersAdminTestJSON-1071157516 tempest-DeleteServersAdminTestJSON-1071157516-project-admin] [instance: bb64fc41-19d7-49dc-a17a-9dd74730130f] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 620.580154] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-a0754b39-9b94-45fe-86fb-b66ad1bd90cb tempest-DeleteServersAdminTestJSON-1071157516 tempest-DeleteServersAdminTestJSON-1071157516-project-admin] [instance: bb64fc41-19d7-49dc-a17a-9dd74730130f] Unregistering the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 620.580154] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2c768100-6ec9-443c-9644-46757f50f9fe {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.654619] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-a0754b39-9b94-45fe-86fb-b66ad1bd90cb tempest-DeleteServersAdminTestJSON-1071157516 tempest-DeleteServersAdminTestJSON-1071157516-project-admin] [instance: bb64fc41-19d7-49dc-a17a-9dd74730130f] Unregistered the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 620.654890] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-a0754b39-9b94-45fe-86fb-b66ad1bd90cb tempest-DeleteServersAdminTestJSON-1071157516 tempest-DeleteServersAdminTestJSON-1071157516-project-admin] [instance: bb64fc41-19d7-49dc-a17a-9dd74730130f] Deleting contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 620.655133] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-a0754b39-9b94-45fe-86fb-b66ad1bd90cb tempest-DeleteServersAdminTestJSON-1071157516 tempest-DeleteServersAdminTestJSON-1071157516-project-admin] Deleting the datastore file [datastore1] bb64fc41-19d7-49dc-a17a-9dd74730130f {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 620.655410] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-58ce87f1-afa0-4d43-b699-41f7e38615aa {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.667972] env[65726]: DEBUG oslo_vmware.api [None req-a0754b39-9b94-45fe-86fb-b66ad1bd90cb tempest-DeleteServersAdminTestJSON-1071157516 tempest-DeleteServersAdminTestJSON-1071157516-project-admin] Waiting for the task: (returnval){ [ 620.667972] env[65726]: value = "task-5115409" [ 620.667972] env[65726]: _type = "Task" [ 620.667972] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 620.678412] env[65726]: DEBUG oslo_vmware.api [None req-a0754b39-9b94-45fe-86fb-b66ad1bd90cb tempest-DeleteServersAdminTestJSON-1071157516 tempest-DeleteServersAdminTestJSON-1071157516-project-admin] Task: {'id': task-5115409, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 620.711011] env[65726]: DEBUG nova.compute.manager [req-8796054c-ca0b-4881-ac77-c6ee3e3219b1 req-cadf388c-1142-44c7-a35f-1f3e9be20c91 service nova] [instance: 703fff2d-5aec-4ed1-b987-9e1cc3a25f67] Received event network-changed-29330b5b-6171-4de8-bf77-6c57f6a8bd58 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 620.711199] env[65726]: DEBUG nova.compute.manager [req-8796054c-ca0b-4881-ac77-c6ee3e3219b1 req-cadf388c-1142-44c7-a35f-1f3e9be20c91 service nova] [instance: 703fff2d-5aec-4ed1-b987-9e1cc3a25f67] Refreshing instance network info cache due to event network-changed-29330b5b-6171-4de8-bf77-6c57f6a8bd58. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 620.711462] env[65726]: DEBUG oslo_concurrency.lockutils [req-8796054c-ca0b-4881-ac77-c6ee3e3219b1 req-cadf388c-1142-44c7-a35f-1f3e9be20c91 service nova] Acquiring lock "refresh_cache-703fff2d-5aec-4ed1-b987-9e1cc3a25f67" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 620.711747] env[65726]: DEBUG oslo_concurrency.lockutils [req-8796054c-ca0b-4881-ac77-c6ee3e3219b1 req-cadf388c-1142-44c7-a35f-1f3e9be20c91 service nova] Acquired lock "refresh_cache-703fff2d-5aec-4ed1-b987-9e1cc3a25f67" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 620.711841] env[65726]: DEBUG nova.network.neutron [req-8796054c-ca0b-4881-ac77-c6ee3e3219b1 req-cadf388c-1142-44c7-a35f-1f3e9be20c91 service nova] [instance: 703fff2d-5aec-4ed1-b987-9e1cc3a25f67] Refreshing network info cache for port 29330b5b-6171-4de8-bf77-6c57f6a8bd58 {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 620.835269] env[65726]: DEBUG nova.network.neutron [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] [instance: 21730277-cef6-4bfe-9f67-de71f6f615ff] Successfully created port: cc7e0970-7e47-45d6-b7fa-6e33947b210c {{(pid=65726) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 620.876497] env[65726]: DEBUG oslo_vmware.api [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Task: {'id': task-5115406, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 620.916728] env[65726]: WARNING openstack [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 620.917248] env[65726]: WARNING openstack [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 620.924512] env[65726]: DEBUG nova.network.neutron [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] [instance: 8493ba52-b950-4727-970a-19b1797b2ca1] Instance cache missing network info. {{(pid=65726) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 620.951034] env[65726]: DEBUG nova.compute.manager [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] [instance: 21730277-cef6-4bfe-9f67-de71f6f615ff] Start building block device mappings for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 620.973030] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-f6d9baae-af6b-4ded-b5be-fe592833addb tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 8e9c27f6-e9fa-4ed0-b6ae-3b5df52d1c44] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 620.981261] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-25527038-c697-4736-bb84-b5e0ee9c27c8 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.992802] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115405, 'name': CreateVM_Task, 'duration_secs': 1.432807} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 620.995420] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ee676e3b-3326-46a3-940d-ebbb8b108991] Created VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 620.996182] env[65726]: DEBUG oslo_vmware.api [None req-f6d9baae-af6b-4ded-b5be-fe592833addb tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Waiting for the task: (returnval){ [ 620.996182] env[65726]: value = "task-5115410" [ 620.996182] env[65726]: _type = "Task" [ 620.996182] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 620.997363] env[65726]: WARNING openstack [None req-9da01a84-79f6-4bc1-8b36-334b89c8de59 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 620.997947] env[65726]: WARNING openstack [None req-9da01a84-79f6-4bc1-8b36-334b89c8de59 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 621.005132] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9da01a84-79f6-4bc1-8b36-334b89c8de59 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 621.005647] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9da01a84-79f6-4bc1-8b36-334b89c8de59 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 621.006145] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9da01a84-79f6-4bc1-8b36-334b89c8de59 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 621.013033] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d3005bc7-17be-4d5d-88fa-a20d51f1f14c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.025304] env[65726]: DEBUG oslo_vmware.api [None req-9da01a84-79f6-4bc1-8b36-334b89c8de59 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Waiting for the task: (returnval){ [ 621.025304] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52decc75-d8ff-80a2-4039-27775acef038" [ 621.025304] env[65726]: _type = "Task" [ 621.025304] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 621.033769] env[65726]: DEBUG oslo_vmware.api [None req-f6d9baae-af6b-4ded-b5be-fe592833addb tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': task-5115410, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 621.054685] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9da01a84-79f6-4bc1-8b36-334b89c8de59 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 621.054685] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-9da01a84-79f6-4bc1-8b36-334b89c8de59 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] [instance: ee676e3b-3326-46a3-940d-ebbb8b108991] Processing image b52362a3-ee8a-4cbf-b06f-513b0cc8f95c {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 621.054685] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9da01a84-79f6-4bc1-8b36-334b89c8de59 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 621.108508] env[65726]: DEBUG nova.compute.manager [None req-32ab2821-307c-47ba-ac96-c38579de1e37 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] [instance: 0e064341-4e4a-407b-8c26-3eb04b409029] Stashing vm_state: active {{(pid=65726) _prep_resize /opt/stack/nova/nova/compute/manager.py:6193}} [ 621.185057] env[65726]: DEBUG oslo_vmware.api [None req-a0754b39-9b94-45fe-86fb-b66ad1bd90cb tempest-DeleteServersAdminTestJSON-1071157516 tempest-DeleteServersAdminTestJSON-1071157516-project-admin] Task: {'id': task-5115409, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.382404} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 621.185057] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-a0754b39-9b94-45fe-86fb-b66ad1bd90cb tempest-DeleteServersAdminTestJSON-1071157516 tempest-DeleteServersAdminTestJSON-1071157516-project-admin] Deleted the datastore file {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 621.185346] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-a0754b39-9b94-45fe-86fb-b66ad1bd90cb tempest-DeleteServersAdminTestJSON-1071157516 tempest-DeleteServersAdminTestJSON-1071157516-project-admin] [instance: bb64fc41-19d7-49dc-a17a-9dd74730130f] Deleted contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 621.185417] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-a0754b39-9b94-45fe-86fb-b66ad1bd90cb tempest-DeleteServersAdminTestJSON-1071157516 tempest-DeleteServersAdminTestJSON-1071157516-project-admin] [instance: bb64fc41-19d7-49dc-a17a-9dd74730130f] Instance destroyed {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 621.185600] env[65726]: INFO nova.compute.manager [None req-a0754b39-9b94-45fe-86fb-b66ad1bd90cb tempest-DeleteServersAdminTestJSON-1071157516 tempest-DeleteServersAdminTestJSON-1071157516-project-admin] [instance: bb64fc41-19d7-49dc-a17a-9dd74730130f] Took 1.14 seconds to destroy the instance on the hypervisor. [ 621.185933] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-a0754b39-9b94-45fe-86fb-b66ad1bd90cb tempest-DeleteServersAdminTestJSON-1071157516 tempest-DeleteServersAdminTestJSON-1071157516-project-admin] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 621.186286] env[65726]: DEBUG nova.compute.manager [-] [instance: bb64fc41-19d7-49dc-a17a-9dd74730130f] Deallocating network for instance {{(pid=65726) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 621.186363] env[65726]: DEBUG nova.network.neutron [-] [instance: bb64fc41-19d7-49dc-a17a-9dd74730130f] deallocate_for_instance() {{(pid=65726) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 621.186922] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 621.187230] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 621.217357] env[65726]: WARNING openstack [req-8796054c-ca0b-4881-ac77-c6ee3e3219b1 req-cadf388c-1142-44c7-a35f-1f3e9be20c91 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 621.217874] env[65726]: WARNING openstack [req-8796054c-ca0b-4881-ac77-c6ee3e3219b1 req-cadf388c-1142-44c7-a35f-1f3e9be20c91 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 621.378805] env[65726]: DEBUG oslo_vmware.api [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Task: {'id': task-5115406, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 621.471017] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47ff44a0-e561-479b-b1c6-c022c347b034 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.479518] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dec7d403-f1d2-4780-bd6b-5d874287e01a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.522612] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ef52f3f-f2e6-4fa6-a5fb-525c041c01bd {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.535353] env[65726]: DEBUG oslo_vmware.api [None req-f6d9baae-af6b-4ded-b5be-fe592833addb tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': task-5115410, 'name': PowerOffVM_Task, 'duration_secs': 0.2141} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 621.537231] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd6f623f-ced3-4140-8eb2-cca3bf65e03d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.544183] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-f6d9baae-af6b-4ded-b5be-fe592833addb tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 8e9c27f6-e9fa-4ed0-b6ae-3b5df52d1c44] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 621.544183] env[65726]: DEBUG nova.compute.manager [None req-f6d9baae-af6b-4ded-b5be-fe592833addb tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 8e9c27f6-e9fa-4ed0-b6ae-3b5df52d1c44] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 621.544371] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d3c4772-3262-4b9c-a692-7f961eafdac7 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.563532] env[65726]: DEBUG nova.compute.provider_tree [None req-d2d75a44-1ded-480a-a8b4-89859e465e33 tempest-ServerDiagnosticsV248Test-347810772 tempest-ServerDiagnosticsV248Test-347810772-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 621.629691] env[65726]: DEBUG oslo_concurrency.lockutils [None req-32ab2821-307c-47ba-ac96-c38579de1e37 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 621.823729] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 621.823946] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 621.879564] env[65726]: DEBUG oslo_vmware.api [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Task: {'id': task-5115406, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.915375} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 621.879801] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Copied Virtual Disk [datastore2] vmware_temp/faf7f68d-e90f-4f24-809d-008da3fe89b9/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/tmp-sparse.vmdk to [datastore2] vmware_temp/faf7f68d-e90f-4f24-809d-008da3fe89b9/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 621.879983] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Deleting the datastore file [datastore2] vmware_temp/faf7f68d-e90f-4f24-809d-008da3fe89b9/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/tmp-sparse.vmdk {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 621.880275] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c50dff43-ac99-4f94-8068-6944860508c7 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.891988] env[65726]: DEBUG oslo_vmware.api [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Waiting for the task: (returnval){ [ 621.891988] env[65726]: value = "task-5115411" [ 621.891988] env[65726]: _type = "Task" [ 621.891988] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 621.903835] env[65726]: DEBUG oslo_vmware.api [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Task: {'id': task-5115411, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 621.917338] env[65726]: WARNING openstack [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 621.917965] env[65726]: WARNING openstack [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 621.963712] env[65726]: DEBUG nova.compute.manager [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] [instance: 21730277-cef6-4bfe-9f67-de71f6f615ff] Start spawning the instance on the hypervisor. {{(pid=65726) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 621.996897] env[65726]: DEBUG nova.virt.hardware [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T19:28:34Z,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1539f06fa3534e90acbc3a60c4b8bd3f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T19:28:35Z,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 621.997164] env[65726]: DEBUG nova.virt.hardware [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 621.997315] env[65726]: DEBUG nova.virt.hardware [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 621.997494] env[65726]: DEBUG nova.virt.hardware [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 621.997629] env[65726]: DEBUG nova.virt.hardware [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 621.997797] env[65726]: DEBUG nova.virt.hardware [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 621.997985] env[65726]: DEBUG nova.virt.hardware [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 621.998158] env[65726]: DEBUG nova.virt.hardware [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 621.998354] env[65726]: DEBUG nova.virt.hardware [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 621.998519] env[65726]: DEBUG nova.virt.hardware [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 621.998689] env[65726]: DEBUG nova.virt.hardware [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 621.999648] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3fe7d05c-0529-46eb-a117-e2e3f788fdf1 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.011407] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f69f812-d25f-42f4-8d03-afae1cd0fced {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.058652] env[65726]: DEBUG oslo_concurrency.lockutils [None req-f6d9baae-af6b-4ded-b5be-fe592833addb tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Lock "8e9c27f6-e9fa-4ed0-b6ae-3b5df52d1c44" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.131s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 622.069022] env[65726]: DEBUG nova.scheduler.client.report [None req-d2d75a44-1ded-480a-a8b4-89859e465e33 tempest-ServerDiagnosticsV248Test-347810772 tempest-ServerDiagnosticsV248Test-347810772-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 622.403902] env[65726]: DEBUG oslo_vmware.api [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Task: {'id': task-5115411, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.072908} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 622.404245] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Deleted the datastore file {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 622.404365] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Moving file from [datastore2] vmware_temp/faf7f68d-e90f-4f24-809d-008da3fe89b9/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c to [datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c. {{(pid=65726) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:334}} [ 622.405410] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MoveDatastoreFile_Task with opID=oslo.vmware-cc35c40e-5aef-4a36-86d5-a28cf7dec312 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.412950] env[65726]: DEBUG oslo_vmware.api [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Waiting for the task: (returnval){ [ 622.412950] env[65726]: value = "task-5115412" [ 622.412950] env[65726]: _type = "Task" [ 622.412950] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 622.423373] env[65726]: DEBUG oslo_vmware.api [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Task: {'id': task-5115412, 'name': MoveDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 622.558359] env[65726]: WARNING openstack [req-8796054c-ca0b-4881-ac77-c6ee3e3219b1 req-cadf388c-1142-44c7-a35f-1f3e9be20c91 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 622.558640] env[65726]: WARNING openstack [req-8796054c-ca0b-4881-ac77-c6ee3e3219b1 req-cadf388c-1142-44c7-a35f-1f3e9be20c91 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 622.572499] env[65726]: DEBUG oslo_concurrency.lockutils [None req-d2d75a44-1ded-480a-a8b4-89859e465e33 tempest-ServerDiagnosticsV248Test-347810772 tempest-ServerDiagnosticsV248Test-347810772-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.637s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 622.572989] env[65726]: DEBUG nova.compute.manager [None req-d2d75a44-1ded-480a-a8b4-89859e465e33 tempest-ServerDiagnosticsV248Test-347810772 tempest-ServerDiagnosticsV248Test-347810772-project-member] [instance: 4d4dc88b-0d6c-430c-ab4f-c81a56e4657b] Start building networks asynchronously for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 622.576222] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b82a7f4d-3603-48cd-8e48-1d790e39dbfb tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.958s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 622.577755] env[65726]: INFO nova.compute.claims [None req-b82a7f4d-3603-48cd-8e48-1d790e39dbfb tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] [instance: cd539d8b-921b-4947-954c-8573380dbdc8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 622.643884] env[65726]: WARNING openstack [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 622.644268] env[65726]: WARNING openstack [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 622.757379] env[65726]: DEBUG nova.network.neutron [-] [instance: bb64fc41-19d7-49dc-a17a-9dd74730130f] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 622.799418] env[65726]: DEBUG nova.network.neutron [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] [instance: 21730277-cef6-4bfe-9f67-de71f6f615ff] Successfully updated port: cc7e0970-7e47-45d6-b7fa-6e33947b210c {{(pid=65726) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 622.868573] env[65726]: WARNING openstack [req-8796054c-ca0b-4881-ac77-c6ee3e3219b1 req-cadf388c-1142-44c7-a35f-1f3e9be20c91 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 622.869156] env[65726]: WARNING openstack [req-8796054c-ca0b-4881-ac77-c6ee3e3219b1 req-cadf388c-1142-44c7-a35f-1f3e9be20c91 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 622.901593] env[65726]: DEBUG nova.network.neutron [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] [instance: 8493ba52-b950-4727-970a-19b1797b2ca1] Updating instance_info_cache with network_info: [{"id": "964ef0b7-9857-4032-ac36-efb1ecd1419d", "address": "fa:16:3e:3e:4b:2f", "network": {"id": "c68b8bcc-840e-459b-bee0-76a9ffdc08cb", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-2061916615-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "7c3641c4a34b4bf8b09403220e153e65", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "aa0d7fd6-4fd3-4451-9208-8a3cf25e30e4", "external-id": "nsx-vlan-transportzone-545", "segmentation_id": 545, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap964ef0b7-98", "ovs_interfaceid": "964ef0b7-9857-4032-ac36-efb1ecd1419d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 622.929788] env[65726]: DEBUG oslo_vmware.api [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Task: {'id': task-5115412, 'name': MoveDatastoreFile_Task, 'duration_secs': 0.056768} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 622.931402] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] File moved {{(pid=65726) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:346}} [ 622.931402] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] [instance: 703fff2d-5aec-4ed1-b987-9e1cc3a25f67] Cleaning up location [datastore2] vmware_temp/faf7f68d-e90f-4f24-809d-008da3fe89b9 {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 622.931402] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Deleting the datastore file [datastore2] vmware_temp/faf7f68d-e90f-4f24-809d-008da3fe89b9 {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 622.931402] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-18b5bf39-7525-4c23-869d-e5247d7ed313 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.940406] env[65726]: DEBUG oslo_vmware.api [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Waiting for the task: (returnval){ [ 622.940406] env[65726]: value = "task-5115413" [ 622.940406] env[65726]: _type = "Task" [ 622.940406] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 622.952354] env[65726]: DEBUG oslo_vmware.api [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Task: {'id': task-5115413, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 623.057977] env[65726]: DEBUG nova.network.neutron [req-8796054c-ca0b-4881-ac77-c6ee3e3219b1 req-cadf388c-1142-44c7-a35f-1f3e9be20c91 service nova] [instance: 703fff2d-5aec-4ed1-b987-9e1cc3a25f67] Updated VIF entry in instance network info cache for port 29330b5b-6171-4de8-bf77-6c57f6a8bd58. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 623.058514] env[65726]: DEBUG nova.network.neutron [req-8796054c-ca0b-4881-ac77-c6ee3e3219b1 req-cadf388c-1142-44c7-a35f-1f3e9be20c91 service nova] [instance: 703fff2d-5aec-4ed1-b987-9e1cc3a25f67] Updating instance_info_cache with network_info: [{"id": "29330b5b-6171-4de8-bf77-6c57f6a8bd58", "address": "fa:16:3e:1d:eb:7e", "network": {"id": "f639fd12-fb47-4b3f-9895-6facebc89726", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-2112264538-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.2", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.3"}}], "meta": {"injected": false, "tenant_id": "3a680eae0e734ea8b2274b504d0a46d8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "257e5ea7-8b80-4301-9900-a754f1fe2031", "external-id": "nsx-vlan-transportzone-682", "segmentation_id": 682, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap29330b5b-61", "ovs_interfaceid": "29330b5b-6171-4de8-bf77-6c57f6a8bd58", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 623.086434] env[65726]: DEBUG nova.compute.utils [None req-d2d75a44-1ded-480a-a8b4-89859e465e33 tempest-ServerDiagnosticsV248Test-347810772 tempest-ServerDiagnosticsV248Test-347810772-project-member] Using /dev/sd instead of None {{(pid=65726) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 623.089095] env[65726]: DEBUG nova.compute.manager [None req-d2d75a44-1ded-480a-a8b4-89859e465e33 tempest-ServerDiagnosticsV248Test-347810772 tempest-ServerDiagnosticsV248Test-347810772-project-member] [instance: 4d4dc88b-0d6c-430c-ab4f-c81a56e4657b] Not allocating networking since 'none' was specified. {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2015}} [ 623.263319] env[65726]: INFO nova.compute.manager [-] [instance: bb64fc41-19d7-49dc-a17a-9dd74730130f] Took 2.08 seconds to deallocate network for instance. [ 623.277766] env[65726]: DEBUG nova.compute.manager [req-93aa6f4f-9c10-41ef-b81e-47851f66fb89 req-7d8801ff-3d1b-481c-94d5-8c54ba593b16 service nova] [instance: ee676e3b-3326-46a3-940d-ebbb8b108991] Received event network-changed-30651f8a-abdc-487b-9044-1c8a17beab23 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 623.277965] env[65726]: DEBUG nova.compute.manager [req-93aa6f4f-9c10-41ef-b81e-47851f66fb89 req-7d8801ff-3d1b-481c-94d5-8c54ba593b16 service nova] [instance: ee676e3b-3326-46a3-940d-ebbb8b108991] Refreshing instance network info cache due to event network-changed-30651f8a-abdc-487b-9044-1c8a17beab23. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 623.278184] env[65726]: DEBUG oslo_concurrency.lockutils [req-93aa6f4f-9c10-41ef-b81e-47851f66fb89 req-7d8801ff-3d1b-481c-94d5-8c54ba593b16 service nova] Acquiring lock "refresh_cache-ee676e3b-3326-46a3-940d-ebbb8b108991" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 623.278395] env[65726]: DEBUG oslo_concurrency.lockutils [req-93aa6f4f-9c10-41ef-b81e-47851f66fb89 req-7d8801ff-3d1b-481c-94d5-8c54ba593b16 service nova] Acquired lock "refresh_cache-ee676e3b-3326-46a3-940d-ebbb8b108991" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 623.278960] env[65726]: DEBUG nova.network.neutron [req-93aa6f4f-9c10-41ef-b81e-47851f66fb89 req-7d8801ff-3d1b-481c-94d5-8c54ba593b16 service nova] [instance: ee676e3b-3326-46a3-940d-ebbb8b108991] Refreshing network info cache for port 30651f8a-abdc-487b-9044-1c8a17beab23 {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 623.302881] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Acquiring lock "refresh_cache-21730277-cef6-4bfe-9f67-de71f6f615ff" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 623.303070] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Acquired lock "refresh_cache-21730277-cef6-4bfe-9f67-de71f6f615ff" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 623.303362] env[65726]: DEBUG nova.network.neutron [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] [instance: 21730277-cef6-4bfe-9f67-de71f6f615ff] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 623.404988] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Releasing lock "refresh_cache-8493ba52-b950-4727-970a-19b1797b2ca1" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 623.404988] env[65726]: DEBUG nova.compute.manager [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] [instance: 8493ba52-b950-4727-970a-19b1797b2ca1] Instance network_info: |[{"id": "964ef0b7-9857-4032-ac36-efb1ecd1419d", "address": "fa:16:3e:3e:4b:2f", "network": {"id": "c68b8bcc-840e-459b-bee0-76a9ffdc08cb", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-2061916615-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "7c3641c4a34b4bf8b09403220e153e65", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "aa0d7fd6-4fd3-4451-9208-8a3cf25e30e4", "external-id": "nsx-vlan-transportzone-545", "segmentation_id": 545, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap964ef0b7-98", "ovs_interfaceid": "964ef0b7-9857-4032-ac36-efb1ecd1419d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 623.405446] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] [instance: 8493ba52-b950-4727-970a-19b1797b2ca1] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:3e:4b:2f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'aa0d7fd6-4fd3-4451-9208-8a3cf25e30e4', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '964ef0b7-9857-4032-ac36-efb1ecd1419d', 'vif_model': 'vmxnet3'}] {{(pid=65726) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 623.413750] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Creating folder: Project (7c3641c4a34b4bf8b09403220e153e65). Parent ref: group-v995008. {{(pid=65726) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 623.414544] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-19dd8d88-ed88-400e-9ac7-ebe54df80ce8 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.427184] env[65726]: INFO nova.virt.vmwareapi.vm_util [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Created folder: Project (7c3641c4a34b4bf8b09403220e153e65) in parent group-v995008. [ 623.427184] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Creating folder: Instances. Parent ref: group-v995031. {{(pid=65726) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 623.427497] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-58fe703f-4d88-47c9-b96b-fccb556da2bf {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.437642] env[65726]: INFO nova.virt.vmwareapi.vm_util [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Created folder: Instances in parent group-v995031. [ 623.437642] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 623.437642] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8493ba52-b950-4727-970a-19b1797b2ca1] Creating VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 623.437642] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e8fa4468-db11-442e-b8e6-9ebb977a5abe {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.462679] env[65726]: DEBUG oslo_vmware.api [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Task: {'id': task-5115413, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.031455} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 623.464167] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Deleted the datastore file {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 623.464869] env[65726]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 623.464869] env[65726]: value = "task-5115416" [ 623.464869] env[65726]: _type = "Task" [ 623.464869] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 623.465358] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-236f9f02-60bc-48f4-b39a-9b507d2cbe10 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.479720] env[65726]: DEBUG oslo_vmware.api [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Waiting for the task: (returnval){ [ 623.479720] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52d4c728-9dbe-c2be-f67f-df772dfdec4b" [ 623.479720] env[65726]: _type = "Task" [ 623.479720] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 623.479911] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115416, 'name': CreateVM_Task} progress is 6%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 623.488605] env[65726]: DEBUG oslo_vmware.api [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52d4c728-9dbe-c2be-f67f-df772dfdec4b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 623.562764] env[65726]: DEBUG oslo_concurrency.lockutils [req-8796054c-ca0b-4881-ac77-c6ee3e3219b1 req-cadf388c-1142-44c7-a35f-1f3e9be20c91 service nova] Releasing lock "refresh_cache-703fff2d-5aec-4ed1-b987-9e1cc3a25f67" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 623.594305] env[65726]: DEBUG nova.compute.manager [None req-d2d75a44-1ded-480a-a8b4-89859e465e33 tempest-ServerDiagnosticsV248Test-347810772 tempest-ServerDiagnosticsV248Test-347810772-project-member] [instance: 4d4dc88b-0d6c-430c-ab4f-c81a56e4657b] Start building block device mappings for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 623.769859] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a0754b39-9b94-45fe-86fb-b66ad1bd90cb tempest-DeleteServersAdminTestJSON-1071157516 tempest-DeleteServersAdminTestJSON-1071157516-project-admin] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 623.783163] env[65726]: WARNING openstack [req-93aa6f4f-9c10-41ef-b81e-47851f66fb89 req-7d8801ff-3d1b-481c-94d5-8c54ba593b16 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 623.783834] env[65726]: WARNING openstack [req-93aa6f4f-9c10-41ef-b81e-47851f66fb89 req-7d8801ff-3d1b-481c-94d5-8c54ba593b16 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 623.810429] env[65726]: WARNING openstack [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 623.810429] env[65726]: WARNING openstack [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 623.820645] env[65726]: DEBUG nova.network.neutron [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] [instance: 21730277-cef6-4bfe-9f67-de71f6f615ff] Instance cache missing network info. {{(pid=65726) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 623.874451] env[65726]: DEBUG nova.compute.manager [req-fde7dfe2-2f08-46e4-af18-49e8c5f6cc8c req-32b5ea0b-1a90-4298-8ffd-9e3c4c1c1ac2 service nova] [instance: 8493ba52-b950-4727-970a-19b1797b2ca1] Received event network-vif-plugged-964ef0b7-9857-4032-ac36-efb1ecd1419d {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 623.874590] env[65726]: DEBUG oslo_concurrency.lockutils [req-fde7dfe2-2f08-46e4-af18-49e8c5f6cc8c req-32b5ea0b-1a90-4298-8ffd-9e3c4c1c1ac2 service nova] Acquiring lock "8493ba52-b950-4727-970a-19b1797b2ca1-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 623.874794] env[65726]: DEBUG oslo_concurrency.lockutils [req-fde7dfe2-2f08-46e4-af18-49e8c5f6cc8c req-32b5ea0b-1a90-4298-8ffd-9e3c4c1c1ac2 service nova] Lock "8493ba52-b950-4727-970a-19b1797b2ca1-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 623.874953] env[65726]: DEBUG oslo_concurrency.lockutils [req-fde7dfe2-2f08-46e4-af18-49e8c5f6cc8c req-32b5ea0b-1a90-4298-8ffd-9e3c4c1c1ac2 service nova] Lock "8493ba52-b950-4727-970a-19b1797b2ca1-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 623.875196] env[65726]: DEBUG nova.compute.manager [req-fde7dfe2-2f08-46e4-af18-49e8c5f6cc8c req-32b5ea0b-1a90-4298-8ffd-9e3c4c1c1ac2 service nova] [instance: 8493ba52-b950-4727-970a-19b1797b2ca1] No waiting events found dispatching network-vif-plugged-964ef0b7-9857-4032-ac36-efb1ecd1419d {{(pid=65726) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 623.875297] env[65726]: WARNING nova.compute.manager [req-fde7dfe2-2f08-46e4-af18-49e8c5f6cc8c req-32b5ea0b-1a90-4298-8ffd-9e3c4c1c1ac2 service nova] [instance: 8493ba52-b950-4727-970a-19b1797b2ca1] Received unexpected event network-vif-plugged-964ef0b7-9857-4032-ac36-efb1ecd1419d for instance with vm_state building and task_state spawning. [ 623.875457] env[65726]: DEBUG nova.compute.manager [req-fde7dfe2-2f08-46e4-af18-49e8c5f6cc8c req-32b5ea0b-1a90-4298-8ffd-9e3c4c1c1ac2 service nova] [instance: 8493ba52-b950-4727-970a-19b1797b2ca1] Received event network-changed-964ef0b7-9857-4032-ac36-efb1ecd1419d {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 623.875603] env[65726]: DEBUG nova.compute.manager [req-fde7dfe2-2f08-46e4-af18-49e8c5f6cc8c req-32b5ea0b-1a90-4298-8ffd-9e3c4c1c1ac2 service nova] [instance: 8493ba52-b950-4727-970a-19b1797b2ca1] Refreshing instance network info cache due to event network-changed-964ef0b7-9857-4032-ac36-efb1ecd1419d. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 623.875821] env[65726]: DEBUG oslo_concurrency.lockutils [req-fde7dfe2-2f08-46e4-af18-49e8c5f6cc8c req-32b5ea0b-1a90-4298-8ffd-9e3c4c1c1ac2 service nova] Acquiring lock "refresh_cache-8493ba52-b950-4727-970a-19b1797b2ca1" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 623.875903] env[65726]: DEBUG oslo_concurrency.lockutils [req-fde7dfe2-2f08-46e4-af18-49e8c5f6cc8c req-32b5ea0b-1a90-4298-8ffd-9e3c4c1c1ac2 service nova] Acquired lock "refresh_cache-8493ba52-b950-4727-970a-19b1797b2ca1" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 623.876059] env[65726]: DEBUG nova.network.neutron [req-fde7dfe2-2f08-46e4-af18-49e8c5f6cc8c req-32b5ea0b-1a90-4298-8ffd-9e3c4c1c1ac2 service nova] [instance: 8493ba52-b950-4727-970a-19b1797b2ca1] Refreshing network info cache for port 964ef0b7-9857-4032-ac36-efb1ecd1419d {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 623.986081] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115416, 'name': CreateVM_Task, 'duration_secs': 0.387414} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 623.989760] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8493ba52-b950-4727-970a-19b1797b2ca1] Created VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 623.990983] env[65726]: WARNING openstack [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 623.991377] env[65726]: WARNING openstack [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 623.996709] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 623.998586] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 623.998902] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 624.002689] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d7ed44b9-3ee1-4769-88f3-395829dce40f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.008769] env[65726]: DEBUG oslo_vmware.api [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52d4c728-9dbe-c2be-f67f-df772dfdec4b, 'name': SearchDatastore_Task, 'duration_secs': 0.027904} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 624.009678] env[65726]: DEBUG oslo_concurrency.lockutils [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 624.009777] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore2] 703fff2d-5aec-4ed1-b987-9e1cc3a25f67/703fff2d-5aec-4ed1-b987-9e1cc3a25f67.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 624.009993] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9da01a84-79f6-4bc1-8b36-334b89c8de59 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 624.010180] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-9da01a84-79f6-4bc1-8b36-334b89c8de59 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 624.010406] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-af40d20a-ea4e-409a-a80a-82b881437c9c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.014135] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-af5b662a-a291-42e1-8622-4588966c4951 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.016127] env[65726]: DEBUG oslo_vmware.api [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Waiting for the task: (returnval){ [ 624.016127] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]526664c6-af4b-2893-1fad-21e8e649059b" [ 624.016127] env[65726]: _type = "Task" [ 624.016127] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 624.031805] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-9da01a84-79f6-4bc1-8b36-334b89c8de59 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 624.032026] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-9da01a84-79f6-4bc1-8b36-334b89c8de59 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65726) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 624.032917] env[65726]: DEBUG oslo_vmware.api [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Waiting for the task: (returnval){ [ 624.032917] env[65726]: value = "task-5115417" [ 624.032917] env[65726]: _type = "Task" [ 624.032917] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 624.033789] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-de87335b-ae75-42f8-9639-99370a390f2c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.045012] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1da2301-9ea2-4300-86b3-db12e257bcf5 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.048207] env[65726]: DEBUG oslo_vmware.api [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]526664c6-af4b-2893-1fad-21e8e649059b, 'name': SearchDatastore_Task, 'duration_secs': 0.017436} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 624.048439] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 624.048592] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] [instance: 8493ba52-b950-4727-970a-19b1797b2ca1] Processing image b52362a3-ee8a-4cbf-b06f-513b0cc8f95c {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 624.048834] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 624.051870] env[65726]: DEBUG oslo_vmware.api [None req-9da01a84-79f6-4bc1-8b36-334b89c8de59 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Waiting for the task: (returnval){ [ 624.051870] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]528249b2-a0b1-2a4d-7c9e-82723aec963e" [ 624.051870] env[65726]: _type = "Task" [ 624.051870] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 624.057035] env[65726]: DEBUG oslo_vmware.api [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Task: {'id': task-5115417, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 624.062606] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2f6c206-6160-4c2c-816c-b60dd7c07908 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.071678] env[65726]: DEBUG oslo_vmware.api [None req-9da01a84-79f6-4bc1-8b36-334b89c8de59 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]528249b2-a0b1-2a4d-7c9e-82723aec963e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 624.098426] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3aab2f0a-ad5b-4686-9928-db637bbb565a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.112030] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e83c131-8975-4988-af5e-a7ee2215df79 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.127999] env[65726]: DEBUG nova.compute.provider_tree [None req-b82a7f4d-3603-48cd-8e48-1d790e39dbfb tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 624.353456] env[65726]: WARNING openstack [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 624.353972] env[65726]: WARNING openstack [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 624.387102] env[65726]: WARNING openstack [req-fde7dfe2-2f08-46e4-af18-49e8c5f6cc8c req-32b5ea0b-1a90-4298-8ffd-9e3c4c1c1ac2 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 624.387690] env[65726]: WARNING openstack [req-fde7dfe2-2f08-46e4-af18-49e8c5f6cc8c req-32b5ea0b-1a90-4298-8ffd-9e3c4c1c1ac2 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 624.549018] env[65726]: DEBUG oslo_vmware.api [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Task: {'id': task-5115417, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 624.569367] env[65726]: DEBUG oslo_vmware.api [None req-9da01a84-79f6-4bc1-8b36-334b89c8de59 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]528249b2-a0b1-2a4d-7c9e-82723aec963e, 'name': SearchDatastore_Task, 'duration_secs': 0.020858} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 624.570762] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2b07eb13-0f64-4fa3-aa53-67ac007fece1 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.578351] env[65726]: DEBUG oslo_vmware.api [None req-9da01a84-79f6-4bc1-8b36-334b89c8de59 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Waiting for the task: (returnval){ [ 624.578351] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52286a98-e485-ef9f-3162-93551818cc23" [ 624.578351] env[65726]: _type = "Task" [ 624.578351] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 624.587662] env[65726]: DEBUG oslo_vmware.api [None req-9da01a84-79f6-4bc1-8b36-334b89c8de59 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52286a98-e485-ef9f-3162-93551818cc23, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 624.606161] env[65726]: DEBUG nova.compute.manager [None req-d2d75a44-1ded-480a-a8b4-89859e465e33 tempest-ServerDiagnosticsV248Test-347810772 tempest-ServerDiagnosticsV248Test-347810772-project-member] [instance: 4d4dc88b-0d6c-430c-ab4f-c81a56e4657b] Start spawning the instance on the hypervisor. {{(pid=65726) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 624.632157] env[65726]: DEBUG nova.scheduler.client.report [None req-b82a7f4d-3603-48cd-8e48-1d790e39dbfb tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 624.645227] env[65726]: DEBUG nova.virt.hardware [None req-d2d75a44-1ded-480a-a8b4-89859e465e33 tempest-ServerDiagnosticsV248Test-347810772 tempest-ServerDiagnosticsV248Test-347810772-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T19:28:34Z,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1539f06fa3534e90acbc3a60c4b8bd3f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T19:28:35Z,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 624.645492] env[65726]: DEBUG nova.virt.hardware [None req-d2d75a44-1ded-480a-a8b4-89859e465e33 tempest-ServerDiagnosticsV248Test-347810772 tempest-ServerDiagnosticsV248Test-347810772-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 624.645646] env[65726]: DEBUG nova.virt.hardware [None req-d2d75a44-1ded-480a-a8b4-89859e465e33 tempest-ServerDiagnosticsV248Test-347810772 tempest-ServerDiagnosticsV248Test-347810772-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 624.645864] env[65726]: DEBUG nova.virt.hardware [None req-d2d75a44-1ded-480a-a8b4-89859e465e33 tempest-ServerDiagnosticsV248Test-347810772 tempest-ServerDiagnosticsV248Test-347810772-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 624.646141] env[65726]: DEBUG nova.virt.hardware [None req-d2d75a44-1ded-480a-a8b4-89859e465e33 tempest-ServerDiagnosticsV248Test-347810772 tempest-ServerDiagnosticsV248Test-347810772-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 624.646374] env[65726]: DEBUG nova.virt.hardware [None req-d2d75a44-1ded-480a-a8b4-89859e465e33 tempest-ServerDiagnosticsV248Test-347810772 tempest-ServerDiagnosticsV248Test-347810772-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 624.646464] env[65726]: DEBUG nova.virt.hardware [None req-d2d75a44-1ded-480a-a8b4-89859e465e33 tempest-ServerDiagnosticsV248Test-347810772 tempest-ServerDiagnosticsV248Test-347810772-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 624.646635] env[65726]: DEBUG nova.virt.hardware [None req-d2d75a44-1ded-480a-a8b4-89859e465e33 tempest-ServerDiagnosticsV248Test-347810772 tempest-ServerDiagnosticsV248Test-347810772-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 624.646806] env[65726]: DEBUG nova.virt.hardware [None req-d2d75a44-1ded-480a-a8b4-89859e465e33 tempest-ServerDiagnosticsV248Test-347810772 tempest-ServerDiagnosticsV248Test-347810772-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 624.647328] env[65726]: DEBUG nova.virt.hardware [None req-d2d75a44-1ded-480a-a8b4-89859e465e33 tempest-ServerDiagnosticsV248Test-347810772 tempest-ServerDiagnosticsV248Test-347810772-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 624.648110] env[65726]: DEBUG nova.virt.hardware [None req-d2d75a44-1ded-480a-a8b4-89859e465e33 tempest-ServerDiagnosticsV248Test-347810772 tempest-ServerDiagnosticsV248Test-347810772-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 624.648855] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53ae6efb-6fd3-4e8e-8ec6-b7d4e709704c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.660881] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2442c8f7-7f7e-41d7-8e04-5e23784053a4 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.677179] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-d2d75a44-1ded-480a-a8b4-89859e465e33 tempest-ServerDiagnosticsV248Test-347810772 tempest-ServerDiagnosticsV248Test-347810772-project-member] [instance: 4d4dc88b-0d6c-430c-ab4f-c81a56e4657b] Instance VIF info [] {{(pid=65726) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 624.683078] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-d2d75a44-1ded-480a-a8b4-89859e465e33 tempest-ServerDiagnosticsV248Test-347810772 tempest-ServerDiagnosticsV248Test-347810772-project-member] Creating folder: Project (a67c79dfde874229b801d5552cc183b8). Parent ref: group-v995008. {{(pid=65726) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 624.683335] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-fc743587-f539-41bc-be12-a29bbaf129be {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.696573] env[65726]: INFO nova.virt.vmwareapi.vm_util [None req-d2d75a44-1ded-480a-a8b4-89859e465e33 tempest-ServerDiagnosticsV248Test-347810772 tempest-ServerDiagnosticsV248Test-347810772-project-member] Created folder: Project (a67c79dfde874229b801d5552cc183b8) in parent group-v995008. [ 624.697402] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-d2d75a44-1ded-480a-a8b4-89859e465e33 tempest-ServerDiagnosticsV248Test-347810772 tempest-ServerDiagnosticsV248Test-347810772-project-member] Creating folder: Instances. Parent ref: group-v995034. {{(pid=65726) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 624.697402] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-78e01aa1-0be0-4e83-a15d-8af6b5e033c5 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.710808] env[65726]: INFO nova.virt.vmwareapi.vm_util [None req-d2d75a44-1ded-480a-a8b4-89859e465e33 tempest-ServerDiagnosticsV248Test-347810772 tempest-ServerDiagnosticsV248Test-347810772-project-member] Created folder: Instances in parent group-v995034. [ 624.711086] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-d2d75a44-1ded-480a-a8b4-89859e465e33 tempest-ServerDiagnosticsV248Test-347810772 tempest-ServerDiagnosticsV248Test-347810772-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 624.711366] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4d4dc88b-0d6c-430c-ab4f-c81a56e4657b] Creating VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 624.711679] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-0f97bfa5-1801-434c-a8f4-7a9affaf79eb {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.739869] env[65726]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 624.739869] env[65726]: value = "task-5115420" [ 624.739869] env[65726]: _type = "Task" [ 624.739869] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 624.751382] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115420, 'name': CreateVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 624.808560] env[65726]: WARNING openstack [req-93aa6f4f-9c10-41ef-b81e-47851f66fb89 req-7d8801ff-3d1b-481c-94d5-8c54ba593b16 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 624.809098] env[65726]: WARNING openstack [req-93aa6f4f-9c10-41ef-b81e-47851f66fb89 req-7d8801ff-3d1b-481c-94d5-8c54ba593b16 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 625.051051] env[65726]: DEBUG oslo_vmware.api [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Task: {'id': task-5115417, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.853171} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 625.051205] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore2] 703fff2d-5aec-4ed1-b987-9e1cc3a25f67/703fff2d-5aec-4ed1-b987-9e1cc3a25f67.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 625.051347] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] [instance: 703fff2d-5aec-4ed1-b987-9e1cc3a25f67] Extending root virtual disk to 1048576 {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 625.051622] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-60688793-73dd-44f8-bf61-4f884b046f6d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.059825] env[65726]: DEBUG oslo_vmware.api [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Waiting for the task: (returnval){ [ 625.059825] env[65726]: value = "task-5115421" [ 625.059825] env[65726]: _type = "Task" [ 625.059825] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 625.074299] env[65726]: DEBUG oslo_vmware.api [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Task: {'id': task-5115421, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 625.089874] env[65726]: DEBUG oslo_vmware.api [None req-9da01a84-79f6-4bc1-8b36-334b89c8de59 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52286a98-e485-ef9f-3162-93551818cc23, 'name': SearchDatastore_Task, 'duration_secs': 0.076489} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 625.090125] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9da01a84-79f6-4bc1-8b36-334b89c8de59 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 625.090482] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-9da01a84-79f6-4bc1-8b36-334b89c8de59 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore2] ee676e3b-3326-46a3-940d-ebbb8b108991/ee676e3b-3326-46a3-940d-ebbb8b108991.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 625.090800] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 625.090990] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 625.091243] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-014886f2-61fe-49b8-8b45-93afa000a0ac {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.093596] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-81397a18-fdc7-4684-b0f1-00fc1088b1f1 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.103061] env[65726]: DEBUG oslo_vmware.api [None req-9da01a84-79f6-4bc1-8b36-334b89c8de59 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Waiting for the task: (returnval){ [ 625.103061] env[65726]: value = "task-5115422" [ 625.103061] env[65726]: _type = "Task" [ 625.103061] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 625.107963] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 625.108168] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65726) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 625.109429] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a2cf405e-b6e1-44a8-84f5-267e149a004b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.115058] env[65726]: DEBUG oslo_vmware.api [None req-9da01a84-79f6-4bc1-8b36-334b89c8de59 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Task: {'id': task-5115422, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 625.120980] env[65726]: DEBUG oslo_vmware.api [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Waiting for the task: (returnval){ [ 625.120980] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52eeba30-7828-ec15-a2d6-ca51efcbcdcf" [ 625.120980] env[65726]: _type = "Task" [ 625.120980] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 625.133474] env[65726]: DEBUG oslo_vmware.api [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52eeba30-7828-ec15-a2d6-ca51efcbcdcf, 'name': SearchDatastore_Task, 'duration_secs': 0.009961} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 625.134495] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6cc499dc-40de-488f-97c7-41b2db74c0fc {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.141089] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b82a7f4d-3603-48cd-8e48-1d790e39dbfb tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.565s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 625.141637] env[65726]: DEBUG nova.compute.manager [None req-b82a7f4d-3603-48cd-8e48-1d790e39dbfb tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] [instance: cd539d8b-921b-4947-954c-8573380dbdc8] Start building networks asynchronously for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 625.144424] env[65726]: DEBUG oslo_vmware.api [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Waiting for the task: (returnval){ [ 625.144424] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]521cd967-b6a2-be77-be93-102c14e4f2a6" [ 625.144424] env[65726]: _type = "Task" [ 625.144424] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 625.144986] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b8e65ecf-6b31-4881-a7b5-2f0a252ee6d8 tempest-VolumesAssistedSnapshotsTest-1875841935 tempest-VolumesAssistedSnapshotsTest-1875841935-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 9.063s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 625.146648] env[65726]: INFO nova.compute.claims [None req-b8e65ecf-6b31-4881-a7b5-2f0a252ee6d8 tempest-VolumesAssistedSnapshotsTest-1875841935 tempest-VolumesAssistedSnapshotsTest-1875841935-project-member] [instance: aa2716a9-2cf8-4e51-800d-1c8f98b2cf05] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 625.162687] env[65726]: DEBUG oslo_vmware.api [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]521cd967-b6a2-be77-be93-102c14e4f2a6, 'name': SearchDatastore_Task, 'duration_secs': 0.011916} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 625.163091] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 625.163960] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore2] 8493ba52-b950-4727-970a-19b1797b2ca1/8493ba52-b950-4727-970a-19b1797b2ca1.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 625.163960] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-320fc07a-3de5-4753-8560-144d15286bd9 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.171329] env[65726]: DEBUG oslo_vmware.api [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Waiting for the task: (returnval){ [ 625.171329] env[65726]: value = "task-5115423" [ 625.171329] env[65726]: _type = "Task" [ 625.171329] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 625.181636] env[65726]: DEBUG oslo_vmware.api [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Task: {'id': task-5115423, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 625.251539] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115420, 'name': CreateVM_Task, 'duration_secs': 0.426172} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 625.251734] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4d4dc88b-0d6c-430c-ab4f-c81a56e4657b] Created VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 625.252148] env[65726]: DEBUG oslo_concurrency.lockutils [None req-d2d75a44-1ded-480a-a8b4-89859e465e33 tempest-ServerDiagnosticsV248Test-347810772 tempest-ServerDiagnosticsV248Test-347810772-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 625.252391] env[65726]: DEBUG oslo_concurrency.lockutils [None req-d2d75a44-1ded-480a-a8b4-89859e465e33 tempest-ServerDiagnosticsV248Test-347810772 tempest-ServerDiagnosticsV248Test-347810772-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 625.252717] env[65726]: DEBUG oslo_concurrency.lockutils [None req-d2d75a44-1ded-480a-a8b4-89859e465e33 tempest-ServerDiagnosticsV248Test-347810772 tempest-ServerDiagnosticsV248Test-347810772-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 625.253127] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-08460d6b-3faa-4b17-9c3c-1d73c466fc8b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.258956] env[65726]: DEBUG oslo_vmware.api [None req-d2d75a44-1ded-480a-a8b4-89859e465e33 tempest-ServerDiagnosticsV248Test-347810772 tempest-ServerDiagnosticsV248Test-347810772-project-member] Waiting for the task: (returnval){ [ 625.258956] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]527b41ca-5148-d28c-4d4f-d945f6679dca" [ 625.258956] env[65726]: _type = "Task" [ 625.258956] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 625.270296] env[65726]: DEBUG oslo_vmware.api [None req-d2d75a44-1ded-480a-a8b4-89859e465e33 tempest-ServerDiagnosticsV248Test-347810772 tempest-ServerDiagnosticsV248Test-347810772-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]527b41ca-5148-d28c-4d4f-d945f6679dca, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 625.571345] env[65726]: DEBUG oslo_vmware.api [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Task: {'id': task-5115421, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.079539} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 625.571704] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] [instance: 703fff2d-5aec-4ed1-b987-9e1cc3a25f67] Extended root virtual disk {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 625.572453] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56a6213e-53dc-4753-a6e3-29999be30a2a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.578258] env[65726]: WARNING openstack [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 625.578774] env[65726]: WARNING openstack [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 625.608729] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] [instance: 703fff2d-5aec-4ed1-b987-9e1cc3a25f67] Reconfiguring VM instance instance-00000005 to attach disk [datastore2] 703fff2d-5aec-4ed1-b987-9e1cc3a25f67/703fff2d-5aec-4ed1-b987-9e1cc3a25f67.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 625.609132] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ebfd69a7-e28e-4dce-bbe5-e968b45cc68b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.634351] env[65726]: DEBUG oslo_vmware.api [None req-9da01a84-79f6-4bc1-8b36-334b89c8de59 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Task: {'id': task-5115422, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.481033} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 625.636042] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-9da01a84-79f6-4bc1-8b36-334b89c8de59 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore2] ee676e3b-3326-46a3-940d-ebbb8b108991/ee676e3b-3326-46a3-940d-ebbb8b108991.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 625.636578] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-9da01a84-79f6-4bc1-8b36-334b89c8de59 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] [instance: ee676e3b-3326-46a3-940d-ebbb8b108991] Extending root virtual disk to 1048576 {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 625.636703] env[65726]: DEBUG oslo_vmware.api [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Waiting for the task: (returnval){ [ 625.636703] env[65726]: value = "task-5115424" [ 625.636703] env[65726]: _type = "Task" [ 625.636703] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 625.636919] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c5d7b1d8-609f-4cde-b2e4-43d3874280a4 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.646721] env[65726]: DEBUG nova.compute.utils [None req-b82a7f4d-3603-48cd-8e48-1d790e39dbfb tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Using /dev/sd instead of None {{(pid=65726) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 625.651497] env[65726]: DEBUG nova.compute.manager [None req-b82a7f4d-3603-48cd-8e48-1d790e39dbfb tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] [instance: cd539d8b-921b-4947-954c-8573380dbdc8] Allocating IP information in the background. {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 625.651764] env[65726]: DEBUG nova.network.neutron [None req-b82a7f4d-3603-48cd-8e48-1d790e39dbfb tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] [instance: cd539d8b-921b-4947-954c-8573380dbdc8] allocate_for_instance() {{(pid=65726) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 625.652287] env[65726]: WARNING neutronclient.v2_0.client [None req-b82a7f4d-3603-48cd-8e48-1d790e39dbfb tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 625.652652] env[65726]: WARNING neutronclient.v2_0.client [None req-b82a7f4d-3603-48cd-8e48-1d790e39dbfb tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 625.653789] env[65726]: WARNING openstack [None req-b82a7f4d-3603-48cd-8e48-1d790e39dbfb tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 625.653789] env[65726]: WARNING openstack [None req-b82a7f4d-3603-48cd-8e48-1d790e39dbfb tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 625.662703] env[65726]: DEBUG oslo_vmware.api [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Task: {'id': task-5115424, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 625.665338] env[65726]: DEBUG oslo_vmware.api [None req-9da01a84-79f6-4bc1-8b36-334b89c8de59 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Waiting for the task: (returnval){ [ 625.665338] env[65726]: value = "task-5115425" [ 625.665338] env[65726]: _type = "Task" [ 625.665338] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 625.681202] env[65726]: DEBUG oslo_vmware.api [None req-9da01a84-79f6-4bc1-8b36-334b89c8de59 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Task: {'id': task-5115425, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 625.688453] env[65726]: DEBUG oslo_vmware.api [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Task: {'id': task-5115423, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 625.770791] env[65726]: DEBUG oslo_vmware.api [None req-d2d75a44-1ded-480a-a8b4-89859e465e33 tempest-ServerDiagnosticsV248Test-347810772 tempest-ServerDiagnosticsV248Test-347810772-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]527b41ca-5148-d28c-4d4f-d945f6679dca, 'name': SearchDatastore_Task, 'duration_secs': 0.010023} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 625.771188] env[65726]: DEBUG oslo_concurrency.lockutils [None req-d2d75a44-1ded-480a-a8b4-89859e465e33 tempest-ServerDiagnosticsV248Test-347810772 tempest-ServerDiagnosticsV248Test-347810772-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 625.771467] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-d2d75a44-1ded-480a-a8b4-89859e465e33 tempest-ServerDiagnosticsV248Test-347810772 tempest-ServerDiagnosticsV248Test-347810772-project-member] [instance: 4d4dc88b-0d6c-430c-ab4f-c81a56e4657b] Processing image b52362a3-ee8a-4cbf-b06f-513b0cc8f95c {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 625.771723] env[65726]: DEBUG oslo_concurrency.lockutils [None req-d2d75a44-1ded-480a-a8b4-89859e465e33 tempest-ServerDiagnosticsV248Test-347810772 tempest-ServerDiagnosticsV248Test-347810772-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 625.771861] env[65726]: DEBUG oslo_concurrency.lockutils [None req-d2d75a44-1ded-480a-a8b4-89859e465e33 tempest-ServerDiagnosticsV248Test-347810772 tempest-ServerDiagnosticsV248Test-347810772-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 625.772045] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-d2d75a44-1ded-480a-a8b4-89859e465e33 tempest-ServerDiagnosticsV248Test-347810772 tempest-ServerDiagnosticsV248Test-347810772-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 625.772373] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7269f9d6-2678-44cc-9ec9-1e0b0cba9fb7 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.788354] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-d2d75a44-1ded-480a-a8b4-89859e465e33 tempest-ServerDiagnosticsV248Test-347810772 tempest-ServerDiagnosticsV248Test-347810772-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 625.788649] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-d2d75a44-1ded-480a-a8b4-89859e465e33 tempest-ServerDiagnosticsV248Test-347810772 tempest-ServerDiagnosticsV248Test-347810772-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65726) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 625.789472] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cf430594-8dc5-46e8-a1f6-891875f63082 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.796302] env[65726]: DEBUG oslo_vmware.api [None req-d2d75a44-1ded-480a-a8b4-89859e465e33 tempest-ServerDiagnosticsV248Test-347810772 tempest-ServerDiagnosticsV248Test-347810772-project-member] Waiting for the task: (returnval){ [ 625.796302] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52e62a6d-77a8-e275-3199-fd64c5644e5f" [ 625.796302] env[65726]: _type = "Task" [ 625.796302] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 625.807052] env[65726]: DEBUG oslo_vmware.api [None req-d2d75a44-1ded-480a-a8b4-89859e465e33 tempest-ServerDiagnosticsV248Test-347810772 tempest-ServerDiagnosticsV248Test-347810772-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52e62a6d-77a8-e275-3199-fd64c5644e5f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 626.150447] env[65726]: DEBUG oslo_vmware.api [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Task: {'id': task-5115424, 'name': ReconfigVM_Task, 'duration_secs': 0.4411} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 626.152107] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] [instance: 703fff2d-5aec-4ed1-b987-9e1cc3a25f67] Reconfigured VM instance instance-00000005 to attach disk [datastore2] 703fff2d-5aec-4ed1-b987-9e1cc3a25f67/703fff2d-5aec-4ed1-b987-9e1cc3a25f67.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 626.152107] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-38129e50-553c-445b-989b-6c248e561707 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.159791] env[65726]: DEBUG oslo_vmware.api [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Waiting for the task: (returnval){ [ 626.159791] env[65726]: value = "task-5115426" [ 626.159791] env[65726]: _type = "Task" [ 626.159791] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 626.164839] env[65726]: DEBUG nova.compute.manager [None req-b82a7f4d-3603-48cd-8e48-1d790e39dbfb tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] [instance: cd539d8b-921b-4947-954c-8573380dbdc8] Start building block device mappings for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 626.175114] env[65726]: DEBUG oslo_vmware.api [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Task: {'id': task-5115426, 'name': Rename_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 626.189813] env[65726]: DEBUG oslo_vmware.api [None req-9da01a84-79f6-4bc1-8b36-334b89c8de59 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Task: {'id': task-5115425, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.098243} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 626.193432] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-9da01a84-79f6-4bc1-8b36-334b89c8de59 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] [instance: ee676e3b-3326-46a3-940d-ebbb8b108991] Extended root virtual disk {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 626.193859] env[65726]: DEBUG oslo_vmware.api [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Task: {'id': task-5115423, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.734} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 626.194902] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-863de601-816f-4e7b-a1ce-9655ec94c2f8 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.197311] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore2] 8493ba52-b950-4727-970a-19b1797b2ca1/8493ba52-b950-4727-970a-19b1797b2ca1.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 626.197578] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] [instance: 8493ba52-b950-4727-970a-19b1797b2ca1] Extending root virtual disk to 1048576 {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 626.197816] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-92577286-b23a-44ad-a6e9-23b784ffb12d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.222128] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-9da01a84-79f6-4bc1-8b36-334b89c8de59 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] [instance: ee676e3b-3326-46a3-940d-ebbb8b108991] Reconfiguring VM instance instance-00000006 to attach disk [datastore2] ee676e3b-3326-46a3-940d-ebbb8b108991/ee676e3b-3326-46a3-940d-ebbb8b108991.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 626.228056] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-855e87ef-7c2e-452c-aa31-7f1a838e1b25 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.244229] env[65726]: DEBUG oslo_vmware.api [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Waiting for the task: (returnval){ [ 626.244229] env[65726]: value = "task-5115427" [ 626.244229] env[65726]: _type = "Task" [ 626.244229] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 626.253595] env[65726]: DEBUG oslo_vmware.api [None req-9da01a84-79f6-4bc1-8b36-334b89c8de59 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Waiting for the task: (returnval){ [ 626.253595] env[65726]: value = "task-5115428" [ 626.253595] env[65726]: _type = "Task" [ 626.253595] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 626.258877] env[65726]: DEBUG oslo_vmware.api [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Task: {'id': task-5115427, 'name': ExtendVirtualDisk_Task} progress is 50%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 626.271189] env[65726]: DEBUG oslo_vmware.api [None req-9da01a84-79f6-4bc1-8b36-334b89c8de59 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Task: {'id': task-5115428, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 626.310636] env[65726]: DEBUG oslo_vmware.api [None req-d2d75a44-1ded-480a-a8b4-89859e465e33 tempest-ServerDiagnosticsV248Test-347810772 tempest-ServerDiagnosticsV248Test-347810772-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52e62a6d-77a8-e275-3199-fd64c5644e5f, 'name': SearchDatastore_Task, 'duration_secs': 0.058505} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 626.311460] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-de938033-0f37-4271-b3fc-999cb53a4020 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.319168] env[65726]: DEBUG oslo_vmware.api [None req-d2d75a44-1ded-480a-a8b4-89859e465e33 tempest-ServerDiagnosticsV248Test-347810772 tempest-ServerDiagnosticsV248Test-347810772-project-member] Waiting for the task: (returnval){ [ 626.319168] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]527dda12-e691-7874-1dd0-1b54ecf44081" [ 626.319168] env[65726]: _type = "Task" [ 626.319168] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 626.334635] env[65726]: DEBUG oslo_vmware.api [None req-d2d75a44-1ded-480a-a8b4-89859e465e33 tempest-ServerDiagnosticsV248Test-347810772 tempest-ServerDiagnosticsV248Test-347810772-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]527dda12-e691-7874-1dd0-1b54ecf44081, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 626.653116] env[65726]: WARNING openstack [req-93aa6f4f-9c10-41ef-b81e-47851f66fb89 req-7d8801ff-3d1b-481c-94d5-8c54ba593b16 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 626.653116] env[65726]: WARNING openstack [req-93aa6f4f-9c10-41ef-b81e-47851f66fb89 req-7d8801ff-3d1b-481c-94d5-8c54ba593b16 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 626.666033] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-220655ee-d00e-44bf-94c8-63a6e8037142 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.679942] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61cdb288-a475-4437-9599-1526716d1276 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.683812] env[65726]: DEBUG oslo_vmware.api [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Task: {'id': task-5115426, 'name': Rename_Task, 'duration_secs': 0.158982} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 626.684437] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] [instance: 703fff2d-5aec-4ed1-b987-9e1cc3a25f67] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 626.684736] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1a7de0fb-d86c-4f6a-8d2c-8c79f70e5fef {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.727886] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5956a6d-8661-4aba-8774-840e7d1cde33 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.731273] env[65726]: DEBUG oslo_vmware.api [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Waiting for the task: (returnval){ [ 626.731273] env[65726]: value = "task-5115429" [ 626.731273] env[65726]: _type = "Task" [ 626.731273] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 626.744275] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ff32447-35fe-4105-b9c6-228cc4e5dfd4 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.751952] env[65726]: DEBUG oslo_vmware.api [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Task: {'id': task-5115429, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 626.766431] env[65726]: DEBUG nova.compute.provider_tree [None req-b8e65ecf-6b31-4881-a7b5-2f0a252ee6d8 tempest-VolumesAssistedSnapshotsTest-1875841935 tempest-VolumesAssistedSnapshotsTest-1875841935-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 626.777029] env[65726]: DEBUG oslo_vmware.api [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Task: {'id': task-5115427, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.076541} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 626.777029] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] [instance: 8493ba52-b950-4727-970a-19b1797b2ca1] Extended root virtual disk {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 626.777029] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71a335e4-d452-4861-880c-bbbca0a16508 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.784237] env[65726]: DEBUG oslo_vmware.api [None req-9da01a84-79f6-4bc1-8b36-334b89c8de59 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Task: {'id': task-5115428, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 626.806717] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] [instance: 8493ba52-b950-4727-970a-19b1797b2ca1] Reconfiguring VM instance instance-00000007 to attach disk [datastore2] 8493ba52-b950-4727-970a-19b1797b2ca1/8493ba52-b950-4727-970a-19b1797b2ca1.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 626.810630] env[65726]: DEBUG nova.policy [None req-b82a7f4d-3603-48cd-8e48-1d790e39dbfb tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a78a1cdb82644c1abe2b0ab2e0f621a5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3a680eae0e734ea8b2274b504d0a46d8', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65726) authorize /opt/stack/nova/nova/policy.py:192}} [ 626.813891] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f4a843fd-a994-4876-8c3d-cae6e8982a3a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.830898] env[65726]: WARNING openstack [req-fde7dfe2-2f08-46e4-af18-49e8c5f6cc8c req-32b5ea0b-1a90-4298-8ffd-9e3c4c1c1ac2 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 626.831256] env[65726]: WARNING openstack [req-fde7dfe2-2f08-46e4-af18-49e8c5f6cc8c req-32b5ea0b-1a90-4298-8ffd-9e3c4c1c1ac2 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 626.850958] env[65726]: DEBUG oslo_vmware.api [None req-d2d75a44-1ded-480a-a8b4-89859e465e33 tempest-ServerDiagnosticsV248Test-347810772 tempest-ServerDiagnosticsV248Test-347810772-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]527dda12-e691-7874-1dd0-1b54ecf44081, 'name': SearchDatastore_Task, 'duration_secs': 0.037883} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 626.853039] env[65726]: DEBUG oslo_concurrency.lockutils [None req-d2d75a44-1ded-480a-a8b4-89859e465e33 tempest-ServerDiagnosticsV248Test-347810772 tempest-ServerDiagnosticsV248Test-347810772-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 626.853147] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-d2d75a44-1ded-480a-a8b4-89859e465e33 tempest-ServerDiagnosticsV248Test-347810772 tempest-ServerDiagnosticsV248Test-347810772-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore2] 4d4dc88b-0d6c-430c-ab4f-c81a56e4657b/4d4dc88b-0d6c-430c-ab4f-c81a56e4657b.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 626.853474] env[65726]: DEBUG oslo_vmware.api [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Waiting for the task: (returnval){ [ 626.853474] env[65726]: value = "task-5115430" [ 626.853474] env[65726]: _type = "Task" [ 626.853474] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 626.853671] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ab5c22e7-c95a-4a6b-874f-2632d123a03c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.867724] env[65726]: DEBUG oslo_vmware.api [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Task: {'id': task-5115430, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 626.871130] env[65726]: DEBUG oslo_vmware.api [None req-d2d75a44-1ded-480a-a8b4-89859e465e33 tempest-ServerDiagnosticsV248Test-347810772 tempest-ServerDiagnosticsV248Test-347810772-project-member] Waiting for the task: (returnval){ [ 626.871130] env[65726]: value = "task-5115431" [ 626.871130] env[65726]: _type = "Task" [ 626.871130] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 626.890934] env[65726]: DEBUG oslo_vmware.api [None req-d2d75a44-1ded-480a-a8b4-89859e465e33 tempest-ServerDiagnosticsV248Test-347810772 tempest-ServerDiagnosticsV248Test-347810772-project-member] Task: {'id': task-5115431, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 627.174309] env[65726]: DEBUG nova.compute.manager [None req-b82a7f4d-3603-48cd-8e48-1d790e39dbfb tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] [instance: cd539d8b-921b-4947-954c-8573380dbdc8] Start spawning the instance on the hypervisor. {{(pid=65726) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 627.204242] env[65726]: DEBUG nova.virt.hardware [None req-b82a7f4d-3603-48cd-8e48-1d790e39dbfb tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T19:28:34Z,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1539f06fa3534e90acbc3a60c4b8bd3f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T19:28:35Z,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 627.204533] env[65726]: DEBUG nova.virt.hardware [None req-b82a7f4d-3603-48cd-8e48-1d790e39dbfb tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 627.204685] env[65726]: DEBUG nova.virt.hardware [None req-b82a7f4d-3603-48cd-8e48-1d790e39dbfb tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 627.205051] env[65726]: DEBUG nova.virt.hardware [None req-b82a7f4d-3603-48cd-8e48-1d790e39dbfb tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 627.205198] env[65726]: DEBUG nova.virt.hardware [None req-b82a7f4d-3603-48cd-8e48-1d790e39dbfb tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 627.205354] env[65726]: DEBUG nova.virt.hardware [None req-b82a7f4d-3603-48cd-8e48-1d790e39dbfb tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 627.205553] env[65726]: DEBUG nova.virt.hardware [None req-b82a7f4d-3603-48cd-8e48-1d790e39dbfb tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 627.205708] env[65726]: DEBUG nova.virt.hardware [None req-b82a7f4d-3603-48cd-8e48-1d790e39dbfb tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 627.205881] env[65726]: DEBUG nova.virt.hardware [None req-b82a7f4d-3603-48cd-8e48-1d790e39dbfb tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 627.206049] env[65726]: DEBUG nova.virt.hardware [None req-b82a7f4d-3603-48cd-8e48-1d790e39dbfb tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 627.206301] env[65726]: DEBUG nova.virt.hardware [None req-b82a7f4d-3603-48cd-8e48-1d790e39dbfb tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 627.207179] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e19e66ad-9edf-4b76-ad53-f32e88984b56 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.216833] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a047610d-5d56-4b2c-bab1-20aed4ca8bc9 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.243855] env[65726]: DEBUG oslo_vmware.api [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Task: {'id': task-5115429, 'name': PowerOnVM_Task, 'duration_secs': 0.508628} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 627.244091] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] [instance: 703fff2d-5aec-4ed1-b987-9e1cc3a25f67] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 627.244349] env[65726]: INFO nova.compute.manager [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] [instance: 703fff2d-5aec-4ed1-b987-9e1cc3a25f67] Took 12.90 seconds to spawn the instance on the hypervisor. [ 627.244567] env[65726]: DEBUG nova.compute.manager [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] [instance: 703fff2d-5aec-4ed1-b987-9e1cc3a25f67] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 627.245748] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b63826ea-5629-4ac9-adfb-92cb2cf5dfaa {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.278929] env[65726]: DEBUG nova.scheduler.client.report [None req-b8e65ecf-6b31-4881-a7b5-2f0a252ee6d8 tempest-VolumesAssistedSnapshotsTest-1875841935 tempest-VolumesAssistedSnapshotsTest-1875841935-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 627.283285] env[65726]: DEBUG oslo_vmware.api [None req-9da01a84-79f6-4bc1-8b36-334b89c8de59 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Task: {'id': task-5115428, 'name': ReconfigVM_Task, 'duration_secs': 0.657147} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 627.283623] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-9da01a84-79f6-4bc1-8b36-334b89c8de59 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] [instance: ee676e3b-3326-46a3-940d-ebbb8b108991] Reconfigured VM instance instance-00000006 to attach disk [datastore2] ee676e3b-3326-46a3-940d-ebbb8b108991/ee676e3b-3326-46a3-940d-ebbb8b108991.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 627.284338] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-70c2daa2-df05-42c8-b72d-87ef7edb0578 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.297282] env[65726]: DEBUG oslo_vmware.api [None req-9da01a84-79f6-4bc1-8b36-334b89c8de59 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Waiting for the task: (returnval){ [ 627.297282] env[65726]: value = "task-5115432" [ 627.297282] env[65726]: _type = "Task" [ 627.297282] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 627.310375] env[65726]: DEBUG oslo_vmware.api [None req-9da01a84-79f6-4bc1-8b36-334b89c8de59 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Task: {'id': task-5115432, 'name': Rename_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 627.367608] env[65726]: DEBUG oslo_vmware.api [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Task: {'id': task-5115430, 'name': ReconfigVM_Task, 'duration_secs': 0.352777} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 627.367916] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] [instance: 8493ba52-b950-4727-970a-19b1797b2ca1] Reconfigured VM instance instance-00000007 to attach disk [datastore2] 8493ba52-b950-4727-970a-19b1797b2ca1/8493ba52-b950-4727-970a-19b1797b2ca1.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 627.368654] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4b84a608-8400-451b-9283-53bfa4c9305f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.379702] env[65726]: DEBUG oslo_vmware.api [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Waiting for the task: (returnval){ [ 627.379702] env[65726]: value = "task-5115433" [ 627.379702] env[65726]: _type = "Task" [ 627.379702] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 627.385443] env[65726]: DEBUG oslo_vmware.api [None req-d2d75a44-1ded-480a-a8b4-89859e465e33 tempest-ServerDiagnosticsV248Test-347810772 tempest-ServerDiagnosticsV248Test-347810772-project-member] Task: {'id': task-5115431, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 627.392307] env[65726]: DEBUG oslo_vmware.api [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Task: {'id': task-5115433, 'name': Rename_Task} progress is 6%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 627.730969] env[65726]: DEBUG nova.network.neutron [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] [instance: 21730277-cef6-4bfe-9f67-de71f6f615ff] Updating instance_info_cache with network_info: [{"id": "cc7e0970-7e47-45d6-b7fa-6e33947b210c", "address": "fa:16:3e:60:60:cb", "network": {"id": "c68b8bcc-840e-459b-bee0-76a9ffdc08cb", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-2061916615-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7c3641c4a34b4bf8b09403220e153e65", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "aa0d7fd6-4fd3-4451-9208-8a3cf25e30e4", "external-id": "nsx-vlan-transportzone-545", "segmentation_id": 545, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcc7e0970-7e", "ovs_interfaceid": "cc7e0970-7e47-45d6-b7fa-6e33947b210c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 627.769829] env[65726]: INFO nova.compute.manager [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] [instance: 703fff2d-5aec-4ed1-b987-9e1cc3a25f67] Took 27.33 seconds to build instance. [ 627.785859] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b8e65ecf-6b31-4881-a7b5-2f0a252ee6d8 tempest-VolumesAssistedSnapshotsTest-1875841935 tempest-VolumesAssistedSnapshotsTest-1875841935-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.639s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 627.787379] env[65726]: DEBUG nova.compute.manager [None req-b8e65ecf-6b31-4881-a7b5-2f0a252ee6d8 tempest-VolumesAssistedSnapshotsTest-1875841935 tempest-VolumesAssistedSnapshotsTest-1875841935-project-member] [instance: aa2716a9-2cf8-4e51-800d-1c8f98b2cf05] Start building networks asynchronously for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 627.795509] env[65726]: DEBUG oslo_concurrency.lockutils [None req-8fd7433e-ebac-43f4-aa60-6301be278066 tempest-ServerDiagnosticsTest-2021442419 tempest-ServerDiagnosticsTest-2021442419-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.037s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 627.798162] env[65726]: INFO nova.compute.claims [None req-8fd7433e-ebac-43f4-aa60-6301be278066 tempest-ServerDiagnosticsTest-2021442419 tempest-ServerDiagnosticsTest-2021442419-project-member] [instance: de5cad99-1dbc-4435-98b4-987e24ba8c3c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 627.812452] env[65726]: DEBUG oslo_vmware.api [None req-9da01a84-79f6-4bc1-8b36-334b89c8de59 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Task: {'id': task-5115432, 'name': Rename_Task, 'duration_secs': 0.242999} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 627.812789] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-9da01a84-79f6-4bc1-8b36-334b89c8de59 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] [instance: ee676e3b-3326-46a3-940d-ebbb8b108991] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 627.813082] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9acb4744-bfc8-4fd1-9493-898a92e4e065 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.823439] env[65726]: DEBUG oslo_vmware.api [None req-9da01a84-79f6-4bc1-8b36-334b89c8de59 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Waiting for the task: (returnval){ [ 627.823439] env[65726]: value = "task-5115434" [ 627.823439] env[65726]: _type = "Task" [ 627.823439] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 627.834408] env[65726]: DEBUG oslo_vmware.api [None req-9da01a84-79f6-4bc1-8b36-334b89c8de59 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Task: {'id': task-5115434, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 627.900747] env[65726]: DEBUG oslo_vmware.api [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Task: {'id': task-5115433, 'name': Rename_Task, 'duration_secs': 0.16065} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 627.900747] env[65726]: DEBUG oslo_vmware.api [None req-d2d75a44-1ded-480a-a8b4-89859e465e33 tempest-ServerDiagnosticsV248Test-347810772 tempest-ServerDiagnosticsV248Test-347810772-project-member] Task: {'id': task-5115431, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.556515} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 627.900747] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] [instance: 8493ba52-b950-4727-970a-19b1797b2ca1] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 627.900747] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-d2d75a44-1ded-480a-a8b4-89859e465e33 tempest-ServerDiagnosticsV248Test-347810772 tempest-ServerDiagnosticsV248Test-347810772-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore2] 4d4dc88b-0d6c-430c-ab4f-c81a56e4657b/4d4dc88b-0d6c-430c-ab4f-c81a56e4657b.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 627.901133] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-d2d75a44-1ded-480a-a8b4-89859e465e33 tempest-ServerDiagnosticsV248Test-347810772 tempest-ServerDiagnosticsV248Test-347810772-project-member] [instance: 4d4dc88b-0d6c-430c-ab4f-c81a56e4657b] Extending root virtual disk to 1048576 {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 627.901133] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-403a8674-67fb-4691-b64f-d0acdfaa1b08 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.902207] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3ae75373-c050-4446-a12b-3c694be9f012 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.910557] env[65726]: DEBUG oslo_vmware.api [None req-d2d75a44-1ded-480a-a8b4-89859e465e33 tempest-ServerDiagnosticsV248Test-347810772 tempest-ServerDiagnosticsV248Test-347810772-project-member] Waiting for the task: (returnval){ [ 627.910557] env[65726]: value = "task-5115436" [ 627.910557] env[65726]: _type = "Task" [ 627.910557] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 627.912601] env[65726]: DEBUG oslo_vmware.api [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Waiting for the task: (returnval){ [ 627.912601] env[65726]: value = "task-5115435" [ 627.912601] env[65726]: _type = "Task" [ 627.912601] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 627.927814] env[65726]: DEBUG oslo_vmware.api [None req-d2d75a44-1ded-480a-a8b4-89859e465e33 tempest-ServerDiagnosticsV248Test-347810772 tempest-ServerDiagnosticsV248Test-347810772-project-member] Task: {'id': task-5115436, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 627.933360] env[65726]: DEBUG oslo_vmware.api [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Task: {'id': task-5115435, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 628.123828] env[65726]: DEBUG nova.network.neutron [None req-b82a7f4d-3603-48cd-8e48-1d790e39dbfb tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] [instance: cd539d8b-921b-4947-954c-8573380dbdc8] Successfully created port: 60d92cf3-683a-4291-ba51-fad4124f3c8c {{(pid=65726) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 628.235918] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Releasing lock "refresh_cache-21730277-cef6-4bfe-9f67-de71f6f615ff" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 628.235918] env[65726]: DEBUG nova.compute.manager [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] [instance: 21730277-cef6-4bfe-9f67-de71f6f615ff] Instance network_info: |[{"id": "cc7e0970-7e47-45d6-b7fa-6e33947b210c", "address": "fa:16:3e:60:60:cb", "network": {"id": "c68b8bcc-840e-459b-bee0-76a9ffdc08cb", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-2061916615-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7c3641c4a34b4bf8b09403220e153e65", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "aa0d7fd6-4fd3-4451-9208-8a3cf25e30e4", "external-id": "nsx-vlan-transportzone-545", "segmentation_id": 545, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcc7e0970-7e", "ovs_interfaceid": "cc7e0970-7e47-45d6-b7fa-6e33947b210c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 628.236102] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] [instance: 21730277-cef6-4bfe-9f67-de71f6f615ff] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:60:60:cb', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'aa0d7fd6-4fd3-4451-9208-8a3cf25e30e4', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'cc7e0970-7e47-45d6-b7fa-6e33947b210c', 'vif_model': 'vmxnet3'}] {{(pid=65726) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 628.245373] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 628.245815] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 21730277-cef6-4bfe-9f67-de71f6f615ff] Creating VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 628.247867] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-54ab3748-4dbb-43f0-8378-fc2d9458dcf9 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.272671] env[65726]: DEBUG oslo_concurrency.lockutils [None req-932e1a19-3446-45c8-978f-c65749a1d5a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Lock "703fff2d-5aec-4ed1-b987-9e1cc3a25f67" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 28.847s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 628.275623] env[65726]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 628.275623] env[65726]: value = "task-5115437" [ 628.275623] env[65726]: _type = "Task" [ 628.275623] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 628.291250] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115437, 'name': CreateVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 628.296645] env[65726]: DEBUG nova.compute.utils [None req-b8e65ecf-6b31-4881-a7b5-2f0a252ee6d8 tempest-VolumesAssistedSnapshotsTest-1875841935 tempest-VolumesAssistedSnapshotsTest-1875841935-project-member] Using /dev/sd instead of None {{(pid=65726) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 628.298931] env[65726]: DEBUG nova.compute.manager [None req-b8e65ecf-6b31-4881-a7b5-2f0a252ee6d8 tempest-VolumesAssistedSnapshotsTest-1875841935 tempest-VolumesAssistedSnapshotsTest-1875841935-project-member] [instance: aa2716a9-2cf8-4e51-800d-1c8f98b2cf05] Allocating IP information in the background. {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 628.299957] env[65726]: DEBUG nova.network.neutron [None req-b8e65ecf-6b31-4881-a7b5-2f0a252ee6d8 tempest-VolumesAssistedSnapshotsTest-1875841935 tempest-VolumesAssistedSnapshotsTest-1875841935-project-member] [instance: aa2716a9-2cf8-4e51-800d-1c8f98b2cf05] allocate_for_instance() {{(pid=65726) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 628.300536] env[65726]: WARNING neutronclient.v2_0.client [None req-b8e65ecf-6b31-4881-a7b5-2f0a252ee6d8 tempest-VolumesAssistedSnapshotsTest-1875841935 tempest-VolumesAssistedSnapshotsTest-1875841935-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 628.300939] env[65726]: WARNING neutronclient.v2_0.client [None req-b8e65ecf-6b31-4881-a7b5-2f0a252ee6d8 tempest-VolumesAssistedSnapshotsTest-1875841935 tempest-VolumesAssistedSnapshotsTest-1875841935-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 628.301616] env[65726]: WARNING openstack [None req-b8e65ecf-6b31-4881-a7b5-2f0a252ee6d8 tempest-VolumesAssistedSnapshotsTest-1875841935 tempest-VolumesAssistedSnapshotsTest-1875841935-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 628.302036] env[65726]: WARNING openstack [None req-b8e65ecf-6b31-4881-a7b5-2f0a252ee6d8 tempest-VolumesAssistedSnapshotsTest-1875841935 tempest-VolumesAssistedSnapshotsTest-1875841935-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 628.338412] env[65726]: DEBUG oslo_vmware.api [None req-9da01a84-79f6-4bc1-8b36-334b89c8de59 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Task: {'id': task-5115434, 'name': PowerOnVM_Task, 'duration_secs': 0.513353} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 628.338756] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-9da01a84-79f6-4bc1-8b36-334b89c8de59 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] [instance: ee676e3b-3326-46a3-940d-ebbb8b108991] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 628.339049] env[65726]: INFO nova.compute.manager [None req-9da01a84-79f6-4bc1-8b36-334b89c8de59 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] [instance: ee676e3b-3326-46a3-940d-ebbb8b108991] Took 11.50 seconds to spawn the instance on the hypervisor. [ 628.339345] env[65726]: DEBUG nova.compute.manager [None req-9da01a84-79f6-4bc1-8b36-334b89c8de59 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] [instance: ee676e3b-3326-46a3-940d-ebbb8b108991] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 628.340377] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c96ede1-b8f9-4fcb-aea6-14b55d44f19c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.430363] env[65726]: DEBUG oslo_vmware.api [None req-d2d75a44-1ded-480a-a8b4-89859e465e33 tempest-ServerDiagnosticsV248Test-347810772 tempest-ServerDiagnosticsV248Test-347810772-project-member] Task: {'id': task-5115436, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.079947} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 628.435372] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-d2d75a44-1ded-480a-a8b4-89859e465e33 tempest-ServerDiagnosticsV248Test-347810772 tempest-ServerDiagnosticsV248Test-347810772-project-member] [instance: 4d4dc88b-0d6c-430c-ab4f-c81a56e4657b] Extended root virtual disk {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 628.435372] env[65726]: DEBUG oslo_vmware.api [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Task: {'id': task-5115435, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 628.436469] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2ec9eef-8e41-4a7b-84f4-442287067233 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.467341] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-d2d75a44-1ded-480a-a8b4-89859e465e33 tempest-ServerDiagnosticsV248Test-347810772 tempest-ServerDiagnosticsV248Test-347810772-project-member] [instance: 4d4dc88b-0d6c-430c-ab4f-c81a56e4657b] Reconfiguring VM instance instance-00000009 to attach disk [datastore2] 4d4dc88b-0d6c-430c-ab4f-c81a56e4657b/4d4dc88b-0d6c-430c-ab4f-c81a56e4657b.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 628.468165] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5b5472bf-97fd-4cda-b392-d5f016fdf720 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.486668] env[65726]: DEBUG nova.network.neutron [req-93aa6f4f-9c10-41ef-b81e-47851f66fb89 req-7d8801ff-3d1b-481c-94d5-8c54ba593b16 service nova] [instance: ee676e3b-3326-46a3-940d-ebbb8b108991] Updated VIF entry in instance network info cache for port 30651f8a-abdc-487b-9044-1c8a17beab23. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 628.486771] env[65726]: DEBUG nova.network.neutron [req-93aa6f4f-9c10-41ef-b81e-47851f66fb89 req-7d8801ff-3d1b-481c-94d5-8c54ba593b16 service nova] [instance: ee676e3b-3326-46a3-940d-ebbb8b108991] Updating instance_info_cache with network_info: [{"id": "30651f8a-abdc-487b-9044-1c8a17beab23", "address": "fa:16:3e:13:1f:1c", "network": {"id": "55d629b8-a541-4720-9289-927bbc6e6f22", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.53", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "1539f06fa3534e90acbc3a60c4b8bd3f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "35342bcb-8b06-472e-b3c0-43fd3d6c4b30", "external-id": "nsx-vlan-transportzone-524", "segmentation_id": 524, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap30651f8a-ab", "ovs_interfaceid": "30651f8a-abdc-487b-9044-1c8a17beab23", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 628.495515] env[65726]: DEBUG oslo_vmware.api [None req-d2d75a44-1ded-480a-a8b4-89859e465e33 tempest-ServerDiagnosticsV248Test-347810772 tempest-ServerDiagnosticsV248Test-347810772-project-member] Waiting for the task: (returnval){ [ 628.495515] env[65726]: value = "task-5115438" [ 628.495515] env[65726]: _type = "Task" [ 628.495515] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 628.505939] env[65726]: DEBUG oslo_vmware.api [None req-d2d75a44-1ded-480a-a8b4-89859e465e33 tempest-ServerDiagnosticsV248Test-347810772 tempest-ServerDiagnosticsV248Test-347810772-project-member] Task: {'id': task-5115438, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 628.521510] env[65726]: WARNING openstack [req-fde7dfe2-2f08-46e4-af18-49e8c5f6cc8c req-32b5ea0b-1a90-4298-8ffd-9e3c4c1c1ac2 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 628.521510] env[65726]: WARNING openstack [req-fde7dfe2-2f08-46e4-af18-49e8c5f6cc8c req-32b5ea0b-1a90-4298-8ffd-9e3c4c1c1ac2 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 628.781667] env[65726]: DEBUG nova.compute.manager [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] [instance: 3a60b719-9db7-436e-9908-25f6a233c465] Starting instance... {{(pid=65726) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 628.795291] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115437, 'name': CreateVM_Task} progress is 25%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 628.814191] env[65726]: DEBUG nova.compute.manager [None req-b8e65ecf-6b31-4881-a7b5-2f0a252ee6d8 tempest-VolumesAssistedSnapshotsTest-1875841935 tempest-VolumesAssistedSnapshotsTest-1875841935-project-member] [instance: aa2716a9-2cf8-4e51-800d-1c8f98b2cf05] Start building block device mappings for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 628.882879] env[65726]: INFO nova.compute.manager [None req-9da01a84-79f6-4bc1-8b36-334b89c8de59 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] [instance: ee676e3b-3326-46a3-940d-ebbb8b108991] Took 23.43 seconds to build instance. [ 628.936419] env[65726]: DEBUG oslo_vmware.api [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Task: {'id': task-5115435, 'name': PowerOnVM_Task, 'duration_secs': 0.555496} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 628.936419] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] [instance: 8493ba52-b950-4727-970a-19b1797b2ca1] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 628.936419] env[65726]: INFO nova.compute.manager [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] [instance: 8493ba52-b950-4727-970a-19b1797b2ca1] Took 10.08 seconds to spawn the instance on the hypervisor. [ 628.936571] env[65726]: DEBUG nova.compute.manager [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] [instance: 8493ba52-b950-4727-970a-19b1797b2ca1] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 628.938560] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b88e0ae-e7f5-484a-85d0-2dcd3c80bd62 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.989724] env[65726]: DEBUG oslo_concurrency.lockutils [req-93aa6f4f-9c10-41ef-b81e-47851f66fb89 req-7d8801ff-3d1b-481c-94d5-8c54ba593b16 service nova] Releasing lock "refresh_cache-ee676e3b-3326-46a3-940d-ebbb8b108991" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 629.015861] env[65726]: DEBUG oslo_vmware.api [None req-d2d75a44-1ded-480a-a8b4-89859e465e33 tempest-ServerDiagnosticsV248Test-347810772 tempest-ServerDiagnosticsV248Test-347810772-project-member] Task: {'id': task-5115438, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 629.229518] env[65726]: DEBUG nova.policy [None req-b8e65ecf-6b31-4881-a7b5-2f0a252ee6d8 tempest-VolumesAssistedSnapshotsTest-1875841935 tempest-VolumesAssistedSnapshotsTest-1875841935-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '14f6d504d8624465bfd17bf5a715a303', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b01c77c483364e23b4a48b1e1c9c9d1c', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65726) authorize /opt/stack/nova/nova/policy.py:192}} [ 629.300415] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-147a51ba-e17e-4336-8748-107e09d25fe1 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.303887] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115437, 'name': CreateVM_Task, 'duration_secs': 0.747373} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 629.304482] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 21730277-cef6-4bfe-9f67-de71f6f615ff] Created VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 629.305187] env[65726]: WARNING openstack [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 629.305708] env[65726]: WARNING openstack [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 629.311474] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 629.312095] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 629.312095] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 629.313149] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 629.315578] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-53ee0e7c-c360-47e3-9c7c-0f16767e7923 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.317966] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a85f3b6e-306b-40fa-a6fa-180c5f826dd3 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.327848] env[65726]: DEBUG oslo_vmware.api [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Waiting for the task: (returnval){ [ 629.327848] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5222ce49-ce97-3b54-a0db-e0a802131e49" [ 629.327848] env[65726]: _type = "Task" [ 629.327848] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 629.366416] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27a5a57e-e10b-4b72-b1d7-966f4a7a05e2 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.374746] env[65726]: DEBUG oslo_vmware.api [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5222ce49-ce97-3b54-a0db-e0a802131e49, 'name': SearchDatastore_Task, 'duration_secs': 0.017033} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 629.375249] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 629.375516] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] [instance: 21730277-cef6-4bfe-9f67-de71f6f615ff] Processing image b52362a3-ee8a-4cbf-b06f-513b0cc8f95c {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 629.375774] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 629.375915] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 629.376100] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 629.376773] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d4376933-311e-43db-81f1-37abef17b920 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.384680] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2805d96-6211-4c00-839e-2b8347d2de0d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.390532] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9da01a84-79f6-4bc1-8b36-334b89c8de59 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Lock "ee676e3b-3326-46a3-940d-ebbb8b108991" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 24.965s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 629.393360] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 629.393532] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65726) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 629.395359] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bdb58507-b96d-4b6f-b43f-d86816636436 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.411419] env[65726]: DEBUG nova.compute.provider_tree [None req-8fd7433e-ebac-43f4-aa60-6301be278066 tempest-ServerDiagnosticsTest-2021442419 tempest-ServerDiagnosticsTest-2021442419-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 629.417024] env[65726]: DEBUG oslo_vmware.api [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Waiting for the task: (returnval){ [ 629.417024] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5251c83e-39bd-2d6c-e189-b024cdd1cd33" [ 629.417024] env[65726]: _type = "Task" [ 629.417024] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 629.426314] env[65726]: DEBUG oslo_vmware.api [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5251c83e-39bd-2d6c-e189-b024cdd1cd33, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 629.466484] env[65726]: INFO nova.compute.manager [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] [instance: 8493ba52-b950-4727-970a-19b1797b2ca1] Took 23.07 seconds to build instance. [ 629.507831] env[65726]: DEBUG oslo_vmware.api [None req-d2d75a44-1ded-480a-a8b4-89859e465e33 tempest-ServerDiagnosticsV248Test-347810772 tempest-ServerDiagnosticsV248Test-347810772-project-member] Task: {'id': task-5115438, 'name': ReconfigVM_Task, 'duration_secs': 0.868674} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 629.508161] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-d2d75a44-1ded-480a-a8b4-89859e465e33 tempest-ServerDiagnosticsV248Test-347810772 tempest-ServerDiagnosticsV248Test-347810772-project-member] [instance: 4d4dc88b-0d6c-430c-ab4f-c81a56e4657b] Reconfigured VM instance instance-00000009 to attach disk [datastore2] 4d4dc88b-0d6c-430c-ab4f-c81a56e4657b/4d4dc88b-0d6c-430c-ab4f-c81a56e4657b.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 629.509216] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3a882168-b2e1-4571-8677-7c1315e74f28 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.516343] env[65726]: DEBUG oslo_vmware.api [None req-d2d75a44-1ded-480a-a8b4-89859e465e33 tempest-ServerDiagnosticsV248Test-347810772 tempest-ServerDiagnosticsV248Test-347810772-project-member] Waiting for the task: (returnval){ [ 629.516343] env[65726]: value = "task-5115439" [ 629.516343] env[65726]: _type = "Task" [ 629.516343] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 629.525157] env[65726]: DEBUG oslo_vmware.api [None req-d2d75a44-1ded-480a-a8b4-89859e465e33 tempest-ServerDiagnosticsV248Test-347810772 tempest-ServerDiagnosticsV248Test-347810772-project-member] Task: {'id': task-5115439, 'name': Rename_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 629.828853] env[65726]: DEBUG nova.compute.manager [None req-b8e65ecf-6b31-4881-a7b5-2f0a252ee6d8 tempest-VolumesAssistedSnapshotsTest-1875841935 tempest-VolumesAssistedSnapshotsTest-1875841935-project-member] [instance: aa2716a9-2cf8-4e51-800d-1c8f98b2cf05] Start spawning the instance on the hypervisor. {{(pid=65726) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 629.867387] env[65726]: DEBUG nova.virt.hardware [None req-b8e65ecf-6b31-4881-a7b5-2f0a252ee6d8 tempest-VolumesAssistedSnapshotsTest-1875841935 tempest-VolumesAssistedSnapshotsTest-1875841935-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T19:28:34Z,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1539f06fa3534e90acbc3a60c4b8bd3f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T19:28:35Z,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 629.867630] env[65726]: DEBUG nova.virt.hardware [None req-b8e65ecf-6b31-4881-a7b5-2f0a252ee6d8 tempest-VolumesAssistedSnapshotsTest-1875841935 tempest-VolumesAssistedSnapshotsTest-1875841935-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 629.867782] env[65726]: DEBUG nova.virt.hardware [None req-b8e65ecf-6b31-4881-a7b5-2f0a252ee6d8 tempest-VolumesAssistedSnapshotsTest-1875841935 tempest-VolumesAssistedSnapshotsTest-1875841935-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 629.867956] env[65726]: DEBUG nova.virt.hardware [None req-b8e65ecf-6b31-4881-a7b5-2f0a252ee6d8 tempest-VolumesAssistedSnapshotsTest-1875841935 tempest-VolumesAssistedSnapshotsTest-1875841935-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 629.868108] env[65726]: DEBUG nova.virt.hardware [None req-b8e65ecf-6b31-4881-a7b5-2f0a252ee6d8 tempest-VolumesAssistedSnapshotsTest-1875841935 tempest-VolumesAssistedSnapshotsTest-1875841935-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 629.868252] env[65726]: DEBUG nova.virt.hardware [None req-b8e65ecf-6b31-4881-a7b5-2f0a252ee6d8 tempest-VolumesAssistedSnapshotsTest-1875841935 tempest-VolumesAssistedSnapshotsTest-1875841935-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 629.868562] env[65726]: DEBUG nova.virt.hardware [None req-b8e65ecf-6b31-4881-a7b5-2f0a252ee6d8 tempest-VolumesAssistedSnapshotsTest-1875841935 tempest-VolumesAssistedSnapshotsTest-1875841935-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 629.868732] env[65726]: DEBUG nova.virt.hardware [None req-b8e65ecf-6b31-4881-a7b5-2f0a252ee6d8 tempest-VolumesAssistedSnapshotsTest-1875841935 tempest-VolumesAssistedSnapshotsTest-1875841935-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 629.868900] env[65726]: DEBUG nova.virt.hardware [None req-b8e65ecf-6b31-4881-a7b5-2f0a252ee6d8 tempest-VolumesAssistedSnapshotsTest-1875841935 tempest-VolumesAssistedSnapshotsTest-1875841935-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 629.869071] env[65726]: DEBUG nova.virt.hardware [None req-b8e65ecf-6b31-4881-a7b5-2f0a252ee6d8 tempest-VolumesAssistedSnapshotsTest-1875841935 tempest-VolumesAssistedSnapshotsTest-1875841935-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 629.869246] env[65726]: DEBUG nova.virt.hardware [None req-b8e65ecf-6b31-4881-a7b5-2f0a252ee6d8 tempest-VolumesAssistedSnapshotsTest-1875841935 tempest-VolumesAssistedSnapshotsTest-1875841935-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 629.870217] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ae03537-e6f9-4036-b5e6-66a6094b1dd0 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.879516] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e33240cd-5ed0-474a-a6e9-3813e7d21235 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.895309] env[65726]: DEBUG nova.compute.manager [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] [instance: f5151062-57b1-4e4e-93f4-aab0e7f504d4] Starting instance... {{(pid=65726) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 629.914691] env[65726]: DEBUG nova.scheduler.client.report [None req-8fd7433e-ebac-43f4-aa60-6301be278066 tempest-ServerDiagnosticsTest-2021442419 tempest-ServerDiagnosticsTest-2021442419-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 629.930209] env[65726]: DEBUG oslo_vmware.api [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5251c83e-39bd-2d6c-e189-b024cdd1cd33, 'name': SearchDatastore_Task, 'duration_secs': 0.017464} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 629.930209] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7193641f-d456-4457-b66d-7970962a41db {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.936237] env[65726]: DEBUG oslo_vmware.api [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Waiting for the task: (returnval){ [ 629.936237] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52b47c4b-f1cc-b642-4a3e-b43660c6171c" [ 629.936237] env[65726]: _type = "Task" [ 629.936237] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 629.948904] env[65726]: DEBUG oslo_vmware.api [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52b47c4b-f1cc-b642-4a3e-b43660c6171c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 629.968183] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Lock "8493ba52-b950-4727-970a-19b1797b2ca1" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 24.586s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 630.033126] env[65726]: DEBUG oslo_vmware.api [None req-d2d75a44-1ded-480a-a8b4-89859e465e33 tempest-ServerDiagnosticsV248Test-347810772 tempest-ServerDiagnosticsV248Test-347810772-project-member] Task: {'id': task-5115439, 'name': Rename_Task, 'duration_secs': 0.36738} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 630.033444] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-d2d75a44-1ded-480a-a8b4-89859e465e33 tempest-ServerDiagnosticsV248Test-347810772 tempest-ServerDiagnosticsV248Test-347810772-project-member] [instance: 4d4dc88b-0d6c-430c-ab4f-c81a56e4657b] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 630.034219] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-cad13cad-a5d3-48fc-b7fb-d23d4584d384 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.042062] env[65726]: DEBUG oslo_vmware.api [None req-d2d75a44-1ded-480a-a8b4-89859e465e33 tempest-ServerDiagnosticsV248Test-347810772 tempest-ServerDiagnosticsV248Test-347810772-project-member] Waiting for the task: (returnval){ [ 630.042062] env[65726]: value = "task-5115440" [ 630.042062] env[65726]: _type = "Task" [ 630.042062] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 630.052052] env[65726]: DEBUG oslo_vmware.api [None req-d2d75a44-1ded-480a-a8b4-89859e465e33 tempest-ServerDiagnosticsV248Test-347810772 tempest-ServerDiagnosticsV248Test-347810772-project-member] Task: {'id': task-5115440, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 630.245117] env[65726]: DEBUG nova.network.neutron [None req-b82a7f4d-3603-48cd-8e48-1d790e39dbfb tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] [instance: cd539d8b-921b-4947-954c-8573380dbdc8] Successfully updated port: 60d92cf3-683a-4291-ba51-fad4124f3c8c {{(pid=65726) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 630.282085] env[65726]: DEBUG nova.network.neutron [req-fde7dfe2-2f08-46e4-af18-49e8c5f6cc8c req-32b5ea0b-1a90-4298-8ffd-9e3c4c1c1ac2 service nova] [instance: 8493ba52-b950-4727-970a-19b1797b2ca1] Updated VIF entry in instance network info cache for port 964ef0b7-9857-4032-ac36-efb1ecd1419d. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 630.282473] env[65726]: DEBUG nova.network.neutron [req-fde7dfe2-2f08-46e4-af18-49e8c5f6cc8c req-32b5ea0b-1a90-4298-8ffd-9e3c4c1c1ac2 service nova] [instance: 8493ba52-b950-4727-970a-19b1797b2ca1] Updating instance_info_cache with network_info: [{"id": "964ef0b7-9857-4032-ac36-efb1ecd1419d", "address": "fa:16:3e:3e:4b:2f", "network": {"id": "c68b8bcc-840e-459b-bee0-76a9ffdc08cb", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-2061916615-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7c3641c4a34b4bf8b09403220e153e65", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "aa0d7fd6-4fd3-4451-9208-8a3cf25e30e4", "external-id": "nsx-vlan-transportzone-545", "segmentation_id": 545, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap964ef0b7-98", "ovs_interfaceid": "964ef0b7-9857-4032-ac36-efb1ecd1419d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 630.426073] env[65726]: DEBUG oslo_concurrency.lockutils [None req-8fd7433e-ebac-43f4-aa60-6301be278066 tempest-ServerDiagnosticsTest-2021442419 tempest-ServerDiagnosticsTest-2021442419-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.630s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 630.427903] env[65726]: DEBUG nova.compute.manager [None req-8fd7433e-ebac-43f4-aa60-6301be278066 tempest-ServerDiagnosticsTest-2021442419 tempest-ServerDiagnosticsTest-2021442419-project-member] [instance: de5cad99-1dbc-4435-98b4-987e24ba8c3c] Start building networks asynchronously for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 630.432636] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 630.432636] env[65726]: DEBUG oslo_concurrency.lockutils [None req-8cc1dec8-17ac-4db3-bea9-0fb861f8b4b8 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.531s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 630.434040] env[65726]: INFO nova.compute.claims [None req-8cc1dec8-17ac-4db3-bea9-0fb861f8b4b8 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] [instance: 04ffdf80-c54c-4587-9bb4-d520dc440501] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 630.450422] env[65726]: DEBUG oslo_vmware.api [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52b47c4b-f1cc-b642-4a3e-b43660c6171c, 'name': SearchDatastore_Task, 'duration_secs': 0.027087} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 630.450684] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 630.450934] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore2] 21730277-cef6-4bfe-9f67-de71f6f615ff/21730277-cef6-4bfe-9f67-de71f6f615ff.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 630.451522] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6f0f1026-4798-4d30-936b-8e5bd64570a2 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.461032] env[65726]: DEBUG oslo_vmware.api [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Waiting for the task: (returnval){ [ 630.461032] env[65726]: value = "task-5115441" [ 630.461032] env[65726]: _type = "Task" [ 630.461032] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 630.471865] env[65726]: DEBUG oslo_vmware.api [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Task: {'id': task-5115441, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 630.471992] env[65726]: DEBUG nova.compute.manager [None req-881e6c53-2eb2-4fe8-9bf1-92c962897b3f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: f2d9090c-988f-43f4-9c81-7aa718a3438a] Starting instance... {{(pid=65726) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 630.511412] env[65726]: DEBUG nova.network.neutron [None req-b8e65ecf-6b31-4881-a7b5-2f0a252ee6d8 tempest-VolumesAssistedSnapshotsTest-1875841935 tempest-VolumesAssistedSnapshotsTest-1875841935-project-member] [instance: aa2716a9-2cf8-4e51-800d-1c8f98b2cf05] Successfully created port: c4f80985-65f1-478f-9944-afb3538efb48 {{(pid=65726) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 630.555026] env[65726]: DEBUG oslo_vmware.api [None req-d2d75a44-1ded-480a-a8b4-89859e465e33 tempest-ServerDiagnosticsV248Test-347810772 tempest-ServerDiagnosticsV248Test-347810772-project-member] Task: {'id': task-5115440, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 630.751513] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b82a7f4d-3603-48cd-8e48-1d790e39dbfb tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Acquiring lock "refresh_cache-cd539d8b-921b-4947-954c-8573380dbdc8" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 630.751709] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b82a7f4d-3603-48cd-8e48-1d790e39dbfb tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Acquired lock "refresh_cache-cd539d8b-921b-4947-954c-8573380dbdc8" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 630.751905] env[65726]: DEBUG nova.network.neutron [None req-b82a7f4d-3603-48cd-8e48-1d790e39dbfb tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] [instance: cd539d8b-921b-4947-954c-8573380dbdc8] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 630.785286] env[65726]: DEBUG oslo_concurrency.lockutils [req-fde7dfe2-2f08-46e4-af18-49e8c5f6cc8c req-32b5ea0b-1a90-4298-8ffd-9e3c4c1c1ac2 service nova] Releasing lock "refresh_cache-8493ba52-b950-4727-970a-19b1797b2ca1" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 630.785554] env[65726]: DEBUG nova.compute.manager [req-fde7dfe2-2f08-46e4-af18-49e8c5f6cc8c req-32b5ea0b-1a90-4298-8ffd-9e3c4c1c1ac2 service nova] [instance: 88fb7235-cd0e-49dc-9d78-3583d5d3f528] Received event network-changed-b915a591-ee09-4390-ae01-09778bd3f2c3 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 630.785711] env[65726]: DEBUG nova.compute.manager [req-fde7dfe2-2f08-46e4-af18-49e8c5f6cc8c req-32b5ea0b-1a90-4298-8ffd-9e3c4c1c1ac2 service nova] [instance: 88fb7235-cd0e-49dc-9d78-3583d5d3f528] Refreshing instance network info cache due to event network-changed-b915a591-ee09-4390-ae01-09778bd3f2c3. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 630.785907] env[65726]: DEBUG oslo_concurrency.lockutils [req-fde7dfe2-2f08-46e4-af18-49e8c5f6cc8c req-32b5ea0b-1a90-4298-8ffd-9e3c4c1c1ac2 service nova] Acquiring lock "refresh_cache-88fb7235-cd0e-49dc-9d78-3583d5d3f528" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 630.786046] env[65726]: DEBUG oslo_concurrency.lockutils [req-fde7dfe2-2f08-46e4-af18-49e8c5f6cc8c req-32b5ea0b-1a90-4298-8ffd-9e3c4c1c1ac2 service nova] Acquired lock "refresh_cache-88fb7235-cd0e-49dc-9d78-3583d5d3f528" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 630.786196] env[65726]: DEBUG nova.network.neutron [req-fde7dfe2-2f08-46e4-af18-49e8c5f6cc8c req-32b5ea0b-1a90-4298-8ffd-9e3c4c1c1ac2 service nova] [instance: 88fb7235-cd0e-49dc-9d78-3583d5d3f528] Refreshing network info cache for port b915a591-ee09-4390-ae01-09778bd3f2c3 {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 630.946542] env[65726]: DEBUG nova.compute.utils [None req-8fd7433e-ebac-43f4-aa60-6301be278066 tempest-ServerDiagnosticsTest-2021442419 tempest-ServerDiagnosticsTest-2021442419-project-member] Using /dev/sd instead of None {{(pid=65726) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 630.946542] env[65726]: DEBUG nova.compute.manager [None req-8fd7433e-ebac-43f4-aa60-6301be278066 tempest-ServerDiagnosticsTest-2021442419 tempest-ServerDiagnosticsTest-2021442419-project-member] [instance: de5cad99-1dbc-4435-98b4-987e24ba8c3c] Allocating IP information in the background. {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 630.946542] env[65726]: DEBUG nova.network.neutron [None req-8fd7433e-ebac-43f4-aa60-6301be278066 tempest-ServerDiagnosticsTest-2021442419 tempest-ServerDiagnosticsTest-2021442419-project-member] [instance: de5cad99-1dbc-4435-98b4-987e24ba8c3c] allocate_for_instance() {{(pid=65726) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 630.948663] env[65726]: WARNING neutronclient.v2_0.client [None req-8fd7433e-ebac-43f4-aa60-6301be278066 tempest-ServerDiagnosticsTest-2021442419 tempest-ServerDiagnosticsTest-2021442419-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 630.949168] env[65726]: WARNING neutronclient.v2_0.client [None req-8fd7433e-ebac-43f4-aa60-6301be278066 tempest-ServerDiagnosticsTest-2021442419 tempest-ServerDiagnosticsTest-2021442419-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 630.950376] env[65726]: WARNING openstack [None req-8fd7433e-ebac-43f4-aa60-6301be278066 tempest-ServerDiagnosticsTest-2021442419 tempest-ServerDiagnosticsTest-2021442419-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 630.950376] env[65726]: WARNING openstack [None req-8fd7433e-ebac-43f4-aa60-6301be278066 tempest-ServerDiagnosticsTest-2021442419 tempest-ServerDiagnosticsTest-2021442419-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 630.978653] env[65726]: DEBUG oslo_vmware.api [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Task: {'id': task-5115441, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 631.014257] env[65726]: DEBUG oslo_concurrency.lockutils [None req-881e6c53-2eb2-4fe8-9bf1-92c962897b3f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 631.059133] env[65726]: DEBUG oslo_vmware.api [None req-d2d75a44-1ded-480a-a8b4-89859e465e33 tempest-ServerDiagnosticsV248Test-347810772 tempest-ServerDiagnosticsV248Test-347810772-project-member] Task: {'id': task-5115440, 'name': PowerOnVM_Task, 'duration_secs': 0.982713} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 631.059953] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-d2d75a44-1ded-480a-a8b4-89859e465e33 tempest-ServerDiagnosticsV248Test-347810772 tempest-ServerDiagnosticsV248Test-347810772-project-member] [instance: 4d4dc88b-0d6c-430c-ab4f-c81a56e4657b] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 631.059953] env[65726]: INFO nova.compute.manager [None req-d2d75a44-1ded-480a-a8b4-89859e465e33 tempest-ServerDiagnosticsV248Test-347810772 tempest-ServerDiagnosticsV248Test-347810772-project-member] [instance: 4d4dc88b-0d6c-430c-ab4f-c81a56e4657b] Took 6.45 seconds to spawn the instance on the hypervisor. [ 631.059953] env[65726]: DEBUG nova.compute.manager [None req-d2d75a44-1ded-480a-a8b4-89859e465e33 tempest-ServerDiagnosticsV248Test-347810772 tempest-ServerDiagnosticsV248Test-347810772-project-member] [instance: 4d4dc88b-0d6c-430c-ab4f-c81a56e4657b] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 631.061061] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b006277-3943-4353-881b-ccf4e82c182c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.257540] env[65726]: WARNING openstack [None req-b82a7f4d-3603-48cd-8e48-1d790e39dbfb tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 631.257540] env[65726]: WARNING openstack [None req-b82a7f4d-3603-48cd-8e48-1d790e39dbfb tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 631.263356] env[65726]: DEBUG nova.network.neutron [None req-b82a7f4d-3603-48cd-8e48-1d790e39dbfb tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] [instance: cd539d8b-921b-4947-954c-8573380dbdc8] Instance cache missing network info. {{(pid=65726) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 631.291392] env[65726]: WARNING openstack [req-fde7dfe2-2f08-46e4-af18-49e8c5f6cc8c req-32b5ea0b-1a90-4298-8ffd-9e3c4c1c1ac2 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 631.291743] env[65726]: WARNING openstack [req-fde7dfe2-2f08-46e4-af18-49e8c5f6cc8c req-32b5ea0b-1a90-4298-8ffd-9e3c4c1c1ac2 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 631.462442] env[65726]: DEBUG nova.compute.manager [None req-8fd7433e-ebac-43f4-aa60-6301be278066 tempest-ServerDiagnosticsTest-2021442419 tempest-ServerDiagnosticsTest-2021442419-project-member] [instance: de5cad99-1dbc-4435-98b4-987e24ba8c3c] Start building block device mappings for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 631.480623] env[65726]: DEBUG oslo_vmware.api [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Task: {'id': task-5115441, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.767018} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 631.481107] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore2] 21730277-cef6-4bfe-9f67-de71f6f615ff/21730277-cef6-4bfe-9f67-de71f6f615ff.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 631.481107] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] [instance: 21730277-cef6-4bfe-9f67-de71f6f615ff] Extending root virtual disk to 1048576 {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 631.481395] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-56d93377-1ce1-4eb2-b12b-539fb5fe90f1 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.492388] env[65726]: DEBUG oslo_vmware.api [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Waiting for the task: (returnval){ [ 631.492388] env[65726]: value = "task-5115442" [ 631.492388] env[65726]: _type = "Task" [ 631.492388] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 631.504195] env[65726]: DEBUG oslo_vmware.api [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Task: {'id': task-5115442, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 631.585614] env[65726]: INFO nova.compute.manager [None req-d2d75a44-1ded-480a-a8b4-89859e465e33 tempest-ServerDiagnosticsV248Test-347810772 tempest-ServerDiagnosticsV248Test-347810772-project-member] [instance: 4d4dc88b-0d6c-430c-ab4f-c81a56e4657b] Took 24.87 seconds to build instance. [ 631.651687] env[65726]: DEBUG oslo_concurrency.lockutils [None req-36828e27-b100-49cc-98c2-ad0b25ca6572 tempest-ServerGroupTestJSON-125182767 tempest-ServerGroupTestJSON-125182767-project-member] Acquiring lock "c2929727-a7ea-4f63-997d-474c29a305fc" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 631.651915] env[65726]: DEBUG oslo_concurrency.lockutils [None req-36828e27-b100-49cc-98c2-ad0b25ca6572 tempest-ServerGroupTestJSON-125182767 tempest-ServerGroupTestJSON-125182767-project-member] Lock "c2929727-a7ea-4f63-997d-474c29a305fc" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 631.849683] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fdb9a151-8a8b-43e2-90ca-dd7aa45dfab0 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.858190] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca1664c3-36a7-4c89-b017-db8d61ad56eb {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.891295] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8342021-fa82-4d1a-99fd-9fd3413bced2 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.900143] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4832014-66c4-4289-b6ef-a8b6abfaa30d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.916788] env[65726]: DEBUG nova.compute.provider_tree [None req-8cc1dec8-17ac-4db3-bea9-0fb861f8b4b8 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 632.007933] env[65726]: DEBUG oslo_vmware.api [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Task: {'id': task-5115442, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.154744} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 632.008526] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] [instance: 21730277-cef6-4bfe-9f67-de71f6f615ff] Extended root virtual disk {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 632.009798] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a6e9ca7-dc7b-40b9-8c6f-7152d918d4bc {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.053283] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] [instance: 21730277-cef6-4bfe-9f67-de71f6f615ff] Reconfiguring VM instance instance-00000008 to attach disk [datastore2] 21730277-cef6-4bfe-9f67-de71f6f615ff/21730277-cef6-4bfe-9f67-de71f6f615ff.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 632.059504] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ab233a5d-1b7f-4bd2-a02b-06c28f6e1da6 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.083323] env[65726]: DEBUG oslo_vmware.api [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Waiting for the task: (returnval){ [ 632.083323] env[65726]: value = "task-5115443" [ 632.083323] env[65726]: _type = "Task" [ 632.083323] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 632.093643] env[65726]: DEBUG oslo_concurrency.lockutils [None req-d2d75a44-1ded-480a-a8b4-89859e465e33 tempest-ServerDiagnosticsV248Test-347810772 tempest-ServerDiagnosticsV248Test-347810772-project-member] Lock "4d4dc88b-0d6c-430c-ab4f-c81a56e4657b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 26.391s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 632.094444] env[65726]: DEBUG oslo_vmware.api [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Task: {'id': task-5115443, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 632.214775] env[65726]: DEBUG nova.policy [None req-8fd7433e-ebac-43f4-aa60-6301be278066 tempest-ServerDiagnosticsTest-2021442419 tempest-ServerDiagnosticsTest-2021442419-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1ee9945864004543b4d5d40111e3548d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0d16e812d7bc4d3285f2930454e8a434', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65726) authorize /opt/stack/nova/nova/policy.py:192}} [ 632.248125] env[65726]: WARNING openstack [None req-b82a7f4d-3603-48cd-8e48-1d790e39dbfb tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 632.248583] env[65726]: WARNING openstack [None req-b82a7f4d-3603-48cd-8e48-1d790e39dbfb tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 632.361970] env[65726]: DEBUG nova.network.neutron [None req-b8e65ecf-6b31-4881-a7b5-2f0a252ee6d8 tempest-VolumesAssistedSnapshotsTest-1875841935 tempest-VolumesAssistedSnapshotsTest-1875841935-project-member] [instance: aa2716a9-2cf8-4e51-800d-1c8f98b2cf05] Successfully updated port: c4f80985-65f1-478f-9944-afb3538efb48 {{(pid=65726) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 632.420179] env[65726]: DEBUG nova.scheduler.client.report [None req-8cc1dec8-17ac-4db3-bea9-0fb861f8b4b8 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 632.483315] env[65726]: DEBUG nova.compute.manager [None req-8fd7433e-ebac-43f4-aa60-6301be278066 tempest-ServerDiagnosticsTest-2021442419 tempest-ServerDiagnosticsTest-2021442419-project-member] [instance: de5cad99-1dbc-4435-98b4-987e24ba8c3c] Start spawning the instance on the hypervisor. {{(pid=65726) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 632.515805] env[65726]: DEBUG nova.virt.hardware [None req-8fd7433e-ebac-43f4-aa60-6301be278066 tempest-ServerDiagnosticsTest-2021442419 tempest-ServerDiagnosticsTest-2021442419-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T19:28:34Z,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1539f06fa3534e90acbc3a60c4b8bd3f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T19:28:35Z,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 632.515805] env[65726]: DEBUG nova.virt.hardware [None req-8fd7433e-ebac-43f4-aa60-6301be278066 tempest-ServerDiagnosticsTest-2021442419 tempest-ServerDiagnosticsTest-2021442419-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 632.515805] env[65726]: DEBUG nova.virt.hardware [None req-8fd7433e-ebac-43f4-aa60-6301be278066 tempest-ServerDiagnosticsTest-2021442419 tempest-ServerDiagnosticsTest-2021442419-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 632.516029] env[65726]: DEBUG nova.virt.hardware [None req-8fd7433e-ebac-43f4-aa60-6301be278066 tempest-ServerDiagnosticsTest-2021442419 tempest-ServerDiagnosticsTest-2021442419-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 632.516029] env[65726]: DEBUG nova.virt.hardware [None req-8fd7433e-ebac-43f4-aa60-6301be278066 tempest-ServerDiagnosticsTest-2021442419 tempest-ServerDiagnosticsTest-2021442419-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 632.516029] env[65726]: DEBUG nova.virt.hardware [None req-8fd7433e-ebac-43f4-aa60-6301be278066 tempest-ServerDiagnosticsTest-2021442419 tempest-ServerDiagnosticsTest-2021442419-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 632.516699] env[65726]: DEBUG nova.virt.hardware [None req-8fd7433e-ebac-43f4-aa60-6301be278066 tempest-ServerDiagnosticsTest-2021442419 tempest-ServerDiagnosticsTest-2021442419-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 632.516699] env[65726]: DEBUG nova.virt.hardware [None req-8fd7433e-ebac-43f4-aa60-6301be278066 tempest-ServerDiagnosticsTest-2021442419 tempest-ServerDiagnosticsTest-2021442419-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 632.516699] env[65726]: DEBUG nova.virt.hardware [None req-8fd7433e-ebac-43f4-aa60-6301be278066 tempest-ServerDiagnosticsTest-2021442419 tempest-ServerDiagnosticsTest-2021442419-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 632.516940] env[65726]: DEBUG nova.virt.hardware [None req-8fd7433e-ebac-43f4-aa60-6301be278066 tempest-ServerDiagnosticsTest-2021442419 tempest-ServerDiagnosticsTest-2021442419-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 632.516940] env[65726]: DEBUG nova.virt.hardware [None req-8fd7433e-ebac-43f4-aa60-6301be278066 tempest-ServerDiagnosticsTest-2021442419 tempest-ServerDiagnosticsTest-2021442419-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 632.517820] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d85a9717-46e3-44c4-b0a6-9917c9b5ed5d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.527246] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3489370c-74d7-44aa-ad28-da2443307f5e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.598613] env[65726]: DEBUG nova.compute.manager [None req-cc552ef7-8733-47e0-9cfd-32896dfc9132 tempest-TenantUsagesTestJSON-327211524 tempest-TenantUsagesTestJSON-327211524-project-member] [instance: 7748c23f-9ea9-4172-8ab7-187678272bb3] Starting instance... {{(pid=65726) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 632.601292] env[65726]: DEBUG oslo_vmware.api [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Task: {'id': task-5115443, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 632.664913] env[65726]: DEBUG nova.network.neutron [None req-8fd7433e-ebac-43f4-aa60-6301be278066 tempest-ServerDiagnosticsTest-2021442419 tempest-ServerDiagnosticsTest-2021442419-project-member] [instance: de5cad99-1dbc-4435-98b4-987e24ba8c3c] Successfully created port: 3f320668-6397-427f-b8c4-3e63282fcb87 {{(pid=65726) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 632.780333] env[65726]: WARNING openstack [req-fde7dfe2-2f08-46e4-af18-49e8c5f6cc8c req-32b5ea0b-1a90-4298-8ffd-9e3c4c1c1ac2 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 632.780333] env[65726]: WARNING openstack [req-fde7dfe2-2f08-46e4-af18-49e8c5f6cc8c req-32b5ea0b-1a90-4298-8ffd-9e3c4c1c1ac2 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 632.809876] env[65726]: WARNING openstack [None req-b82a7f4d-3603-48cd-8e48-1d790e39dbfb tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 632.810396] env[65726]: WARNING openstack [None req-b82a7f4d-3603-48cd-8e48-1d790e39dbfb tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 632.863681] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b8e65ecf-6b31-4881-a7b5-2f0a252ee6d8 tempest-VolumesAssistedSnapshotsTest-1875841935 tempest-VolumesAssistedSnapshotsTest-1875841935-project-member] Acquiring lock "refresh_cache-aa2716a9-2cf8-4e51-800d-1c8f98b2cf05" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 632.863929] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b8e65ecf-6b31-4881-a7b5-2f0a252ee6d8 tempest-VolumesAssistedSnapshotsTest-1875841935 tempest-VolumesAssistedSnapshotsTest-1875841935-project-member] Acquired lock "refresh_cache-aa2716a9-2cf8-4e51-800d-1c8f98b2cf05" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 632.864128] env[65726]: DEBUG nova.network.neutron [None req-b8e65ecf-6b31-4881-a7b5-2f0a252ee6d8 tempest-VolumesAssistedSnapshotsTest-1875841935 tempest-VolumesAssistedSnapshotsTest-1875841935-project-member] [instance: aa2716a9-2cf8-4e51-800d-1c8f98b2cf05] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 632.927431] env[65726]: DEBUG oslo_concurrency.lockutils [None req-8cc1dec8-17ac-4db3-bea9-0fb861f8b4b8 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.495s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 632.928246] env[65726]: DEBUG nova.compute.manager [None req-8cc1dec8-17ac-4db3-bea9-0fb861f8b4b8 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] [instance: 04ffdf80-c54c-4587-9bb4-d520dc440501] Start building networks asynchronously for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 632.932370] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.698s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 632.933828] env[65726]: INFO nova.compute.claims [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] [instance: 811c1bbd-259c-4f5e-9a61-581bfba0ec4d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 633.096266] env[65726]: DEBUG oslo_vmware.api [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Task: {'id': task-5115443, 'name': ReconfigVM_Task, 'duration_secs': 1.001409} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 633.097407] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] [instance: 21730277-cef6-4bfe-9f67-de71f6f615ff] Reconfigured VM instance instance-00000008 to attach disk [datastore2] 21730277-cef6-4bfe-9f67-de71f6f615ff/21730277-cef6-4bfe-9f67-de71f6f615ff.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 633.098378] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-eb2a8deb-90e2-417c-bfe3-9cb80bcbafdd {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.110117] env[65726]: DEBUG oslo_vmware.api [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Waiting for the task: (returnval){ [ 633.110117] env[65726]: value = "task-5115444" [ 633.110117] env[65726]: _type = "Task" [ 633.110117] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 633.124995] env[65726]: DEBUG oslo_vmware.api [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Task: {'id': task-5115444, 'name': Rename_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 633.172691] env[65726]: DEBUG oslo_concurrency.lockutils [None req-cc552ef7-8733-47e0-9cfd-32896dfc9132 tempest-TenantUsagesTestJSON-327211524 tempest-TenantUsagesTestJSON-327211524-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 633.372408] env[65726]: WARNING openstack [None req-b8e65ecf-6b31-4881-a7b5-2f0a252ee6d8 tempest-VolumesAssistedSnapshotsTest-1875841935 tempest-VolumesAssistedSnapshotsTest-1875841935-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 633.372408] env[65726]: WARNING openstack [None req-b8e65ecf-6b31-4881-a7b5-2f0a252ee6d8 tempest-VolumesAssistedSnapshotsTest-1875841935 tempest-VolumesAssistedSnapshotsTest-1875841935-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 633.376516] env[65726]: DEBUG nova.network.neutron [None req-b8e65ecf-6b31-4881-a7b5-2f0a252ee6d8 tempest-VolumesAssistedSnapshotsTest-1875841935 tempest-VolumesAssistedSnapshotsTest-1875841935-project-member] [instance: aa2716a9-2cf8-4e51-800d-1c8f98b2cf05] Instance cache missing network info. {{(pid=65726) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 633.444286] env[65726]: DEBUG nova.compute.utils [None req-8cc1dec8-17ac-4db3-bea9-0fb861f8b4b8 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Using /dev/sd instead of None {{(pid=65726) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 633.453578] env[65726]: DEBUG nova.compute.manager [None req-8cc1dec8-17ac-4db3-bea9-0fb861f8b4b8 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] [instance: 04ffdf80-c54c-4587-9bb4-d520dc440501] Allocating IP information in the background. {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 633.454023] env[65726]: DEBUG nova.network.neutron [None req-8cc1dec8-17ac-4db3-bea9-0fb861f8b4b8 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] [instance: 04ffdf80-c54c-4587-9bb4-d520dc440501] allocate_for_instance() {{(pid=65726) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 633.454780] env[65726]: WARNING neutronclient.v2_0.client [None req-8cc1dec8-17ac-4db3-bea9-0fb861f8b4b8 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 633.454880] env[65726]: WARNING neutronclient.v2_0.client [None req-8cc1dec8-17ac-4db3-bea9-0fb861f8b4b8 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 633.455549] env[65726]: WARNING openstack [None req-8cc1dec8-17ac-4db3-bea9-0fb861f8b4b8 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 633.455954] env[65726]: WARNING openstack [None req-8cc1dec8-17ac-4db3-bea9-0fb861f8b4b8 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 633.623156] env[65726]: DEBUG oslo_vmware.api [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Task: {'id': task-5115444, 'name': Rename_Task, 'duration_secs': 0.213649} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 633.623436] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] [instance: 21730277-cef6-4bfe-9f67-de71f6f615ff] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 633.623695] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-93ba7883-7788-45a8-9393-b242b7d6468f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.633434] env[65726]: DEBUG oslo_vmware.api [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Waiting for the task: (returnval){ [ 633.633434] env[65726]: value = "task-5115445" [ 633.633434] env[65726]: _type = "Task" [ 633.633434] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 633.644013] env[65726]: DEBUG oslo_vmware.api [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Task: {'id': task-5115445, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 633.879786] env[65726]: WARNING openstack [req-fde7dfe2-2f08-46e4-af18-49e8c5f6cc8c req-32b5ea0b-1a90-4298-8ffd-9e3c4c1c1ac2 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 633.881704] env[65726]: WARNING openstack [req-fde7dfe2-2f08-46e4-af18-49e8c5f6cc8c req-32b5ea0b-1a90-4298-8ffd-9e3c4c1c1ac2 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 633.893987] env[65726]: DEBUG nova.compute.manager [req-a1685450-20f0-432e-8d51-f059ae0e8b68 req-6ba2c67d-2799-4b3f-99ef-94b5740db0d1 service nova] [instance: 21730277-cef6-4bfe-9f67-de71f6f615ff] Received event network-vif-plugged-cc7e0970-7e47-45d6-b7fa-6e33947b210c {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 633.894285] env[65726]: DEBUG oslo_concurrency.lockutils [req-a1685450-20f0-432e-8d51-f059ae0e8b68 req-6ba2c67d-2799-4b3f-99ef-94b5740db0d1 service nova] Acquiring lock "21730277-cef6-4bfe-9f67-de71f6f615ff-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 633.894983] env[65726]: DEBUG oslo_concurrency.lockutils [req-a1685450-20f0-432e-8d51-f059ae0e8b68 req-6ba2c67d-2799-4b3f-99ef-94b5740db0d1 service nova] Lock "21730277-cef6-4bfe-9f67-de71f6f615ff-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 633.894983] env[65726]: DEBUG oslo_concurrency.lockutils [req-a1685450-20f0-432e-8d51-f059ae0e8b68 req-6ba2c67d-2799-4b3f-99ef-94b5740db0d1 service nova] Lock "21730277-cef6-4bfe-9f67-de71f6f615ff-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 633.895156] env[65726]: DEBUG nova.compute.manager [req-a1685450-20f0-432e-8d51-f059ae0e8b68 req-6ba2c67d-2799-4b3f-99ef-94b5740db0d1 service nova] [instance: 21730277-cef6-4bfe-9f67-de71f6f615ff] No waiting events found dispatching network-vif-plugged-cc7e0970-7e47-45d6-b7fa-6e33947b210c {{(pid=65726) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 633.895304] env[65726]: WARNING nova.compute.manager [req-a1685450-20f0-432e-8d51-f059ae0e8b68 req-6ba2c67d-2799-4b3f-99ef-94b5740db0d1 service nova] [instance: 21730277-cef6-4bfe-9f67-de71f6f615ff] Received unexpected event network-vif-plugged-cc7e0970-7e47-45d6-b7fa-6e33947b210c for instance with vm_state building and task_state spawning. [ 633.895421] env[65726]: DEBUG nova.compute.manager [req-a1685450-20f0-432e-8d51-f059ae0e8b68 req-6ba2c67d-2799-4b3f-99ef-94b5740db0d1 service nova] [instance: 21730277-cef6-4bfe-9f67-de71f6f615ff] Received event network-changed-cc7e0970-7e47-45d6-b7fa-6e33947b210c {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 633.895571] env[65726]: DEBUG nova.compute.manager [req-a1685450-20f0-432e-8d51-f059ae0e8b68 req-6ba2c67d-2799-4b3f-99ef-94b5740db0d1 service nova] [instance: 21730277-cef6-4bfe-9f67-de71f6f615ff] Refreshing instance network info cache due to event network-changed-cc7e0970-7e47-45d6-b7fa-6e33947b210c. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 633.895751] env[65726]: DEBUG oslo_concurrency.lockutils [req-a1685450-20f0-432e-8d51-f059ae0e8b68 req-6ba2c67d-2799-4b3f-99ef-94b5740db0d1 service nova] Acquiring lock "refresh_cache-21730277-cef6-4bfe-9f67-de71f6f615ff" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 633.895872] env[65726]: DEBUG oslo_concurrency.lockutils [req-a1685450-20f0-432e-8d51-f059ae0e8b68 req-6ba2c67d-2799-4b3f-99ef-94b5740db0d1 service nova] Acquired lock "refresh_cache-21730277-cef6-4bfe-9f67-de71f6f615ff" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 633.896031] env[65726]: DEBUG nova.network.neutron [req-a1685450-20f0-432e-8d51-f059ae0e8b68 req-6ba2c67d-2799-4b3f-99ef-94b5740db0d1 service nova] [instance: 21730277-cef6-4bfe-9f67-de71f6f615ff] Refreshing network info cache for port cc7e0970-7e47-45d6-b7fa-6e33947b210c {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 633.916248] env[65726]: DEBUG nova.network.neutron [None req-b82a7f4d-3603-48cd-8e48-1d790e39dbfb tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] [instance: cd539d8b-921b-4947-954c-8573380dbdc8] Updating instance_info_cache with network_info: [{"id": "60d92cf3-683a-4291-ba51-fad4124f3c8c", "address": "fa:16:3e:03:6c:08", "network": {"id": "f639fd12-fb47-4b3f-9895-6facebc89726", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-2112264538-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.3"}}], "meta": {"injected": false, "tenant_id": "3a680eae0e734ea8b2274b504d0a46d8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "257e5ea7-8b80-4301-9900-a754f1fe2031", "external-id": "nsx-vlan-transportzone-682", "segmentation_id": 682, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap60d92cf3-68", "ovs_interfaceid": "60d92cf3-683a-4291-ba51-fad4124f3c8c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 633.956405] env[65726]: DEBUG nova.compute.manager [None req-8cc1dec8-17ac-4db3-bea9-0fb861f8b4b8 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] [instance: 04ffdf80-c54c-4587-9bb4-d520dc440501] Start building block device mappings for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 634.145645] env[65726]: DEBUG oslo_vmware.api [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Task: {'id': task-5115445, 'name': PowerOnVM_Task} progress is 88%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 634.201194] env[65726]: DEBUG nova.policy [None req-8cc1dec8-17ac-4db3-bea9-0fb861f8b4b8 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a78a1cdb82644c1abe2b0ab2e0f621a5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3a680eae0e734ea8b2274b504d0a46d8', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65726) authorize /opt/stack/nova/nova/policy.py:192}} [ 634.400669] env[65726]: WARNING openstack [req-a1685450-20f0-432e-8d51-f059ae0e8b68 req-6ba2c67d-2799-4b3f-99ef-94b5740db0d1 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 634.401110] env[65726]: WARNING openstack [req-a1685450-20f0-432e-8d51-f059ae0e8b68 req-6ba2c67d-2799-4b3f-99ef-94b5740db0d1 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 634.410235] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d61384d-d196-494f-a7b2-3fea3faa4904 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.421593] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b82a7f4d-3603-48cd-8e48-1d790e39dbfb tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Releasing lock "refresh_cache-cd539d8b-921b-4947-954c-8573380dbdc8" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 634.422342] env[65726]: DEBUG nova.compute.manager [None req-b82a7f4d-3603-48cd-8e48-1d790e39dbfb tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] [instance: cd539d8b-921b-4947-954c-8573380dbdc8] Instance network_info: |[{"id": "60d92cf3-683a-4291-ba51-fad4124f3c8c", "address": "fa:16:3e:03:6c:08", "network": {"id": "f639fd12-fb47-4b3f-9895-6facebc89726", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-2112264538-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.3"}}], "meta": {"injected": false, "tenant_id": "3a680eae0e734ea8b2274b504d0a46d8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "257e5ea7-8b80-4301-9900-a754f1fe2031", "external-id": "nsx-vlan-transportzone-682", "segmentation_id": 682, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap60d92cf3-68", "ovs_interfaceid": "60d92cf3-683a-4291-ba51-fad4124f3c8c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 634.422698] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-b82a7f4d-3603-48cd-8e48-1d790e39dbfb tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] [instance: cd539d8b-921b-4947-954c-8573380dbdc8] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:03:6c:08', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '257e5ea7-8b80-4301-9900-a754f1fe2031', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '60d92cf3-683a-4291-ba51-fad4124f3c8c', 'vif_model': 'vmxnet3'}] {{(pid=65726) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 634.431052] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-b82a7f4d-3603-48cd-8e48-1d790e39dbfb tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 634.432286] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2052c51e-4772-4222-baf4-ad11739369c4 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.435160] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: cd539d8b-921b-4947-954c-8573380dbdc8] Creating VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 634.435446] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-01bf4c2d-9ce6-4f18-a794-16822f7c799c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.486329] env[65726]: DEBUG nova.compute.manager [None req-4240f7e0-4ab1-4718-b82f-a2d7b6caacae tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 8e9c27f6-e9fa-4ed0-b6ae-3b5df52d1c44] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 634.495179] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f5e283b-9cc4-4cb3-8e4f-0e66dd003e55 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.498861] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba3f235c-350a-4c2f-ab9b-e6525073607c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.501457] env[65726]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 634.501457] env[65726]: value = "task-5115446" [ 634.501457] env[65726]: _type = "Task" [ 634.501457] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 634.514887] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c75685c6-0b61-4260-9add-ff835a8a697d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.525430] env[65726]: DEBUG nova.network.neutron [None req-8fd7433e-ebac-43f4-aa60-6301be278066 tempest-ServerDiagnosticsTest-2021442419 tempest-ServerDiagnosticsTest-2021442419-project-member] [instance: de5cad99-1dbc-4435-98b4-987e24ba8c3c] Successfully updated port: 3f320668-6397-427f-b8c4-3e63282fcb87 {{(pid=65726) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 634.526858] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115446, 'name': CreateVM_Task} progress is 6%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 634.540589] env[65726]: DEBUG nova.compute.provider_tree [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 634.646453] env[65726]: DEBUG oslo_vmware.api [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Task: {'id': task-5115445, 'name': PowerOnVM_Task, 'duration_secs': 0.886351} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 634.646696] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] [instance: 21730277-cef6-4bfe-9f67-de71f6f615ff] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 634.647455] env[65726]: INFO nova.compute.manager [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] [instance: 21730277-cef6-4bfe-9f67-de71f6f615ff] Took 12.68 seconds to spawn the instance on the hypervisor. [ 634.647455] env[65726]: DEBUG nova.compute.manager [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] [instance: 21730277-cef6-4bfe-9f67-de71f6f615ff] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 634.648344] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-078f9683-bef0-40bc-bb6f-6106541c542f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.705763] env[65726]: WARNING openstack [None req-b8e65ecf-6b31-4881-a7b5-2f0a252ee6d8 tempest-VolumesAssistedSnapshotsTest-1875841935 tempest-VolumesAssistedSnapshotsTest-1875841935-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 634.706146] env[65726]: WARNING openstack [None req-b8e65ecf-6b31-4881-a7b5-2f0a252ee6d8 tempest-VolumesAssistedSnapshotsTest-1875841935 tempest-VolumesAssistedSnapshotsTest-1875841935-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 634.939991] env[65726]: DEBUG nova.network.neutron [None req-8cc1dec8-17ac-4db3-bea9-0fb861f8b4b8 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] [instance: 04ffdf80-c54c-4587-9bb4-d520dc440501] Successfully created port: ece7a121-1068-4cb7-b500-9b6a1e8c4a1c {{(pid=65726) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 635.004139] env[65726]: DEBUG nova.compute.manager [None req-8cc1dec8-17ac-4db3-bea9-0fb861f8b4b8 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] [instance: 04ffdf80-c54c-4587-9bb4-d520dc440501] Start spawning the instance on the hypervisor. {{(pid=65726) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 635.021605] env[65726]: INFO nova.compute.manager [None req-4240f7e0-4ab1-4718-b82f-a2d7b6caacae tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 8e9c27f6-e9fa-4ed0-b6ae-3b5df52d1c44] instance snapshotting [ 635.024037] env[65726]: WARNING nova.compute.manager [None req-4240f7e0-4ab1-4718-b82f-a2d7b6caacae tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 8e9c27f6-e9fa-4ed0-b6ae-3b5df52d1c44] trying to snapshot a non-running instance: (state: 4 expected: 1) [ 635.030994] env[65726]: DEBUG oslo_concurrency.lockutils [None req-8fd7433e-ebac-43f4-aa60-6301be278066 tempest-ServerDiagnosticsTest-2021442419 tempest-ServerDiagnosticsTest-2021442419-project-member] Acquiring lock "refresh_cache-de5cad99-1dbc-4435-98b4-987e24ba8c3c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 635.032022] env[65726]: DEBUG oslo_concurrency.lockutils [None req-8fd7433e-ebac-43f4-aa60-6301be278066 tempest-ServerDiagnosticsTest-2021442419 tempest-ServerDiagnosticsTest-2021442419-project-member] Acquired lock "refresh_cache-de5cad99-1dbc-4435-98b4-987e24ba8c3c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 635.032022] env[65726]: DEBUG nova.network.neutron [None req-8fd7433e-ebac-43f4-aa60-6301be278066 tempest-ServerDiagnosticsTest-2021442419 tempest-ServerDiagnosticsTest-2021442419-project-member] [instance: de5cad99-1dbc-4435-98b4-987e24ba8c3c] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 635.033606] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115446, 'name': CreateVM_Task} progress is 25%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 635.039421] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-daf5ddcf-f0a0-40eb-ac95-667d8065626d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.046212] env[65726]: DEBUG nova.scheduler.client.report [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 635.052686] env[65726]: DEBUG nova.virt.hardware [None req-8cc1dec8-17ac-4db3-bea9-0fb861f8b4b8 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T19:28:34Z,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1539f06fa3534e90acbc3a60c4b8bd3f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T19:28:35Z,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 635.052686] env[65726]: DEBUG nova.virt.hardware [None req-8cc1dec8-17ac-4db3-bea9-0fb861f8b4b8 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 635.052686] env[65726]: DEBUG nova.virt.hardware [None req-8cc1dec8-17ac-4db3-bea9-0fb861f8b4b8 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 635.052987] env[65726]: DEBUG nova.virt.hardware [None req-8cc1dec8-17ac-4db3-bea9-0fb861f8b4b8 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 635.053047] env[65726]: DEBUG nova.virt.hardware [None req-8cc1dec8-17ac-4db3-bea9-0fb861f8b4b8 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 635.053185] env[65726]: DEBUG nova.virt.hardware [None req-8cc1dec8-17ac-4db3-bea9-0fb861f8b4b8 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 635.053412] env[65726]: DEBUG nova.virt.hardware [None req-8cc1dec8-17ac-4db3-bea9-0fb861f8b4b8 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 635.053629] env[65726]: DEBUG nova.virt.hardware [None req-8cc1dec8-17ac-4db3-bea9-0fb861f8b4b8 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 635.053732] env[65726]: DEBUG nova.virt.hardware [None req-8cc1dec8-17ac-4db3-bea9-0fb861f8b4b8 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 635.053913] env[65726]: DEBUG nova.virt.hardware [None req-8cc1dec8-17ac-4db3-bea9-0fb861f8b4b8 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 635.054145] env[65726]: DEBUG nova.virt.hardware [None req-8cc1dec8-17ac-4db3-bea9-0fb861f8b4b8 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 635.069270] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83ed67b3-2753-42f2-a43f-bcb1b3aee224 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.073222] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.142s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 635.073806] env[65726]: DEBUG nova.compute.manager [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] [instance: 811c1bbd-259c-4f5e-9a61-581bfba0ec4d] Start building networks asynchronously for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 635.077362] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc508c98-8fdc-4ec1-a9a8-a82af5f35ddd {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.080406] env[65726]: DEBUG oslo_concurrency.lockutils [None req-32ab2821-307c-47ba-ac96-c38579de1e37 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 13.452s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 635.092790] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41b3eac9-924c-4c4f-8995-9c7b59ed1c82 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.166921] env[65726]: INFO nova.compute.manager [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] [instance: 21730277-cef6-4bfe-9f67-de71f6f615ff] Took 28.73 seconds to build instance. [ 635.204589] env[65726]: DEBUG nova.network.neutron [req-fde7dfe2-2f08-46e4-af18-49e8c5f6cc8c req-32b5ea0b-1a90-4298-8ffd-9e3c4c1c1ac2 service nova] [instance: 88fb7235-cd0e-49dc-9d78-3583d5d3f528] Updated VIF entry in instance network info cache for port b915a591-ee09-4390-ae01-09778bd3f2c3. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 635.204948] env[65726]: DEBUG nova.network.neutron [req-fde7dfe2-2f08-46e4-af18-49e8c5f6cc8c req-32b5ea0b-1a90-4298-8ffd-9e3c4c1c1ac2 service nova] [instance: 88fb7235-cd0e-49dc-9d78-3583d5d3f528] Updating instance_info_cache with network_info: [{"id": "b915a591-ee09-4390-ae01-09778bd3f2c3", "address": "fa:16:3e:e3:6b:b7", "network": {"id": "ae04280a-5cd5-46c3-a6f2-8020d38faea4", "bridge": "br-int", "label": "tempest-ServersTestJSON-1476935398-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.187", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "53f76b7cd812445da88b09acb430da70", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "49b5df12-d801-4140-8816-2fd401608c7d", "external-id": "nsx-vlan-transportzone-326", "segmentation_id": 326, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb915a591-ee", "ovs_interfaceid": "b915a591-ee09-4390-ae01-09778bd3f2c3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 635.522552] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115446, 'name': CreateVM_Task, 'duration_secs': 0.711659} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 635.522771] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: cd539d8b-921b-4947-954c-8573380dbdc8] Created VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 635.523378] env[65726]: WARNING openstack [None req-b82a7f4d-3603-48cd-8e48-1d790e39dbfb tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 635.523764] env[65726]: WARNING openstack [None req-b82a7f4d-3603-48cd-8e48-1d790e39dbfb tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 635.532362] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b82a7f4d-3603-48cd-8e48-1d790e39dbfb tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 635.532362] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b82a7f4d-3603-48cd-8e48-1d790e39dbfb tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 635.535355] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b82a7f4d-3603-48cd-8e48-1d790e39dbfb tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 635.542946] env[65726]: WARNING openstack [None req-8fd7433e-ebac-43f4-aa60-6301be278066 tempest-ServerDiagnosticsTest-2021442419 tempest-ServerDiagnosticsTest-2021442419-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 635.544475] env[65726]: WARNING openstack [None req-8fd7433e-ebac-43f4-aa60-6301be278066 tempest-ServerDiagnosticsTest-2021442419 tempest-ServerDiagnosticsTest-2021442419-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 635.550220] env[65726]: DEBUG nova.network.neutron [None req-8fd7433e-ebac-43f4-aa60-6301be278066 tempest-ServerDiagnosticsTest-2021442419 tempest-ServerDiagnosticsTest-2021442419-project-member] [instance: de5cad99-1dbc-4435-98b4-987e24ba8c3c] Instance cache missing network info. {{(pid=65726) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 635.551945] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8203e9e9-9133-40bb-8385-592eff2b04b8 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.559194] env[65726]: DEBUG oslo_vmware.api [None req-b82a7f4d-3603-48cd-8e48-1d790e39dbfb tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Waiting for the task: (returnval){ [ 635.559194] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52bb9320-eb05-7522-3d0d-ff1d0b83c381" [ 635.559194] env[65726]: _type = "Task" [ 635.559194] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 635.576579] env[65726]: DEBUG oslo_vmware.api [None req-b82a7f4d-3603-48cd-8e48-1d790e39dbfb tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52bb9320-eb05-7522-3d0d-ff1d0b83c381, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 635.584244] env[65726]: DEBUG nova.compute.utils [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Using /dev/sd instead of None {{(pid=65726) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 635.593467] env[65726]: INFO nova.compute.claims [None req-32ab2821-307c-47ba-ac96-c38579de1e37 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] [instance: 0e064341-4e4a-407b-8c26-3eb04b409029] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 635.598896] env[65726]: DEBUG nova.compute.manager [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] [instance: 811c1bbd-259c-4f5e-9a61-581bfba0ec4d] Allocating IP information in the background. {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 635.598896] env[65726]: DEBUG nova.network.neutron [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] [instance: 811c1bbd-259c-4f5e-9a61-581bfba0ec4d] allocate_for_instance() {{(pid=65726) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 635.599101] env[65726]: WARNING neutronclient.v2_0.client [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 635.599655] env[65726]: WARNING neutronclient.v2_0.client [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 635.600472] env[65726]: WARNING openstack [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 635.603092] env[65726]: WARNING openstack [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 635.612555] env[65726]: DEBUG nova.compute.manager [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] [instance: 811c1bbd-259c-4f5e-9a61-581bfba0ec4d] Start building block device mappings for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 635.615925] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-4240f7e0-4ab1-4718-b82f-a2d7b6caacae tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 8e9c27f6-e9fa-4ed0-b6ae-3b5df52d1c44] Creating Snapshot of the VM instance {{(pid=65726) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 635.620027] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-503998eb-862b-4369-a623-08abde4f589b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.626305] env[65726]: DEBUG oslo_vmware.api [None req-4240f7e0-4ab1-4718-b82f-a2d7b6caacae tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Waiting for the task: (returnval){ [ 635.626305] env[65726]: value = "task-5115447" [ 635.626305] env[65726]: _type = "Task" [ 635.626305] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 635.639054] env[65726]: DEBUG oslo_vmware.api [None req-4240f7e0-4ab1-4718-b82f-a2d7b6caacae tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': task-5115447, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 635.670407] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9854c11b-f610-49b4-83de-8393ff92aa0b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Lock "21730277-cef6-4bfe-9f67-de71f6f615ff" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 30.239s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 635.708245] env[65726]: DEBUG oslo_concurrency.lockutils [req-fde7dfe2-2f08-46e4-af18-49e8c5f6cc8c req-32b5ea0b-1a90-4298-8ffd-9e3c4c1c1ac2 service nova] Releasing lock "refresh_cache-88fb7235-cd0e-49dc-9d78-3583d5d3f528" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 635.708622] env[65726]: DEBUG nova.compute.manager [req-fde7dfe2-2f08-46e4-af18-49e8c5f6cc8c req-32b5ea0b-1a90-4298-8ffd-9e3c4c1c1ac2 service nova] [instance: bb64fc41-19d7-49dc-a17a-9dd74730130f] Received event network-vif-deleted-b6351d8e-b77c-4b29-b18b-02fc5e9e0c72 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 635.753639] env[65726]: WARNING openstack [None req-b8e65ecf-6b31-4881-a7b5-2f0a252ee6d8 tempest-VolumesAssistedSnapshotsTest-1875841935 tempest-VolumesAssistedSnapshotsTest-1875841935-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 635.754178] env[65726]: WARNING openstack [None req-b8e65ecf-6b31-4881-a7b5-2f0a252ee6d8 tempest-VolumesAssistedSnapshotsTest-1875841935 tempest-VolumesAssistedSnapshotsTest-1875841935-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 635.795969] env[65726]: WARNING openstack [req-a1685450-20f0-432e-8d51-f059ae0e8b68 req-6ba2c67d-2799-4b3f-99ef-94b5740db0d1 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 635.798940] env[65726]: WARNING openstack [req-a1685450-20f0-432e-8d51-f059ae0e8b68 req-6ba2c67d-2799-4b3f-99ef-94b5740db0d1 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 636.072230] env[65726]: DEBUG oslo_vmware.api [None req-b82a7f4d-3603-48cd-8e48-1d790e39dbfb tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52bb9320-eb05-7522-3d0d-ff1d0b83c381, 'name': SearchDatastore_Task, 'duration_secs': 0.025796} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 636.072542] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b82a7f4d-3603-48cd-8e48-1d790e39dbfb tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 636.072779] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-b82a7f4d-3603-48cd-8e48-1d790e39dbfb tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] [instance: cd539d8b-921b-4947-954c-8573380dbdc8] Processing image b52362a3-ee8a-4cbf-b06f-513b0cc8f95c {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 636.073083] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b82a7f4d-3603-48cd-8e48-1d790e39dbfb tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 636.073198] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b82a7f4d-3603-48cd-8e48-1d790e39dbfb tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 636.073375] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-b82a7f4d-3603-48cd-8e48-1d790e39dbfb tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 636.073699] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ba708b58-8630-4cd2-8ea5-54977100c5e3 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.086208] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-b82a7f4d-3603-48cd-8e48-1d790e39dbfb tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 636.086415] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-b82a7f4d-3603-48cd-8e48-1d790e39dbfb tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65726) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 636.087422] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f413b815-25b5-4818-9bed-fc6aa6ade6a7 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.094404] env[65726]: DEBUG oslo_vmware.api [None req-b82a7f4d-3603-48cd-8e48-1d790e39dbfb tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Waiting for the task: (returnval){ [ 636.094404] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52a06312-0576-bffc-3d6f-5958ac0ca24b" [ 636.094404] env[65726]: _type = "Task" [ 636.094404] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 636.106630] env[65726]: DEBUG oslo_vmware.api [None req-b82a7f4d-3603-48cd-8e48-1d790e39dbfb tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52a06312-0576-bffc-3d6f-5958ac0ca24b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 636.114697] env[65726]: INFO nova.compute.resource_tracker [None req-32ab2821-307c-47ba-ac96-c38579de1e37 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] [instance: 0e064341-4e4a-407b-8c26-3eb04b409029] Updating resource usage from migration 8f64b022-cb5e-47d0-89a3-b1455f9404a6 [ 636.137500] env[65726]: DEBUG oslo_vmware.api [None req-4240f7e0-4ab1-4718-b82f-a2d7b6caacae tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': task-5115447, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 636.174077] env[65726]: DEBUG nova.compute.manager [None req-40525054-994c-41ab-bed7-c877834cae18 tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] [instance: 3b933684-ac19-44b0-a49d-6af45501e166] Starting instance... {{(pid=65726) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 636.435425] env[65726]: DEBUG nova.policy [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '568156063bd14e60b728431e93d2665d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4d508c846ae64e2c84ee281fe59af610', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65726) authorize /opt/stack/nova/nova/policy.py:192}} [ 636.524737] env[65726]: WARNING openstack [None req-8fd7433e-ebac-43f4-aa60-6301be278066 tempest-ServerDiagnosticsTest-2021442419 tempest-ServerDiagnosticsTest-2021442419-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 636.525269] env[65726]: WARNING openstack [None req-8fd7433e-ebac-43f4-aa60-6301be278066 tempest-ServerDiagnosticsTest-2021442419 tempest-ServerDiagnosticsTest-2021442419-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 636.615403] env[65726]: DEBUG oslo_vmware.api [None req-b82a7f4d-3603-48cd-8e48-1d790e39dbfb tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52a06312-0576-bffc-3d6f-5958ac0ca24b, 'name': SearchDatastore_Task, 'duration_secs': 0.011271} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 636.617183] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7b30746f-eb0d-4f9c-8742-1c72f1427833 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.625984] env[65726]: DEBUG nova.compute.manager [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] [instance: 811c1bbd-259c-4f5e-9a61-581bfba0ec4d] Start spawning the instance on the hypervisor. {{(pid=65726) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 636.639035] env[65726]: DEBUG oslo_vmware.api [None req-b82a7f4d-3603-48cd-8e48-1d790e39dbfb tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Waiting for the task: (returnval){ [ 636.639035] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52bbb803-03d4-8a5e-ef6b-dc743a1b850f" [ 636.639035] env[65726]: _type = "Task" [ 636.639035] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 636.653096] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92b9d128-f7f5-43d8-b709-bc21cf41a3e8 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.657476] env[65726]: DEBUG oslo_vmware.api [None req-4240f7e0-4ab1-4718-b82f-a2d7b6caacae tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': task-5115447, 'name': CreateSnapshot_Task, 'duration_secs': 0.895376} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 636.666731] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-4240f7e0-4ab1-4718-b82f-a2d7b6caacae tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 8e9c27f6-e9fa-4ed0-b6ae-3b5df52d1c44] Created Snapshot of the VM instance {{(pid=65726) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 636.674826] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03e62a2e-cc88-47ed-8dc4-dd74aea72739 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.676666] env[65726]: DEBUG oslo_vmware.api [None req-b82a7f4d-3603-48cd-8e48-1d790e39dbfb tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52bbb803-03d4-8a5e-ef6b-dc743a1b850f, 'name': SearchDatastore_Task, 'duration_secs': 0.013646} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 636.679766] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b82a7f4d-3603-48cd-8e48-1d790e39dbfb tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 636.680401] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-b82a7f4d-3603-48cd-8e48-1d790e39dbfb tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] cd539d8b-921b-4947-954c-8573380dbdc8/cd539d8b-921b-4947-954c-8573380dbdc8.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 636.682673] env[65726]: DEBUG nova.virt.hardware [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T19:28:34Z,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1539f06fa3534e90acbc3a60c4b8bd3f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T19:28:35Z,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 636.682828] env[65726]: DEBUG nova.virt.hardware [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 636.682965] env[65726]: DEBUG nova.virt.hardware [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 636.683199] env[65726]: DEBUG nova.virt.hardware [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 636.683425] env[65726]: DEBUG nova.virt.hardware [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 636.683535] env[65726]: DEBUG nova.virt.hardware [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 636.683788] env[65726]: DEBUG nova.virt.hardware [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 636.683961] env[65726]: DEBUG nova.virt.hardware [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 636.684189] env[65726]: DEBUG nova.virt.hardware [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 636.684384] env[65726]: DEBUG nova.virt.hardware [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 636.684594] env[65726]: DEBUG nova.virt.hardware [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 636.691616] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5c6c8b32-abed-4cb6-aaad-2c2c050d4cb4 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.691616] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06afbce7-7493-45b2-8c31-cf0216fec821 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.695284] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b366777a-4acd-49b9-ba7d-2f9043ac8915 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.715068] env[65726]: DEBUG oslo_vmware.api [None req-b82a7f4d-3603-48cd-8e48-1d790e39dbfb tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Waiting for the task: (returnval){ [ 636.715068] env[65726]: value = "task-5115448" [ 636.715068] env[65726]: _type = "Task" [ 636.715068] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 636.744028] env[65726]: DEBUG oslo_concurrency.lockutils [None req-40525054-994c-41ab-bed7-c877834cae18 tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 636.747169] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dde32bb9-e922-497b-a218-78d3c258df2d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.750869] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfd6ac5a-54c8-4644-9fa0-c61c51c7063d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.760868] env[65726]: DEBUG oslo_vmware.api [None req-b82a7f4d-3603-48cd-8e48-1d790e39dbfb tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Task: {'id': task-5115448, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 636.773717] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9952fde8-19f3-4f99-a59d-fa05c7262735 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.790657] env[65726]: DEBUG nova.compute.provider_tree [None req-32ab2821-307c-47ba-ac96-c38579de1e37 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 637.022958] env[65726]: DEBUG nova.network.neutron [None req-8cc1dec8-17ac-4db3-bea9-0fb861f8b4b8 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] [instance: 04ffdf80-c54c-4587-9bb4-d520dc440501] Successfully updated port: ece7a121-1068-4cb7-b500-9b6a1e8c4a1c {{(pid=65726) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 637.050081] env[65726]: DEBUG nova.network.neutron [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] [instance: 811c1bbd-259c-4f5e-9a61-581bfba0ec4d] Successfully created port: 205b7b13-8d99-402c-9587-c84362a7af94 {{(pid=65726) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 637.087483] env[65726]: DEBUG nova.network.neutron [None req-b8e65ecf-6b31-4881-a7b5-2f0a252ee6d8 tempest-VolumesAssistedSnapshotsTest-1875841935 tempest-VolumesAssistedSnapshotsTest-1875841935-project-member] [instance: aa2716a9-2cf8-4e51-800d-1c8f98b2cf05] Updating instance_info_cache with network_info: [{"id": "c4f80985-65f1-478f-9944-afb3538efb48", "address": "fa:16:3e:35:5f:a5", "network": {"id": "da8b1180-a4d3-4444-b571-050c941634d4", "bridge": "br-int", "label": "tempest-VolumesAssistedSnapshotsTest-1753385467-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b01c77c483364e23b4a48b1e1c9c9d1c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c297fe21-cd0b-4226-813b-a65d2358d034", "external-id": "nsx-vlan-transportzone-98", "segmentation_id": 98, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc4f80985-65", "ovs_interfaceid": "c4f80985-65f1-478f-9944-afb3538efb48", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 637.141013] env[65726]: WARNING openstack [req-a1685450-20f0-432e-8d51-f059ae0e8b68 req-6ba2c67d-2799-4b3f-99ef-94b5740db0d1 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 637.141180] env[65726]: WARNING openstack [req-a1685450-20f0-432e-8d51-f059ae0e8b68 req-6ba2c67d-2799-4b3f-99ef-94b5740db0d1 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 637.222348] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-4240f7e0-4ab1-4718-b82f-a2d7b6caacae tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 8e9c27f6-e9fa-4ed0-b6ae-3b5df52d1c44] Creating linked-clone VM from snapshot {{(pid=65726) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 637.222992] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-86e4382a-139f-4754-94d0-e8498f129515 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.236287] env[65726]: DEBUG oslo_vmware.api [None req-4240f7e0-4ab1-4718-b82f-a2d7b6caacae tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Waiting for the task: (returnval){ [ 637.236287] env[65726]: value = "task-5115449" [ 637.236287] env[65726]: _type = "Task" [ 637.236287] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 637.255082] env[65726]: DEBUG oslo_vmware.api [None req-4240f7e0-4ab1-4718-b82f-a2d7b6caacae tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': task-5115449, 'name': CloneVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 637.261891] env[65726]: DEBUG oslo_vmware.api [None req-b82a7f4d-3603-48cd-8e48-1d790e39dbfb tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Task: {'id': task-5115448, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 637.296736] env[65726]: DEBUG nova.scheduler.client.report [None req-32ab2821-307c-47ba-ac96-c38579de1e37 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 637.525648] env[65726]: DEBUG oslo_concurrency.lockutils [None req-8cc1dec8-17ac-4db3-bea9-0fb861f8b4b8 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Acquiring lock "refresh_cache-04ffdf80-c54c-4587-9bb4-d520dc440501" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 637.525895] env[65726]: DEBUG oslo_concurrency.lockutils [None req-8cc1dec8-17ac-4db3-bea9-0fb861f8b4b8 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Acquired lock "refresh_cache-04ffdf80-c54c-4587-9bb4-d520dc440501" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 637.526147] env[65726]: DEBUG nova.network.neutron [None req-8cc1dec8-17ac-4db3-bea9-0fb861f8b4b8 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] [instance: 04ffdf80-c54c-4587-9bb4-d520dc440501] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 637.591185] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b8e65ecf-6b31-4881-a7b5-2f0a252ee6d8 tempest-VolumesAssistedSnapshotsTest-1875841935 tempest-VolumesAssistedSnapshotsTest-1875841935-project-member] Releasing lock "refresh_cache-aa2716a9-2cf8-4e51-800d-1c8f98b2cf05" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 637.591567] env[65726]: DEBUG nova.compute.manager [None req-b8e65ecf-6b31-4881-a7b5-2f0a252ee6d8 tempest-VolumesAssistedSnapshotsTest-1875841935 tempest-VolumesAssistedSnapshotsTest-1875841935-project-member] [instance: aa2716a9-2cf8-4e51-800d-1c8f98b2cf05] Instance network_info: |[{"id": "c4f80985-65f1-478f-9944-afb3538efb48", "address": "fa:16:3e:35:5f:a5", "network": {"id": "da8b1180-a4d3-4444-b571-050c941634d4", "bridge": "br-int", "label": "tempest-VolumesAssistedSnapshotsTest-1753385467-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b01c77c483364e23b4a48b1e1c9c9d1c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c297fe21-cd0b-4226-813b-a65d2358d034", "external-id": "nsx-vlan-transportzone-98", "segmentation_id": 98, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc4f80985-65", "ovs_interfaceid": "c4f80985-65f1-478f-9944-afb3538efb48", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 637.592066] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-b8e65ecf-6b31-4881-a7b5-2f0a252ee6d8 tempest-VolumesAssistedSnapshotsTest-1875841935 tempest-VolumesAssistedSnapshotsTest-1875841935-project-member] [instance: aa2716a9-2cf8-4e51-800d-1c8f98b2cf05] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:35:5f:a5', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c297fe21-cd0b-4226-813b-a65d2358d034', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c4f80985-65f1-478f-9944-afb3538efb48', 'vif_model': 'vmxnet3'}] {{(pid=65726) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 637.599978] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-b8e65ecf-6b31-4881-a7b5-2f0a252ee6d8 tempest-VolumesAssistedSnapshotsTest-1875841935 tempest-VolumesAssistedSnapshotsTest-1875841935-project-member] Creating folder: Project (b01c77c483364e23b4a48b1e1c9c9d1c). Parent ref: group-v995008. {{(pid=65726) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 637.600283] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7289e635-def6-4db5-8e22-913fdf058b3a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.612814] env[65726]: INFO nova.virt.vmwareapi.vm_util [None req-b8e65ecf-6b31-4881-a7b5-2f0a252ee6d8 tempest-VolumesAssistedSnapshotsTest-1875841935 tempest-VolumesAssistedSnapshotsTest-1875841935-project-member] Created folder: Project (b01c77c483364e23b4a48b1e1c9c9d1c) in parent group-v995008. [ 637.612994] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-b8e65ecf-6b31-4881-a7b5-2f0a252ee6d8 tempest-VolumesAssistedSnapshotsTest-1875841935 tempest-VolumesAssistedSnapshotsTest-1875841935-project-member] Creating folder: Instances. Parent ref: group-v995041. {{(pid=65726) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 637.613254] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d9c4ecc0-7a5e-4d76-938b-fd21955df99b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.624945] env[65726]: INFO nova.virt.vmwareapi.vm_util [None req-b8e65ecf-6b31-4881-a7b5-2f0a252ee6d8 tempest-VolumesAssistedSnapshotsTest-1875841935 tempest-VolumesAssistedSnapshotsTest-1875841935-project-member] Created folder: Instances in parent group-v995041. [ 637.625233] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-b8e65ecf-6b31-4881-a7b5-2f0a252ee6d8 tempest-VolumesAssistedSnapshotsTest-1875841935 tempest-VolumesAssistedSnapshotsTest-1875841935-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 637.625440] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: aa2716a9-2cf8-4e51-800d-1c8f98b2cf05] Creating VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 637.625892] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ce93f6d5-e19e-422b-9749-f0ac1107d466 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.647782] env[65726]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 637.647782] env[65726]: value = "task-5115452" [ 637.647782] env[65726]: _type = "Task" [ 637.647782] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 637.656750] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115452, 'name': CreateVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 637.751531] env[65726]: DEBUG oslo_vmware.api [None req-4240f7e0-4ab1-4718-b82f-a2d7b6caacae tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': task-5115449, 'name': CloneVM_Task} progress is 94%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 637.755473] env[65726]: DEBUG oslo_vmware.api [None req-b82a7f4d-3603-48cd-8e48-1d790e39dbfb tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Task: {'id': task-5115448, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.660106} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 637.755793] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-b82a7f4d-3603-48cd-8e48-1d790e39dbfb tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] cd539d8b-921b-4947-954c-8573380dbdc8/cd539d8b-921b-4947-954c-8573380dbdc8.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 637.756013] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-b82a7f4d-3603-48cd-8e48-1d790e39dbfb tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] [instance: cd539d8b-921b-4947-954c-8573380dbdc8] Extending root virtual disk to 1048576 {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 637.756293] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-6e03197c-d875-4328-962a-20a42b4f71b6 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.765039] env[65726]: DEBUG oslo_vmware.api [None req-b82a7f4d-3603-48cd-8e48-1d790e39dbfb tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Waiting for the task: (returnval){ [ 637.765039] env[65726]: value = "task-5115453" [ 637.765039] env[65726]: _type = "Task" [ 637.765039] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 637.775293] env[65726]: DEBUG oslo_vmware.api [None req-b82a7f4d-3603-48cd-8e48-1d790e39dbfb tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Task: {'id': task-5115453, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 637.803804] env[65726]: DEBUG oslo_concurrency.lockutils [None req-32ab2821-307c-47ba-ac96-c38579de1e37 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.723s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 637.804016] env[65726]: INFO nova.compute.manager [None req-32ab2821-307c-47ba-ac96-c38579de1e37 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] [instance: 0e064341-4e4a-407b-8c26-3eb04b409029] Migrating [ 637.804247] env[65726]: DEBUG oslo_concurrency.lockutils [None req-32ab2821-307c-47ba-ac96-c38579de1e37 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Acquiring lock "compute-rpcapi-router" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 637.804385] env[65726]: DEBUG oslo_concurrency.lockutils [None req-32ab2821-307c-47ba-ac96-c38579de1e37 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Acquired lock "compute-rpcapi-router" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 637.805785] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a0754b39-9b94-45fe-86fb-b66ad1bd90cb tempest-DeleteServersAdminTestJSON-1071157516 tempest-DeleteServersAdminTestJSON-1071157516-project-admin] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 14.036s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 637.806011] env[65726]: DEBUG nova.objects.instance [None req-a0754b39-9b94-45fe-86fb-b66ad1bd90cb tempest-DeleteServersAdminTestJSON-1071157516 tempest-DeleteServersAdminTestJSON-1071157516-project-admin] Lazy-loading 'resources' on Instance uuid bb64fc41-19d7-49dc-a17a-9dd74730130f {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 637.922050] env[65726]: DEBUG oslo_concurrency.lockutils [None req-fb0f14b3-fd50-4c8d-a984-877b235dcb32 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Acquiring lock "d7905c4f-1ba3-43b3-b4b3-aebacd2f9cc3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 637.922329] env[65726]: DEBUG oslo_concurrency.lockutils [None req-fb0f14b3-fd50-4c8d-a984-877b235dcb32 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Lock "d7905c4f-1ba3-43b3-b4b3-aebacd2f9cc3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 638.023977] env[65726]: WARNING openstack [None req-8fd7433e-ebac-43f4-aa60-6301be278066 tempest-ServerDiagnosticsTest-2021442419 tempest-ServerDiagnosticsTest-2021442419-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 638.024388] env[65726]: WARNING openstack [None req-8fd7433e-ebac-43f4-aa60-6301be278066 tempest-ServerDiagnosticsTest-2021442419 tempest-ServerDiagnosticsTest-2021442419-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 638.033418] env[65726]: WARNING openstack [None req-8cc1dec8-17ac-4db3-bea9-0fb861f8b4b8 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 638.033857] env[65726]: WARNING openstack [None req-8cc1dec8-17ac-4db3-bea9-0fb861f8b4b8 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 638.038748] env[65726]: DEBUG nova.network.neutron [None req-8cc1dec8-17ac-4db3-bea9-0fb861f8b4b8 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] [instance: 04ffdf80-c54c-4587-9bb4-d520dc440501] Instance cache missing network info. {{(pid=65726) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 638.159449] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115452, 'name': CreateVM_Task} progress is 99%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 638.163858] env[65726]: DEBUG oslo_concurrency.lockutils [None req-efa812b8-b008-4879-b549-6ccf454f374f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] Acquiring lock "ddd36db0-28ad-4582-994b-7eef8b08ed46" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 638.164136] env[65726]: DEBUG oslo_concurrency.lockutils [None req-efa812b8-b008-4879-b549-6ccf454f374f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] Lock "ddd36db0-28ad-4582-994b-7eef8b08ed46" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 638.248158] env[65726]: DEBUG oslo_vmware.api [None req-4240f7e0-4ab1-4718-b82f-a2d7b6caacae tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': task-5115449, 'name': CloneVM_Task} progress is 94%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 638.276596] env[65726]: DEBUG oslo_vmware.api [None req-b82a7f4d-3603-48cd-8e48-1d790e39dbfb tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Task: {'id': task-5115453, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.226063} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 638.276676] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-b82a7f4d-3603-48cd-8e48-1d790e39dbfb tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] [instance: cd539d8b-921b-4947-954c-8573380dbdc8] Extended root virtual disk {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 638.277425] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8d8b4dc-9030-4544-bc3a-053cfe74ff4a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.300431] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-b82a7f4d-3603-48cd-8e48-1d790e39dbfb tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] [instance: cd539d8b-921b-4947-954c-8573380dbdc8] Reconfiguring VM instance instance-0000000a to attach disk [datastore1] cd539d8b-921b-4947-954c-8573380dbdc8/cd539d8b-921b-4947-954c-8573380dbdc8.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 638.301160] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9ca69c90-01a2-46ea-b53f-22f1a236025b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.316655] env[65726]: INFO nova.compute.rpcapi [None req-32ab2821-307c-47ba-ac96-c38579de1e37 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Automatically selected compute RPC version 6.4 from minimum service version 70 [ 638.317193] env[65726]: DEBUG oslo_concurrency.lockutils [None req-32ab2821-307c-47ba-ac96-c38579de1e37 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Releasing lock "compute-rpcapi-router" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 638.337449] env[65726]: DEBUG oslo_vmware.api [None req-b82a7f4d-3603-48cd-8e48-1d790e39dbfb tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Waiting for the task: (returnval){ [ 638.337449] env[65726]: value = "task-5115454" [ 638.337449] env[65726]: _type = "Task" [ 638.337449] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 638.348663] env[65726]: DEBUG oslo_vmware.api [None req-b82a7f4d-3603-48cd-8e48-1d790e39dbfb tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Task: {'id': task-5115454, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 638.663870] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115452, 'name': CreateVM_Task, 'duration_secs': 0.77046} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 638.664126] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: aa2716a9-2cf8-4e51-800d-1c8f98b2cf05] Created VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 638.665311] env[65726]: WARNING openstack [None req-b8e65ecf-6b31-4881-a7b5-2f0a252ee6d8 tempest-VolumesAssistedSnapshotsTest-1875841935 tempest-VolumesAssistedSnapshotsTest-1875841935-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 638.665924] env[65726]: WARNING openstack [None req-b8e65ecf-6b31-4881-a7b5-2f0a252ee6d8 tempest-VolumesAssistedSnapshotsTest-1875841935 tempest-VolumesAssistedSnapshotsTest-1875841935-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 638.673231] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b8e65ecf-6b31-4881-a7b5-2f0a252ee6d8 tempest-VolumesAssistedSnapshotsTest-1875841935 tempest-VolumesAssistedSnapshotsTest-1875841935-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 638.673454] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b8e65ecf-6b31-4881-a7b5-2f0a252ee6d8 tempest-VolumesAssistedSnapshotsTest-1875841935 tempest-VolumesAssistedSnapshotsTest-1875841935-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 638.674138] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b8e65ecf-6b31-4881-a7b5-2f0a252ee6d8 tempest-VolumesAssistedSnapshotsTest-1875841935 tempest-VolumesAssistedSnapshotsTest-1875841935-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 638.677409] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0177a354-f9bc-49fe-a7d7-03982dd1b571 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.684355] env[65726]: DEBUG oslo_vmware.api [None req-b8e65ecf-6b31-4881-a7b5-2f0a252ee6d8 tempest-VolumesAssistedSnapshotsTest-1875841935 tempest-VolumesAssistedSnapshotsTest-1875841935-project-member] Waiting for the task: (returnval){ [ 638.684355] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52b49ab6-66a4-8ac5-2883-07fc1a1b6f42" [ 638.684355] env[65726]: _type = "Task" [ 638.684355] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 638.696966] env[65726]: DEBUG oslo_vmware.api [None req-b8e65ecf-6b31-4881-a7b5-2f0a252ee6d8 tempest-VolumesAssistedSnapshotsTest-1875841935 tempest-VolumesAssistedSnapshotsTest-1875841935-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52b49ab6-66a4-8ac5-2883-07fc1a1b6f42, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 638.735212] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64777720-f3ed-43be-b4b1-52155deb9ada {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.750943] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94e9947b-571d-4828-8d65-de37c74451a0 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.754409] env[65726]: DEBUG oslo_vmware.api [None req-4240f7e0-4ab1-4718-b82f-a2d7b6caacae tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': task-5115449, 'name': CloneVM_Task} progress is 94%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 638.800267] env[65726]: DEBUG nova.network.neutron [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] [instance: 811c1bbd-259c-4f5e-9a61-581bfba0ec4d] Successfully updated port: 205b7b13-8d99-402c-9587-c84362a7af94 {{(pid=65726) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 638.802035] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-924fc3ee-3fc0-48bd-bd16-4bbbf53fbbca {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.817072] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eca7f773-1d21-4569-987d-8a1896bd682d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.832998] env[65726]: DEBUG nova.compute.provider_tree [None req-a0754b39-9b94-45fe-86fb-b66ad1bd90cb tempest-DeleteServersAdminTestJSON-1071157516 tempest-DeleteServersAdminTestJSON-1071157516-project-admin] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 638.837929] env[65726]: DEBUG oslo_concurrency.lockutils [None req-32ab2821-307c-47ba-ac96-c38579de1e37 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Acquiring lock "refresh_cache-0e064341-4e4a-407b-8c26-3eb04b409029" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 638.838191] env[65726]: DEBUG oslo_concurrency.lockutils [None req-32ab2821-307c-47ba-ac96-c38579de1e37 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Acquired lock "refresh_cache-0e064341-4e4a-407b-8c26-3eb04b409029" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 638.838396] env[65726]: DEBUG nova.network.neutron [None req-32ab2821-307c-47ba-ac96-c38579de1e37 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] [instance: 0e064341-4e4a-407b-8c26-3eb04b409029] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 638.856673] env[65726]: DEBUG oslo_vmware.api [None req-b82a7f4d-3603-48cd-8e48-1d790e39dbfb tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Task: {'id': task-5115454, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 638.911744] env[65726]: DEBUG nova.network.neutron [req-a1685450-20f0-432e-8d51-f059ae0e8b68 req-6ba2c67d-2799-4b3f-99ef-94b5740db0d1 service nova] [instance: 21730277-cef6-4bfe-9f67-de71f6f615ff] Updated VIF entry in instance network info cache for port cc7e0970-7e47-45d6-b7fa-6e33947b210c. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 638.912347] env[65726]: DEBUG nova.network.neutron [req-a1685450-20f0-432e-8d51-f059ae0e8b68 req-6ba2c67d-2799-4b3f-99ef-94b5740db0d1 service nova] [instance: 21730277-cef6-4bfe-9f67-de71f6f615ff] Updating instance_info_cache with network_info: [{"id": "cc7e0970-7e47-45d6-b7fa-6e33947b210c", "address": "fa:16:3e:60:60:cb", "network": {"id": "c68b8bcc-840e-459b-bee0-76a9ffdc08cb", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-2061916615-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7c3641c4a34b4bf8b09403220e153e65", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "aa0d7fd6-4fd3-4451-9208-8a3cf25e30e4", "external-id": "nsx-vlan-transportzone-545", "segmentation_id": 545, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcc7e0970-7e", "ovs_interfaceid": "cc7e0970-7e47-45d6-b7fa-6e33947b210c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 639.075240] env[65726]: WARNING openstack [None req-8cc1dec8-17ac-4db3-bea9-0fb861f8b4b8 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 639.075569] env[65726]: WARNING openstack [None req-8cc1dec8-17ac-4db3-bea9-0fb861f8b4b8 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 639.125247] env[65726]: DEBUG nova.network.neutron [None req-8fd7433e-ebac-43f4-aa60-6301be278066 tempest-ServerDiagnosticsTest-2021442419 tempest-ServerDiagnosticsTest-2021442419-project-member] [instance: de5cad99-1dbc-4435-98b4-987e24ba8c3c] Updating instance_info_cache with network_info: [{"id": "3f320668-6397-427f-b8c4-3e63282fcb87", "address": "fa:16:3e:ca:a0:6d", "network": {"id": "55d629b8-a541-4720-9289-927bbc6e6f22", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.83", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "1539f06fa3534e90acbc3a60c4b8bd3f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "35342bcb-8b06-472e-b3c0-43fd3d6c4b30", "external-id": "nsx-vlan-transportzone-524", "segmentation_id": 524, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3f320668-63", "ovs_interfaceid": "3f320668-6397-427f-b8c4-3e63282fcb87", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 639.197061] env[65726]: DEBUG oslo_vmware.api [None req-b8e65ecf-6b31-4881-a7b5-2f0a252ee6d8 tempest-VolumesAssistedSnapshotsTest-1875841935 tempest-VolumesAssistedSnapshotsTest-1875841935-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52b49ab6-66a4-8ac5-2883-07fc1a1b6f42, 'name': SearchDatastore_Task, 'duration_secs': 0.018448} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 639.197304] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b8e65ecf-6b31-4881-a7b5-2f0a252ee6d8 tempest-VolumesAssistedSnapshotsTest-1875841935 tempest-VolumesAssistedSnapshotsTest-1875841935-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 639.198174] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-b8e65ecf-6b31-4881-a7b5-2f0a252ee6d8 tempest-VolumesAssistedSnapshotsTest-1875841935 tempest-VolumesAssistedSnapshotsTest-1875841935-project-member] [instance: aa2716a9-2cf8-4e51-800d-1c8f98b2cf05] Processing image b52362a3-ee8a-4cbf-b06f-513b0cc8f95c {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 639.198174] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b8e65ecf-6b31-4881-a7b5-2f0a252ee6d8 tempest-VolumesAssistedSnapshotsTest-1875841935 tempest-VolumesAssistedSnapshotsTest-1875841935-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 639.198174] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b8e65ecf-6b31-4881-a7b5-2f0a252ee6d8 tempest-VolumesAssistedSnapshotsTest-1875841935 tempest-VolumesAssistedSnapshotsTest-1875841935-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 639.198174] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-b8e65ecf-6b31-4881-a7b5-2f0a252ee6d8 tempest-VolumesAssistedSnapshotsTest-1875841935 tempest-VolumesAssistedSnapshotsTest-1875841935-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 639.198305] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c6929feb-38df-40e2-8156-6fdf5f473d21 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.209425] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-b8e65ecf-6b31-4881-a7b5-2f0a252ee6d8 tempest-VolumesAssistedSnapshotsTest-1875841935 tempest-VolumesAssistedSnapshotsTest-1875841935-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 639.209615] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-b8e65ecf-6b31-4881-a7b5-2f0a252ee6d8 tempest-VolumesAssistedSnapshotsTest-1875841935 tempest-VolumesAssistedSnapshotsTest-1875841935-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65726) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 639.210376] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-08e2ccaf-1506-4034-86e7-d9ceef58e2a0 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.216636] env[65726]: DEBUG oslo_vmware.api [None req-b8e65ecf-6b31-4881-a7b5-2f0a252ee6d8 tempest-VolumesAssistedSnapshotsTest-1875841935 tempest-VolumesAssistedSnapshotsTest-1875841935-project-member] Waiting for the task: (returnval){ [ 639.216636] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52c71cb7-1369-adc2-e167-39b606899334" [ 639.216636] env[65726]: _type = "Task" [ 639.216636] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 639.224941] env[65726]: DEBUG oslo_vmware.api [None req-b8e65ecf-6b31-4881-a7b5-2f0a252ee6d8 tempest-VolumesAssistedSnapshotsTest-1875841935 tempest-VolumesAssistedSnapshotsTest-1875841935-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52c71cb7-1369-adc2-e167-39b606899334, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 639.250663] env[65726]: DEBUG oslo_vmware.api [None req-4240f7e0-4ab1-4718-b82f-a2d7b6caacae tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': task-5115449, 'name': CloneVM_Task, 'duration_secs': 1.971219} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 639.251142] env[65726]: INFO nova.virt.vmwareapi.vmops [None req-4240f7e0-4ab1-4718-b82f-a2d7b6caacae tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 8e9c27f6-e9fa-4ed0-b6ae-3b5df52d1c44] Created linked-clone VM from snapshot [ 639.252058] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6975e71-05e2-42b6-8529-7aba3fe63008 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.261310] env[65726]: DEBUG nova.virt.vmwareapi.images [None req-4240f7e0-4ab1-4718-b82f-a2d7b6caacae tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 8e9c27f6-e9fa-4ed0-b6ae-3b5df52d1c44] Uploading image 699731f0-c4b1-40f5-9f6e-780cb60b3ce3 {{(pid=65726) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 639.298500] env[65726]: DEBUG oslo_vmware.rw_handles [None req-4240f7e0-4ab1-4718-b82f-a2d7b6caacae tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 639.298500] env[65726]: value = "vm-995040" [ 639.298500] env[65726]: _type = "VirtualMachine" [ 639.298500] env[65726]: }. {{(pid=65726) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 639.298825] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-df923337-235d-4fd2-92d9-d8bc5fe9c329 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.311471] env[65726]: DEBUG oslo_vmware.rw_handles [None req-4240f7e0-4ab1-4718-b82f-a2d7b6caacae tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Lease: (returnval){ [ 639.311471] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52332b92-be9b-d680-de8d-45137ac26750" [ 639.311471] env[65726]: _type = "HttpNfcLease" [ 639.311471] env[65726]: } obtained for exporting VM: (result){ [ 639.311471] env[65726]: value = "vm-995040" [ 639.311471] env[65726]: _type = "VirtualMachine" [ 639.311471] env[65726]: }. {{(pid=65726) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 639.311471] env[65726]: DEBUG oslo_vmware.api [None req-4240f7e0-4ab1-4718-b82f-a2d7b6caacae tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Waiting for the lease: (returnval){ [ 639.311471] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52332b92-be9b-d680-de8d-45137ac26750" [ 639.311471] env[65726]: _type = "HttpNfcLease" [ 639.311471] env[65726]: } to be ready. {{(pid=65726) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 639.313154] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Acquiring lock "refresh_cache-811c1bbd-259c-4f5e-9a61-581bfba0ec4d" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 639.313419] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Acquired lock "refresh_cache-811c1bbd-259c-4f5e-9a61-581bfba0ec4d" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 639.314607] env[65726]: DEBUG nova.network.neutron [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] [instance: 811c1bbd-259c-4f5e-9a61-581bfba0ec4d] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 639.320915] env[65726]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 639.320915] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52332b92-be9b-d680-de8d-45137ac26750" [ 639.320915] env[65726]: _type = "HttpNfcLease" [ 639.320915] env[65726]: } is initializing. {{(pid=65726) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 639.336253] env[65726]: DEBUG nova.scheduler.client.report [None req-a0754b39-9b94-45fe-86fb-b66ad1bd90cb tempest-DeleteServersAdminTestJSON-1071157516 tempest-DeleteServersAdminTestJSON-1071157516-project-admin] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 639.344627] env[65726]: WARNING openstack [None req-32ab2821-307c-47ba-ac96-c38579de1e37 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 639.344627] env[65726]: WARNING openstack [None req-32ab2821-307c-47ba-ac96-c38579de1e37 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 639.358943] env[65726]: DEBUG oslo_vmware.api [None req-b82a7f4d-3603-48cd-8e48-1d790e39dbfb tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Task: {'id': task-5115454, 'name': ReconfigVM_Task, 'duration_secs': 0.579832} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 639.360058] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-b82a7f4d-3603-48cd-8e48-1d790e39dbfb tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] [instance: cd539d8b-921b-4947-954c-8573380dbdc8] Reconfigured VM instance instance-0000000a to attach disk [datastore1] cd539d8b-921b-4947-954c-8573380dbdc8/cd539d8b-921b-4947-954c-8573380dbdc8.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 639.360767] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-fc7fe2bd-0725-4415-9790-6a344f4bb67d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.368942] env[65726]: DEBUG oslo_vmware.api [None req-b82a7f4d-3603-48cd-8e48-1d790e39dbfb tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Waiting for the task: (returnval){ [ 639.368942] env[65726]: value = "task-5115456" [ 639.368942] env[65726]: _type = "Task" [ 639.368942] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 639.387582] env[65726]: DEBUG oslo_vmware.api [None req-b82a7f4d-3603-48cd-8e48-1d790e39dbfb tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Task: {'id': task-5115456, 'name': Rename_Task} progress is 6%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 639.417021] env[65726]: DEBUG oslo_concurrency.lockutils [req-a1685450-20f0-432e-8d51-f059ae0e8b68 req-6ba2c67d-2799-4b3f-99ef-94b5740db0d1 service nova] Releasing lock "refresh_cache-21730277-cef6-4bfe-9f67-de71f6f615ff" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 639.487093] env[65726]: WARNING openstack [None req-8cc1dec8-17ac-4db3-bea9-0fb861f8b4b8 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 639.492362] env[65726]: WARNING openstack [None req-8cc1dec8-17ac-4db3-bea9-0fb861f8b4b8 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 639.629249] env[65726]: DEBUG oslo_concurrency.lockutils [None req-8fd7433e-ebac-43f4-aa60-6301be278066 tempest-ServerDiagnosticsTest-2021442419 tempest-ServerDiagnosticsTest-2021442419-project-member] Releasing lock "refresh_cache-de5cad99-1dbc-4435-98b4-987e24ba8c3c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 639.629618] env[65726]: DEBUG nova.compute.manager [None req-8fd7433e-ebac-43f4-aa60-6301be278066 tempest-ServerDiagnosticsTest-2021442419 tempest-ServerDiagnosticsTest-2021442419-project-member] [instance: de5cad99-1dbc-4435-98b4-987e24ba8c3c] Instance network_info: |[{"id": "3f320668-6397-427f-b8c4-3e63282fcb87", "address": "fa:16:3e:ca:a0:6d", "network": {"id": "55d629b8-a541-4720-9289-927bbc6e6f22", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.83", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "1539f06fa3534e90acbc3a60c4b8bd3f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "35342bcb-8b06-472e-b3c0-43fd3d6c4b30", "external-id": "nsx-vlan-transportzone-524", "segmentation_id": 524, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3f320668-63", "ovs_interfaceid": "3f320668-6397-427f-b8c4-3e63282fcb87", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 639.630148] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-8fd7433e-ebac-43f4-aa60-6301be278066 tempest-ServerDiagnosticsTest-2021442419 tempest-ServerDiagnosticsTest-2021442419-project-member] [instance: de5cad99-1dbc-4435-98b4-987e24ba8c3c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ca:a0:6d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '35342bcb-8b06-472e-b3c0-43fd3d6c4b30', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3f320668-6397-427f-b8c4-3e63282fcb87', 'vif_model': 'vmxnet3'}] {{(pid=65726) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 639.637925] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-8fd7433e-ebac-43f4-aa60-6301be278066 tempest-ServerDiagnosticsTest-2021442419 tempest-ServerDiagnosticsTest-2021442419-project-member] Creating folder: Project (0d16e812d7bc4d3285f2930454e8a434). Parent ref: group-v995008. {{(pid=65726) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 639.638113] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-6ef3189c-69c3-4dc7-9328-d53d7d8b8dfb {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.646756] env[65726]: DEBUG nova.network.neutron [None req-8cc1dec8-17ac-4db3-bea9-0fb861f8b4b8 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] [instance: 04ffdf80-c54c-4587-9bb4-d520dc440501] Updating instance_info_cache with network_info: [{"id": "ece7a121-1068-4cb7-b500-9b6a1e8c4a1c", "address": "fa:16:3e:7a:97:f3", "network": {"id": "f639fd12-fb47-4b3f-9895-6facebc89726", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-2112264538-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.3"}}], "meta": {"injected": false, "tenant_id": "3a680eae0e734ea8b2274b504d0a46d8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "257e5ea7-8b80-4301-9900-a754f1fe2031", "external-id": "nsx-vlan-transportzone-682", "segmentation_id": 682, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapece7a121-10", "ovs_interfaceid": "ece7a121-1068-4cb7-b500-9b6a1e8c4a1c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 639.650886] env[65726]: INFO nova.virt.vmwareapi.vm_util [None req-8fd7433e-ebac-43f4-aa60-6301be278066 tempest-ServerDiagnosticsTest-2021442419 tempest-ServerDiagnosticsTest-2021442419-project-member] Created folder: Project (0d16e812d7bc4d3285f2930454e8a434) in parent group-v995008. [ 639.651114] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-8fd7433e-ebac-43f4-aa60-6301be278066 tempest-ServerDiagnosticsTest-2021442419 tempest-ServerDiagnosticsTest-2021442419-project-member] Creating folder: Instances. Parent ref: group-v995044. {{(pid=65726) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 639.651380] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-cf02821a-3bee-4e1b-9c48-2e1a77429119 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.663752] env[65726]: INFO nova.virt.vmwareapi.vm_util [None req-8fd7433e-ebac-43f4-aa60-6301be278066 tempest-ServerDiagnosticsTest-2021442419 tempest-ServerDiagnosticsTest-2021442419-project-member] Created folder: Instances in parent group-v995044. [ 639.664137] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-8fd7433e-ebac-43f4-aa60-6301be278066 tempest-ServerDiagnosticsTest-2021442419 tempest-ServerDiagnosticsTest-2021442419-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 639.666182] env[65726]: WARNING openstack [None req-32ab2821-307c-47ba-ac96-c38579de1e37 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 639.666857] env[65726]: WARNING openstack [None req-32ab2821-307c-47ba-ac96-c38579de1e37 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 639.675555] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: de5cad99-1dbc-4435-98b4-987e24ba8c3c] Creating VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 639.676910] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b38f2956-6b95-4140-98ca-a71cdcdb8c1e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.703564] env[65726]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 639.703564] env[65726]: value = "task-5115459" [ 639.703564] env[65726]: _type = "Task" [ 639.703564] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 639.713843] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115459, 'name': CreateVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 639.734091] env[65726]: DEBUG oslo_vmware.api [None req-b8e65ecf-6b31-4881-a7b5-2f0a252ee6d8 tempest-VolumesAssistedSnapshotsTest-1875841935 tempest-VolumesAssistedSnapshotsTest-1875841935-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52c71cb7-1369-adc2-e167-39b606899334, 'name': SearchDatastore_Task, 'duration_secs': 0.025407} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 639.735209] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5d3ce37b-cea3-4963-a3c4-eef567604bb5 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.744068] env[65726]: DEBUG oslo_vmware.api [None req-b8e65ecf-6b31-4881-a7b5-2f0a252ee6d8 tempest-VolumesAssistedSnapshotsTest-1875841935 tempest-VolumesAssistedSnapshotsTest-1875841935-project-member] Waiting for the task: (returnval){ [ 639.744068] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52e262ca-2fbd-623f-7c53-8216b3588966" [ 639.744068] env[65726]: _type = "Task" [ 639.744068] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 639.751557] env[65726]: DEBUG oslo_vmware.api [None req-b8e65ecf-6b31-4881-a7b5-2f0a252ee6d8 tempest-VolumesAssistedSnapshotsTest-1875841935 tempest-VolumesAssistedSnapshotsTest-1875841935-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52e262ca-2fbd-623f-7c53-8216b3588966, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 639.772353] env[65726]: WARNING openstack [None req-32ab2821-307c-47ba-ac96-c38579de1e37 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 639.772802] env[65726]: WARNING openstack [None req-32ab2821-307c-47ba-ac96-c38579de1e37 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 639.822361] env[65726]: WARNING openstack [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 639.822709] env[65726]: WARNING openstack [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 639.830021] env[65726]: DEBUG nova.network.neutron [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] [instance: 811c1bbd-259c-4f5e-9a61-581bfba0ec4d] Instance cache missing network info. {{(pid=65726) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 639.831301] env[65726]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 639.831301] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52332b92-be9b-d680-de8d-45137ac26750" [ 639.831301] env[65726]: _type = "HttpNfcLease" [ 639.831301] env[65726]: } is ready. {{(pid=65726) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 639.839287] env[65726]: DEBUG oslo_vmware.rw_handles [None req-4240f7e0-4ab1-4718-b82f-a2d7b6caacae tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 639.839287] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52332b92-be9b-d680-de8d-45137ac26750" [ 639.839287] env[65726]: _type = "HttpNfcLease" [ 639.839287] env[65726]: }. {{(pid=65726) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 639.840537] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a4c1cea-816b-4ca2-b24d-8e065b9d2dcf {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.844456] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a0754b39-9b94-45fe-86fb-b66ad1bd90cb tempest-DeleteServersAdminTestJSON-1071157516 tempest-DeleteServersAdminTestJSON-1071157516-project-admin] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.038s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 639.847249] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 10.534s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 639.848744] env[65726]: INFO nova.compute.claims [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] [instance: 3a60b719-9db7-436e-9908-25f6a233c465] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 639.858921] env[65726]: DEBUG oslo_vmware.rw_handles [None req-4240f7e0-4ab1-4718-b82f-a2d7b6caacae tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52a00c2f-2a99-27a5-f5a2-5fcf9b09ffeb/disk-0.vmdk from lease info. {{(pid=65726) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 639.860057] env[65726]: DEBUG oslo_vmware.rw_handles [None req-4240f7e0-4ab1-4718-b82f-a2d7b6caacae tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52a00c2f-2a99-27a5-f5a2-5fcf9b09ffeb/disk-0.vmdk for reading. {{(pid=65726) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 639.920570] env[65726]: INFO nova.scheduler.client.report [None req-a0754b39-9b94-45fe-86fb-b66ad1bd90cb tempest-DeleteServersAdminTestJSON-1071157516 tempest-DeleteServersAdminTestJSON-1071157516-project-admin] Deleted allocations for instance bb64fc41-19d7-49dc-a17a-9dd74730130f [ 639.928751] env[65726]: DEBUG nova.network.neutron [None req-32ab2821-307c-47ba-ac96-c38579de1e37 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] [instance: 0e064341-4e4a-407b-8c26-3eb04b409029] Updating instance_info_cache with network_info: [{"id": "9d3c5bc9-aa25-4013-b8dc-f565887d2736", "address": "fa:16:3e:f9:95:de", "network": {"id": "55d629b8-a541-4720-9289-927bbc6e6f22", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.93", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "1539f06fa3534e90acbc3a60c4b8bd3f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "35342bcb-8b06-472e-b3c0-43fd3d6c4b30", "external-id": "nsx-vlan-transportzone-524", "segmentation_id": 524, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9d3c5bc9-aa", "ovs_interfaceid": "9d3c5bc9-aa25-4013-b8dc-f565887d2736", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 639.947426] env[65726]: DEBUG oslo_vmware.api [None req-b82a7f4d-3603-48cd-8e48-1d790e39dbfb tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Task: {'id': task-5115456, 'name': Rename_Task, 'duration_secs': 0.162525} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 639.947426] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-b82a7f4d-3603-48cd-8e48-1d790e39dbfb tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] [instance: cd539d8b-921b-4947-954c-8573380dbdc8] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 639.947682] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-22089a8d-1514-47cc-b067-807d3d1689b7 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.956802] env[65726]: WARNING openstack [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 639.959112] env[65726]: WARNING openstack [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 639.968753] env[65726]: DEBUG oslo_vmware.api [None req-b82a7f4d-3603-48cd-8e48-1d790e39dbfb tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Waiting for the task: (returnval){ [ 639.968753] env[65726]: value = "task-5115460" [ 639.968753] env[65726]: _type = "Task" [ 639.968753] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 639.976077] env[65726]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-47824ff1-f530-493f-96f3-9095c0d2b1e8 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.985039] env[65726]: DEBUG oslo_vmware.api [None req-b82a7f4d-3603-48cd-8e48-1d790e39dbfb tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Task: {'id': task-5115460, 'name': PowerOnVM_Task} progress is 33%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 640.086581] env[65726]: WARNING openstack [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 640.086983] env[65726]: WARNING openstack [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 640.151720] env[65726]: DEBUG oslo_concurrency.lockutils [None req-8cc1dec8-17ac-4db3-bea9-0fb861f8b4b8 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Releasing lock "refresh_cache-04ffdf80-c54c-4587-9bb4-d520dc440501" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 640.152139] env[65726]: DEBUG nova.compute.manager [None req-8cc1dec8-17ac-4db3-bea9-0fb861f8b4b8 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] [instance: 04ffdf80-c54c-4587-9bb4-d520dc440501] Instance network_info: |[{"id": "ece7a121-1068-4cb7-b500-9b6a1e8c4a1c", "address": "fa:16:3e:7a:97:f3", "network": {"id": "f639fd12-fb47-4b3f-9895-6facebc89726", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-2112264538-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.3"}}], "meta": {"injected": false, "tenant_id": "3a680eae0e734ea8b2274b504d0a46d8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "257e5ea7-8b80-4301-9900-a754f1fe2031", "external-id": "nsx-vlan-transportzone-682", "segmentation_id": 682, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapece7a121-10", "ovs_interfaceid": "ece7a121-1068-4cb7-b500-9b6a1e8c4a1c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 640.152637] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-8cc1dec8-17ac-4db3-bea9-0fb861f8b4b8 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] [instance: 04ffdf80-c54c-4587-9bb4-d520dc440501] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:7a:97:f3', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '257e5ea7-8b80-4301-9900-a754f1fe2031', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ece7a121-1068-4cb7-b500-9b6a1e8c4a1c', 'vif_model': 'vmxnet3'}] {{(pid=65726) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 640.161809] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-8cc1dec8-17ac-4db3-bea9-0fb861f8b4b8 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 640.162517] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 04ffdf80-c54c-4587-9bb4-d520dc440501] Creating VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 640.162916] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1d64a40b-2332-4e6a-b4d5-ce0e1aa3925f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.187211] env[65726]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 640.187211] env[65726]: value = "task-5115461" [ 640.187211] env[65726]: _type = "Task" [ 640.187211] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 640.205448] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115461, 'name': CreateVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 640.220072] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115459, 'name': CreateVM_Task, 'duration_secs': 0.403806} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 640.220072] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: de5cad99-1dbc-4435-98b4-987e24ba8c3c] Created VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 640.220908] env[65726]: WARNING openstack [None req-8fd7433e-ebac-43f4-aa60-6301be278066 tempest-ServerDiagnosticsTest-2021442419 tempest-ServerDiagnosticsTest-2021442419-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 640.221342] env[65726]: WARNING openstack [None req-8fd7433e-ebac-43f4-aa60-6301be278066 tempest-ServerDiagnosticsTest-2021442419 tempest-ServerDiagnosticsTest-2021442419-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 640.231901] env[65726]: DEBUG oslo_concurrency.lockutils [None req-8fd7433e-ebac-43f4-aa60-6301be278066 tempest-ServerDiagnosticsTest-2021442419 tempest-ServerDiagnosticsTest-2021442419-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 640.231901] env[65726]: DEBUG oslo_concurrency.lockutils [None req-8fd7433e-ebac-43f4-aa60-6301be278066 tempest-ServerDiagnosticsTest-2021442419 tempest-ServerDiagnosticsTest-2021442419-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 640.231901] env[65726]: DEBUG oslo_concurrency.lockutils [None req-8fd7433e-ebac-43f4-aa60-6301be278066 tempest-ServerDiagnosticsTest-2021442419 tempest-ServerDiagnosticsTest-2021442419-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 640.233575] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-36e7c22e-3e52-4018-83c0-a810ac7fc7bc {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.236915] env[65726]: DEBUG nova.compute.manager [req-9d2ca3dc-c2dc-4dff-84d7-ded6698a1c31 req-8bbed885-1517-4a79-af77-fedddb7d8a2a service nova] [instance: de5cad99-1dbc-4435-98b4-987e24ba8c3c] Received event network-vif-plugged-3f320668-6397-427f-b8c4-3e63282fcb87 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 640.237185] env[65726]: DEBUG oslo_concurrency.lockutils [req-9d2ca3dc-c2dc-4dff-84d7-ded6698a1c31 req-8bbed885-1517-4a79-af77-fedddb7d8a2a service nova] Acquiring lock "de5cad99-1dbc-4435-98b4-987e24ba8c3c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 640.237384] env[65726]: DEBUG oslo_concurrency.lockutils [req-9d2ca3dc-c2dc-4dff-84d7-ded6698a1c31 req-8bbed885-1517-4a79-af77-fedddb7d8a2a service nova] Lock "de5cad99-1dbc-4435-98b4-987e24ba8c3c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 640.237582] env[65726]: DEBUG oslo_concurrency.lockutils [req-9d2ca3dc-c2dc-4dff-84d7-ded6698a1c31 req-8bbed885-1517-4a79-af77-fedddb7d8a2a service nova] Lock "de5cad99-1dbc-4435-98b4-987e24ba8c3c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 640.237851] env[65726]: DEBUG nova.compute.manager [req-9d2ca3dc-c2dc-4dff-84d7-ded6698a1c31 req-8bbed885-1517-4a79-af77-fedddb7d8a2a service nova] [instance: de5cad99-1dbc-4435-98b4-987e24ba8c3c] No waiting events found dispatching network-vif-plugged-3f320668-6397-427f-b8c4-3e63282fcb87 {{(pid=65726) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 640.238102] env[65726]: WARNING nova.compute.manager [req-9d2ca3dc-c2dc-4dff-84d7-ded6698a1c31 req-8bbed885-1517-4a79-af77-fedddb7d8a2a service nova] [instance: de5cad99-1dbc-4435-98b4-987e24ba8c3c] Received unexpected event network-vif-plugged-3f320668-6397-427f-b8c4-3e63282fcb87 for instance with vm_state building and task_state spawning. [ 640.243951] env[65726]: DEBUG oslo_vmware.api [None req-8fd7433e-ebac-43f4-aa60-6301be278066 tempest-ServerDiagnosticsTest-2021442419 tempest-ServerDiagnosticsTest-2021442419-project-member] Waiting for the task: (returnval){ [ 640.243951] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52f17334-55e4-7d04-5be9-9120c3f2a418" [ 640.243951] env[65726]: _type = "Task" [ 640.243951] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 640.261835] env[65726]: DEBUG oslo_vmware.api [None req-8fd7433e-ebac-43f4-aa60-6301be278066 tempest-ServerDiagnosticsTest-2021442419 tempest-ServerDiagnosticsTest-2021442419-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52f17334-55e4-7d04-5be9-9120c3f2a418, 'name': SearchDatastore_Task, 'duration_secs': 0.012146} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 640.266199] env[65726]: DEBUG oslo_concurrency.lockutils [None req-8fd7433e-ebac-43f4-aa60-6301be278066 tempest-ServerDiagnosticsTest-2021442419 tempest-ServerDiagnosticsTest-2021442419-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 640.266375] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-8fd7433e-ebac-43f4-aa60-6301be278066 tempest-ServerDiagnosticsTest-2021442419 tempest-ServerDiagnosticsTest-2021442419-project-member] [instance: de5cad99-1dbc-4435-98b4-987e24ba8c3c] Processing image b52362a3-ee8a-4cbf-b06f-513b0cc8f95c {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 640.266684] env[65726]: DEBUG oslo_concurrency.lockutils [None req-8fd7433e-ebac-43f4-aa60-6301be278066 tempest-ServerDiagnosticsTest-2021442419 tempest-ServerDiagnosticsTest-2021442419-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 640.266779] env[65726]: DEBUG oslo_vmware.api [None req-b8e65ecf-6b31-4881-a7b5-2f0a252ee6d8 tempest-VolumesAssistedSnapshotsTest-1875841935 tempest-VolumesAssistedSnapshotsTest-1875841935-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52e262ca-2fbd-623f-7c53-8216b3588966, 'name': SearchDatastore_Task, 'duration_secs': 0.018466} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 640.269032] env[65726]: DEBUG nova.network.neutron [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] [instance: 811c1bbd-259c-4f5e-9a61-581bfba0ec4d] Updating instance_info_cache with network_info: [{"id": "205b7b13-8d99-402c-9587-c84362a7af94", "address": "fa:16:3e:fd:96:82", "network": {"id": "b579021f-99ae-447a-b69d-293190930add", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1055463010-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4d508c846ae64e2c84ee281fe59af610", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5fb6e5b3-7be7-4018-8d9e-2133d926178c", "external-id": "nsx-vlan-transportzone-188", "segmentation_id": 188, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap205b7b13-8d", "ovs_interfaceid": "205b7b13-8d99-402c-9587-c84362a7af94", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 640.270467] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b8e65ecf-6b31-4881-a7b5-2f0a252ee6d8 tempest-VolumesAssistedSnapshotsTest-1875841935 tempest-VolumesAssistedSnapshotsTest-1875841935-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 640.270467] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-b8e65ecf-6b31-4881-a7b5-2f0a252ee6d8 tempest-VolumesAssistedSnapshotsTest-1875841935 tempest-VolumesAssistedSnapshotsTest-1875841935-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] aa2716a9-2cf8-4e51-800d-1c8f98b2cf05/aa2716a9-2cf8-4e51-800d-1c8f98b2cf05.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 640.270977] env[65726]: DEBUG oslo_concurrency.lockutils [None req-8fd7433e-ebac-43f4-aa60-6301be278066 tempest-ServerDiagnosticsTest-2021442419 tempest-ServerDiagnosticsTest-2021442419-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 640.271183] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-8fd7433e-ebac-43f4-aa60-6301be278066 tempest-ServerDiagnosticsTest-2021442419 tempest-ServerDiagnosticsTest-2021442419-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 640.271403] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7a7cc833-74ec-416c-88a2-108c7cce8560 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.273518] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-bb2c078d-54f7-4acb-aa2d-3a0b2788cfa6 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.282540] env[65726]: DEBUG oslo_vmware.api [None req-b8e65ecf-6b31-4881-a7b5-2f0a252ee6d8 tempest-VolumesAssistedSnapshotsTest-1875841935 tempest-VolumesAssistedSnapshotsTest-1875841935-project-member] Waiting for the task: (returnval){ [ 640.282540] env[65726]: value = "task-5115462" [ 640.282540] env[65726]: _type = "Task" [ 640.282540] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 640.283964] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-8fd7433e-ebac-43f4-aa60-6301be278066 tempest-ServerDiagnosticsTest-2021442419 tempest-ServerDiagnosticsTest-2021442419-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 640.283964] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-8fd7433e-ebac-43f4-aa60-6301be278066 tempest-ServerDiagnosticsTest-2021442419 tempest-ServerDiagnosticsTest-2021442419-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65726) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 640.288685] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e3ade2e2-f2bd-450a-ac9c-08f330693111 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.308167] env[65726]: DEBUG oslo_vmware.api [None req-b8e65ecf-6b31-4881-a7b5-2f0a252ee6d8 tempest-VolumesAssistedSnapshotsTest-1875841935 tempest-VolumesAssistedSnapshotsTest-1875841935-project-member] Task: {'id': task-5115462, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 640.311178] env[65726]: DEBUG oslo_vmware.api [None req-8fd7433e-ebac-43f4-aa60-6301be278066 tempest-ServerDiagnosticsTest-2021442419 tempest-ServerDiagnosticsTest-2021442419-project-member] Waiting for the task: (returnval){ [ 640.311178] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52b225ea-97b5-f2b6-ba05-b871543ba652" [ 640.311178] env[65726]: _type = "Task" [ 640.311178] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 640.324185] env[65726]: DEBUG oslo_vmware.api [None req-8fd7433e-ebac-43f4-aa60-6301be278066 tempest-ServerDiagnosticsTest-2021442419 tempest-ServerDiagnosticsTest-2021442419-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52b225ea-97b5-f2b6-ba05-b871543ba652, 'name': SearchDatastore_Task, 'duration_secs': 0.010583} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 640.325219] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-512e3b5e-29a1-445a-a365-34266c5eb884 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.333025] env[65726]: DEBUG oslo_vmware.api [None req-8fd7433e-ebac-43f4-aa60-6301be278066 tempest-ServerDiagnosticsTest-2021442419 tempest-ServerDiagnosticsTest-2021442419-project-member] Waiting for the task: (returnval){ [ 640.333025] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]526d0285-4b4d-0d10-c9fe-6241207dfee5" [ 640.333025] env[65726]: _type = "Task" [ 640.333025] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 640.345199] env[65726]: DEBUG oslo_vmware.api [None req-8fd7433e-ebac-43f4-aa60-6301be278066 tempest-ServerDiagnosticsTest-2021442419 tempest-ServerDiagnosticsTest-2021442419-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]526d0285-4b4d-0d10-c9fe-6241207dfee5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 640.437856] env[65726]: DEBUG oslo_concurrency.lockutils [None req-32ab2821-307c-47ba-ac96-c38579de1e37 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Releasing lock "refresh_cache-0e064341-4e4a-407b-8c26-3eb04b409029" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 640.447656] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a0754b39-9b94-45fe-86fb-b66ad1bd90cb tempest-DeleteServersAdminTestJSON-1071157516 tempest-DeleteServersAdminTestJSON-1071157516-project-admin] Lock "bb64fc41-19d7-49dc-a17a-9dd74730130f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 20.915s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 640.454525] env[65726]: DEBUG nova.compute.manager [None req-252fca3b-d488-4d9b-bee5-2db98c37e592 tempest-ServerDiagnosticsV248Test-1827630017 tempest-ServerDiagnosticsV248Test-1827630017-project-admin] [instance: 4d4dc88b-0d6c-430c-ab4f-c81a56e4657b] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 640.456307] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f9a6f79-69b6-43c3-ac98-d0ab28e768a6 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.470240] env[65726]: INFO nova.compute.manager [None req-252fca3b-d488-4d9b-bee5-2db98c37e592 tempest-ServerDiagnosticsV248Test-1827630017 tempest-ServerDiagnosticsV248Test-1827630017-project-admin] [instance: 4d4dc88b-0d6c-430c-ab4f-c81a56e4657b] Retrieving diagnostics [ 640.475472] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2dade8ea-6ab0-451d-907f-4cd3263b89b6 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.531941] env[65726]: DEBUG oslo_vmware.api [None req-b82a7f4d-3603-48cd-8e48-1d790e39dbfb tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Task: {'id': task-5115460, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 640.605237] env[65726]: DEBUG oslo_concurrency.lockutils [None req-bb345b26-9f49-4e52-904f-99ccc71299ac tempest-AttachInterfacesV270Test-133387050 tempest-AttachInterfacesV270Test-133387050-project-member] Acquiring lock "c977257e-dce0-4ecc-b42b-0f6ebab38797" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 640.606603] env[65726]: DEBUG oslo_concurrency.lockutils [None req-bb345b26-9f49-4e52-904f-99ccc71299ac tempest-AttachInterfacesV270Test-133387050 tempest-AttachInterfacesV270Test-133387050-project-member] Lock "c977257e-dce0-4ecc-b42b-0f6ebab38797" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 640.696852] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115461, 'name': CreateVM_Task, 'duration_secs': 0.437537} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 640.697059] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 04ffdf80-c54c-4587-9bb4-d520dc440501] Created VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 640.697847] env[65726]: WARNING openstack [None req-8cc1dec8-17ac-4db3-bea9-0fb861f8b4b8 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 640.698236] env[65726]: WARNING openstack [None req-8cc1dec8-17ac-4db3-bea9-0fb861f8b4b8 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 640.706284] env[65726]: DEBUG oslo_concurrency.lockutils [None req-8cc1dec8-17ac-4db3-bea9-0fb861f8b4b8 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 640.706284] env[65726]: DEBUG oslo_concurrency.lockutils [None req-8cc1dec8-17ac-4db3-bea9-0fb861f8b4b8 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 640.706284] env[65726]: DEBUG oslo_concurrency.lockutils [None req-8cc1dec8-17ac-4db3-bea9-0fb861f8b4b8 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 640.706284] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-51107582-8172-42d7-bf78-0648b4b3b1d3 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.716872] env[65726]: DEBUG oslo_vmware.api [None req-8cc1dec8-17ac-4db3-bea9-0fb861f8b4b8 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Waiting for the task: (returnval){ [ 640.716872] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52c83eb5-77dd-b4bd-4d36-d2ba196588e4" [ 640.716872] env[65726]: _type = "Task" [ 640.716872] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 640.728278] env[65726]: DEBUG oslo_vmware.api [None req-8cc1dec8-17ac-4db3-bea9-0fb861f8b4b8 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52c83eb5-77dd-b4bd-4d36-d2ba196588e4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 640.773532] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Releasing lock "refresh_cache-811c1bbd-259c-4f5e-9a61-581bfba0ec4d" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 640.773938] env[65726]: DEBUG nova.compute.manager [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] [instance: 811c1bbd-259c-4f5e-9a61-581bfba0ec4d] Instance network_info: |[{"id": "205b7b13-8d99-402c-9587-c84362a7af94", "address": "fa:16:3e:fd:96:82", "network": {"id": "b579021f-99ae-447a-b69d-293190930add", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1055463010-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4d508c846ae64e2c84ee281fe59af610", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5fb6e5b3-7be7-4018-8d9e-2133d926178c", "external-id": "nsx-vlan-transportzone-188", "segmentation_id": 188, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap205b7b13-8d", "ovs_interfaceid": "205b7b13-8d99-402c-9587-c84362a7af94", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 640.774440] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] [instance: 811c1bbd-259c-4f5e-9a61-581bfba0ec4d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:fd:96:82', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '5fb6e5b3-7be7-4018-8d9e-2133d926178c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '205b7b13-8d99-402c-9587-c84362a7af94', 'vif_model': 'vmxnet3'}] {{(pid=65726) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 640.782486] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Creating folder: Project (4d508c846ae64e2c84ee281fe59af610). Parent ref: group-v995008. {{(pid=65726) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 640.783220] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-48fa7c04-0b20-4a90-9ce0-de919f8f2bef {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.795813] env[65726]: DEBUG oslo_vmware.api [None req-b8e65ecf-6b31-4881-a7b5-2f0a252ee6d8 tempest-VolumesAssistedSnapshotsTest-1875841935 tempest-VolumesAssistedSnapshotsTest-1875841935-project-member] Task: {'id': task-5115462, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.497282} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 640.796164] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-b8e65ecf-6b31-4881-a7b5-2f0a252ee6d8 tempest-VolumesAssistedSnapshotsTest-1875841935 tempest-VolumesAssistedSnapshotsTest-1875841935-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] aa2716a9-2cf8-4e51-800d-1c8f98b2cf05/aa2716a9-2cf8-4e51-800d-1c8f98b2cf05.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 640.796390] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-b8e65ecf-6b31-4881-a7b5-2f0a252ee6d8 tempest-VolumesAssistedSnapshotsTest-1875841935 tempest-VolumesAssistedSnapshotsTest-1875841935-project-member] [instance: aa2716a9-2cf8-4e51-800d-1c8f98b2cf05] Extending root virtual disk to 1048576 {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 640.796742] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-797242bd-e565-4ad1-8f0b-f6955e30969a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.803087] env[65726]: DEBUG oslo_vmware.api [None req-b8e65ecf-6b31-4881-a7b5-2f0a252ee6d8 tempest-VolumesAssistedSnapshotsTest-1875841935 tempest-VolumesAssistedSnapshotsTest-1875841935-project-member] Waiting for the task: (returnval){ [ 640.803087] env[65726]: value = "task-5115464" [ 640.803087] env[65726]: _type = "Task" [ 640.803087] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 640.807556] env[65726]: INFO nova.virt.vmwareapi.vm_util [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Created folder: Project (4d508c846ae64e2c84ee281fe59af610) in parent group-v995008. [ 640.807736] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Creating folder: Instances. Parent ref: group-v995048. {{(pid=65726) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 640.808444] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-1cd57a65-f631-46fc-b7e5-9f87f9e988ad {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.818203] env[65726]: DEBUG nova.compute.manager [req-01a7e274-ce8b-49bc-8cf6-8bcdfd822dac req-4e41e538-b25d-4feb-9034-55389ca8e11e service nova] [instance: cd539d8b-921b-4947-954c-8573380dbdc8] Received event network-vif-plugged-60d92cf3-683a-4291-ba51-fad4124f3c8c {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 640.818203] env[65726]: DEBUG oslo_concurrency.lockutils [req-01a7e274-ce8b-49bc-8cf6-8bcdfd822dac req-4e41e538-b25d-4feb-9034-55389ca8e11e service nova] Acquiring lock "cd539d8b-921b-4947-954c-8573380dbdc8-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 640.818630] env[65726]: DEBUG oslo_concurrency.lockutils [req-01a7e274-ce8b-49bc-8cf6-8bcdfd822dac req-4e41e538-b25d-4feb-9034-55389ca8e11e service nova] Lock "cd539d8b-921b-4947-954c-8573380dbdc8-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 640.818630] env[65726]: DEBUG oslo_concurrency.lockutils [req-01a7e274-ce8b-49bc-8cf6-8bcdfd822dac req-4e41e538-b25d-4feb-9034-55389ca8e11e service nova] Lock "cd539d8b-921b-4947-954c-8573380dbdc8-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 640.818630] env[65726]: DEBUG nova.compute.manager [req-01a7e274-ce8b-49bc-8cf6-8bcdfd822dac req-4e41e538-b25d-4feb-9034-55389ca8e11e service nova] [instance: cd539d8b-921b-4947-954c-8573380dbdc8] No waiting events found dispatching network-vif-plugged-60d92cf3-683a-4291-ba51-fad4124f3c8c {{(pid=65726) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 640.818888] env[65726]: WARNING nova.compute.manager [req-01a7e274-ce8b-49bc-8cf6-8bcdfd822dac req-4e41e538-b25d-4feb-9034-55389ca8e11e service nova] [instance: cd539d8b-921b-4947-954c-8573380dbdc8] Received unexpected event network-vif-plugged-60d92cf3-683a-4291-ba51-fad4124f3c8c for instance with vm_state building and task_state spawning. [ 640.818953] env[65726]: DEBUG nova.compute.manager [req-01a7e274-ce8b-49bc-8cf6-8bcdfd822dac req-4e41e538-b25d-4feb-9034-55389ca8e11e service nova] [instance: cd539d8b-921b-4947-954c-8573380dbdc8] Received event network-changed-60d92cf3-683a-4291-ba51-fad4124f3c8c {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 640.819142] env[65726]: DEBUG nova.compute.manager [req-01a7e274-ce8b-49bc-8cf6-8bcdfd822dac req-4e41e538-b25d-4feb-9034-55389ca8e11e service nova] [instance: cd539d8b-921b-4947-954c-8573380dbdc8] Refreshing instance network info cache due to event network-changed-60d92cf3-683a-4291-ba51-fad4124f3c8c. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 640.819325] env[65726]: DEBUG oslo_concurrency.lockutils [req-01a7e274-ce8b-49bc-8cf6-8bcdfd822dac req-4e41e538-b25d-4feb-9034-55389ca8e11e service nova] Acquiring lock "refresh_cache-cd539d8b-921b-4947-954c-8573380dbdc8" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 640.819447] env[65726]: DEBUG oslo_concurrency.lockutils [req-01a7e274-ce8b-49bc-8cf6-8bcdfd822dac req-4e41e538-b25d-4feb-9034-55389ca8e11e service nova] Acquired lock "refresh_cache-cd539d8b-921b-4947-954c-8573380dbdc8" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 640.819596] env[65726]: DEBUG nova.network.neutron [req-01a7e274-ce8b-49bc-8cf6-8bcdfd822dac req-4e41e538-b25d-4feb-9034-55389ca8e11e service nova] [instance: cd539d8b-921b-4947-954c-8573380dbdc8] Refreshing network info cache for port 60d92cf3-683a-4291-ba51-fad4124f3c8c {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 640.821490] env[65726]: DEBUG oslo_vmware.api [None req-b8e65ecf-6b31-4881-a7b5-2f0a252ee6d8 tempest-VolumesAssistedSnapshotsTest-1875841935 tempest-VolumesAssistedSnapshotsTest-1875841935-project-member] Task: {'id': task-5115464, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 640.823618] env[65726]: INFO nova.virt.vmwareapi.vm_util [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Created folder: Instances in parent group-v995048. [ 640.823830] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 640.824028] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 811c1bbd-259c-4f5e-9a61-581bfba0ec4d] Creating VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 640.824245] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-74791831-83d4-4290-bf87-54b370953b3a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.851112] env[65726]: DEBUG oslo_vmware.api [None req-8fd7433e-ebac-43f4-aa60-6301be278066 tempest-ServerDiagnosticsTest-2021442419 tempest-ServerDiagnosticsTest-2021442419-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]526d0285-4b4d-0d10-c9fe-6241207dfee5, 'name': SearchDatastore_Task, 'duration_secs': 0.012215} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 640.852846] env[65726]: DEBUG oslo_concurrency.lockutils [None req-8fd7433e-ebac-43f4-aa60-6301be278066 tempest-ServerDiagnosticsTest-2021442419 tempest-ServerDiagnosticsTest-2021442419-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 640.853302] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-8fd7433e-ebac-43f4-aa60-6301be278066 tempest-ServerDiagnosticsTest-2021442419 tempest-ServerDiagnosticsTest-2021442419-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] de5cad99-1dbc-4435-98b4-987e24ba8c3c/de5cad99-1dbc-4435-98b4-987e24ba8c3c.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 640.853302] env[65726]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 640.853302] env[65726]: value = "task-5115466" [ 640.853302] env[65726]: _type = "Task" [ 640.853302] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 640.853609] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d594a234-cdf3-47d7-84a5-58b6dc5cffe9 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.870028] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115466, 'name': CreateVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 640.871359] env[65726]: DEBUG oslo_vmware.api [None req-8fd7433e-ebac-43f4-aa60-6301be278066 tempest-ServerDiagnosticsTest-2021442419 tempest-ServerDiagnosticsTest-2021442419-project-member] Waiting for the task: (returnval){ [ 640.871359] env[65726]: value = "task-5115467" [ 640.871359] env[65726]: _type = "Task" [ 640.871359] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 640.882565] env[65726]: DEBUG oslo_vmware.api [None req-8fd7433e-ebac-43f4-aa60-6301be278066 tempest-ServerDiagnosticsTest-2021442419 tempest-ServerDiagnosticsTest-2021442419-project-member] Task: {'id': task-5115467, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 640.983841] env[65726]: DEBUG oslo_vmware.api [None req-b82a7f4d-3603-48cd-8e48-1d790e39dbfb tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Task: {'id': task-5115460, 'name': PowerOnVM_Task, 'duration_secs': 0.673196} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 640.985125] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-b82a7f4d-3603-48cd-8e48-1d790e39dbfb tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] [instance: cd539d8b-921b-4947-954c-8573380dbdc8] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 640.985371] env[65726]: INFO nova.compute.manager [None req-b82a7f4d-3603-48cd-8e48-1d790e39dbfb tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] [instance: cd539d8b-921b-4947-954c-8573380dbdc8] Took 13.81 seconds to spawn the instance on the hypervisor. [ 640.985659] env[65726]: DEBUG nova.compute.manager [None req-b82a7f4d-3603-48cd-8e48-1d790e39dbfb tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] [instance: cd539d8b-921b-4947-954c-8573380dbdc8] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 640.987898] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-413da761-da26-43b4-8c1f-ceb9ecd02616 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.233057] env[65726]: DEBUG oslo_vmware.api [None req-8cc1dec8-17ac-4db3-bea9-0fb861f8b4b8 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52c83eb5-77dd-b4bd-4d36-d2ba196588e4, 'name': SearchDatastore_Task, 'duration_secs': 0.051356} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 641.234029] env[65726]: DEBUG oslo_concurrency.lockutils [None req-8cc1dec8-17ac-4db3-bea9-0fb861f8b4b8 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 641.234148] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-8cc1dec8-17ac-4db3-bea9-0fb861f8b4b8 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] [instance: 04ffdf80-c54c-4587-9bb4-d520dc440501] Processing image b52362a3-ee8a-4cbf-b06f-513b0cc8f95c {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 641.234386] env[65726]: DEBUG oslo_concurrency.lockutils [None req-8cc1dec8-17ac-4db3-bea9-0fb861f8b4b8 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 641.237042] env[65726]: DEBUG oslo_concurrency.lockutils [None req-8cc1dec8-17ac-4db3-bea9-0fb861f8b4b8 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 641.237042] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-8cc1dec8-17ac-4db3-bea9-0fb861f8b4b8 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 641.237042] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-71e28c76-aea4-4770-8cd6-7290b83af466 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.253913] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-8cc1dec8-17ac-4db3-bea9-0fb861f8b4b8 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 641.254134] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-8cc1dec8-17ac-4db3-bea9-0fb861f8b4b8 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65726) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 641.255104] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-00820d5d-e562-4039-a070-cbb5f1e4b758 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.264455] env[65726]: DEBUG oslo_vmware.api [None req-8cc1dec8-17ac-4db3-bea9-0fb861f8b4b8 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Waiting for the task: (returnval){ [ 641.264455] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52ade42b-c408-daeb-e997-6d3320d539f1" [ 641.264455] env[65726]: _type = "Task" [ 641.264455] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 641.282203] env[65726]: DEBUG oslo_vmware.api [None req-8cc1dec8-17ac-4db3-bea9-0fb861f8b4b8 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52ade42b-c408-daeb-e997-6d3320d539f1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 641.320892] env[65726]: DEBUG oslo_vmware.api [None req-b8e65ecf-6b31-4881-a7b5-2f0a252ee6d8 tempest-VolumesAssistedSnapshotsTest-1875841935 tempest-VolumesAssistedSnapshotsTest-1875841935-project-member] Task: {'id': task-5115464, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.107207} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 641.321915] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-b8e65ecf-6b31-4881-a7b5-2f0a252ee6d8 tempest-VolumesAssistedSnapshotsTest-1875841935 tempest-VolumesAssistedSnapshotsTest-1875841935-project-member] [instance: aa2716a9-2cf8-4e51-800d-1c8f98b2cf05] Extended root virtual disk {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 641.322383] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3120e067-539d-48d8-b5cb-5627edeb8b24 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.326531] env[65726]: WARNING openstack [req-01a7e274-ce8b-49bc-8cf6-8bcdfd822dac req-4e41e538-b25d-4feb-9034-55389ca8e11e service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 641.330638] env[65726]: WARNING openstack [req-01a7e274-ce8b-49bc-8cf6-8bcdfd822dac req-4e41e538-b25d-4feb-9034-55389ca8e11e service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 641.370617] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-b8e65ecf-6b31-4881-a7b5-2f0a252ee6d8 tempest-VolumesAssistedSnapshotsTest-1875841935 tempest-VolumesAssistedSnapshotsTest-1875841935-project-member] [instance: aa2716a9-2cf8-4e51-800d-1c8f98b2cf05] Reconfiguring VM instance instance-0000000b to attach disk [datastore1] aa2716a9-2cf8-4e51-800d-1c8f98b2cf05/aa2716a9-2cf8-4e51-800d-1c8f98b2cf05.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 641.376755] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1ec5b997-55eb-41ce-8fc2-7ffaedcdc0f6 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.403502] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115466, 'name': CreateVM_Task} progress is 99%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 641.408409] env[65726]: DEBUG oslo_vmware.api [None req-b8e65ecf-6b31-4881-a7b5-2f0a252ee6d8 tempest-VolumesAssistedSnapshotsTest-1875841935 tempest-VolumesAssistedSnapshotsTest-1875841935-project-member] Waiting for the task: (returnval){ [ 641.408409] env[65726]: value = "task-5115468" [ 641.408409] env[65726]: _type = "Task" [ 641.408409] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 641.408752] env[65726]: DEBUG oslo_vmware.api [None req-8fd7433e-ebac-43f4-aa60-6301be278066 tempest-ServerDiagnosticsTest-2021442419 tempest-ServerDiagnosticsTest-2021442419-project-member] Task: {'id': task-5115467, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 641.410986] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99c1a241-3088-4146-bcb4-cec9f5cc30fd {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.426263] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c3d7ece-08f7-4567-ac30-c08c7ad7f091 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.431276] env[65726]: DEBUG oslo_vmware.api [None req-b8e65ecf-6b31-4881-a7b5-2f0a252ee6d8 tempest-VolumesAssistedSnapshotsTest-1875841935 tempest-VolumesAssistedSnapshotsTest-1875841935-project-member] Task: {'id': task-5115468, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 641.478173] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c59b9a0-3fd6-4ca3-8151-a45aa32e86d7 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.487306] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c155b77b-b634-4da7-9a37-6819fc1f5e92 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.504020] env[65726]: DEBUG nova.compute.provider_tree [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 641.512922] env[65726]: INFO nova.compute.manager [None req-b82a7f4d-3603-48cd-8e48-1d790e39dbfb tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] [instance: cd539d8b-921b-4947-954c-8573380dbdc8] Took 33.91 seconds to build instance. [ 641.778673] env[65726]: DEBUG oslo_vmware.api [None req-8cc1dec8-17ac-4db3-bea9-0fb861f8b4b8 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52ade42b-c408-daeb-e997-6d3320d539f1, 'name': SearchDatastore_Task, 'duration_secs': 0.055018} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 641.779427] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-624d79b7-487c-4579-84b5-43a0220c6241 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.786442] env[65726]: DEBUG oslo_vmware.api [None req-8cc1dec8-17ac-4db3-bea9-0fb861f8b4b8 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Waiting for the task: (returnval){ [ 641.786442] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]529de47d-9966-6c71-9008-9d1878ee5887" [ 641.786442] env[65726]: _type = "Task" [ 641.786442] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 641.805039] env[65726]: DEBUG oslo_vmware.api [None req-8cc1dec8-17ac-4db3-bea9-0fb861f8b4b8 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]529de47d-9966-6c71-9008-9d1878ee5887, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 641.849740] env[65726]: WARNING openstack [req-01a7e274-ce8b-49bc-8cf6-8bcdfd822dac req-4e41e538-b25d-4feb-9034-55389ca8e11e service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 641.850194] env[65726]: WARNING openstack [req-01a7e274-ce8b-49bc-8cf6-8bcdfd822dac req-4e41e538-b25d-4feb-9034-55389ca8e11e service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 641.872812] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115466, 'name': CreateVM_Task, 'duration_secs': 0.626786} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 641.874153] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 811c1bbd-259c-4f5e-9a61-581bfba0ec4d] Created VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 641.875261] env[65726]: WARNING openstack [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 641.875901] env[65726]: WARNING openstack [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 641.882039] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 641.882601] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 641.882963] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 641.883275] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9f2ba3a5-ab23-4653-8747-2af8b81f6478 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.889166] env[65726]: DEBUG oslo_vmware.api [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Waiting for the task: (returnval){ [ 641.889166] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]522024c0-4905-0f90-e363-3cdff4248374" [ 641.889166] env[65726]: _type = "Task" [ 641.889166] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 641.907425] env[65726]: DEBUG oslo_vmware.api [None req-8fd7433e-ebac-43f4-aa60-6301be278066 tempest-ServerDiagnosticsTest-2021442419 tempest-ServerDiagnosticsTest-2021442419-project-member] Task: {'id': task-5115467, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.565629} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 641.912232] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-8fd7433e-ebac-43f4-aa60-6301be278066 tempest-ServerDiagnosticsTest-2021442419 tempest-ServerDiagnosticsTest-2021442419-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] de5cad99-1dbc-4435-98b4-987e24ba8c3c/de5cad99-1dbc-4435-98b4-987e24ba8c3c.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 641.912658] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-8fd7433e-ebac-43f4-aa60-6301be278066 tempest-ServerDiagnosticsTest-2021442419 tempest-ServerDiagnosticsTest-2021442419-project-member] [instance: de5cad99-1dbc-4435-98b4-987e24ba8c3c] Extending root virtual disk to 1048576 {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 641.912834] env[65726]: DEBUG oslo_vmware.api [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]522024c0-4905-0f90-e363-3cdff4248374, 'name': SearchDatastore_Task, 'duration_secs': 0.011379} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 641.913082] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d9ad0cdd-6711-4dc6-9dae-10f383806561 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.921476] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 641.921476] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] [instance: 811c1bbd-259c-4f5e-9a61-581bfba0ec4d] Processing image b52362a3-ee8a-4cbf-b06f-513b0cc8f95c {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 641.921476] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 641.927620] env[65726]: DEBUG oslo_vmware.api [None req-b8e65ecf-6b31-4881-a7b5-2f0a252ee6d8 tempest-VolumesAssistedSnapshotsTest-1875841935 tempest-VolumesAssistedSnapshotsTest-1875841935-project-member] Task: {'id': task-5115468, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 641.929394] env[65726]: DEBUG oslo_vmware.api [None req-8fd7433e-ebac-43f4-aa60-6301be278066 tempest-ServerDiagnosticsTest-2021442419 tempest-ServerDiagnosticsTest-2021442419-project-member] Waiting for the task: (returnval){ [ 641.929394] env[65726]: value = "task-5115469" [ 641.929394] env[65726]: _type = "Task" [ 641.929394] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 641.933307] env[65726]: WARNING openstack [req-01a7e274-ce8b-49bc-8cf6-8bcdfd822dac req-4e41e538-b25d-4feb-9034-55389ca8e11e service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 641.933531] env[65726]: WARNING openstack [req-01a7e274-ce8b-49bc-8cf6-8bcdfd822dac req-4e41e538-b25d-4feb-9034-55389ca8e11e service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 641.953594] env[65726]: DEBUG oslo_vmware.api [None req-8fd7433e-ebac-43f4-aa60-6301be278066 tempest-ServerDiagnosticsTest-2021442419 tempest-ServerDiagnosticsTest-2021442419-project-member] Task: {'id': task-5115469, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 641.978678] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca85ae70-8737-46da-b229-3e3859e3ee00 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.000431] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-32ab2821-307c-47ba-ac96-c38579de1e37 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] [instance: 0e064341-4e4a-407b-8c26-3eb04b409029] Updating instance '0e064341-4e4a-407b-8c26-3eb04b409029' progress to 0 {{(pid=65726) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 642.007704] env[65726]: DEBUG nova.scheduler.client.report [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 642.015931] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b82a7f4d-3603-48cd-8e48-1d790e39dbfb tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Lock "cd539d8b-921b-4947-954c-8573380dbdc8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 35.427s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 642.122611] env[65726]: DEBUG nova.network.neutron [req-01a7e274-ce8b-49bc-8cf6-8bcdfd822dac req-4e41e538-b25d-4feb-9034-55389ca8e11e service nova] [instance: cd539d8b-921b-4947-954c-8573380dbdc8] Updated VIF entry in instance network info cache for port 60d92cf3-683a-4291-ba51-fad4124f3c8c. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 642.123201] env[65726]: DEBUG nova.network.neutron [req-01a7e274-ce8b-49bc-8cf6-8bcdfd822dac req-4e41e538-b25d-4feb-9034-55389ca8e11e service nova] [instance: cd539d8b-921b-4947-954c-8573380dbdc8] Updating instance_info_cache with network_info: [{"id": "60d92cf3-683a-4291-ba51-fad4124f3c8c", "address": "fa:16:3e:03:6c:08", "network": {"id": "f639fd12-fb47-4b3f-9895-6facebc89726", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-2112264538-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.3"}}], "meta": {"injected": false, "tenant_id": "3a680eae0e734ea8b2274b504d0a46d8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "257e5ea7-8b80-4301-9900-a754f1fe2031", "external-id": "nsx-vlan-transportzone-682", "segmentation_id": 682, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap60d92cf3-68", "ovs_interfaceid": "60d92cf3-683a-4291-ba51-fad4124f3c8c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 642.305701] env[65726]: DEBUG oslo_vmware.api [None req-8cc1dec8-17ac-4db3-bea9-0fb861f8b4b8 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]529de47d-9966-6c71-9008-9d1878ee5887, 'name': SearchDatastore_Task, 'duration_secs': 0.011624} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 642.305701] env[65726]: DEBUG oslo_concurrency.lockutils [None req-8cc1dec8-17ac-4db3-bea9-0fb861f8b4b8 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 642.305701] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-8cc1dec8-17ac-4db3-bea9-0fb861f8b4b8 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] 04ffdf80-c54c-4587-9bb4-d520dc440501/04ffdf80-c54c-4587-9bb4-d520dc440501.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 642.307702] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 642.307952] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 642.308316] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8ad90e8d-3901-483c-9a9a-bbc5198d563e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.311327] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a553fba9-9026-48b0-83cc-26e605882d13 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.321178] env[65726]: DEBUG oslo_vmware.api [None req-8cc1dec8-17ac-4db3-bea9-0fb861f8b4b8 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Waiting for the task: (returnval){ [ 642.321178] env[65726]: value = "task-5115470" [ 642.321178] env[65726]: _type = "Task" [ 642.321178] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 642.323672] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 642.323672] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65726) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 642.328632] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-78c0c0d3-3263-400d-a0ed-f24812a64382 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.343545] env[65726]: DEBUG oslo_vmware.api [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Waiting for the task: (returnval){ [ 642.343545] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52a53fc9-5be5-ab54-a0ee-515460cac2d9" [ 642.343545] env[65726]: _type = "Task" [ 642.343545] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 642.344899] env[65726]: DEBUG oslo_vmware.api [None req-8cc1dec8-17ac-4db3-bea9-0fb861f8b4b8 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Task: {'id': task-5115470, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 642.357838] env[65726]: DEBUG oslo_vmware.api [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52a53fc9-5be5-ab54-a0ee-515460cac2d9, 'name': SearchDatastore_Task, 'duration_secs': 0.012418} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 642.358821] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0cc30f81-61f1-479e-80aa-3363a5a890fa {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.365710] env[65726]: DEBUG oslo_vmware.api [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Waiting for the task: (returnval){ [ 642.365710] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]520adfd9-99a2-3705-bdf4-463b48d3e45e" [ 642.365710] env[65726]: _type = "Task" [ 642.365710] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 642.376952] env[65726]: DEBUG oslo_vmware.api [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]520adfd9-99a2-3705-bdf4-463b48d3e45e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 642.428542] env[65726]: DEBUG oslo_vmware.api [None req-b8e65ecf-6b31-4881-a7b5-2f0a252ee6d8 tempest-VolumesAssistedSnapshotsTest-1875841935 tempest-VolumesAssistedSnapshotsTest-1875841935-project-member] Task: {'id': task-5115468, 'name': ReconfigVM_Task, 'duration_secs': 0.794626} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 642.428886] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-b8e65ecf-6b31-4881-a7b5-2f0a252ee6d8 tempest-VolumesAssistedSnapshotsTest-1875841935 tempest-VolumesAssistedSnapshotsTest-1875841935-project-member] [instance: aa2716a9-2cf8-4e51-800d-1c8f98b2cf05] Reconfigured VM instance instance-0000000b to attach disk [datastore1] aa2716a9-2cf8-4e51-800d-1c8f98b2cf05/aa2716a9-2cf8-4e51-800d-1c8f98b2cf05.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 642.429829] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ad587be5-b52d-40dd-89fa-1796ffe7c3ca {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.440351] env[65726]: DEBUG oslo_vmware.api [None req-b8e65ecf-6b31-4881-a7b5-2f0a252ee6d8 tempest-VolumesAssistedSnapshotsTest-1875841935 tempest-VolumesAssistedSnapshotsTest-1875841935-project-member] Waiting for the task: (returnval){ [ 642.440351] env[65726]: value = "task-5115471" [ 642.440351] env[65726]: _type = "Task" [ 642.440351] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 642.460423] env[65726]: DEBUG oslo_vmware.api [None req-b8e65ecf-6b31-4881-a7b5-2f0a252ee6d8 tempest-VolumesAssistedSnapshotsTest-1875841935 tempest-VolumesAssistedSnapshotsTest-1875841935-project-member] Task: {'id': task-5115471, 'name': Rename_Task} progress is 10%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 642.464472] env[65726]: DEBUG oslo_vmware.api [None req-8fd7433e-ebac-43f4-aa60-6301be278066 tempest-ServerDiagnosticsTest-2021442419 tempest-ServerDiagnosticsTest-2021442419-project-member] Task: {'id': task-5115469, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.14736} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 642.464737] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-8fd7433e-ebac-43f4-aa60-6301be278066 tempest-ServerDiagnosticsTest-2021442419 tempest-ServerDiagnosticsTest-2021442419-project-member] [instance: de5cad99-1dbc-4435-98b4-987e24ba8c3c] Extended root virtual disk {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 642.466341] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20a5bcf5-92cc-4489-a00e-d877e61ac58b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.494329] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-8fd7433e-ebac-43f4-aa60-6301be278066 tempest-ServerDiagnosticsTest-2021442419 tempest-ServerDiagnosticsTest-2021442419-project-member] [instance: de5cad99-1dbc-4435-98b4-987e24ba8c3c] Reconfiguring VM instance instance-0000000c to attach disk [datastore1] de5cad99-1dbc-4435-98b4-987e24ba8c3c/de5cad99-1dbc-4435-98b4-987e24ba8c3c.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 642.494746] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-67dd354a-02d7-4aee-beba-84a5c1adc75a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.517256] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-32ab2821-307c-47ba-ac96-c38579de1e37 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] [instance: 0e064341-4e4a-407b-8c26-3eb04b409029] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 642.520146] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.673s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 642.522074] env[65726]: DEBUG nova.compute.manager [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] [instance: 3a60b719-9db7-436e-9908-25f6a233c465] Start building networks asynchronously for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 642.531028] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8472c861-990e-4081-8cd8-503b37346f00 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.533039] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.101s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 642.537385] env[65726]: INFO nova.compute.claims [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] [instance: f5151062-57b1-4e4e-93f4-aab0e7f504d4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 642.538846] env[65726]: DEBUG nova.compute.manager [None req-362ec7a6-07fd-4c6b-b5fb-9865bdcb8c1a tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: 274ab469-61a9-4b7e-852c-074c871e3abf] Starting instance... {{(pid=65726) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 642.549235] env[65726]: DEBUG oslo_vmware.api [None req-8fd7433e-ebac-43f4-aa60-6301be278066 tempest-ServerDiagnosticsTest-2021442419 tempest-ServerDiagnosticsTest-2021442419-project-member] Waiting for the task: (returnval){ [ 642.549235] env[65726]: value = "task-5115472" [ 642.549235] env[65726]: _type = "Task" [ 642.549235] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 642.557911] env[65726]: DEBUG oslo_vmware.api [None req-32ab2821-307c-47ba-ac96-c38579de1e37 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Waiting for the task: (returnval){ [ 642.557911] env[65726]: value = "task-5115473" [ 642.557911] env[65726]: _type = "Task" [ 642.557911] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 642.572588] env[65726]: DEBUG oslo_vmware.api [None req-8fd7433e-ebac-43f4-aa60-6301be278066 tempest-ServerDiagnosticsTest-2021442419 tempest-ServerDiagnosticsTest-2021442419-project-member] Task: {'id': task-5115472, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 642.583969] env[65726]: DEBUG oslo_vmware.api [None req-32ab2821-307c-47ba-ac96-c38579de1e37 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Task: {'id': task-5115473, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 642.626219] env[65726]: DEBUG oslo_concurrency.lockutils [req-01a7e274-ce8b-49bc-8cf6-8bcdfd822dac req-4e41e538-b25d-4feb-9034-55389ca8e11e service nova] Releasing lock "refresh_cache-cd539d8b-921b-4947-954c-8573380dbdc8" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 642.626516] env[65726]: DEBUG nova.compute.manager [req-01a7e274-ce8b-49bc-8cf6-8bcdfd822dac req-4e41e538-b25d-4feb-9034-55389ca8e11e service nova] [instance: aa2716a9-2cf8-4e51-800d-1c8f98b2cf05] Received event network-vif-plugged-c4f80985-65f1-478f-9944-afb3538efb48 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 642.626712] env[65726]: DEBUG oslo_concurrency.lockutils [req-01a7e274-ce8b-49bc-8cf6-8bcdfd822dac req-4e41e538-b25d-4feb-9034-55389ca8e11e service nova] Acquiring lock "aa2716a9-2cf8-4e51-800d-1c8f98b2cf05-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 642.628845] env[65726]: DEBUG oslo_concurrency.lockutils [req-01a7e274-ce8b-49bc-8cf6-8bcdfd822dac req-4e41e538-b25d-4feb-9034-55389ca8e11e service nova] Lock "aa2716a9-2cf8-4e51-800d-1c8f98b2cf05-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 642.628845] env[65726]: DEBUG oslo_concurrency.lockutils [req-01a7e274-ce8b-49bc-8cf6-8bcdfd822dac req-4e41e538-b25d-4feb-9034-55389ca8e11e service nova] Lock "aa2716a9-2cf8-4e51-800d-1c8f98b2cf05-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 642.628845] env[65726]: DEBUG nova.compute.manager [req-01a7e274-ce8b-49bc-8cf6-8bcdfd822dac req-4e41e538-b25d-4feb-9034-55389ca8e11e service nova] [instance: aa2716a9-2cf8-4e51-800d-1c8f98b2cf05] No waiting events found dispatching network-vif-plugged-c4f80985-65f1-478f-9944-afb3538efb48 {{(pid=65726) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 642.628845] env[65726]: WARNING nova.compute.manager [req-01a7e274-ce8b-49bc-8cf6-8bcdfd822dac req-4e41e538-b25d-4feb-9034-55389ca8e11e service nova] [instance: aa2716a9-2cf8-4e51-800d-1c8f98b2cf05] Received unexpected event network-vif-plugged-c4f80985-65f1-478f-9944-afb3538efb48 for instance with vm_state building and task_state spawning. [ 642.628845] env[65726]: DEBUG nova.compute.manager [req-01a7e274-ce8b-49bc-8cf6-8bcdfd822dac req-4e41e538-b25d-4feb-9034-55389ca8e11e service nova] [instance: aa2716a9-2cf8-4e51-800d-1c8f98b2cf05] Received event network-changed-c4f80985-65f1-478f-9944-afb3538efb48 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 642.629324] env[65726]: DEBUG nova.compute.manager [req-01a7e274-ce8b-49bc-8cf6-8bcdfd822dac req-4e41e538-b25d-4feb-9034-55389ca8e11e service nova] [instance: aa2716a9-2cf8-4e51-800d-1c8f98b2cf05] Refreshing instance network info cache due to event network-changed-c4f80985-65f1-478f-9944-afb3538efb48. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 642.629324] env[65726]: DEBUG oslo_concurrency.lockutils [req-01a7e274-ce8b-49bc-8cf6-8bcdfd822dac req-4e41e538-b25d-4feb-9034-55389ca8e11e service nova] Acquiring lock "refresh_cache-aa2716a9-2cf8-4e51-800d-1c8f98b2cf05" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 642.629324] env[65726]: DEBUG oslo_concurrency.lockutils [req-01a7e274-ce8b-49bc-8cf6-8bcdfd822dac req-4e41e538-b25d-4feb-9034-55389ca8e11e service nova] Acquired lock "refresh_cache-aa2716a9-2cf8-4e51-800d-1c8f98b2cf05" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 642.629324] env[65726]: DEBUG nova.network.neutron [req-01a7e274-ce8b-49bc-8cf6-8bcdfd822dac req-4e41e538-b25d-4feb-9034-55389ca8e11e service nova] [instance: aa2716a9-2cf8-4e51-800d-1c8f98b2cf05] Refreshing network info cache for port c4f80985-65f1-478f-9944-afb3538efb48 {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 642.837425] env[65726]: DEBUG oslo_vmware.api [None req-8cc1dec8-17ac-4db3-bea9-0fb861f8b4b8 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Task: {'id': task-5115470, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 642.863534] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e9359f40-01a5-4a39-b7b9-47a0a48e2ede tempest-VolumesAdminNegativeTest-2004477850 tempest-VolumesAdminNegativeTest-2004477850-project-member] Acquiring lock "30b17a17-229c-4ca7-9ae6-c67ccdcd3963" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 642.863855] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e9359f40-01a5-4a39-b7b9-47a0a48e2ede tempest-VolumesAdminNegativeTest-2004477850 tempest-VolumesAdminNegativeTest-2004477850-project-member] Lock "30b17a17-229c-4ca7-9ae6-c67ccdcd3963" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 642.877587] env[65726]: DEBUG oslo_vmware.api [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]520adfd9-99a2-3705-bdf4-463b48d3e45e, 'name': SearchDatastore_Task, 'duration_secs': 0.012055} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 642.878020] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 642.879275] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] 811c1bbd-259c-4f5e-9a61-581bfba0ec4d/811c1bbd-259c-4f5e-9a61-581bfba0ec4d.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 642.880610] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f8a550b8-ab7b-4fb0-9003-ea13dc5277db {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.888546] env[65726]: DEBUG oslo_vmware.api [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Waiting for the task: (returnval){ [ 642.888546] env[65726]: value = "task-5115474" [ 642.888546] env[65726]: _type = "Task" [ 642.888546] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 642.897377] env[65726]: DEBUG oslo_concurrency.lockutils [None req-60511481-93fa-4f8b-817d-01a577f85a26 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Acquiring lock "49bcab1f-7b64-4999-abff-37771c58a271" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 642.897679] env[65726]: DEBUG oslo_concurrency.lockutils [None req-60511481-93fa-4f8b-817d-01a577f85a26 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Lock "49bcab1f-7b64-4999-abff-37771c58a271" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 642.905437] env[65726]: DEBUG oslo_vmware.api [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Task: {'id': task-5115474, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 642.953425] env[65726]: DEBUG oslo_vmware.api [None req-b8e65ecf-6b31-4881-a7b5-2f0a252ee6d8 tempest-VolumesAssistedSnapshotsTest-1875841935 tempest-VolumesAssistedSnapshotsTest-1875841935-project-member] Task: {'id': task-5115471, 'name': Rename_Task, 'duration_secs': 0.275853} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 642.953706] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-b8e65ecf-6b31-4881-a7b5-2f0a252ee6d8 tempest-VolumesAssistedSnapshotsTest-1875841935 tempest-VolumesAssistedSnapshotsTest-1875841935-project-member] [instance: aa2716a9-2cf8-4e51-800d-1c8f98b2cf05] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 642.953955] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-07041752-82aa-45b9-a5e9-f9b90b29010b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.961575] env[65726]: DEBUG oslo_vmware.api [None req-b8e65ecf-6b31-4881-a7b5-2f0a252ee6d8 tempest-VolumesAssistedSnapshotsTest-1875841935 tempest-VolumesAssistedSnapshotsTest-1875841935-project-member] Waiting for the task: (returnval){ [ 642.961575] env[65726]: value = "task-5115475" [ 642.961575] env[65726]: _type = "Task" [ 642.961575] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 642.971686] env[65726]: DEBUG oslo_vmware.api [None req-b8e65ecf-6b31-4881-a7b5-2f0a252ee6d8 tempest-VolumesAssistedSnapshotsTest-1875841935 tempest-VolumesAssistedSnapshotsTest-1875841935-project-member] Task: {'id': task-5115475, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 643.051590] env[65726]: DEBUG nova.compute.utils [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Using /dev/sd instead of None {{(pid=65726) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 643.058155] env[65726]: DEBUG nova.compute.manager [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] [instance: 3a60b719-9db7-436e-9908-25f6a233c465] Allocating IP information in the background. {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 643.062101] env[65726]: DEBUG nova.network.neutron [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] [instance: 3a60b719-9db7-436e-9908-25f6a233c465] allocate_for_instance() {{(pid=65726) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 643.062101] env[65726]: WARNING neutronclient.v2_0.client [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 643.062101] env[65726]: WARNING neutronclient.v2_0.client [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 643.062101] env[65726]: WARNING openstack [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 643.062251] env[65726]: WARNING openstack [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 643.099793] env[65726]: DEBUG oslo_vmware.api [None req-8fd7433e-ebac-43f4-aa60-6301be278066 tempest-ServerDiagnosticsTest-2021442419 tempest-ServerDiagnosticsTest-2021442419-project-member] Task: {'id': task-5115472, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 643.102653] env[65726]: DEBUG oslo_vmware.api [None req-32ab2821-307c-47ba-ac96-c38579de1e37 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Task: {'id': task-5115473, 'name': PowerOffVM_Task, 'duration_secs': 0.377186} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 643.103892] env[65726]: DEBUG oslo_concurrency.lockutils [None req-362ec7a6-07fd-4c6b-b5fb-9865bdcb8c1a tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 643.104225] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-32ab2821-307c-47ba-ac96-c38579de1e37 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] [instance: 0e064341-4e4a-407b-8c26-3eb04b409029] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 643.104536] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-32ab2821-307c-47ba-ac96-c38579de1e37 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] [instance: 0e064341-4e4a-407b-8c26-3eb04b409029] Updating instance '0e064341-4e4a-407b-8c26-3eb04b409029' progress to 17 {{(pid=65726) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 643.134327] env[65726]: WARNING openstack [req-01a7e274-ce8b-49bc-8cf6-8bcdfd822dac req-4e41e538-b25d-4feb-9034-55389ca8e11e service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 643.134894] env[65726]: WARNING openstack [req-01a7e274-ce8b-49bc-8cf6-8bcdfd822dac req-4e41e538-b25d-4feb-9034-55389ca8e11e service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 643.335179] env[65726]: DEBUG oslo_vmware.api [None req-8cc1dec8-17ac-4db3-bea9-0fb861f8b4b8 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Task: {'id': task-5115470, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.528146} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 643.335485] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-8cc1dec8-17ac-4db3-bea9-0fb861f8b4b8 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] 04ffdf80-c54c-4587-9bb4-d520dc440501/04ffdf80-c54c-4587-9bb4-d520dc440501.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 643.335697] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-8cc1dec8-17ac-4db3-bea9-0fb861f8b4b8 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] [instance: 04ffdf80-c54c-4587-9bb4-d520dc440501] Extending root virtual disk to 1048576 {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 643.335959] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c686e2d5-d081-42dc-aba7-f63122ac64a3 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.345340] env[65726]: DEBUG oslo_vmware.api [None req-8cc1dec8-17ac-4db3-bea9-0fb861f8b4b8 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Waiting for the task: (returnval){ [ 643.345340] env[65726]: value = "task-5115476" [ 643.345340] env[65726]: _type = "Task" [ 643.345340] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 643.357853] env[65726]: DEBUG oslo_vmware.api [None req-8cc1dec8-17ac-4db3-bea9-0fb861f8b4b8 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Task: {'id': task-5115476, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 643.401827] env[65726]: DEBUG oslo_vmware.api [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Task: {'id': task-5115474, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 643.451027] env[65726]: DEBUG nova.policy [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '568156063bd14e60b728431e93d2665d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4d508c846ae64e2c84ee281fe59af610', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65726) authorize /opt/stack/nova/nova/policy.py:192}} [ 643.476567] env[65726]: DEBUG oslo_vmware.api [None req-b8e65ecf-6b31-4881-a7b5-2f0a252ee6d8 tempest-VolumesAssistedSnapshotsTest-1875841935 tempest-VolumesAssistedSnapshotsTest-1875841935-project-member] Task: {'id': task-5115475, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 643.560316] env[65726]: DEBUG nova.compute.manager [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] [instance: 3a60b719-9db7-436e-9908-25f6a233c465] Start building block device mappings for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 643.586934] env[65726]: DEBUG oslo_vmware.api [None req-8fd7433e-ebac-43f4-aa60-6301be278066 tempest-ServerDiagnosticsTest-2021442419 tempest-ServerDiagnosticsTest-2021442419-project-member] Task: {'id': task-5115472, 'name': ReconfigVM_Task, 'duration_secs': 0.559259} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 643.587335] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-8fd7433e-ebac-43f4-aa60-6301be278066 tempest-ServerDiagnosticsTest-2021442419 tempest-ServerDiagnosticsTest-2021442419-project-member] [instance: de5cad99-1dbc-4435-98b4-987e24ba8c3c] Reconfigured VM instance instance-0000000c to attach disk [datastore1] de5cad99-1dbc-4435-98b4-987e24ba8c3c/de5cad99-1dbc-4435-98b4-987e24ba8c3c.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 643.588306] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f57b2597-f00d-4427-9309-dba457b9eca7 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.601719] env[65726]: DEBUG oslo_vmware.api [None req-8fd7433e-ebac-43f4-aa60-6301be278066 tempest-ServerDiagnosticsTest-2021442419 tempest-ServerDiagnosticsTest-2021442419-project-member] Waiting for the task: (returnval){ [ 643.601719] env[65726]: value = "task-5115477" [ 643.601719] env[65726]: _type = "Task" [ 643.601719] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 643.619787] env[65726]: DEBUG nova.virt.hardware [None req-32ab2821-307c-47ba-ac96-c38579de1e37 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 643.620157] env[65726]: DEBUG nova.virt.hardware [None req-32ab2821-307c-47ba-ac96-c38579de1e37 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 643.620447] env[65726]: DEBUG nova.virt.hardware [None req-32ab2821-307c-47ba-ac96-c38579de1e37 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 643.620699] env[65726]: DEBUG nova.virt.hardware [None req-32ab2821-307c-47ba-ac96-c38579de1e37 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 643.620932] env[65726]: DEBUG nova.virt.hardware [None req-32ab2821-307c-47ba-ac96-c38579de1e37 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 643.621180] env[65726]: DEBUG nova.virt.hardware [None req-32ab2821-307c-47ba-ac96-c38579de1e37 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 643.621620] env[65726]: DEBUG nova.virt.hardware [None req-32ab2821-307c-47ba-ac96-c38579de1e37 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 643.621620] env[65726]: DEBUG nova.virt.hardware [None req-32ab2821-307c-47ba-ac96-c38579de1e37 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 643.621808] env[65726]: DEBUG nova.virt.hardware [None req-32ab2821-307c-47ba-ac96-c38579de1e37 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 643.622066] env[65726]: DEBUG nova.virt.hardware [None req-32ab2821-307c-47ba-ac96-c38579de1e37 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 643.622402] env[65726]: DEBUG nova.virt.hardware [None req-32ab2821-307c-47ba-ac96-c38579de1e37 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 643.629386] env[65726]: DEBUG oslo_vmware.api [None req-8fd7433e-ebac-43f4-aa60-6301be278066 tempest-ServerDiagnosticsTest-2021442419 tempest-ServerDiagnosticsTest-2021442419-project-member] Task: {'id': task-5115477, 'name': Rename_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 643.635964] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d72f905e-ec67-423c-85dc-46926e352115 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.658050] env[65726]: DEBUG oslo_vmware.api [None req-32ab2821-307c-47ba-ac96-c38579de1e37 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Waiting for the task: (returnval){ [ 643.658050] env[65726]: value = "task-5115478" [ 643.658050] env[65726]: _type = "Task" [ 643.658050] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 643.671380] env[65726]: DEBUG oslo_vmware.api [None req-32ab2821-307c-47ba-ac96-c38579de1e37 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Task: {'id': task-5115478, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 643.816722] env[65726]: DEBUG nova.network.neutron [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] [instance: 3a60b719-9db7-436e-9908-25f6a233c465] Successfully created port: 5c88141b-7089-49fb-afaa-a38f2cc47aaa {{(pid=65726) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 643.868744] env[65726]: DEBUG oslo_vmware.api [None req-8cc1dec8-17ac-4db3-bea9-0fb861f8b4b8 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Task: {'id': task-5115476, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.116403} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 643.869103] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-8cc1dec8-17ac-4db3-bea9-0fb861f8b4b8 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] [instance: 04ffdf80-c54c-4587-9bb4-d520dc440501] Extended root virtual disk {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 643.870084] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12611ffc-f60f-422f-b8f2-31feb07ec4a8 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.899648] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-8cc1dec8-17ac-4db3-bea9-0fb861f8b4b8 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] [instance: 04ffdf80-c54c-4587-9bb4-d520dc440501] Reconfiguring VM instance instance-0000000d to attach disk [datastore1] 04ffdf80-c54c-4587-9bb4-d520dc440501/04ffdf80-c54c-4587-9bb4-d520dc440501.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 643.902910] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-13a6b4eb-3c15-4e1d-8096-f6c1124078e8 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.931448] env[65726]: DEBUG oslo_vmware.api [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Task: {'id': task-5115474, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.540665} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 643.933216] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] 811c1bbd-259c-4f5e-9a61-581bfba0ec4d/811c1bbd-259c-4f5e-9a61-581bfba0ec4d.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 643.933492] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] [instance: 811c1bbd-259c-4f5e-9a61-581bfba0ec4d] Extending root virtual disk to 1048576 {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 643.933875] env[65726]: DEBUG oslo_vmware.api [None req-8cc1dec8-17ac-4db3-bea9-0fb861f8b4b8 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Waiting for the task: (returnval){ [ 643.933875] env[65726]: value = "task-5115479" [ 643.933875] env[65726]: _type = "Task" [ 643.933875] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 643.936475] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-7cfc1ca7-0b8d-4238-9ea3-25de2ec196ec {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.947840] env[65726]: DEBUG oslo_vmware.api [None req-8cc1dec8-17ac-4db3-bea9-0fb861f8b4b8 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Task: {'id': task-5115479, 'name': ReconfigVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 643.949334] env[65726]: DEBUG oslo_vmware.api [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Waiting for the task: (returnval){ [ 643.949334] env[65726]: value = "task-5115480" [ 643.949334] env[65726]: _type = "Task" [ 643.949334] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 643.961980] env[65726]: DEBUG oslo_vmware.api [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Task: {'id': task-5115480, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 643.980038] env[65726]: DEBUG oslo_vmware.api [None req-b8e65ecf-6b31-4881-a7b5-2f0a252ee6d8 tempest-VolumesAssistedSnapshotsTest-1875841935 tempest-VolumesAssistedSnapshotsTest-1875841935-project-member] Task: {'id': task-5115475, 'name': PowerOnVM_Task, 'duration_secs': 0.677152} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 643.980038] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-b8e65ecf-6b31-4881-a7b5-2f0a252ee6d8 tempest-VolumesAssistedSnapshotsTest-1875841935 tempest-VolumesAssistedSnapshotsTest-1875841935-project-member] [instance: aa2716a9-2cf8-4e51-800d-1c8f98b2cf05] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 643.980038] env[65726]: INFO nova.compute.manager [None req-b8e65ecf-6b31-4881-a7b5-2f0a252ee6d8 tempest-VolumesAssistedSnapshotsTest-1875841935 tempest-VolumesAssistedSnapshotsTest-1875841935-project-member] [instance: aa2716a9-2cf8-4e51-800d-1c8f98b2cf05] Took 14.15 seconds to spawn the instance on the hypervisor. [ 643.980038] env[65726]: DEBUG nova.compute.manager [None req-b8e65ecf-6b31-4881-a7b5-2f0a252ee6d8 tempest-VolumesAssistedSnapshotsTest-1875841935 tempest-VolumesAssistedSnapshotsTest-1875841935-project-member] [instance: aa2716a9-2cf8-4e51-800d-1c8f98b2cf05] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 643.980038] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8db4c1ca-f388-43c1-b882-4e5320a5913a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.118743] env[65726]: DEBUG oslo_vmware.api [None req-8fd7433e-ebac-43f4-aa60-6301be278066 tempest-ServerDiagnosticsTest-2021442419 tempest-ServerDiagnosticsTest-2021442419-project-member] Task: {'id': task-5115477, 'name': Rename_Task, 'duration_secs': 0.203443} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 644.119470] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-8fd7433e-ebac-43f4-aa60-6301be278066 tempest-ServerDiagnosticsTest-2021442419 tempest-ServerDiagnosticsTest-2021442419-project-member] [instance: de5cad99-1dbc-4435-98b4-987e24ba8c3c] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 644.119470] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d639a1ff-d847-4d7b-8ad0-761d02f6dcab {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.128648] env[65726]: DEBUG oslo_vmware.api [None req-8fd7433e-ebac-43f4-aa60-6301be278066 tempest-ServerDiagnosticsTest-2021442419 tempest-ServerDiagnosticsTest-2021442419-project-member] Waiting for the task: (returnval){ [ 644.128648] env[65726]: value = "task-5115481" [ 644.128648] env[65726]: _type = "Task" [ 644.128648] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 644.141035] env[65726]: DEBUG oslo_vmware.api [None req-8fd7433e-ebac-43f4-aa60-6301be278066 tempest-ServerDiagnosticsTest-2021442419 tempest-ServerDiagnosticsTest-2021442419-project-member] Task: {'id': task-5115481, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 644.155031] env[65726]: WARNING openstack [req-01a7e274-ce8b-49bc-8cf6-8bcdfd822dac req-4e41e538-b25d-4feb-9034-55389ca8e11e service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 644.155384] env[65726]: WARNING openstack [req-01a7e274-ce8b-49bc-8cf6-8bcdfd822dac req-4e41e538-b25d-4feb-9034-55389ca8e11e service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 644.167492] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-888f1953-ee6a-4faf-81cb-48856ab6c1b1 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.177489] env[65726]: DEBUG oslo_vmware.api [None req-32ab2821-307c-47ba-ac96-c38579de1e37 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Task: {'id': task-5115478, 'name': ReconfigVM_Task, 'duration_secs': 0.309677} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 644.179954] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-32ab2821-307c-47ba-ac96-c38579de1e37 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] [instance: 0e064341-4e4a-407b-8c26-3eb04b409029] Updating instance '0e064341-4e4a-407b-8c26-3eb04b409029' progress to 33 {{(pid=65726) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 644.184423] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f52dee16-c517-4835-9e52-943311ea89fc {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.224133] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e6924e1-f75a-46f5-a18a-f6c282ce9e4b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.233271] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e3392a6-b311-4259-a630-b2493d5d895b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.249094] env[65726]: DEBUG nova.compute.provider_tree [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 644.325980] env[65726]: DEBUG oslo_concurrency.lockutils [None req-f3878371-4599-418c-a420-84e54e420a24 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Acquiring lock "423af936-a383-4652-8887-9c8b8f507909" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 644.326235] env[65726]: DEBUG oslo_concurrency.lockutils [None req-f3878371-4599-418c-a420-84e54e420a24 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Lock "423af936-a383-4652-8887-9c8b8f507909" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 644.335135] env[65726]: WARNING openstack [req-01a7e274-ce8b-49bc-8cf6-8bcdfd822dac req-4e41e538-b25d-4feb-9034-55389ca8e11e service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 644.335135] env[65726]: WARNING openstack [req-01a7e274-ce8b-49bc-8cf6-8bcdfd822dac req-4e41e538-b25d-4feb-9034-55389ca8e11e service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 644.449862] env[65726]: DEBUG oslo_vmware.api [None req-8cc1dec8-17ac-4db3-bea9-0fb861f8b4b8 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Task: {'id': task-5115479, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 644.459787] env[65726]: DEBUG oslo_vmware.api [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Task: {'id': task-5115480, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.117785} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 644.459982] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] [instance: 811c1bbd-259c-4f5e-9a61-581bfba0ec4d] Extended root virtual disk {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 644.460915] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5b98548-2acf-43cd-afb7-7c4f00fff469 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.485674] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] [instance: 811c1bbd-259c-4f5e-9a61-581bfba0ec4d] Reconfiguring VM instance instance-0000000e to attach disk [datastore1] 811c1bbd-259c-4f5e-9a61-581bfba0ec4d/811c1bbd-259c-4f5e-9a61-581bfba0ec4d.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 644.485954] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3721245c-eb13-47f7-a1ed-7a615030f111 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.509448] env[65726]: INFO nova.compute.manager [None req-b8e65ecf-6b31-4881-a7b5-2f0a252ee6d8 tempest-VolumesAssistedSnapshotsTest-1875841935 tempest-VolumesAssistedSnapshotsTest-1875841935-project-member] [instance: aa2716a9-2cf8-4e51-800d-1c8f98b2cf05] Took 28.46 seconds to build instance. [ 644.517564] env[65726]: DEBUG oslo_vmware.api [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Waiting for the task: (returnval){ [ 644.517564] env[65726]: value = "task-5115482" [ 644.517564] env[65726]: _type = "Task" [ 644.517564] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 644.529230] env[65726]: DEBUG oslo_vmware.api [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Task: {'id': task-5115482, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 644.584976] env[65726]: DEBUG nova.compute.manager [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] [instance: 3a60b719-9db7-436e-9908-25f6a233c465] Start spawning the instance on the hypervisor. {{(pid=65726) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 644.622078] env[65726]: DEBUG nova.virt.hardware [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T19:28:34Z,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1539f06fa3534e90acbc3a60c4b8bd3f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T19:28:35Z,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 644.622515] env[65726]: DEBUG nova.virt.hardware [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 644.622856] env[65726]: DEBUG nova.virt.hardware [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 644.623214] env[65726]: DEBUG nova.virt.hardware [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 644.623531] env[65726]: DEBUG nova.virt.hardware [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 644.623791] env[65726]: DEBUG nova.virt.hardware [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 644.624170] env[65726]: DEBUG nova.virt.hardware [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 644.624404] env[65726]: DEBUG nova.virt.hardware [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 644.624800] env[65726]: DEBUG nova.virt.hardware [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 644.625120] env[65726]: DEBUG nova.virt.hardware [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 644.625471] env[65726]: DEBUG nova.virt.hardware [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 644.626799] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c97a6e2-6114-47d4-ad06-f8f31c5a5d51 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.646746] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b531f75f-ee91-4101-b071-10187f3659ec {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.650662] env[65726]: DEBUG oslo_vmware.api [None req-8fd7433e-ebac-43f4-aa60-6301be278066 tempest-ServerDiagnosticsTest-2021442419 tempest-ServerDiagnosticsTest-2021442419-project-member] Task: {'id': task-5115481, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 644.690863] env[65726]: DEBUG nova.virt.hardware [None req-32ab2821-307c-47ba-ac96-c38579de1e37 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 644.691138] env[65726]: DEBUG nova.virt.hardware [None req-32ab2821-307c-47ba-ac96-c38579de1e37 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 644.691308] env[65726]: DEBUG nova.virt.hardware [None req-32ab2821-307c-47ba-ac96-c38579de1e37 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 644.691509] env[65726]: DEBUG nova.virt.hardware [None req-32ab2821-307c-47ba-ac96-c38579de1e37 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 644.691657] env[65726]: DEBUG nova.virt.hardware [None req-32ab2821-307c-47ba-ac96-c38579de1e37 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 644.691814] env[65726]: DEBUG nova.virt.hardware [None req-32ab2821-307c-47ba-ac96-c38579de1e37 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 644.692054] env[65726]: DEBUG nova.virt.hardware [None req-32ab2821-307c-47ba-ac96-c38579de1e37 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 644.692232] env[65726]: DEBUG nova.virt.hardware [None req-32ab2821-307c-47ba-ac96-c38579de1e37 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 644.692397] env[65726]: DEBUG nova.virt.hardware [None req-32ab2821-307c-47ba-ac96-c38579de1e37 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 644.692553] env[65726]: DEBUG nova.virt.hardware [None req-32ab2821-307c-47ba-ac96-c38579de1e37 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 644.692718] env[65726]: DEBUG nova.virt.hardware [None req-32ab2821-307c-47ba-ac96-c38579de1e37 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 644.698147] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-32ab2821-307c-47ba-ac96-c38579de1e37 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] [instance: 0e064341-4e4a-407b-8c26-3eb04b409029] Reconfiguring VM instance instance-00000002 to detach disk 2000 {{(pid=65726) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 644.698460] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6746a24f-45fe-4856-aa80-26d07e6588da {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.725096] env[65726]: DEBUG oslo_vmware.api [None req-32ab2821-307c-47ba-ac96-c38579de1e37 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Waiting for the task: (returnval){ [ 644.725096] env[65726]: value = "task-5115483" [ 644.725096] env[65726]: _type = "Task" [ 644.725096] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 644.734459] env[65726]: DEBUG oslo_vmware.api [None req-32ab2821-307c-47ba-ac96-c38579de1e37 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Task: {'id': task-5115483, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 644.756023] env[65726]: DEBUG nova.scheduler.client.report [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 644.956369] env[65726]: DEBUG oslo_vmware.api [None req-8cc1dec8-17ac-4db3-bea9-0fb861f8b4b8 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Task: {'id': task-5115479, 'name': ReconfigVM_Task, 'duration_secs': 0.728389} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 644.957238] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-8cc1dec8-17ac-4db3-bea9-0fb861f8b4b8 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] [instance: 04ffdf80-c54c-4587-9bb4-d520dc440501] Reconfigured VM instance instance-0000000d to attach disk [datastore1] 04ffdf80-c54c-4587-9bb4-d520dc440501/04ffdf80-c54c-4587-9bb4-d520dc440501.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 644.957927] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-92c58738-f30b-425d-91b3-362f08a4f936 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.967919] env[65726]: DEBUG oslo_vmware.api [None req-8cc1dec8-17ac-4db3-bea9-0fb861f8b4b8 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Waiting for the task: (returnval){ [ 644.967919] env[65726]: value = "task-5115484" [ 644.967919] env[65726]: _type = "Task" [ 644.967919] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 644.983703] env[65726]: DEBUG oslo_vmware.api [None req-8cc1dec8-17ac-4db3-bea9-0fb861f8b4b8 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Task: {'id': task-5115484, 'name': Rename_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 645.012653] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b8e65ecf-6b31-4881-a7b5-2f0a252ee6d8 tempest-VolumesAssistedSnapshotsTest-1875841935 tempest-VolumesAssistedSnapshotsTest-1875841935-project-member] Lock "aa2716a9-2cf8-4e51-800d-1c8f98b2cf05" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 36.396s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 645.034631] env[65726]: DEBUG oslo_vmware.api [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Task: {'id': task-5115482, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 645.097222] env[65726]: DEBUG nova.network.neutron [req-01a7e274-ce8b-49bc-8cf6-8bcdfd822dac req-4e41e538-b25d-4feb-9034-55389ca8e11e service nova] [instance: aa2716a9-2cf8-4e51-800d-1c8f98b2cf05] Updated VIF entry in instance network info cache for port c4f80985-65f1-478f-9944-afb3538efb48. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 645.097577] env[65726]: DEBUG nova.network.neutron [req-01a7e274-ce8b-49bc-8cf6-8bcdfd822dac req-4e41e538-b25d-4feb-9034-55389ca8e11e service nova] [instance: aa2716a9-2cf8-4e51-800d-1c8f98b2cf05] Updating instance_info_cache with network_info: [{"id": "c4f80985-65f1-478f-9944-afb3538efb48", "address": "fa:16:3e:35:5f:a5", "network": {"id": "da8b1180-a4d3-4444-b571-050c941634d4", "bridge": "br-int", "label": "tempest-VolumesAssistedSnapshotsTest-1753385467-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b01c77c483364e23b4a48b1e1c9c9d1c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c297fe21-cd0b-4226-813b-a65d2358d034", "external-id": "nsx-vlan-transportzone-98", "segmentation_id": 98, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc4f80985-65", "ovs_interfaceid": "c4f80985-65f1-478f-9944-afb3538efb48", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 645.141881] env[65726]: DEBUG oslo_vmware.api [None req-8fd7433e-ebac-43f4-aa60-6301be278066 tempest-ServerDiagnosticsTest-2021442419 tempest-ServerDiagnosticsTest-2021442419-project-member] Task: {'id': task-5115481, 'name': PowerOnVM_Task, 'duration_secs': 0.787336} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 645.142169] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-8fd7433e-ebac-43f4-aa60-6301be278066 tempest-ServerDiagnosticsTest-2021442419 tempest-ServerDiagnosticsTest-2021442419-project-member] [instance: de5cad99-1dbc-4435-98b4-987e24ba8c3c] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 645.142367] env[65726]: INFO nova.compute.manager [None req-8fd7433e-ebac-43f4-aa60-6301be278066 tempest-ServerDiagnosticsTest-2021442419 tempest-ServerDiagnosticsTest-2021442419-project-member] [instance: de5cad99-1dbc-4435-98b4-987e24ba8c3c] Took 12.66 seconds to spawn the instance on the hypervisor. [ 645.142600] env[65726]: DEBUG nova.compute.manager [None req-8fd7433e-ebac-43f4-aa60-6301be278066 tempest-ServerDiagnosticsTest-2021442419 tempest-ServerDiagnosticsTest-2021442419-project-member] [instance: de5cad99-1dbc-4435-98b4-987e24ba8c3c] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 645.143419] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32b23a39-2d0f-4625-9616-6d9d7cdf9e13 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.240365] env[65726]: DEBUG oslo_vmware.api [None req-32ab2821-307c-47ba-ac96-c38579de1e37 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Task: {'id': task-5115483, 'name': ReconfigVM_Task, 'duration_secs': 0.273039} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 645.240365] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-32ab2821-307c-47ba-ac96-c38579de1e37 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] [instance: 0e064341-4e4a-407b-8c26-3eb04b409029] Reconfigured VM instance instance-00000002 to detach disk 2000 {{(pid=65726) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 645.240989] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba143347-a612-4590-9c48-1efcc7372b18 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.260988] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.728s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 645.261570] env[65726]: DEBUG nova.compute.manager [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] [instance: f5151062-57b1-4e4e-93f4-aab0e7f504d4] Start building networks asynchronously for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 645.274099] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-32ab2821-307c-47ba-ac96-c38579de1e37 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] [instance: 0e064341-4e4a-407b-8c26-3eb04b409029] Reconfiguring VM instance instance-00000002 to attach disk [datastore1] 0e064341-4e4a-407b-8c26-3eb04b409029/0e064341-4e4a-407b-8c26-3eb04b409029.vmdk or device None with type thin {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 645.274099] env[65726]: DEBUG oslo_concurrency.lockutils [None req-881e6c53-2eb2-4fe8-9bf1-92c962897b3f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.259s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 645.274888] env[65726]: INFO nova.compute.claims [None req-881e6c53-2eb2-4fe8-9bf1-92c962897b3f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: f2d9090c-988f-43f4-9c81-7aa718a3438a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 645.277978] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9d16e4b4-bb54-425d-8d2d-ab5dad825a39 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.301745] env[65726]: DEBUG oslo_vmware.api [None req-32ab2821-307c-47ba-ac96-c38579de1e37 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Waiting for the task: (returnval){ [ 645.301745] env[65726]: value = "task-5115485" [ 645.301745] env[65726]: _type = "Task" [ 645.301745] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 645.317296] env[65726]: DEBUG oslo_vmware.api [None req-32ab2821-307c-47ba-ac96-c38579de1e37 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Task: {'id': task-5115485, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 645.478876] env[65726]: DEBUG oslo_vmware.api [None req-8cc1dec8-17ac-4db3-bea9-0fb861f8b4b8 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Task: {'id': task-5115484, 'name': Rename_Task, 'duration_secs': 0.195562} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 645.479264] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-8cc1dec8-17ac-4db3-bea9-0fb861f8b4b8 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] [instance: 04ffdf80-c54c-4587-9bb4-d520dc440501] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 645.479474] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1ad84530-d355-4063-95a9-107fb8bdeb96 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.488044] env[65726]: DEBUG oslo_vmware.api [None req-8cc1dec8-17ac-4db3-bea9-0fb861f8b4b8 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Waiting for the task: (returnval){ [ 645.488044] env[65726]: value = "task-5115486" [ 645.488044] env[65726]: _type = "Task" [ 645.488044] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 645.497080] env[65726]: DEBUG oslo_vmware.api [None req-8cc1dec8-17ac-4db3-bea9-0fb861f8b4b8 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Task: {'id': task-5115486, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 645.508666] env[65726]: DEBUG nova.network.neutron [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] [instance: 3a60b719-9db7-436e-9908-25f6a233c465] Successfully updated port: 5c88141b-7089-49fb-afaa-a38f2cc47aaa {{(pid=65726) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 645.515549] env[65726]: DEBUG nova.compute.manager [None req-621af3d8-d8b6-4b45-86f5-d66b91010e2e tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] [instance: fd481728-3ef0-4a96-affd-ab2dd3f596bb] Starting instance... {{(pid=65726) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 645.538280] env[65726]: DEBUG oslo_vmware.api [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Task: {'id': task-5115482, 'name': ReconfigVM_Task, 'duration_secs': 0.547171} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 645.539819] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] [instance: 811c1bbd-259c-4f5e-9a61-581bfba0ec4d] Reconfigured VM instance instance-0000000e to attach disk [datastore1] 811c1bbd-259c-4f5e-9a61-581bfba0ec4d/811c1bbd-259c-4f5e-9a61-581bfba0ec4d.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 645.541177] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e21cbb6f-eb74-4f90-ac62-17b2ad087ee6 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.550764] env[65726]: DEBUG oslo_vmware.api [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Waiting for the task: (returnval){ [ 645.550764] env[65726]: value = "task-5115487" [ 645.550764] env[65726]: _type = "Task" [ 645.550764] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 645.562516] env[65726]: DEBUG oslo_vmware.api [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Task: {'id': task-5115487, 'name': Rename_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 645.574197] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5dd7a34f-2002-4a12-9099-b0b630d2774d tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Acquiring lock "815f056b-2d88-40dd-9ef7-a891ec8d92cf" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 645.574601] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5dd7a34f-2002-4a12-9099-b0b630d2774d tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Lock "815f056b-2d88-40dd-9ef7-a891ec8d92cf" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 645.601559] env[65726]: DEBUG oslo_concurrency.lockutils [req-01a7e274-ce8b-49bc-8cf6-8bcdfd822dac req-4e41e538-b25d-4feb-9034-55389ca8e11e service nova] Releasing lock "refresh_cache-aa2716a9-2cf8-4e51-800d-1c8f98b2cf05" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 645.666069] env[65726]: INFO nova.compute.manager [None req-8fd7433e-ebac-43f4-aa60-6301be278066 tempest-ServerDiagnosticsTest-2021442419 tempest-ServerDiagnosticsTest-2021442419-project-member] [instance: de5cad99-1dbc-4435-98b4-987e24ba8c3c] Took 28.93 seconds to build instance. [ 645.776228] env[65726]: DEBUG nova.compute.utils [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Using /dev/sd instead of None {{(pid=65726) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 645.779160] env[65726]: DEBUG nova.compute.manager [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] [instance: f5151062-57b1-4e4e-93f4-aab0e7f504d4] Allocating IP information in the background. {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 645.779879] env[65726]: DEBUG nova.network.neutron [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] [instance: f5151062-57b1-4e4e-93f4-aab0e7f504d4] allocate_for_instance() {{(pid=65726) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 645.780168] env[65726]: WARNING neutronclient.v2_0.client [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 645.780660] env[65726]: WARNING neutronclient.v2_0.client [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 645.781434] env[65726]: WARNING openstack [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 645.781944] env[65726]: WARNING openstack [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 645.793971] env[65726]: DEBUG nova.compute.manager [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] [instance: f5151062-57b1-4e4e-93f4-aab0e7f504d4] Start building block device mappings for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 645.816461] env[65726]: DEBUG oslo_vmware.api [None req-32ab2821-307c-47ba-ac96-c38579de1e37 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Task: {'id': task-5115485, 'name': ReconfigVM_Task, 'duration_secs': 0.487228} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 645.816835] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-32ab2821-307c-47ba-ac96-c38579de1e37 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] [instance: 0e064341-4e4a-407b-8c26-3eb04b409029] Reconfigured VM instance instance-00000002 to attach disk [datastore1] 0e064341-4e4a-407b-8c26-3eb04b409029/0e064341-4e4a-407b-8c26-3eb04b409029.vmdk or device None with type thin {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 645.817314] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-32ab2821-307c-47ba-ac96-c38579de1e37 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] [instance: 0e064341-4e4a-407b-8c26-3eb04b409029] Updating instance '0e064341-4e4a-407b-8c26-3eb04b409029' progress to 50 {{(pid=65726) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 645.870813] env[65726]: DEBUG nova.policy [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '568156063bd14e60b728431e93d2665d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4d508c846ae64e2c84ee281fe59af610', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65726) authorize /opt/stack/nova/nova/policy.py:192}} [ 645.882721] env[65726]: DEBUG nova.compute.manager [req-a90995a5-d731-4249-85b8-09bdde3732a0 req-a1f2baad-ff22-43d1-8421-2b62b9ae2bb6 service nova] [instance: de5cad99-1dbc-4435-98b4-987e24ba8c3c] Received event network-changed-3f320668-6397-427f-b8c4-3e63282fcb87 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 645.883523] env[65726]: DEBUG nova.compute.manager [req-a90995a5-d731-4249-85b8-09bdde3732a0 req-a1f2baad-ff22-43d1-8421-2b62b9ae2bb6 service nova] [instance: de5cad99-1dbc-4435-98b4-987e24ba8c3c] Refreshing instance network info cache due to event network-changed-3f320668-6397-427f-b8c4-3e63282fcb87. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 645.883938] env[65726]: DEBUG oslo_concurrency.lockutils [req-a90995a5-d731-4249-85b8-09bdde3732a0 req-a1f2baad-ff22-43d1-8421-2b62b9ae2bb6 service nova] Acquiring lock "refresh_cache-de5cad99-1dbc-4435-98b4-987e24ba8c3c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 645.884336] env[65726]: DEBUG oslo_concurrency.lockutils [req-a90995a5-d731-4249-85b8-09bdde3732a0 req-a1f2baad-ff22-43d1-8421-2b62b9ae2bb6 service nova] Acquired lock "refresh_cache-de5cad99-1dbc-4435-98b4-987e24ba8c3c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 645.886440] env[65726]: DEBUG nova.network.neutron [req-a90995a5-d731-4249-85b8-09bdde3732a0 req-a1f2baad-ff22-43d1-8421-2b62b9ae2bb6 service nova] [instance: de5cad99-1dbc-4435-98b4-987e24ba8c3c] Refreshing network info cache for port 3f320668-6397-427f-b8c4-3e63282fcb87 {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 645.900346] env[65726]: DEBUG oslo_concurrency.lockutils [None req-4b216162-e1e7-4270-8d1f-482978b0681a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Acquiring lock "8493ba52-b950-4727-970a-19b1797b2ca1" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 645.900527] env[65726]: DEBUG oslo_concurrency.lockutils [None req-4b216162-e1e7-4270-8d1f-482978b0681a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Lock "8493ba52-b950-4727-970a-19b1797b2ca1" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 645.901212] env[65726]: DEBUG oslo_concurrency.lockutils [None req-4b216162-e1e7-4270-8d1f-482978b0681a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Acquiring lock "8493ba52-b950-4727-970a-19b1797b2ca1-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 645.901439] env[65726]: DEBUG oslo_concurrency.lockutils [None req-4b216162-e1e7-4270-8d1f-482978b0681a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Lock "8493ba52-b950-4727-970a-19b1797b2ca1-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 645.901704] env[65726]: DEBUG oslo_concurrency.lockutils [None req-4b216162-e1e7-4270-8d1f-482978b0681a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Lock "8493ba52-b950-4727-970a-19b1797b2ca1-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 645.905237] env[65726]: INFO nova.compute.manager [None req-4b216162-e1e7-4270-8d1f-482978b0681a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] [instance: 8493ba52-b950-4727-970a-19b1797b2ca1] Terminating instance [ 645.955514] env[65726]: DEBUG nova.compute.manager [req-0393c08a-a0f7-4006-9890-55aec2c8ba2a req-059d1391-338d-4ce7-8bf2-d1c94c047e32 service nova] [instance: 811c1bbd-259c-4f5e-9a61-581bfba0ec4d] Received event network-vif-plugged-205b7b13-8d99-402c-9587-c84362a7af94 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 645.955514] env[65726]: DEBUG oslo_concurrency.lockutils [req-0393c08a-a0f7-4006-9890-55aec2c8ba2a req-059d1391-338d-4ce7-8bf2-d1c94c047e32 service nova] Acquiring lock "811c1bbd-259c-4f5e-9a61-581bfba0ec4d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 645.955514] env[65726]: DEBUG oslo_concurrency.lockutils [req-0393c08a-a0f7-4006-9890-55aec2c8ba2a req-059d1391-338d-4ce7-8bf2-d1c94c047e32 service nova] Lock "811c1bbd-259c-4f5e-9a61-581bfba0ec4d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 645.955514] env[65726]: DEBUG oslo_concurrency.lockutils [req-0393c08a-a0f7-4006-9890-55aec2c8ba2a req-059d1391-338d-4ce7-8bf2-d1c94c047e32 service nova] Lock "811c1bbd-259c-4f5e-9a61-581bfba0ec4d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 645.956029] env[65726]: DEBUG nova.compute.manager [req-0393c08a-a0f7-4006-9890-55aec2c8ba2a req-059d1391-338d-4ce7-8bf2-d1c94c047e32 service nova] [instance: 811c1bbd-259c-4f5e-9a61-581bfba0ec4d] No waiting events found dispatching network-vif-plugged-205b7b13-8d99-402c-9587-c84362a7af94 {{(pid=65726) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 645.956029] env[65726]: WARNING nova.compute.manager [req-0393c08a-a0f7-4006-9890-55aec2c8ba2a req-059d1391-338d-4ce7-8bf2-d1c94c047e32 service nova] [instance: 811c1bbd-259c-4f5e-9a61-581bfba0ec4d] Received unexpected event network-vif-plugged-205b7b13-8d99-402c-9587-c84362a7af94 for instance with vm_state building and task_state spawning. [ 645.956029] env[65726]: DEBUG nova.compute.manager [req-0393c08a-a0f7-4006-9890-55aec2c8ba2a req-059d1391-338d-4ce7-8bf2-d1c94c047e32 service nova] [instance: 811c1bbd-259c-4f5e-9a61-581bfba0ec4d] Received event network-changed-205b7b13-8d99-402c-9587-c84362a7af94 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 645.956163] env[65726]: DEBUG nova.compute.manager [req-0393c08a-a0f7-4006-9890-55aec2c8ba2a req-059d1391-338d-4ce7-8bf2-d1c94c047e32 service nova] [instance: 811c1bbd-259c-4f5e-9a61-581bfba0ec4d] Refreshing instance network info cache due to event network-changed-205b7b13-8d99-402c-9587-c84362a7af94. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 645.956684] env[65726]: DEBUG oslo_concurrency.lockutils [req-0393c08a-a0f7-4006-9890-55aec2c8ba2a req-059d1391-338d-4ce7-8bf2-d1c94c047e32 service nova] Acquiring lock "refresh_cache-811c1bbd-259c-4f5e-9a61-581bfba0ec4d" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 645.956684] env[65726]: DEBUG oslo_concurrency.lockutils [req-0393c08a-a0f7-4006-9890-55aec2c8ba2a req-059d1391-338d-4ce7-8bf2-d1c94c047e32 service nova] Acquired lock "refresh_cache-811c1bbd-259c-4f5e-9a61-581bfba0ec4d" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 645.957635] env[65726]: DEBUG nova.network.neutron [req-0393c08a-a0f7-4006-9890-55aec2c8ba2a req-059d1391-338d-4ce7-8bf2-d1c94c047e32 service nova] [instance: 811c1bbd-259c-4f5e-9a61-581bfba0ec4d] Refreshing network info cache for port 205b7b13-8d99-402c-9587-c84362a7af94 {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 646.000960] env[65726]: DEBUG oslo_vmware.api [None req-8cc1dec8-17ac-4db3-bea9-0fb861f8b4b8 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Task: {'id': task-5115486, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 646.015060] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Acquiring lock "refresh_cache-3a60b719-9db7-436e-9908-25f6a233c465" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 646.015591] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Acquired lock "refresh_cache-3a60b719-9db7-436e-9908-25f6a233c465" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 646.015591] env[65726]: DEBUG nova.network.neutron [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] [instance: 3a60b719-9db7-436e-9908-25f6a233c465] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 646.065623] env[65726]: DEBUG oslo_vmware.api [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Task: {'id': task-5115487, 'name': Rename_Task, 'duration_secs': 0.247624} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 646.069263] env[65726]: DEBUG oslo_concurrency.lockutils [None req-621af3d8-d8b6-4b45-86f5-d66b91010e2e tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 646.069896] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] [instance: 811c1bbd-259c-4f5e-9a61-581bfba0ec4d] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 646.070222] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-123a1508-82da-48df-9366-b0e80d6897cb {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.080671] env[65726]: DEBUG oslo_vmware.api [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Waiting for the task: (returnval){ [ 646.080671] env[65726]: value = "task-5115488" [ 646.080671] env[65726]: _type = "Task" [ 646.080671] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 646.096810] env[65726]: DEBUG oslo_vmware.api [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Task: {'id': task-5115488, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 646.172185] env[65726]: DEBUG oslo_concurrency.lockutils [None req-8fd7433e-ebac-43f4-aa60-6301be278066 tempest-ServerDiagnosticsTest-2021442419 tempest-ServerDiagnosticsTest-2021442419-project-member] Lock "de5cad99-1dbc-4435-98b4-987e24ba8c3c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 36.629s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 646.237146] env[65726]: DEBUG nova.network.neutron [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] [instance: f5151062-57b1-4e4e-93f4-aab0e7f504d4] Successfully created port: 83750abb-1d26-4e23-a675-61ef785b858b {{(pid=65726) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 646.325927] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7691fc38-a813-4c0d-a0da-a2df46d231a6 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.367150] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08ac0813-d57a-4324-a2b2-d532e60072fa {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.394903] env[65726]: WARNING openstack [req-a90995a5-d731-4249-85b8-09bdde3732a0 req-a1f2baad-ff22-43d1-8421-2b62b9ae2bb6 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 646.394903] env[65726]: WARNING openstack [req-a90995a5-d731-4249-85b8-09bdde3732a0 req-a1f2baad-ff22-43d1-8421-2b62b9ae2bb6 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 646.400525] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-32ab2821-307c-47ba-ac96-c38579de1e37 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] [instance: 0e064341-4e4a-407b-8c26-3eb04b409029] Updating instance '0e064341-4e4a-407b-8c26-3eb04b409029' progress to 67 {{(pid=65726) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 646.409425] env[65726]: DEBUG nova.compute.manager [None req-4b216162-e1e7-4270-8d1f-482978b0681a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] [instance: 8493ba52-b950-4727-970a-19b1797b2ca1] Start destroying the instance on the hypervisor. {{(pid=65726) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 646.409777] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-4b216162-e1e7-4270-8d1f-482978b0681a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] [instance: 8493ba52-b950-4727-970a-19b1797b2ca1] Destroying instance {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 646.411565] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec941f06-ed55-4096-93f9-34a6c6421f0c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.427105] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-4b216162-e1e7-4270-8d1f-482978b0681a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] [instance: 8493ba52-b950-4727-970a-19b1797b2ca1] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 646.427465] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3e191b0d-02c2-42bd-b32e-7d7b5f5863cb {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.436681] env[65726]: DEBUG oslo_vmware.api [None req-4b216162-e1e7-4270-8d1f-482978b0681a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Waiting for the task: (returnval){ [ 646.436681] env[65726]: value = "task-5115489" [ 646.436681] env[65726]: _type = "Task" [ 646.436681] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 646.455022] env[65726]: DEBUG oslo_vmware.api [None req-4b216162-e1e7-4270-8d1f-482978b0681a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Task: {'id': task-5115489, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 646.461252] env[65726]: WARNING openstack [req-0393c08a-a0f7-4006-9890-55aec2c8ba2a req-059d1391-338d-4ce7-8bf2-d1c94c047e32 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 646.461702] env[65726]: WARNING openstack [req-0393c08a-a0f7-4006-9890-55aec2c8ba2a req-059d1391-338d-4ce7-8bf2-d1c94c047e32 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 646.504385] env[65726]: DEBUG oslo_vmware.api [None req-8cc1dec8-17ac-4db3-bea9-0fb861f8b4b8 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Task: {'id': task-5115486, 'name': PowerOnVM_Task, 'duration_secs': 0.664351} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 646.504777] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-8cc1dec8-17ac-4db3-bea9-0fb861f8b4b8 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] [instance: 04ffdf80-c54c-4587-9bb4-d520dc440501] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 646.504929] env[65726]: INFO nova.compute.manager [None req-8cc1dec8-17ac-4db3-bea9-0fb861f8b4b8 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] [instance: 04ffdf80-c54c-4587-9bb4-d520dc440501] Took 11.50 seconds to spawn the instance on the hypervisor. [ 646.505242] env[65726]: DEBUG nova.compute.manager [None req-8cc1dec8-17ac-4db3-bea9-0fb861f8b4b8 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] [instance: 04ffdf80-c54c-4587-9bb4-d520dc440501] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 646.506166] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a1f060b-5464-4196-a1c4-8ea375eb7b0f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.522082] env[65726]: WARNING openstack [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 646.522082] env[65726]: WARNING openstack [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 646.526545] env[65726]: DEBUG nova.network.neutron [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] [instance: 3a60b719-9db7-436e-9908-25f6a233c465] Instance cache missing network info. {{(pid=65726) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 646.593952] env[65726]: DEBUG oslo_vmware.api [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Task: {'id': task-5115488, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 646.676357] env[65726]: DEBUG nova.compute.manager [None req-06f67a27-e6e0-4078-804d-cff1137abcad tempest-ServersTestBootFromVolume-762458528 tempest-ServersTestBootFromVolume-762458528-project-member] [instance: 5a8d4357-4b44-4a19-b1da-42d188c38adc] Starting instance... {{(pid=65726) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 646.811840] env[65726]: DEBUG nova.compute.manager [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] [instance: f5151062-57b1-4e4e-93f4-aab0e7f504d4] Start spawning the instance on the hypervisor. {{(pid=65726) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 646.846059] env[65726]: DEBUG nova.virt.hardware [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T19:28:34Z,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1539f06fa3534e90acbc3a60c4b8bd3f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T19:28:35Z,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 646.846302] env[65726]: DEBUG nova.virt.hardware [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 646.846478] env[65726]: DEBUG nova.virt.hardware [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 646.846629] env[65726]: DEBUG nova.virt.hardware [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 646.846760] env[65726]: DEBUG nova.virt.hardware [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 646.847046] env[65726]: DEBUG nova.virt.hardware [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 646.847476] env[65726]: DEBUG nova.virt.hardware [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 646.847476] env[65726]: DEBUG nova.virt.hardware [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 646.847476] env[65726]: DEBUG nova.virt.hardware [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 646.847739] env[65726]: DEBUG nova.virt.hardware [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 646.847739] env[65726]: DEBUG nova.virt.hardware [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 646.848776] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce7707db-0bed-434b-a610-54c9b40ce4b8 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.865234] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b44196d6-be0b-48b9-b570-fb73714849c0 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.915669] env[65726]: WARNING neutronclient.v2_0.client [None req-32ab2821-307c-47ba-ac96-c38579de1e37 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 646.946405] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b61dc040-de4f-409d-a641-fa7ac0f96770 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.955144] env[65726]: DEBUG oslo_vmware.api [None req-4b216162-e1e7-4270-8d1f-482978b0681a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Task: {'id': task-5115489, 'name': PowerOffVM_Task, 'duration_secs': 0.380602} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 646.957139] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-4b216162-e1e7-4270-8d1f-482978b0681a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] [instance: 8493ba52-b950-4727-970a-19b1797b2ca1] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 646.957331] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-4b216162-e1e7-4270-8d1f-482978b0681a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] [instance: 8493ba52-b950-4727-970a-19b1797b2ca1] Unregistering the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 646.957640] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f5c6a4da-99e2-4a96-8ac5-046a89d39dd3 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.960172] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5db8ede8-0aaa-48f3-9f9c-fe249de4fd97 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.998848] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84e8b8a3-8459-4c20-b72b-e57173da6092 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.009710] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbfba739-bdeb-4f00-ab77-ea5d5153c65f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.041424] env[65726]: DEBUG nova.compute.provider_tree [None req-881e6c53-2eb2-4fe8-9bf1-92c962897b3f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 647.045990] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-4b216162-e1e7-4270-8d1f-482978b0681a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] [instance: 8493ba52-b950-4727-970a-19b1797b2ca1] Unregistered the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 647.045990] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-4b216162-e1e7-4270-8d1f-482978b0681a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] [instance: 8493ba52-b950-4727-970a-19b1797b2ca1] Deleting contents of the VM from datastore datastore2 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 647.045990] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-4b216162-e1e7-4270-8d1f-482978b0681a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Deleting the datastore file [datastore2] 8493ba52-b950-4727-970a-19b1797b2ca1 {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 647.049637] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2054e971-322a-411e-969d-fd9aae1b181e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.053266] env[65726]: INFO nova.compute.manager [None req-8cc1dec8-17ac-4db3-bea9-0fb861f8b4b8 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] [instance: 04ffdf80-c54c-4587-9bb4-d520dc440501] Took 28.19 seconds to build instance. [ 647.066038] env[65726]: DEBUG oslo_vmware.api [None req-4b216162-e1e7-4270-8d1f-482978b0681a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Waiting for the task: (returnval){ [ 647.066038] env[65726]: value = "task-5115491" [ 647.066038] env[65726]: _type = "Task" [ 647.066038] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 647.077952] env[65726]: DEBUG oslo_vmware.api [None req-4b216162-e1e7-4270-8d1f-482978b0681a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Task: {'id': task-5115491, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 647.096806] env[65726]: DEBUG oslo_vmware.api [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Task: {'id': task-5115488, 'name': PowerOnVM_Task, 'duration_secs': 0.701106} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 647.097278] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] [instance: 811c1bbd-259c-4f5e-9a61-581bfba0ec4d] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 647.097606] env[65726]: INFO nova.compute.manager [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] [instance: 811c1bbd-259c-4f5e-9a61-581bfba0ec4d] Took 10.47 seconds to spawn the instance on the hypervisor. [ 647.097919] env[65726]: DEBUG nova.compute.manager [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] [instance: 811c1bbd-259c-4f5e-9a61-581bfba0ec4d] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 647.099387] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e94b0e3a-7764-4a76-92a8-e41e220de53a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.204506] env[65726]: DEBUG oslo_concurrency.lockutils [None req-06f67a27-e6e0-4078-804d-cff1137abcad tempest-ServersTestBootFromVolume-762458528 tempest-ServersTestBootFromVolume-762458528-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 647.525555] env[65726]: DEBUG nova.network.neutron [None req-32ab2821-307c-47ba-ac96-c38579de1e37 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] [instance: 0e064341-4e4a-407b-8c26-3eb04b409029] Port 9d3c5bc9-aa25-4013-b8dc-f565887d2736 binding to destination host cpu-1 is already ACTIVE {{(pid=65726) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3236}} [ 647.550437] env[65726]: DEBUG nova.scheduler.client.report [None req-881e6c53-2eb2-4fe8-9bf1-92c962897b3f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 647.557178] env[65726]: DEBUG oslo_concurrency.lockutils [None req-8cc1dec8-17ac-4db3-bea9-0fb861f8b4b8 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Lock "04ffdf80-c54c-4587-9bb4-d520dc440501" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 37.038s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 647.576630] env[65726]: DEBUG oslo_vmware.api [None req-4b216162-e1e7-4270-8d1f-482978b0681a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Task: {'id': task-5115491, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.286032} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 647.576913] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-4b216162-e1e7-4270-8d1f-482978b0681a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Deleted the datastore file {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 647.577105] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-4b216162-e1e7-4270-8d1f-482978b0681a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] [instance: 8493ba52-b950-4727-970a-19b1797b2ca1] Deleted contents of the VM from datastore datastore2 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 647.577277] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-4b216162-e1e7-4270-8d1f-482978b0681a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] [instance: 8493ba52-b950-4727-970a-19b1797b2ca1] Instance destroyed {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 647.577448] env[65726]: INFO nova.compute.manager [None req-4b216162-e1e7-4270-8d1f-482978b0681a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] [instance: 8493ba52-b950-4727-970a-19b1797b2ca1] Took 1.17 seconds to destroy the instance on the hypervisor. [ 647.577685] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-4b216162-e1e7-4270-8d1f-482978b0681a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 647.577877] env[65726]: DEBUG nova.compute.manager [-] [instance: 8493ba52-b950-4727-970a-19b1797b2ca1] Deallocating network for instance {{(pid=65726) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 647.577973] env[65726]: DEBUG nova.network.neutron [-] [instance: 8493ba52-b950-4727-970a-19b1797b2ca1] deallocate_for_instance() {{(pid=65726) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 647.578503] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 647.578807] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 647.621884] env[65726]: INFO nova.compute.manager [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] [instance: 811c1bbd-259c-4f5e-9a61-581bfba0ec4d] Took 28.41 seconds to build instance. [ 647.800037] env[65726]: WARNING openstack [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 647.801089] env[65726]: WARNING openstack [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 647.923664] env[65726]: DEBUG nova.network.neutron [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] [instance: f5151062-57b1-4e4e-93f4-aab0e7f504d4] Successfully updated port: 83750abb-1d26-4e23-a675-61ef785b858b {{(pid=65726) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 647.978256] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 647.978588] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 648.055325] env[65726]: DEBUG oslo_concurrency.lockutils [None req-881e6c53-2eb2-4fe8-9bf1-92c962897b3f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.782s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 648.055900] env[65726]: DEBUG nova.compute.manager [None req-881e6c53-2eb2-4fe8-9bf1-92c962897b3f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: f2d9090c-988f-43f4-9c81-7aa718a3438a] Start building networks asynchronously for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 648.058620] env[65726]: DEBUG oslo_concurrency.lockutils [None req-cc552ef7-8733-47e0-9cfd-32896dfc9132 tempest-TenantUsagesTestJSON-327211524 tempest-TenantUsagesTestJSON-327211524-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.886s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 648.060268] env[65726]: INFO nova.compute.claims [None req-cc552ef7-8733-47e0-9cfd-32896dfc9132 tempest-TenantUsagesTestJSON-327211524 tempest-TenantUsagesTestJSON-327211524-project-member] [instance: 7748c23f-9ea9-4172-8ab7-187678272bb3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 648.063324] env[65726]: DEBUG nova.compute.manager [None req-36828e27-b100-49cc-98c2-ad0b25ca6572 tempest-ServerGroupTestJSON-125182767 tempest-ServerGroupTestJSON-125182767-project-member] [instance: c2929727-a7ea-4f63-997d-474c29a305fc] Starting instance... {{(pid=65726) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 648.107878] env[65726]: WARNING openstack [req-a90995a5-d731-4249-85b8-09bdde3732a0 req-a1f2baad-ff22-43d1-8421-2b62b9ae2bb6 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 648.108441] env[65726]: WARNING openstack [req-a90995a5-d731-4249-85b8-09bdde3732a0 req-a1f2baad-ff22-43d1-8421-2b62b9ae2bb6 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 648.123126] env[65726]: WARNING openstack [req-0393c08a-a0f7-4006-9890-55aec2c8ba2a req-059d1391-338d-4ce7-8bf2-d1c94c047e32 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 648.123511] env[65726]: WARNING openstack [req-0393c08a-a0f7-4006-9890-55aec2c8ba2a req-059d1391-338d-4ce7-8bf2-d1c94c047e32 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 648.130160] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Lock "811c1bbd-259c-4f5e-9a61-581bfba0ec4d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 34.849s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 648.287788] env[65726]: DEBUG oslo_concurrency.lockutils [None req-64ef4520-b00a-42b8-937b-fbfc7c9b2298 tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Acquiring lock "21730277-cef6-4bfe-9f67-de71f6f615ff" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 648.290255] env[65726]: DEBUG oslo_concurrency.lockutils [None req-64ef4520-b00a-42b8-937b-fbfc7c9b2298 tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Lock "21730277-cef6-4bfe-9f67-de71f6f615ff" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 648.290255] env[65726]: DEBUG oslo_concurrency.lockutils [None req-64ef4520-b00a-42b8-937b-fbfc7c9b2298 tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Acquiring lock "21730277-cef6-4bfe-9f67-de71f6f615ff-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 648.290255] env[65726]: DEBUG oslo_concurrency.lockutils [None req-64ef4520-b00a-42b8-937b-fbfc7c9b2298 tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Lock "21730277-cef6-4bfe-9f67-de71f6f615ff-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 648.290255] env[65726]: DEBUG oslo_concurrency.lockutils [None req-64ef4520-b00a-42b8-937b-fbfc7c9b2298 tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Lock "21730277-cef6-4bfe-9f67-de71f6f615ff-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 648.292885] env[65726]: INFO nova.compute.manager [None req-64ef4520-b00a-42b8-937b-fbfc7c9b2298 tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] [instance: 21730277-cef6-4bfe-9f67-de71f6f615ff] Terminating instance [ 648.426887] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Acquiring lock "refresh_cache-f5151062-57b1-4e4e-93f4-aab0e7f504d4" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 648.426887] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Acquired lock "refresh_cache-f5151062-57b1-4e4e-93f4-aab0e7f504d4" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 648.427938] env[65726]: DEBUG nova.network.neutron [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] [instance: f5151062-57b1-4e4e-93f4-aab0e7f504d4] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 648.558740] env[65726]: DEBUG oslo_concurrency.lockutils [None req-32ab2821-307c-47ba-ac96-c38579de1e37 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Acquiring lock "0e064341-4e4a-407b-8c26-3eb04b409029-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 648.559118] env[65726]: DEBUG oslo_concurrency.lockutils [None req-32ab2821-307c-47ba-ac96-c38579de1e37 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Lock "0e064341-4e4a-407b-8c26-3eb04b409029-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 648.559235] env[65726]: DEBUG oslo_concurrency.lockutils [None req-32ab2821-307c-47ba-ac96-c38579de1e37 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Lock "0e064341-4e4a-407b-8c26-3eb04b409029-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 648.568218] env[65726]: DEBUG nova.compute.utils [None req-881e6c53-2eb2-4fe8-9bf1-92c962897b3f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Using /dev/sd instead of None {{(pid=65726) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 648.577890] env[65726]: DEBUG nova.compute.manager [None req-881e6c53-2eb2-4fe8-9bf1-92c962897b3f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: f2d9090c-988f-43f4-9c81-7aa718a3438a] Allocating IP information in the background. {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 648.578605] env[65726]: DEBUG nova.network.neutron [None req-881e6c53-2eb2-4fe8-9bf1-92c962897b3f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: f2d9090c-988f-43f4-9c81-7aa718a3438a] allocate_for_instance() {{(pid=65726) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 648.578966] env[65726]: WARNING neutronclient.v2_0.client [None req-881e6c53-2eb2-4fe8-9bf1-92c962897b3f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 648.579319] env[65726]: WARNING neutronclient.v2_0.client [None req-881e6c53-2eb2-4fe8-9bf1-92c962897b3f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 648.579909] env[65726]: WARNING openstack [None req-881e6c53-2eb2-4fe8-9bf1-92c962897b3f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 648.580299] env[65726]: WARNING openstack [None req-881e6c53-2eb2-4fe8-9bf1-92c962897b3f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 648.619649] env[65726]: DEBUG oslo_concurrency.lockutils [None req-36828e27-b100-49cc-98c2-ad0b25ca6572 tempest-ServerGroupTestJSON-125182767 tempest-ServerGroupTestJSON-125182767-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 648.636322] env[65726]: DEBUG nova.compute.manager [None req-fb0f14b3-fd50-4c8d-a984-877b235dcb32 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] [instance: d7905c4f-1ba3-43b3-b4b3-aebacd2f9cc3] Starting instance... {{(pid=65726) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 648.682025] env[65726]: WARNING openstack [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 648.682025] env[65726]: WARNING openstack [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 648.800626] env[65726]: DEBUG nova.compute.manager [None req-64ef4520-b00a-42b8-937b-fbfc7c9b2298 tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] [instance: 21730277-cef6-4bfe-9f67-de71f6f615ff] Start destroying the instance on the hypervisor. {{(pid=65726) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 648.800928] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-64ef4520-b00a-42b8-937b-fbfc7c9b2298 tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] [instance: 21730277-cef6-4bfe-9f67-de71f6f615ff] Destroying instance {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 648.802288] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffcba533-e81e-4083-b4eb-7e8c21f5266d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.814619] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-64ef4520-b00a-42b8-937b-fbfc7c9b2298 tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] [instance: 21730277-cef6-4bfe-9f67-de71f6f615ff] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 648.814619] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fb97a53e-b8a3-4197-89c0-342909567c2c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.823237] env[65726]: DEBUG oslo_vmware.api [None req-64ef4520-b00a-42b8-937b-fbfc7c9b2298 tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Waiting for the task: (returnval){ [ 648.823237] env[65726]: value = "task-5115492" [ 648.823237] env[65726]: _type = "Task" [ 648.823237] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 648.835382] env[65726]: DEBUG oslo_vmware.api [None req-64ef4520-b00a-42b8-937b-fbfc7c9b2298 tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Task: {'id': task-5115492, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 648.934438] env[65726]: WARNING openstack [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 648.935105] env[65726]: WARNING openstack [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 648.941761] env[65726]: DEBUG nova.network.neutron [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] [instance: f5151062-57b1-4e4e-93f4-aab0e7f504d4] Instance cache missing network info. {{(pid=65726) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 649.061362] env[65726]: DEBUG nova.network.neutron [-] [instance: 8493ba52-b950-4727-970a-19b1797b2ca1] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 649.083048] env[65726]: DEBUG nova.compute.manager [None req-881e6c53-2eb2-4fe8-9bf1-92c962897b3f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: f2d9090c-988f-43f4-9c81-7aa718a3438a] Start building block device mappings for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 649.161821] env[65726]: DEBUG nova.policy [None req-881e6c53-2eb2-4fe8-9bf1-92c962897b3f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4b6c9cc4ac3c487693240e375a83ea19', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b794c99309e94c1181fc9d8eeeb84702', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65726) authorize /opt/stack/nova/nova/policy.py:192}} [ 649.170249] env[65726]: DEBUG oslo_concurrency.lockutils [None req-fb0f14b3-fd50-4c8d-a984-877b235dcb32 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 649.307669] env[65726]: WARNING openstack [req-a90995a5-d731-4249-85b8-09bdde3732a0 req-a1f2baad-ff22-43d1-8421-2b62b9ae2bb6 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 649.310817] env[65726]: WARNING openstack [req-a90995a5-d731-4249-85b8-09bdde3732a0 req-a1f2baad-ff22-43d1-8421-2b62b9ae2bb6 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 649.340146] env[65726]: WARNING openstack [req-0393c08a-a0f7-4006-9890-55aec2c8ba2a req-059d1391-338d-4ce7-8bf2-d1c94c047e32 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 649.340730] env[65726]: WARNING openstack [req-0393c08a-a0f7-4006-9890-55aec2c8ba2a req-059d1391-338d-4ce7-8bf2-d1c94c047e32 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 649.369707] env[65726]: DEBUG oslo_vmware.api [None req-64ef4520-b00a-42b8-937b-fbfc7c9b2298 tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Task: {'id': task-5115492, 'name': PowerOffVM_Task, 'duration_secs': 0.283177} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 649.375604] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-64ef4520-b00a-42b8-937b-fbfc7c9b2298 tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] [instance: 21730277-cef6-4bfe-9f67-de71f6f615ff] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 649.375604] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-64ef4520-b00a-42b8-937b-fbfc7c9b2298 tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] [instance: 21730277-cef6-4bfe-9f67-de71f6f615ff] Unregistering the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 649.375604] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5141090d-9f1a-4e87-a57a-ad15f09d8449 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.456777] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-64ef4520-b00a-42b8-937b-fbfc7c9b2298 tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] [instance: 21730277-cef6-4bfe-9f67-de71f6f615ff] Unregistered the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 649.457060] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-64ef4520-b00a-42b8-937b-fbfc7c9b2298 tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] [instance: 21730277-cef6-4bfe-9f67-de71f6f615ff] Deleting contents of the VM from datastore datastore2 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 649.457191] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-64ef4520-b00a-42b8-937b-fbfc7c9b2298 tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Deleting the datastore file [datastore2] 21730277-cef6-4bfe-9f67-de71f6f615ff {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 649.457481] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1021064e-8ab8-496a-b168-cd6c7e51e338 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.466604] env[65726]: DEBUG oslo_vmware.api [None req-64ef4520-b00a-42b8-937b-fbfc7c9b2298 tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Waiting for the task: (returnval){ [ 649.466604] env[65726]: value = "task-5115494" [ 649.466604] env[65726]: _type = "Task" [ 649.466604] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 649.480696] env[65726]: DEBUG oslo_vmware.api [None req-64ef4520-b00a-42b8-937b-fbfc7c9b2298 tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Task: {'id': task-5115494, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 649.567701] env[65726]: WARNING openstack [None req-32ab2821-307c-47ba-ac96-c38579de1e37 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 649.568276] env[65726]: WARNING openstack [None req-32ab2821-307c-47ba-ac96-c38579de1e37 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 649.575269] env[65726]: INFO nova.compute.manager [-] [instance: 8493ba52-b950-4727-970a-19b1797b2ca1] Took 2.00 seconds to deallocate network for instance. [ 649.699324] env[65726]: DEBUG nova.network.neutron [None req-881e6c53-2eb2-4fe8-9bf1-92c962897b3f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: f2d9090c-988f-43f4-9c81-7aa718a3438a] Successfully created port: 7117cd21-ee2a-497d-b789-65e43b068258 {{(pid=65726) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 649.745746] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af7af273-48da-49dd-80a0-42427d959a8a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.754638] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26b46152-e6dd-406a-bcdb-88433972ef67 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.798615] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-831e62c3-238f-4341-b8b9-8f48b0a35c5d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.808521] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23c5f355-efd2-4de4-9ce7-e8291a23da64 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.815568] env[65726]: DEBUG nova.network.neutron [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] [instance: 3a60b719-9db7-436e-9908-25f6a233c465] Updating instance_info_cache with network_info: [{"id": "5c88141b-7089-49fb-afaa-a38f2cc47aaa", "address": "fa:16:3e:1f:a2:2e", "network": {"id": "b579021f-99ae-447a-b69d-293190930add", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1055463010-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4d508c846ae64e2c84ee281fe59af610", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5fb6e5b3-7be7-4018-8d9e-2133d926178c", "external-id": "nsx-vlan-transportzone-188", "segmentation_id": 188, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5c88141b-70", "ovs_interfaceid": "5c88141b-7089-49fb-afaa-a38f2cc47aaa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 649.832556] env[65726]: DEBUG nova.compute.provider_tree [None req-cc552ef7-8733-47e0-9cfd-32896dfc9132 tempest-TenantUsagesTestJSON-327211524 tempest-TenantUsagesTestJSON-327211524-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 649.836133] env[65726]: WARNING openstack [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 649.836586] env[65726]: WARNING openstack [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 649.913226] env[65726]: DEBUG oslo_concurrency.lockutils [None req-32ab2821-307c-47ba-ac96-c38579de1e37 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Acquiring lock "refresh_cache-0e064341-4e4a-407b-8c26-3eb04b409029" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 649.913226] env[65726]: DEBUG oslo_concurrency.lockutils [None req-32ab2821-307c-47ba-ac96-c38579de1e37 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Acquired lock "refresh_cache-0e064341-4e4a-407b-8c26-3eb04b409029" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 649.914256] env[65726]: DEBUG nova.network.neutron [None req-32ab2821-307c-47ba-ac96-c38579de1e37 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] [instance: 0e064341-4e4a-407b-8c26-3eb04b409029] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 649.981876] env[65726]: DEBUG oslo_vmware.api [None req-64ef4520-b00a-42b8-937b-fbfc7c9b2298 tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Task: {'id': task-5115494, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.259187} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 649.982153] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-64ef4520-b00a-42b8-937b-fbfc7c9b2298 tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Deleted the datastore file {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 649.982336] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-64ef4520-b00a-42b8-937b-fbfc7c9b2298 tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] [instance: 21730277-cef6-4bfe-9f67-de71f6f615ff] Deleted contents of the VM from datastore datastore2 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 649.982506] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-64ef4520-b00a-42b8-937b-fbfc7c9b2298 tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] [instance: 21730277-cef6-4bfe-9f67-de71f6f615ff] Instance destroyed {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 649.982785] env[65726]: INFO nova.compute.manager [None req-64ef4520-b00a-42b8-937b-fbfc7c9b2298 tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] [instance: 21730277-cef6-4bfe-9f67-de71f6f615ff] Took 1.18 seconds to destroy the instance on the hypervisor. [ 649.983043] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-64ef4520-b00a-42b8-937b-fbfc7c9b2298 tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 649.983909] env[65726]: DEBUG nova.compute.manager [-] [instance: 21730277-cef6-4bfe-9f67-de71f6f615ff] Deallocating network for instance {{(pid=65726) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 649.983909] env[65726]: DEBUG nova.network.neutron [-] [instance: 21730277-cef6-4bfe-9f67-de71f6f615ff] deallocate_for_instance() {{(pid=65726) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 649.984463] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 649.984800] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 650.087343] env[65726]: DEBUG oslo_concurrency.lockutils [None req-4b216162-e1e7-4270-8d1f-482978b0681a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 650.093800] env[65726]: DEBUG nova.compute.manager [None req-881e6c53-2eb2-4fe8-9bf1-92c962897b3f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: f2d9090c-988f-43f4-9c81-7aa718a3438a] Start spawning the instance on the hypervisor. {{(pid=65726) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 650.122929] env[65726]: DEBUG nova.virt.hardware [None req-881e6c53-2eb2-4fe8-9bf1-92c962897b3f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T19:28:34Z,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1539f06fa3534e90acbc3a60c4b8bd3f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T19:28:35Z,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 650.123180] env[65726]: DEBUG nova.virt.hardware [None req-881e6c53-2eb2-4fe8-9bf1-92c962897b3f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 650.123352] env[65726]: DEBUG nova.virt.hardware [None req-881e6c53-2eb2-4fe8-9bf1-92c962897b3f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 650.123555] env[65726]: DEBUG nova.virt.hardware [None req-881e6c53-2eb2-4fe8-9bf1-92c962897b3f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 650.123715] env[65726]: DEBUG nova.virt.hardware [None req-881e6c53-2eb2-4fe8-9bf1-92c962897b3f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 650.123877] env[65726]: DEBUG nova.virt.hardware [None req-881e6c53-2eb2-4fe8-9bf1-92c962897b3f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 650.124128] env[65726]: DEBUG nova.virt.hardware [None req-881e6c53-2eb2-4fe8-9bf1-92c962897b3f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 650.124305] env[65726]: DEBUG nova.virt.hardware [None req-881e6c53-2eb2-4fe8-9bf1-92c962897b3f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 650.124492] env[65726]: DEBUG nova.virt.hardware [None req-881e6c53-2eb2-4fe8-9bf1-92c962897b3f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 650.124744] env[65726]: DEBUG nova.virt.hardware [None req-881e6c53-2eb2-4fe8-9bf1-92c962897b3f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 650.124883] env[65726]: DEBUG nova.virt.hardware [None req-881e6c53-2eb2-4fe8-9bf1-92c962897b3f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 650.125881] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31ae28e6-2ac8-4797-854d-df79ec33af68 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.135603] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70720c52-af1f-4bcf-93ce-dc62407e7f0e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.215619] env[65726]: DEBUG nova.network.neutron [req-a90995a5-d731-4249-85b8-09bdde3732a0 req-a1f2baad-ff22-43d1-8421-2b62b9ae2bb6 service nova] [instance: de5cad99-1dbc-4435-98b4-987e24ba8c3c] Updated VIF entry in instance network info cache for port 3f320668-6397-427f-b8c4-3e63282fcb87. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 650.216033] env[65726]: DEBUG nova.network.neutron [req-a90995a5-d731-4249-85b8-09bdde3732a0 req-a1f2baad-ff22-43d1-8421-2b62b9ae2bb6 service nova] [instance: de5cad99-1dbc-4435-98b4-987e24ba8c3c] Updating instance_info_cache with network_info: [{"id": "3f320668-6397-427f-b8c4-3e63282fcb87", "address": "fa:16:3e:ca:a0:6d", "network": {"id": "55d629b8-a541-4720-9289-927bbc6e6f22", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.83", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "1539f06fa3534e90acbc3a60c4b8bd3f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "35342bcb-8b06-472e-b3c0-43fd3d6c4b30", "external-id": "nsx-vlan-transportzone-524", "segmentation_id": 524, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3f320668-63", "ovs_interfaceid": "3f320668-6397-427f-b8c4-3e63282fcb87", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 650.256195] env[65726]: DEBUG nova.network.neutron [req-0393c08a-a0f7-4006-9890-55aec2c8ba2a req-059d1391-338d-4ce7-8bf2-d1c94c047e32 service nova] [instance: 811c1bbd-259c-4f5e-9a61-581bfba0ec4d] Updated VIF entry in instance network info cache for port 205b7b13-8d99-402c-9587-c84362a7af94. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 650.257227] env[65726]: DEBUG nova.network.neutron [req-0393c08a-a0f7-4006-9890-55aec2c8ba2a req-059d1391-338d-4ce7-8bf2-d1c94c047e32 service nova] [instance: 811c1bbd-259c-4f5e-9a61-581bfba0ec4d] Updating instance_info_cache with network_info: [{"id": "205b7b13-8d99-402c-9587-c84362a7af94", "address": "fa:16:3e:fd:96:82", "network": {"id": "b579021f-99ae-447a-b69d-293190930add", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1055463010-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4d508c846ae64e2c84ee281fe59af610", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5fb6e5b3-7be7-4018-8d9e-2133d926178c", "external-id": "nsx-vlan-transportzone-188", "segmentation_id": 188, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap205b7b13-8d", "ovs_interfaceid": "205b7b13-8d99-402c-9587-c84362a7af94", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 650.319236] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Releasing lock "refresh_cache-3a60b719-9db7-436e-9908-25f6a233c465" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 650.319671] env[65726]: DEBUG nova.compute.manager [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] [instance: 3a60b719-9db7-436e-9908-25f6a233c465] Instance network_info: |[{"id": "5c88141b-7089-49fb-afaa-a38f2cc47aaa", "address": "fa:16:3e:1f:a2:2e", "network": {"id": "b579021f-99ae-447a-b69d-293190930add", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1055463010-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4d508c846ae64e2c84ee281fe59af610", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5fb6e5b3-7be7-4018-8d9e-2133d926178c", "external-id": "nsx-vlan-transportzone-188", "segmentation_id": 188, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5c88141b-70", "ovs_interfaceid": "5c88141b-7089-49fb-afaa-a38f2cc47aaa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 650.320204] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] [instance: 3a60b719-9db7-436e-9908-25f6a233c465] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:1f:a2:2e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '5fb6e5b3-7be7-4018-8d9e-2133d926178c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '5c88141b-7089-49fb-afaa-a38f2cc47aaa', 'vif_model': 'vmxnet3'}] {{(pid=65726) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 650.328709] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 650.328947] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3a60b719-9db7-436e-9908-25f6a233c465] Creating VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 650.329212] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ee0ebc7f-0d92-45d2-b590-d6381cd3f33f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.348766] env[65726]: DEBUG nova.scheduler.client.report [None req-cc552ef7-8733-47e0-9cfd-32896dfc9132 tempest-TenantUsagesTestJSON-327211524 tempest-TenantUsagesTestJSON-327211524-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 650.360348] env[65726]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 650.360348] env[65726]: value = "task-5115495" [ 650.360348] env[65726]: _type = "Task" [ 650.360348] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 650.369656] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115495, 'name': CreateVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 650.383515] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 650.384047] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 650.410663] env[65726]: WARNING openstack [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 650.411327] env[65726]: WARNING openstack [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 650.419278] env[65726]: WARNING openstack [None req-32ab2821-307c-47ba-ac96-c38579de1e37 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 650.419648] env[65726]: WARNING openstack [None req-32ab2821-307c-47ba-ac96-c38579de1e37 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 650.720670] env[65726]: DEBUG oslo_concurrency.lockutils [req-a90995a5-d731-4249-85b8-09bdde3732a0 req-a1f2baad-ff22-43d1-8421-2b62b9ae2bb6 service nova] Releasing lock "refresh_cache-de5cad99-1dbc-4435-98b4-987e24ba8c3c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 650.720670] env[65726]: DEBUG nova.compute.manager [req-a90995a5-d731-4249-85b8-09bdde3732a0 req-a1f2baad-ff22-43d1-8421-2b62b9ae2bb6 service nova] [instance: 04ffdf80-c54c-4587-9bb4-d520dc440501] Received event network-vif-plugged-ece7a121-1068-4cb7-b500-9b6a1e8c4a1c {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 650.720670] env[65726]: DEBUG oslo_concurrency.lockutils [req-a90995a5-d731-4249-85b8-09bdde3732a0 req-a1f2baad-ff22-43d1-8421-2b62b9ae2bb6 service nova] Acquiring lock "04ffdf80-c54c-4587-9bb4-d520dc440501-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 650.721120] env[65726]: DEBUG oslo_concurrency.lockutils [req-a90995a5-d731-4249-85b8-09bdde3732a0 req-a1f2baad-ff22-43d1-8421-2b62b9ae2bb6 service nova] Lock "04ffdf80-c54c-4587-9bb4-d520dc440501-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 650.721120] env[65726]: DEBUG oslo_concurrency.lockutils [req-a90995a5-d731-4249-85b8-09bdde3732a0 req-a1f2baad-ff22-43d1-8421-2b62b9ae2bb6 service nova] Lock "04ffdf80-c54c-4587-9bb4-d520dc440501-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 650.721409] env[65726]: DEBUG nova.compute.manager [req-a90995a5-d731-4249-85b8-09bdde3732a0 req-a1f2baad-ff22-43d1-8421-2b62b9ae2bb6 service nova] [instance: 04ffdf80-c54c-4587-9bb4-d520dc440501] No waiting events found dispatching network-vif-plugged-ece7a121-1068-4cb7-b500-9b6a1e8c4a1c {{(pid=65726) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 650.721409] env[65726]: WARNING nova.compute.manager [req-a90995a5-d731-4249-85b8-09bdde3732a0 req-a1f2baad-ff22-43d1-8421-2b62b9ae2bb6 service nova] [instance: 04ffdf80-c54c-4587-9bb4-d520dc440501] Received unexpected event network-vif-plugged-ece7a121-1068-4cb7-b500-9b6a1e8c4a1c for instance with vm_state building and task_state spawning. [ 650.721529] env[65726]: DEBUG nova.compute.manager [req-a90995a5-d731-4249-85b8-09bdde3732a0 req-a1f2baad-ff22-43d1-8421-2b62b9ae2bb6 service nova] [instance: 04ffdf80-c54c-4587-9bb4-d520dc440501] Received event network-changed-ece7a121-1068-4cb7-b500-9b6a1e8c4a1c {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 650.721677] env[65726]: DEBUG nova.compute.manager [req-a90995a5-d731-4249-85b8-09bdde3732a0 req-a1f2baad-ff22-43d1-8421-2b62b9ae2bb6 service nova] [instance: 04ffdf80-c54c-4587-9bb4-d520dc440501] Refreshing instance network info cache due to event network-changed-ece7a121-1068-4cb7-b500-9b6a1e8c4a1c. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 650.721854] env[65726]: DEBUG oslo_concurrency.lockutils [req-a90995a5-d731-4249-85b8-09bdde3732a0 req-a1f2baad-ff22-43d1-8421-2b62b9ae2bb6 service nova] Acquiring lock "refresh_cache-04ffdf80-c54c-4587-9bb4-d520dc440501" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 650.721984] env[65726]: DEBUG oslo_concurrency.lockutils [req-a90995a5-d731-4249-85b8-09bdde3732a0 req-a1f2baad-ff22-43d1-8421-2b62b9ae2bb6 service nova] Acquired lock "refresh_cache-04ffdf80-c54c-4587-9bb4-d520dc440501" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 650.722148] env[65726]: DEBUG nova.network.neutron [req-a90995a5-d731-4249-85b8-09bdde3732a0 req-a1f2baad-ff22-43d1-8421-2b62b9ae2bb6 service nova] [instance: 04ffdf80-c54c-4587-9bb4-d520dc440501] Refreshing network info cache for port ece7a121-1068-4cb7-b500-9b6a1e8c4a1c {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 650.760325] env[65726]: DEBUG oslo_concurrency.lockutils [req-0393c08a-a0f7-4006-9890-55aec2c8ba2a req-059d1391-338d-4ce7-8bf2-d1c94c047e32 service nova] Releasing lock "refresh_cache-811c1bbd-259c-4f5e-9a61-581bfba0ec4d" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 650.853805] env[65726]: DEBUG oslo_concurrency.lockutils [None req-cc552ef7-8733-47e0-9cfd-32896dfc9132 tempest-TenantUsagesTestJSON-327211524 tempest-TenantUsagesTestJSON-327211524-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.795s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 650.854424] env[65726]: DEBUG nova.compute.manager [None req-cc552ef7-8733-47e0-9cfd-32896dfc9132 tempest-TenantUsagesTestJSON-327211524 tempest-TenantUsagesTestJSON-327211524-project-member] [instance: 7748c23f-9ea9-4172-8ab7-187678272bb3] Start building networks asynchronously for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 650.857874] env[65726]: DEBUG oslo_concurrency.lockutils [None req-40525054-994c-41ab-bed7-c877834cae18 tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.114s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 650.862456] env[65726]: INFO nova.compute.claims [None req-40525054-994c-41ab-bed7-c877834cae18 tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] [instance: 3b933684-ac19-44b0-a49d-6af45501e166] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 650.872949] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115495, 'name': CreateVM_Task} progress is 25%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 650.962725] env[65726]: DEBUG nova.network.neutron [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] [instance: f5151062-57b1-4e4e-93f4-aab0e7f504d4] Updating instance_info_cache with network_info: [{"id": "83750abb-1d26-4e23-a675-61ef785b858b", "address": "fa:16:3e:42:77:9f", "network": {"id": "b579021f-99ae-447a-b69d-293190930add", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1055463010-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4d508c846ae64e2c84ee281fe59af610", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5fb6e5b3-7be7-4018-8d9e-2133d926178c", "external-id": "nsx-vlan-transportzone-188", "segmentation_id": 188, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap83750abb-1d", "ovs_interfaceid": "83750abb-1d26-4e23-a675-61ef785b858b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 651.018295] env[65726]: WARNING openstack [None req-32ab2821-307c-47ba-ac96-c38579de1e37 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 651.018961] env[65726]: WARNING openstack [None req-32ab2821-307c-47ba-ac96-c38579de1e37 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 651.135601] env[65726]: WARNING openstack [None req-32ab2821-307c-47ba-ac96-c38579de1e37 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 651.136119] env[65726]: WARNING openstack [None req-32ab2821-307c-47ba-ac96-c38579de1e37 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 651.155390] env[65726]: DEBUG nova.network.neutron [-] [instance: 21730277-cef6-4bfe-9f67-de71f6f615ff] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 651.225592] env[65726]: WARNING openstack [req-a90995a5-d731-4249-85b8-09bdde3732a0 req-a1f2baad-ff22-43d1-8421-2b62b9ae2bb6 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 651.226103] env[65726]: WARNING openstack [req-a90995a5-d731-4249-85b8-09bdde3732a0 req-a1f2baad-ff22-43d1-8421-2b62b9ae2bb6 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 651.247397] env[65726]: DEBUG nova.compute.manager [req-3637bfed-a5e8-48d7-8845-d70917530ab6 req-c5ac2a4e-436a-47b6-811c-41a214deb69f service nova] [instance: 3a60b719-9db7-436e-9908-25f6a233c465] Received event network-vif-plugged-5c88141b-7089-49fb-afaa-a38f2cc47aaa {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 651.247636] env[65726]: DEBUG oslo_concurrency.lockutils [req-3637bfed-a5e8-48d7-8845-d70917530ab6 req-c5ac2a4e-436a-47b6-811c-41a214deb69f service nova] Acquiring lock "3a60b719-9db7-436e-9908-25f6a233c465-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 651.247870] env[65726]: DEBUG oslo_concurrency.lockutils [req-3637bfed-a5e8-48d7-8845-d70917530ab6 req-c5ac2a4e-436a-47b6-811c-41a214deb69f service nova] Lock "3a60b719-9db7-436e-9908-25f6a233c465-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 651.248053] env[65726]: DEBUG oslo_concurrency.lockutils [req-3637bfed-a5e8-48d7-8845-d70917530ab6 req-c5ac2a4e-436a-47b6-811c-41a214deb69f service nova] Lock "3a60b719-9db7-436e-9908-25f6a233c465-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 651.248632] env[65726]: DEBUG nova.compute.manager [req-3637bfed-a5e8-48d7-8845-d70917530ab6 req-c5ac2a4e-436a-47b6-811c-41a214deb69f service nova] [instance: 3a60b719-9db7-436e-9908-25f6a233c465] No waiting events found dispatching network-vif-plugged-5c88141b-7089-49fb-afaa-a38f2cc47aaa {{(pid=65726) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 651.248632] env[65726]: WARNING nova.compute.manager [req-3637bfed-a5e8-48d7-8845-d70917530ab6 req-c5ac2a4e-436a-47b6-811c-41a214deb69f service nova] [instance: 3a60b719-9db7-436e-9908-25f6a233c465] Received unexpected event network-vif-plugged-5c88141b-7089-49fb-afaa-a38f2cc47aaa for instance with vm_state building and task_state spawning. [ 651.248632] env[65726]: DEBUG nova.compute.manager [req-3637bfed-a5e8-48d7-8845-d70917530ab6 req-c5ac2a4e-436a-47b6-811c-41a214deb69f service nova] [instance: 3a60b719-9db7-436e-9908-25f6a233c465] Received event network-changed-5c88141b-7089-49fb-afaa-a38f2cc47aaa {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 651.248856] env[65726]: DEBUG nova.compute.manager [req-3637bfed-a5e8-48d7-8845-d70917530ab6 req-c5ac2a4e-436a-47b6-811c-41a214deb69f service nova] [instance: 3a60b719-9db7-436e-9908-25f6a233c465] Refreshing instance network info cache due to event network-changed-5c88141b-7089-49fb-afaa-a38f2cc47aaa. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 651.248896] env[65726]: DEBUG oslo_concurrency.lockutils [req-3637bfed-a5e8-48d7-8845-d70917530ab6 req-c5ac2a4e-436a-47b6-811c-41a214deb69f service nova] Acquiring lock "refresh_cache-3a60b719-9db7-436e-9908-25f6a233c465" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 651.249090] env[65726]: DEBUG oslo_concurrency.lockutils [req-3637bfed-a5e8-48d7-8845-d70917530ab6 req-c5ac2a4e-436a-47b6-811c-41a214deb69f service nova] Acquired lock "refresh_cache-3a60b719-9db7-436e-9908-25f6a233c465" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 651.249187] env[65726]: DEBUG nova.network.neutron [req-3637bfed-a5e8-48d7-8845-d70917530ab6 req-c5ac2a4e-436a-47b6-811c-41a214deb69f service nova] [instance: 3a60b719-9db7-436e-9908-25f6a233c465] Refreshing network info cache for port 5c88141b-7089-49fb-afaa-a38f2cc47aaa {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 651.267282] env[65726]: DEBUG nova.compute.manager [req-58cf80ad-4388-4990-9b3f-6f7d2c8dd658 req-142cbe85-245d-473f-9481-cb3143bca571 service nova] [instance: f5151062-57b1-4e4e-93f4-aab0e7f504d4] Received event network-vif-plugged-83750abb-1d26-4e23-a675-61ef785b858b {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 651.267532] env[65726]: DEBUG oslo_concurrency.lockutils [req-58cf80ad-4388-4990-9b3f-6f7d2c8dd658 req-142cbe85-245d-473f-9481-cb3143bca571 service nova] Acquiring lock "f5151062-57b1-4e4e-93f4-aab0e7f504d4-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 651.267665] env[65726]: DEBUG oslo_concurrency.lockutils [req-58cf80ad-4388-4990-9b3f-6f7d2c8dd658 req-142cbe85-245d-473f-9481-cb3143bca571 service nova] Lock "f5151062-57b1-4e4e-93f4-aab0e7f504d4-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 651.267872] env[65726]: DEBUG oslo_concurrency.lockutils [req-58cf80ad-4388-4990-9b3f-6f7d2c8dd658 req-142cbe85-245d-473f-9481-cb3143bca571 service nova] Lock "f5151062-57b1-4e4e-93f4-aab0e7f504d4-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 651.267971] env[65726]: DEBUG nova.compute.manager [req-58cf80ad-4388-4990-9b3f-6f7d2c8dd658 req-142cbe85-245d-473f-9481-cb3143bca571 service nova] [instance: f5151062-57b1-4e4e-93f4-aab0e7f504d4] No waiting events found dispatching network-vif-plugged-83750abb-1d26-4e23-a675-61ef785b858b {{(pid=65726) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 651.268142] env[65726]: WARNING nova.compute.manager [req-58cf80ad-4388-4990-9b3f-6f7d2c8dd658 req-142cbe85-245d-473f-9481-cb3143bca571 service nova] [instance: f5151062-57b1-4e4e-93f4-aab0e7f504d4] Received unexpected event network-vif-plugged-83750abb-1d26-4e23-a675-61ef785b858b for instance with vm_state building and task_state spawning. [ 651.268287] env[65726]: DEBUG nova.compute.manager [req-58cf80ad-4388-4990-9b3f-6f7d2c8dd658 req-142cbe85-245d-473f-9481-cb3143bca571 service nova] [instance: f5151062-57b1-4e4e-93f4-aab0e7f504d4] Received event network-changed-83750abb-1d26-4e23-a675-61ef785b858b {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 651.268430] env[65726]: DEBUG nova.compute.manager [req-58cf80ad-4388-4990-9b3f-6f7d2c8dd658 req-142cbe85-245d-473f-9481-cb3143bca571 service nova] [instance: f5151062-57b1-4e4e-93f4-aab0e7f504d4] Refreshing instance network info cache due to event network-changed-83750abb-1d26-4e23-a675-61ef785b858b. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 651.268634] env[65726]: DEBUG oslo_concurrency.lockutils [req-58cf80ad-4388-4990-9b3f-6f7d2c8dd658 req-142cbe85-245d-473f-9481-cb3143bca571 service nova] Acquiring lock "refresh_cache-f5151062-57b1-4e4e-93f4-aab0e7f504d4" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 651.290552] env[65726]: DEBUG nova.network.neutron [None req-32ab2821-307c-47ba-ac96-c38579de1e37 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] [instance: 0e064341-4e4a-407b-8c26-3eb04b409029] Updating instance_info_cache with network_info: [{"id": "9d3c5bc9-aa25-4013-b8dc-f565887d2736", "address": "fa:16:3e:f9:95:de", "network": {"id": "55d629b8-a541-4720-9289-927bbc6e6f22", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.93", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "1539f06fa3534e90acbc3a60c4b8bd3f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "35342bcb-8b06-472e-b3c0-43fd3d6c4b30", "external-id": "nsx-vlan-transportzone-524", "segmentation_id": 524, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9d3c5bc9-aa", "ovs_interfaceid": "9d3c5bc9-aa25-4013-b8dc-f565887d2736", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 651.359677] env[65726]: DEBUG nova.compute.utils [None req-cc552ef7-8733-47e0-9cfd-32896dfc9132 tempest-TenantUsagesTestJSON-327211524 tempest-TenantUsagesTestJSON-327211524-project-member] Using /dev/sd instead of None {{(pid=65726) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 651.361158] env[65726]: DEBUG nova.compute.manager [None req-cc552ef7-8733-47e0-9cfd-32896dfc9132 tempest-TenantUsagesTestJSON-327211524 tempest-TenantUsagesTestJSON-327211524-project-member] [instance: 7748c23f-9ea9-4172-8ab7-187678272bb3] Allocating IP information in the background. {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 651.365043] env[65726]: DEBUG nova.network.neutron [None req-cc552ef7-8733-47e0-9cfd-32896dfc9132 tempest-TenantUsagesTestJSON-327211524 tempest-TenantUsagesTestJSON-327211524-project-member] [instance: 7748c23f-9ea9-4172-8ab7-187678272bb3] allocate_for_instance() {{(pid=65726) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 651.365043] env[65726]: WARNING neutronclient.v2_0.client [None req-cc552ef7-8733-47e0-9cfd-32896dfc9132 tempest-TenantUsagesTestJSON-327211524 tempest-TenantUsagesTestJSON-327211524-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 651.365043] env[65726]: WARNING neutronclient.v2_0.client [None req-cc552ef7-8733-47e0-9cfd-32896dfc9132 tempest-TenantUsagesTestJSON-327211524 tempest-TenantUsagesTestJSON-327211524-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 651.365043] env[65726]: WARNING openstack [None req-cc552ef7-8733-47e0-9cfd-32896dfc9132 tempest-TenantUsagesTestJSON-327211524 tempest-TenantUsagesTestJSON-327211524-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 651.365043] env[65726]: WARNING openstack [None req-cc552ef7-8733-47e0-9cfd-32896dfc9132 tempest-TenantUsagesTestJSON-327211524 tempest-TenantUsagesTestJSON-327211524-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 651.386270] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115495, 'name': CreateVM_Task, 'duration_secs': 0.6985} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 651.386945] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3a60b719-9db7-436e-9908-25f6a233c465] Created VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 651.387377] env[65726]: WARNING openstack [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 651.387696] env[65726]: WARNING openstack [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 651.393170] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 651.393170] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 651.394180] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 651.394453] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5be67a13-848b-41bc-a5da-782574ee895a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.404200] env[65726]: DEBUG oslo_vmware.api [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Waiting for the task: (returnval){ [ 651.404200] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52eadf8a-d033-b6b4-d101-f9ab8b17ca5e" [ 651.404200] env[65726]: _type = "Task" [ 651.404200] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 651.417031] env[65726]: DEBUG oslo_vmware.api [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52eadf8a-d033-b6b4-d101-f9ab8b17ca5e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 651.444813] env[65726]: WARNING openstack [req-a90995a5-d731-4249-85b8-09bdde3732a0 req-a1f2baad-ff22-43d1-8421-2b62b9ae2bb6 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 651.446019] env[65726]: WARNING openstack [req-a90995a5-d731-4249-85b8-09bdde3732a0 req-a1f2baad-ff22-43d1-8421-2b62b9ae2bb6 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 651.465109] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Releasing lock "refresh_cache-f5151062-57b1-4e4e-93f4-aab0e7f504d4" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 651.465674] env[65726]: DEBUG nova.compute.manager [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] [instance: f5151062-57b1-4e4e-93f4-aab0e7f504d4] Instance network_info: |[{"id": "83750abb-1d26-4e23-a675-61ef785b858b", "address": "fa:16:3e:42:77:9f", "network": {"id": "b579021f-99ae-447a-b69d-293190930add", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1055463010-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4d508c846ae64e2c84ee281fe59af610", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5fb6e5b3-7be7-4018-8d9e-2133d926178c", "external-id": "nsx-vlan-transportzone-188", "segmentation_id": 188, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap83750abb-1d", "ovs_interfaceid": "83750abb-1d26-4e23-a675-61ef785b858b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 651.465852] env[65726]: DEBUG oslo_concurrency.lockutils [req-58cf80ad-4388-4990-9b3f-6f7d2c8dd658 req-142cbe85-245d-473f-9481-cb3143bca571 service nova] Acquired lock "refresh_cache-f5151062-57b1-4e4e-93f4-aab0e7f504d4" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 651.465924] env[65726]: DEBUG nova.network.neutron [req-58cf80ad-4388-4990-9b3f-6f7d2c8dd658 req-142cbe85-245d-473f-9481-cb3143bca571 service nova] [instance: f5151062-57b1-4e4e-93f4-aab0e7f504d4] Refreshing network info cache for port 83750abb-1d26-4e23-a675-61ef785b858b {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 651.467467] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] [instance: f5151062-57b1-4e4e-93f4-aab0e7f504d4] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:42:77:9f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '5fb6e5b3-7be7-4018-8d9e-2133d926178c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '83750abb-1d26-4e23-a675-61ef785b858b', 'vif_model': 'vmxnet3'}] {{(pid=65726) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 651.479069] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 651.480235] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f5151062-57b1-4e4e-93f4-aab0e7f504d4] Creating VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 651.480570] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b74c588d-4165-4d1d-b7f0-10d4519ed3db {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.498855] env[65726]: DEBUG nova.policy [None req-cc552ef7-8733-47e0-9cfd-32896dfc9132 tempest-TenantUsagesTestJSON-327211524 tempest-TenantUsagesTestJSON-327211524-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5212b8de543d494c87c78c6984a7f534', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a2b0d29fd55f4116acb5331831eb23e9', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65726) authorize /opt/stack/nova/nova/policy.py:192}} [ 651.507664] env[65726]: DEBUG nova.network.neutron [None req-881e6c53-2eb2-4fe8-9bf1-92c962897b3f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: f2d9090c-988f-43f4-9c81-7aa718a3438a] Successfully updated port: 7117cd21-ee2a-497d-b789-65e43b068258 {{(pid=65726) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 651.512203] env[65726]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 651.512203] env[65726]: value = "task-5115496" [ 651.512203] env[65726]: _type = "Task" [ 651.512203] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 651.523779] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115496, 'name': CreateVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 651.576517] env[65726]: WARNING openstack [req-a90995a5-d731-4249-85b8-09bdde3732a0 req-a1f2baad-ff22-43d1-8421-2b62b9ae2bb6 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 651.577241] env[65726]: WARNING openstack [req-a90995a5-d731-4249-85b8-09bdde3732a0 req-a1f2baad-ff22-43d1-8421-2b62b9ae2bb6 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 651.658801] env[65726]: INFO nova.compute.manager [-] [instance: 21730277-cef6-4bfe-9f67-de71f6f615ff] Took 1.68 seconds to deallocate network for instance. [ 651.743873] env[65726]: DEBUG nova.network.neutron [req-a90995a5-d731-4249-85b8-09bdde3732a0 req-a1f2baad-ff22-43d1-8421-2b62b9ae2bb6 service nova] [instance: 04ffdf80-c54c-4587-9bb4-d520dc440501] Updated VIF entry in instance network info cache for port ece7a121-1068-4cb7-b500-9b6a1e8c4a1c. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 651.744363] env[65726]: DEBUG nova.network.neutron [req-a90995a5-d731-4249-85b8-09bdde3732a0 req-a1f2baad-ff22-43d1-8421-2b62b9ae2bb6 service nova] [instance: 04ffdf80-c54c-4587-9bb4-d520dc440501] Updating instance_info_cache with network_info: [{"id": "ece7a121-1068-4cb7-b500-9b6a1e8c4a1c", "address": "fa:16:3e:7a:97:f3", "network": {"id": "f639fd12-fb47-4b3f-9895-6facebc89726", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-2112264538-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.3"}}], "meta": {"injected": false, "tenant_id": "3a680eae0e734ea8b2274b504d0a46d8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "257e5ea7-8b80-4301-9900-a754f1fe2031", "external-id": "nsx-vlan-transportzone-682", "segmentation_id": 682, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapece7a121-10", "ovs_interfaceid": "ece7a121-1068-4cb7-b500-9b6a1e8c4a1c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 651.752973] env[65726]: WARNING openstack [req-3637bfed-a5e8-48d7-8845-d70917530ab6 req-c5ac2a4e-436a-47b6-811c-41a214deb69f service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 651.753500] env[65726]: WARNING openstack [req-3637bfed-a5e8-48d7-8845-d70917530ab6 req-c5ac2a4e-436a-47b6-811c-41a214deb69f service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 651.794193] env[65726]: DEBUG oslo_concurrency.lockutils [None req-32ab2821-307c-47ba-ac96-c38579de1e37 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Releasing lock "refresh_cache-0e064341-4e4a-407b-8c26-3eb04b409029" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 651.875942] env[65726]: DEBUG nova.compute.manager [None req-cc552ef7-8733-47e0-9cfd-32896dfc9132 tempest-TenantUsagesTestJSON-327211524 tempest-TenantUsagesTestJSON-327211524-project-member] [instance: 7748c23f-9ea9-4172-8ab7-187678272bb3] Start building block device mappings for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 651.918065] env[65726]: DEBUG oslo_vmware.api [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52eadf8a-d033-b6b4-d101-f9ab8b17ca5e, 'name': SearchDatastore_Task, 'duration_secs': 0.01485} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 651.920929] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 651.921707] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] [instance: 3a60b719-9db7-436e-9908-25f6a233c465] Processing image b52362a3-ee8a-4cbf-b06f-513b0cc8f95c {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 651.922116] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 651.924022] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 651.924022] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 651.927274] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-717e2792-73d4-4731-a0f0-b36f9fcade18 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.937248] env[65726]: DEBUG oslo_vmware.rw_handles [None req-4240f7e0-4ab1-4718-b82f-a2d7b6caacae tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52a00c2f-2a99-27a5-f5a2-5fcf9b09ffeb/disk-0.vmdk. {{(pid=65726) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 651.939399] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e518c73b-b745-47c2-a0ea-1860bc1117b5 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.945132] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 651.945689] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65726) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 651.947108] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-27784902-c274-4714-a18c-ff567bc2c722 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.953418] env[65726]: DEBUG nova.compute.manager [None req-f9784253-3c11-4245-a226-656651a2901a tempest-ServerDiagnosticsTest-578350078 tempest-ServerDiagnosticsTest-578350078-project-admin] [instance: de5cad99-1dbc-4435-98b4-987e24ba8c3c] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 651.954068] env[65726]: DEBUG oslo_vmware.rw_handles [None req-4240f7e0-4ab1-4718-b82f-a2d7b6caacae tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52a00c2f-2a99-27a5-f5a2-5fcf9b09ffeb/disk-0.vmdk is in state: ready. {{(pid=65726) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 651.954242] env[65726]: ERROR oslo_vmware.rw_handles [None req-4240f7e0-4ab1-4718-b82f-a2d7b6caacae tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52a00c2f-2a99-27a5-f5a2-5fcf9b09ffeb/disk-0.vmdk due to incomplete transfer. [ 651.957104] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59fe9d9a-b7f8-4bcf-9f3e-885c789738f0 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.962316] env[65726]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-2b08db38-dd55-4d19-a9c8-2052fe6fe19f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.965816] env[65726]: DEBUG nova.network.neutron [None req-cc552ef7-8733-47e0-9cfd-32896dfc9132 tempest-TenantUsagesTestJSON-327211524 tempest-TenantUsagesTestJSON-327211524-project-member] [instance: 7748c23f-9ea9-4172-8ab7-187678272bb3] Successfully created port: 8bb2d672-e56e-4f01-bfbf-74a279d9da7d {{(pid=65726) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 651.975000] env[65726]: DEBUG oslo_vmware.api [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Waiting for the task: (returnval){ [ 651.975000] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]522d04ea-cea6-da23-0db2-ea78e59ef11a" [ 651.975000] env[65726]: _type = "Task" [ 651.975000] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 651.980639] env[65726]: INFO nova.compute.manager [None req-f9784253-3c11-4245-a226-656651a2901a tempest-ServerDiagnosticsTest-578350078 tempest-ServerDiagnosticsTest-578350078-project-admin] [instance: de5cad99-1dbc-4435-98b4-987e24ba8c3c] Retrieving diagnostics [ 651.983353] env[65726]: WARNING openstack [req-3637bfed-a5e8-48d7-8845-d70917530ab6 req-c5ac2a4e-436a-47b6-811c-41a214deb69f service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 651.984156] env[65726]: WARNING openstack [req-3637bfed-a5e8-48d7-8845-d70917530ab6 req-c5ac2a4e-436a-47b6-811c-41a214deb69f service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 651.991734] env[65726]: WARNING openstack [req-58cf80ad-4388-4990-9b3f-6f7d2c8dd658 req-142cbe85-245d-473f-9481-cb3143bca571 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 651.992195] env[65726]: WARNING openstack [req-58cf80ad-4388-4990-9b3f-6f7d2c8dd658 req-142cbe85-245d-473f-9481-cb3143bca571 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 652.000562] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19a28c32-d4c5-4859-8323-ece74e26edad {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.003629] env[65726]: DEBUG oslo_vmware.rw_handles [None req-4240f7e0-4ab1-4718-b82f-a2d7b6caacae tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52a00c2f-2a99-27a5-f5a2-5fcf9b09ffeb/disk-0.vmdk. {{(pid=65726) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 652.003784] env[65726]: DEBUG nova.virt.vmwareapi.images [None req-4240f7e0-4ab1-4718-b82f-a2d7b6caacae tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 8e9c27f6-e9fa-4ed0-b6ae-3b5df52d1c44] Uploaded image 699731f0-c4b1-40f5-9f6e-780cb60b3ce3 to the Glance image server {{(pid=65726) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 652.007078] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-4240f7e0-4ab1-4718-b82f-a2d7b6caacae tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 8e9c27f6-e9fa-4ed0-b6ae-3b5df52d1c44] Destroying the VM {{(pid=65726) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 652.013382] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-8b6672eb-727e-468c-a4cf-72bcb0f13f6c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.016839] env[65726]: DEBUG oslo_concurrency.lockutils [None req-881e6c53-2eb2-4fe8-9bf1-92c962897b3f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Acquiring lock "refresh_cache-f2d9090c-988f-43f4-9c81-7aa718a3438a" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 652.016839] env[65726]: DEBUG oslo_concurrency.lockutils [None req-881e6c53-2eb2-4fe8-9bf1-92c962897b3f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Acquired lock "refresh_cache-f2d9090c-988f-43f4-9c81-7aa718a3438a" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 652.016839] env[65726]: DEBUG nova.network.neutron [None req-881e6c53-2eb2-4fe8-9bf1-92c962897b3f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: f2d9090c-988f-43f4-9c81-7aa718a3438a] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 652.017310] env[65726]: DEBUG oslo_vmware.api [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]522d04ea-cea6-da23-0db2-ea78e59ef11a, 'name': SearchDatastore_Task, 'duration_secs': 0.034727} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 652.049781] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-50b66e09-bbf0-45b8-9624-4fc49be4a920 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.059811] env[65726]: DEBUG oslo_vmware.api [None req-4240f7e0-4ab1-4718-b82f-a2d7b6caacae tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Waiting for the task: (returnval){ [ 652.059811] env[65726]: value = "task-5115497" [ 652.059811] env[65726]: _type = "Task" [ 652.059811] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 652.070066] env[65726]: DEBUG oslo_vmware.api [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Waiting for the task: (returnval){ [ 652.070066] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52031a33-80d4-83e1-6718-bcdeb6ba9633" [ 652.070066] env[65726]: _type = "Task" [ 652.070066] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 652.070434] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115496, 'name': CreateVM_Task, 'duration_secs': 0.44382} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 652.071252] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f5151062-57b1-4e4e-93f4-aab0e7f504d4] Created VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 652.075227] env[65726]: WARNING openstack [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 652.075783] env[65726]: WARNING openstack [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 652.081078] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 652.081298] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 652.081748] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 652.088813] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c067522a-aabe-43f2-96aa-ecb3a9263d2f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.091159] env[65726]: DEBUG oslo_vmware.api [None req-4240f7e0-4ab1-4718-b82f-a2d7b6caacae tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': task-5115497, 'name': Destroy_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 652.095564] env[65726]: DEBUG oslo_vmware.api [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Waiting for the task: (returnval){ [ 652.095564] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5235a295-f978-422d-bc76-033825a497f2" [ 652.095564] env[65726]: _type = "Task" [ 652.095564] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 652.099838] env[65726]: DEBUG oslo_vmware.api [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52031a33-80d4-83e1-6718-bcdeb6ba9633, 'name': SearchDatastore_Task, 'duration_secs': 0.01998} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 652.103694] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 652.104210] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore2] 3a60b719-9db7-436e-9908-25f6a233c465/3a60b719-9db7-436e-9908-25f6a233c465.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 652.104210] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7488ff47-f539-4d70-b2af-75b4638520f2 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.113323] env[65726]: DEBUG oslo_vmware.api [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5235a295-f978-422d-bc76-033825a497f2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 652.115528] env[65726]: DEBUG oslo_vmware.api [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Waiting for the task: (returnval){ [ 652.115528] env[65726]: value = "task-5115498" [ 652.115528] env[65726]: _type = "Task" [ 652.115528] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 652.124487] env[65726]: DEBUG oslo_vmware.api [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Task: {'id': task-5115498, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 652.167406] env[65726]: DEBUG oslo_concurrency.lockutils [None req-64ef4520-b00a-42b8-937b-fbfc7c9b2298 tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 652.248438] env[65726]: DEBUG oslo_concurrency.lockutils [req-a90995a5-d731-4249-85b8-09bdde3732a0 req-a1f2baad-ff22-43d1-8421-2b62b9ae2bb6 service nova] Releasing lock "refresh_cache-04ffdf80-c54c-4587-9bb4-d520dc440501" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 652.339241] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8813ac9-14d5-4865-b01e-b7333421fd86 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.364063] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a590efc2-e8df-4dbd-a9c2-17829719ebb4 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.377239] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-32ab2821-307c-47ba-ac96-c38579de1e37 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] [instance: 0e064341-4e4a-407b-8c26-3eb04b409029] Updating instance '0e064341-4e4a-407b-8c26-3eb04b409029' progress to 83 {{(pid=65726) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 652.429013] env[65726]: WARNING openstack [req-3637bfed-a5e8-48d7-8845-d70917530ab6 req-c5ac2a4e-436a-47b6-811c-41a214deb69f service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 652.430166] env[65726]: WARNING openstack [req-3637bfed-a5e8-48d7-8845-d70917530ab6 req-c5ac2a4e-436a-47b6-811c-41a214deb69f service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 652.445202] env[65726]: WARNING openstack [req-58cf80ad-4388-4990-9b3f-6f7d2c8dd658 req-142cbe85-245d-473f-9481-cb3143bca571 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 652.445492] env[65726]: WARNING openstack [req-58cf80ad-4388-4990-9b3f-6f7d2c8dd658 req-142cbe85-245d-473f-9481-cb3143bca571 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 652.521434] env[65726]: WARNING openstack [None req-881e6c53-2eb2-4fe8-9bf1-92c962897b3f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 652.521537] env[65726]: WARNING openstack [None req-881e6c53-2eb2-4fe8-9bf1-92c962897b3f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 652.526579] env[65726]: DEBUG nova.network.neutron [None req-881e6c53-2eb2-4fe8-9bf1-92c962897b3f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: f2d9090c-988f-43f4-9c81-7aa718a3438a] Instance cache missing network info. {{(pid=65726) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 652.573302] env[65726]: DEBUG oslo_vmware.api [None req-4240f7e0-4ab1-4718-b82f-a2d7b6caacae tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': task-5115497, 'name': Destroy_Task, 'duration_secs': 0.34371} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 652.573302] env[65726]: INFO nova.virt.vmwareapi.vm_util [None req-4240f7e0-4ab1-4718-b82f-a2d7b6caacae tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 8e9c27f6-e9fa-4ed0-b6ae-3b5df52d1c44] Destroyed the VM [ 652.573469] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-4240f7e0-4ab1-4718-b82f-a2d7b6caacae tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 8e9c27f6-e9fa-4ed0-b6ae-3b5df52d1c44] Deleting Snapshot of the VM instance {{(pid=65726) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 652.573612] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-b3cc6eaa-84ad-44d4-a6a8-07456fb878ae {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.582129] env[65726]: DEBUG oslo_vmware.api [None req-4240f7e0-4ab1-4718-b82f-a2d7b6caacae tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Waiting for the task: (returnval){ [ 652.582129] env[65726]: value = "task-5115499" [ 652.582129] env[65726]: _type = "Task" [ 652.582129] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 652.591721] env[65726]: DEBUG oslo_vmware.api [None req-4240f7e0-4ab1-4718-b82f-a2d7b6caacae tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': task-5115499, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 652.593351] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47ba0535-a6fe-4445-accc-75924479cdcb {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.607205] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3533f26a-a8ee-4a8f-90f1-3378c21e5e28 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.619847] env[65726]: DEBUG oslo_vmware.api [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5235a295-f978-422d-bc76-033825a497f2, 'name': SearchDatastore_Task, 'duration_secs': 0.017245} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 652.649991] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 652.650574] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] [instance: f5151062-57b1-4e4e-93f4-aab0e7f504d4] Processing image b52362a3-ee8a-4cbf-b06f-513b0cc8f95c {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 652.650724] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 652.650917] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 652.651180] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 652.651842] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1d9c14d1-ec5e-4fa5-98b7-b39f0ef1cdce {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.654259] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1601b76a-44a2-4919-aa76-3b2fc38f03f7 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.669739] env[65726]: DEBUG oslo_vmware.api [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Task: {'id': task-5115498, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 652.671211] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ffe87f4-689a-422a-ba94-ba14a9fe29ad {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.679483] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 652.679848] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65726) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 652.689040] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e0edb95a-2d54-4dcf-b733-fbed1bc0828c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.692989] env[65726]: DEBUG nova.compute.provider_tree [None req-40525054-994c-41ab-bed7-c877834cae18 tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 652.697146] env[65726]: WARNING openstack [req-58cf80ad-4388-4990-9b3f-6f7d2c8dd658 req-142cbe85-245d-473f-9481-cb3143bca571 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 652.697571] env[65726]: WARNING openstack [req-58cf80ad-4388-4990-9b3f-6f7d2c8dd658 req-142cbe85-245d-473f-9481-cb3143bca571 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 652.712621] env[65726]: DEBUG oslo_vmware.api [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Waiting for the task: (returnval){ [ 652.712621] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52b686b9-b1ec-da2c-3e4f-c8ac167348ef" [ 652.712621] env[65726]: _type = "Task" [ 652.712621] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 652.723596] env[65726]: DEBUG oslo_vmware.api [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52b686b9-b1ec-da2c-3e4f-c8ac167348ef, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 652.737351] env[65726]: DEBUG nova.network.neutron [req-3637bfed-a5e8-48d7-8845-d70917530ab6 req-c5ac2a4e-436a-47b6-811c-41a214deb69f service nova] [instance: 3a60b719-9db7-436e-9908-25f6a233c465] Updated VIF entry in instance network info cache for port 5c88141b-7089-49fb-afaa-a38f2cc47aaa. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 652.737727] env[65726]: DEBUG nova.network.neutron [req-3637bfed-a5e8-48d7-8845-d70917530ab6 req-c5ac2a4e-436a-47b6-811c-41a214deb69f service nova] [instance: 3a60b719-9db7-436e-9908-25f6a233c465] Updating instance_info_cache with network_info: [{"id": "5c88141b-7089-49fb-afaa-a38f2cc47aaa", "address": "fa:16:3e:1f:a2:2e", "network": {"id": "b579021f-99ae-447a-b69d-293190930add", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1055463010-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4d508c846ae64e2c84ee281fe59af610", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5fb6e5b3-7be7-4018-8d9e-2133d926178c", "external-id": "nsx-vlan-transportzone-188", "segmentation_id": 188, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5c88141b-70", "ovs_interfaceid": "5c88141b-7089-49fb-afaa-a38f2cc47aaa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 652.787166] env[65726]: WARNING openstack [None req-881e6c53-2eb2-4fe8-9bf1-92c962897b3f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 652.789372] env[65726]: WARNING openstack [None req-881e6c53-2eb2-4fe8-9bf1-92c962897b3f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 652.894277] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-32ab2821-307c-47ba-ac96-c38579de1e37 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] [instance: 0e064341-4e4a-407b-8c26-3eb04b409029] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 652.894338] env[65726]: DEBUG nova.compute.manager [None req-cc552ef7-8733-47e0-9cfd-32896dfc9132 tempest-TenantUsagesTestJSON-327211524 tempest-TenantUsagesTestJSON-327211524-project-member] [instance: 7748c23f-9ea9-4172-8ab7-187678272bb3] Start spawning the instance on the hypervisor. {{(pid=65726) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 652.896472] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d403c808-88fc-4ca1-8bf5-339a57737092 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.907367] env[65726]: DEBUG oslo_vmware.api [None req-32ab2821-307c-47ba-ac96-c38579de1e37 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Waiting for the task: (returnval){ [ 652.907367] env[65726]: value = "task-5115500" [ 652.907367] env[65726]: _type = "Task" [ 652.907367] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 652.919365] env[65726]: DEBUG oslo_vmware.api [None req-32ab2821-307c-47ba-ac96-c38579de1e37 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Task: {'id': task-5115500, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 652.943350] env[65726]: DEBUG nova.virt.hardware [None req-cc552ef7-8733-47e0-9cfd-32896dfc9132 tempest-TenantUsagesTestJSON-327211524 tempest-TenantUsagesTestJSON-327211524-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T19:28:34Z,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1539f06fa3534e90acbc3a60c4b8bd3f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T19:28:35Z,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 652.943701] env[65726]: DEBUG nova.virt.hardware [None req-cc552ef7-8733-47e0-9cfd-32896dfc9132 tempest-TenantUsagesTestJSON-327211524 tempest-TenantUsagesTestJSON-327211524-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 652.943866] env[65726]: DEBUG nova.virt.hardware [None req-cc552ef7-8733-47e0-9cfd-32896dfc9132 tempest-TenantUsagesTestJSON-327211524 tempest-TenantUsagesTestJSON-327211524-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 652.943999] env[65726]: DEBUG nova.virt.hardware [None req-cc552ef7-8733-47e0-9cfd-32896dfc9132 tempest-TenantUsagesTestJSON-327211524 tempest-TenantUsagesTestJSON-327211524-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 652.945728] env[65726]: DEBUG nova.virt.hardware [None req-cc552ef7-8733-47e0-9cfd-32896dfc9132 tempest-TenantUsagesTestJSON-327211524 tempest-TenantUsagesTestJSON-327211524-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 652.945728] env[65726]: DEBUG nova.virt.hardware [None req-cc552ef7-8733-47e0-9cfd-32896dfc9132 tempest-TenantUsagesTestJSON-327211524 tempest-TenantUsagesTestJSON-327211524-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 652.945728] env[65726]: DEBUG nova.virt.hardware [None req-cc552ef7-8733-47e0-9cfd-32896dfc9132 tempest-TenantUsagesTestJSON-327211524 tempest-TenantUsagesTestJSON-327211524-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 652.945728] env[65726]: DEBUG nova.virt.hardware [None req-cc552ef7-8733-47e0-9cfd-32896dfc9132 tempest-TenantUsagesTestJSON-327211524 tempest-TenantUsagesTestJSON-327211524-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 652.945728] env[65726]: DEBUG nova.virt.hardware [None req-cc552ef7-8733-47e0-9cfd-32896dfc9132 tempest-TenantUsagesTestJSON-327211524 tempest-TenantUsagesTestJSON-327211524-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 652.945984] env[65726]: DEBUG nova.virt.hardware [None req-cc552ef7-8733-47e0-9cfd-32896dfc9132 tempest-TenantUsagesTestJSON-327211524 tempest-TenantUsagesTestJSON-327211524-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 652.945984] env[65726]: DEBUG nova.virt.hardware [None req-cc552ef7-8733-47e0-9cfd-32896dfc9132 tempest-TenantUsagesTestJSON-327211524 tempest-TenantUsagesTestJSON-327211524-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 652.946245] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbf1470d-feef-4abf-9026-d6e261d036e2 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.951379] env[65726]: DEBUG nova.network.neutron [req-58cf80ad-4388-4990-9b3f-6f7d2c8dd658 req-142cbe85-245d-473f-9481-cb3143bca571 service nova] [instance: f5151062-57b1-4e4e-93f4-aab0e7f504d4] Updated VIF entry in instance network info cache for port 83750abb-1d26-4e23-a675-61ef785b858b. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 652.951739] env[65726]: DEBUG nova.network.neutron [req-58cf80ad-4388-4990-9b3f-6f7d2c8dd658 req-142cbe85-245d-473f-9481-cb3143bca571 service nova] [instance: f5151062-57b1-4e4e-93f4-aab0e7f504d4] Updating instance_info_cache with network_info: [{"id": "83750abb-1d26-4e23-a675-61ef785b858b", "address": "fa:16:3e:42:77:9f", "network": {"id": "b579021f-99ae-447a-b69d-293190930add", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1055463010-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4d508c846ae64e2c84ee281fe59af610", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5fb6e5b3-7be7-4018-8d9e-2133d926178c", "external-id": "nsx-vlan-transportzone-188", "segmentation_id": 188, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap83750abb-1d", "ovs_interfaceid": "83750abb-1d26-4e23-a675-61ef785b858b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 652.960355] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fff82c2d-9702-4968-8d36-12b112d9cd5d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.998664] env[65726]: WARNING openstack [None req-881e6c53-2eb2-4fe8-9bf1-92c962897b3f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 652.999027] env[65726]: WARNING openstack [None req-881e6c53-2eb2-4fe8-9bf1-92c962897b3f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 653.085645] env[65726]: DEBUG nova.compute.manager [None req-8873e006-42c1-4c1b-8e78-853b15a3788b tempest-ServerDiagnosticsV248Test-1827630017 tempest-ServerDiagnosticsV248Test-1827630017-project-admin] [instance: 4d4dc88b-0d6c-430c-ab4f-c81a56e4657b] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 653.093957] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71260ee6-c805-414d-91c7-a50e2db814dc {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.109465] env[65726]: INFO nova.compute.manager [None req-8873e006-42c1-4c1b-8e78-853b15a3788b tempest-ServerDiagnosticsV248Test-1827630017 tempest-ServerDiagnosticsV248Test-1827630017-project-admin] [instance: 4d4dc88b-0d6c-430c-ab4f-c81a56e4657b] Retrieving diagnostics [ 653.110316] env[65726]: DEBUG oslo_vmware.api [None req-4240f7e0-4ab1-4718-b82f-a2d7b6caacae tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': task-5115499, 'name': RemoveSnapshot_Task, 'duration_secs': 0.462032} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 653.111094] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69ec088c-e3f8-4ca6-beb6-68e0712bbe56 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.114015] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-4240f7e0-4ab1-4718-b82f-a2d7b6caacae tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 8e9c27f6-e9fa-4ed0-b6ae-3b5df52d1c44] Deleted Snapshot of the VM instance {{(pid=65726) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 653.114209] env[65726]: INFO nova.compute.manager [None req-4240f7e0-4ab1-4718-b82f-a2d7b6caacae tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 8e9c27f6-e9fa-4ed0-b6ae-3b5df52d1c44] Took 18.08 seconds to snapshot the instance on the hypervisor. [ 653.154112] env[65726]: DEBUG oslo_vmware.api [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Task: {'id': task-5115498, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.910364} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 653.154334] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore2] 3a60b719-9db7-436e-9908-25f6a233c465/3a60b719-9db7-436e-9908-25f6a233c465.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 653.154536] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] [instance: 3a60b719-9db7-436e-9908-25f6a233c465] Extending root virtual disk to 1048576 {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 653.154794] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-38fb6a55-7846-47b4-bbcd-20827dd8ff71 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.163657] env[65726]: DEBUG oslo_vmware.api [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Waiting for the task: (returnval){ [ 653.163657] env[65726]: value = "task-5115501" [ 653.163657] env[65726]: _type = "Task" [ 653.163657] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 653.174701] env[65726]: DEBUG oslo_vmware.api [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Task: {'id': task-5115501, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 653.188923] env[65726]: DEBUG nova.network.neutron [None req-881e6c53-2eb2-4fe8-9bf1-92c962897b3f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: f2d9090c-988f-43f4-9c81-7aa718a3438a] Updating instance_info_cache with network_info: [{"id": "7117cd21-ee2a-497d-b789-65e43b068258", "address": "fa:16:3e:10:2a:24", "network": {"id": "3f7916f7-ee9c-4629-b3b3-1367de9f98ea", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-579713617-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b794c99309e94c1181fc9d8eeeb84702", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d0bdd11b-58af-4cc0-9d38-8322e1bb4e74", "external-id": "nsx-vlan-transportzone-398", "segmentation_id": 398, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7117cd21-ee", "ovs_interfaceid": "7117cd21-ee2a-497d-b789-65e43b068258", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 653.207174] env[65726]: DEBUG nova.scheduler.client.report [None req-40525054-994c-41ab-bed7-c877834cae18 tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 653.225506] env[65726]: DEBUG oslo_vmware.api [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52b686b9-b1ec-da2c-3e4f-c8ac167348ef, 'name': SearchDatastore_Task, 'duration_secs': 0.062763} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 653.227289] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6d5c6596-6097-4950-927d-ee7a65b61aef {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.234465] env[65726]: DEBUG oslo_vmware.api [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Waiting for the task: (returnval){ [ 653.234465] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52230773-875f-c28e-e7e3-7e1721a534d7" [ 653.234465] env[65726]: _type = "Task" [ 653.234465] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 653.242223] env[65726]: DEBUG oslo_concurrency.lockutils [req-3637bfed-a5e8-48d7-8845-d70917530ab6 req-c5ac2a4e-436a-47b6-811c-41a214deb69f service nova] Releasing lock "refresh_cache-3a60b719-9db7-436e-9908-25f6a233c465" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 653.247309] env[65726]: DEBUG oslo_vmware.api [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52230773-875f-c28e-e7e3-7e1721a534d7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 653.346699] env[65726]: DEBUG oslo_concurrency.lockutils [None req-f3171f49-8536-4923-b5cf-31399a616a07 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Acquiring lock "608cc0ea-3f6b-4b4f-83c1-01aa50999d63" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 653.346955] env[65726]: DEBUG oslo_concurrency.lockutils [None req-f3171f49-8536-4923-b5cf-31399a616a07 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Lock "608cc0ea-3f6b-4b4f-83c1-01aa50999d63" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 653.423463] env[65726]: DEBUG oslo_vmware.api [None req-32ab2821-307c-47ba-ac96-c38579de1e37 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Task: {'id': task-5115500, 'name': PowerOnVM_Task} progress is 88%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 653.456330] env[65726]: DEBUG oslo_concurrency.lockutils [req-58cf80ad-4388-4990-9b3f-6f7d2c8dd658 req-142cbe85-245d-473f-9481-cb3143bca571 service nova] Releasing lock "refresh_cache-f5151062-57b1-4e4e-93f4-aab0e7f504d4" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 653.653050] env[65726]: DEBUG nova.network.neutron [None req-cc552ef7-8733-47e0-9cfd-32896dfc9132 tempest-TenantUsagesTestJSON-327211524 tempest-TenantUsagesTestJSON-327211524-project-member] [instance: 7748c23f-9ea9-4172-8ab7-187678272bb3] Successfully updated port: 8bb2d672-e56e-4f01-bfbf-74a279d9da7d {{(pid=65726) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 653.675880] env[65726]: DEBUG oslo_vmware.api [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Task: {'id': task-5115501, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.150498} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 653.675880] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] [instance: 3a60b719-9db7-436e-9908-25f6a233c465] Extended root virtual disk {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 653.676945] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88c66717-f153-476e-b762-9e929e2db0e3 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.696348] env[65726]: DEBUG oslo_concurrency.lockutils [None req-881e6c53-2eb2-4fe8-9bf1-92c962897b3f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Releasing lock "refresh_cache-f2d9090c-988f-43f4-9c81-7aa718a3438a" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 653.697235] env[65726]: DEBUG nova.compute.manager [None req-881e6c53-2eb2-4fe8-9bf1-92c962897b3f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: f2d9090c-988f-43f4-9c81-7aa718a3438a] Instance network_info: |[{"id": "7117cd21-ee2a-497d-b789-65e43b068258", "address": "fa:16:3e:10:2a:24", "network": {"id": "3f7916f7-ee9c-4629-b3b3-1367de9f98ea", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-579713617-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b794c99309e94c1181fc9d8eeeb84702", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d0bdd11b-58af-4cc0-9d38-8322e1bb4e74", "external-id": "nsx-vlan-transportzone-398", "segmentation_id": 398, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7117cd21-ee", "ovs_interfaceid": "7117cd21-ee2a-497d-b789-65e43b068258", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 653.711316] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] [instance: 3a60b719-9db7-436e-9908-25f6a233c465] Reconfiguring VM instance instance-0000000f to attach disk [datastore2] 3a60b719-9db7-436e-9908-25f6a233c465/3a60b719-9db7-436e-9908-25f6a233c465.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 653.711316] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-881e6c53-2eb2-4fe8-9bf1-92c962897b3f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: f2d9090c-988f-43f4-9c81-7aa718a3438a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:10:2a:24', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd0bdd11b-58af-4cc0-9d38-8322e1bb4e74', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7117cd21-ee2a-497d-b789-65e43b068258', 'vif_model': 'vmxnet3'}] {{(pid=65726) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 653.718931] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-881e6c53-2eb2-4fe8-9bf1-92c962897b3f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Creating folder: Project (b794c99309e94c1181fc9d8eeeb84702). Parent ref: group-v995008. {{(pid=65726) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 653.719758] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6f604af6-32a0-41cf-b916-f342cd9f71c6 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.735716] env[65726]: DEBUG oslo_concurrency.lockutils [None req-40525054-994c-41ab-bed7-c877834cae18 tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.878s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 653.736251] env[65726]: DEBUG nova.compute.manager [None req-40525054-994c-41ab-bed7-c877834cae18 tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] [instance: 3b933684-ac19-44b0-a49d-6af45501e166] Start building networks asynchronously for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 653.739341] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-05d0dd25-73a3-46c8-8283-bc86586a8b3e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.742042] env[65726]: DEBUG oslo_concurrency.lockutils [None req-362ec7a6-07fd-4c6b-b5fb-9865bdcb8c1a tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 10.638s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 653.743625] env[65726]: INFO nova.compute.claims [None req-362ec7a6-07fd-4c6b-b5fb-9865bdcb8c1a tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: 274ab469-61a9-4b7e-852c-074c871e3abf] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 653.761827] env[65726]: DEBUG oslo_vmware.api [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52230773-875f-c28e-e7e3-7e1721a534d7, 'name': SearchDatastore_Task, 'duration_secs': 0.017824} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 653.765082] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 653.765082] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore2] f5151062-57b1-4e4e-93f4-aab0e7f504d4/f5151062-57b1-4e4e-93f4-aab0e7f504d4.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 653.765587] env[65726]: DEBUG oslo_vmware.api [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Waiting for the task: (returnval){ [ 653.765587] env[65726]: value = "task-5115503" [ 653.765587] env[65726]: _type = "Task" [ 653.765587] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 653.766964] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-59ec2d05-385b-4e4a-bba9-34aa68345b0e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.769560] env[65726]: INFO nova.virt.vmwareapi.vm_util [None req-881e6c53-2eb2-4fe8-9bf1-92c962897b3f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Created folder: Project (b794c99309e94c1181fc9d8eeeb84702) in parent group-v995008. [ 653.769892] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-881e6c53-2eb2-4fe8-9bf1-92c962897b3f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Creating folder: Instances. Parent ref: group-v995053. {{(pid=65726) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 653.774252] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-69c065b1-5869-419d-88f3-8010411d1254 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.784713] env[65726]: DEBUG oslo_vmware.api [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Task: {'id': task-5115503, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 653.786681] env[65726]: DEBUG oslo_vmware.api [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Waiting for the task: (returnval){ [ 653.786681] env[65726]: value = "task-5115504" [ 653.786681] env[65726]: _type = "Task" [ 653.786681] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 653.798134] env[65726]: INFO nova.virt.vmwareapi.vm_util [None req-881e6c53-2eb2-4fe8-9bf1-92c962897b3f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Created folder: Instances in parent group-v995053. [ 653.798134] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-881e6c53-2eb2-4fe8-9bf1-92c962897b3f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 653.798708] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f2d9090c-988f-43f4-9c81-7aa718a3438a] Creating VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 653.798958] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-04579f43-3ce7-4215-b14a-4f7b9f8baa95 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.820392] env[65726]: DEBUG oslo_vmware.api [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Task: {'id': task-5115504, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 653.826935] env[65726]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 653.826935] env[65726]: value = "task-5115506" [ 653.826935] env[65726]: _type = "Task" [ 653.826935] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 653.839146] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115506, 'name': CreateVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 653.923379] env[65726]: DEBUG oslo_vmware.api [None req-32ab2821-307c-47ba-ac96-c38579de1e37 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Task: {'id': task-5115500, 'name': PowerOnVM_Task} progress is 88%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 654.159025] env[65726]: DEBUG oslo_concurrency.lockutils [None req-cc552ef7-8733-47e0-9cfd-32896dfc9132 tempest-TenantUsagesTestJSON-327211524 tempest-TenantUsagesTestJSON-327211524-project-member] Acquiring lock "refresh_cache-7748c23f-9ea9-4172-8ab7-187678272bb3" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 654.159025] env[65726]: DEBUG oslo_concurrency.lockutils [None req-cc552ef7-8733-47e0-9cfd-32896dfc9132 tempest-TenantUsagesTestJSON-327211524 tempest-TenantUsagesTestJSON-327211524-project-member] Acquired lock "refresh_cache-7748c23f-9ea9-4172-8ab7-187678272bb3" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 654.159183] env[65726]: DEBUG nova.network.neutron [None req-cc552ef7-8733-47e0-9cfd-32896dfc9132 tempest-TenantUsagesTestJSON-327211524 tempest-TenantUsagesTestJSON-327211524-project-member] [instance: 7748c23f-9ea9-4172-8ab7-187678272bb3] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 654.244590] env[65726]: DEBUG nova.compute.utils [None req-40525054-994c-41ab-bed7-c877834cae18 tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] Using /dev/sd instead of None {{(pid=65726) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 654.246796] env[65726]: DEBUG nova.compute.manager [None req-40525054-994c-41ab-bed7-c877834cae18 tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] [instance: 3b933684-ac19-44b0-a49d-6af45501e166] Not allocating networking since 'none' was specified. {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2015}} [ 654.284942] env[65726]: DEBUG oslo_vmware.api [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Task: {'id': task-5115503, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 654.303920] env[65726]: DEBUG oslo_vmware.api [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Task: {'id': task-5115504, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 654.341350] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115506, 'name': CreateVM_Task} progress is 99%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 654.421731] env[65726]: DEBUG oslo_vmware.api [None req-32ab2821-307c-47ba-ac96-c38579de1e37 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Task: {'id': task-5115500, 'name': PowerOnVM_Task, 'duration_secs': 1.145097} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 654.422321] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-32ab2821-307c-47ba-ac96-c38579de1e37 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] [instance: 0e064341-4e4a-407b-8c26-3eb04b409029] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 654.422321] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-32ab2821-307c-47ba-ac96-c38579de1e37 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] [instance: 0e064341-4e4a-407b-8c26-3eb04b409029] Updating instance '0e064341-4e4a-407b-8c26-3eb04b409029' progress to 100 {{(pid=65726) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 654.529345] env[65726]: DEBUG oslo_concurrency.lockutils [None req-296fd45e-ee7a-4e44-912a-f3b5a22cbbf0 tempest-ServerExternalEventsTest-1845125219 tempest-ServerExternalEventsTest-1845125219-project-member] Acquiring lock "96acc5f8-b411-408e-929d-d1035cfd50db" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 654.529345] env[65726]: DEBUG oslo_concurrency.lockutils [None req-296fd45e-ee7a-4e44-912a-f3b5a22cbbf0 tempest-ServerExternalEventsTest-1845125219 tempest-ServerExternalEventsTest-1845125219-project-member] Lock "96acc5f8-b411-408e-929d-d1035cfd50db" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 654.667528] env[65726]: WARNING openstack [None req-cc552ef7-8733-47e0-9cfd-32896dfc9132 tempest-TenantUsagesTestJSON-327211524 tempest-TenantUsagesTestJSON-327211524-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 654.667528] env[65726]: WARNING openstack [None req-cc552ef7-8733-47e0-9cfd-32896dfc9132 tempest-TenantUsagesTestJSON-327211524 tempest-TenantUsagesTestJSON-327211524-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 654.673233] env[65726]: DEBUG nova.network.neutron [None req-cc552ef7-8733-47e0-9cfd-32896dfc9132 tempest-TenantUsagesTestJSON-327211524 tempest-TenantUsagesTestJSON-327211524-project-member] [instance: 7748c23f-9ea9-4172-8ab7-187678272bb3] Instance cache missing network info. {{(pid=65726) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 654.738321] env[65726]: WARNING openstack [None req-cc552ef7-8733-47e0-9cfd-32896dfc9132 tempest-TenantUsagesTestJSON-327211524 tempest-TenantUsagesTestJSON-327211524-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 654.738564] env[65726]: WARNING openstack [None req-cc552ef7-8733-47e0-9cfd-32896dfc9132 tempest-TenantUsagesTestJSON-327211524 tempest-TenantUsagesTestJSON-327211524-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 654.748613] env[65726]: DEBUG nova.compute.manager [None req-40525054-994c-41ab-bed7-c877834cae18 tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] [instance: 3b933684-ac19-44b0-a49d-6af45501e166] Start building block device mappings for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 654.754165] env[65726]: DEBUG nova.compute.manager [req-0f31bbc3-93ad-4deb-8036-b7cc96fecf02 req-a2260251-52ad-40fb-aaf8-2159f9516fd6 service nova] [instance: 8493ba52-b950-4727-970a-19b1797b2ca1] Received event network-vif-deleted-964ef0b7-9857-4032-ac36-efb1ecd1419d {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 654.754378] env[65726]: DEBUG nova.compute.manager [req-0f31bbc3-93ad-4deb-8036-b7cc96fecf02 req-a2260251-52ad-40fb-aaf8-2159f9516fd6 service nova] [instance: 21730277-cef6-4bfe-9f67-de71f6f615ff] Received event network-vif-deleted-cc7e0970-7e47-45d6-b7fa-6e33947b210c {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 654.781850] env[65726]: DEBUG nova.compute.manager [req-21237459-e2d1-4f37-bebf-e40793b06175 req-69557957-f522-464c-bd7e-ffbde47a6481 service nova] [instance: aa2716a9-2cf8-4e51-800d-1c8f98b2cf05] Received event network-changed-c4f80985-65f1-478f-9944-afb3538efb48 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 654.782044] env[65726]: DEBUG nova.compute.manager [req-21237459-e2d1-4f37-bebf-e40793b06175 req-69557957-f522-464c-bd7e-ffbde47a6481 service nova] [instance: aa2716a9-2cf8-4e51-800d-1c8f98b2cf05] Refreshing instance network info cache due to event network-changed-c4f80985-65f1-478f-9944-afb3538efb48. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 654.782255] env[65726]: DEBUG oslo_concurrency.lockutils [req-21237459-e2d1-4f37-bebf-e40793b06175 req-69557957-f522-464c-bd7e-ffbde47a6481 service nova] Acquiring lock "refresh_cache-aa2716a9-2cf8-4e51-800d-1c8f98b2cf05" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 654.782400] env[65726]: DEBUG oslo_concurrency.lockutils [req-21237459-e2d1-4f37-bebf-e40793b06175 req-69557957-f522-464c-bd7e-ffbde47a6481 service nova] Acquired lock "refresh_cache-aa2716a9-2cf8-4e51-800d-1c8f98b2cf05" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 654.782545] env[65726]: DEBUG nova.network.neutron [req-21237459-e2d1-4f37-bebf-e40793b06175 req-69557957-f522-464c-bd7e-ffbde47a6481 service nova] [instance: aa2716a9-2cf8-4e51-800d-1c8f98b2cf05] Refreshing network info cache for port c4f80985-65f1-478f-9944-afb3538efb48 {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 654.791288] env[65726]: DEBUG oslo_vmware.api [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Task: {'id': task-5115503, 'name': ReconfigVM_Task, 'duration_secs': 0.616153} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 654.794081] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] [instance: 3a60b719-9db7-436e-9908-25f6a233c465] Reconfigured VM instance instance-0000000f to attach disk [datastore2] 3a60b719-9db7-436e-9908-25f6a233c465/3a60b719-9db7-436e-9908-25f6a233c465.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 654.795113] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e243cfac-673e-49f8-a0c5-8280b3554b18 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.803788] env[65726]: DEBUG oslo_vmware.api [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Task: {'id': task-5115504, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.781332} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 654.807972] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore2] f5151062-57b1-4e4e-93f4-aab0e7f504d4/f5151062-57b1-4e4e-93f4-aab0e7f504d4.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 654.808365] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] [instance: f5151062-57b1-4e4e-93f4-aab0e7f504d4] Extending root virtual disk to 1048576 {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 654.809282] env[65726]: DEBUG oslo_vmware.api [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Waiting for the task: (returnval){ [ 654.809282] env[65726]: value = "task-5115507" [ 654.809282] env[65726]: _type = "Task" [ 654.809282] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 654.809708] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1e946808-2a3d-4011-b645-867dba361447 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.821509] env[65726]: DEBUG oslo_vmware.api [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Task: {'id': task-5115507, 'name': Rename_Task} progress is 10%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 654.826180] env[65726]: DEBUG oslo_vmware.api [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Waiting for the task: (returnval){ [ 654.826180] env[65726]: value = "task-5115508" [ 654.826180] env[65726]: _type = "Task" [ 654.826180] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 654.830594] env[65726]: WARNING openstack [None req-cc552ef7-8733-47e0-9cfd-32896dfc9132 tempest-TenantUsagesTestJSON-327211524 tempest-TenantUsagesTestJSON-327211524-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 654.831011] env[65726]: WARNING openstack [None req-cc552ef7-8733-47e0-9cfd-32896dfc9132 tempest-TenantUsagesTestJSON-327211524 tempest-TenantUsagesTestJSON-327211524-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 654.859162] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115506, 'name': CreateVM_Task, 'duration_secs': 0.561895} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 654.862423] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f2d9090c-988f-43f4-9c81-7aa718a3438a] Created VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 654.866028] env[65726]: WARNING openstack [None req-881e6c53-2eb2-4fe8-9bf1-92c962897b3f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 654.866028] env[65726]: WARNING openstack [None req-881e6c53-2eb2-4fe8-9bf1-92c962897b3f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 654.869461] env[65726]: DEBUG oslo_concurrency.lockutils [None req-881e6c53-2eb2-4fe8-9bf1-92c962897b3f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 654.869534] env[65726]: DEBUG oslo_concurrency.lockutils [None req-881e6c53-2eb2-4fe8-9bf1-92c962897b3f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 654.869848] env[65726]: DEBUG oslo_concurrency.lockutils [None req-881e6c53-2eb2-4fe8-9bf1-92c962897b3f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 654.871039] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1af4ff8e-56f1-42d5-8659-fd0592c83c4f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.880383] env[65726]: DEBUG oslo_vmware.api [None req-881e6c53-2eb2-4fe8-9bf1-92c962897b3f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Waiting for the task: (returnval){ [ 654.880383] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52cbb417-8c2e-4b7a-b129-af6183a50b71" [ 654.880383] env[65726]: _type = "Task" [ 654.880383] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 654.895731] env[65726]: DEBUG oslo_vmware.api [None req-881e6c53-2eb2-4fe8-9bf1-92c962897b3f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52cbb417-8c2e-4b7a-b129-af6183a50b71, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 654.965263] env[65726]: DEBUG nova.network.neutron [None req-cc552ef7-8733-47e0-9cfd-32896dfc9132 tempest-TenantUsagesTestJSON-327211524 tempest-TenantUsagesTestJSON-327211524-project-member] [instance: 7748c23f-9ea9-4172-8ab7-187678272bb3] Updating instance_info_cache with network_info: [{"id": "8bb2d672-e56e-4f01-bfbf-74a279d9da7d", "address": "fa:16:3e:01:7e:88", "network": {"id": "55d629b8-a541-4720-9289-927bbc6e6f22", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.177", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "1539f06fa3534e90acbc3a60c4b8bd3f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "35342bcb-8b06-472e-b3c0-43fd3d6c4b30", "external-id": "nsx-vlan-transportzone-524", "segmentation_id": 524, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8bb2d672-e5", "ovs_interfaceid": "8bb2d672-e56e-4f01-bfbf-74a279d9da7d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 655.084330] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b83663a4-2303-46f3-8690-72e6256a8051 tempest-ServerDiagnosticsTest-2021442419 tempest-ServerDiagnosticsTest-2021442419-project-member] Acquiring lock "de5cad99-1dbc-4435-98b4-987e24ba8c3c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 655.084626] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b83663a4-2303-46f3-8690-72e6256a8051 tempest-ServerDiagnosticsTest-2021442419 tempest-ServerDiagnosticsTest-2021442419-project-member] Lock "de5cad99-1dbc-4435-98b4-987e24ba8c3c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 655.084949] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b83663a4-2303-46f3-8690-72e6256a8051 tempest-ServerDiagnosticsTest-2021442419 tempest-ServerDiagnosticsTest-2021442419-project-member] Acquiring lock "de5cad99-1dbc-4435-98b4-987e24ba8c3c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 655.085207] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b83663a4-2303-46f3-8690-72e6256a8051 tempest-ServerDiagnosticsTest-2021442419 tempest-ServerDiagnosticsTest-2021442419-project-member] Lock "de5cad99-1dbc-4435-98b4-987e24ba8c3c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 655.085207] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b83663a4-2303-46f3-8690-72e6256a8051 tempest-ServerDiagnosticsTest-2021442419 tempest-ServerDiagnosticsTest-2021442419-project-member] Lock "de5cad99-1dbc-4435-98b4-987e24ba8c3c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 655.088707] env[65726]: INFO nova.compute.manager [None req-b83663a4-2303-46f3-8690-72e6256a8051 tempest-ServerDiagnosticsTest-2021442419 tempest-ServerDiagnosticsTest-2021442419-project-member] [instance: de5cad99-1dbc-4435-98b4-987e24ba8c3c] Terminating instance [ 655.294238] env[65726]: WARNING openstack [req-21237459-e2d1-4f37-bebf-e40793b06175 req-69557957-f522-464c-bd7e-ffbde47a6481 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 655.294440] env[65726]: WARNING openstack [req-21237459-e2d1-4f37-bebf-e40793b06175 req-69557957-f522-464c-bd7e-ffbde47a6481 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 655.326027] env[65726]: DEBUG oslo_vmware.api [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Task: {'id': task-5115507, 'name': Rename_Task, 'duration_secs': 0.288678} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 655.329315] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] [instance: 3a60b719-9db7-436e-9908-25f6a233c465] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 655.330348] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ec70e9c2-d2c6-4a64-92c7-2ddf735cef4b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.342138] env[65726]: DEBUG oslo_vmware.api [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Waiting for the task: (returnval){ [ 655.342138] env[65726]: value = "task-5115509" [ 655.342138] env[65726]: _type = "Task" [ 655.342138] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 655.367024] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5fd4688c-49eb-401e-afbd-116789f6a402 tempest-ServerDiagnosticsV248Test-347810772 tempest-ServerDiagnosticsV248Test-347810772-project-member] Acquiring lock "4d4dc88b-0d6c-430c-ab4f-c81a56e4657b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 655.367024] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5fd4688c-49eb-401e-afbd-116789f6a402 tempest-ServerDiagnosticsV248Test-347810772 tempest-ServerDiagnosticsV248Test-347810772-project-member] Lock "4d4dc88b-0d6c-430c-ab4f-c81a56e4657b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 655.367024] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5fd4688c-49eb-401e-afbd-116789f6a402 tempest-ServerDiagnosticsV248Test-347810772 tempest-ServerDiagnosticsV248Test-347810772-project-member] Acquiring lock "4d4dc88b-0d6c-430c-ab4f-c81a56e4657b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 655.367024] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5fd4688c-49eb-401e-afbd-116789f6a402 tempest-ServerDiagnosticsV248Test-347810772 tempest-ServerDiagnosticsV248Test-347810772-project-member] Lock "4d4dc88b-0d6c-430c-ab4f-c81a56e4657b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 655.367491] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5fd4688c-49eb-401e-afbd-116789f6a402 tempest-ServerDiagnosticsV248Test-347810772 tempest-ServerDiagnosticsV248Test-347810772-project-member] Lock "4d4dc88b-0d6c-430c-ab4f-c81a56e4657b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 655.372555] env[65726]: DEBUG oslo_vmware.api [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Task: {'id': task-5115508, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.198052} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 655.372816] env[65726]: DEBUG oslo_vmware.api [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Task: {'id': task-5115509, 'name': PowerOnVM_Task} progress is 33%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 655.373495] env[65726]: INFO nova.compute.manager [None req-5fd4688c-49eb-401e-afbd-116789f6a402 tempest-ServerDiagnosticsV248Test-347810772 tempest-ServerDiagnosticsV248Test-347810772-project-member] [instance: 4d4dc88b-0d6c-430c-ab4f-c81a56e4657b] Terminating instance [ 655.375336] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] [instance: f5151062-57b1-4e4e-93f4-aab0e7f504d4] Extended root virtual disk {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 655.377459] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a87a5eb-1f29-4827-af36-d5b12aef06dd {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.409277] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] [instance: f5151062-57b1-4e4e-93f4-aab0e7f504d4] Reconfiguring VM instance instance-00000010 to attach disk [datastore2] f5151062-57b1-4e4e-93f4-aab0e7f504d4/f5151062-57b1-4e4e-93f4-aab0e7f504d4.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 655.418081] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-77f4e47c-d1fc-450f-ba3d-d2335581430a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.432919] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6daea59f-1a35-44eb-93fc-3798961f9df7 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.436259] env[65726]: DEBUG oslo_vmware.api [None req-881e6c53-2eb2-4fe8-9bf1-92c962897b3f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52cbb417-8c2e-4b7a-b129-af6183a50b71, 'name': SearchDatastore_Task, 'duration_secs': 0.020833} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 655.439211] env[65726]: DEBUG oslo_concurrency.lockutils [None req-881e6c53-2eb2-4fe8-9bf1-92c962897b3f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 655.439211] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-881e6c53-2eb2-4fe8-9bf1-92c962897b3f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: f2d9090c-988f-43f4-9c81-7aa718a3438a] Processing image b52362a3-ee8a-4cbf-b06f-513b0cc8f95c {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 655.439211] env[65726]: DEBUG oslo_concurrency.lockutils [None req-881e6c53-2eb2-4fe8-9bf1-92c962897b3f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 655.439211] env[65726]: DEBUG oslo_concurrency.lockutils [None req-881e6c53-2eb2-4fe8-9bf1-92c962897b3f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 655.439518] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-881e6c53-2eb2-4fe8-9bf1-92c962897b3f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 655.442618] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7252caf8-1897-477f-a4d0-d4452a4cdef2 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.450811] env[65726]: DEBUG oslo_vmware.api [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Waiting for the task: (returnval){ [ 655.450811] env[65726]: value = "task-5115510" [ 655.450811] env[65726]: _type = "Task" [ 655.450811] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 655.453133] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7cb58332-70df-42f6-ba5a-8265a89ec130 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.464859] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-881e6c53-2eb2-4fe8-9bf1-92c962897b3f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 655.465204] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-881e6c53-2eb2-4fe8-9bf1-92c962897b3f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65726) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 655.466783] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ae3d08fa-a03c-4c17-b2bd-9ba732a1db67 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.516905] env[65726]: DEBUG oslo_concurrency.lockutils [None req-cc552ef7-8733-47e0-9cfd-32896dfc9132 tempest-TenantUsagesTestJSON-327211524 tempest-TenantUsagesTestJSON-327211524-project-member] Releasing lock "refresh_cache-7748c23f-9ea9-4172-8ab7-187678272bb3" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 655.517092] env[65726]: DEBUG nova.compute.manager [None req-cc552ef7-8733-47e0-9cfd-32896dfc9132 tempest-TenantUsagesTestJSON-327211524 tempest-TenantUsagesTestJSON-327211524-project-member] [instance: 7748c23f-9ea9-4172-8ab7-187678272bb3] Instance network_info: |[{"id": "8bb2d672-e56e-4f01-bfbf-74a279d9da7d", "address": "fa:16:3e:01:7e:88", "network": {"id": "55d629b8-a541-4720-9289-927bbc6e6f22", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.177", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "1539f06fa3534e90acbc3a60c4b8bd3f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "35342bcb-8b06-472e-b3c0-43fd3d6c4b30", "external-id": "nsx-vlan-transportzone-524", "segmentation_id": 524, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8bb2d672-e5", "ovs_interfaceid": "8bb2d672-e56e-4f01-bfbf-74a279d9da7d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 655.517730] env[65726]: DEBUG oslo_vmware.api [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Task: {'id': task-5115510, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 655.520901] env[65726]: WARNING openstack [req-21237459-e2d1-4f37-bebf-e40793b06175 req-69557957-f522-464c-bd7e-ffbde47a6481 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 655.521272] env[65726]: WARNING openstack [req-21237459-e2d1-4f37-bebf-e40793b06175 req-69557957-f522-464c-bd7e-ffbde47a6481 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 655.528048] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-cc552ef7-8733-47e0-9cfd-32896dfc9132 tempest-TenantUsagesTestJSON-327211524 tempest-TenantUsagesTestJSON-327211524-project-member] [instance: 7748c23f-9ea9-4172-8ab7-187678272bb3] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:01:7e:88', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '35342bcb-8b06-472e-b3c0-43fd3d6c4b30', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '8bb2d672-e56e-4f01-bfbf-74a279d9da7d', 'vif_model': 'vmxnet3'}] {{(pid=65726) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 655.535969] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-cc552ef7-8733-47e0-9cfd-32896dfc9132 tempest-TenantUsagesTestJSON-327211524 tempest-TenantUsagesTestJSON-327211524-project-member] Creating folder: Project (a2b0d29fd55f4116acb5331831eb23e9). Parent ref: group-v995008. {{(pid=65726) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 655.537754] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13ccdc62-641a-4b64-9c79-cb6e4ca386bd {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.548023] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-8065e9e6-da92-4e73-8f4b-8b7e6cd0959b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.548291] env[65726]: DEBUG oslo_vmware.api [None req-881e6c53-2eb2-4fe8-9bf1-92c962897b3f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Waiting for the task: (returnval){ [ 655.548291] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]528160ad-6e66-ba25-2886-dfd191e394ea" [ 655.548291] env[65726]: _type = "Task" [ 655.548291] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 655.558196] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c30d3b57-2c21-40e7-9ac5-97057215c22c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.567437] env[65726]: DEBUG oslo_vmware.api [None req-881e6c53-2eb2-4fe8-9bf1-92c962897b3f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]528160ad-6e66-ba25-2886-dfd191e394ea, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 655.567599] env[65726]: INFO nova.virt.vmwareapi.vm_util [None req-cc552ef7-8733-47e0-9cfd-32896dfc9132 tempest-TenantUsagesTestJSON-327211524 tempest-TenantUsagesTestJSON-327211524-project-member] Created folder: Project (a2b0d29fd55f4116acb5331831eb23e9) in parent group-v995008. [ 655.568665] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-cc552ef7-8733-47e0-9cfd-32896dfc9132 tempest-TenantUsagesTestJSON-327211524 tempest-TenantUsagesTestJSON-327211524-project-member] Creating folder: Instances. Parent ref: group-v995056. {{(pid=65726) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 655.568665] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f28ac43c-0a26-4e42-a982-c94bd020a1d0 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.579659] env[65726]: DEBUG nova.compute.provider_tree [None req-362ec7a6-07fd-4c6b-b5fb-9865bdcb8c1a tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 655.586343] env[65726]: INFO nova.virt.vmwareapi.vm_util [None req-cc552ef7-8733-47e0-9cfd-32896dfc9132 tempest-TenantUsagesTestJSON-327211524 tempest-TenantUsagesTestJSON-327211524-project-member] Created folder: Instances in parent group-v995056. [ 655.586641] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-cc552ef7-8733-47e0-9cfd-32896dfc9132 tempest-TenantUsagesTestJSON-327211524 tempest-TenantUsagesTestJSON-327211524-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 655.587355] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7748c23f-9ea9-4172-8ab7-187678272bb3] Creating VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 655.588368] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-38aa4237-c3e3-4568-9de9-5a6eb196b796 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.605093] env[65726]: DEBUG nova.compute.manager [None req-b83663a4-2303-46f3-8690-72e6256a8051 tempest-ServerDiagnosticsTest-2021442419 tempest-ServerDiagnosticsTest-2021442419-project-member] [instance: de5cad99-1dbc-4435-98b4-987e24ba8c3c] Start destroying the instance on the hypervisor. {{(pid=65726) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 655.605599] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-b83663a4-2303-46f3-8690-72e6256a8051 tempest-ServerDiagnosticsTest-2021442419 tempest-ServerDiagnosticsTest-2021442419-project-member] [instance: de5cad99-1dbc-4435-98b4-987e24ba8c3c] Destroying instance {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 655.606215] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-737fde9a-44da-4981-8fa0-244a00f7d776 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.615765] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-b83663a4-2303-46f3-8690-72e6256a8051 tempest-ServerDiagnosticsTest-2021442419 tempest-ServerDiagnosticsTest-2021442419-project-member] [instance: de5cad99-1dbc-4435-98b4-987e24ba8c3c] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 655.617128] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-cf948572-090b-4617-b0e6-b6f92c601a26 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.619518] env[65726]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 655.619518] env[65726]: value = "task-5115513" [ 655.619518] env[65726]: _type = "Task" [ 655.619518] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 655.626580] env[65726]: WARNING openstack [req-21237459-e2d1-4f37-bebf-e40793b06175 req-69557957-f522-464c-bd7e-ffbde47a6481 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 655.627446] env[65726]: WARNING openstack [req-21237459-e2d1-4f37-bebf-e40793b06175 req-69557957-f522-464c-bd7e-ffbde47a6481 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 655.640350] env[65726]: DEBUG oslo_vmware.api [None req-b83663a4-2303-46f3-8690-72e6256a8051 tempest-ServerDiagnosticsTest-2021442419 tempest-ServerDiagnosticsTest-2021442419-project-member] Waiting for the task: (returnval){ [ 655.640350] env[65726]: value = "task-5115514" [ 655.640350] env[65726]: _type = "Task" [ 655.640350] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 655.648485] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115513, 'name': CreateVM_Task} progress is 6%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 655.655789] env[65726]: DEBUG oslo_vmware.api [None req-b83663a4-2303-46f3-8690-72e6256a8051 tempest-ServerDiagnosticsTest-2021442419 tempest-ServerDiagnosticsTest-2021442419-project-member] Task: {'id': task-5115514, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 655.747139] env[65726]: DEBUG nova.network.neutron [req-21237459-e2d1-4f37-bebf-e40793b06175 req-69557957-f522-464c-bd7e-ffbde47a6481 service nova] [instance: aa2716a9-2cf8-4e51-800d-1c8f98b2cf05] Updated VIF entry in instance network info cache for port c4f80985-65f1-478f-9944-afb3538efb48. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 655.747139] env[65726]: DEBUG nova.network.neutron [req-21237459-e2d1-4f37-bebf-e40793b06175 req-69557957-f522-464c-bd7e-ffbde47a6481 service nova] [instance: aa2716a9-2cf8-4e51-800d-1c8f98b2cf05] Updating instance_info_cache with network_info: [{"id": "c4f80985-65f1-478f-9944-afb3538efb48", "address": "fa:16:3e:35:5f:a5", "network": {"id": "da8b1180-a4d3-4444-b571-050c941634d4", "bridge": "br-int", "label": "tempest-VolumesAssistedSnapshotsTest-1753385467-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.234", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b01c77c483364e23b4a48b1e1c9c9d1c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c297fe21-cd0b-4226-813b-a65d2358d034", "external-id": "nsx-vlan-transportzone-98", "segmentation_id": 98, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc4f80985-65", "ovs_interfaceid": "c4f80985-65f1-478f-9944-afb3538efb48", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 655.762223] env[65726]: DEBUG nova.compute.manager [None req-40525054-994c-41ab-bed7-c877834cae18 tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] [instance: 3b933684-ac19-44b0-a49d-6af45501e166] Start spawning the instance on the hypervisor. {{(pid=65726) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 655.795978] env[65726]: DEBUG nova.virt.hardware [None req-40525054-994c-41ab-bed7-c877834cae18 tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T19:28:34Z,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1539f06fa3534e90acbc3a60c4b8bd3f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T19:28:35Z,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 655.796304] env[65726]: DEBUG nova.virt.hardware [None req-40525054-994c-41ab-bed7-c877834cae18 tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 655.796482] env[65726]: DEBUG nova.virt.hardware [None req-40525054-994c-41ab-bed7-c877834cae18 tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 655.796707] env[65726]: DEBUG nova.virt.hardware [None req-40525054-994c-41ab-bed7-c877834cae18 tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 655.796852] env[65726]: DEBUG nova.virt.hardware [None req-40525054-994c-41ab-bed7-c877834cae18 tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 655.797038] env[65726]: DEBUG nova.virt.hardware [None req-40525054-994c-41ab-bed7-c877834cae18 tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 655.797247] env[65726]: DEBUG nova.virt.hardware [None req-40525054-994c-41ab-bed7-c877834cae18 tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 655.797400] env[65726]: DEBUG nova.virt.hardware [None req-40525054-994c-41ab-bed7-c877834cae18 tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 655.797556] env[65726]: DEBUG nova.virt.hardware [None req-40525054-994c-41ab-bed7-c877834cae18 tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 655.797727] env[65726]: DEBUG nova.virt.hardware [None req-40525054-994c-41ab-bed7-c877834cae18 tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 655.797898] env[65726]: DEBUG nova.virt.hardware [None req-40525054-994c-41ab-bed7-c877834cae18 tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 655.799040] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f70b315b-043a-480d-8c71-413f9d4d8507 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.809249] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb176b35-eaba-4a7a-b9bc-e57a64164ed2 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.827065] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-40525054-994c-41ab-bed7-c877834cae18 tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] [instance: 3b933684-ac19-44b0-a49d-6af45501e166] Instance VIF info [] {{(pid=65726) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 655.833297] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-40525054-994c-41ab-bed7-c877834cae18 tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] Creating folder: Project (c18594ded60e455e8c1e2e90f3d5cc7f). Parent ref: group-v995008. {{(pid=65726) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 655.833717] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a594c0c8-4886-40fb-accb-141c2d89a901 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.848114] env[65726]: INFO nova.virt.vmwareapi.vm_util [None req-40525054-994c-41ab-bed7-c877834cae18 tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] Created folder: Project (c18594ded60e455e8c1e2e90f3d5cc7f) in parent group-v995008. [ 655.848363] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-40525054-994c-41ab-bed7-c877834cae18 tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] Creating folder: Instances. Parent ref: group-v995059. {{(pid=65726) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 655.848557] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c87f3d59-beca-437e-a98d-898f156e7699 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.862017] env[65726]: DEBUG oslo_vmware.api [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Task: {'id': task-5115509, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 655.885833] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5fd4688c-49eb-401e-afbd-116789f6a402 tempest-ServerDiagnosticsV248Test-347810772 tempest-ServerDiagnosticsV248Test-347810772-project-member] Acquiring lock "refresh_cache-4d4dc88b-0d6c-430c-ab4f-c81a56e4657b" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 655.886110] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5fd4688c-49eb-401e-afbd-116789f6a402 tempest-ServerDiagnosticsV248Test-347810772 tempest-ServerDiagnosticsV248Test-347810772-project-member] Acquired lock "refresh_cache-4d4dc88b-0d6c-430c-ab4f-c81a56e4657b" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 655.886339] env[65726]: DEBUG nova.network.neutron [None req-5fd4688c-49eb-401e-afbd-116789f6a402 tempest-ServerDiagnosticsV248Test-347810772 tempest-ServerDiagnosticsV248Test-347810772-project-member] [instance: 4d4dc88b-0d6c-430c-ab4f-c81a56e4657b] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 655.909320] env[65726]: INFO nova.virt.vmwareapi.vm_util [None req-40525054-994c-41ab-bed7-c877834cae18 tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] Created folder: Instances in parent group-v995059. [ 655.909787] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-40525054-994c-41ab-bed7-c877834cae18 tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 655.909950] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3b933684-ac19-44b0-a49d-6af45501e166] Creating VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 655.910120] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e584ddc9-1aed-4684-ba0c-edb23db6fdb5 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.930691] env[65726]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 655.930691] env[65726]: value = "task-5115517" [ 655.930691] env[65726]: _type = "Task" [ 655.930691] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 655.940425] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115517, 'name': CreateVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 655.968765] env[65726]: DEBUG oslo_vmware.api [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Task: {'id': task-5115510, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 656.062532] env[65726]: DEBUG oslo_vmware.api [None req-881e6c53-2eb2-4fe8-9bf1-92c962897b3f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]528160ad-6e66-ba25-2886-dfd191e394ea, 'name': SearchDatastore_Task, 'duration_secs': 0.029329} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 656.063845] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1dff56c7-1831-47f3-9ba2-88de5007a7dd {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.072647] env[65726]: DEBUG oslo_vmware.api [None req-881e6c53-2eb2-4fe8-9bf1-92c962897b3f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Waiting for the task: (returnval){ [ 656.072647] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5229088b-5f33-fc41-11e1-2e7a2b750407" [ 656.072647] env[65726]: _type = "Task" [ 656.072647] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 656.085544] env[65726]: DEBUG oslo_vmware.api [None req-881e6c53-2eb2-4fe8-9bf1-92c962897b3f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5229088b-5f33-fc41-11e1-2e7a2b750407, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 656.089423] env[65726]: DEBUG nova.scheduler.client.report [None req-362ec7a6-07fd-4c6b-b5fb-9865bdcb8c1a tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 656.137389] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115513, 'name': CreateVM_Task, 'duration_secs': 0.495772} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 656.137389] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7748c23f-9ea9-4172-8ab7-187678272bb3] Created VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 656.137389] env[65726]: WARNING openstack [None req-cc552ef7-8733-47e0-9cfd-32896dfc9132 tempest-TenantUsagesTestJSON-327211524 tempest-TenantUsagesTestJSON-327211524-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 656.137389] env[65726]: WARNING openstack [None req-cc552ef7-8733-47e0-9cfd-32896dfc9132 tempest-TenantUsagesTestJSON-327211524 tempest-TenantUsagesTestJSON-327211524-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 656.143687] env[65726]: DEBUG oslo_concurrency.lockutils [None req-cc552ef7-8733-47e0-9cfd-32896dfc9132 tempest-TenantUsagesTestJSON-327211524 tempest-TenantUsagesTestJSON-327211524-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 656.143959] env[65726]: DEBUG oslo_concurrency.lockutils [None req-cc552ef7-8733-47e0-9cfd-32896dfc9132 tempest-TenantUsagesTestJSON-327211524 tempest-TenantUsagesTestJSON-327211524-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 656.144341] env[65726]: DEBUG oslo_concurrency.lockutils [None req-cc552ef7-8733-47e0-9cfd-32896dfc9132 tempest-TenantUsagesTestJSON-327211524 tempest-TenantUsagesTestJSON-327211524-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 656.145670] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-26096060-b044-4aa4-9eb2-b03133bcb5d6 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.158325] env[65726]: DEBUG oslo_vmware.api [None req-cc552ef7-8733-47e0-9cfd-32896dfc9132 tempest-TenantUsagesTestJSON-327211524 tempest-TenantUsagesTestJSON-327211524-project-member] Waiting for the task: (returnval){ [ 656.158325] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]525860f1-792a-d440-587b-035f07ea7ec4" [ 656.158325] env[65726]: _type = "Task" [ 656.158325] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 656.163222] env[65726]: DEBUG oslo_vmware.api [None req-b83663a4-2303-46f3-8690-72e6256a8051 tempest-ServerDiagnosticsTest-2021442419 tempest-ServerDiagnosticsTest-2021442419-project-member] Task: {'id': task-5115514, 'name': PowerOffVM_Task, 'duration_secs': 0.293348} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 656.166170] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-b83663a4-2303-46f3-8690-72e6256a8051 tempest-ServerDiagnosticsTest-2021442419 tempest-ServerDiagnosticsTest-2021442419-project-member] [instance: de5cad99-1dbc-4435-98b4-987e24ba8c3c] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 656.166346] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-b83663a4-2303-46f3-8690-72e6256a8051 tempest-ServerDiagnosticsTest-2021442419 tempest-ServerDiagnosticsTest-2021442419-project-member] [instance: de5cad99-1dbc-4435-98b4-987e24ba8c3c] Unregistering the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 656.166643] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-dc6e0407-d5c6-45a2-b797-82204b27079c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.175062] env[65726]: DEBUG oslo_vmware.api [None req-cc552ef7-8733-47e0-9cfd-32896dfc9132 tempest-TenantUsagesTestJSON-327211524 tempest-TenantUsagesTestJSON-327211524-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]525860f1-792a-d440-587b-035f07ea7ec4, 'name': SearchDatastore_Task, 'duration_secs': 0.012795} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 656.175375] env[65726]: DEBUG oslo_concurrency.lockutils [None req-cc552ef7-8733-47e0-9cfd-32896dfc9132 tempest-TenantUsagesTestJSON-327211524 tempest-TenantUsagesTestJSON-327211524-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 656.175594] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-cc552ef7-8733-47e0-9cfd-32896dfc9132 tempest-TenantUsagesTestJSON-327211524 tempest-TenantUsagesTestJSON-327211524-project-member] [instance: 7748c23f-9ea9-4172-8ab7-187678272bb3] Processing image b52362a3-ee8a-4cbf-b06f-513b0cc8f95c {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 656.175799] env[65726]: DEBUG oslo_concurrency.lockutils [None req-cc552ef7-8733-47e0-9cfd-32896dfc9132 tempest-TenantUsagesTestJSON-327211524 tempest-TenantUsagesTestJSON-327211524-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 656.249032] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-b83663a4-2303-46f3-8690-72e6256a8051 tempest-ServerDiagnosticsTest-2021442419 tempest-ServerDiagnosticsTest-2021442419-project-member] [instance: de5cad99-1dbc-4435-98b4-987e24ba8c3c] Unregistered the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 656.249032] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-b83663a4-2303-46f3-8690-72e6256a8051 tempest-ServerDiagnosticsTest-2021442419 tempest-ServerDiagnosticsTest-2021442419-project-member] [instance: de5cad99-1dbc-4435-98b4-987e24ba8c3c] Deleting contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 656.249032] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-b83663a4-2303-46f3-8690-72e6256a8051 tempest-ServerDiagnosticsTest-2021442419 tempest-ServerDiagnosticsTest-2021442419-project-member] Deleting the datastore file [datastore1] de5cad99-1dbc-4435-98b4-987e24ba8c3c {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 656.249032] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7ac984c3-242c-491b-be4b-99a468a0fa4f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.251788] env[65726]: DEBUG oslo_concurrency.lockutils [req-21237459-e2d1-4f37-bebf-e40793b06175 req-69557957-f522-464c-bd7e-ffbde47a6481 service nova] Releasing lock "refresh_cache-aa2716a9-2cf8-4e51-800d-1c8f98b2cf05" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 656.252513] env[65726]: DEBUG nova.compute.manager [req-21237459-e2d1-4f37-bebf-e40793b06175 req-69557957-f522-464c-bd7e-ffbde47a6481 service nova] [instance: f2d9090c-988f-43f4-9c81-7aa718a3438a] Received event network-vif-plugged-7117cd21-ee2a-497d-b789-65e43b068258 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 656.253888] env[65726]: DEBUG oslo_concurrency.lockutils [req-21237459-e2d1-4f37-bebf-e40793b06175 req-69557957-f522-464c-bd7e-ffbde47a6481 service nova] Acquiring lock "f2d9090c-988f-43f4-9c81-7aa718a3438a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 656.253888] env[65726]: DEBUG oslo_concurrency.lockutils [req-21237459-e2d1-4f37-bebf-e40793b06175 req-69557957-f522-464c-bd7e-ffbde47a6481 service nova] Lock "f2d9090c-988f-43f4-9c81-7aa718a3438a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 656.253888] env[65726]: DEBUG oslo_concurrency.lockutils [req-21237459-e2d1-4f37-bebf-e40793b06175 req-69557957-f522-464c-bd7e-ffbde47a6481 service nova] Lock "f2d9090c-988f-43f4-9c81-7aa718a3438a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 656.256834] env[65726]: DEBUG nova.compute.manager [req-21237459-e2d1-4f37-bebf-e40793b06175 req-69557957-f522-464c-bd7e-ffbde47a6481 service nova] [instance: f2d9090c-988f-43f4-9c81-7aa718a3438a] No waiting events found dispatching network-vif-plugged-7117cd21-ee2a-497d-b789-65e43b068258 {{(pid=65726) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 656.256834] env[65726]: WARNING nova.compute.manager [req-21237459-e2d1-4f37-bebf-e40793b06175 req-69557957-f522-464c-bd7e-ffbde47a6481 service nova] [instance: f2d9090c-988f-43f4-9c81-7aa718a3438a] Received unexpected event network-vif-plugged-7117cd21-ee2a-497d-b789-65e43b068258 for instance with vm_state building and task_state spawning. [ 656.256834] env[65726]: DEBUG nova.compute.manager [req-21237459-e2d1-4f37-bebf-e40793b06175 req-69557957-f522-464c-bd7e-ffbde47a6481 service nova] [instance: f2d9090c-988f-43f4-9c81-7aa718a3438a] Received event network-changed-7117cd21-ee2a-497d-b789-65e43b068258 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 656.256834] env[65726]: DEBUG nova.compute.manager [req-21237459-e2d1-4f37-bebf-e40793b06175 req-69557957-f522-464c-bd7e-ffbde47a6481 service nova] [instance: f2d9090c-988f-43f4-9c81-7aa718a3438a] Refreshing instance network info cache due to event network-changed-7117cd21-ee2a-497d-b789-65e43b068258. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 656.256834] env[65726]: DEBUG oslo_concurrency.lockutils [req-21237459-e2d1-4f37-bebf-e40793b06175 req-69557957-f522-464c-bd7e-ffbde47a6481 service nova] Acquiring lock "refresh_cache-f2d9090c-988f-43f4-9c81-7aa718a3438a" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 656.258673] env[65726]: DEBUG oslo_concurrency.lockutils [req-21237459-e2d1-4f37-bebf-e40793b06175 req-69557957-f522-464c-bd7e-ffbde47a6481 service nova] Acquired lock "refresh_cache-f2d9090c-988f-43f4-9c81-7aa718a3438a" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 656.258673] env[65726]: DEBUG nova.network.neutron [req-21237459-e2d1-4f37-bebf-e40793b06175 req-69557957-f522-464c-bd7e-ffbde47a6481 service nova] [instance: f2d9090c-988f-43f4-9c81-7aa718a3438a] Refreshing network info cache for port 7117cd21-ee2a-497d-b789-65e43b068258 {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 656.270388] env[65726]: DEBUG oslo_vmware.api [None req-b83663a4-2303-46f3-8690-72e6256a8051 tempest-ServerDiagnosticsTest-2021442419 tempest-ServerDiagnosticsTest-2021442419-project-member] Waiting for the task: (returnval){ [ 656.270388] env[65726]: value = "task-5115519" [ 656.270388] env[65726]: _type = "Task" [ 656.270388] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 656.285751] env[65726]: DEBUG oslo_vmware.api [None req-b83663a4-2303-46f3-8690-72e6256a8051 tempest-ServerDiagnosticsTest-2021442419 tempest-ServerDiagnosticsTest-2021442419-project-member] Task: {'id': task-5115519, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 656.367173] env[65726]: DEBUG oslo_vmware.api [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Task: {'id': task-5115509, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 656.392449] env[65726]: WARNING openstack [None req-5fd4688c-49eb-401e-afbd-116789f6a402 tempest-ServerDiagnosticsV248Test-347810772 tempest-ServerDiagnosticsV248Test-347810772-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 656.392872] env[65726]: WARNING openstack [None req-5fd4688c-49eb-401e-afbd-116789f6a402 tempest-ServerDiagnosticsV248Test-347810772 tempest-ServerDiagnosticsV248Test-347810772-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 656.398327] env[65726]: DEBUG nova.network.neutron [None req-5fd4688c-49eb-401e-afbd-116789f6a402 tempest-ServerDiagnosticsV248Test-347810772 tempest-ServerDiagnosticsV248Test-347810772-project-member] [instance: 4d4dc88b-0d6c-430c-ab4f-c81a56e4657b] Instance cache missing network info. {{(pid=65726) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 656.445823] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115517, 'name': CreateVM_Task} progress is 99%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 656.470110] env[65726]: DEBUG oslo_vmware.api [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Task: {'id': task-5115510, 'name': ReconfigVM_Task, 'duration_secs': 0.70827} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 656.470411] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] [instance: f5151062-57b1-4e4e-93f4-aab0e7f504d4] Reconfigured VM instance instance-00000010 to attach disk [datastore2] f5151062-57b1-4e4e-93f4-aab0e7f504d4/f5151062-57b1-4e4e-93f4-aab0e7f504d4.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 656.471249] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3cb002f2-98f9-4b3b-8da6-6446b14e92b5 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.483444] env[65726]: DEBUG oslo_vmware.api [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Waiting for the task: (returnval){ [ 656.483444] env[65726]: value = "task-5115520" [ 656.483444] env[65726]: _type = "Task" [ 656.483444] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 656.494407] env[65726]: DEBUG oslo_vmware.api [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Task: {'id': task-5115520, 'name': Rename_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 656.530930] env[65726]: DEBUG nova.network.neutron [None req-5fd4688c-49eb-401e-afbd-116789f6a402 tempest-ServerDiagnosticsV248Test-347810772 tempest-ServerDiagnosticsV248Test-347810772-project-member] [instance: 4d4dc88b-0d6c-430c-ab4f-c81a56e4657b] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 656.588477] env[65726]: DEBUG oslo_vmware.api [None req-881e6c53-2eb2-4fe8-9bf1-92c962897b3f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5229088b-5f33-fc41-11e1-2e7a2b750407, 'name': SearchDatastore_Task, 'duration_secs': 0.012308} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 656.588477] env[65726]: DEBUG oslo_concurrency.lockutils [None req-881e6c53-2eb2-4fe8-9bf1-92c962897b3f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 656.588477] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-881e6c53-2eb2-4fe8-9bf1-92c962897b3f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore2] f2d9090c-988f-43f4-9c81-7aa718a3438a/f2d9090c-988f-43f4-9c81-7aa718a3438a.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 656.588477] env[65726]: DEBUG oslo_concurrency.lockutils [None req-cc552ef7-8733-47e0-9cfd-32896dfc9132 tempest-TenantUsagesTestJSON-327211524 tempest-TenantUsagesTestJSON-327211524-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 656.588848] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-cc552ef7-8733-47e0-9cfd-32896dfc9132 tempest-TenantUsagesTestJSON-327211524 tempest-TenantUsagesTestJSON-327211524-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 656.588848] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-14fd8ffd-1e61-4920-93ce-fa3f34d87899 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.591901] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3f0a1a91-f5a5-4086-8747-a10160d16959 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.596914] env[65726]: DEBUG oslo_concurrency.lockutils [None req-362ec7a6-07fd-4c6b-b5fb-9865bdcb8c1a tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.854s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 656.597133] env[65726]: DEBUG nova.compute.manager [None req-362ec7a6-07fd-4c6b-b5fb-9865bdcb8c1a tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: 274ab469-61a9-4b7e-852c-074c871e3abf] Start building networks asynchronously for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 656.601740] env[65726]: DEBUG oslo_concurrency.lockutils [None req-621af3d8-d8b6-4b45-86f5-d66b91010e2e tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 10.535s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 656.604039] env[65726]: INFO nova.compute.claims [None req-621af3d8-d8b6-4b45-86f5-d66b91010e2e tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] [instance: fd481728-3ef0-4a96-affd-ab2dd3f596bb] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 656.607593] env[65726]: DEBUG oslo_vmware.api [None req-881e6c53-2eb2-4fe8-9bf1-92c962897b3f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Waiting for the task: (returnval){ [ 656.607593] env[65726]: value = "task-5115521" [ 656.607593] env[65726]: _type = "Task" [ 656.607593] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 656.610229] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-cc552ef7-8733-47e0-9cfd-32896dfc9132 tempest-TenantUsagesTestJSON-327211524 tempest-TenantUsagesTestJSON-327211524-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 656.610484] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-cc552ef7-8733-47e0-9cfd-32896dfc9132 tempest-TenantUsagesTestJSON-327211524 tempest-TenantUsagesTestJSON-327211524-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65726) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 656.615021] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8d55015e-bea4-4977-b455-1e663061b6dd {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.625165] env[65726]: DEBUG oslo_vmware.api [None req-cc552ef7-8733-47e0-9cfd-32896dfc9132 tempest-TenantUsagesTestJSON-327211524 tempest-TenantUsagesTestJSON-327211524-project-member] Waiting for the task: (returnval){ [ 656.625165] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]525f59b7-5363-b4b5-3f6d-48f5f76b66e5" [ 656.625165] env[65726]: _type = "Task" [ 656.625165] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 656.630202] env[65726]: DEBUG oslo_vmware.api [None req-881e6c53-2eb2-4fe8-9bf1-92c962897b3f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Task: {'id': task-5115521, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 656.643636] env[65726]: DEBUG oslo_vmware.api [None req-cc552ef7-8733-47e0-9cfd-32896dfc9132 tempest-TenantUsagesTestJSON-327211524 tempest-TenantUsagesTestJSON-327211524-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]525f59b7-5363-b4b5-3f6d-48f5f76b66e5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 656.759293] env[65726]: WARNING openstack [req-21237459-e2d1-4f37-bebf-e40793b06175 req-69557957-f522-464c-bd7e-ffbde47a6481 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 656.759683] env[65726]: WARNING openstack [req-21237459-e2d1-4f37-bebf-e40793b06175 req-69557957-f522-464c-bd7e-ffbde47a6481 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 656.781558] env[65726]: DEBUG oslo_vmware.api [None req-b83663a4-2303-46f3-8690-72e6256a8051 tempest-ServerDiagnosticsTest-2021442419 tempest-ServerDiagnosticsTest-2021442419-project-member] Task: {'id': task-5115519, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.413587} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 656.781974] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-b83663a4-2303-46f3-8690-72e6256a8051 tempest-ServerDiagnosticsTest-2021442419 tempest-ServerDiagnosticsTest-2021442419-project-member] Deleted the datastore file {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 656.782076] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-b83663a4-2303-46f3-8690-72e6256a8051 tempest-ServerDiagnosticsTest-2021442419 tempest-ServerDiagnosticsTest-2021442419-project-member] [instance: de5cad99-1dbc-4435-98b4-987e24ba8c3c] Deleted contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 656.782492] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-b83663a4-2303-46f3-8690-72e6256a8051 tempest-ServerDiagnosticsTest-2021442419 tempest-ServerDiagnosticsTest-2021442419-project-member] [instance: de5cad99-1dbc-4435-98b4-987e24ba8c3c] Instance destroyed {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 656.782492] env[65726]: INFO nova.compute.manager [None req-b83663a4-2303-46f3-8690-72e6256a8051 tempest-ServerDiagnosticsTest-2021442419 tempest-ServerDiagnosticsTest-2021442419-project-member] [instance: de5cad99-1dbc-4435-98b4-987e24ba8c3c] Took 1.18 seconds to destroy the instance on the hypervisor. [ 656.782688] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-b83663a4-2303-46f3-8690-72e6256a8051 tempest-ServerDiagnosticsTest-2021442419 tempest-ServerDiagnosticsTest-2021442419-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 656.782907] env[65726]: DEBUG nova.compute.manager [-] [instance: de5cad99-1dbc-4435-98b4-987e24ba8c3c] Deallocating network for instance {{(pid=65726) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 656.782907] env[65726]: DEBUG nova.network.neutron [-] [instance: de5cad99-1dbc-4435-98b4-987e24ba8c3c] deallocate_for_instance() {{(pid=65726) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 656.783442] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 656.783698] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 656.859409] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 656.859753] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 656.875436] env[65726]: DEBUG oslo_vmware.api [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Task: {'id': task-5115509, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 656.926347] env[65726]: WARNING openstack [req-21237459-e2d1-4f37-bebf-e40793b06175 req-69557957-f522-464c-bd7e-ffbde47a6481 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 656.926776] env[65726]: WARNING openstack [req-21237459-e2d1-4f37-bebf-e40793b06175 req-69557957-f522-464c-bd7e-ffbde47a6481 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 656.951464] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115517, 'name': CreateVM_Task} progress is 99%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 656.994870] env[65726]: DEBUG oslo_vmware.api [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Task: {'id': task-5115520, 'name': Rename_Task, 'duration_secs': 0.232803} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 656.995166] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] [instance: f5151062-57b1-4e4e-93f4-aab0e7f504d4] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 656.995452] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9ead63f8-c9cd-437b-9cb9-8971396e0657 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.004449] env[65726]: DEBUG oslo_vmware.api [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Waiting for the task: (returnval){ [ 657.004449] env[65726]: value = "task-5115522" [ 657.004449] env[65726]: _type = "Task" [ 657.004449] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 657.016561] env[65726]: DEBUG oslo_vmware.api [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Task: {'id': task-5115522, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 657.034979] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5fd4688c-49eb-401e-afbd-116789f6a402 tempest-ServerDiagnosticsV248Test-347810772 tempest-ServerDiagnosticsV248Test-347810772-project-member] Releasing lock "refresh_cache-4d4dc88b-0d6c-430c-ab4f-c81a56e4657b" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 657.035936] env[65726]: DEBUG nova.compute.manager [None req-5fd4688c-49eb-401e-afbd-116789f6a402 tempest-ServerDiagnosticsV248Test-347810772 tempest-ServerDiagnosticsV248Test-347810772-project-member] [instance: 4d4dc88b-0d6c-430c-ab4f-c81a56e4657b] Start destroying the instance on the hypervisor. {{(pid=65726) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 657.035936] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-5fd4688c-49eb-401e-afbd-116789f6a402 tempest-ServerDiagnosticsV248Test-347810772 tempest-ServerDiagnosticsV248Test-347810772-project-member] [instance: 4d4dc88b-0d6c-430c-ab4f-c81a56e4657b] Destroying instance {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 657.038663] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f933a06e-32c8-4371-9047-9f424b27fe97 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.050234] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-5fd4688c-49eb-401e-afbd-116789f6a402 tempest-ServerDiagnosticsV248Test-347810772 tempest-ServerDiagnosticsV248Test-347810772-project-member] [instance: 4d4dc88b-0d6c-430c-ab4f-c81a56e4657b] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 657.050419] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1a07ba42-e6c4-4d18-af2b-1be89fb7c9a0 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.063080] env[65726]: DEBUG oslo_vmware.api [None req-5fd4688c-49eb-401e-afbd-116789f6a402 tempest-ServerDiagnosticsV248Test-347810772 tempest-ServerDiagnosticsV248Test-347810772-project-member] Waiting for the task: (returnval){ [ 657.063080] env[65726]: value = "task-5115523" [ 657.063080] env[65726]: _type = "Task" [ 657.063080] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 657.083332] env[65726]: DEBUG oslo_vmware.api [None req-5fd4688c-49eb-401e-afbd-116789f6a402 tempest-ServerDiagnosticsV248Test-347810772 tempest-ServerDiagnosticsV248Test-347810772-project-member] Task: {'id': task-5115523, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 657.113493] env[65726]: DEBUG nova.compute.utils [None req-362ec7a6-07fd-4c6b-b5fb-9865bdcb8c1a tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Using /dev/sd instead of None {{(pid=65726) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 657.120376] env[65726]: DEBUG nova.compute.manager [None req-362ec7a6-07fd-4c6b-b5fb-9865bdcb8c1a tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: 274ab469-61a9-4b7e-852c-074c871e3abf] Allocating IP information in the background. {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 657.120507] env[65726]: DEBUG nova.network.neutron [None req-362ec7a6-07fd-4c6b-b5fb-9865bdcb8c1a tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: 274ab469-61a9-4b7e-852c-074c871e3abf] allocate_for_instance() {{(pid=65726) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 657.121050] env[65726]: WARNING neutronclient.v2_0.client [None req-362ec7a6-07fd-4c6b-b5fb-9865bdcb8c1a tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 657.121590] env[65726]: WARNING neutronclient.v2_0.client [None req-362ec7a6-07fd-4c6b-b5fb-9865bdcb8c1a tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 657.123038] env[65726]: WARNING openstack [None req-362ec7a6-07fd-4c6b-b5fb-9865bdcb8c1a tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 657.125167] env[65726]: WARNING openstack [None req-362ec7a6-07fd-4c6b-b5fb-9865bdcb8c1a tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 657.169035] env[65726]: DEBUG oslo_vmware.api [None req-cc552ef7-8733-47e0-9cfd-32896dfc9132 tempest-TenantUsagesTestJSON-327211524 tempest-TenantUsagesTestJSON-327211524-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]525f59b7-5363-b4b5-3f6d-48f5f76b66e5, 'name': SearchDatastore_Task, 'duration_secs': 0.029909} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 657.176470] env[65726]: DEBUG oslo_vmware.api [None req-881e6c53-2eb2-4fe8-9bf1-92c962897b3f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Task: {'id': task-5115521, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 657.177165] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cbf2724d-0b9d-4c1d-be30-a85efda1eda1 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.186991] env[65726]: DEBUG oslo_vmware.api [None req-cc552ef7-8733-47e0-9cfd-32896dfc9132 tempest-TenantUsagesTestJSON-327211524 tempest-TenantUsagesTestJSON-327211524-project-member] Waiting for the task: (returnval){ [ 657.186991] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]525c9223-4cb2-cb6e-784d-350b55e2c729" [ 657.186991] env[65726]: _type = "Task" [ 657.186991] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 657.203404] env[65726]: DEBUG oslo_vmware.api [None req-cc552ef7-8733-47e0-9cfd-32896dfc9132 tempest-TenantUsagesTestJSON-327211524 tempest-TenantUsagesTestJSON-327211524-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]525c9223-4cb2-cb6e-784d-350b55e2c729, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 657.214856] env[65726]: WARNING openstack [req-21237459-e2d1-4f37-bebf-e40793b06175 req-69557957-f522-464c-bd7e-ffbde47a6481 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 657.215378] env[65726]: WARNING openstack [req-21237459-e2d1-4f37-bebf-e40793b06175 req-69557957-f522-464c-bd7e-ffbde47a6481 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 657.269127] env[65726]: DEBUG nova.policy [None req-362ec7a6-07fd-4c6b-b5fb-9865bdcb8c1a tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4b6c9cc4ac3c487693240e375a83ea19', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b794c99309e94c1181fc9d8eeeb84702', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65726) authorize /opt/stack/nova/nova/policy.py:192}} [ 657.367161] env[65726]: DEBUG oslo_vmware.api [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Task: {'id': task-5115509, 'name': PowerOnVM_Task, 'duration_secs': 1.860475} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 657.367582] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] [instance: 3a60b719-9db7-436e-9908-25f6a233c465] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 657.367847] env[65726]: INFO nova.compute.manager [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] [instance: 3a60b719-9db7-436e-9908-25f6a233c465] Took 12.78 seconds to spawn the instance on the hypervisor. [ 657.368092] env[65726]: DEBUG nova.compute.manager [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] [instance: 3a60b719-9db7-436e-9908-25f6a233c465] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 657.369045] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b06efba4-c0d1-4b54-b88e-9a2fbf4f2e03 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.420397] env[65726]: DEBUG nova.network.neutron [req-21237459-e2d1-4f37-bebf-e40793b06175 req-69557957-f522-464c-bd7e-ffbde47a6481 service nova] [instance: f2d9090c-988f-43f4-9c81-7aa718a3438a] Updated VIF entry in instance network info cache for port 7117cd21-ee2a-497d-b789-65e43b068258. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 657.420397] env[65726]: DEBUG nova.network.neutron [req-21237459-e2d1-4f37-bebf-e40793b06175 req-69557957-f522-464c-bd7e-ffbde47a6481 service nova] [instance: f2d9090c-988f-43f4-9c81-7aa718a3438a] Updating instance_info_cache with network_info: [{"id": "7117cd21-ee2a-497d-b789-65e43b068258", "address": "fa:16:3e:10:2a:24", "network": {"id": "3f7916f7-ee9c-4629-b3b3-1367de9f98ea", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-579713617-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b794c99309e94c1181fc9d8eeeb84702", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d0bdd11b-58af-4cc0-9d38-8322e1bb4e74", "external-id": "nsx-vlan-transportzone-398", "segmentation_id": 398, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7117cd21-ee", "ovs_interfaceid": "7117cd21-ee2a-497d-b789-65e43b068258", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 657.454331] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115517, 'name': CreateVM_Task, 'duration_secs': 1.31879} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 657.454331] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3b933684-ac19-44b0-a49d-6af45501e166] Created VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 657.454331] env[65726]: DEBUG oslo_concurrency.lockutils [None req-40525054-994c-41ab-bed7-c877834cae18 tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 657.454331] env[65726]: DEBUG oslo_concurrency.lockutils [None req-40525054-994c-41ab-bed7-c877834cae18 tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 657.454331] env[65726]: DEBUG oslo_concurrency.lockutils [None req-40525054-994c-41ab-bed7-c877834cae18 tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 657.454571] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8d3a5c09-7d4d-4010-abce-1886bc6b18c9 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.461761] env[65726]: DEBUG oslo_vmware.api [None req-40525054-994c-41ab-bed7-c877834cae18 tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] Waiting for the task: (returnval){ [ 657.461761] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52423c66-2d34-a7bc-9103-e8b255733749" [ 657.461761] env[65726]: _type = "Task" [ 657.461761] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 657.473539] env[65726]: DEBUG oslo_vmware.api [None req-40525054-994c-41ab-bed7-c877834cae18 tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52423c66-2d34-a7bc-9103-e8b255733749, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 657.517323] env[65726]: DEBUG oslo_vmware.api [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Task: {'id': task-5115522, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 657.585466] env[65726]: DEBUG oslo_vmware.api [None req-5fd4688c-49eb-401e-afbd-116789f6a402 tempest-ServerDiagnosticsV248Test-347810772 tempest-ServerDiagnosticsV248Test-347810772-project-member] Task: {'id': task-5115523, 'name': PowerOffVM_Task, 'duration_secs': 0.457757} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 657.585817] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-5fd4688c-49eb-401e-afbd-116789f6a402 tempest-ServerDiagnosticsV248Test-347810772 tempest-ServerDiagnosticsV248Test-347810772-project-member] [instance: 4d4dc88b-0d6c-430c-ab4f-c81a56e4657b] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 657.586016] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-5fd4688c-49eb-401e-afbd-116789f6a402 tempest-ServerDiagnosticsV248Test-347810772 tempest-ServerDiagnosticsV248Test-347810772-project-member] [instance: 4d4dc88b-0d6c-430c-ab4f-c81a56e4657b] Unregistering the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 657.586293] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-087e5d94-6777-4d21-8f18-a8ab1cba7ae7 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.616907] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-5fd4688c-49eb-401e-afbd-116789f6a402 tempest-ServerDiagnosticsV248Test-347810772 tempest-ServerDiagnosticsV248Test-347810772-project-member] [instance: 4d4dc88b-0d6c-430c-ab4f-c81a56e4657b] Unregistered the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 657.617167] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-5fd4688c-49eb-401e-afbd-116789f6a402 tempest-ServerDiagnosticsV248Test-347810772 tempest-ServerDiagnosticsV248Test-347810772-project-member] [instance: 4d4dc88b-0d6c-430c-ab4f-c81a56e4657b] Deleting contents of the VM from datastore datastore2 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 657.617354] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-5fd4688c-49eb-401e-afbd-116789f6a402 tempest-ServerDiagnosticsV248Test-347810772 tempest-ServerDiagnosticsV248Test-347810772-project-member] Deleting the datastore file [datastore2] 4d4dc88b-0d6c-430c-ab4f-c81a56e4657b {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 657.617661] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ba1b8444-83c2-40ef-98ba-8f03ca9adfb9 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.620464] env[65726]: DEBUG nova.compute.manager [None req-362ec7a6-07fd-4c6b-b5fb-9865bdcb8c1a tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: 274ab469-61a9-4b7e-852c-074c871e3abf] Start building block device mappings for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 657.635121] env[65726]: DEBUG oslo_vmware.api [None req-881e6c53-2eb2-4fe8-9bf1-92c962897b3f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Task: {'id': task-5115521, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.832492} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 657.636675] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-881e6c53-2eb2-4fe8-9bf1-92c962897b3f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore2] f2d9090c-988f-43f4-9c81-7aa718a3438a/f2d9090c-988f-43f4-9c81-7aa718a3438a.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 657.636984] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-881e6c53-2eb2-4fe8-9bf1-92c962897b3f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: f2d9090c-988f-43f4-9c81-7aa718a3438a] Extending root virtual disk to 1048576 {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 657.637346] env[65726]: DEBUG oslo_vmware.api [None req-5fd4688c-49eb-401e-afbd-116789f6a402 tempest-ServerDiagnosticsV248Test-347810772 tempest-ServerDiagnosticsV248Test-347810772-project-member] Waiting for the task: (returnval){ [ 657.637346] env[65726]: value = "task-5115525" [ 657.637346] env[65726]: _type = "Task" [ 657.637346] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 657.637538] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5b2232a8-8b18-43f6-b56f-10914cfdfa06 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.655084] env[65726]: DEBUG oslo_vmware.api [None req-5fd4688c-49eb-401e-afbd-116789f6a402 tempest-ServerDiagnosticsV248Test-347810772 tempest-ServerDiagnosticsV248Test-347810772-project-member] Task: {'id': task-5115525, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 657.658021] env[65726]: DEBUG oslo_vmware.api [None req-881e6c53-2eb2-4fe8-9bf1-92c962897b3f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Waiting for the task: (returnval){ [ 657.658021] env[65726]: value = "task-5115526" [ 657.658021] env[65726]: _type = "Task" [ 657.658021] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 657.667549] env[65726]: DEBUG oslo_vmware.api [None req-881e6c53-2eb2-4fe8-9bf1-92c962897b3f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Task: {'id': task-5115526, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 657.678013] env[65726]: DEBUG nova.network.neutron [-] [instance: de5cad99-1dbc-4435-98b4-987e24ba8c3c] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 657.688190] env[65726]: DEBUG nova.network.neutron [None req-362ec7a6-07fd-4c6b-b5fb-9865bdcb8c1a tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: 274ab469-61a9-4b7e-852c-074c871e3abf] Successfully created port: 757d86b3-4474-41c5-aaf8-14fc7216cf15 {{(pid=65726) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 657.704793] env[65726]: DEBUG oslo_vmware.api [None req-cc552ef7-8733-47e0-9cfd-32896dfc9132 tempest-TenantUsagesTestJSON-327211524 tempest-TenantUsagesTestJSON-327211524-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]525c9223-4cb2-cb6e-784d-350b55e2c729, 'name': SearchDatastore_Task, 'duration_secs': 0.06871} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 657.705127] env[65726]: DEBUG oslo_concurrency.lockutils [None req-cc552ef7-8733-47e0-9cfd-32896dfc9132 tempest-TenantUsagesTestJSON-327211524 tempest-TenantUsagesTestJSON-327211524-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 657.705604] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-cc552ef7-8733-47e0-9cfd-32896dfc9132 tempest-TenantUsagesTestJSON-327211524 tempest-TenantUsagesTestJSON-327211524-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore2] 7748c23f-9ea9-4172-8ab7-187678272bb3/7748c23f-9ea9-4172-8ab7-187678272bb3.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 657.705799] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c0eb2acf-7c75-4acf-9f15-3da81095f928 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.716231] env[65726]: DEBUG oslo_vmware.api [None req-cc552ef7-8733-47e0-9cfd-32896dfc9132 tempest-TenantUsagesTestJSON-327211524 tempest-TenantUsagesTestJSON-327211524-project-member] Waiting for the task: (returnval){ [ 657.716231] env[65726]: value = "task-5115527" [ 657.716231] env[65726]: _type = "Task" [ 657.716231] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 657.728469] env[65726]: DEBUG oslo_vmware.api [None req-cc552ef7-8733-47e0-9cfd-32896dfc9132 tempest-TenantUsagesTestJSON-327211524 tempest-TenantUsagesTestJSON-327211524-project-member] Task: {'id': task-5115527, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 657.815641] env[65726]: DEBUG nova.compute.manager [req-ca6fc5f3-49e6-44cf-8dae-7e569705410e req-b86824fb-b135-4333-9c5a-e3ab63dad036 service nova] [instance: 7748c23f-9ea9-4172-8ab7-187678272bb3] Received event network-vif-plugged-8bb2d672-e56e-4f01-bfbf-74a279d9da7d {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 657.815849] env[65726]: DEBUG oslo_concurrency.lockutils [req-ca6fc5f3-49e6-44cf-8dae-7e569705410e req-b86824fb-b135-4333-9c5a-e3ab63dad036 service nova] Acquiring lock "7748c23f-9ea9-4172-8ab7-187678272bb3-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 657.816054] env[65726]: DEBUG oslo_concurrency.lockutils [req-ca6fc5f3-49e6-44cf-8dae-7e569705410e req-b86824fb-b135-4333-9c5a-e3ab63dad036 service nova] Lock "7748c23f-9ea9-4172-8ab7-187678272bb3-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 657.816206] env[65726]: DEBUG oslo_concurrency.lockutils [req-ca6fc5f3-49e6-44cf-8dae-7e569705410e req-b86824fb-b135-4333-9c5a-e3ab63dad036 service nova] Lock "7748c23f-9ea9-4172-8ab7-187678272bb3-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 657.816359] env[65726]: DEBUG nova.compute.manager [req-ca6fc5f3-49e6-44cf-8dae-7e569705410e req-b86824fb-b135-4333-9c5a-e3ab63dad036 service nova] [instance: 7748c23f-9ea9-4172-8ab7-187678272bb3] No waiting events found dispatching network-vif-plugged-8bb2d672-e56e-4f01-bfbf-74a279d9da7d {{(pid=65726) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 657.816508] env[65726]: WARNING nova.compute.manager [req-ca6fc5f3-49e6-44cf-8dae-7e569705410e req-b86824fb-b135-4333-9c5a-e3ab63dad036 service nova] [instance: 7748c23f-9ea9-4172-8ab7-187678272bb3] Received unexpected event network-vif-plugged-8bb2d672-e56e-4f01-bfbf-74a279d9da7d for instance with vm_state building and task_state spawning. [ 657.816812] env[65726]: DEBUG nova.compute.manager [req-ca6fc5f3-49e6-44cf-8dae-7e569705410e req-b86824fb-b135-4333-9c5a-e3ab63dad036 service nova] [instance: 7748c23f-9ea9-4172-8ab7-187678272bb3] Received event network-changed-8bb2d672-e56e-4f01-bfbf-74a279d9da7d {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 657.816970] env[65726]: DEBUG nova.compute.manager [req-ca6fc5f3-49e6-44cf-8dae-7e569705410e req-b86824fb-b135-4333-9c5a-e3ab63dad036 service nova] [instance: 7748c23f-9ea9-4172-8ab7-187678272bb3] Refreshing instance network info cache due to event network-changed-8bb2d672-e56e-4f01-bfbf-74a279d9da7d. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 657.817161] env[65726]: DEBUG oslo_concurrency.lockutils [req-ca6fc5f3-49e6-44cf-8dae-7e569705410e req-b86824fb-b135-4333-9c5a-e3ab63dad036 service nova] Acquiring lock "refresh_cache-7748c23f-9ea9-4172-8ab7-187678272bb3" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 657.817290] env[65726]: DEBUG oslo_concurrency.lockutils [req-ca6fc5f3-49e6-44cf-8dae-7e569705410e req-b86824fb-b135-4333-9c5a-e3ab63dad036 service nova] Acquired lock "refresh_cache-7748c23f-9ea9-4172-8ab7-187678272bb3" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 657.818030] env[65726]: DEBUG nova.network.neutron [req-ca6fc5f3-49e6-44cf-8dae-7e569705410e req-b86824fb-b135-4333-9c5a-e3ab63dad036 service nova] [instance: 7748c23f-9ea9-4172-8ab7-187678272bb3] Refreshing network info cache for port 8bb2d672-e56e-4f01-bfbf-74a279d9da7d {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 657.896678] env[65726]: INFO nova.compute.manager [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] [instance: 3a60b719-9db7-436e-9908-25f6a233c465] Took 28.61 seconds to build instance. [ 657.914267] env[65726]: DEBUG oslo_concurrency.lockutils [None req-f72e4138-ba89-47ec-80a1-52d74fe64f4e tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Acquiring lock "8e9c27f6-e9fa-4ed0-b6ae-3b5df52d1c44" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 657.914267] env[65726]: DEBUG oslo_concurrency.lockutils [None req-f72e4138-ba89-47ec-80a1-52d74fe64f4e tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Lock "8e9c27f6-e9fa-4ed0-b6ae-3b5df52d1c44" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 657.914267] env[65726]: DEBUG oslo_concurrency.lockutils [None req-f72e4138-ba89-47ec-80a1-52d74fe64f4e tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Acquiring lock "8e9c27f6-e9fa-4ed0-b6ae-3b5df52d1c44-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 657.914267] env[65726]: DEBUG oslo_concurrency.lockutils [None req-f72e4138-ba89-47ec-80a1-52d74fe64f4e tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Lock "8e9c27f6-e9fa-4ed0-b6ae-3b5df52d1c44-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 657.914499] env[65726]: DEBUG oslo_concurrency.lockutils [None req-f72e4138-ba89-47ec-80a1-52d74fe64f4e tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Lock "8e9c27f6-e9fa-4ed0-b6ae-3b5df52d1c44-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 657.917231] env[65726]: INFO nova.compute.manager [None req-f72e4138-ba89-47ec-80a1-52d74fe64f4e tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 8e9c27f6-e9fa-4ed0-b6ae-3b5df52d1c44] Terminating instance [ 657.919777] env[65726]: DEBUG oslo_concurrency.lockutils [req-21237459-e2d1-4f37-bebf-e40793b06175 req-69557957-f522-464c-bd7e-ffbde47a6481 service nova] Releasing lock "refresh_cache-f2d9090c-988f-43f4-9c81-7aa718a3438a" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 657.980551] env[65726]: DEBUG oslo_vmware.api [None req-40525054-994c-41ab-bed7-c877834cae18 tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52423c66-2d34-a7bc-9103-e8b255733749, 'name': SearchDatastore_Task, 'duration_secs': 0.015754} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 657.980551] env[65726]: DEBUG oslo_concurrency.lockutils [None req-40525054-994c-41ab-bed7-c877834cae18 tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 657.980551] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-40525054-994c-41ab-bed7-c877834cae18 tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] [instance: 3b933684-ac19-44b0-a49d-6af45501e166] Processing image b52362a3-ee8a-4cbf-b06f-513b0cc8f95c {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 657.980551] env[65726]: DEBUG oslo_concurrency.lockutils [None req-40525054-994c-41ab-bed7-c877834cae18 tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 657.980723] env[65726]: DEBUG oslo_concurrency.lockutils [None req-40525054-994c-41ab-bed7-c877834cae18 tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 657.980723] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-40525054-994c-41ab-bed7-c877834cae18 tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 657.980723] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-cc447632-cf2a-49c9-b6cd-6c91d818d4f5 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.006239] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-40525054-994c-41ab-bed7-c877834cae18 tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 658.006239] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-40525054-994c-41ab-bed7-c877834cae18 tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65726) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 658.006239] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bbf51aac-6da0-41e2-945a-68f6d0950b71 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.034322] env[65726]: DEBUG oslo_vmware.api [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Task: {'id': task-5115522, 'name': PowerOnVM_Task, 'duration_secs': 0.61322} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 658.036238] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] [instance: f5151062-57b1-4e4e-93f4-aab0e7f504d4] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 658.036596] env[65726]: INFO nova.compute.manager [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] [instance: f5151062-57b1-4e4e-93f4-aab0e7f504d4] Took 11.22 seconds to spawn the instance on the hypervisor. [ 658.036919] env[65726]: DEBUG nova.compute.manager [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] [instance: f5151062-57b1-4e4e-93f4-aab0e7f504d4] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 658.037330] env[65726]: DEBUG oslo_vmware.api [None req-40525054-994c-41ab-bed7-c877834cae18 tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] Waiting for the task: (returnval){ [ 658.037330] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5294ce9e-2906-9103-2cac-a05b2ca8e3eb" [ 658.037330] env[65726]: _type = "Task" [ 658.037330] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 658.045063] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01645bd8-1cf3-45f5-bff4-d65004ba87b7 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.068283] env[65726]: DEBUG oslo_vmware.api [None req-40525054-994c-41ab-bed7-c877834cae18 tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5294ce9e-2906-9103-2cac-a05b2ca8e3eb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 658.162116] env[65726]: DEBUG oslo_vmware.api [None req-5fd4688c-49eb-401e-afbd-116789f6a402 tempest-ServerDiagnosticsV248Test-347810772 tempest-ServerDiagnosticsV248Test-347810772-project-member] Task: {'id': task-5115525, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.124305} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 658.167308] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-5fd4688c-49eb-401e-afbd-116789f6a402 tempest-ServerDiagnosticsV248Test-347810772 tempest-ServerDiagnosticsV248Test-347810772-project-member] Deleted the datastore file {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 658.167566] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-5fd4688c-49eb-401e-afbd-116789f6a402 tempest-ServerDiagnosticsV248Test-347810772 tempest-ServerDiagnosticsV248Test-347810772-project-member] [instance: 4d4dc88b-0d6c-430c-ab4f-c81a56e4657b] Deleted contents of the VM from datastore datastore2 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 658.167778] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-5fd4688c-49eb-401e-afbd-116789f6a402 tempest-ServerDiagnosticsV248Test-347810772 tempest-ServerDiagnosticsV248Test-347810772-project-member] [instance: 4d4dc88b-0d6c-430c-ab4f-c81a56e4657b] Instance destroyed {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 658.167936] env[65726]: INFO nova.compute.manager [None req-5fd4688c-49eb-401e-afbd-116789f6a402 tempest-ServerDiagnosticsV248Test-347810772 tempest-ServerDiagnosticsV248Test-347810772-project-member] [instance: 4d4dc88b-0d6c-430c-ab4f-c81a56e4657b] Took 1.13 seconds to destroy the instance on the hypervisor. [ 658.168348] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-5fd4688c-49eb-401e-afbd-116789f6a402 tempest-ServerDiagnosticsV248Test-347810772 tempest-ServerDiagnosticsV248Test-347810772-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 658.169094] env[65726]: DEBUG nova.compute.manager [-] [instance: 4d4dc88b-0d6c-430c-ab4f-c81a56e4657b] Deallocating network for instance {{(pid=65726) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 658.169253] env[65726]: DEBUG nova.network.neutron [-] [instance: 4d4dc88b-0d6c-430c-ab4f-c81a56e4657b] deallocate_for_instance() {{(pid=65726) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 658.170088] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 658.170088] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 658.181705] env[65726]: INFO nova.compute.manager [-] [instance: de5cad99-1dbc-4435-98b4-987e24ba8c3c] Took 1.40 seconds to deallocate network for instance. [ 658.189034] env[65726]: DEBUG oslo_vmware.api [None req-881e6c53-2eb2-4fe8-9bf1-92c962897b3f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Task: {'id': task-5115526, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.109412} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 658.192535] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-881e6c53-2eb2-4fe8-9bf1-92c962897b3f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: f2d9090c-988f-43f4-9c81-7aa718a3438a] Extended root virtual disk {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 658.193823] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eef5b4d0-5eaa-4af9-bd65-990c84ebb7d5 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.235460] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-881e6c53-2eb2-4fe8-9bf1-92c962897b3f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: f2d9090c-988f-43f4-9c81-7aa718a3438a] Reconfiguring VM instance instance-00000011 to attach disk [datastore2] f2d9090c-988f-43f4-9c81-7aa718a3438a/f2d9090c-988f-43f4-9c81-7aa718a3438a.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 658.242379] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-280f097d-8341-4b44-a5c5-faeb10c13528 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.265128] env[65726]: DEBUG oslo_vmware.api [None req-cc552ef7-8733-47e0-9cfd-32896dfc9132 tempest-TenantUsagesTestJSON-327211524 tempest-TenantUsagesTestJSON-327211524-project-member] Task: {'id': task-5115527, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.539016} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 658.266717] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-cc552ef7-8733-47e0-9cfd-32896dfc9132 tempest-TenantUsagesTestJSON-327211524 tempest-TenantUsagesTestJSON-327211524-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore2] 7748c23f-9ea9-4172-8ab7-187678272bb3/7748c23f-9ea9-4172-8ab7-187678272bb3.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 658.266941] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-cc552ef7-8733-47e0-9cfd-32896dfc9132 tempest-TenantUsagesTestJSON-327211524 tempest-TenantUsagesTestJSON-327211524-project-member] [instance: 7748c23f-9ea9-4172-8ab7-187678272bb3] Extending root virtual disk to 1048576 {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 658.267302] env[65726]: DEBUG oslo_vmware.api [None req-881e6c53-2eb2-4fe8-9bf1-92c962897b3f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Waiting for the task: (returnval){ [ 658.267302] env[65726]: value = "task-5115528" [ 658.267302] env[65726]: _type = "Task" [ 658.267302] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 658.267518] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3ea0185a-d675-4a84-8403-e380597cadbe {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.282131] env[65726]: DEBUG oslo_vmware.api [None req-881e6c53-2eb2-4fe8-9bf1-92c962897b3f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Task: {'id': task-5115528, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 658.283641] env[65726]: DEBUG oslo_vmware.api [None req-cc552ef7-8733-47e0-9cfd-32896dfc9132 tempest-TenantUsagesTestJSON-327211524 tempest-TenantUsagesTestJSON-327211524-project-member] Waiting for the task: (returnval){ [ 658.283641] env[65726]: value = "task-5115529" [ 658.283641] env[65726]: _type = "Task" [ 658.283641] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 658.296721] env[65726]: DEBUG oslo_vmware.api [None req-cc552ef7-8733-47e0-9cfd-32896dfc9132 tempest-TenantUsagesTestJSON-327211524 tempest-TenantUsagesTestJSON-327211524-project-member] Task: {'id': task-5115529, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 658.320692] env[65726]: WARNING openstack [req-ca6fc5f3-49e6-44cf-8dae-7e569705410e req-b86824fb-b135-4333-9c5a-e3ab63dad036 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 658.321886] env[65726]: WARNING openstack [req-ca6fc5f3-49e6-44cf-8dae-7e569705410e req-b86824fb-b135-4333-9c5a-e3ab63dad036 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 658.371663] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61b61488-f9dd-47b5-8953-37dc1e66c596 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.380694] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d46091f3-fb3f-40e4-b044-28a1baa6c45d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.412842] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Lock "3a60b719-9db7-436e-9908-25f6a233c465" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 45.077s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 658.415184] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-046ddeb1-8824-4927-9c73-936a30bb7274 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.423994] env[65726]: DEBUG nova.compute.manager [None req-f72e4138-ba89-47ec-80a1-52d74fe64f4e tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 8e9c27f6-e9fa-4ed0-b6ae-3b5df52d1c44] Start destroying the instance on the hypervisor. {{(pid=65726) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 658.424330] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-f72e4138-ba89-47ec-80a1-52d74fe64f4e tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 8e9c27f6-e9fa-4ed0-b6ae-3b5df52d1c44] Destroying instance {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 658.426033] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bff237a1-ac05-45e4-bae5-20b1ae3d8e70 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.429382] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff67f9e5-e448-4c4f-bf5a-3c2fb77a96fc {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.448319] env[65726]: DEBUG nova.compute.provider_tree [None req-621af3d8-d8b6-4b45-86f5-d66b91010e2e tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 658.453167] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-f72e4138-ba89-47ec-80a1-52d74fe64f4e tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 8e9c27f6-e9fa-4ed0-b6ae-3b5df52d1c44] Unregistering the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 658.453675] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d4424bae-5824-4f22-a581-4f75739d927b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.547942] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-f72e4138-ba89-47ec-80a1-52d74fe64f4e tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 8e9c27f6-e9fa-4ed0-b6ae-3b5df52d1c44] Unregistered the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 658.548184] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-f72e4138-ba89-47ec-80a1-52d74fe64f4e tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 8e9c27f6-e9fa-4ed0-b6ae-3b5df52d1c44] Deleting contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 658.548360] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-f72e4138-ba89-47ec-80a1-52d74fe64f4e tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Deleting the datastore file [datastore1] 8e9c27f6-e9fa-4ed0-b6ae-3b5df52d1c44 {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 658.548703] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-603cf417-6d76-4811-8ee3-8f06da03a48e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.561069] env[65726]: DEBUG oslo_vmware.api [None req-40525054-994c-41ab-bed7-c877834cae18 tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5294ce9e-2906-9103-2cac-a05b2ca8e3eb, 'name': SearchDatastore_Task, 'duration_secs': 0.067617} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 658.563353] env[65726]: DEBUG oslo_vmware.api [None req-f72e4138-ba89-47ec-80a1-52d74fe64f4e tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Waiting for the task: (returnval){ [ 658.563353] env[65726]: value = "task-5115531" [ 658.563353] env[65726]: _type = "Task" [ 658.563353] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 658.563634] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-07f352fc-8aec-4ba0-b27f-e90daf6cf34f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.573215] env[65726]: DEBUG oslo_vmware.api [None req-40525054-994c-41ab-bed7-c877834cae18 tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] Waiting for the task: (returnval){ [ 658.573215] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]522ed211-bd02-078e-135a-481b9d386111" [ 658.573215] env[65726]: _type = "Task" [ 658.573215] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 658.582220] env[65726]: DEBUG oslo_vmware.api [None req-f72e4138-ba89-47ec-80a1-52d74fe64f4e tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': task-5115531, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 658.588208] env[65726]: INFO nova.compute.manager [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] [instance: f5151062-57b1-4e4e-93f4-aab0e7f504d4] Took 28.19 seconds to build instance. [ 658.600681] env[65726]: DEBUG oslo_vmware.api [None req-40525054-994c-41ab-bed7-c877834cae18 tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]522ed211-bd02-078e-135a-481b9d386111, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 658.638504] env[65726]: DEBUG nova.compute.manager [None req-362ec7a6-07fd-4c6b-b5fb-9865bdcb8c1a tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: 274ab469-61a9-4b7e-852c-074c871e3abf] Start spawning the instance on the hypervisor. {{(pid=65726) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 658.669636] env[65726]: DEBUG nova.virt.hardware [None req-362ec7a6-07fd-4c6b-b5fb-9865bdcb8c1a tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T19:28:34Z,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1539f06fa3534e90acbc3a60c4b8bd3f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T19:28:35Z,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 658.669901] env[65726]: DEBUG nova.virt.hardware [None req-362ec7a6-07fd-4c6b-b5fb-9865bdcb8c1a tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 658.670071] env[65726]: DEBUG nova.virt.hardware [None req-362ec7a6-07fd-4c6b-b5fb-9865bdcb8c1a tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 658.670255] env[65726]: DEBUG nova.virt.hardware [None req-362ec7a6-07fd-4c6b-b5fb-9865bdcb8c1a tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 658.670413] env[65726]: DEBUG nova.virt.hardware [None req-362ec7a6-07fd-4c6b-b5fb-9865bdcb8c1a tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 658.670555] env[65726]: DEBUG nova.virt.hardware [None req-362ec7a6-07fd-4c6b-b5fb-9865bdcb8c1a tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 658.670779] env[65726]: DEBUG nova.virt.hardware [None req-362ec7a6-07fd-4c6b-b5fb-9865bdcb8c1a tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 658.671157] env[65726]: DEBUG nova.virt.hardware [None req-362ec7a6-07fd-4c6b-b5fb-9865bdcb8c1a tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 658.671330] env[65726]: DEBUG nova.virt.hardware [None req-362ec7a6-07fd-4c6b-b5fb-9865bdcb8c1a tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 658.671487] env[65726]: DEBUG nova.virt.hardware [None req-362ec7a6-07fd-4c6b-b5fb-9865bdcb8c1a tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 658.671650] env[65726]: DEBUG nova.virt.hardware [None req-362ec7a6-07fd-4c6b-b5fb-9865bdcb8c1a tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 658.672591] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81c8e7c5-f58b-4b47-9850-610d66f510c1 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.683837] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c684f6e3-fc09-4d90-9c31-48be039f36bf {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.700491] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b83663a4-2303-46f3-8690-72e6256a8051 tempest-ServerDiagnosticsTest-2021442419 tempest-ServerDiagnosticsTest-2021442419-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 658.779783] env[65726]: DEBUG oslo_vmware.api [None req-881e6c53-2eb2-4fe8-9bf1-92c962897b3f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Task: {'id': task-5115528, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 658.797078] env[65726]: DEBUG oslo_vmware.api [None req-cc552ef7-8733-47e0-9cfd-32896dfc9132 tempest-TenantUsagesTestJSON-327211524 tempest-TenantUsagesTestJSON-327211524-project-member] Task: {'id': task-5115529, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.071848} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 658.797704] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-cc552ef7-8733-47e0-9cfd-32896dfc9132 tempest-TenantUsagesTestJSON-327211524 tempest-TenantUsagesTestJSON-327211524-project-member] [instance: 7748c23f-9ea9-4172-8ab7-187678272bb3] Extended root virtual disk {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 658.798341] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72638add-9bfa-4ed5-a609-477cd5b59f36 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.824631] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-cc552ef7-8733-47e0-9cfd-32896dfc9132 tempest-TenantUsagesTestJSON-327211524 tempest-TenantUsagesTestJSON-327211524-project-member] [instance: 7748c23f-9ea9-4172-8ab7-187678272bb3] Reconfiguring VM instance instance-00000012 to attach disk [datastore2] 7748c23f-9ea9-4172-8ab7-187678272bb3/7748c23f-9ea9-4172-8ab7-187678272bb3.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 658.824984] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f67ddef6-6a5f-4ac6-b7ec-431fa62b9da0 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.848952] env[65726]: DEBUG oslo_vmware.api [None req-cc552ef7-8733-47e0-9cfd-32896dfc9132 tempest-TenantUsagesTestJSON-327211524 tempest-TenantUsagesTestJSON-327211524-project-member] Waiting for the task: (returnval){ [ 658.848952] env[65726]: value = "task-5115532" [ 658.848952] env[65726]: _type = "Task" [ 658.848952] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 658.860533] env[65726]: DEBUG oslo_vmware.api [None req-cc552ef7-8733-47e0-9cfd-32896dfc9132 tempest-TenantUsagesTestJSON-327211524 tempest-TenantUsagesTestJSON-327211524-project-member] Task: {'id': task-5115532, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 658.919976] env[65726]: DEBUG nova.compute.manager [None req-efa812b8-b008-4879-b549-6ccf454f374f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] [instance: ddd36db0-28ad-4582-994b-7eef8b08ed46] Starting instance... {{(pid=65726) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 658.955428] env[65726]: DEBUG nova.scheduler.client.report [None req-621af3d8-d8b6-4b45-86f5-d66b91010e2e tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 659.077769] env[65726]: DEBUG oslo_vmware.api [None req-f72e4138-ba89-47ec-80a1-52d74fe64f4e tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': task-5115531, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.177197} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 659.081124] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-f72e4138-ba89-47ec-80a1-52d74fe64f4e tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Deleted the datastore file {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 659.081328] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-f72e4138-ba89-47ec-80a1-52d74fe64f4e tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 8e9c27f6-e9fa-4ed0-b6ae-3b5df52d1c44] Deleted contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 659.081502] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-f72e4138-ba89-47ec-80a1-52d74fe64f4e tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 8e9c27f6-e9fa-4ed0-b6ae-3b5df52d1c44] Instance destroyed {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 659.081680] env[65726]: INFO nova.compute.manager [None req-f72e4138-ba89-47ec-80a1-52d74fe64f4e tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 8e9c27f6-e9fa-4ed0-b6ae-3b5df52d1c44] Took 0.66 seconds to destroy the instance on the hypervisor. [ 659.081951] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-f72e4138-ba89-47ec-80a1-52d74fe64f4e tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 659.082270] env[65726]: DEBUG nova.compute.manager [-] [instance: 8e9c27f6-e9fa-4ed0-b6ae-3b5df52d1c44] Deallocating network for instance {{(pid=65726) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 659.082373] env[65726]: DEBUG nova.network.neutron [-] [instance: 8e9c27f6-e9fa-4ed0-b6ae-3b5df52d1c44] deallocate_for_instance() {{(pid=65726) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 659.082905] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 659.083188] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 659.090694] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a312f663-1ca0-47a3-ab6b-a56fbdbf35d7 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Lock "f5151062-57b1-4e4e-93f4-aab0e7f504d4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 45.671s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 659.099236] env[65726]: DEBUG oslo_vmware.api [None req-40525054-994c-41ab-bed7-c877834cae18 tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]522ed211-bd02-078e-135a-481b9d386111, 'name': SearchDatastore_Task, 'duration_secs': 0.021789} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 659.100632] env[65726]: DEBUG oslo_concurrency.lockutils [None req-40525054-994c-41ab-bed7-c877834cae18 tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 659.100632] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-40525054-994c-41ab-bed7-c877834cae18 tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore2] 3b933684-ac19-44b0-a49d-6af45501e166/3b933684-ac19-44b0-a49d-6af45501e166.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 659.100860] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c829cf72-f0c1-4a27-a157-dd70f339eb00 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.110829] env[65726]: DEBUG oslo_vmware.api [None req-40525054-994c-41ab-bed7-c877834cae18 tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] Waiting for the task: (returnval){ [ 659.110829] env[65726]: value = "task-5115533" [ 659.110829] env[65726]: _type = "Task" [ 659.110829] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 659.120840] env[65726]: DEBUG oslo_vmware.api [None req-40525054-994c-41ab-bed7-c877834cae18 tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] Task: {'id': task-5115533, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 659.282776] env[65726]: DEBUG oslo_vmware.api [None req-881e6c53-2eb2-4fe8-9bf1-92c962897b3f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Task: {'id': task-5115528, 'name': ReconfigVM_Task, 'duration_secs': 0.919672} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 659.283764] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-881e6c53-2eb2-4fe8-9bf1-92c962897b3f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: f2d9090c-988f-43f4-9c81-7aa718a3438a] Reconfigured VM instance instance-00000011 to attach disk [datastore2] f2d9090c-988f-43f4-9c81-7aa718a3438a/f2d9090c-988f-43f4-9c81-7aa718a3438a.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 659.285272] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-93295552-ec5a-491c-8ea2-eb21915c0271 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.295290] env[65726]: DEBUG oslo_vmware.api [None req-881e6c53-2eb2-4fe8-9bf1-92c962897b3f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Waiting for the task: (returnval){ [ 659.295290] env[65726]: value = "task-5115534" [ 659.295290] env[65726]: _type = "Task" [ 659.295290] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 659.308410] env[65726]: DEBUG oslo_vmware.api [None req-881e6c53-2eb2-4fe8-9bf1-92c962897b3f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Task: {'id': task-5115534, 'name': Rename_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 659.326050] env[65726]: DEBUG nova.network.neutron [-] [instance: 4d4dc88b-0d6c-430c-ab4f-c81a56e4657b] Instance cache missing network info. {{(pid=65726) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 659.326423] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 659.326829] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 659.345865] env[65726]: DEBUG nova.compute.manager [req-53ca3179-afc3-44a3-85e5-86acad674899 req-59453eac-374e-48b2-87cb-f0b98ccf4a78 service nova] [instance: de5cad99-1dbc-4435-98b4-987e24ba8c3c] Received event network-vif-deleted-3f320668-6397-427f-b8c4-3e63282fcb87 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 659.361958] env[65726]: DEBUG oslo_vmware.api [None req-cc552ef7-8733-47e0-9cfd-32896dfc9132 tempest-TenantUsagesTestJSON-327211524 tempest-TenantUsagesTestJSON-327211524-project-member] Task: {'id': task-5115532, 'name': ReconfigVM_Task, 'duration_secs': 0.489889} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 659.362227] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-cc552ef7-8733-47e0-9cfd-32896dfc9132 tempest-TenantUsagesTestJSON-327211524 tempest-TenantUsagesTestJSON-327211524-project-member] [instance: 7748c23f-9ea9-4172-8ab7-187678272bb3] Reconfigured VM instance instance-00000012 to attach disk [datastore2] 7748c23f-9ea9-4172-8ab7-187678272bb3/7748c23f-9ea9-4172-8ab7-187678272bb3.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 659.363147] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ffe389d9-e9c2-4796-b36f-f11e0c465651 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.373020] env[65726]: DEBUG nova.network.neutron [None req-362ec7a6-07fd-4c6b-b5fb-9865bdcb8c1a tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: 274ab469-61a9-4b7e-852c-074c871e3abf] Successfully updated port: 757d86b3-4474-41c5-aaf8-14fc7216cf15 {{(pid=65726) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 659.381470] env[65726]: DEBUG oslo_vmware.api [None req-cc552ef7-8733-47e0-9cfd-32896dfc9132 tempest-TenantUsagesTestJSON-327211524 tempest-TenantUsagesTestJSON-327211524-project-member] Waiting for the task: (returnval){ [ 659.381470] env[65726]: value = "task-5115535" [ 659.381470] env[65726]: _type = "Task" [ 659.381470] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 659.394296] env[65726]: DEBUG oslo_vmware.api [None req-cc552ef7-8733-47e0-9cfd-32896dfc9132 tempest-TenantUsagesTestJSON-327211524 tempest-TenantUsagesTestJSON-327211524-project-member] Task: {'id': task-5115535, 'name': Rename_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 659.440142] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 659.440547] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 659.450539] env[65726]: DEBUG oslo_concurrency.lockutils [None req-efa812b8-b008-4879-b549-6ccf454f374f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 659.462402] env[65726]: DEBUG oslo_concurrency.lockutils [None req-621af3d8-d8b6-4b45-86f5-d66b91010e2e tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.861s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 659.462949] env[65726]: DEBUG nova.compute.manager [None req-621af3d8-d8b6-4b45-86f5-d66b91010e2e tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] [instance: fd481728-3ef0-4a96-affd-ab2dd3f596bb] Start building networks asynchronously for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 659.467154] env[65726]: DEBUG oslo_concurrency.lockutils [None req-06f67a27-e6e0-4078-804d-cff1137abcad tempest-ServersTestBootFromVolume-762458528 tempest-ServersTestBootFromVolume-762458528-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.262s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 659.468840] env[65726]: INFO nova.compute.claims [None req-06f67a27-e6e0-4078-804d-cff1137abcad tempest-ServersTestBootFromVolume-762458528 tempest-ServersTestBootFromVolume-762458528-project-member] [instance: 5a8d4357-4b44-4a19-b1da-42d188c38adc] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 659.601941] env[65726]: DEBUG nova.compute.manager [None req-bb345b26-9f49-4e52-904f-99ccc71299ac tempest-AttachInterfacesV270Test-133387050 tempest-AttachInterfacesV270Test-133387050-project-member] [instance: c977257e-dce0-4ecc-b42b-0f6ebab38797] Starting instance... {{(pid=65726) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 659.627253] env[65726]: DEBUG oslo_vmware.api [None req-40525054-994c-41ab-bed7-c877834cae18 tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] Task: {'id': task-5115533, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 659.809572] env[65726]: DEBUG oslo_vmware.api [None req-881e6c53-2eb2-4fe8-9bf1-92c962897b3f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Task: {'id': task-5115534, 'name': Rename_Task, 'duration_secs': 0.262158} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 659.809870] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-881e6c53-2eb2-4fe8-9bf1-92c962897b3f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: f2d9090c-988f-43f4-9c81-7aa718a3438a] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 659.810490] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d9a94b23-99c6-4a56-896e-c2b09d6eb926 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.817899] env[65726]: DEBUG oslo_vmware.api [None req-881e6c53-2eb2-4fe8-9bf1-92c962897b3f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Waiting for the task: (returnval){ [ 659.817899] env[65726]: value = "task-5115536" [ 659.817899] env[65726]: _type = "Task" [ 659.817899] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 659.829008] env[65726]: DEBUG oslo_vmware.api [None req-881e6c53-2eb2-4fe8-9bf1-92c962897b3f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Task: {'id': task-5115536, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 659.839760] env[65726]: DEBUG nova.network.neutron [-] [instance: 4d4dc88b-0d6c-430c-ab4f-c81a56e4657b] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 659.877833] env[65726]: DEBUG oslo_concurrency.lockutils [None req-362ec7a6-07fd-4c6b-b5fb-9865bdcb8c1a tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Acquiring lock "refresh_cache-274ab469-61a9-4b7e-852c-074c871e3abf" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 659.878052] env[65726]: DEBUG oslo_concurrency.lockutils [None req-362ec7a6-07fd-4c6b-b5fb-9865bdcb8c1a tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Acquired lock "refresh_cache-274ab469-61a9-4b7e-852c-074c871e3abf" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 659.878209] env[65726]: DEBUG nova.network.neutron [None req-362ec7a6-07fd-4c6b-b5fb-9865bdcb8c1a tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: 274ab469-61a9-4b7e-852c-074c871e3abf] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 659.898357] env[65726]: DEBUG oslo_vmware.api [None req-cc552ef7-8733-47e0-9cfd-32896dfc9132 tempest-TenantUsagesTestJSON-327211524 tempest-TenantUsagesTestJSON-327211524-project-member] Task: {'id': task-5115535, 'name': Rename_Task, 'duration_secs': 0.498328} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 659.898714] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-cc552ef7-8733-47e0-9cfd-32896dfc9132 tempest-TenantUsagesTestJSON-327211524 tempest-TenantUsagesTestJSON-327211524-project-member] [instance: 7748c23f-9ea9-4172-8ab7-187678272bb3] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 659.900321] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-07704248-67aa-41b8-96cf-677b76a9833e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.913638] env[65726]: DEBUG oslo_vmware.api [None req-cc552ef7-8733-47e0-9cfd-32896dfc9132 tempest-TenantUsagesTestJSON-327211524 tempest-TenantUsagesTestJSON-327211524-project-member] Waiting for the task: (returnval){ [ 659.913638] env[65726]: value = "task-5115537" [ 659.913638] env[65726]: _type = "Task" [ 659.913638] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 659.925043] env[65726]: DEBUG oslo_vmware.api [None req-cc552ef7-8733-47e0-9cfd-32896dfc9132 tempest-TenantUsagesTestJSON-327211524 tempest-TenantUsagesTestJSON-327211524-project-member] Task: {'id': task-5115537, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 659.973750] env[65726]: DEBUG nova.compute.utils [None req-621af3d8-d8b6-4b45-86f5-d66b91010e2e tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Using /dev/sd instead of None {{(pid=65726) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 659.978406] env[65726]: DEBUG nova.compute.manager [None req-621af3d8-d8b6-4b45-86f5-d66b91010e2e tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] [instance: fd481728-3ef0-4a96-affd-ab2dd3f596bb] Allocating IP information in the background. {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 659.978754] env[65726]: DEBUG nova.network.neutron [None req-621af3d8-d8b6-4b45-86f5-d66b91010e2e tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] [instance: fd481728-3ef0-4a96-affd-ab2dd3f596bb] allocate_for_instance() {{(pid=65726) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 659.979014] env[65726]: WARNING neutronclient.v2_0.client [None req-621af3d8-d8b6-4b45-86f5-d66b91010e2e tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 659.982466] env[65726]: WARNING neutronclient.v2_0.client [None req-621af3d8-d8b6-4b45-86f5-d66b91010e2e tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 659.983888] env[65726]: WARNING openstack [None req-621af3d8-d8b6-4b45-86f5-d66b91010e2e tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 659.984255] env[65726]: WARNING openstack [None req-621af3d8-d8b6-4b45-86f5-d66b91010e2e tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 659.999688] env[65726]: DEBUG nova.compute.manager [None req-621af3d8-d8b6-4b45-86f5-d66b91010e2e tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] [instance: fd481728-3ef0-4a96-affd-ab2dd3f596bb] Start building block device mappings for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 660.123917] env[65726]: DEBUG oslo_vmware.api [None req-40525054-994c-41ab-bed7-c877834cae18 tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] Task: {'id': task-5115533, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.69453} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 660.125269] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-40525054-994c-41ab-bed7-c877834cae18 tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore2] 3b933684-ac19-44b0-a49d-6af45501e166/3b933684-ac19-44b0-a49d-6af45501e166.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 660.128634] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-40525054-994c-41ab-bed7-c877834cae18 tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] [instance: 3b933684-ac19-44b0-a49d-6af45501e166] Extending root virtual disk to 1048576 {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 660.129113] env[65726]: DEBUG oslo_concurrency.lockutils [None req-bb345b26-9f49-4e52-904f-99ccc71299ac tempest-AttachInterfacesV270Test-133387050 tempest-AttachInterfacesV270Test-133387050-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 660.129403] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-18102c78-bec9-49a3-8936-6061208543ff {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.137120] env[65726]: DEBUG oslo_vmware.api [None req-40525054-994c-41ab-bed7-c877834cae18 tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] Waiting for the task: (returnval){ [ 660.137120] env[65726]: value = "task-5115538" [ 660.137120] env[65726]: _type = "Task" [ 660.137120] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 660.150833] env[65726]: DEBUG oslo_vmware.api [None req-40525054-994c-41ab-bed7-c877834cae18 tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] Task: {'id': task-5115538, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 660.204873] env[65726]: WARNING openstack [req-ca6fc5f3-49e6-44cf-8dae-7e569705410e req-b86824fb-b135-4333-9c5a-e3ab63dad036 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 660.205234] env[65726]: WARNING openstack [req-ca6fc5f3-49e6-44cf-8dae-7e569705410e req-b86824fb-b135-4333-9c5a-e3ab63dad036 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 660.322839] env[65726]: DEBUG nova.policy [None req-621af3d8-d8b6-4b45-86f5-d66b91010e2e tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'cea37b69d6cc4784a4c6a8889a718bfd', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '79d3e960f190439fae7c61d693e9356e', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65726) authorize /opt/stack/nova/nova/policy.py:192}} [ 660.339272] env[65726]: DEBUG oslo_vmware.api [None req-881e6c53-2eb2-4fe8-9bf1-92c962897b3f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Task: {'id': task-5115536, 'name': PowerOnVM_Task} progress is 88%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 660.343130] env[65726]: INFO nova.compute.manager [-] [instance: 4d4dc88b-0d6c-430c-ab4f-c81a56e4657b] Took 2.17 seconds to deallocate network for instance. [ 660.381865] env[65726]: WARNING openstack [None req-362ec7a6-07fd-4c6b-b5fb-9865bdcb8c1a tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 660.382464] env[65726]: WARNING openstack [None req-362ec7a6-07fd-4c6b-b5fb-9865bdcb8c1a tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 660.388184] env[65726]: DEBUG nova.network.neutron [None req-362ec7a6-07fd-4c6b-b5fb-9865bdcb8c1a tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: 274ab469-61a9-4b7e-852c-074c871e3abf] Instance cache missing network info. {{(pid=65726) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 660.435827] env[65726]: DEBUG nova.network.neutron [-] [instance: 8e9c27f6-e9fa-4ed0-b6ae-3b5df52d1c44] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 660.440873] env[65726]: DEBUG oslo_vmware.api [None req-cc552ef7-8733-47e0-9cfd-32896dfc9132 tempest-TenantUsagesTestJSON-327211524 tempest-TenantUsagesTestJSON-327211524-project-member] Task: {'id': task-5115537, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 660.607413] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c8072ff-f8d7-4fab-b0fe-dcbfed4e7b40 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.617098] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a520f141-740b-4226-8493-6feb3516d16d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.661352] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb89e453-871b-4899-b744-f85e707fc4e9 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.669714] env[65726]: DEBUG oslo_vmware.api [None req-40525054-994c-41ab-bed7-c877834cae18 tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] Task: {'id': task-5115538, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.08203} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 660.672151] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-40525054-994c-41ab-bed7-c877834cae18 tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] [instance: 3b933684-ac19-44b0-a49d-6af45501e166] Extended root virtual disk {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 660.672988] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-faec8176-61cb-4411-b26c-a8cb72ba52f7 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.676672] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d182e3ec-c7ba-48ac-8545-881b15f0df9d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.700039] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-40525054-994c-41ab-bed7-c877834cae18 tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] [instance: 3b933684-ac19-44b0-a49d-6af45501e166] Reconfiguring VM instance instance-00000013 to attach disk [datastore2] 3b933684-ac19-44b0-a49d-6af45501e166/3b933684-ac19-44b0-a49d-6af45501e166.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 660.708473] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-abaf7797-10b4-44dd-bb85-3193b5eb85c4 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.723348] env[65726]: DEBUG nova.compute.provider_tree [None req-06f67a27-e6e0-4078-804d-cff1137abcad tempest-ServersTestBootFromVolume-762458528 tempest-ServersTestBootFromVolume-762458528-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 660.731899] env[65726]: DEBUG oslo_vmware.api [None req-40525054-994c-41ab-bed7-c877834cae18 tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] Waiting for the task: (returnval){ [ 660.731899] env[65726]: value = "task-5115539" [ 660.731899] env[65726]: _type = "Task" [ 660.731899] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 660.743203] env[65726]: DEBUG oslo_vmware.api [None req-40525054-994c-41ab-bed7-c877834cae18 tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] Task: {'id': task-5115539, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 660.783858] env[65726]: WARNING openstack [req-ca6fc5f3-49e6-44cf-8dae-7e569705410e req-b86824fb-b135-4333-9c5a-e3ab63dad036 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 660.783858] env[65726]: WARNING openstack [req-ca6fc5f3-49e6-44cf-8dae-7e569705410e req-b86824fb-b135-4333-9c5a-e3ab63dad036 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 660.833020] env[65726]: DEBUG oslo_vmware.api [None req-881e6c53-2eb2-4fe8-9bf1-92c962897b3f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Task: {'id': task-5115536, 'name': PowerOnVM_Task, 'duration_secs': 0.848186} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 660.833020] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-881e6c53-2eb2-4fe8-9bf1-92c962897b3f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: f2d9090c-988f-43f4-9c81-7aa718a3438a] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 660.833020] env[65726]: INFO nova.compute.manager [None req-881e6c53-2eb2-4fe8-9bf1-92c962897b3f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: f2d9090c-988f-43f4-9c81-7aa718a3438a] Took 10.74 seconds to spawn the instance on the hypervisor. [ 660.833020] env[65726]: DEBUG nova.compute.manager [None req-881e6c53-2eb2-4fe8-9bf1-92c962897b3f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: f2d9090c-988f-43f4-9c81-7aa718a3438a] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 660.833020] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab46b29c-29a5-4cfc-9b40-c2ad0feae985 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.855377] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5fd4688c-49eb-401e-afbd-116789f6a402 tempest-ServerDiagnosticsV248Test-347810772 tempest-ServerDiagnosticsV248Test-347810772-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 660.925358] env[65726]: DEBUG oslo_vmware.api [None req-cc552ef7-8733-47e0-9cfd-32896dfc9132 tempest-TenantUsagesTestJSON-327211524 tempest-TenantUsagesTestJSON-327211524-project-member] Task: {'id': task-5115537, 'name': PowerOnVM_Task, 'duration_secs': 0.567026} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 660.926276] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-cc552ef7-8733-47e0-9cfd-32896dfc9132 tempest-TenantUsagesTestJSON-327211524 tempest-TenantUsagesTestJSON-327211524-project-member] [instance: 7748c23f-9ea9-4172-8ab7-187678272bb3] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 660.926441] env[65726]: INFO nova.compute.manager [None req-cc552ef7-8733-47e0-9cfd-32896dfc9132 tempest-TenantUsagesTestJSON-327211524 tempest-TenantUsagesTestJSON-327211524-project-member] [instance: 7748c23f-9ea9-4172-8ab7-187678272bb3] Took 8.03 seconds to spawn the instance on the hypervisor. [ 660.926616] env[65726]: DEBUG nova.compute.manager [None req-cc552ef7-8733-47e0-9cfd-32896dfc9132 tempest-TenantUsagesTestJSON-327211524 tempest-TenantUsagesTestJSON-327211524-project-member] [instance: 7748c23f-9ea9-4172-8ab7-187678272bb3] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 660.927623] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81f8dab9-0ac1-4f21-9b1a-2a02bc004b69 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.939682] env[65726]: INFO nova.compute.manager [-] [instance: 8e9c27f6-e9fa-4ed0-b6ae-3b5df52d1c44] Took 1.86 seconds to deallocate network for instance. [ 661.016171] env[65726]: DEBUG nova.compute.manager [None req-621af3d8-d8b6-4b45-86f5-d66b91010e2e tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] [instance: fd481728-3ef0-4a96-affd-ab2dd3f596bb] Start spawning the instance on the hypervisor. {{(pid=65726) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 661.054835] env[65726]: DEBUG nova.virt.hardware [None req-621af3d8-d8b6-4b45-86f5-d66b91010e2e tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:31:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='382792978',id=22,is_public=True,memory_mb=192,name='tempest-flavor_with_ephemeral_0-83177855',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T19:28:34Z,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1539f06fa3534e90acbc3a60c4b8bd3f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T19:28:35Z,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 661.055044] env[65726]: DEBUG nova.virt.hardware [None req-621af3d8-d8b6-4b45-86f5-d66b91010e2e tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 661.056377] env[65726]: DEBUG nova.virt.hardware [None req-621af3d8-d8b6-4b45-86f5-d66b91010e2e tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 661.056836] env[65726]: DEBUG nova.virt.hardware [None req-621af3d8-d8b6-4b45-86f5-d66b91010e2e tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 661.057077] env[65726]: DEBUG nova.virt.hardware [None req-621af3d8-d8b6-4b45-86f5-d66b91010e2e tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 661.057305] env[65726]: DEBUG nova.virt.hardware [None req-621af3d8-d8b6-4b45-86f5-d66b91010e2e tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 661.057709] env[65726]: DEBUG nova.virt.hardware [None req-621af3d8-d8b6-4b45-86f5-d66b91010e2e tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 661.058059] env[65726]: DEBUG nova.virt.hardware [None req-621af3d8-d8b6-4b45-86f5-d66b91010e2e tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 661.058399] env[65726]: DEBUG nova.virt.hardware [None req-621af3d8-d8b6-4b45-86f5-d66b91010e2e tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 661.058763] env[65726]: DEBUG nova.virt.hardware [None req-621af3d8-d8b6-4b45-86f5-d66b91010e2e tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 661.059145] env[65726]: DEBUG nova.virt.hardware [None req-621af3d8-d8b6-4b45-86f5-d66b91010e2e tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 661.061150] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0bdd48bc-3b3f-437c-a2b0-6bb1ba01a3f9 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.080028] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9547a064-f921-4bd9-af65-2f6d0f4a5fc0 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.227128] env[65726]: DEBUG nova.scheduler.client.report [None req-06f67a27-e6e0-4078-804d-cff1137abcad tempest-ServersTestBootFromVolume-762458528 tempest-ServersTestBootFromVolume-762458528-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 661.244031] env[65726]: DEBUG oslo_vmware.api [None req-40525054-994c-41ab-bed7-c877834cae18 tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] Task: {'id': task-5115539, 'name': ReconfigVM_Task, 'duration_secs': 0.446661} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 661.244316] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-40525054-994c-41ab-bed7-c877834cae18 tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] [instance: 3b933684-ac19-44b0-a49d-6af45501e166] Reconfigured VM instance instance-00000013 to attach disk [datastore2] 3b933684-ac19-44b0-a49d-6af45501e166/3b933684-ac19-44b0-a49d-6af45501e166.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 661.245607] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4bedd9f6-d387-4620-a779-1726a24107e0 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.253665] env[65726]: DEBUG oslo_vmware.api [None req-40525054-994c-41ab-bed7-c877834cae18 tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] Waiting for the task: (returnval){ [ 661.253665] env[65726]: value = "task-5115540" [ 661.253665] env[65726]: _type = "Task" [ 661.253665] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 661.264251] env[65726]: DEBUG oslo_vmware.api [None req-40525054-994c-41ab-bed7-c877834cae18 tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] Task: {'id': task-5115540, 'name': Rename_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 661.345229] env[65726]: DEBUG nova.network.neutron [None req-621af3d8-d8b6-4b45-86f5-d66b91010e2e tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] [instance: fd481728-3ef0-4a96-affd-ab2dd3f596bb] Successfully created port: 7df3f02b-e4e5-4840-bf9f-3498771b9184 {{(pid=65726) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 661.359252] env[65726]: INFO nova.compute.manager [None req-881e6c53-2eb2-4fe8-9bf1-92c962897b3f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: f2d9090c-988f-43f4-9c81-7aa718a3438a] Took 30.38 seconds to build instance. [ 661.365163] env[65726]: WARNING openstack [None req-362ec7a6-07fd-4c6b-b5fb-9865bdcb8c1a tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 661.365427] env[65726]: WARNING openstack [None req-362ec7a6-07fd-4c6b-b5fb-9865bdcb8c1a tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 661.447233] env[65726]: DEBUG oslo_concurrency.lockutils [None req-f72e4138-ba89-47ec-80a1-52d74fe64f4e tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 661.447233] env[65726]: INFO nova.compute.manager [None req-cc552ef7-8733-47e0-9cfd-32896dfc9132 tempest-TenantUsagesTestJSON-327211524 tempest-TenantUsagesTestJSON-327211524-project-member] [instance: 7748c23f-9ea9-4172-8ab7-187678272bb3] Took 28.34 seconds to build instance. [ 661.733113] env[65726]: DEBUG oslo_concurrency.lockutils [None req-06f67a27-e6e0-4078-804d-cff1137abcad tempest-ServersTestBootFromVolume-762458528 tempest-ServersTestBootFromVolume-762458528-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.266s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 661.733542] env[65726]: DEBUG nova.compute.manager [None req-06f67a27-e6e0-4078-804d-cff1137abcad tempest-ServersTestBootFromVolume-762458528 tempest-ServersTestBootFromVolume-762458528-project-member] [instance: 5a8d4357-4b44-4a19-b1da-42d188c38adc] Start building networks asynchronously for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 661.736650] env[65726]: DEBUG oslo_concurrency.lockutils [None req-36828e27-b100-49cc-98c2-ad0b25ca6572 tempest-ServerGroupTestJSON-125182767 tempest-ServerGroupTestJSON-125182767-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.119s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 661.738172] env[65726]: INFO nova.compute.claims [None req-36828e27-b100-49cc-98c2-ad0b25ca6572 tempest-ServerGroupTestJSON-125182767 tempest-ServerGroupTestJSON-125182767-project-member] [instance: c2929727-a7ea-4f63-997d-474c29a305fc] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 661.766076] env[65726]: DEBUG oslo_vmware.api [None req-40525054-994c-41ab-bed7-c877834cae18 tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] Task: {'id': task-5115540, 'name': Rename_Task, 'duration_secs': 0.354922} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 661.766161] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-40525054-994c-41ab-bed7-c877834cae18 tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] [instance: 3b933684-ac19-44b0-a49d-6af45501e166] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 661.766423] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-466d2c8c-f02d-4fb2-b39d-a6deed7dda56 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.778674] env[65726]: DEBUG oslo_vmware.api [None req-40525054-994c-41ab-bed7-c877834cae18 tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] Waiting for the task: (returnval){ [ 661.778674] env[65726]: value = "task-5115541" [ 661.778674] env[65726]: _type = "Task" [ 661.778674] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 661.786939] env[65726]: DEBUG oslo_vmware.api [None req-40525054-994c-41ab-bed7-c877834cae18 tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] Task: {'id': task-5115541, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 661.873878] env[65726]: DEBUG oslo_concurrency.lockutils [None req-881e6c53-2eb2-4fe8-9bf1-92c962897b3f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Lock "f2d9090c-988f-43f4-9c81-7aa718a3438a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 47.227s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 661.949420] env[65726]: DEBUG oslo_concurrency.lockutils [None req-cc552ef7-8733-47e0-9cfd-32896dfc9132 tempest-TenantUsagesTestJSON-327211524 tempest-TenantUsagesTestJSON-327211524-project-member] Lock "7748c23f-9ea9-4172-8ab7-187678272bb3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 45.889s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 662.243620] env[65726]: DEBUG nova.compute.utils [None req-06f67a27-e6e0-4078-804d-cff1137abcad tempest-ServersTestBootFromVolume-762458528 tempest-ServersTestBootFromVolume-762458528-project-member] Using /dev/sd instead of None {{(pid=65726) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 662.248103] env[65726]: DEBUG nova.compute.manager [None req-06f67a27-e6e0-4078-804d-cff1137abcad tempest-ServersTestBootFromVolume-762458528 tempest-ServersTestBootFromVolume-762458528-project-member] [instance: 5a8d4357-4b44-4a19-b1da-42d188c38adc] Allocating IP information in the background. {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 662.248103] env[65726]: DEBUG nova.network.neutron [None req-06f67a27-e6e0-4078-804d-cff1137abcad tempest-ServersTestBootFromVolume-762458528 tempest-ServersTestBootFromVolume-762458528-project-member] [instance: 5a8d4357-4b44-4a19-b1da-42d188c38adc] allocate_for_instance() {{(pid=65726) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 662.248103] env[65726]: WARNING neutronclient.v2_0.client [None req-06f67a27-e6e0-4078-804d-cff1137abcad tempest-ServersTestBootFromVolume-762458528 tempest-ServersTestBootFromVolume-762458528-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 662.248103] env[65726]: WARNING neutronclient.v2_0.client [None req-06f67a27-e6e0-4078-804d-cff1137abcad tempest-ServersTestBootFromVolume-762458528 tempest-ServersTestBootFromVolume-762458528-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 662.248395] env[65726]: WARNING openstack [None req-06f67a27-e6e0-4078-804d-cff1137abcad tempest-ServersTestBootFromVolume-762458528 tempest-ServersTestBootFromVolume-762458528-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 662.248783] env[65726]: WARNING openstack [None req-06f67a27-e6e0-4078-804d-cff1137abcad tempest-ServersTestBootFromVolume-762458528 tempest-ServersTestBootFromVolume-762458528-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 662.302103] env[65726]: DEBUG oslo_vmware.api [None req-40525054-994c-41ab-bed7-c877834cae18 tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] Task: {'id': task-5115541, 'name': PowerOnVM_Task} progress is 87%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 662.380807] env[65726]: DEBUG nova.compute.manager [None req-e9359f40-01a5-4a39-b7b9-47a0a48e2ede tempest-VolumesAdminNegativeTest-2004477850 tempest-VolumesAdminNegativeTest-2004477850-project-member] [instance: 30b17a17-229c-4ca7-9ae6-c67ccdcd3963] Starting instance... {{(pid=65726) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 662.459047] env[65726]: DEBUG nova.compute.manager [None req-60511481-93fa-4f8b-817d-01a577f85a26 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: 49bcab1f-7b64-4999-abff-37771c58a271] Starting instance... {{(pid=65726) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 662.738101] env[65726]: DEBUG nova.network.neutron [req-ca6fc5f3-49e6-44cf-8dae-7e569705410e req-b86824fb-b135-4333-9c5a-e3ab63dad036 service nova] [instance: 7748c23f-9ea9-4172-8ab7-187678272bb3] Updated VIF entry in instance network info cache for port 8bb2d672-e56e-4f01-bfbf-74a279d9da7d. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 662.738790] env[65726]: DEBUG nova.network.neutron [req-ca6fc5f3-49e6-44cf-8dae-7e569705410e req-b86824fb-b135-4333-9c5a-e3ab63dad036 service nova] [instance: 7748c23f-9ea9-4172-8ab7-187678272bb3] Updating instance_info_cache with network_info: [{"id": "8bb2d672-e56e-4f01-bfbf-74a279d9da7d", "address": "fa:16:3e:01:7e:88", "network": {"id": "55d629b8-a541-4720-9289-927bbc6e6f22", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.177", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "1539f06fa3534e90acbc3a60c4b8bd3f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "35342bcb-8b06-472e-b3c0-43fd3d6c4b30", "external-id": "nsx-vlan-transportzone-524", "segmentation_id": 524, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8bb2d672-e5", "ovs_interfaceid": "8bb2d672-e56e-4f01-bfbf-74a279d9da7d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 662.762043] env[65726]: DEBUG nova.compute.manager [None req-06f67a27-e6e0-4078-804d-cff1137abcad tempest-ServersTestBootFromVolume-762458528 tempest-ServersTestBootFromVolume-762458528-project-member] [instance: 5a8d4357-4b44-4a19-b1da-42d188c38adc] Start building block device mappings for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 662.772955] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a225664-947b-4394-93f0-a12652553beb {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.784434] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42507438-ff4d-437c-acd5-97d98e8db40e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.794341] env[65726]: DEBUG oslo_vmware.api [None req-40525054-994c-41ab-bed7-c877834cae18 tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] Task: {'id': task-5115541, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 662.823718] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a4f24fc-a36a-4cfb-84b2-2d7b3e335e90 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.833684] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b7e77b5-e299-4423-8ff5-5876a9046af9 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.851444] env[65726]: DEBUG nova.compute.provider_tree [None req-36828e27-b100-49cc-98c2-ad0b25ca6572 tempest-ServerGroupTestJSON-125182767 tempest-ServerGroupTestJSON-125182767-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 662.894363] env[65726]: DEBUG nova.policy [None req-06f67a27-e6e0-4078-804d-cff1137abcad tempest-ServersTestBootFromVolume-762458528 tempest-ServersTestBootFromVolume-762458528-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '73ae04dc18d2480ea2903c94032069e2', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c52b5289c78f45d3942a6c0a4b026207', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65726) authorize /opt/stack/nova/nova/policy.py:192}} [ 662.916926] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e9359f40-01a5-4a39-b7b9-47a0a48e2ede tempest-VolumesAdminNegativeTest-2004477850 tempest-VolumesAdminNegativeTest-2004477850-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 662.942896] env[65726]: WARNING openstack [None req-362ec7a6-07fd-4c6b-b5fb-9865bdcb8c1a tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 662.943395] env[65726]: WARNING openstack [None req-362ec7a6-07fd-4c6b-b5fb-9865bdcb8c1a tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 662.981858] env[65726]: DEBUG oslo_concurrency.lockutils [None req-60511481-93fa-4f8b-817d-01a577f85a26 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 663.198848] env[65726]: DEBUG nova.network.neutron [None req-621af3d8-d8b6-4b45-86f5-d66b91010e2e tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] [instance: fd481728-3ef0-4a96-affd-ab2dd3f596bb] Successfully updated port: 7df3f02b-e4e5-4840-bf9f-3498771b9184 {{(pid=65726) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 663.241736] env[65726]: DEBUG oslo_concurrency.lockutils [req-ca6fc5f3-49e6-44cf-8dae-7e569705410e req-b86824fb-b135-4333-9c5a-e3ab63dad036 service nova] Releasing lock "refresh_cache-7748c23f-9ea9-4172-8ab7-187678272bb3" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 663.268600] env[65726]: INFO nova.virt.block_device [None req-06f67a27-e6e0-4078-804d-cff1137abcad tempest-ServersTestBootFromVolume-762458528 tempest-ServersTestBootFromVolume-762458528-project-member] [instance: 5a8d4357-4b44-4a19-b1da-42d188c38adc] Booting with volume b11cb57f-92f1-416d-b59b-70cffe1353ff at /dev/sda [ 663.307677] env[65726]: DEBUG oslo_vmware.api [None req-40525054-994c-41ab-bed7-c877834cae18 tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] Task: {'id': task-5115541, 'name': PowerOnVM_Task, 'duration_secs': 1.515599} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 663.307677] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-40525054-994c-41ab-bed7-c877834cae18 tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] [instance: 3b933684-ac19-44b0-a49d-6af45501e166] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 663.307677] env[65726]: INFO nova.compute.manager [None req-40525054-994c-41ab-bed7-c877834cae18 tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] [instance: 3b933684-ac19-44b0-a49d-6af45501e166] Took 7.55 seconds to spawn the instance on the hypervisor. [ 663.307677] env[65726]: DEBUG nova.compute.manager [None req-40525054-994c-41ab-bed7-c877834cae18 tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] [instance: 3b933684-ac19-44b0-a49d-6af45501e166] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 663.311335] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-049dc16a-262d-4d48-abee-50056445257a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.354607] env[65726]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7196aad4-afe6-472e-a03c-b34be2668c33 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.356171] env[65726]: DEBUG nova.scheduler.client.report [None req-36828e27-b100-49cc-98c2-ad0b25ca6572 tempest-ServerGroupTestJSON-125182767 tempest-ServerGroupTestJSON-125182767-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 663.370864] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e28f68a4-c851-4935-b051-2d118709ecd8 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.418143] env[65726]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5e919f08-c1e2-49ee-a8b8-e32054b7d5f0 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.429114] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72527322-ef9a-4cbb-aa39-2d0f15cc18e3 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.475636] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd68aa56-41c4-4ed5-b693-a11d2332745e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.486055] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc416615-9fc4-47c7-9a9b-4523be3453c8 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.503736] env[65726]: DEBUG nova.virt.block_device [None req-06f67a27-e6e0-4078-804d-cff1137abcad tempest-ServersTestBootFromVolume-762458528 tempest-ServersTestBootFromVolume-762458528-project-member] [instance: 5a8d4357-4b44-4a19-b1da-42d188c38adc] Updating existing volume attachment record: 647e2ade-9cc7-4618-838d-500431f01095 {{(pid=65726) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 663.700628] env[65726]: DEBUG nova.network.neutron [None req-362ec7a6-07fd-4c6b-b5fb-9865bdcb8c1a tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: 274ab469-61a9-4b7e-852c-074c871e3abf] Updating instance_info_cache with network_info: [{"id": "757d86b3-4474-41c5-aaf8-14fc7216cf15", "address": "fa:16:3e:b9:96:d6", "network": {"id": "3f7916f7-ee9c-4629-b3b3-1367de9f98ea", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-579713617-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b794c99309e94c1181fc9d8eeeb84702", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d0bdd11b-58af-4cc0-9d38-8322e1bb4e74", "external-id": "nsx-vlan-transportzone-398", "segmentation_id": 398, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap757d86b3-44", "ovs_interfaceid": "757d86b3-4474-41c5-aaf8-14fc7216cf15", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 663.705092] env[65726]: DEBUG oslo_concurrency.lockutils [None req-621af3d8-d8b6-4b45-86f5-d66b91010e2e tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Acquiring lock "refresh_cache-fd481728-3ef0-4a96-affd-ab2dd3f596bb" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 663.705092] env[65726]: DEBUG oslo_concurrency.lockutils [None req-621af3d8-d8b6-4b45-86f5-d66b91010e2e tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Acquired lock "refresh_cache-fd481728-3ef0-4a96-affd-ab2dd3f596bb" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 663.705092] env[65726]: DEBUG nova.network.neutron [None req-621af3d8-d8b6-4b45-86f5-d66b91010e2e tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] [instance: fd481728-3ef0-4a96-affd-ab2dd3f596bb] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 663.819052] env[65726]: DEBUG nova.network.neutron [None req-06f67a27-e6e0-4078-804d-cff1137abcad tempest-ServersTestBootFromVolume-762458528 tempest-ServersTestBootFromVolume-762458528-project-member] [instance: 5a8d4357-4b44-4a19-b1da-42d188c38adc] Successfully created port: 9aaacd82-31c1-4cfa-bbce-860b30db74af {{(pid=65726) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 663.839413] env[65726]: INFO nova.compute.manager [None req-40525054-994c-41ab-bed7-c877834cae18 tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] [instance: 3b933684-ac19-44b0-a49d-6af45501e166] Took 27.15 seconds to build instance. [ 663.863672] env[65726]: DEBUG oslo_concurrency.lockutils [None req-36828e27-b100-49cc-98c2-ad0b25ca6572 tempest-ServerGroupTestJSON-125182767 tempest-ServerGroupTestJSON-125182767-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.127s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 663.867918] env[65726]: DEBUG oslo_concurrency.lockutils [None req-fb0f14b3-fd50-4c8d-a984-877b235dcb32 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.695s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 663.867918] env[65726]: INFO nova.compute.claims [None req-fb0f14b3-fd50-4c8d-a984-877b235dcb32 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] [instance: d7905c4f-1ba3-43b3-b4b3-aebacd2f9cc3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 663.904027] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9b3c93a2-43e7-4858-b887-85486b2f5f2a tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Acquiring lock "0e064341-4e4a-407b-8c26-3eb04b409029" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 663.904027] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9b3c93a2-43e7-4858-b887-85486b2f5f2a tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Lock "0e064341-4e4a-407b-8c26-3eb04b409029" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 663.904027] env[65726]: DEBUG nova.compute.manager [None req-9b3c93a2-43e7-4858-b887-85486b2f5f2a tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] [instance: 0e064341-4e4a-407b-8c26-3eb04b409029] Going to confirm migration 1 {{(pid=65726) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:5307}} [ 664.462868] env[65726]: DEBUG oslo_concurrency.lockutils [None req-362ec7a6-07fd-4c6b-b5fb-9865bdcb8c1a tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Releasing lock "refresh_cache-274ab469-61a9-4b7e-852c-074c871e3abf" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 664.462868] env[65726]: DEBUG nova.compute.manager [None req-362ec7a6-07fd-4c6b-b5fb-9865bdcb8c1a tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: 274ab469-61a9-4b7e-852c-074c871e3abf] Instance network_info: |[{"id": "757d86b3-4474-41c5-aaf8-14fc7216cf15", "address": "fa:16:3e:b9:96:d6", "network": {"id": "3f7916f7-ee9c-4629-b3b3-1367de9f98ea", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-579713617-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b794c99309e94c1181fc9d8eeeb84702", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d0bdd11b-58af-4cc0-9d38-8322e1bb4e74", "external-id": "nsx-vlan-transportzone-398", "segmentation_id": 398, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap757d86b3-44", "ovs_interfaceid": "757d86b3-4474-41c5-aaf8-14fc7216cf15", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 664.465919] env[65726]: WARNING openstack [None req-621af3d8-d8b6-4b45-86f5-d66b91010e2e tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 664.465919] env[65726]: WARNING openstack [None req-621af3d8-d8b6-4b45-86f5-d66b91010e2e tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 664.465919] env[65726]: DEBUG nova.network.neutron [None req-621af3d8-d8b6-4b45-86f5-d66b91010e2e tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] [instance: fd481728-3ef0-4a96-affd-ab2dd3f596bb] Instance cache missing network info. {{(pid=65726) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 664.470427] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-362ec7a6-07fd-4c6b-b5fb-9865bdcb8c1a tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: 274ab469-61a9-4b7e-852c-074c871e3abf] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b9:96:d6', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd0bdd11b-58af-4cc0-9d38-8322e1bb4e74', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '757d86b3-4474-41c5-aaf8-14fc7216cf15', 'vif_model': 'vmxnet3'}] {{(pid=65726) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 664.470427] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-362ec7a6-07fd-4c6b-b5fb-9865bdcb8c1a tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 664.470427] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 274ab469-61a9-4b7e-852c-074c871e3abf] Creating VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 664.470427] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-16c2fe6b-527a-4f0c-a000-2705d5a4a98f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.470427] env[65726]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 664.470427] env[65726]: value = "task-5115542" [ 664.470427] env[65726]: _type = "Task" [ 664.470427] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 664.473480] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115542, 'name': CreateVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 664.473480] env[65726]: DEBUG oslo_concurrency.lockutils [None req-40525054-994c-41ab-bed7-c877834cae18 tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] Lock "3b933684-ac19-44b0-a49d-6af45501e166" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 47.895s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 664.473480] env[65726]: DEBUG oslo_concurrency.lockutils [None req-36828e27-b100-49cc-98c2-ad0b25ca6572 tempest-ServerGroupTestJSON-125182767 tempest-ServerGroupTestJSON-125182767-project-member] Acquiring lock "c2a697ce-7305-4c27-bbba-a0aa22cc0978" by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 664.473480] env[65726]: DEBUG oslo_concurrency.lockutils [None req-36828e27-b100-49cc-98c2-ad0b25ca6572 tempest-ServerGroupTestJSON-125182767 tempest-ServerGroupTestJSON-125182767-project-member] Lock "c2a697ce-7305-4c27-bbba-a0aa22cc0978" acquired by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" :: waited 0.001s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 664.474419] env[65726]: WARNING openstack [None req-9b3c93a2-43e7-4858-b887-85486b2f5f2a tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 664.474419] env[65726]: WARNING openstack [None req-9b3c93a2-43e7-4858-b887-85486b2f5f2a tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 664.474419] env[65726]: WARNING openstack [None req-621af3d8-d8b6-4b45-86f5-d66b91010e2e tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 664.474419] env[65726]: WARNING openstack [None req-621af3d8-d8b6-4b45-86f5-d66b91010e2e tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 664.502576] env[65726]: WARNING openstack [None req-9b3c93a2-43e7-4858-b887-85486b2f5f2a tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 664.503034] env[65726]: WARNING openstack [None req-9b3c93a2-43e7-4858-b887-85486b2f5f2a tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 664.510101] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9b3c93a2-43e7-4858-b887-85486b2f5f2a tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Acquiring lock "refresh_cache-0e064341-4e4a-407b-8c26-3eb04b409029" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 664.511478] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9b3c93a2-43e7-4858-b887-85486b2f5f2a tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Acquired lock "refresh_cache-0e064341-4e4a-407b-8c26-3eb04b409029" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 664.511478] env[65726]: DEBUG nova.network.neutron [None req-9b3c93a2-43e7-4858-b887-85486b2f5f2a tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] [instance: 0e064341-4e4a-407b-8c26-3eb04b409029] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 664.511478] env[65726]: DEBUG nova.objects.instance [None req-9b3c93a2-43e7-4858-b887-85486b2f5f2a tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Lazy-loading 'info_cache' on Instance uuid 0e064341-4e4a-407b-8c26-3eb04b409029 {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 664.582907] env[65726]: WARNING openstack [None req-621af3d8-d8b6-4b45-86f5-d66b91010e2e tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 664.585270] env[65726]: WARNING openstack [None req-621af3d8-d8b6-4b45-86f5-d66b91010e2e tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 664.648953] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c24265b5-a0c6-4266-b4c0-64535a550ddd tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Acquiring lock "bd839fe1-8801-4ba1-9c23-288453258b66" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 664.650154] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c24265b5-a0c6-4266-b4c0-64535a550ddd tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Lock "bd839fe1-8801-4ba1-9c23-288453258b66" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 664.733847] env[65726]: DEBUG nova.network.neutron [None req-621af3d8-d8b6-4b45-86f5-d66b91010e2e tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] [instance: fd481728-3ef0-4a96-affd-ab2dd3f596bb] Updating instance_info_cache with network_info: [{"id": "7df3f02b-e4e5-4840-bf9f-3498771b9184", "address": "fa:16:3e:6b:c1:e9", "network": {"id": "d37113a1-409c-4c2c-81f9-8f80822b85b9", "bridge": "br-int", "label": "tempest-ServersWithSpecificFlavorTestJSON-471231219-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "79d3e960f190439fae7c61d693e9356e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "13af9422-d668-4413-b63a-766558d83a3b", "external-id": "nsx-vlan-transportzone-842", "segmentation_id": 842, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7df3f02b-e4", "ovs_interfaceid": "7df3f02b-e4e5-4840-bf9f-3498771b9184", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 665.494036] env[65726]: DEBUG nova.compute.manager [None req-f3878371-4599-418c-a420-84e54e420a24 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] [instance: 423af936-a383-4652-8887-9c8b8f507909] Starting instance... {{(pid=65726) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 665.499019] env[65726]: DEBUG oslo_concurrency.lockutils [None req-36828e27-b100-49cc-98c2-ad0b25ca6572 tempest-ServerGroupTestJSON-125182767 tempest-ServerGroupTestJSON-125182767-project-member] Lock "c2a697ce-7305-4c27-bbba-a0aa22cc0978" "released" by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" :: held 1.124s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 665.499019] env[65726]: DEBUG nova.compute.manager [None req-36828e27-b100-49cc-98c2-ad0b25ca6572 tempest-ServerGroupTestJSON-125182767 tempest-ServerGroupTestJSON-125182767-project-member] [instance: c2929727-a7ea-4f63-997d-474c29a305fc] Start building networks asynchronously for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 665.510280] env[65726]: DEBUG oslo_concurrency.lockutils [None req-621af3d8-d8b6-4b45-86f5-d66b91010e2e tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Releasing lock "refresh_cache-fd481728-3ef0-4a96-affd-ab2dd3f596bb" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 665.510578] env[65726]: DEBUG nova.compute.manager [None req-621af3d8-d8b6-4b45-86f5-d66b91010e2e tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] [instance: fd481728-3ef0-4a96-affd-ab2dd3f596bb] Instance network_info: |[{"id": "7df3f02b-e4e5-4840-bf9f-3498771b9184", "address": "fa:16:3e:6b:c1:e9", "network": {"id": "d37113a1-409c-4c2c-81f9-8f80822b85b9", "bridge": "br-int", "label": "tempest-ServersWithSpecificFlavorTestJSON-471231219-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "79d3e960f190439fae7c61d693e9356e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "13af9422-d668-4413-b63a-766558d83a3b", "external-id": "nsx-vlan-transportzone-842", "segmentation_id": 842, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7df3f02b-e4", "ovs_interfaceid": "7df3f02b-e4e5-4840-bf9f-3498771b9184", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 665.510843] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115542, 'name': CreateVM_Task} progress is 99%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 665.510970] env[65726]: WARNING oslo_vmware.common.loopingcall [-] task run outlasted interval by 0.25475099999999995 sec [ 665.512881] env[65726]: DEBUG nova.compute.manager [req-8b636882-6005-4ef8-a9e8-0be7180a48a1 req-bc040676-7517-4749-8c51-095b0251a8b6 service nova] [instance: 274ab469-61a9-4b7e-852c-074c871e3abf] Received event network-vif-plugged-757d86b3-4474-41c5-aaf8-14fc7216cf15 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 665.513163] env[65726]: DEBUG oslo_concurrency.lockutils [req-8b636882-6005-4ef8-a9e8-0be7180a48a1 req-bc040676-7517-4749-8c51-095b0251a8b6 service nova] Acquiring lock "274ab469-61a9-4b7e-852c-074c871e3abf-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 665.513210] env[65726]: DEBUG oslo_concurrency.lockutils [req-8b636882-6005-4ef8-a9e8-0be7180a48a1 req-bc040676-7517-4749-8c51-095b0251a8b6 service nova] Lock "274ab469-61a9-4b7e-852c-074c871e3abf-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 665.516019] env[65726]: DEBUG oslo_concurrency.lockutils [req-8b636882-6005-4ef8-a9e8-0be7180a48a1 req-bc040676-7517-4749-8c51-095b0251a8b6 service nova] Lock "274ab469-61a9-4b7e-852c-074c871e3abf-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 665.516019] env[65726]: DEBUG nova.compute.manager [req-8b636882-6005-4ef8-a9e8-0be7180a48a1 req-bc040676-7517-4749-8c51-095b0251a8b6 service nova] [instance: 274ab469-61a9-4b7e-852c-074c871e3abf] No waiting events found dispatching network-vif-plugged-757d86b3-4474-41c5-aaf8-14fc7216cf15 {{(pid=65726) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 665.516019] env[65726]: WARNING nova.compute.manager [req-8b636882-6005-4ef8-a9e8-0be7180a48a1 req-bc040676-7517-4749-8c51-095b0251a8b6 service nova] [instance: 274ab469-61a9-4b7e-852c-074c871e3abf] Received unexpected event network-vif-plugged-757d86b3-4474-41c5-aaf8-14fc7216cf15 for instance with vm_state building and task_state spawning. [ 665.516019] env[65726]: DEBUG nova.compute.manager [req-8b636882-6005-4ef8-a9e8-0be7180a48a1 req-bc040676-7517-4749-8c51-095b0251a8b6 service nova] [instance: 274ab469-61a9-4b7e-852c-074c871e3abf] Received event network-changed-757d86b3-4474-41c5-aaf8-14fc7216cf15 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 665.516019] env[65726]: DEBUG nova.compute.manager [req-8b636882-6005-4ef8-a9e8-0be7180a48a1 req-bc040676-7517-4749-8c51-095b0251a8b6 service nova] [instance: 274ab469-61a9-4b7e-852c-074c871e3abf] Refreshing instance network info cache due to event network-changed-757d86b3-4474-41c5-aaf8-14fc7216cf15. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 665.516318] env[65726]: DEBUG oslo_concurrency.lockutils [req-8b636882-6005-4ef8-a9e8-0be7180a48a1 req-bc040676-7517-4749-8c51-095b0251a8b6 service nova] Acquiring lock "refresh_cache-274ab469-61a9-4b7e-852c-074c871e3abf" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 665.516318] env[65726]: DEBUG oslo_concurrency.lockutils [req-8b636882-6005-4ef8-a9e8-0be7180a48a1 req-bc040676-7517-4749-8c51-095b0251a8b6 service nova] Acquired lock "refresh_cache-274ab469-61a9-4b7e-852c-074c871e3abf" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 665.516318] env[65726]: DEBUG nova.network.neutron [req-8b636882-6005-4ef8-a9e8-0be7180a48a1 req-bc040676-7517-4749-8c51-095b0251a8b6 service nova] [instance: 274ab469-61a9-4b7e-852c-074c871e3abf] Refreshing network info cache for port 757d86b3-4474-41c5-aaf8-14fc7216cf15 {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 665.516318] env[65726]: DEBUG nova.compute.manager [req-50a0a1bd-8b36-447a-b018-324091111ac2 req-af6744f1-5027-4605-9cb7-199c94378ca1 service nova] [instance: 8e9c27f6-e9fa-4ed0-b6ae-3b5df52d1c44] Received event network-vif-deleted-cc1dce8f-441c-42a7-b3ba-7f3d3fc9ca66 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 665.516975] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-621af3d8-d8b6-4b45-86f5-d66b91010e2e tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] [instance: fd481728-3ef0-4a96-affd-ab2dd3f596bb] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:6b:c1:e9', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '13af9422-d668-4413-b63a-766558d83a3b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7df3f02b-e4e5-4840-bf9f-3498771b9184', 'vif_model': 'vmxnet3'}] {{(pid=65726) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 665.524352] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-621af3d8-d8b6-4b45-86f5-d66b91010e2e tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Creating folder: Project (79d3e960f190439fae7c61d693e9356e). Parent ref: group-v995008. {{(pid=65726) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 665.526453] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9407779b-08dd-4f06-81eb-dae9421e8c43 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.538178] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115542, 'name': CreateVM_Task, 'duration_secs': 0.551779} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 665.538466] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 274ab469-61a9-4b7e-852c-074c871e3abf] Created VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 665.539502] env[65726]: WARNING openstack [None req-362ec7a6-07fd-4c6b-b5fb-9865bdcb8c1a tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 665.539855] env[65726]: WARNING openstack [None req-362ec7a6-07fd-4c6b-b5fb-9865bdcb8c1a tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 665.544910] env[65726]: DEBUG oslo_concurrency.lockutils [None req-362ec7a6-07fd-4c6b-b5fb-9865bdcb8c1a tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 665.545074] env[65726]: DEBUG oslo_concurrency.lockutils [None req-362ec7a6-07fd-4c6b-b5fb-9865bdcb8c1a tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 665.545393] env[65726]: DEBUG oslo_concurrency.lockutils [None req-362ec7a6-07fd-4c6b-b5fb-9865bdcb8c1a tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 665.550527] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a35aebb1-03f2-47cb-9087-40d88c5a770f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.552857] env[65726]: INFO nova.virt.vmwareapi.vm_util [None req-621af3d8-d8b6-4b45-86f5-d66b91010e2e tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Created folder: Project (79d3e960f190439fae7c61d693e9356e) in parent group-v995008. [ 665.553041] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-621af3d8-d8b6-4b45-86f5-d66b91010e2e tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Creating folder: Instances. Parent ref: group-v995063. {{(pid=65726) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 665.553654] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5f527ce8-7b26-45bf-9632-e4a77c3d10ad {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.560370] env[65726]: DEBUG oslo_vmware.api [None req-362ec7a6-07fd-4c6b-b5fb-9865bdcb8c1a tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Waiting for the task: (returnval){ [ 665.560370] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]521c6d7e-c824-8fbf-53c6-e14f2bc16d7a" [ 665.560370] env[65726]: _type = "Task" [ 665.560370] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 665.565980] env[65726]: INFO nova.virt.vmwareapi.vm_util [None req-621af3d8-d8b6-4b45-86f5-d66b91010e2e tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Created folder: Instances in parent group-v995063. [ 665.566277] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-621af3d8-d8b6-4b45-86f5-d66b91010e2e tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 665.566469] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fd481728-3ef0-4a96-affd-ab2dd3f596bb] Creating VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 665.569983] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6d64d62b-f15f-401d-9e80-38f9fc21a17b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.587733] env[65726]: DEBUG oslo_vmware.api [None req-362ec7a6-07fd-4c6b-b5fb-9865bdcb8c1a tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]521c6d7e-c824-8fbf-53c6-e14f2bc16d7a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 665.592945] env[65726]: DEBUG nova.network.neutron [None req-06f67a27-e6e0-4078-804d-cff1137abcad tempest-ServersTestBootFromVolume-762458528 tempest-ServersTestBootFromVolume-762458528-project-member] [instance: 5a8d4357-4b44-4a19-b1da-42d188c38adc] Successfully updated port: 9aaacd82-31c1-4cfa-bbce-860b30db74af {{(pid=65726) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 665.601952] env[65726]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 665.601952] env[65726]: value = "task-5115545" [ 665.601952] env[65726]: _type = "Task" [ 665.601952] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 665.613279] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115545, 'name': CreateVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 666.012717] env[65726]: WARNING openstack [None req-9b3c93a2-43e7-4858-b887-85486b2f5f2a tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 666.013320] env[65726]: WARNING openstack [None req-9b3c93a2-43e7-4858-b887-85486b2f5f2a tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 666.028072] env[65726]: DEBUG nova.compute.utils [None req-36828e27-b100-49cc-98c2-ad0b25ca6572 tempest-ServerGroupTestJSON-125182767 tempest-ServerGroupTestJSON-125182767-project-member] Using /dev/sd instead of None {{(pid=65726) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 666.030266] env[65726]: WARNING openstack [req-8b636882-6005-4ef8-a9e8-0be7180a48a1 req-bc040676-7517-4749-8c51-095b0251a8b6 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 666.030716] env[65726]: WARNING openstack [req-8b636882-6005-4ef8-a9e8-0be7180a48a1 req-bc040676-7517-4749-8c51-095b0251a8b6 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 666.041606] env[65726]: DEBUG nova.compute.manager [None req-36828e27-b100-49cc-98c2-ad0b25ca6572 tempest-ServerGroupTestJSON-125182767 tempest-ServerGroupTestJSON-125182767-project-member] [instance: c2929727-a7ea-4f63-997d-474c29a305fc] Allocating IP information in the background. {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 666.041875] env[65726]: DEBUG nova.network.neutron [None req-36828e27-b100-49cc-98c2-ad0b25ca6572 tempest-ServerGroupTestJSON-125182767 tempest-ServerGroupTestJSON-125182767-project-member] [instance: c2929727-a7ea-4f63-997d-474c29a305fc] allocate_for_instance() {{(pid=65726) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 666.042281] env[65726]: WARNING neutronclient.v2_0.client [None req-36828e27-b100-49cc-98c2-ad0b25ca6572 tempest-ServerGroupTestJSON-125182767 tempest-ServerGroupTestJSON-125182767-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 666.043192] env[65726]: WARNING neutronclient.v2_0.client [None req-36828e27-b100-49cc-98c2-ad0b25ca6572 tempest-ServerGroupTestJSON-125182767 tempest-ServerGroupTestJSON-125182767-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 666.043192] env[65726]: WARNING openstack [None req-36828e27-b100-49cc-98c2-ad0b25ca6572 tempest-ServerGroupTestJSON-125182767 tempest-ServerGroupTestJSON-125182767-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 666.043536] env[65726]: WARNING openstack [None req-36828e27-b100-49cc-98c2-ad0b25ca6572 tempest-ServerGroupTestJSON-125182767 tempest-ServerGroupTestJSON-125182767-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 666.056058] env[65726]: DEBUG nova.compute.manager [None req-06f67a27-e6e0-4078-804d-cff1137abcad tempest-ServersTestBootFromVolume-762458528 tempest-ServersTestBootFromVolume-762458528-project-member] [instance: 5a8d4357-4b44-4a19-b1da-42d188c38adc] Start spawning the instance on the hypervisor. {{(pid=65726) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 666.056058] env[65726]: DEBUG nova.virt.hardware [None req-06f67a27-e6e0-4078-804d-cff1137abcad tempest-ServersTestBootFromVolume-762458528 tempest-ServersTestBootFromVolume-762458528-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 666.056058] env[65726]: DEBUG nova.virt.hardware [None req-06f67a27-e6e0-4078-804d-cff1137abcad tempest-ServersTestBootFromVolume-762458528 tempest-ServersTestBootFromVolume-762458528-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 666.056058] env[65726]: DEBUG nova.virt.hardware [None req-06f67a27-e6e0-4078-804d-cff1137abcad tempest-ServersTestBootFromVolume-762458528 tempest-ServersTestBootFromVolume-762458528-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 666.056283] env[65726]: DEBUG nova.virt.hardware [None req-06f67a27-e6e0-4078-804d-cff1137abcad tempest-ServersTestBootFromVolume-762458528 tempest-ServersTestBootFromVolume-762458528-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 666.056283] env[65726]: DEBUG nova.virt.hardware [None req-06f67a27-e6e0-4078-804d-cff1137abcad tempest-ServersTestBootFromVolume-762458528 tempest-ServersTestBootFromVolume-762458528-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 666.056283] env[65726]: DEBUG nova.virt.hardware [None req-06f67a27-e6e0-4078-804d-cff1137abcad tempest-ServersTestBootFromVolume-762458528 tempest-ServersTestBootFromVolume-762458528-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 666.058640] env[65726]: DEBUG nova.virt.hardware [None req-06f67a27-e6e0-4078-804d-cff1137abcad tempest-ServersTestBootFromVolume-762458528 tempest-ServersTestBootFromVolume-762458528-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 666.058640] env[65726]: DEBUG nova.virt.hardware [None req-06f67a27-e6e0-4078-804d-cff1137abcad tempest-ServersTestBootFromVolume-762458528 tempest-ServersTestBootFromVolume-762458528-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 666.058640] env[65726]: DEBUG nova.virt.hardware [None req-06f67a27-e6e0-4078-804d-cff1137abcad tempest-ServersTestBootFromVolume-762458528 tempest-ServersTestBootFromVolume-762458528-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 666.058640] env[65726]: DEBUG nova.virt.hardware [None req-06f67a27-e6e0-4078-804d-cff1137abcad tempest-ServersTestBootFromVolume-762458528 tempest-ServersTestBootFromVolume-762458528-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 666.058640] env[65726]: DEBUG nova.virt.hardware [None req-06f67a27-e6e0-4078-804d-cff1137abcad tempest-ServersTestBootFromVolume-762458528 tempest-ServersTestBootFromVolume-762458528-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 666.061910] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01b016f3-5d23-4cae-baf9-828274f591bc {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.066419] env[65726]: DEBUG oslo_concurrency.lockutils [None req-f3878371-4599-418c-a420-84e54e420a24 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 666.084688] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-baf8e8d1-f301-43a7-9a9f-0047e670f364 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.094725] env[65726]: DEBUG oslo_vmware.api [None req-362ec7a6-07fd-4c6b-b5fb-9865bdcb8c1a tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]521c6d7e-c824-8fbf-53c6-e14f2bc16d7a, 'name': SearchDatastore_Task, 'duration_secs': 0.018389} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 666.096120] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5366109-8ac1-4f2e-8f1b-3449324e3f5b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.103574] env[65726]: DEBUG oslo_concurrency.lockutils [None req-362ec7a6-07fd-4c6b-b5fb-9865bdcb8c1a tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 666.103824] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-362ec7a6-07fd-4c6b-b5fb-9865bdcb8c1a tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: 274ab469-61a9-4b7e-852c-074c871e3abf] Processing image b52362a3-ee8a-4cbf-b06f-513b0cc8f95c {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 666.104406] env[65726]: DEBUG oslo_concurrency.lockutils [None req-362ec7a6-07fd-4c6b-b5fb-9865bdcb8c1a tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 666.104406] env[65726]: DEBUG oslo_concurrency.lockutils [None req-362ec7a6-07fd-4c6b-b5fb-9865bdcb8c1a tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 666.104506] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-362ec7a6-07fd-4c6b-b5fb-9865bdcb8c1a tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 666.105080] env[65726]: DEBUG oslo_concurrency.lockutils [None req-06f67a27-e6e0-4078-804d-cff1137abcad tempest-ServersTestBootFromVolume-762458528 tempest-ServersTestBootFromVolume-762458528-project-member] Acquiring lock "refresh_cache-5a8d4357-4b44-4a19-b1da-42d188c38adc" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 666.105283] env[65726]: DEBUG oslo_concurrency.lockutils [None req-06f67a27-e6e0-4078-804d-cff1137abcad tempest-ServersTestBootFromVolume-762458528 tempest-ServersTestBootFromVolume-762458528-project-member] Acquired lock "refresh_cache-5a8d4357-4b44-4a19-b1da-42d188c38adc" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 666.105479] env[65726]: DEBUG nova.network.neutron [None req-06f67a27-e6e0-4078-804d-cff1137abcad tempest-ServersTestBootFromVolume-762458528 tempest-ServersTestBootFromVolume-762458528-project-member] [instance: 5a8d4357-4b44-4a19-b1da-42d188c38adc] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 666.107673] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ff7cff1b-f78c-44d1-b186-36a4fdb6738b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.128410] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02c4d1ff-cced-4337-9f77-797994132e61 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.138290] env[65726]: DEBUG nova.policy [None req-36828e27-b100-49cc-98c2-ad0b25ca6572 tempest-ServerGroupTestJSON-125182767 tempest-ServerGroupTestJSON-125182767-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1f6ed7f5930d4a5fbde6625cce66f675', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '34534927100c4f7a91a1073de9047674', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65726) authorize /opt/stack/nova/nova/policy.py:192}} [ 666.141823] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115545, 'name': CreateVM_Task, 'duration_secs': 0.422532} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 666.143889] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fd481728-3ef0-4a96-affd-ab2dd3f596bb] Created VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 666.144146] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-362ec7a6-07fd-4c6b-b5fb-9865bdcb8c1a tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 666.144314] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-362ec7a6-07fd-4c6b-b5fb-9865bdcb8c1a tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65726) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 666.145862] env[65726]: WARNING openstack [None req-621af3d8-d8b6-4b45-86f5-d66b91010e2e tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 666.146220] env[65726]: WARNING openstack [None req-621af3d8-d8b6-4b45-86f5-d66b91010e2e tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 666.151491] env[65726]: DEBUG oslo_concurrency.lockutils [None req-621af3d8-d8b6-4b45-86f5-d66b91010e2e tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 666.151622] env[65726]: DEBUG oslo_concurrency.lockutils [None req-621af3d8-d8b6-4b45-86f5-d66b91010e2e tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 666.151919] env[65726]: DEBUG oslo_concurrency.lockutils [None req-621af3d8-d8b6-4b45-86f5-d66b91010e2e tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 666.152205] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-265b4f28-e99c-4969-be27-ac18b398b301 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.184219] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-99cdd766-edec-40f9-9765-4b19143371d7 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.191361] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8195f381-de3f-42d7-8c7a-651a1a42c728 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.196672] env[65726]: DEBUG oslo_vmware.api [None req-362ec7a6-07fd-4c6b-b5fb-9865bdcb8c1a tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Waiting for the task: (returnval){ [ 666.196672] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]526fa25c-4916-dec9-72d7-59817ba0f8c3" [ 666.196672] env[65726]: _type = "Task" [ 666.196672] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 666.204899] env[65726]: DEBUG oslo_vmware.api [None req-621af3d8-d8b6-4b45-86f5-d66b91010e2e tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Waiting for the task: (returnval){ [ 666.204899] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]523996d3-26a5-bc70-0463-0684e4421cf4" [ 666.204899] env[65726]: _type = "Task" [ 666.204899] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 666.206633] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66b22c52-7928-4ad1-849b-fffc6a77c11c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.217944] env[65726]: DEBUG oslo_vmware.api [None req-362ec7a6-07fd-4c6b-b5fb-9865bdcb8c1a tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]526fa25c-4916-dec9-72d7-59817ba0f8c3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 666.219484] env[65726]: WARNING openstack [None req-9b3c93a2-43e7-4858-b887-85486b2f5f2a tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 666.219887] env[65726]: WARNING openstack [None req-9b3c93a2-43e7-4858-b887-85486b2f5f2a tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 666.237219] env[65726]: DEBUG nova.compute.provider_tree [None req-fb0f14b3-fd50-4c8d-a984-877b235dcb32 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 666.243141] env[65726]: WARNING openstack [req-8b636882-6005-4ef8-a9e8-0be7180a48a1 req-bc040676-7517-4749-8c51-095b0251a8b6 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 666.243508] env[65726]: WARNING openstack [req-8b636882-6005-4ef8-a9e8-0be7180a48a1 req-bc040676-7517-4749-8c51-095b0251a8b6 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 666.250165] env[65726]: DEBUG oslo_vmware.api [None req-621af3d8-d8b6-4b45-86f5-d66b91010e2e tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]523996d3-26a5-bc70-0463-0684e4421cf4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 666.252020] env[65726]: DEBUG nova.scheduler.client.report [None req-fb0f14b3-fd50-4c8d-a984-877b235dcb32 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 666.365401] env[65726]: WARNING openstack [None req-9b3c93a2-43e7-4858-b887-85486b2f5f2a tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 666.365401] env[65726]: WARNING openstack [None req-9b3c93a2-43e7-4858-b887-85486b2f5f2a tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 666.417593] env[65726]: WARNING openstack [req-8b636882-6005-4ef8-a9e8-0be7180a48a1 req-bc040676-7517-4749-8c51-095b0251a8b6 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 666.417967] env[65726]: WARNING openstack [req-8b636882-6005-4ef8-a9e8-0be7180a48a1 req-bc040676-7517-4749-8c51-095b0251a8b6 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 666.476940] env[65726]: DEBUG nova.network.neutron [None req-36828e27-b100-49cc-98c2-ad0b25ca6572 tempest-ServerGroupTestJSON-125182767 tempest-ServerGroupTestJSON-125182767-project-member] [instance: c2929727-a7ea-4f63-997d-474c29a305fc] Successfully created port: 27cddbbc-b8ac-4cf0-8baf-d03643c3de8d {{(pid=65726) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 666.537681] env[65726]: DEBUG nova.network.neutron [None req-9b3c93a2-43e7-4858-b887-85486b2f5f2a tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] [instance: 0e064341-4e4a-407b-8c26-3eb04b409029] Updating instance_info_cache with network_info: [{"id": "9d3c5bc9-aa25-4013-b8dc-f565887d2736", "address": "fa:16:3e:f9:95:de", "network": {"id": "55d629b8-a541-4720-9289-927bbc6e6f22", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.93", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "1539f06fa3534e90acbc3a60c4b8bd3f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "35342bcb-8b06-472e-b3c0-43fd3d6c4b30", "external-id": "nsx-vlan-transportzone-524", "segmentation_id": 524, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9d3c5bc9-aa", "ovs_interfaceid": "9d3c5bc9-aa25-4013-b8dc-f565887d2736", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 666.542214] env[65726]: DEBUG nova.compute.manager [None req-36828e27-b100-49cc-98c2-ad0b25ca6572 tempest-ServerGroupTestJSON-125182767 tempest-ServerGroupTestJSON-125182767-project-member] [instance: c2929727-a7ea-4f63-997d-474c29a305fc] Start building block device mappings for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 666.611684] env[65726]: DEBUG oslo_concurrency.lockutils [None req-2737dc3f-715d-481c-b9a9-998472382a0d tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] Acquiring lock "83573c2b-9448-456f-8fd2-b19661dd6cc4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 666.611921] env[65726]: DEBUG oslo_concurrency.lockutils [None req-2737dc3f-715d-481c-b9a9-998472382a0d tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] Lock "83573c2b-9448-456f-8fd2-b19661dd6cc4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 666.624113] env[65726]: WARNING openstack [None req-06f67a27-e6e0-4078-804d-cff1137abcad tempest-ServersTestBootFromVolume-762458528 tempest-ServersTestBootFromVolume-762458528-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 666.624481] env[65726]: WARNING openstack [None req-06f67a27-e6e0-4078-804d-cff1137abcad tempest-ServersTestBootFromVolume-762458528 tempest-ServersTestBootFromVolume-762458528-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 666.629578] env[65726]: DEBUG nova.network.neutron [None req-06f67a27-e6e0-4078-804d-cff1137abcad tempest-ServersTestBootFromVolume-762458528 tempest-ServersTestBootFromVolume-762458528-project-member] [instance: 5a8d4357-4b44-4a19-b1da-42d188c38adc] Instance cache missing network info. {{(pid=65726) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 666.711775] env[65726]: DEBUG oslo_vmware.api [None req-362ec7a6-07fd-4c6b-b5fb-9865bdcb8c1a tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]526fa25c-4916-dec9-72d7-59817ba0f8c3, 'name': SearchDatastore_Task, 'duration_secs': 0.021132} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 666.716716] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-482d5eb0-1f73-4c7c-9c5e-97fd9cd6b087 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.723593] env[65726]: DEBUG oslo_vmware.api [None req-362ec7a6-07fd-4c6b-b5fb-9865bdcb8c1a tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Waiting for the task: (returnval){ [ 666.723593] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5298f225-e2e9-686f-78c9-a5b0b8dad923" [ 666.723593] env[65726]: _type = "Task" [ 666.723593] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 666.727772] env[65726]: DEBUG oslo_vmware.api [None req-621af3d8-d8b6-4b45-86f5-d66b91010e2e tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]523996d3-26a5-bc70-0463-0684e4421cf4, 'name': SearchDatastore_Task, 'duration_secs': 0.04239} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 666.731657] env[65726]: DEBUG oslo_concurrency.lockutils [None req-621af3d8-d8b6-4b45-86f5-d66b91010e2e tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 666.731909] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-621af3d8-d8b6-4b45-86f5-d66b91010e2e tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] [instance: fd481728-3ef0-4a96-affd-ab2dd3f596bb] Processing image b52362a3-ee8a-4cbf-b06f-513b0cc8f95c {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 666.732496] env[65726]: DEBUG oslo_concurrency.lockutils [None req-621af3d8-d8b6-4b45-86f5-d66b91010e2e tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 666.732656] env[65726]: DEBUG oslo_concurrency.lockutils [None req-621af3d8-d8b6-4b45-86f5-d66b91010e2e tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 666.733067] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-621af3d8-d8b6-4b45-86f5-d66b91010e2e tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 666.733175] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-96bd99b8-0257-4d2b-a7bf-66cb27c31fac {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.742940] env[65726]: DEBUG oslo_vmware.api [None req-362ec7a6-07fd-4c6b-b5fb-9865bdcb8c1a tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5298f225-e2e9-686f-78c9-a5b0b8dad923, 'name': SearchDatastore_Task, 'duration_secs': 0.017816} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 666.743407] env[65726]: DEBUG oslo_concurrency.lockutils [None req-362ec7a6-07fd-4c6b-b5fb-9865bdcb8c1a tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 666.743725] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-362ec7a6-07fd-4c6b-b5fb-9865bdcb8c1a tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore2] 274ab469-61a9-4b7e-852c-074c871e3abf/274ab469-61a9-4b7e-852c-074c871e3abf.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 666.744086] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-af59ee67-4f31-49e2-8525-1e887f5bdb56 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.749310] env[65726]: DEBUG nova.network.neutron [req-8b636882-6005-4ef8-a9e8-0be7180a48a1 req-bc040676-7517-4749-8c51-095b0251a8b6 service nova] [instance: 274ab469-61a9-4b7e-852c-074c871e3abf] Updated VIF entry in instance network info cache for port 757d86b3-4474-41c5-aaf8-14fc7216cf15. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 666.749584] env[65726]: DEBUG nova.network.neutron [req-8b636882-6005-4ef8-a9e8-0be7180a48a1 req-bc040676-7517-4749-8c51-095b0251a8b6 service nova] [instance: 274ab469-61a9-4b7e-852c-074c871e3abf] Updating instance_info_cache with network_info: [{"id": "757d86b3-4474-41c5-aaf8-14fc7216cf15", "address": "fa:16:3e:b9:96:d6", "network": {"id": "3f7916f7-ee9c-4629-b3b3-1367de9f98ea", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-579713617-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b794c99309e94c1181fc9d8eeeb84702", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d0bdd11b-58af-4cc0-9d38-8322e1bb4e74", "external-id": "nsx-vlan-transportzone-398", "segmentation_id": 398, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap757d86b3-44", "ovs_interfaceid": "757d86b3-4474-41c5-aaf8-14fc7216cf15", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 666.750885] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-621af3d8-d8b6-4b45-86f5-d66b91010e2e tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 666.751028] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-621af3d8-d8b6-4b45-86f5-d66b91010e2e tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65726) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 666.752593] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9024d4d5-f82c-47a3-8be9-b826cd629bc6 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.755117] env[65726]: DEBUG oslo_service.periodic_task [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=65726) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 666.755397] env[65726]: DEBUG oslo_service.periodic_task [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=65726) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 666.759983] env[65726]: DEBUG oslo_concurrency.lockutils [None req-fb0f14b3-fd50-4c8d-a984-877b235dcb32 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.893s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 666.759983] env[65726]: DEBUG nova.compute.manager [None req-fb0f14b3-fd50-4c8d-a984-877b235dcb32 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] [instance: d7905c4f-1ba3-43b3-b4b3-aebacd2f9cc3] Start building networks asynchronously for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 666.762200] env[65726]: DEBUG oslo_vmware.api [None req-362ec7a6-07fd-4c6b-b5fb-9865bdcb8c1a tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Waiting for the task: (returnval){ [ 666.762200] env[65726]: value = "task-5115546" [ 666.762200] env[65726]: _type = "Task" [ 666.762200] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 666.766139] env[65726]: DEBUG oslo_concurrency.lockutils [None req-4b216162-e1e7-4270-8d1f-482978b0681a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 16.677s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 666.766139] env[65726]: DEBUG nova.objects.instance [None req-4b216162-e1e7-4270-8d1f-482978b0681a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Lazy-loading 'resources' on Instance uuid 8493ba52-b950-4727-970a-19b1797b2ca1 {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 666.766139] env[65726]: DEBUG oslo_vmware.api [None req-621af3d8-d8b6-4b45-86f5-d66b91010e2e tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Waiting for the task: (returnval){ [ 666.766139] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52432eb1-6918-8737-af8b-aa9e09ab4e65" [ 666.766139] env[65726]: _type = "Task" [ 666.766139] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 666.779834] env[65726]: DEBUG oslo_vmware.api [None req-362ec7a6-07fd-4c6b-b5fb-9865bdcb8c1a tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Task: {'id': task-5115546, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 666.784252] env[65726]: DEBUG oslo_vmware.api [None req-621af3d8-d8b6-4b45-86f5-d66b91010e2e tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52432eb1-6918-8737-af8b-aa9e09ab4e65, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 666.827878] env[65726]: WARNING openstack [None req-06f67a27-e6e0-4078-804d-cff1137abcad tempest-ServersTestBootFromVolume-762458528 tempest-ServersTestBootFromVolume-762458528-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 666.828470] env[65726]: WARNING openstack [None req-06f67a27-e6e0-4078-804d-cff1137abcad tempest-ServersTestBootFromVolume-762458528 tempest-ServersTestBootFromVolume-762458528-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 666.893843] env[65726]: WARNING openstack [None req-06f67a27-e6e0-4078-804d-cff1137abcad tempest-ServersTestBootFromVolume-762458528 tempest-ServersTestBootFromVolume-762458528-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 666.894443] env[65726]: WARNING openstack [None req-06f67a27-e6e0-4078-804d-cff1137abcad tempest-ServersTestBootFromVolume-762458528 tempest-ServersTestBootFromVolume-762458528-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 667.041474] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9b3c93a2-43e7-4858-b887-85486b2f5f2a tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Releasing lock "refresh_cache-0e064341-4e4a-407b-8c26-3eb04b409029" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 667.042143] env[65726]: DEBUG nova.objects.instance [None req-9b3c93a2-43e7-4858-b887-85486b2f5f2a tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Lazy-loading 'migration_context' on Instance uuid 0e064341-4e4a-407b-8c26-3eb04b409029 {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 667.051682] env[65726]: DEBUG nova.network.neutron [None req-06f67a27-e6e0-4078-804d-cff1137abcad tempest-ServersTestBootFromVolume-762458528 tempest-ServersTestBootFromVolume-762458528-project-member] [instance: 5a8d4357-4b44-4a19-b1da-42d188c38adc] Updating instance_info_cache with network_info: [{"id": "9aaacd82-31c1-4cfa-bbce-860b30db74af", "address": "fa:16:3e:a2:ff:16", "network": {"id": "bee3bacb-6a14-41ab-a9a3-a82d9ca0b175", "bridge": "br-int", "label": "tempest-ServersTestBootFromVolume-514317700-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "c52b5289c78f45d3942a6c0a4b026207", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dad4f433-bb0b-45c7-8040-972ef2277f75", "external-id": "nsx-vlan-transportzone-451", "segmentation_id": 451, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9aaacd82-31", "ovs_interfaceid": "9aaacd82-31c1-4cfa-bbce-860b30db74af", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 667.253945] env[65726]: DEBUG oslo_concurrency.lockutils [req-8b636882-6005-4ef8-a9e8-0be7180a48a1 req-bc040676-7517-4749-8c51-095b0251a8b6 service nova] Releasing lock "refresh_cache-274ab469-61a9-4b7e-852c-074c871e3abf" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 667.271091] env[65726]: DEBUG nova.compute.utils [None req-fb0f14b3-fd50-4c8d-a984-877b235dcb32 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Using /dev/sd instead of None {{(pid=65726) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 667.278844] env[65726]: DEBUG nova.compute.manager [None req-fb0f14b3-fd50-4c8d-a984-877b235dcb32 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] [instance: d7905c4f-1ba3-43b3-b4b3-aebacd2f9cc3] Allocating IP information in the background. {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 667.279408] env[65726]: DEBUG nova.network.neutron [None req-fb0f14b3-fd50-4c8d-a984-877b235dcb32 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] [instance: d7905c4f-1ba3-43b3-b4b3-aebacd2f9cc3] allocate_for_instance() {{(pid=65726) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 667.279852] env[65726]: WARNING neutronclient.v2_0.client [None req-fb0f14b3-fd50-4c8d-a984-877b235dcb32 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 667.280342] env[65726]: WARNING neutronclient.v2_0.client [None req-fb0f14b3-fd50-4c8d-a984-877b235dcb32 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 667.280942] env[65726]: WARNING openstack [None req-fb0f14b3-fd50-4c8d-a984-877b235dcb32 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 667.281370] env[65726]: WARNING openstack [None req-fb0f14b3-fd50-4c8d-a984-877b235dcb32 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 667.299218] env[65726]: DEBUG oslo_service.periodic_task [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=65726) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 667.300532] env[65726]: DEBUG oslo_service.periodic_task [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=65726) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 667.301785] env[65726]: DEBUG oslo_service.periodic_task [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=65726) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 667.303768] env[65726]: DEBUG oslo_service.periodic_task [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=65726) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 667.303768] env[65726]: DEBUG oslo_service.periodic_task [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=65726) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 667.303768] env[65726]: DEBUG oslo_service.periodic_task [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=65726) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 667.303768] env[65726]: DEBUG nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=65726) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11251}} [ 667.303768] env[65726]: DEBUG oslo_service.periodic_task [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Running periodic task ComputeManager.update_available_resource {{(pid=65726) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 667.308583] env[65726]: DEBUG oslo_vmware.api [None req-362ec7a6-07fd-4c6b-b5fb-9865bdcb8c1a tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Task: {'id': task-5115546, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 667.315284] env[65726]: DEBUG oslo_vmware.api [None req-621af3d8-d8b6-4b45-86f5-d66b91010e2e tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52432eb1-6918-8737-af8b-aa9e09ab4e65, 'name': SearchDatastore_Task, 'duration_secs': 0.023909} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 667.316524] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-11ecd0cb-49b1-4ad2-afd7-33f91e01e464 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.331622] env[65726]: DEBUG oslo_vmware.api [None req-621af3d8-d8b6-4b45-86f5-d66b91010e2e tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Waiting for the task: (returnval){ [ 667.331622] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5276422b-4a3a-23a1-86f5-c59efd0eded2" [ 667.331622] env[65726]: _type = "Task" [ 667.331622] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 667.353048] env[65726]: DEBUG oslo_vmware.api [None req-621af3d8-d8b6-4b45-86f5-d66b91010e2e tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5276422b-4a3a-23a1-86f5-c59efd0eded2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 667.357097] env[65726]: DEBUG nova.policy [None req-fb0f14b3-fd50-4c8d-a984-877b235dcb32 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b23d26af16b54e6d9374fe28e5162d86', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'aac5fe6c99f740ada3747088b2b0ad89', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65726) authorize /opt/stack/nova/nova/policy.py:192}} [ 667.546025] env[65726]: DEBUG nova.objects.base [None req-9b3c93a2-43e7-4858-b887-85486b2f5f2a tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Object Instance<0e064341-4e4a-407b-8c26-3eb04b409029> lazy-loaded attributes: info_cache,migration_context {{(pid=65726) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 667.550379] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07b7fada-b36c-48c3-b7af-c720b4ae8027 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.555230] env[65726]: DEBUG nova.compute.manager [None req-36828e27-b100-49cc-98c2-ad0b25ca6572 tempest-ServerGroupTestJSON-125182767 tempest-ServerGroupTestJSON-125182767-project-member] [instance: c2929727-a7ea-4f63-997d-474c29a305fc] Start spawning the instance on the hypervisor. {{(pid=65726) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 667.557843] env[65726]: DEBUG oslo_concurrency.lockutils [None req-06f67a27-e6e0-4078-804d-cff1137abcad tempest-ServersTestBootFromVolume-762458528 tempest-ServersTestBootFromVolume-762458528-project-member] Releasing lock "refresh_cache-5a8d4357-4b44-4a19-b1da-42d188c38adc" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 667.558258] env[65726]: DEBUG nova.compute.manager [None req-06f67a27-e6e0-4078-804d-cff1137abcad tempest-ServersTestBootFromVolume-762458528 tempest-ServersTestBootFromVolume-762458528-project-member] [instance: 5a8d4357-4b44-4a19-b1da-42d188c38adc] Instance network_info: |[{"id": "9aaacd82-31c1-4cfa-bbce-860b30db74af", "address": "fa:16:3e:a2:ff:16", "network": {"id": "bee3bacb-6a14-41ab-a9a3-a82d9ca0b175", "bridge": "br-int", "label": "tempest-ServersTestBootFromVolume-514317700-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "c52b5289c78f45d3942a6c0a4b026207", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dad4f433-bb0b-45c7-8040-972ef2277f75", "external-id": "nsx-vlan-transportzone-451", "segmentation_id": 451, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9aaacd82-31", "ovs_interfaceid": "9aaacd82-31c1-4cfa-bbce-860b30db74af", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 667.560195] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-06f67a27-e6e0-4078-804d-cff1137abcad tempest-ServersTestBootFromVolume-762458528 tempest-ServersTestBootFromVolume-762458528-project-member] [instance: 5a8d4357-4b44-4a19-b1da-42d188c38adc] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a2:ff:16', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'dad4f433-bb0b-45c7-8040-972ef2277f75', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9aaacd82-31c1-4cfa-bbce-860b30db74af', 'vif_model': 'vmxnet3'}] {{(pid=65726) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 667.567930] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-06f67a27-e6e0-4078-804d-cff1137abcad tempest-ServersTestBootFromVolume-762458528 tempest-ServersTestBootFromVolume-762458528-project-member] Creating folder: Project (c52b5289c78f45d3942a6c0a4b026207). Parent ref: group-v995008. {{(pid=65726) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 667.584942] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5a160b52-2552-45ca-90c0-c2c368c6e7e1 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.590461] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0894dd9b-df26-486d-812a-6bdedce4625e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.598036] env[65726]: DEBUG oslo_vmware.api [None req-9b3c93a2-43e7-4858-b887-85486b2f5f2a tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Waiting for the task: (returnval){ [ 667.598036] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52468d4a-ac0a-c1cc-9928-5327961fb421" [ 667.598036] env[65726]: _type = "Task" [ 667.598036] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 667.603728] env[65726]: DEBUG nova.virt.hardware [None req-36828e27-b100-49cc-98c2-ad0b25ca6572 tempest-ServerGroupTestJSON-125182767 tempest-ServerGroupTestJSON-125182767-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T19:28:34Z,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1539f06fa3534e90acbc3a60c4b8bd3f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T19:28:35Z,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 667.603991] env[65726]: DEBUG nova.virt.hardware [None req-36828e27-b100-49cc-98c2-ad0b25ca6572 tempest-ServerGroupTestJSON-125182767 tempest-ServerGroupTestJSON-125182767-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 667.604157] env[65726]: DEBUG nova.virt.hardware [None req-36828e27-b100-49cc-98c2-ad0b25ca6572 tempest-ServerGroupTestJSON-125182767 tempest-ServerGroupTestJSON-125182767-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 667.604329] env[65726]: DEBUG nova.virt.hardware [None req-36828e27-b100-49cc-98c2-ad0b25ca6572 tempest-ServerGroupTestJSON-125182767 tempest-ServerGroupTestJSON-125182767-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 667.604469] env[65726]: DEBUG nova.virt.hardware [None req-36828e27-b100-49cc-98c2-ad0b25ca6572 tempest-ServerGroupTestJSON-125182767 tempest-ServerGroupTestJSON-125182767-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 667.604630] env[65726]: DEBUG nova.virt.hardware [None req-36828e27-b100-49cc-98c2-ad0b25ca6572 tempest-ServerGroupTestJSON-125182767 tempest-ServerGroupTestJSON-125182767-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 667.604984] env[65726]: DEBUG nova.virt.hardware [None req-36828e27-b100-49cc-98c2-ad0b25ca6572 tempest-ServerGroupTestJSON-125182767 tempest-ServerGroupTestJSON-125182767-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 667.604984] env[65726]: DEBUG nova.virt.hardware [None req-36828e27-b100-49cc-98c2-ad0b25ca6572 tempest-ServerGroupTestJSON-125182767 tempest-ServerGroupTestJSON-125182767-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 667.605292] env[65726]: DEBUG nova.virt.hardware [None req-36828e27-b100-49cc-98c2-ad0b25ca6572 tempest-ServerGroupTestJSON-125182767 tempest-ServerGroupTestJSON-125182767-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 667.605360] env[65726]: DEBUG nova.virt.hardware [None req-36828e27-b100-49cc-98c2-ad0b25ca6572 tempest-ServerGroupTestJSON-125182767 tempest-ServerGroupTestJSON-125182767-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 667.605505] env[65726]: DEBUG nova.virt.hardware [None req-36828e27-b100-49cc-98c2-ad0b25ca6572 tempest-ServerGroupTestJSON-125182767 tempest-ServerGroupTestJSON-125182767-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 667.606420] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20267559-45fa-492d-9408-4de30bbf86b2 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.621775] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1068391-5051-4e7f-9787-18db4bf85a62 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.626180] env[65726]: DEBUG oslo_vmware.api [None req-9b3c93a2-43e7-4858-b887-85486b2f5f2a tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52468d4a-ac0a-c1cc-9928-5327961fb421, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 667.630155] env[65726]: WARNING suds.client [-] Web service reported a SOAP processing fault using an unexpected HTTP status code 200. Reporting as an internal server error. [ 667.630336] env[65726]: DEBUG oslo_vmware.api [-] Fault list: [DuplicateName] {{(pid=65726) _invoke_api /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:337}} [ 667.630750] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-06f67a27-e6e0-4078-804d-cff1137abcad tempest-ServersTestBootFromVolume-762458528 tempest-ServersTestBootFromVolume-762458528-project-member] Folder already exists: Project (c52b5289c78f45d3942a6c0a4b026207). Parent ref: group-v995008. {{(pid=65726) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1609}} [ 667.630946] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-06f67a27-e6e0-4078-804d-cff1137abcad tempest-ServersTestBootFromVolume-762458528 tempest-ServersTestBootFromVolume-762458528-project-member] Creating folder: Instances. Parent ref: group-v995009. {{(pid=65726) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 667.632274] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a7951cd3-7941-4264-85ed-b46131265746 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.648281] env[65726]: INFO nova.virt.vmwareapi.vm_util [None req-06f67a27-e6e0-4078-804d-cff1137abcad tempest-ServersTestBootFromVolume-762458528 tempest-ServersTestBootFromVolume-762458528-project-member] Created folder: Instances in parent group-v995009. [ 667.648539] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-06f67a27-e6e0-4078-804d-cff1137abcad tempest-ServersTestBootFromVolume-762458528 tempest-ServersTestBootFromVolume-762458528-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 667.649212] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5a8d4357-4b44-4a19-b1da-42d188c38adc] Creating VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 667.649212] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-4e0ecd55-f870-4845-a8ff-6deeae499616 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.666054] env[65726]: DEBUG nova.network.neutron [None req-fb0f14b3-fd50-4c8d-a984-877b235dcb32 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] [instance: d7905c4f-1ba3-43b3-b4b3-aebacd2f9cc3] Successfully created port: ee24bb04-7df3-4d06-9751-289b4bdeb753 {{(pid=65726) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 667.678558] env[65726]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 667.678558] env[65726]: value = "task-5115549" [ 667.678558] env[65726]: _type = "Task" [ 667.678558] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 667.691226] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115549, 'name': CreateVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 667.786173] env[65726]: DEBUG oslo_vmware.api [None req-362ec7a6-07fd-4c6b-b5fb-9865bdcb8c1a tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Task: {'id': task-5115546, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.93062} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 667.786482] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-362ec7a6-07fd-4c6b-b5fb-9865bdcb8c1a tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore2] 274ab469-61a9-4b7e-852c-074c871e3abf/274ab469-61a9-4b7e-852c-074c871e3abf.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 667.786672] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-362ec7a6-07fd-4c6b-b5fb-9865bdcb8c1a tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: 274ab469-61a9-4b7e-852c-074c871e3abf] Extending root virtual disk to 1048576 {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 667.786929] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3c20c972-ee1c-4c02-8e0f-6ba120374dad {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.792537] env[65726]: DEBUG nova.compute.manager [None req-fb0f14b3-fd50-4c8d-a984-877b235dcb32 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] [instance: d7905c4f-1ba3-43b3-b4b3-aebacd2f9cc3] Start building block device mappings for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 667.802161] env[65726]: DEBUG oslo_vmware.api [None req-362ec7a6-07fd-4c6b-b5fb-9865bdcb8c1a tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Waiting for the task: (returnval){ [ 667.802161] env[65726]: value = "task-5115550" [ 667.802161] env[65726]: _type = "Task" [ 667.802161] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 667.814829] env[65726]: DEBUG oslo_vmware.api [None req-362ec7a6-07fd-4c6b-b5fb-9865bdcb8c1a tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Task: {'id': task-5115550, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 667.816352] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 667.847255] env[65726]: DEBUG oslo_vmware.api [None req-621af3d8-d8b6-4b45-86f5-d66b91010e2e tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5276422b-4a3a-23a1-86f5-c59efd0eded2, 'name': SearchDatastore_Task, 'duration_secs': 0.031223} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 667.851199] env[65726]: DEBUG oslo_concurrency.lockutils [None req-621af3d8-d8b6-4b45-86f5-d66b91010e2e tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 667.851579] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-621af3d8-d8b6-4b45-86f5-d66b91010e2e tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] fd481728-3ef0-4a96-affd-ab2dd3f596bb/fd481728-3ef0-4a96-affd-ab2dd3f596bb.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 667.855251] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-83cef25c-1d72-4170-8fb7-95a06942aea9 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.862086] env[65726]: DEBUG oslo_vmware.api [None req-621af3d8-d8b6-4b45-86f5-d66b91010e2e tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Waiting for the task: (returnval){ [ 667.862086] env[65726]: value = "task-5115551" [ 667.862086] env[65726]: _type = "Task" [ 667.862086] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 667.879588] env[65726]: DEBUG oslo_vmware.api [None req-621af3d8-d8b6-4b45-86f5-d66b91010e2e tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Task: {'id': task-5115551, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 667.915414] env[65726]: INFO nova.compute.manager [None req-3d632edf-0ed8-46ec-b4fd-7092ea7f050e tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] [instance: 3b933684-ac19-44b0-a49d-6af45501e166] Rebuilding instance [ 667.966820] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ceb93408-d192-427d-82f2-40f6c7f5306a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.977806] env[65726]: DEBUG nova.compute.manager [None req-3d632edf-0ed8-46ec-b4fd-7092ea7f050e tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] [instance: 3b933684-ac19-44b0-a49d-6af45501e166] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 667.979057] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4057a3ee-79a7-4a99-bad6-352a22479816 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.984356] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f416cc0e-83ec-4b9e-be10-37b70b6f9170 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.024052] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1e13cd7-1a97-431f-99ed-566b8bad16cc {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.034847] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6d31b6c-94e0-418b-b451-a74e22850b36 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.054777] env[65726]: DEBUG nova.compute.provider_tree [None req-4b216162-e1e7-4270-8d1f-482978b0681a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 668.109359] env[65726]: DEBUG oslo_vmware.api [None req-9b3c93a2-43e7-4858-b887-85486b2f5f2a tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52468d4a-ac0a-c1cc-9928-5327961fb421, 'name': SearchDatastore_Task, 'duration_secs': 0.027734} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 668.110412] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9b3c93a2-43e7-4858-b887-85486b2f5f2a tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 668.181195] env[65726]: DEBUG nova.network.neutron [None req-36828e27-b100-49cc-98c2-ad0b25ca6572 tempest-ServerGroupTestJSON-125182767 tempest-ServerGroupTestJSON-125182767-project-member] [instance: c2929727-a7ea-4f63-997d-474c29a305fc] Successfully updated port: 27cddbbc-b8ac-4cf0-8baf-d03643c3de8d {{(pid=65726) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 668.194475] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115549, 'name': CreateVM_Task} progress is 99%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 668.317974] env[65726]: DEBUG oslo_vmware.api [None req-362ec7a6-07fd-4c6b-b5fb-9865bdcb8c1a tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Task: {'id': task-5115550, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.080091} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 668.318306] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-362ec7a6-07fd-4c6b-b5fb-9865bdcb8c1a tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: 274ab469-61a9-4b7e-852c-074c871e3abf] Extended root virtual disk {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 668.319714] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-292967ec-dab4-4437-aa7b-30dfa076381f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.351860] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-362ec7a6-07fd-4c6b-b5fb-9865bdcb8c1a tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: 274ab469-61a9-4b7e-852c-074c871e3abf] Reconfiguring VM instance instance-00000014 to attach disk [datastore2] 274ab469-61a9-4b7e-852c-074c871e3abf/274ab469-61a9-4b7e-852c-074c871e3abf.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 668.353461] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3e9d1e06-6030-4b9d-9306-786d1e5ac9e2 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.391638] env[65726]: DEBUG oslo_vmware.api [None req-621af3d8-d8b6-4b45-86f5-d66b91010e2e tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Task: {'id': task-5115551, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 668.393387] env[65726]: DEBUG oslo_vmware.api [None req-362ec7a6-07fd-4c6b-b5fb-9865bdcb8c1a tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Waiting for the task: (returnval){ [ 668.393387] env[65726]: value = "task-5115552" [ 668.393387] env[65726]: _type = "Task" [ 668.393387] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 668.405083] env[65726]: DEBUG oslo_vmware.api [None req-362ec7a6-07fd-4c6b-b5fb-9865bdcb8c1a tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Task: {'id': task-5115552, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 668.557225] env[65726]: DEBUG nova.scheduler.client.report [None req-4b216162-e1e7-4270-8d1f-482978b0681a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 668.686838] env[65726]: DEBUG oslo_concurrency.lockutils [None req-36828e27-b100-49cc-98c2-ad0b25ca6572 tempest-ServerGroupTestJSON-125182767 tempest-ServerGroupTestJSON-125182767-project-member] Acquiring lock "refresh_cache-c2929727-a7ea-4f63-997d-474c29a305fc" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 668.689222] env[65726]: DEBUG oslo_concurrency.lockutils [None req-36828e27-b100-49cc-98c2-ad0b25ca6572 tempest-ServerGroupTestJSON-125182767 tempest-ServerGroupTestJSON-125182767-project-member] Acquired lock "refresh_cache-c2929727-a7ea-4f63-997d-474c29a305fc" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 668.689222] env[65726]: DEBUG nova.network.neutron [None req-36828e27-b100-49cc-98c2-ad0b25ca6572 tempest-ServerGroupTestJSON-125182767 tempest-ServerGroupTestJSON-125182767-project-member] [instance: c2929727-a7ea-4f63-997d-474c29a305fc] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 668.695266] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115549, 'name': CreateVM_Task, 'duration_secs': 0.553185} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 668.695486] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5a8d4357-4b44-4a19-b1da-42d188c38adc] Created VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 668.697151] env[65726]: WARNING openstack [None req-06f67a27-e6e0-4078-804d-cff1137abcad tempest-ServersTestBootFromVolume-762458528 tempest-ServersTestBootFromVolume-762458528-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 668.697435] env[65726]: WARNING openstack [None req-06f67a27-e6e0-4078-804d-cff1137abcad tempest-ServersTestBootFromVolume-762458528 tempest-ServersTestBootFromVolume-762458528-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 668.703308] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-06f67a27-e6e0-4078-804d-cff1137abcad tempest-ServersTestBootFromVolume-762458528 tempest-ServersTestBootFromVolume-762458528-project-member] [instance: 5a8d4357-4b44-4a19-b1da-42d188c38adc] Block device information present: {'root_device_name': '/dev/sda', 'image': [], 'ephemerals': [], 'block_device_mapping': [{'disk_bus': None, 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-995020', 'volume_id': 'b11cb57f-92f1-416d-b59b-70cffe1353ff', 'name': 'volume-b11cb57f-92f1-416d-b59b-70cffe1353ff', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '5a8d4357-4b44-4a19-b1da-42d188c38adc', 'attached_at': '', 'detached_at': '', 'volume_id': 'b11cb57f-92f1-416d-b59b-70cffe1353ff', 'serial': 'b11cb57f-92f1-416d-b59b-70cffe1353ff'}, 'delete_on_termination': True, 'device_type': None, 'boot_index': 0, 'mount_device': '/dev/sda', 'attachment_id': '647e2ade-9cc7-4618-838d-500431f01095', 'guest_format': None, 'volume_type': None}], 'swap': None} {{(pid=65726) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 668.703571] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-06f67a27-e6e0-4078-804d-cff1137abcad tempest-ServersTestBootFromVolume-762458528 tempest-ServersTestBootFromVolume-762458528-project-member] [instance: 5a8d4357-4b44-4a19-b1da-42d188c38adc] Root volume attach. Driver type: vmdk {{(pid=65726) attach_root_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:661}} [ 668.704847] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acbace79-ae35-4237-882f-168917c3c054 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.718463] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ffcbeb1-360a-410b-9db1-3e16b8752492 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.728860] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2009ea2-98e6-4c25-9c46-3cc1b09ccf58 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.745331] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.RelocateVM_Task with opID=oslo.vmware-e5e719db-e93c-4ddd-9184-f195ccc494f0 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.757783] env[65726]: DEBUG oslo_vmware.api [None req-06f67a27-e6e0-4078-804d-cff1137abcad tempest-ServersTestBootFromVolume-762458528 tempest-ServersTestBootFromVolume-762458528-project-member] Waiting for the task: (returnval){ [ 668.757783] env[65726]: value = "task-5115553" [ 668.757783] env[65726]: _type = "Task" [ 668.757783] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 668.771855] env[65726]: DEBUG oslo_vmware.api [None req-06f67a27-e6e0-4078-804d-cff1137abcad tempest-ServersTestBootFromVolume-762458528 tempest-ServersTestBootFromVolume-762458528-project-member] Task: {'id': task-5115553, 'name': RelocateVM_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 668.813431] env[65726]: DEBUG nova.compute.manager [None req-fb0f14b3-fd50-4c8d-a984-877b235dcb32 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] [instance: d7905c4f-1ba3-43b3-b4b3-aebacd2f9cc3] Start spawning the instance on the hypervisor. {{(pid=65726) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 668.856396] env[65726]: DEBUG nova.virt.hardware [None req-fb0f14b3-fd50-4c8d-a984-877b235dcb32 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T19:28:34Z,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1539f06fa3534e90acbc3a60c4b8bd3f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T19:28:35Z,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 668.856933] env[65726]: DEBUG nova.virt.hardware [None req-fb0f14b3-fd50-4c8d-a984-877b235dcb32 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 668.857339] env[65726]: DEBUG nova.virt.hardware [None req-fb0f14b3-fd50-4c8d-a984-877b235dcb32 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 668.857644] env[65726]: DEBUG nova.virt.hardware [None req-fb0f14b3-fd50-4c8d-a984-877b235dcb32 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 668.857911] env[65726]: DEBUG nova.virt.hardware [None req-fb0f14b3-fd50-4c8d-a984-877b235dcb32 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 668.858143] env[65726]: DEBUG nova.virt.hardware [None req-fb0f14b3-fd50-4c8d-a984-877b235dcb32 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 668.858434] env[65726]: DEBUG nova.virt.hardware [None req-fb0f14b3-fd50-4c8d-a984-877b235dcb32 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 668.858641] env[65726]: DEBUG nova.virt.hardware [None req-fb0f14b3-fd50-4c8d-a984-877b235dcb32 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 668.858978] env[65726]: DEBUG nova.virt.hardware [None req-fb0f14b3-fd50-4c8d-a984-877b235dcb32 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 668.859217] env[65726]: DEBUG nova.virt.hardware [None req-fb0f14b3-fd50-4c8d-a984-877b235dcb32 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 668.859485] env[65726]: DEBUG nova.virt.hardware [None req-fb0f14b3-fd50-4c8d-a984-877b235dcb32 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 668.860552] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee04102e-cf01-42ea-9d85-81578c61ea45 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.871017] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec68ea24-199c-4873-bf3a-d8b5a0cb934a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.898840] env[65726]: DEBUG oslo_vmware.api [None req-621af3d8-d8b6-4b45-86f5-d66b91010e2e tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Task: {'id': task-5115551, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.65798} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 668.902577] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-621af3d8-d8b6-4b45-86f5-d66b91010e2e tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] fd481728-3ef0-4a96-affd-ab2dd3f596bb/fd481728-3ef0-4a96-affd-ab2dd3f596bb.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 668.902826] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-621af3d8-d8b6-4b45-86f5-d66b91010e2e tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] [instance: fd481728-3ef0-4a96-affd-ab2dd3f596bb] Extending root virtual disk to 1048576 {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 668.903138] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e8e9b0d2-41e2-47ba-89c9-c1d183598fe2 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.911629] env[65726]: DEBUG oslo_vmware.api [None req-362ec7a6-07fd-4c6b-b5fb-9865bdcb8c1a tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Task: {'id': task-5115552, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 668.913267] env[65726]: DEBUG oslo_vmware.api [None req-621af3d8-d8b6-4b45-86f5-d66b91010e2e tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Waiting for the task: (returnval){ [ 668.913267] env[65726]: value = "task-5115554" [ 668.913267] env[65726]: _type = "Task" [ 668.913267] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 668.923108] env[65726]: DEBUG oslo_vmware.api [None req-621af3d8-d8b6-4b45-86f5-d66b91010e2e tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Task: {'id': task-5115554, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 668.981588] env[65726]: DEBUG oslo_concurrency.lockutils [None req-34fc70a6-ee2b-4816-b933-3db54eee74ae tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Acquiring lock "811c1bbd-259c-4f5e-9a61-581bfba0ec4d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 668.982157] env[65726]: DEBUG oslo_concurrency.lockutils [None req-34fc70a6-ee2b-4816-b933-3db54eee74ae tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Lock "811c1bbd-259c-4f5e-9a61-581bfba0ec4d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 668.982157] env[65726]: DEBUG oslo_concurrency.lockutils [None req-34fc70a6-ee2b-4816-b933-3db54eee74ae tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Acquiring lock "811c1bbd-259c-4f5e-9a61-581bfba0ec4d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 668.982416] env[65726]: DEBUG oslo_concurrency.lockutils [None req-34fc70a6-ee2b-4816-b933-3db54eee74ae tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Lock "811c1bbd-259c-4f5e-9a61-581bfba0ec4d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 668.982520] env[65726]: DEBUG oslo_concurrency.lockutils [None req-34fc70a6-ee2b-4816-b933-3db54eee74ae tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Lock "811c1bbd-259c-4f5e-9a61-581bfba0ec4d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 668.984973] env[65726]: INFO nova.compute.manager [None req-34fc70a6-ee2b-4816-b933-3db54eee74ae tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] [instance: 811c1bbd-259c-4f5e-9a61-581bfba0ec4d] Terminating instance [ 669.029366] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-3d632edf-0ed8-46ec-b4fd-7092ea7f050e tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] [instance: 3b933684-ac19-44b0-a49d-6af45501e166] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 669.029906] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-796c3a9b-2849-4c6b-aa9d-5f622854d15d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.039684] env[65726]: DEBUG oslo_vmware.api [None req-3d632edf-0ed8-46ec-b4fd-7092ea7f050e tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] Waiting for the task: (returnval){ [ 669.039684] env[65726]: value = "task-5115555" [ 669.039684] env[65726]: _type = "Task" [ 669.039684] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 669.049572] env[65726]: DEBUG oslo_vmware.api [None req-3d632edf-0ed8-46ec-b4fd-7092ea7f050e tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] Task: {'id': task-5115555, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 669.064730] env[65726]: DEBUG oslo_concurrency.lockutils [None req-4b216162-e1e7-4270-8d1f-482978b0681a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.300s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 669.071702] env[65726]: DEBUG oslo_concurrency.lockutils [None req-64ef4520-b00a-42b8-937b-fbfc7c9b2298 tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 16.902s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 669.071702] env[65726]: DEBUG nova.objects.instance [None req-64ef4520-b00a-42b8-937b-fbfc7c9b2298 tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Lazy-loading 'resources' on Instance uuid 21730277-cef6-4bfe-9f67-de71f6f615ff {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 669.093156] env[65726]: INFO nova.scheduler.client.report [None req-4b216162-e1e7-4270-8d1f-482978b0681a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Deleted allocations for instance 8493ba52-b950-4727-970a-19b1797b2ca1 [ 669.191810] env[65726]: WARNING openstack [None req-36828e27-b100-49cc-98c2-ad0b25ca6572 tempest-ServerGroupTestJSON-125182767 tempest-ServerGroupTestJSON-125182767-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 669.192813] env[65726]: WARNING openstack [None req-36828e27-b100-49cc-98c2-ad0b25ca6572 tempest-ServerGroupTestJSON-125182767 tempest-ServerGroupTestJSON-125182767-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 669.199464] env[65726]: DEBUG nova.network.neutron [None req-36828e27-b100-49cc-98c2-ad0b25ca6572 tempest-ServerGroupTestJSON-125182767 tempest-ServerGroupTestJSON-125182767-project-member] [instance: c2929727-a7ea-4f63-997d-474c29a305fc] Instance cache missing network info. {{(pid=65726) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 669.270638] env[65726]: DEBUG oslo_vmware.api [None req-06f67a27-e6e0-4078-804d-cff1137abcad tempest-ServersTestBootFromVolume-762458528 tempest-ServersTestBootFromVolume-762458528-project-member] Task: {'id': task-5115553, 'name': RelocateVM_Task} progress is 19%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 669.379974] env[65726]: DEBUG nova.network.neutron [None req-fb0f14b3-fd50-4c8d-a984-877b235dcb32 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] [instance: d7905c4f-1ba3-43b3-b4b3-aebacd2f9cc3] Successfully updated port: ee24bb04-7df3-4d06-9751-289b4bdeb753 {{(pid=65726) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 669.408727] env[65726]: DEBUG oslo_vmware.api [None req-362ec7a6-07fd-4c6b-b5fb-9865bdcb8c1a tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Task: {'id': task-5115552, 'name': ReconfigVM_Task, 'duration_secs': 0.635727} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 669.409229] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-362ec7a6-07fd-4c6b-b5fb-9865bdcb8c1a tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: 274ab469-61a9-4b7e-852c-074c871e3abf] Reconfigured VM instance instance-00000014 to attach disk [datastore2] 274ab469-61a9-4b7e-852c-074c871e3abf/274ab469-61a9-4b7e-852c-074c871e3abf.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 669.410034] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-38119095-627a-497b-851d-9b37efcd9379 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.423871] env[65726]: DEBUG oslo_vmware.api [None req-362ec7a6-07fd-4c6b-b5fb-9865bdcb8c1a tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Waiting for the task: (returnval){ [ 669.423871] env[65726]: value = "task-5115556" [ 669.423871] env[65726]: _type = "Task" [ 669.423871] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 669.431470] env[65726]: DEBUG oslo_vmware.api [None req-621af3d8-d8b6-4b45-86f5-d66b91010e2e tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Task: {'id': task-5115554, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.092395} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 669.432187] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-621af3d8-d8b6-4b45-86f5-d66b91010e2e tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] [instance: fd481728-3ef0-4a96-affd-ab2dd3f596bb] Extended root virtual disk {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 669.433038] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54475cf8-de68-44cb-b293-a8c2915bec60 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.439328] env[65726]: DEBUG oslo_vmware.api [None req-362ec7a6-07fd-4c6b-b5fb-9865bdcb8c1a tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Task: {'id': task-5115556, 'name': Rename_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 669.460563] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-621af3d8-d8b6-4b45-86f5-d66b91010e2e tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] [instance: fd481728-3ef0-4a96-affd-ab2dd3f596bb] Reconfiguring VM instance instance-00000016 to attach disk [datastore1] fd481728-3ef0-4a96-affd-ab2dd3f596bb/fd481728-3ef0-4a96-affd-ab2dd3f596bb.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 669.461026] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-67423966-b684-4d38-868b-227fa41de33e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.486241] env[65726]: DEBUG oslo_vmware.api [None req-621af3d8-d8b6-4b45-86f5-d66b91010e2e tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Waiting for the task: (returnval){ [ 669.486241] env[65726]: value = "task-5115557" [ 669.486241] env[65726]: _type = "Task" [ 669.486241] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 669.491800] env[65726]: DEBUG nova.compute.manager [None req-34fc70a6-ee2b-4816-b933-3db54eee74ae tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] [instance: 811c1bbd-259c-4f5e-9a61-581bfba0ec4d] Start destroying the instance on the hypervisor. {{(pid=65726) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 669.491800] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-34fc70a6-ee2b-4816-b933-3db54eee74ae tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] [instance: 811c1bbd-259c-4f5e-9a61-581bfba0ec4d] Destroying instance {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 669.493532] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6960ce6-6446-409f-9a1b-69769438e0f6 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.508959] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-34fc70a6-ee2b-4816-b933-3db54eee74ae tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] [instance: 811c1bbd-259c-4f5e-9a61-581bfba0ec4d] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 669.508959] env[65726]: DEBUG oslo_vmware.api [None req-621af3d8-d8b6-4b45-86f5-d66b91010e2e tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Task: {'id': task-5115557, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 669.508959] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3f1b8060-9721-4384-a2e2-bf90a09a9f3e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.516371] env[65726]: DEBUG oslo_vmware.api [None req-34fc70a6-ee2b-4816-b933-3db54eee74ae tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Waiting for the task: (returnval){ [ 669.516371] env[65726]: value = "task-5115558" [ 669.516371] env[65726]: _type = "Task" [ 669.516371] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 669.525268] env[65726]: DEBUG oslo_vmware.api [None req-34fc70a6-ee2b-4816-b933-3db54eee74ae tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Task: {'id': task-5115558, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 669.550866] env[65726]: DEBUG oslo_vmware.api [None req-3d632edf-0ed8-46ec-b4fd-7092ea7f050e tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] Task: {'id': task-5115555, 'name': PowerOffVM_Task, 'duration_secs': 0.286865} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 669.551276] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-3d632edf-0ed8-46ec-b4fd-7092ea7f050e tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] [instance: 3b933684-ac19-44b0-a49d-6af45501e166] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 669.551555] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-3d632edf-0ed8-46ec-b4fd-7092ea7f050e tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] [instance: 3b933684-ac19-44b0-a49d-6af45501e166] Destroying instance {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 669.552500] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d61f9fdd-8b5f-41f3-baad-dfe96cf16f82 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.561180] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-3d632edf-0ed8-46ec-b4fd-7092ea7f050e tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] [instance: 3b933684-ac19-44b0-a49d-6af45501e166] Unregistering the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 669.561488] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7befb94d-c833-48df-aeab-6e09883a2f11 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.570422] env[65726]: WARNING openstack [None req-36828e27-b100-49cc-98c2-ad0b25ca6572 tempest-ServerGroupTestJSON-125182767 tempest-ServerGroupTestJSON-125182767-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 669.570873] env[65726]: WARNING openstack [None req-36828e27-b100-49cc-98c2-ad0b25ca6572 tempest-ServerGroupTestJSON-125182767 tempest-ServerGroupTestJSON-125182767-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 669.593976] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-3d632edf-0ed8-46ec-b4fd-7092ea7f050e tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] [instance: 3b933684-ac19-44b0-a49d-6af45501e166] Unregistered the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 669.594449] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-3d632edf-0ed8-46ec-b4fd-7092ea7f050e tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] [instance: 3b933684-ac19-44b0-a49d-6af45501e166] Deleting contents of the VM from datastore datastore2 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 669.594665] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-3d632edf-0ed8-46ec-b4fd-7092ea7f050e tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] Deleting the datastore file [datastore2] 3b933684-ac19-44b0-a49d-6af45501e166 {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 669.594942] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d31c42e0-7bf5-4bc6-a9a4-9e01d7ab0fb2 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.603491] env[65726]: DEBUG nova.compute.manager [req-889d19b5-b49d-4352-848b-37f235b066e0 req-7307af5f-8ede-44db-8223-5db2e208a88d service nova] [instance: fd481728-3ef0-4a96-affd-ab2dd3f596bb] Received event network-vif-plugged-7df3f02b-e4e5-4840-bf9f-3498771b9184 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 669.604260] env[65726]: DEBUG oslo_concurrency.lockutils [req-889d19b5-b49d-4352-848b-37f235b066e0 req-7307af5f-8ede-44db-8223-5db2e208a88d service nova] Acquiring lock "fd481728-3ef0-4a96-affd-ab2dd3f596bb-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 669.604260] env[65726]: DEBUG oslo_concurrency.lockutils [req-889d19b5-b49d-4352-848b-37f235b066e0 req-7307af5f-8ede-44db-8223-5db2e208a88d service nova] Lock "fd481728-3ef0-4a96-affd-ab2dd3f596bb-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 669.604533] env[65726]: DEBUG oslo_concurrency.lockutils [req-889d19b5-b49d-4352-848b-37f235b066e0 req-7307af5f-8ede-44db-8223-5db2e208a88d service nova] Lock "fd481728-3ef0-4a96-affd-ab2dd3f596bb-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 669.604533] env[65726]: DEBUG nova.compute.manager [req-889d19b5-b49d-4352-848b-37f235b066e0 req-7307af5f-8ede-44db-8223-5db2e208a88d service nova] [instance: fd481728-3ef0-4a96-affd-ab2dd3f596bb] No waiting events found dispatching network-vif-plugged-7df3f02b-e4e5-4840-bf9f-3498771b9184 {{(pid=65726) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 669.604700] env[65726]: WARNING nova.compute.manager [req-889d19b5-b49d-4352-848b-37f235b066e0 req-7307af5f-8ede-44db-8223-5db2e208a88d service nova] [instance: fd481728-3ef0-4a96-affd-ab2dd3f596bb] Received unexpected event network-vif-plugged-7df3f02b-e4e5-4840-bf9f-3498771b9184 for instance with vm_state building and task_state spawning. [ 669.604772] env[65726]: DEBUG nova.compute.manager [req-889d19b5-b49d-4352-848b-37f235b066e0 req-7307af5f-8ede-44db-8223-5db2e208a88d service nova] [instance: fd481728-3ef0-4a96-affd-ab2dd3f596bb] Received event network-changed-7df3f02b-e4e5-4840-bf9f-3498771b9184 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 669.604914] env[65726]: DEBUG nova.compute.manager [req-889d19b5-b49d-4352-848b-37f235b066e0 req-7307af5f-8ede-44db-8223-5db2e208a88d service nova] [instance: fd481728-3ef0-4a96-affd-ab2dd3f596bb] Refreshing instance network info cache due to event network-changed-7df3f02b-e4e5-4840-bf9f-3498771b9184. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 669.605135] env[65726]: DEBUG oslo_concurrency.lockutils [req-889d19b5-b49d-4352-848b-37f235b066e0 req-7307af5f-8ede-44db-8223-5db2e208a88d service nova] Acquiring lock "refresh_cache-fd481728-3ef0-4a96-affd-ab2dd3f596bb" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 669.605284] env[65726]: DEBUG oslo_concurrency.lockutils [req-889d19b5-b49d-4352-848b-37f235b066e0 req-7307af5f-8ede-44db-8223-5db2e208a88d service nova] Acquired lock "refresh_cache-fd481728-3ef0-4a96-affd-ab2dd3f596bb" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 669.605427] env[65726]: DEBUG nova.network.neutron [req-889d19b5-b49d-4352-848b-37f235b066e0 req-7307af5f-8ede-44db-8223-5db2e208a88d service nova] [instance: fd481728-3ef0-4a96-affd-ab2dd3f596bb] Refreshing network info cache for port 7df3f02b-e4e5-4840-bf9f-3498771b9184 {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 669.611758] env[65726]: DEBUG oslo_concurrency.lockutils [None req-4b216162-e1e7-4270-8d1f-482978b0681a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Lock "8493ba52-b950-4727-970a-19b1797b2ca1" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 23.711s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 669.617019] env[65726]: DEBUG oslo_vmware.api [None req-3d632edf-0ed8-46ec-b4fd-7092ea7f050e tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] Waiting for the task: (returnval){ [ 669.617019] env[65726]: value = "task-5115560" [ 669.617019] env[65726]: _type = "Task" [ 669.617019] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 669.629272] env[65726]: DEBUG oslo_vmware.api [None req-3d632edf-0ed8-46ec-b4fd-7092ea7f050e tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] Task: {'id': task-5115560, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 669.674922] env[65726]: WARNING openstack [None req-36828e27-b100-49cc-98c2-ad0b25ca6572 tempest-ServerGroupTestJSON-125182767 tempest-ServerGroupTestJSON-125182767-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 669.675524] env[65726]: WARNING openstack [None req-36828e27-b100-49cc-98c2-ad0b25ca6572 tempest-ServerGroupTestJSON-125182767 tempest-ServerGroupTestJSON-125182767-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 669.771297] env[65726]: DEBUG oslo_vmware.api [None req-06f67a27-e6e0-4078-804d-cff1137abcad tempest-ServersTestBootFromVolume-762458528 tempest-ServersTestBootFromVolume-762458528-project-member] Task: {'id': task-5115553, 'name': RelocateVM_Task, 'duration_secs': 0.96775} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 669.772360] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-06f67a27-e6e0-4078-804d-cff1137abcad tempest-ServersTestBootFromVolume-762458528 tempest-ServersTestBootFromVolume-762458528-project-member] [instance: 5a8d4357-4b44-4a19-b1da-42d188c38adc] Volume attach. Driver type: vmdk {{(pid=65726) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 669.772360] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-06f67a27-e6e0-4078-804d-cff1137abcad tempest-ServersTestBootFromVolume-762458528 tempest-ServersTestBootFromVolume-762458528-project-member] [instance: 5a8d4357-4b44-4a19-b1da-42d188c38adc] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-995020', 'volume_id': 'b11cb57f-92f1-416d-b59b-70cffe1353ff', 'name': 'volume-b11cb57f-92f1-416d-b59b-70cffe1353ff', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '5a8d4357-4b44-4a19-b1da-42d188c38adc', 'attached_at': '', 'detached_at': '', 'volume_id': 'b11cb57f-92f1-416d-b59b-70cffe1353ff', 'serial': 'b11cb57f-92f1-416d-b59b-70cffe1353ff'} {{(pid=65726) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 669.772824] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7a13211-bb84-46a1-b6c5-59dd839699e0 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.797848] env[65726]: DEBUG nova.network.neutron [None req-36828e27-b100-49cc-98c2-ad0b25ca6572 tempest-ServerGroupTestJSON-125182767 tempest-ServerGroupTestJSON-125182767-project-member] [instance: c2929727-a7ea-4f63-997d-474c29a305fc] Updating instance_info_cache with network_info: [{"id": "27cddbbc-b8ac-4cf0-8baf-d03643c3de8d", "address": "fa:16:3e:53:19:47", "network": {"id": "1401b82d-0e46-47ea-bf6d-ed2a2e5f7f24", "bridge": "br-int", "label": "tempest-ServerGroupTestJSON-933777645-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "34534927100c4f7a91a1073de9047674", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "40859343-2baa-45fd-88e3-ebf8aaed2b19", "external-id": "nsx-vlan-transportzone-10", "segmentation_id": 10, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap27cddbbc-b8", "ovs_interfaceid": "27cddbbc-b8ac-4cf0-8baf-d03643c3de8d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 669.797848] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f32417a-c23c-43fb-aef6-bd5a5053f602 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.824093] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-06f67a27-e6e0-4078-804d-cff1137abcad tempest-ServersTestBootFromVolume-762458528 tempest-ServersTestBootFromVolume-762458528-project-member] [instance: 5a8d4357-4b44-4a19-b1da-42d188c38adc] Reconfiguring VM instance instance-00000015 to attach disk [datastore1] volume-b11cb57f-92f1-416d-b59b-70cffe1353ff/volume-b11cb57f-92f1-416d-b59b-70cffe1353ff.vmdk or device None with type thin {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 669.826643] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8e63cb72-0ad6-4865-8aec-eb2c9fc36da3 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.850810] env[65726]: DEBUG oslo_vmware.api [None req-06f67a27-e6e0-4078-804d-cff1137abcad tempest-ServersTestBootFromVolume-762458528 tempest-ServersTestBootFromVolume-762458528-project-member] Waiting for the task: (returnval){ [ 669.850810] env[65726]: value = "task-5115561" [ 669.850810] env[65726]: _type = "Task" [ 669.850810] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 669.863206] env[65726]: DEBUG oslo_vmware.api [None req-06f67a27-e6e0-4078-804d-cff1137abcad tempest-ServersTestBootFromVolume-762458528 tempest-ServersTestBootFromVolume-762458528-project-member] Task: {'id': task-5115561, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 669.887841] env[65726]: DEBUG oslo_concurrency.lockutils [None req-fb0f14b3-fd50-4c8d-a984-877b235dcb32 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Acquiring lock "refresh_cache-d7905c4f-1ba3-43b3-b4b3-aebacd2f9cc3" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 669.888025] env[65726]: DEBUG oslo_concurrency.lockutils [None req-fb0f14b3-fd50-4c8d-a984-877b235dcb32 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Acquired lock "refresh_cache-d7905c4f-1ba3-43b3-b4b3-aebacd2f9cc3" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 669.889526] env[65726]: DEBUG nova.network.neutron [None req-fb0f14b3-fd50-4c8d-a984-877b235dcb32 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] [instance: d7905c4f-1ba3-43b3-b4b3-aebacd2f9cc3] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 669.939173] env[65726]: DEBUG oslo_vmware.api [None req-362ec7a6-07fd-4c6b-b5fb-9865bdcb8c1a tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Task: {'id': task-5115556, 'name': Rename_Task, 'duration_secs': 0.156379} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 669.939360] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-362ec7a6-07fd-4c6b-b5fb-9865bdcb8c1a tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: 274ab469-61a9-4b7e-852c-074c871e3abf] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 669.939602] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-746a3639-f1b5-4704-81e8-ff8dfccc957a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.948546] env[65726]: DEBUG oslo_vmware.api [None req-362ec7a6-07fd-4c6b-b5fb-9865bdcb8c1a tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Waiting for the task: (returnval){ [ 669.948546] env[65726]: value = "task-5115562" [ 669.948546] env[65726]: _type = "Task" [ 669.948546] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 669.958733] env[65726]: DEBUG oslo_vmware.api [None req-362ec7a6-07fd-4c6b-b5fb-9865bdcb8c1a tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Task: {'id': task-5115562, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 670.000933] env[65726]: DEBUG oslo_vmware.api [None req-621af3d8-d8b6-4b45-86f5-d66b91010e2e tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Task: {'id': task-5115557, 'name': ReconfigVM_Task, 'duration_secs': 0.482383} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 670.001594] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-621af3d8-d8b6-4b45-86f5-d66b91010e2e tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] [instance: fd481728-3ef0-4a96-affd-ab2dd3f596bb] Reconfigured VM instance instance-00000016 to attach disk [datastore1] fd481728-3ef0-4a96-affd-ab2dd3f596bb/fd481728-3ef0-4a96-affd-ab2dd3f596bb.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 670.002391] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-fa4552ae-612b-49dc-a161-ad54f2e2791e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.014872] env[65726]: DEBUG oslo_vmware.api [None req-621af3d8-d8b6-4b45-86f5-d66b91010e2e tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Waiting for the task: (returnval){ [ 670.014872] env[65726]: value = "task-5115563" [ 670.014872] env[65726]: _type = "Task" [ 670.014872] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 670.036456] env[65726]: DEBUG oslo_vmware.api [None req-621af3d8-d8b6-4b45-86f5-d66b91010e2e tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Task: {'id': task-5115563, 'name': Rename_Task} progress is 10%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 670.037232] env[65726]: DEBUG oslo_vmware.api [None req-34fc70a6-ee2b-4816-b933-3db54eee74ae tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Task: {'id': task-5115558, 'name': PowerOffVM_Task, 'duration_secs': 0.242762} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 670.040733] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-34fc70a6-ee2b-4816-b933-3db54eee74ae tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] [instance: 811c1bbd-259c-4f5e-9a61-581bfba0ec4d] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 670.040920] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-34fc70a6-ee2b-4816-b933-3db54eee74ae tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] [instance: 811c1bbd-259c-4f5e-9a61-581bfba0ec4d] Unregistering the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 670.041980] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3dc65efd-34da-4c71-b3c2-3948464fb927 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.115088] env[65726]: WARNING openstack [req-889d19b5-b49d-4352-848b-37f235b066e0 req-7307af5f-8ede-44db-8223-5db2e208a88d service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 670.116871] env[65726]: WARNING openstack [req-889d19b5-b49d-4352-848b-37f235b066e0 req-7307af5f-8ede-44db-8223-5db2e208a88d service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 670.145984] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-34fc70a6-ee2b-4816-b933-3db54eee74ae tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] [instance: 811c1bbd-259c-4f5e-9a61-581bfba0ec4d] Unregistered the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 670.145984] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-34fc70a6-ee2b-4816-b933-3db54eee74ae tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] [instance: 811c1bbd-259c-4f5e-9a61-581bfba0ec4d] Deleting contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 670.145984] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-34fc70a6-ee2b-4816-b933-3db54eee74ae tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Deleting the datastore file [datastore1] 811c1bbd-259c-4f5e-9a61-581bfba0ec4d {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 670.146295] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-329e86ee-cb92-49e1-ae71-1af588c366d0 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.156592] env[65726]: DEBUG oslo_vmware.api [None req-3d632edf-0ed8-46ec-b4fd-7092ea7f050e tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] Task: {'id': task-5115560, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.1583} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 670.161276] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-3d632edf-0ed8-46ec-b4fd-7092ea7f050e tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] Deleted the datastore file {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 670.162041] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-3d632edf-0ed8-46ec-b4fd-7092ea7f050e tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] [instance: 3b933684-ac19-44b0-a49d-6af45501e166] Deleted contents of the VM from datastore datastore2 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 670.162041] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-3d632edf-0ed8-46ec-b4fd-7092ea7f050e tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] [instance: 3b933684-ac19-44b0-a49d-6af45501e166] Instance destroyed {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 670.168031] env[65726]: DEBUG oslo_vmware.api [None req-34fc70a6-ee2b-4816-b933-3db54eee74ae tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Waiting for the task: (returnval){ [ 670.168031] env[65726]: value = "task-5115565" [ 670.168031] env[65726]: _type = "Task" [ 670.168031] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 670.178936] env[65726]: DEBUG oslo_vmware.api [None req-34fc70a6-ee2b-4816-b933-3db54eee74ae tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Task: {'id': task-5115565, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 670.302582] env[65726]: DEBUG oslo_concurrency.lockutils [None req-36828e27-b100-49cc-98c2-ad0b25ca6572 tempest-ServerGroupTestJSON-125182767 tempest-ServerGroupTestJSON-125182767-project-member] Releasing lock "refresh_cache-c2929727-a7ea-4f63-997d-474c29a305fc" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 670.302982] env[65726]: DEBUG nova.compute.manager [None req-36828e27-b100-49cc-98c2-ad0b25ca6572 tempest-ServerGroupTestJSON-125182767 tempest-ServerGroupTestJSON-125182767-project-member] [instance: c2929727-a7ea-4f63-997d-474c29a305fc] Instance network_info: |[{"id": "27cddbbc-b8ac-4cf0-8baf-d03643c3de8d", "address": "fa:16:3e:53:19:47", "network": {"id": "1401b82d-0e46-47ea-bf6d-ed2a2e5f7f24", "bridge": "br-int", "label": "tempest-ServerGroupTestJSON-933777645-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "34534927100c4f7a91a1073de9047674", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "40859343-2baa-45fd-88e3-ebf8aaed2b19", "external-id": "nsx-vlan-transportzone-10", "segmentation_id": 10, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap27cddbbc-b8", "ovs_interfaceid": "27cddbbc-b8ac-4cf0-8baf-d03643c3de8d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 670.303482] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-36828e27-b100-49cc-98c2-ad0b25ca6572 tempest-ServerGroupTestJSON-125182767 tempest-ServerGroupTestJSON-125182767-project-member] [instance: c2929727-a7ea-4f63-997d-474c29a305fc] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:53:19:47', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '40859343-2baa-45fd-88e3-ebf8aaed2b19', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '27cddbbc-b8ac-4cf0-8baf-d03643c3de8d', 'vif_model': 'vmxnet3'}] {{(pid=65726) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 670.311743] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-36828e27-b100-49cc-98c2-ad0b25ca6572 tempest-ServerGroupTestJSON-125182767 tempest-ServerGroupTestJSON-125182767-project-member] Creating folder: Project (34534927100c4f7a91a1073de9047674). Parent ref: group-v995008. {{(pid=65726) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 670.315216] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d04a5b2c-6c97-4c6e-9d88-5980fcf4a930 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.329743] env[65726]: INFO nova.virt.vmwareapi.vm_util [None req-36828e27-b100-49cc-98c2-ad0b25ca6572 tempest-ServerGroupTestJSON-125182767 tempest-ServerGroupTestJSON-125182767-project-member] Created folder: Project (34534927100c4f7a91a1073de9047674) in parent group-v995008. [ 670.330048] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-36828e27-b100-49cc-98c2-ad0b25ca6572 tempest-ServerGroupTestJSON-125182767 tempest-ServerGroupTestJSON-125182767-project-member] Creating folder: Instances. Parent ref: group-v995068. {{(pid=65726) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 670.330266] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-56b3ccd6-40da-4366-9e60-ce54a213c8f2 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.343800] env[65726]: INFO nova.virt.vmwareapi.vm_util [None req-36828e27-b100-49cc-98c2-ad0b25ca6572 tempest-ServerGroupTestJSON-125182767 tempest-ServerGroupTestJSON-125182767-project-member] Created folder: Instances in parent group-v995068. [ 670.344159] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-36828e27-b100-49cc-98c2-ad0b25ca6572 tempest-ServerGroupTestJSON-125182767 tempest-ServerGroupTestJSON-125182767-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 670.344482] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c2929727-a7ea-4f63-997d-474c29a305fc] Creating VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 670.344748] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d720ff06-baed-4615-8b4a-329d4c745e6a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.362399] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf4c51a6-cece-4b60-875f-c2644c0e0be0 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.378552] env[65726]: DEBUG oslo_vmware.api [None req-06f67a27-e6e0-4078-804d-cff1137abcad tempest-ServersTestBootFromVolume-762458528 tempest-ServersTestBootFromVolume-762458528-project-member] Task: {'id': task-5115561, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 670.381916] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0748a989-b6fc-4f24-abde-97f3df7294b1 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.385510] env[65726]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 670.385510] env[65726]: value = "task-5115568" [ 670.385510] env[65726]: _type = "Task" [ 670.385510] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 670.417512] env[65726]: WARNING openstack [None req-fb0f14b3-fd50-4c8d-a984-877b235dcb32 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 670.417758] env[65726]: WARNING openstack [None req-fb0f14b3-fd50-4c8d-a984-877b235dcb32 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 670.423341] env[65726]: DEBUG nova.network.neutron [None req-fb0f14b3-fd50-4c8d-a984-877b235dcb32 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] [instance: d7905c4f-1ba3-43b3-b4b3-aebacd2f9cc3] Instance cache missing network info. {{(pid=65726) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 670.426666] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de8686fb-942b-4c55-9ee2-1b62fb71d030 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.432746] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115568, 'name': CreateVM_Task} progress is 15%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 670.438786] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7f7a2f7-a075-4fb3-93c6-4bcc4c152d28 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.454758] env[65726]: DEBUG nova.compute.provider_tree [None req-64ef4520-b00a-42b8-937b-fbfc7c9b2298 tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 670.466555] env[65726]: DEBUG oslo_vmware.api [None req-362ec7a6-07fd-4c6b-b5fb-9865bdcb8c1a tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Task: {'id': task-5115562, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 670.527187] env[65726]: DEBUG oslo_vmware.api [None req-621af3d8-d8b6-4b45-86f5-d66b91010e2e tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Task: {'id': task-5115563, 'name': Rename_Task, 'duration_secs': 0.348079} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 670.527384] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-621af3d8-d8b6-4b45-86f5-d66b91010e2e tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] [instance: fd481728-3ef0-4a96-affd-ab2dd3f596bb] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 670.527644] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-547d888c-ddff-449b-8e24-ca5420fe542a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.536165] env[65726]: DEBUG oslo_vmware.api [None req-621af3d8-d8b6-4b45-86f5-d66b91010e2e tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Waiting for the task: (returnval){ [ 670.536165] env[65726]: value = "task-5115569" [ 670.536165] env[65726]: _type = "Task" [ 670.536165] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 670.547567] env[65726]: DEBUG oslo_vmware.api [None req-621af3d8-d8b6-4b45-86f5-d66b91010e2e tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Task: {'id': task-5115569, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 670.682961] env[65726]: DEBUG oslo_vmware.api [None req-34fc70a6-ee2b-4816-b933-3db54eee74ae tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Task: {'id': task-5115565, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.337723} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 670.684512] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-34fc70a6-ee2b-4816-b933-3db54eee74ae tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Deleted the datastore file {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 670.684512] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-34fc70a6-ee2b-4816-b933-3db54eee74ae tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] [instance: 811c1bbd-259c-4f5e-9a61-581bfba0ec4d] Deleted contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 670.684512] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-34fc70a6-ee2b-4816-b933-3db54eee74ae tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] [instance: 811c1bbd-259c-4f5e-9a61-581bfba0ec4d] Instance destroyed {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 670.684512] env[65726]: INFO nova.compute.manager [None req-34fc70a6-ee2b-4816-b933-3db54eee74ae tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] [instance: 811c1bbd-259c-4f5e-9a61-581bfba0ec4d] Took 1.19 seconds to destroy the instance on the hypervisor. [ 670.684512] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-34fc70a6-ee2b-4816-b933-3db54eee74ae tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 670.684936] env[65726]: DEBUG nova.compute.manager [-] [instance: 811c1bbd-259c-4f5e-9a61-581bfba0ec4d] Deallocating network for instance {{(pid=65726) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 670.684936] env[65726]: DEBUG nova.network.neutron [-] [instance: 811c1bbd-259c-4f5e-9a61-581bfba0ec4d] deallocate_for_instance() {{(pid=65726) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 670.685813] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 670.685813] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 670.720543] env[65726]: WARNING openstack [req-889d19b5-b49d-4352-848b-37f235b066e0 req-7307af5f-8ede-44db-8223-5db2e208a88d service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 670.720977] env[65726]: WARNING openstack [req-889d19b5-b49d-4352-848b-37f235b066e0 req-7307af5f-8ede-44db-8223-5db2e208a88d service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 670.859303] env[65726]: WARNING openstack [None req-fb0f14b3-fd50-4c8d-a984-877b235dcb32 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 670.859671] env[65726]: WARNING openstack [None req-fb0f14b3-fd50-4c8d-a984-877b235dcb32 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 670.876645] env[65726]: DEBUG oslo_vmware.api [None req-06f67a27-e6e0-4078-804d-cff1137abcad tempest-ServersTestBootFromVolume-762458528 tempest-ServersTestBootFromVolume-762458528-project-member] Task: {'id': task-5115561, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 670.885657] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 670.885932] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 670.904696] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115568, 'name': CreateVM_Task} progress is 25%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 670.960665] env[65726]: DEBUG nova.scheduler.client.report [None req-64ef4520-b00a-42b8-937b-fbfc7c9b2298 tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 670.967068] env[65726]: DEBUG oslo_vmware.api [None req-362ec7a6-07fd-4c6b-b5fb-9865bdcb8c1a tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Task: {'id': task-5115562, 'name': PowerOnVM_Task, 'duration_secs': 0.760216} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 670.967903] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-362ec7a6-07fd-4c6b-b5fb-9865bdcb8c1a tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: 274ab469-61a9-4b7e-852c-074c871e3abf] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 670.967903] env[65726]: INFO nova.compute.manager [None req-362ec7a6-07fd-4c6b-b5fb-9865bdcb8c1a tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: 274ab469-61a9-4b7e-852c-074c871e3abf] Took 12.33 seconds to spawn the instance on the hypervisor. [ 670.968050] env[65726]: DEBUG nova.compute.manager [None req-362ec7a6-07fd-4c6b-b5fb-9865bdcb8c1a tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: 274ab469-61a9-4b7e-852c-074c871e3abf] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 670.968944] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb76be9d-a604-4e9f-9749-031fd1df730f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.049323] env[65726]: DEBUG oslo_vmware.api [None req-621af3d8-d8b6-4b45-86f5-d66b91010e2e tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Task: {'id': task-5115569, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 671.114116] env[65726]: WARNING openstack [req-889d19b5-b49d-4352-848b-37f235b066e0 req-7307af5f-8ede-44db-8223-5db2e208a88d service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 671.114489] env[65726]: WARNING openstack [req-889d19b5-b49d-4352-848b-37f235b066e0 req-7307af5f-8ede-44db-8223-5db2e208a88d service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 671.210846] env[65726]: DEBUG nova.virt.hardware [None req-3d632edf-0ed8-46ec-b4fd-7092ea7f050e tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T19:28:34Z,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1539f06fa3534e90acbc3a60c4b8bd3f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T19:28:35Z,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 671.211226] env[65726]: DEBUG nova.virt.hardware [None req-3d632edf-0ed8-46ec-b4fd-7092ea7f050e tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 671.211523] env[65726]: DEBUG nova.virt.hardware [None req-3d632edf-0ed8-46ec-b4fd-7092ea7f050e tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 671.211893] env[65726]: DEBUG nova.virt.hardware [None req-3d632edf-0ed8-46ec-b4fd-7092ea7f050e tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 671.212113] env[65726]: DEBUG nova.virt.hardware [None req-3d632edf-0ed8-46ec-b4fd-7092ea7f050e tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 671.212423] env[65726]: DEBUG nova.virt.hardware [None req-3d632edf-0ed8-46ec-b4fd-7092ea7f050e tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 671.212501] env[65726]: DEBUG nova.virt.hardware [None req-3d632edf-0ed8-46ec-b4fd-7092ea7f050e tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 671.212632] env[65726]: DEBUG nova.virt.hardware [None req-3d632edf-0ed8-46ec-b4fd-7092ea7f050e tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 671.212790] env[65726]: DEBUG nova.virt.hardware [None req-3d632edf-0ed8-46ec-b4fd-7092ea7f050e tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 671.212961] env[65726]: DEBUG nova.virt.hardware [None req-3d632edf-0ed8-46ec-b4fd-7092ea7f050e tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 671.213207] env[65726]: DEBUG nova.virt.hardware [None req-3d632edf-0ed8-46ec-b4fd-7092ea7f050e tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 671.214284] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d748ea5-5fa6-42ab-b572-69d2c9c4b71d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.223750] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc9d4cb9-4bb3-4546-899a-adb08b96a731 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.239958] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-3d632edf-0ed8-46ec-b4fd-7092ea7f050e tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] [instance: 3b933684-ac19-44b0-a49d-6af45501e166] Instance VIF info [] {{(pid=65726) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 671.246169] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-3d632edf-0ed8-46ec-b4fd-7092ea7f050e tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 671.246169] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3b933684-ac19-44b0-a49d-6af45501e166] Creating VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 671.246417] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-673593f4-c94a-414b-a593-f210fd91e884 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.265809] env[65726]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 671.265809] env[65726]: value = "task-5115570" [ 671.265809] env[65726]: _type = "Task" [ 671.265809] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 671.275684] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115570, 'name': CreateVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 671.381077] env[65726]: DEBUG oslo_vmware.api [None req-06f67a27-e6e0-4078-804d-cff1137abcad tempest-ServersTestBootFromVolume-762458528 tempest-ServersTestBootFromVolume-762458528-project-member] Task: {'id': task-5115561, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 671.384165] env[65726]: WARNING openstack [None req-fb0f14b3-fd50-4c8d-a984-877b235dcb32 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 671.384707] env[65726]: WARNING openstack [None req-fb0f14b3-fd50-4c8d-a984-877b235dcb32 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 671.408298] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115568, 'name': CreateVM_Task} progress is 25%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 671.469667] env[65726]: DEBUG oslo_concurrency.lockutils [None req-64ef4520-b00a-42b8-937b-fbfc7c9b2298 tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.400s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 671.474355] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b83663a4-2303-46f3-8690-72e6256a8051 tempest-ServerDiagnosticsTest-2021442419 tempest-ServerDiagnosticsTest-2021442419-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 12.772s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 671.474355] env[65726]: DEBUG nova.objects.instance [None req-b83663a4-2303-46f3-8690-72e6256a8051 tempest-ServerDiagnosticsTest-2021442419 tempest-ServerDiagnosticsTest-2021442419-project-member] Lazy-loading 'resources' on Instance uuid de5cad99-1dbc-4435-98b4-987e24ba8c3c {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 671.494154] env[65726]: INFO nova.compute.manager [None req-362ec7a6-07fd-4c6b-b5fb-9865bdcb8c1a tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: 274ab469-61a9-4b7e-852c-074c871e3abf] Took 28.44 seconds to build instance. [ 671.511533] env[65726]: INFO nova.scheduler.client.report [None req-64ef4520-b00a-42b8-937b-fbfc7c9b2298 tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Deleted allocations for instance 21730277-cef6-4bfe-9f67-de71f6f615ff [ 671.550977] env[65726]: DEBUG oslo_vmware.api [None req-621af3d8-d8b6-4b45-86f5-d66b91010e2e tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Task: {'id': task-5115569, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 671.564632] env[65726]: DEBUG nova.network.neutron [req-889d19b5-b49d-4352-848b-37f235b066e0 req-7307af5f-8ede-44db-8223-5db2e208a88d service nova] [instance: fd481728-3ef0-4a96-affd-ab2dd3f596bb] Updated VIF entry in instance network info cache for port 7df3f02b-e4e5-4840-bf9f-3498771b9184. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 671.565323] env[65726]: DEBUG nova.network.neutron [req-889d19b5-b49d-4352-848b-37f235b066e0 req-7307af5f-8ede-44db-8223-5db2e208a88d service nova] [instance: fd481728-3ef0-4a96-affd-ab2dd3f596bb] Updating instance_info_cache with network_info: [{"id": "7df3f02b-e4e5-4840-bf9f-3498771b9184", "address": "fa:16:3e:6b:c1:e9", "network": {"id": "d37113a1-409c-4c2c-81f9-8f80822b85b9", "bridge": "br-int", "label": "tempest-ServersWithSpecificFlavorTestJSON-471231219-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "79d3e960f190439fae7c61d693e9356e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "13af9422-d668-4413-b63a-766558d83a3b", "external-id": "nsx-vlan-transportzone-842", "segmentation_id": 842, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7df3f02b-e4", "ovs_interfaceid": "7df3f02b-e4e5-4840-bf9f-3498771b9184", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 671.621743] env[65726]: DEBUG nova.compute.manager [req-79f2e59d-799c-49c7-8bb1-f567afa0e67e req-22f1d995-d778-4677-a3ee-1b28a0e68f77 service nova] [instance: d7905c4f-1ba3-43b3-b4b3-aebacd2f9cc3] Received event network-vif-plugged-ee24bb04-7df3-4d06-9751-289b4bdeb753 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 671.622126] env[65726]: DEBUG oslo_concurrency.lockutils [req-79f2e59d-799c-49c7-8bb1-f567afa0e67e req-22f1d995-d778-4677-a3ee-1b28a0e68f77 service nova] Acquiring lock "d7905c4f-1ba3-43b3-b4b3-aebacd2f9cc3-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 671.623211] env[65726]: DEBUG oslo_concurrency.lockutils [req-79f2e59d-799c-49c7-8bb1-f567afa0e67e req-22f1d995-d778-4677-a3ee-1b28a0e68f77 service nova] Lock "d7905c4f-1ba3-43b3-b4b3-aebacd2f9cc3-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 671.623211] env[65726]: DEBUG oslo_concurrency.lockutils [req-79f2e59d-799c-49c7-8bb1-f567afa0e67e req-22f1d995-d778-4677-a3ee-1b28a0e68f77 service nova] Lock "d7905c4f-1ba3-43b3-b4b3-aebacd2f9cc3-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 671.623211] env[65726]: DEBUG nova.compute.manager [req-79f2e59d-799c-49c7-8bb1-f567afa0e67e req-22f1d995-d778-4677-a3ee-1b28a0e68f77 service nova] [instance: d7905c4f-1ba3-43b3-b4b3-aebacd2f9cc3] No waiting events found dispatching network-vif-plugged-ee24bb04-7df3-4d06-9751-289b4bdeb753 {{(pid=65726) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 671.623211] env[65726]: WARNING nova.compute.manager [req-79f2e59d-799c-49c7-8bb1-f567afa0e67e req-22f1d995-d778-4677-a3ee-1b28a0e68f77 service nova] [instance: d7905c4f-1ba3-43b3-b4b3-aebacd2f9cc3] Received unexpected event network-vif-plugged-ee24bb04-7df3-4d06-9751-289b4bdeb753 for instance with vm_state building and task_state spawning. [ 671.784294] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115570, 'name': CreateVM_Task, 'duration_secs': 0.452792} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 671.789029] env[65726]: DEBUG nova.network.neutron [None req-fb0f14b3-fd50-4c8d-a984-877b235dcb32 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] [instance: d7905c4f-1ba3-43b3-b4b3-aebacd2f9cc3] Updating instance_info_cache with network_info: [{"id": "ee24bb04-7df3-4d06-9751-289b4bdeb753", "address": "fa:16:3e:82:91:65", "network": {"id": "27801f9a-9912-41cf-be65-bc9903f4952f", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-781436951-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "aac5fe6c99f740ada3747088b2b0ad89", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dd7d0d95-6848-4e69-ac21-75f8db82a3b5", "external-id": "nsx-vlan-transportzone-272", "segmentation_id": 272, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapee24bb04-7d", "ovs_interfaceid": "ee24bb04-7df3-4d06-9751-289b4bdeb753", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 671.795127] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3b933684-ac19-44b0-a49d-6af45501e166] Created VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 671.795127] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3d632edf-0ed8-46ec-b4fd-7092ea7f050e tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 671.795127] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3d632edf-0ed8-46ec-b4fd-7092ea7f050e tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 671.795127] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3d632edf-0ed8-46ec-b4fd-7092ea7f050e tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 671.795127] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-82892fb0-8f9a-437a-a28c-9879f7fef0d1 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.799738] env[65726]: DEBUG oslo_vmware.api [None req-3d632edf-0ed8-46ec-b4fd-7092ea7f050e tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] Waiting for the task: (returnval){ [ 671.799738] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]526e7e17-64b1-4c27-292c-a88a37945c62" [ 671.799738] env[65726]: _type = "Task" [ 671.799738] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 671.810891] env[65726]: DEBUG oslo_vmware.api [None req-3d632edf-0ed8-46ec-b4fd-7092ea7f050e tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]526e7e17-64b1-4c27-292c-a88a37945c62, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 671.833758] env[65726]: DEBUG nova.network.neutron [-] [instance: 811c1bbd-259c-4f5e-9a61-581bfba0ec4d] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 671.878117] env[65726]: DEBUG oslo_vmware.api [None req-06f67a27-e6e0-4078-804d-cff1137abcad tempest-ServersTestBootFromVolume-762458528 tempest-ServersTestBootFromVolume-762458528-project-member] Task: {'id': task-5115561, 'name': ReconfigVM_Task, 'duration_secs': 1.601633} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 671.879097] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-06f67a27-e6e0-4078-804d-cff1137abcad tempest-ServersTestBootFromVolume-762458528 tempest-ServersTestBootFromVolume-762458528-project-member] [instance: 5a8d4357-4b44-4a19-b1da-42d188c38adc] Reconfigured VM instance instance-00000015 to attach disk [datastore1] volume-b11cb57f-92f1-416d-b59b-70cffe1353ff/volume-b11cb57f-92f1-416d-b59b-70cffe1353ff.vmdk or device None with type thin {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 671.884670] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8e9dd251-2430-4d17-af9b-0e233d630584 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.910653] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115568, 'name': CreateVM_Task, 'duration_secs': 1.35611} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 671.910653] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c2929727-a7ea-4f63-997d-474c29a305fc] Created VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 671.910653] env[65726]: DEBUG oslo_vmware.api [None req-06f67a27-e6e0-4078-804d-cff1137abcad tempest-ServersTestBootFromVolume-762458528 tempest-ServersTestBootFromVolume-762458528-project-member] Waiting for the task: (returnval){ [ 671.910653] env[65726]: value = "task-5115571" [ 671.910653] env[65726]: _type = "Task" [ 671.910653] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 671.910653] env[65726]: WARNING openstack [None req-36828e27-b100-49cc-98c2-ad0b25ca6572 tempest-ServerGroupTestJSON-125182767 tempest-ServerGroupTestJSON-125182767-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 671.911257] env[65726]: WARNING openstack [None req-36828e27-b100-49cc-98c2-ad0b25ca6572 tempest-ServerGroupTestJSON-125182767 tempest-ServerGroupTestJSON-125182767-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 671.916915] env[65726]: DEBUG oslo_concurrency.lockutils [None req-36828e27-b100-49cc-98c2-ad0b25ca6572 tempest-ServerGroupTestJSON-125182767 tempest-ServerGroupTestJSON-125182767-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 671.927147] env[65726]: DEBUG oslo_vmware.api [None req-06f67a27-e6e0-4078-804d-cff1137abcad tempest-ServersTestBootFromVolume-762458528 tempest-ServersTestBootFromVolume-762458528-project-member] Task: {'id': task-5115571, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 671.996259] env[65726]: DEBUG oslo_concurrency.lockutils [None req-362ec7a6-07fd-4c6b-b5fb-9865bdcb8c1a tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Lock "274ab469-61a9-4b7e-852c-074c871e3abf" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 54.014s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 672.029444] env[65726]: DEBUG oslo_concurrency.lockutils [None req-64ef4520-b00a-42b8-937b-fbfc7c9b2298 tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Lock "21730277-cef6-4bfe-9f67-de71f6f615ff" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 23.741s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 672.057396] env[65726]: DEBUG oslo_vmware.api [None req-621af3d8-d8b6-4b45-86f5-d66b91010e2e tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Task: {'id': task-5115569, 'name': PowerOnVM_Task} progress is 94%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 672.071030] env[65726]: DEBUG oslo_concurrency.lockutils [req-889d19b5-b49d-4352-848b-37f235b066e0 req-7307af5f-8ede-44db-8223-5db2e208a88d service nova] Releasing lock "refresh_cache-fd481728-3ef0-4a96-affd-ab2dd3f596bb" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 672.071624] env[65726]: DEBUG nova.compute.manager [req-889d19b5-b49d-4352-848b-37f235b066e0 req-7307af5f-8ede-44db-8223-5db2e208a88d service nova] [instance: 5a8d4357-4b44-4a19-b1da-42d188c38adc] Received event network-vif-plugged-9aaacd82-31c1-4cfa-bbce-860b30db74af {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 672.071624] env[65726]: DEBUG oslo_concurrency.lockutils [req-889d19b5-b49d-4352-848b-37f235b066e0 req-7307af5f-8ede-44db-8223-5db2e208a88d service nova] Acquiring lock "5a8d4357-4b44-4a19-b1da-42d188c38adc-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 672.071805] env[65726]: DEBUG oslo_concurrency.lockutils [req-889d19b5-b49d-4352-848b-37f235b066e0 req-7307af5f-8ede-44db-8223-5db2e208a88d service nova] Lock "5a8d4357-4b44-4a19-b1da-42d188c38adc-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 672.072249] env[65726]: DEBUG oslo_concurrency.lockutils [req-889d19b5-b49d-4352-848b-37f235b066e0 req-7307af5f-8ede-44db-8223-5db2e208a88d service nova] Lock "5a8d4357-4b44-4a19-b1da-42d188c38adc-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 672.072374] env[65726]: DEBUG nova.compute.manager [req-889d19b5-b49d-4352-848b-37f235b066e0 req-7307af5f-8ede-44db-8223-5db2e208a88d service nova] [instance: 5a8d4357-4b44-4a19-b1da-42d188c38adc] No waiting events found dispatching network-vif-plugged-9aaacd82-31c1-4cfa-bbce-860b30db74af {{(pid=65726) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 672.073452] env[65726]: WARNING nova.compute.manager [req-889d19b5-b49d-4352-848b-37f235b066e0 req-7307af5f-8ede-44db-8223-5db2e208a88d service nova] [instance: 5a8d4357-4b44-4a19-b1da-42d188c38adc] Received unexpected event network-vif-plugged-9aaacd82-31c1-4cfa-bbce-860b30db74af for instance with vm_state building and task_state spawning. [ 672.073452] env[65726]: DEBUG nova.compute.manager [req-889d19b5-b49d-4352-848b-37f235b066e0 req-7307af5f-8ede-44db-8223-5db2e208a88d service nova] [instance: 5a8d4357-4b44-4a19-b1da-42d188c38adc] Received event network-changed-9aaacd82-31c1-4cfa-bbce-860b30db74af {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 672.073452] env[65726]: DEBUG nova.compute.manager [req-889d19b5-b49d-4352-848b-37f235b066e0 req-7307af5f-8ede-44db-8223-5db2e208a88d service nova] [instance: 5a8d4357-4b44-4a19-b1da-42d188c38adc] Refreshing instance network info cache due to event network-changed-9aaacd82-31c1-4cfa-bbce-860b30db74af. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 672.073452] env[65726]: DEBUG oslo_concurrency.lockutils [req-889d19b5-b49d-4352-848b-37f235b066e0 req-7307af5f-8ede-44db-8223-5db2e208a88d service nova] Acquiring lock "refresh_cache-5a8d4357-4b44-4a19-b1da-42d188c38adc" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 672.073661] env[65726]: DEBUG oslo_concurrency.lockutils [req-889d19b5-b49d-4352-848b-37f235b066e0 req-7307af5f-8ede-44db-8223-5db2e208a88d service nova] Acquired lock "refresh_cache-5a8d4357-4b44-4a19-b1da-42d188c38adc" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 672.073701] env[65726]: DEBUG nova.network.neutron [req-889d19b5-b49d-4352-848b-37f235b066e0 req-7307af5f-8ede-44db-8223-5db2e208a88d service nova] [instance: 5a8d4357-4b44-4a19-b1da-42d188c38adc] Refreshing network info cache for port 9aaacd82-31c1-4cfa-bbce-860b30db74af {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 672.295598] env[65726]: DEBUG oslo_concurrency.lockutils [None req-fb0f14b3-fd50-4c8d-a984-877b235dcb32 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Releasing lock "refresh_cache-d7905c4f-1ba3-43b3-b4b3-aebacd2f9cc3" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 672.295995] env[65726]: DEBUG nova.compute.manager [None req-fb0f14b3-fd50-4c8d-a984-877b235dcb32 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] [instance: d7905c4f-1ba3-43b3-b4b3-aebacd2f9cc3] Instance network_info: |[{"id": "ee24bb04-7df3-4d06-9751-289b4bdeb753", "address": "fa:16:3e:82:91:65", "network": {"id": "27801f9a-9912-41cf-be65-bc9903f4952f", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-781436951-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "aac5fe6c99f740ada3747088b2b0ad89", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dd7d0d95-6848-4e69-ac21-75f8db82a3b5", "external-id": "nsx-vlan-transportzone-272", "segmentation_id": 272, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapee24bb04-7d", "ovs_interfaceid": "ee24bb04-7df3-4d06-9751-289b4bdeb753", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 672.296587] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-fb0f14b3-fd50-4c8d-a984-877b235dcb32 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] [instance: d7905c4f-1ba3-43b3-b4b3-aebacd2f9cc3] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:82:91:65', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'dd7d0d95-6848-4e69-ac21-75f8db82a3b5', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ee24bb04-7df3-4d06-9751-289b4bdeb753', 'vif_model': 'vmxnet3'}] {{(pid=65726) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 672.306306] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-fb0f14b3-fd50-4c8d-a984-877b235dcb32 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Creating folder: Project (aac5fe6c99f740ada3747088b2b0ad89). Parent ref: group-v995008. {{(pid=65726) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 672.309635] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-92a1a0c2-b17e-428c-96f0-6db69e63d210 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.324295] env[65726]: DEBUG oslo_vmware.api [None req-3d632edf-0ed8-46ec-b4fd-7092ea7f050e tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]526e7e17-64b1-4c27-292c-a88a37945c62, 'name': SearchDatastore_Task, 'duration_secs': 0.013173} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 672.324660] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3d632edf-0ed8-46ec-b4fd-7092ea7f050e tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 672.324931] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-3d632edf-0ed8-46ec-b4fd-7092ea7f050e tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] [instance: 3b933684-ac19-44b0-a49d-6af45501e166] Processing image b52362a3-ee8a-4cbf-b06f-513b0cc8f95c {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 672.325410] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3d632edf-0ed8-46ec-b4fd-7092ea7f050e tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 672.325410] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3d632edf-0ed8-46ec-b4fd-7092ea7f050e tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 672.325573] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-3d632edf-0ed8-46ec-b4fd-7092ea7f050e tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 672.325968] env[65726]: DEBUG oslo_concurrency.lockutils [None req-36828e27-b100-49cc-98c2-ad0b25ca6572 tempest-ServerGroupTestJSON-125182767 tempest-ServerGroupTestJSON-125182767-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 672.326240] env[65726]: DEBUG oslo_concurrency.lockutils [None req-36828e27-b100-49cc-98c2-ad0b25ca6572 tempest-ServerGroupTestJSON-125182767 tempest-ServerGroupTestJSON-125182767-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 672.326485] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-cd5407a0-1c9c-4a5d-8a97-f082efdcbab2 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.331829] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5deee222-7cd5-4c93-a1cc-e484680f2931 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.332037] env[65726]: INFO nova.virt.vmwareapi.vm_util [None req-fb0f14b3-fd50-4c8d-a984-877b235dcb32 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Created folder: Project (aac5fe6c99f740ada3747088b2b0ad89) in parent group-v995008. [ 672.332224] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-fb0f14b3-fd50-4c8d-a984-877b235dcb32 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Creating folder: Instances. Parent ref: group-v995072. {{(pid=65726) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 672.336605] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-6a0737f2-cc0b-4fbf-b8cc-da099e262e91 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.340252] env[65726]: INFO nova.compute.manager [-] [instance: 811c1bbd-259c-4f5e-9a61-581bfba0ec4d] Took 1.66 seconds to deallocate network for instance. [ 672.344540] env[65726]: DEBUG oslo_vmware.api [None req-36828e27-b100-49cc-98c2-ad0b25ca6572 tempest-ServerGroupTestJSON-125182767 tempest-ServerGroupTestJSON-125182767-project-member] Waiting for the task: (returnval){ [ 672.344540] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52ecdc45-07a2-3c3f-923a-9b47ba28d3f4" [ 672.344540] env[65726]: _type = "Task" [ 672.344540] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 672.350322] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-3d632edf-0ed8-46ec-b4fd-7092ea7f050e tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 672.350940] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-3d632edf-0ed8-46ec-b4fd-7092ea7f050e tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65726) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 672.359365] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9ebece1f-ca30-4988-8843-fc605f5a4a7a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.364156] env[65726]: INFO nova.virt.vmwareapi.vm_util [None req-fb0f14b3-fd50-4c8d-a984-877b235dcb32 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Created folder: Instances in parent group-v995072. [ 672.364390] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-fb0f14b3-fd50-4c8d-a984-877b235dcb32 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 672.368316] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d7905c4f-1ba3-43b3-b4b3-aebacd2f9cc3] Creating VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 672.368316] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-09a59e4a-7cd1-46a6-a485-9be224d864f0 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.389832] env[65726]: DEBUG oslo_vmware.api [None req-36828e27-b100-49cc-98c2-ad0b25ca6572 tempest-ServerGroupTestJSON-125182767 tempest-ServerGroupTestJSON-125182767-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52ecdc45-07a2-3c3f-923a-9b47ba28d3f4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 672.390221] env[65726]: DEBUG oslo_vmware.api [None req-3d632edf-0ed8-46ec-b4fd-7092ea7f050e tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] Waiting for the task: (returnval){ [ 672.390221] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]527380cc-ec40-34bf-a5db-6abd55b5f778" [ 672.390221] env[65726]: _type = "Task" [ 672.390221] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 672.401721] env[65726]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 672.401721] env[65726]: value = "task-5115574" [ 672.401721] env[65726]: _type = "Task" [ 672.401721] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 672.410647] env[65726]: DEBUG oslo_vmware.api [None req-3d632edf-0ed8-46ec-b4fd-7092ea7f050e tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]527380cc-ec40-34bf-a5db-6abd55b5f778, 'name': SearchDatastore_Task, 'duration_secs': 0.020752} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 672.412245] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5d209ddc-f1ee-4753-b137-50ca13d7c476 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.418498] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115574, 'name': CreateVM_Task} progress is 6%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 672.424532] env[65726]: DEBUG oslo_vmware.api [None req-3d632edf-0ed8-46ec-b4fd-7092ea7f050e tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] Waiting for the task: (returnval){ [ 672.424532] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]523724a8-564e-ca25-5fef-92616d915c97" [ 672.424532] env[65726]: _type = "Task" [ 672.424532] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 672.436163] env[65726]: DEBUG oslo_vmware.api [None req-06f67a27-e6e0-4078-804d-cff1137abcad tempest-ServersTestBootFromVolume-762458528 tempest-ServersTestBootFromVolume-762458528-project-member] Task: {'id': task-5115571, 'name': ReconfigVM_Task, 'duration_secs': 0.2783} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 672.437103] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-06f67a27-e6e0-4078-804d-cff1137abcad tempest-ServersTestBootFromVolume-762458528 tempest-ServersTestBootFromVolume-762458528-project-member] [instance: 5a8d4357-4b44-4a19-b1da-42d188c38adc] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-995020', 'volume_id': 'b11cb57f-92f1-416d-b59b-70cffe1353ff', 'name': 'volume-b11cb57f-92f1-416d-b59b-70cffe1353ff', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '5a8d4357-4b44-4a19-b1da-42d188c38adc', 'attached_at': '', 'detached_at': '', 'volume_id': 'b11cb57f-92f1-416d-b59b-70cffe1353ff', 'serial': 'b11cb57f-92f1-416d-b59b-70cffe1353ff'} {{(pid=65726) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 672.437726] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-65d98796-c51a-43f0-9a0b-f2252fe971e3 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.443096] env[65726]: DEBUG oslo_vmware.api [None req-3d632edf-0ed8-46ec-b4fd-7092ea7f050e tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]523724a8-564e-ca25-5fef-92616d915c97, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 672.450165] env[65726]: DEBUG oslo_vmware.api [None req-06f67a27-e6e0-4078-804d-cff1137abcad tempest-ServersTestBootFromVolume-762458528 tempest-ServersTestBootFromVolume-762458528-project-member] Waiting for the task: (returnval){ [ 672.450165] env[65726]: value = "task-5115575" [ 672.450165] env[65726]: _type = "Task" [ 672.450165] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 672.463446] env[65726]: DEBUG oslo_vmware.api [None req-06f67a27-e6e0-4078-804d-cff1137abcad tempest-ServersTestBootFromVolume-762458528 tempest-ServersTestBootFromVolume-762458528-project-member] Task: {'id': task-5115575, 'name': Rename_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 672.500902] env[65726]: DEBUG nova.compute.manager [None req-5dd7a34f-2002-4a12-9099-b0b630d2774d tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] [instance: 815f056b-2d88-40dd-9ef7-a891ec8d92cf] Starting instance... {{(pid=65726) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 672.559098] env[65726]: DEBUG oslo_vmware.api [None req-621af3d8-d8b6-4b45-86f5-d66b91010e2e tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Task: {'id': task-5115569, 'name': PowerOnVM_Task, 'duration_secs': 1.690052} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 672.559098] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-621af3d8-d8b6-4b45-86f5-d66b91010e2e tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] [instance: fd481728-3ef0-4a96-affd-ab2dd3f596bb] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 672.559256] env[65726]: INFO nova.compute.manager [None req-621af3d8-d8b6-4b45-86f5-d66b91010e2e tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] [instance: fd481728-3ef0-4a96-affd-ab2dd3f596bb] Took 11.54 seconds to spawn the instance on the hypervisor. [ 672.559545] env[65726]: DEBUG nova.compute.manager [None req-621af3d8-d8b6-4b45-86f5-d66b91010e2e tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] [instance: fd481728-3ef0-4a96-affd-ab2dd3f596bb] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 672.560719] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27fad8b3-b54f-4bb1-8f38-3374a822e0bd {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.582380] env[65726]: WARNING openstack [req-889d19b5-b49d-4352-848b-37f235b066e0 req-7307af5f-8ede-44db-8223-5db2e208a88d service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 672.584066] env[65726]: WARNING openstack [req-889d19b5-b49d-4352-848b-37f235b066e0 req-7307af5f-8ede-44db-8223-5db2e208a88d service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 672.590441] env[65726]: DEBUG oslo_concurrency.lockutils [None req-fd996099-26ef-437f-83c3-98cb140e9de7 tempest-TenantUsagesTestJSON-327211524 tempest-TenantUsagesTestJSON-327211524-project-member] Acquiring lock "7748c23f-9ea9-4172-8ab7-187678272bb3" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 672.590829] env[65726]: DEBUG oslo_concurrency.lockutils [None req-fd996099-26ef-437f-83c3-98cb140e9de7 tempest-TenantUsagesTestJSON-327211524 tempest-TenantUsagesTestJSON-327211524-project-member] Lock "7748c23f-9ea9-4172-8ab7-187678272bb3" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 672.590991] env[65726]: DEBUG oslo_concurrency.lockutils [None req-fd996099-26ef-437f-83c3-98cb140e9de7 tempest-TenantUsagesTestJSON-327211524 tempest-TenantUsagesTestJSON-327211524-project-member] Acquiring lock "7748c23f-9ea9-4172-8ab7-187678272bb3-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 672.591243] env[65726]: DEBUG oslo_concurrency.lockutils [None req-fd996099-26ef-437f-83c3-98cb140e9de7 tempest-TenantUsagesTestJSON-327211524 tempest-TenantUsagesTestJSON-327211524-project-member] Lock "7748c23f-9ea9-4172-8ab7-187678272bb3-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 672.591352] env[65726]: DEBUG oslo_concurrency.lockutils [None req-fd996099-26ef-437f-83c3-98cb140e9de7 tempest-TenantUsagesTestJSON-327211524 tempest-TenantUsagesTestJSON-327211524-project-member] Lock "7748c23f-9ea9-4172-8ab7-187678272bb3-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 672.593329] env[65726]: INFO nova.compute.manager [None req-fd996099-26ef-437f-83c3-98cb140e9de7 tempest-TenantUsagesTestJSON-327211524 tempest-TenantUsagesTestJSON-327211524-project-member] [instance: 7748c23f-9ea9-4172-8ab7-187678272bb3] Terminating instance [ 672.699661] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-682da9e7-50a1-41e4-884d-1916aabf56c7 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.708402] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ffa4191-e635-4a62-9c40-4d3911813fd1 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.746737] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c68cccb3-1006-4979-a921-429e7396a4f7 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.757517] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e7a48c5-a11c-4784-921f-c00bf75e7295 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.771690] env[65726]: DEBUG nova.compute.provider_tree [None req-b83663a4-2303-46f3-8690-72e6256a8051 tempest-ServerDiagnosticsTest-2021442419 tempest-ServerDiagnosticsTest-2021442419-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 672.854945] env[65726]: DEBUG oslo_concurrency.lockutils [None req-34fc70a6-ee2b-4816-b933-3db54eee74ae tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 672.863463] env[65726]: DEBUG oslo_vmware.api [None req-36828e27-b100-49cc-98c2-ad0b25ca6572 tempest-ServerGroupTestJSON-125182767 tempest-ServerGroupTestJSON-125182767-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52ecdc45-07a2-3c3f-923a-9b47ba28d3f4, 'name': SearchDatastore_Task, 'duration_secs': 0.026894} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 672.863928] env[65726]: DEBUG oslo_concurrency.lockutils [None req-36828e27-b100-49cc-98c2-ad0b25ca6572 tempest-ServerGroupTestJSON-125182767 tempest-ServerGroupTestJSON-125182767-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 672.864350] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-36828e27-b100-49cc-98c2-ad0b25ca6572 tempest-ServerGroupTestJSON-125182767 tempest-ServerGroupTestJSON-125182767-project-member] [instance: c2929727-a7ea-4f63-997d-474c29a305fc] Processing image b52362a3-ee8a-4cbf-b06f-513b0cc8f95c {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 672.864667] env[65726]: DEBUG oslo_concurrency.lockutils [None req-36828e27-b100-49cc-98c2-ad0b25ca6572 tempest-ServerGroupTestJSON-125182767 tempest-ServerGroupTestJSON-125182767-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 672.921737] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115574, 'name': CreateVM_Task} progress is 99%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 672.943687] env[65726]: DEBUG oslo_vmware.api [None req-3d632edf-0ed8-46ec-b4fd-7092ea7f050e tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]523724a8-564e-ca25-5fef-92616d915c97, 'name': SearchDatastore_Task, 'duration_secs': 0.020373} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 672.943687] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3d632edf-0ed8-46ec-b4fd-7092ea7f050e tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 672.943687] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-3d632edf-0ed8-46ec-b4fd-7092ea7f050e tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] 3b933684-ac19-44b0-a49d-6af45501e166/3b933684-ac19-44b0-a49d-6af45501e166.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 672.943969] env[65726]: DEBUG oslo_concurrency.lockutils [None req-36828e27-b100-49cc-98c2-ad0b25ca6572 tempest-ServerGroupTestJSON-125182767 tempest-ServerGroupTestJSON-125182767-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 672.944149] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-36828e27-b100-49cc-98c2-ad0b25ca6572 tempest-ServerGroupTestJSON-125182767 tempest-ServerGroupTestJSON-125182767-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 672.944379] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3043fdba-36b2-4dae-b522-d90034a174f9 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.947356] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-41ead90d-fe0b-4f4e-9a12-9871d7d3542f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.957564] env[65726]: WARNING openstack [req-889d19b5-b49d-4352-848b-37f235b066e0 req-7307af5f-8ede-44db-8223-5db2e208a88d service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 672.957980] env[65726]: WARNING openstack [req-889d19b5-b49d-4352-848b-37f235b066e0 req-7307af5f-8ede-44db-8223-5db2e208a88d service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 672.968194] env[65726]: DEBUG oslo_vmware.api [None req-3d632edf-0ed8-46ec-b4fd-7092ea7f050e tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] Waiting for the task: (returnval){ [ 672.968194] env[65726]: value = "task-5115576" [ 672.968194] env[65726]: _type = "Task" [ 672.968194] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 672.974761] env[65726]: DEBUG oslo_vmware.api [None req-06f67a27-e6e0-4078-804d-cff1137abcad tempest-ServersTestBootFromVolume-762458528 tempest-ServersTestBootFromVolume-762458528-project-member] Task: {'id': task-5115575, 'name': Rename_Task, 'duration_secs': 0.226746} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 672.974761] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-36828e27-b100-49cc-98c2-ad0b25ca6572 tempest-ServerGroupTestJSON-125182767 tempest-ServerGroupTestJSON-125182767-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 672.974761] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-36828e27-b100-49cc-98c2-ad0b25ca6572 tempest-ServerGroupTestJSON-125182767 tempest-ServerGroupTestJSON-125182767-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65726) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 672.977808] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-06f67a27-e6e0-4078-804d-cff1137abcad tempest-ServersTestBootFromVolume-762458528 tempest-ServersTestBootFromVolume-762458528-project-member] [instance: 5a8d4357-4b44-4a19-b1da-42d188c38adc] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 672.978082] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e364dcad-3c04-4850-b0f4-c17658ac4f79 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.981057] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7a34a322-52f8-4222-a71c-f54c50160d12 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.988649] env[65726]: DEBUG oslo_vmware.api [None req-3d632edf-0ed8-46ec-b4fd-7092ea7f050e tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] Task: {'id': task-5115576, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 672.991068] env[65726]: DEBUG oslo_vmware.api [None req-36828e27-b100-49cc-98c2-ad0b25ca6572 tempest-ServerGroupTestJSON-125182767 tempest-ServerGroupTestJSON-125182767-project-member] Waiting for the task: (returnval){ [ 672.991068] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5258902f-371e-cdff-6216-9badad040b9d" [ 672.991068] env[65726]: _type = "Task" [ 672.991068] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 672.992078] env[65726]: DEBUG oslo_vmware.api [None req-06f67a27-e6e0-4078-804d-cff1137abcad tempest-ServersTestBootFromVolume-762458528 tempest-ServersTestBootFromVolume-762458528-project-member] Waiting for the task: (returnval){ [ 672.992078] env[65726]: value = "task-5115577" [ 672.992078] env[65726]: _type = "Task" [ 672.992078] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 673.009601] env[65726]: DEBUG oslo_vmware.api [None req-36828e27-b100-49cc-98c2-ad0b25ca6572 tempest-ServerGroupTestJSON-125182767 tempest-ServerGroupTestJSON-125182767-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5258902f-371e-cdff-6216-9badad040b9d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 673.011606] env[65726]: DEBUG oslo_vmware.api [None req-06f67a27-e6e0-4078-804d-cff1137abcad tempest-ServersTestBootFromVolume-762458528 tempest-ServersTestBootFromVolume-762458528-project-member] Task: {'id': task-5115577, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 673.033951] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5dd7a34f-2002-4a12-9099-b0b630d2774d tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 673.090144] env[65726]: INFO nova.compute.manager [None req-621af3d8-d8b6-4b45-86f5-d66b91010e2e tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] [instance: fd481728-3ef0-4a96-affd-ab2dd3f596bb] Took 27.06 seconds to build instance. [ 673.102746] env[65726]: DEBUG nova.compute.manager [None req-fd996099-26ef-437f-83c3-98cb140e9de7 tempest-TenantUsagesTestJSON-327211524 tempest-TenantUsagesTestJSON-327211524-project-member] [instance: 7748c23f-9ea9-4172-8ab7-187678272bb3] Start destroying the instance on the hypervisor. {{(pid=65726) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 673.103487] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-fd996099-26ef-437f-83c3-98cb140e9de7 tempest-TenantUsagesTestJSON-327211524 tempest-TenantUsagesTestJSON-327211524-project-member] [instance: 7748c23f-9ea9-4172-8ab7-187678272bb3] Destroying instance {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 673.103939] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfb0516a-8437-4e84-b994-bf6aea1d344c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.112714] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd996099-26ef-437f-83c3-98cb140e9de7 tempest-TenantUsagesTestJSON-327211524 tempest-TenantUsagesTestJSON-327211524-project-member] [instance: 7748c23f-9ea9-4172-8ab7-187678272bb3] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 673.113018] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-112d1047-67a4-4d3e-a6e5-f4a16edeb656 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.124687] env[65726]: DEBUG oslo_vmware.api [None req-fd996099-26ef-437f-83c3-98cb140e9de7 tempest-TenantUsagesTestJSON-327211524 tempest-TenantUsagesTestJSON-327211524-project-member] Waiting for the task: (returnval){ [ 673.124687] env[65726]: value = "task-5115578" [ 673.124687] env[65726]: _type = "Task" [ 673.124687] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 673.135809] env[65726]: DEBUG oslo_vmware.api [None req-fd996099-26ef-437f-83c3-98cb140e9de7 tempest-TenantUsagesTestJSON-327211524 tempest-TenantUsagesTestJSON-327211524-project-member] Task: {'id': task-5115578, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 673.275619] env[65726]: DEBUG nova.scheduler.client.report [None req-b83663a4-2303-46f3-8690-72e6256a8051 tempest-ServerDiagnosticsTest-2021442419 tempest-ServerDiagnosticsTest-2021442419-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 673.343979] env[65726]: WARNING openstack [req-889d19b5-b49d-4352-848b-37f235b066e0 req-7307af5f-8ede-44db-8223-5db2e208a88d service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 673.343979] env[65726]: WARNING openstack [req-889d19b5-b49d-4352-848b-37f235b066e0 req-7307af5f-8ede-44db-8223-5db2e208a88d service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 673.415861] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115574, 'name': CreateVM_Task, 'duration_secs': 0.571537} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 673.416707] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d7905c4f-1ba3-43b3-b4b3-aebacd2f9cc3] Created VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 673.417987] env[65726]: WARNING openstack [None req-fb0f14b3-fd50-4c8d-a984-877b235dcb32 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 673.419024] env[65726]: WARNING openstack [None req-fb0f14b3-fd50-4c8d-a984-877b235dcb32 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 673.431038] env[65726]: DEBUG oslo_concurrency.lockutils [None req-fb0f14b3-fd50-4c8d-a984-877b235dcb32 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 673.431038] env[65726]: DEBUG oslo_concurrency.lockutils [None req-fb0f14b3-fd50-4c8d-a984-877b235dcb32 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 673.431038] env[65726]: DEBUG oslo_concurrency.lockutils [None req-fb0f14b3-fd50-4c8d-a984-877b235dcb32 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 673.431038] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-178339c9-7799-4b11-a792-f4f92187f226 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.438110] env[65726]: DEBUG oslo_vmware.api [None req-fb0f14b3-fd50-4c8d-a984-877b235dcb32 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Waiting for the task: (returnval){ [ 673.438110] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52e6f445-4058-3e16-5628-b2f0c1c8e86e" [ 673.438110] env[65726]: _type = "Task" [ 673.438110] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 673.451570] env[65726]: DEBUG oslo_vmware.api [None req-fb0f14b3-fd50-4c8d-a984-877b235dcb32 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52e6f445-4058-3e16-5628-b2f0c1c8e86e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 673.487730] env[65726]: DEBUG oslo_vmware.api [None req-3d632edf-0ed8-46ec-b4fd-7092ea7f050e tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] Task: {'id': task-5115576, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 673.510234] env[65726]: DEBUG oslo_vmware.api [None req-36828e27-b100-49cc-98c2-ad0b25ca6572 tempest-ServerGroupTestJSON-125182767 tempest-ServerGroupTestJSON-125182767-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5258902f-371e-cdff-6216-9badad040b9d, 'name': SearchDatastore_Task, 'duration_secs': 0.027265} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 673.516032] env[65726]: DEBUG oslo_vmware.api [None req-06f67a27-e6e0-4078-804d-cff1137abcad tempest-ServersTestBootFromVolume-762458528 tempest-ServersTestBootFromVolume-762458528-project-member] Task: {'id': task-5115577, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 673.516032] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5f98d078-9691-4b94-853c-c312da2ae5c1 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.523281] env[65726]: DEBUG oslo_vmware.api [None req-36828e27-b100-49cc-98c2-ad0b25ca6572 tempest-ServerGroupTestJSON-125182767 tempest-ServerGroupTestJSON-125182767-project-member] Waiting for the task: (returnval){ [ 673.523281] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5212130e-5ebf-ad80-4ec5-9c15c0af383d" [ 673.523281] env[65726]: _type = "Task" [ 673.523281] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 673.539255] env[65726]: DEBUG oslo_vmware.api [None req-36828e27-b100-49cc-98c2-ad0b25ca6572 tempest-ServerGroupTestJSON-125182767 tempest-ServerGroupTestJSON-125182767-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5212130e-5ebf-ad80-4ec5-9c15c0af383d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 673.598760] env[65726]: DEBUG oslo_concurrency.lockutils [None req-621af3d8-d8b6-4b45-86f5-d66b91010e2e tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Lock "fd481728-3ef0-4a96-affd-ab2dd3f596bb" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 54.533s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 673.638843] env[65726]: DEBUG oslo_vmware.api [None req-fd996099-26ef-437f-83c3-98cb140e9de7 tempest-TenantUsagesTestJSON-327211524 tempest-TenantUsagesTestJSON-327211524-project-member] Task: {'id': task-5115578, 'name': PowerOffVM_Task, 'duration_secs': 0.280247} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 673.639230] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd996099-26ef-437f-83c3-98cb140e9de7 tempest-TenantUsagesTestJSON-327211524 tempest-TenantUsagesTestJSON-327211524-project-member] [instance: 7748c23f-9ea9-4172-8ab7-187678272bb3] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 673.639507] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-fd996099-26ef-437f-83c3-98cb140e9de7 tempest-TenantUsagesTestJSON-327211524 tempest-TenantUsagesTestJSON-327211524-project-member] [instance: 7748c23f-9ea9-4172-8ab7-187678272bb3] Unregistering the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 673.639866] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b70f3d98-7d0c-43bb-8d96-06c8c2f46028 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.712498] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-fd996099-26ef-437f-83c3-98cb140e9de7 tempest-TenantUsagesTestJSON-327211524 tempest-TenantUsagesTestJSON-327211524-project-member] [instance: 7748c23f-9ea9-4172-8ab7-187678272bb3] Unregistered the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 673.712498] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-fd996099-26ef-437f-83c3-98cb140e9de7 tempest-TenantUsagesTestJSON-327211524 tempest-TenantUsagesTestJSON-327211524-project-member] [instance: 7748c23f-9ea9-4172-8ab7-187678272bb3] Deleting contents of the VM from datastore datastore2 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 673.712498] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-fd996099-26ef-437f-83c3-98cb140e9de7 tempest-TenantUsagesTestJSON-327211524 tempest-TenantUsagesTestJSON-327211524-project-member] Deleting the datastore file [datastore2] 7748c23f-9ea9-4172-8ab7-187678272bb3 {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 673.712498] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ed44f13c-a059-4db9-a6b1-dc2d97513632 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.722052] env[65726]: DEBUG oslo_vmware.api [None req-fd996099-26ef-437f-83c3-98cb140e9de7 tempest-TenantUsagesTestJSON-327211524 tempest-TenantUsagesTestJSON-327211524-project-member] Waiting for the task: (returnval){ [ 673.722052] env[65726]: value = "task-5115580" [ 673.722052] env[65726]: _type = "Task" [ 673.722052] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 673.736596] env[65726]: DEBUG oslo_vmware.api [None req-fd996099-26ef-437f-83c3-98cb140e9de7 tempest-TenantUsagesTestJSON-327211524 tempest-TenantUsagesTestJSON-327211524-project-member] Task: {'id': task-5115580, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 673.784810] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b83663a4-2303-46f3-8690-72e6256a8051 tempest-ServerDiagnosticsTest-2021442419 tempest-ServerDiagnosticsTest-2021442419-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.313s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 673.790031] env[65726]: DEBUG oslo_concurrency.lockutils [None req-efa812b8-b008-4879-b549-6ccf454f374f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.337s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 673.790031] env[65726]: INFO nova.compute.claims [None req-efa812b8-b008-4879-b549-6ccf454f374f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] [instance: ddd36db0-28ad-4582-994b-7eef8b08ed46] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 673.822162] env[65726]: INFO nova.scheduler.client.report [None req-b83663a4-2303-46f3-8690-72e6256a8051 tempest-ServerDiagnosticsTest-2021442419 tempest-ServerDiagnosticsTest-2021442419-project-member] Deleted allocations for instance de5cad99-1dbc-4435-98b4-987e24ba8c3c [ 673.950934] env[65726]: DEBUG nova.network.neutron [req-889d19b5-b49d-4352-848b-37f235b066e0 req-7307af5f-8ede-44db-8223-5db2e208a88d service nova] [instance: 5a8d4357-4b44-4a19-b1da-42d188c38adc] Updated VIF entry in instance network info cache for port 9aaacd82-31c1-4cfa-bbce-860b30db74af. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 673.952138] env[65726]: DEBUG nova.network.neutron [req-889d19b5-b49d-4352-848b-37f235b066e0 req-7307af5f-8ede-44db-8223-5db2e208a88d service nova] [instance: 5a8d4357-4b44-4a19-b1da-42d188c38adc] Updating instance_info_cache with network_info: [{"id": "9aaacd82-31c1-4cfa-bbce-860b30db74af", "address": "fa:16:3e:a2:ff:16", "network": {"id": "bee3bacb-6a14-41ab-a9a3-a82d9ca0b175", "bridge": "br-int", "label": "tempest-ServersTestBootFromVolume-514317700-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "c52b5289c78f45d3942a6c0a4b026207", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dad4f433-bb0b-45c7-8040-972ef2277f75", "external-id": "nsx-vlan-transportzone-451", "segmentation_id": 451, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9aaacd82-31", "ovs_interfaceid": "9aaacd82-31c1-4cfa-bbce-860b30db74af", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 673.957580] env[65726]: DEBUG oslo_vmware.api [None req-fb0f14b3-fd50-4c8d-a984-877b235dcb32 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52e6f445-4058-3e16-5628-b2f0c1c8e86e, 'name': SearchDatastore_Task, 'duration_secs': 0.066659} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 673.959173] env[65726]: DEBUG oslo_concurrency.lockutils [None req-fb0f14b3-fd50-4c8d-a984-877b235dcb32 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 673.960357] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-fb0f14b3-fd50-4c8d-a984-877b235dcb32 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] [instance: d7905c4f-1ba3-43b3-b4b3-aebacd2f9cc3] Processing image b52362a3-ee8a-4cbf-b06f-513b0cc8f95c {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 673.960357] env[65726]: DEBUG oslo_concurrency.lockutils [None req-fb0f14b3-fd50-4c8d-a984-877b235dcb32 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 673.988031] env[65726]: DEBUG oslo_vmware.api [None req-3d632edf-0ed8-46ec-b4fd-7092ea7f050e tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] Task: {'id': task-5115576, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.648828} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 673.988031] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-3d632edf-0ed8-46ec-b4fd-7092ea7f050e tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] 3b933684-ac19-44b0-a49d-6af45501e166/3b933684-ac19-44b0-a49d-6af45501e166.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 673.988233] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-3d632edf-0ed8-46ec-b4fd-7092ea7f050e tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] [instance: 3b933684-ac19-44b0-a49d-6af45501e166] Extending root virtual disk to 1048576 {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 673.988707] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-7ec83ff3-75d7-42de-8c98-d5fd5986a2e8 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.996279] env[65726]: DEBUG oslo_vmware.api [None req-3d632edf-0ed8-46ec-b4fd-7092ea7f050e tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] Waiting for the task: (returnval){ [ 673.996279] env[65726]: value = "task-5115581" [ 673.996279] env[65726]: _type = "Task" [ 673.996279] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 674.011117] env[65726]: DEBUG oslo_vmware.api [None req-06f67a27-e6e0-4078-804d-cff1137abcad tempest-ServersTestBootFromVolume-762458528 tempest-ServersTestBootFromVolume-762458528-project-member] Task: {'id': task-5115577, 'name': PowerOnVM_Task, 'duration_secs': 0.657588} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 674.014122] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-06f67a27-e6e0-4078-804d-cff1137abcad tempest-ServersTestBootFromVolume-762458528 tempest-ServersTestBootFromVolume-762458528-project-member] [instance: 5a8d4357-4b44-4a19-b1da-42d188c38adc] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 674.014367] env[65726]: INFO nova.compute.manager [None req-06f67a27-e6e0-4078-804d-cff1137abcad tempest-ServersTestBootFromVolume-762458528 tempest-ServersTestBootFromVolume-762458528-project-member] [instance: 5a8d4357-4b44-4a19-b1da-42d188c38adc] Took 7.96 seconds to spawn the instance on the hypervisor. [ 674.014557] env[65726]: DEBUG nova.compute.manager [None req-06f67a27-e6e0-4078-804d-cff1137abcad tempest-ServersTestBootFromVolume-762458528 tempest-ServersTestBootFromVolume-762458528-project-member] [instance: 5a8d4357-4b44-4a19-b1da-42d188c38adc] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 674.014877] env[65726]: DEBUG oslo_vmware.api [None req-3d632edf-0ed8-46ec-b4fd-7092ea7f050e tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] Task: {'id': task-5115581, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 674.016229] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65c6c6dc-b062-4e3e-abf7-76397ac94c53 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.046219] env[65726]: DEBUG oslo_vmware.api [None req-36828e27-b100-49cc-98c2-ad0b25ca6572 tempest-ServerGroupTestJSON-125182767 tempest-ServerGroupTestJSON-125182767-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5212130e-5ebf-ad80-4ec5-9c15c0af383d, 'name': SearchDatastore_Task, 'duration_secs': 0.060937} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 674.046219] env[65726]: DEBUG oslo_concurrency.lockutils [None req-36828e27-b100-49cc-98c2-ad0b25ca6572 tempest-ServerGroupTestJSON-125182767 tempest-ServerGroupTestJSON-125182767-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 674.046219] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-36828e27-b100-49cc-98c2-ad0b25ca6572 tempest-ServerGroupTestJSON-125182767 tempest-ServerGroupTestJSON-125182767-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] c2929727-a7ea-4f63-997d-474c29a305fc/c2929727-a7ea-4f63-997d-474c29a305fc.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 674.046219] env[65726]: DEBUG oslo_concurrency.lockutils [None req-fb0f14b3-fd50-4c8d-a984-877b235dcb32 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 674.046487] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-fb0f14b3-fd50-4c8d-a984-877b235dcb32 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 674.046487] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3ac4fdaa-39ac-4401-a519-91a8b9e3cf70 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.051680] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-fcaf6246-815d-464c-bddb-c759516d7554 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.063863] env[65726]: DEBUG oslo_vmware.api [None req-36828e27-b100-49cc-98c2-ad0b25ca6572 tempest-ServerGroupTestJSON-125182767 tempest-ServerGroupTestJSON-125182767-project-member] Waiting for the task: (returnval){ [ 674.063863] env[65726]: value = "task-5115582" [ 674.063863] env[65726]: _type = "Task" [ 674.063863] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 674.065561] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-fb0f14b3-fd50-4c8d-a984-877b235dcb32 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 674.065813] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-fb0f14b3-fd50-4c8d-a984-877b235dcb32 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65726) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 674.070184] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-207b93dc-f818-46e7-9e1a-5cffd65f7999 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.079281] env[65726]: DEBUG oslo_vmware.api [None req-fb0f14b3-fd50-4c8d-a984-877b235dcb32 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Waiting for the task: (returnval){ [ 674.079281] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52c1034c-1e73-2a38-e6ab-5b4c0993a83c" [ 674.079281] env[65726]: _type = "Task" [ 674.079281] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 674.084679] env[65726]: DEBUG oslo_vmware.api [None req-36828e27-b100-49cc-98c2-ad0b25ca6572 tempest-ServerGroupTestJSON-125182767 tempest-ServerGroupTestJSON-125182767-project-member] Task: {'id': task-5115582, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 674.098675] env[65726]: DEBUG oslo_vmware.api [None req-fb0f14b3-fd50-4c8d-a984-877b235dcb32 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52c1034c-1e73-2a38-e6ab-5b4c0993a83c, 'name': SearchDatastore_Task, 'duration_secs': 0.014616} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 674.099739] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bc639b37-086d-43c1-aeaa-6cd857494937 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.104529] env[65726]: DEBUG nova.compute.manager [None req-f3171f49-8536-4923-b5cf-31399a616a07 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 608cc0ea-3f6b-4b4f-83c1-01aa50999d63] Starting instance... {{(pid=65726) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 674.116376] env[65726]: DEBUG oslo_vmware.api [None req-fb0f14b3-fd50-4c8d-a984-877b235dcb32 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Waiting for the task: (returnval){ [ 674.116376] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]525ded3c-adcb-c90f-9e4d-29a14fa8a1f0" [ 674.116376] env[65726]: _type = "Task" [ 674.116376] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 674.129035] env[65726]: DEBUG oslo_vmware.api [None req-fb0f14b3-fd50-4c8d-a984-877b235dcb32 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]525ded3c-adcb-c90f-9e4d-29a14fa8a1f0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 674.237624] env[65726]: DEBUG oslo_vmware.api [None req-fd996099-26ef-437f-83c3-98cb140e9de7 tempest-TenantUsagesTestJSON-327211524 tempest-TenantUsagesTestJSON-327211524-project-member] Task: {'id': task-5115580, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.312535} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 674.237932] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-fd996099-26ef-437f-83c3-98cb140e9de7 tempest-TenantUsagesTestJSON-327211524 tempest-TenantUsagesTestJSON-327211524-project-member] Deleted the datastore file {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 674.238161] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-fd996099-26ef-437f-83c3-98cb140e9de7 tempest-TenantUsagesTestJSON-327211524 tempest-TenantUsagesTestJSON-327211524-project-member] [instance: 7748c23f-9ea9-4172-8ab7-187678272bb3] Deleted contents of the VM from datastore datastore2 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 674.238334] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-fd996099-26ef-437f-83c3-98cb140e9de7 tempest-TenantUsagesTestJSON-327211524 tempest-TenantUsagesTestJSON-327211524-project-member] [instance: 7748c23f-9ea9-4172-8ab7-187678272bb3] Instance destroyed {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 674.238498] env[65726]: INFO nova.compute.manager [None req-fd996099-26ef-437f-83c3-98cb140e9de7 tempest-TenantUsagesTestJSON-327211524 tempest-TenantUsagesTestJSON-327211524-project-member] [instance: 7748c23f-9ea9-4172-8ab7-187678272bb3] Took 1.14 seconds to destroy the instance on the hypervisor. [ 674.238748] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-fd996099-26ef-437f-83c3-98cb140e9de7 tempest-TenantUsagesTestJSON-327211524 tempest-TenantUsagesTestJSON-327211524-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 674.238995] env[65726]: DEBUG nova.compute.manager [-] [instance: 7748c23f-9ea9-4172-8ab7-187678272bb3] Deallocating network for instance {{(pid=65726) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 674.239365] env[65726]: DEBUG nova.network.neutron [-] [instance: 7748c23f-9ea9-4172-8ab7-187678272bb3] deallocate_for_instance() {{(pid=65726) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 674.239966] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 674.240255] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 674.286506] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 674.286506] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 674.300312] env[65726]: DEBUG nova.compute.manager [req-2ce2f299-88bd-421c-8733-a44c523ab631 req-c0208851-aaa3-4be0-a60b-eafce7d89eff service nova] [instance: c2929727-a7ea-4f63-997d-474c29a305fc] Received event network-vif-plugged-27cddbbc-b8ac-4cf0-8baf-d03643c3de8d {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 674.301782] env[65726]: DEBUG oslo_concurrency.lockutils [req-2ce2f299-88bd-421c-8733-a44c523ab631 req-c0208851-aaa3-4be0-a60b-eafce7d89eff service nova] Acquiring lock "c2929727-a7ea-4f63-997d-474c29a305fc-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 674.301782] env[65726]: DEBUG oslo_concurrency.lockutils [req-2ce2f299-88bd-421c-8733-a44c523ab631 req-c0208851-aaa3-4be0-a60b-eafce7d89eff service nova] Lock "c2929727-a7ea-4f63-997d-474c29a305fc-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 674.301782] env[65726]: DEBUG oslo_concurrency.lockutils [req-2ce2f299-88bd-421c-8733-a44c523ab631 req-c0208851-aaa3-4be0-a60b-eafce7d89eff service nova] Lock "c2929727-a7ea-4f63-997d-474c29a305fc-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 674.301782] env[65726]: DEBUG nova.compute.manager [req-2ce2f299-88bd-421c-8733-a44c523ab631 req-c0208851-aaa3-4be0-a60b-eafce7d89eff service nova] [instance: c2929727-a7ea-4f63-997d-474c29a305fc] No waiting events found dispatching network-vif-plugged-27cddbbc-b8ac-4cf0-8baf-d03643c3de8d {{(pid=65726) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 674.301782] env[65726]: WARNING nova.compute.manager [req-2ce2f299-88bd-421c-8733-a44c523ab631 req-c0208851-aaa3-4be0-a60b-eafce7d89eff service nova] [instance: c2929727-a7ea-4f63-997d-474c29a305fc] Received unexpected event network-vif-plugged-27cddbbc-b8ac-4cf0-8baf-d03643c3de8d for instance with vm_state building and task_state spawning. [ 674.302059] env[65726]: DEBUG nova.compute.manager [req-2ce2f299-88bd-421c-8733-a44c523ab631 req-c0208851-aaa3-4be0-a60b-eafce7d89eff service nova] [instance: c2929727-a7ea-4f63-997d-474c29a305fc] Received event network-changed-27cddbbc-b8ac-4cf0-8baf-d03643c3de8d {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 674.302059] env[65726]: DEBUG nova.compute.manager [req-2ce2f299-88bd-421c-8733-a44c523ab631 req-c0208851-aaa3-4be0-a60b-eafce7d89eff service nova] [instance: c2929727-a7ea-4f63-997d-474c29a305fc] Refreshing instance network info cache due to event network-changed-27cddbbc-b8ac-4cf0-8baf-d03643c3de8d. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 674.302059] env[65726]: DEBUG oslo_concurrency.lockutils [req-2ce2f299-88bd-421c-8733-a44c523ab631 req-c0208851-aaa3-4be0-a60b-eafce7d89eff service nova] Acquiring lock "refresh_cache-c2929727-a7ea-4f63-997d-474c29a305fc" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 674.302059] env[65726]: DEBUG oslo_concurrency.lockutils [req-2ce2f299-88bd-421c-8733-a44c523ab631 req-c0208851-aaa3-4be0-a60b-eafce7d89eff service nova] Acquired lock "refresh_cache-c2929727-a7ea-4f63-997d-474c29a305fc" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 674.302059] env[65726]: DEBUG nova.network.neutron [req-2ce2f299-88bd-421c-8733-a44c523ab631 req-c0208851-aaa3-4be0-a60b-eafce7d89eff service nova] [instance: c2929727-a7ea-4f63-997d-474c29a305fc] Refreshing network info cache for port 27cddbbc-b8ac-4cf0-8baf-d03643c3de8d {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 674.333501] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b83663a4-2303-46f3-8690-72e6256a8051 tempest-ServerDiagnosticsTest-2021442419 tempest-ServerDiagnosticsTest-2021442419-project-member] Lock "de5cad99-1dbc-4435-98b4-987e24ba8c3c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 19.249s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 674.460951] env[65726]: DEBUG oslo_concurrency.lockutils [req-889d19b5-b49d-4352-848b-37f235b066e0 req-7307af5f-8ede-44db-8223-5db2e208a88d service nova] Releasing lock "refresh_cache-5a8d4357-4b44-4a19-b1da-42d188c38adc" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 674.507644] env[65726]: DEBUG oslo_vmware.api [None req-3d632edf-0ed8-46ec-b4fd-7092ea7f050e tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] Task: {'id': task-5115581, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.098669} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 674.507920] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-3d632edf-0ed8-46ec-b4fd-7092ea7f050e tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] [instance: 3b933684-ac19-44b0-a49d-6af45501e166] Extended root virtual disk {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 674.509294] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-955ded52-3c8e-4426-bbb8-64f0ae949a05 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.536056] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-3d632edf-0ed8-46ec-b4fd-7092ea7f050e tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] [instance: 3b933684-ac19-44b0-a49d-6af45501e166] Reconfiguring VM instance instance-00000013 to attach disk [datastore1] 3b933684-ac19-44b0-a49d-6af45501e166/3b933684-ac19-44b0-a49d-6af45501e166.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 674.536428] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-948562f8-51d3-495f-9dab-a462f26a45d9 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.568247] env[65726]: DEBUG oslo_vmware.api [None req-3d632edf-0ed8-46ec-b4fd-7092ea7f050e tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] Waiting for the task: (returnval){ [ 674.568247] env[65726]: value = "task-5115583" [ 674.568247] env[65726]: _type = "Task" [ 674.568247] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 674.568247] env[65726]: INFO nova.compute.manager [None req-06f67a27-e6e0-4078-804d-cff1137abcad tempest-ServersTestBootFromVolume-762458528 tempest-ServersTestBootFromVolume-762458528-project-member] [instance: 5a8d4357-4b44-4a19-b1da-42d188c38adc] Took 27.38 seconds to build instance. [ 674.586223] env[65726]: DEBUG oslo_vmware.api [None req-36828e27-b100-49cc-98c2-ad0b25ca6572 tempest-ServerGroupTestJSON-125182767 tempest-ServerGroupTestJSON-125182767-project-member] Task: {'id': task-5115582, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 674.590012] env[65726]: DEBUG oslo_vmware.api [None req-3d632edf-0ed8-46ec-b4fd-7092ea7f050e tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] Task: {'id': task-5115583, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 674.633888] env[65726]: DEBUG oslo_vmware.api [None req-fb0f14b3-fd50-4c8d-a984-877b235dcb32 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]525ded3c-adcb-c90f-9e4d-29a14fa8a1f0, 'name': SearchDatastore_Task, 'duration_secs': 0.012585} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 674.634291] env[65726]: DEBUG oslo_concurrency.lockutils [None req-fb0f14b3-fd50-4c8d-a984-877b235dcb32 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 674.634577] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-fb0f14b3-fd50-4c8d-a984-877b235dcb32 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] d7905c4f-1ba3-43b3-b4b3-aebacd2f9cc3/d7905c4f-1ba3-43b3-b4b3-aebacd2f9cc3.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 674.636794] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-322eb065-34d1-4a28-922f-9b261b825b38 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.647278] env[65726]: DEBUG oslo_vmware.api [None req-fb0f14b3-fd50-4c8d-a984-877b235dcb32 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Waiting for the task: (returnval){ [ 674.647278] env[65726]: value = "task-5115584" [ 674.647278] env[65726]: _type = "Task" [ 674.647278] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 674.648545] env[65726]: DEBUG oslo_concurrency.lockutils [None req-f3171f49-8536-4923-b5cf-31399a616a07 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 674.659409] env[65726]: DEBUG oslo_vmware.api [None req-fb0f14b3-fd50-4c8d-a984-877b235dcb32 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Task: {'id': task-5115584, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 674.810437] env[65726]: WARNING openstack [req-2ce2f299-88bd-421c-8733-a44c523ab631 req-c0208851-aaa3-4be0-a60b-eafce7d89eff service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 674.810922] env[65726]: WARNING openstack [req-2ce2f299-88bd-421c-8733-a44c523ab631 req-c0208851-aaa3-4be0-a60b-eafce7d89eff service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 675.012479] env[65726]: WARNING openstack [req-2ce2f299-88bd-421c-8733-a44c523ab631 req-c0208851-aaa3-4be0-a60b-eafce7d89eff service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 675.012889] env[65726]: WARNING openstack [req-2ce2f299-88bd-421c-8733-a44c523ab631 req-c0208851-aaa3-4be0-a60b-eafce7d89eff service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 675.083154] env[65726]: DEBUG oslo_concurrency.lockutils [None req-06f67a27-e6e0-4078-804d-cff1137abcad tempest-ServersTestBootFromVolume-762458528 tempest-ServersTestBootFromVolume-762458528-project-member] Lock "5a8d4357-4b44-4a19-b1da-42d188c38adc" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 56.018s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 675.083154] env[65726]: DEBUG nova.network.neutron [-] [instance: 7748c23f-9ea9-4172-8ab7-187678272bb3] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 675.100467] env[65726]: DEBUG oslo_vmware.api [None req-36828e27-b100-49cc-98c2-ad0b25ca6572 tempest-ServerGroupTestJSON-125182767 tempest-ServerGroupTestJSON-125182767-project-member] Task: {'id': task-5115582, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.892802} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 675.100467] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-36828e27-b100-49cc-98c2-ad0b25ca6572 tempest-ServerGroupTestJSON-125182767 tempest-ServerGroupTestJSON-125182767-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] c2929727-a7ea-4f63-997d-474c29a305fc/c2929727-a7ea-4f63-997d-474c29a305fc.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 675.100795] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-36828e27-b100-49cc-98c2-ad0b25ca6572 tempest-ServerGroupTestJSON-125182767 tempest-ServerGroupTestJSON-125182767-project-member] [instance: c2929727-a7ea-4f63-997d-474c29a305fc] Extending root virtual disk to 1048576 {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 675.101204] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f2ec16f7-090b-4005-a566-38872daddc09 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.109304] env[65726]: DEBUG oslo_vmware.api [None req-3d632edf-0ed8-46ec-b4fd-7092ea7f050e tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] Task: {'id': task-5115583, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 675.116293] env[65726]: DEBUG oslo_vmware.api [None req-36828e27-b100-49cc-98c2-ad0b25ca6572 tempest-ServerGroupTestJSON-125182767 tempest-ServerGroupTestJSON-125182767-project-member] Waiting for the task: (returnval){ [ 675.116293] env[65726]: value = "task-5115585" [ 675.116293] env[65726]: _type = "Task" [ 675.116293] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 675.132862] env[65726]: DEBUG oslo_vmware.api [None req-36828e27-b100-49cc-98c2-ad0b25ca6572 tempest-ServerGroupTestJSON-125182767 tempest-ServerGroupTestJSON-125182767-project-member] Task: {'id': task-5115585, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 675.141410] env[65726]: WARNING openstack [req-2ce2f299-88bd-421c-8733-a44c523ab631 req-c0208851-aaa3-4be0-a60b-eafce7d89eff service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 675.141877] env[65726]: WARNING openstack [req-2ce2f299-88bd-421c-8733-a44c523ab631 req-c0208851-aaa3-4be0-a60b-eafce7d89eff service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 675.171215] env[65726]: DEBUG oslo_vmware.api [None req-fb0f14b3-fd50-4c8d-a984-877b235dcb32 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Task: {'id': task-5115584, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 675.390347] env[65726]: DEBUG nova.network.neutron [req-2ce2f299-88bd-421c-8733-a44c523ab631 req-c0208851-aaa3-4be0-a60b-eafce7d89eff service nova] [instance: c2929727-a7ea-4f63-997d-474c29a305fc] Updated VIF entry in instance network info cache for port 27cddbbc-b8ac-4cf0-8baf-d03643c3de8d. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 675.390741] env[65726]: DEBUG nova.network.neutron [req-2ce2f299-88bd-421c-8733-a44c523ab631 req-c0208851-aaa3-4be0-a60b-eafce7d89eff service nova] [instance: c2929727-a7ea-4f63-997d-474c29a305fc] Updating instance_info_cache with network_info: [{"id": "27cddbbc-b8ac-4cf0-8baf-d03643c3de8d", "address": "fa:16:3e:53:19:47", "network": {"id": "1401b82d-0e46-47ea-bf6d-ed2a2e5f7f24", "bridge": "br-int", "label": "tempest-ServerGroupTestJSON-933777645-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "34534927100c4f7a91a1073de9047674", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "40859343-2baa-45fd-88e3-ebf8aaed2b19", "external-id": "nsx-vlan-transportzone-10", "segmentation_id": 10, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap27cddbbc-b8", "ovs_interfaceid": "27cddbbc-b8ac-4cf0-8baf-d03643c3de8d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 675.536255] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79419da0-b5d4-4abe-9a6c-aca6e02c05ef {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.545277] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59320f7d-d3ea-4a9f-82c3-3bf1e53e8579 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.586450] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2830e466-7627-438d-94e3-be3427062896 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.590972] env[65726]: INFO nova.compute.manager [-] [instance: 7748c23f-9ea9-4172-8ab7-187678272bb3] Took 1.35 seconds to deallocate network for instance. [ 675.591442] env[65726]: DEBUG nova.compute.manager [None req-296fd45e-ee7a-4e44-912a-f3b5a22cbbf0 tempest-ServerExternalEventsTest-1845125219 tempest-ServerExternalEventsTest-1845125219-project-member] [instance: 96acc5f8-b411-408e-929d-d1035cfd50db] Starting instance... {{(pid=65726) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 675.601098] env[65726]: DEBUG oslo_vmware.api [None req-3d632edf-0ed8-46ec-b4fd-7092ea7f050e tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] Task: {'id': task-5115583, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 675.604965] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa5a6fda-15ec-4299-addc-843f91c8e4fa {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.621793] env[65726]: DEBUG nova.compute.provider_tree [None req-efa812b8-b008-4879-b549-6ccf454f374f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 675.636047] env[65726]: DEBUG oslo_vmware.api [None req-36828e27-b100-49cc-98c2-ad0b25ca6572 tempest-ServerGroupTestJSON-125182767 tempest-ServerGroupTestJSON-125182767-project-member] Task: {'id': task-5115585, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.112736} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 675.636047] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-36828e27-b100-49cc-98c2-ad0b25ca6572 tempest-ServerGroupTestJSON-125182767 tempest-ServerGroupTestJSON-125182767-project-member] [instance: c2929727-a7ea-4f63-997d-474c29a305fc] Extended root virtual disk {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 675.636047] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fef20220-b5bf-4ea7-91c5-93b04a5246b3 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.658967] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-36828e27-b100-49cc-98c2-ad0b25ca6572 tempest-ServerGroupTestJSON-125182767 tempest-ServerGroupTestJSON-125182767-project-member] [instance: c2929727-a7ea-4f63-997d-474c29a305fc] Reconfiguring VM instance instance-00000017 to attach disk [datastore1] c2929727-a7ea-4f63-997d-474c29a305fc/c2929727-a7ea-4f63-997d-474c29a305fc.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 675.663462] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7a822081-7417-410f-88e9-b34f82cd8319 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.686051] env[65726]: DEBUG oslo_vmware.api [None req-fb0f14b3-fd50-4c8d-a984-877b235dcb32 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Task: {'id': task-5115584, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.878762} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 675.687555] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-fb0f14b3-fd50-4c8d-a984-877b235dcb32 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] d7905c4f-1ba3-43b3-b4b3-aebacd2f9cc3/d7905c4f-1ba3-43b3-b4b3-aebacd2f9cc3.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 675.687756] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-fb0f14b3-fd50-4c8d-a984-877b235dcb32 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] [instance: d7905c4f-1ba3-43b3-b4b3-aebacd2f9cc3] Extending root virtual disk to 1048576 {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 675.688083] env[65726]: DEBUG oslo_vmware.api [None req-36828e27-b100-49cc-98c2-ad0b25ca6572 tempest-ServerGroupTestJSON-125182767 tempest-ServerGroupTestJSON-125182767-project-member] Waiting for the task: (returnval){ [ 675.688083] env[65726]: value = "task-5115586" [ 675.688083] env[65726]: _type = "Task" [ 675.688083] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 675.688274] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-8bd06a9e-2587-4f6d-895b-b32ceb39fb21 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.699797] env[65726]: DEBUG oslo_vmware.api [None req-36828e27-b100-49cc-98c2-ad0b25ca6572 tempest-ServerGroupTestJSON-125182767 tempest-ServerGroupTestJSON-125182767-project-member] Task: {'id': task-5115586, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 675.701577] env[65726]: DEBUG oslo_vmware.api [None req-fb0f14b3-fd50-4c8d-a984-877b235dcb32 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Waiting for the task: (returnval){ [ 675.701577] env[65726]: value = "task-5115587" [ 675.701577] env[65726]: _type = "Task" [ 675.701577] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 675.714787] env[65726]: DEBUG nova.compute.manager [req-5aed8986-c9fa-4c14-8370-f4385ef86371 req-377628b7-9e48-4829-bcbc-48b106d7ad04 service nova] [instance: d7905c4f-1ba3-43b3-b4b3-aebacd2f9cc3] Received event network-changed-ee24bb04-7df3-4d06-9751-289b4bdeb753 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 675.715229] env[65726]: DEBUG nova.compute.manager [req-5aed8986-c9fa-4c14-8370-f4385ef86371 req-377628b7-9e48-4829-bcbc-48b106d7ad04 service nova] [instance: d7905c4f-1ba3-43b3-b4b3-aebacd2f9cc3] Refreshing instance network info cache due to event network-changed-ee24bb04-7df3-4d06-9751-289b4bdeb753. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 675.715340] env[65726]: DEBUG oslo_concurrency.lockutils [req-5aed8986-c9fa-4c14-8370-f4385ef86371 req-377628b7-9e48-4829-bcbc-48b106d7ad04 service nova] Acquiring lock "refresh_cache-d7905c4f-1ba3-43b3-b4b3-aebacd2f9cc3" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 675.715506] env[65726]: DEBUG oslo_concurrency.lockutils [req-5aed8986-c9fa-4c14-8370-f4385ef86371 req-377628b7-9e48-4829-bcbc-48b106d7ad04 service nova] Acquired lock "refresh_cache-d7905c4f-1ba3-43b3-b4b3-aebacd2f9cc3" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 675.715696] env[65726]: DEBUG nova.network.neutron [req-5aed8986-c9fa-4c14-8370-f4385ef86371 req-377628b7-9e48-4829-bcbc-48b106d7ad04 service nova] [instance: d7905c4f-1ba3-43b3-b4b3-aebacd2f9cc3] Refreshing network info cache for port ee24bb04-7df3-4d06-9751-289b4bdeb753 {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 675.896023] env[65726]: DEBUG oslo_concurrency.lockutils [req-2ce2f299-88bd-421c-8733-a44c523ab631 req-c0208851-aaa3-4be0-a60b-eafce7d89eff service nova] Releasing lock "refresh_cache-c2929727-a7ea-4f63-997d-474c29a305fc" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 675.896023] env[65726]: DEBUG nova.compute.manager [req-2ce2f299-88bd-421c-8733-a44c523ab631 req-c0208851-aaa3-4be0-a60b-eafce7d89eff service nova] [instance: 811c1bbd-259c-4f5e-9a61-581bfba0ec4d] Received event network-vif-deleted-205b7b13-8d99-402c-9587-c84362a7af94 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 676.100693] env[65726]: DEBUG oslo_vmware.api [None req-3d632edf-0ed8-46ec-b4fd-7092ea7f050e tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] Task: {'id': task-5115583, 'name': ReconfigVM_Task, 'duration_secs': 1.221717} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 676.106029] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-3d632edf-0ed8-46ec-b4fd-7092ea7f050e tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] [instance: 3b933684-ac19-44b0-a49d-6af45501e166] Reconfigured VM instance instance-00000013 to attach disk [datastore1] 3b933684-ac19-44b0-a49d-6af45501e166/3b933684-ac19-44b0-a49d-6af45501e166.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 676.106029] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-81d4f2a3-80b5-4f17-a5a7-eb7700f8538d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.112837] env[65726]: DEBUG oslo_concurrency.lockutils [None req-fd996099-26ef-437f-83c3-98cb140e9de7 tempest-TenantUsagesTestJSON-327211524 tempest-TenantUsagesTestJSON-327211524-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 676.120924] env[65726]: DEBUG oslo_vmware.api [None req-3d632edf-0ed8-46ec-b4fd-7092ea7f050e tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] Waiting for the task: (returnval){ [ 676.120924] env[65726]: value = "task-5115588" [ 676.120924] env[65726]: _type = "Task" [ 676.120924] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 676.131454] env[65726]: DEBUG nova.scheduler.client.report [None req-efa812b8-b008-4879-b549-6ccf454f374f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 676.138297] env[65726]: DEBUG oslo_vmware.api [None req-3d632edf-0ed8-46ec-b4fd-7092ea7f050e tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] Task: {'id': task-5115588, 'name': Rename_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 676.139328] env[65726]: DEBUG oslo_concurrency.lockutils [None req-296fd45e-ee7a-4e44-912a-f3b5a22cbbf0 tempest-ServerExternalEventsTest-1845125219 tempest-ServerExternalEventsTest-1845125219-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 676.201740] env[65726]: DEBUG oslo_vmware.api [None req-36828e27-b100-49cc-98c2-ad0b25ca6572 tempest-ServerGroupTestJSON-125182767 tempest-ServerGroupTestJSON-125182767-project-member] Task: {'id': task-5115586, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 676.210378] env[65726]: DEBUG oslo_vmware.api [None req-fb0f14b3-fd50-4c8d-a984-877b235dcb32 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Task: {'id': task-5115587, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.09895} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 676.210763] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-fb0f14b3-fd50-4c8d-a984-877b235dcb32 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] [instance: d7905c4f-1ba3-43b3-b4b3-aebacd2f9cc3] Extended root virtual disk {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 676.211678] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3253c66a-5c07-4a05-9f8e-4b3a38515c25 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.227384] env[65726]: WARNING openstack [req-5aed8986-c9fa-4c14-8370-f4385ef86371 req-377628b7-9e48-4829-bcbc-48b106d7ad04 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 676.227761] env[65726]: WARNING openstack [req-5aed8986-c9fa-4c14-8370-f4385ef86371 req-377628b7-9e48-4829-bcbc-48b106d7ad04 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 676.245116] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-fb0f14b3-fd50-4c8d-a984-877b235dcb32 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] [instance: d7905c4f-1ba3-43b3-b4b3-aebacd2f9cc3] Reconfiguring VM instance instance-00000018 to attach disk [datastore1] d7905c4f-1ba3-43b3-b4b3-aebacd2f9cc3/d7905c4f-1ba3-43b3-b4b3-aebacd2f9cc3.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 676.246310] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6121695d-3fdf-4bb9-9083-e95578f4a576 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.269407] env[65726]: DEBUG oslo_vmware.api [None req-fb0f14b3-fd50-4c8d-a984-877b235dcb32 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Waiting for the task: (returnval){ [ 676.269407] env[65726]: value = "task-5115589" [ 676.269407] env[65726]: _type = "Task" [ 676.269407] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 676.278618] env[65726]: DEBUG oslo_vmware.api [None req-fb0f14b3-fd50-4c8d-a984-877b235dcb32 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Task: {'id': task-5115589, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 676.634224] env[65726]: DEBUG oslo_vmware.api [None req-3d632edf-0ed8-46ec-b4fd-7092ea7f050e tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] Task: {'id': task-5115588, 'name': Rename_Task, 'duration_secs': 0.328158} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 676.634353] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-3d632edf-0ed8-46ec-b4fd-7092ea7f050e tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] [instance: 3b933684-ac19-44b0-a49d-6af45501e166] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 676.634528] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3fa32654-4ddb-4d05-b840-d8d888c632b3 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.643415] env[65726]: DEBUG oslo_concurrency.lockutils [None req-efa812b8-b008-4879-b549-6ccf454f374f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.856s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 676.643924] env[65726]: DEBUG nova.compute.manager [None req-efa812b8-b008-4879-b549-6ccf454f374f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] [instance: ddd36db0-28ad-4582-994b-7eef8b08ed46] Start building networks asynchronously for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 676.647519] env[65726]: DEBUG oslo_vmware.api [None req-3d632edf-0ed8-46ec-b4fd-7092ea7f050e tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] Waiting for the task: (returnval){ [ 676.647519] env[65726]: value = "task-5115590" [ 676.647519] env[65726]: _type = "Task" [ 676.647519] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 676.651129] env[65726]: DEBUG oslo_concurrency.lockutils [None req-bb345b26-9f49-4e52-904f-99ccc71299ac tempest-AttachInterfacesV270Test-133387050 tempest-AttachInterfacesV270Test-133387050-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.519s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 676.651129] env[65726]: INFO nova.compute.claims [None req-bb345b26-9f49-4e52-904f-99ccc71299ac tempest-AttachInterfacesV270Test-133387050 tempest-AttachInterfacesV270Test-133387050-project-member] [instance: c977257e-dce0-4ecc-b42b-0f6ebab38797] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 676.661790] env[65726]: DEBUG oslo_vmware.api [None req-3d632edf-0ed8-46ec-b4fd-7092ea7f050e tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] Task: {'id': task-5115590, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 676.703255] env[65726]: DEBUG oslo_vmware.api [None req-36828e27-b100-49cc-98c2-ad0b25ca6572 tempest-ServerGroupTestJSON-125182767 tempest-ServerGroupTestJSON-125182767-project-member] Task: {'id': task-5115586, 'name': ReconfigVM_Task, 'duration_secs': 0.808696} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 676.703582] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-36828e27-b100-49cc-98c2-ad0b25ca6572 tempest-ServerGroupTestJSON-125182767 tempest-ServerGroupTestJSON-125182767-project-member] [instance: c2929727-a7ea-4f63-997d-474c29a305fc] Reconfigured VM instance instance-00000017 to attach disk [datastore1] c2929727-a7ea-4f63-997d-474c29a305fc/c2929727-a7ea-4f63-997d-474c29a305fc.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 676.704244] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c5fe4e24-74ae-456c-bb26-1ad2e856f239 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.708039] env[65726]: WARNING openstack [req-5aed8986-c9fa-4c14-8370-f4385ef86371 req-377628b7-9e48-4829-bcbc-48b106d7ad04 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 676.708443] env[65726]: WARNING openstack [req-5aed8986-c9fa-4c14-8370-f4385ef86371 req-377628b7-9e48-4829-bcbc-48b106d7ad04 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 676.724483] env[65726]: DEBUG oslo_vmware.api [None req-36828e27-b100-49cc-98c2-ad0b25ca6572 tempest-ServerGroupTestJSON-125182767 tempest-ServerGroupTestJSON-125182767-project-member] Waiting for the task: (returnval){ [ 676.724483] env[65726]: value = "task-5115591" [ 676.724483] env[65726]: _type = "Task" [ 676.724483] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 676.738495] env[65726]: DEBUG oslo_vmware.api [None req-36828e27-b100-49cc-98c2-ad0b25ca6572 tempest-ServerGroupTestJSON-125182767 tempest-ServerGroupTestJSON-125182767-project-member] Task: {'id': task-5115591, 'name': Rename_Task} progress is 6%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 676.772102] env[65726]: WARNING openstack [req-5aed8986-c9fa-4c14-8370-f4385ef86371 req-377628b7-9e48-4829-bcbc-48b106d7ad04 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 676.772102] env[65726]: WARNING openstack [req-5aed8986-c9fa-4c14-8370-f4385ef86371 req-377628b7-9e48-4829-bcbc-48b106d7ad04 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 676.797988] env[65726]: DEBUG oslo_vmware.api [None req-fb0f14b3-fd50-4c8d-a984-877b235dcb32 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Task: {'id': task-5115589, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 676.924322] env[65726]: DEBUG nova.network.neutron [req-5aed8986-c9fa-4c14-8370-f4385ef86371 req-377628b7-9e48-4829-bcbc-48b106d7ad04 service nova] [instance: d7905c4f-1ba3-43b3-b4b3-aebacd2f9cc3] Updated VIF entry in instance network info cache for port ee24bb04-7df3-4d06-9751-289b4bdeb753. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 676.924322] env[65726]: DEBUG nova.network.neutron [req-5aed8986-c9fa-4c14-8370-f4385ef86371 req-377628b7-9e48-4829-bcbc-48b106d7ad04 service nova] [instance: d7905c4f-1ba3-43b3-b4b3-aebacd2f9cc3] Updating instance_info_cache with network_info: [{"id": "ee24bb04-7df3-4d06-9751-289b4bdeb753", "address": "fa:16:3e:82:91:65", "network": {"id": "27801f9a-9912-41cf-be65-bc9903f4952f", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-781436951-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "aac5fe6c99f740ada3747088b2b0ad89", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dd7d0d95-6848-4e69-ac21-75f8db82a3b5", "external-id": "nsx-vlan-transportzone-272", "segmentation_id": 272, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapee24bb04-7d", "ovs_interfaceid": "ee24bb04-7df3-4d06-9751-289b4bdeb753", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 677.154013] env[65726]: DEBUG nova.compute.utils [None req-efa812b8-b008-4879-b549-6ccf454f374f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] Using /dev/sd instead of None {{(pid=65726) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 677.162760] env[65726]: DEBUG nova.compute.manager [None req-efa812b8-b008-4879-b549-6ccf454f374f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] [instance: ddd36db0-28ad-4582-994b-7eef8b08ed46] Not allocating networking since 'none' was specified. {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2015}} [ 677.174316] env[65726]: DEBUG oslo_vmware.api [None req-3d632edf-0ed8-46ec-b4fd-7092ea7f050e tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] Task: {'id': task-5115590, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 677.235024] env[65726]: DEBUG oslo_vmware.api [None req-36828e27-b100-49cc-98c2-ad0b25ca6572 tempest-ServerGroupTestJSON-125182767 tempest-ServerGroupTestJSON-125182767-project-member] Task: {'id': task-5115591, 'name': Rename_Task, 'duration_secs': 0.261735} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 677.235238] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-36828e27-b100-49cc-98c2-ad0b25ca6572 tempest-ServerGroupTestJSON-125182767 tempest-ServerGroupTestJSON-125182767-project-member] [instance: c2929727-a7ea-4f63-997d-474c29a305fc] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 677.235534] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-82ba03b4-8cf2-4ca9-8287-f7e7fef4e6f6 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.248211] env[65726]: DEBUG oslo_vmware.api [None req-36828e27-b100-49cc-98c2-ad0b25ca6572 tempest-ServerGroupTestJSON-125182767 tempest-ServerGroupTestJSON-125182767-project-member] Waiting for the task: (returnval){ [ 677.248211] env[65726]: value = "task-5115592" [ 677.248211] env[65726]: _type = "Task" [ 677.248211] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 677.255110] env[65726]: DEBUG oslo_vmware.api [None req-36828e27-b100-49cc-98c2-ad0b25ca6572 tempest-ServerGroupTestJSON-125182767 tempest-ServerGroupTestJSON-125182767-project-member] Task: {'id': task-5115592, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 677.281316] env[65726]: DEBUG oslo_vmware.api [None req-fb0f14b3-fd50-4c8d-a984-877b235dcb32 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Task: {'id': task-5115589, 'name': ReconfigVM_Task, 'duration_secs': 0.702528} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 677.281631] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-fb0f14b3-fd50-4c8d-a984-877b235dcb32 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] [instance: d7905c4f-1ba3-43b3-b4b3-aebacd2f9cc3] Reconfigured VM instance instance-00000018 to attach disk [datastore1] d7905c4f-1ba3-43b3-b4b3-aebacd2f9cc3/d7905c4f-1ba3-43b3-b4b3-aebacd2f9cc3.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 677.282431] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e0cc6447-4b51-46af-b070-316a97e9dcbc {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.290189] env[65726]: DEBUG oslo_vmware.api [None req-fb0f14b3-fd50-4c8d-a984-877b235dcb32 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Waiting for the task: (returnval){ [ 677.290189] env[65726]: value = "task-5115593" [ 677.290189] env[65726]: _type = "Task" [ 677.290189] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 677.301893] env[65726]: DEBUG oslo_vmware.api [None req-fb0f14b3-fd50-4c8d-a984-877b235dcb32 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Task: {'id': task-5115593, 'name': Rename_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 677.428706] env[65726]: DEBUG oslo_concurrency.lockutils [req-5aed8986-c9fa-4c14-8370-f4385ef86371 req-377628b7-9e48-4829-bcbc-48b106d7ad04 service nova] Releasing lock "refresh_cache-d7905c4f-1ba3-43b3-b4b3-aebacd2f9cc3" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 677.667168] env[65726]: DEBUG nova.compute.manager [None req-efa812b8-b008-4879-b549-6ccf454f374f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] [instance: ddd36db0-28ad-4582-994b-7eef8b08ed46] Start building block device mappings for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 677.670231] env[65726]: DEBUG oslo_vmware.api [None req-3d632edf-0ed8-46ec-b4fd-7092ea7f050e tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] Task: {'id': task-5115590, 'name': PowerOnVM_Task, 'duration_secs': 0.870004} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 677.674251] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-3d632edf-0ed8-46ec-b4fd-7092ea7f050e tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] [instance: 3b933684-ac19-44b0-a49d-6af45501e166] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 677.674454] env[65726]: DEBUG nova.compute.manager [None req-3d632edf-0ed8-46ec-b4fd-7092ea7f050e tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] [instance: 3b933684-ac19-44b0-a49d-6af45501e166] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 677.675783] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1864e619-d0cb-4e89-9b37-3267bd3186ae {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.738646] env[65726]: DEBUG oslo_concurrency.lockutils [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Acquiring lock "02351120-8385-4403-8464-a154f3da9380" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 677.739527] env[65726]: DEBUG oslo_concurrency.lockutils [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Lock "02351120-8385-4403-8464-a154f3da9380" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 677.762211] env[65726]: DEBUG oslo_vmware.api [None req-36828e27-b100-49cc-98c2-ad0b25ca6572 tempest-ServerGroupTestJSON-125182767 tempest-ServerGroupTestJSON-125182767-project-member] Task: {'id': task-5115592, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 677.808901] env[65726]: DEBUG oslo_concurrency.lockutils [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Acquiring lock "ca67d1a4-d304-4d06-a436-f257b60ca080" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 677.809347] env[65726]: DEBUG oslo_concurrency.lockutils [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Lock "ca67d1a4-d304-4d06-a436-f257b60ca080" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 677.817108] env[65726]: DEBUG oslo_vmware.api [None req-fb0f14b3-fd50-4c8d-a984-877b235dcb32 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Task: {'id': task-5115593, 'name': Rename_Task, 'duration_secs': 0.210443} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 677.817108] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-fb0f14b3-fd50-4c8d-a984-877b235dcb32 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] [instance: d7905c4f-1ba3-43b3-b4b3-aebacd2f9cc3] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 677.817108] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ec3346fd-4a07-405e-bd21-fcae286d9253 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.823711] env[65726]: DEBUG oslo_vmware.api [None req-fb0f14b3-fd50-4c8d-a984-877b235dcb32 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Waiting for the task: (returnval){ [ 677.823711] env[65726]: value = "task-5115594" [ 677.823711] env[65726]: _type = "Task" [ 677.823711] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 677.839548] env[65726]: DEBUG oslo_vmware.api [None req-fb0f14b3-fd50-4c8d-a984-877b235dcb32 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Task: {'id': task-5115594, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 678.143611] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b0adde83-84f2-4fa3-84bf-8f387eee1293 tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] Acquiring lock "3827eed1-b7a1-4922-a4de-f0c038cb8566" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 678.144152] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b0adde83-84f2-4fa3-84bf-8f387eee1293 tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] Lock "3827eed1-b7a1-4922-a4de-f0c038cb8566" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 678.213175] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3d632edf-0ed8-46ec-b4fd-7092ea7f050e tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 678.266539] env[65726]: DEBUG oslo_vmware.api [None req-36828e27-b100-49cc-98c2-ad0b25ca6572 tempest-ServerGroupTestJSON-125182767 tempest-ServerGroupTestJSON-125182767-project-member] Task: {'id': task-5115592, 'name': PowerOnVM_Task, 'duration_secs': 0.66966} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 678.266796] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-36828e27-b100-49cc-98c2-ad0b25ca6572 tempest-ServerGroupTestJSON-125182767 tempest-ServerGroupTestJSON-125182767-project-member] [instance: c2929727-a7ea-4f63-997d-474c29a305fc] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 678.269873] env[65726]: INFO nova.compute.manager [None req-36828e27-b100-49cc-98c2-ad0b25ca6572 tempest-ServerGroupTestJSON-125182767 tempest-ServerGroupTestJSON-125182767-project-member] [instance: c2929727-a7ea-4f63-997d-474c29a305fc] Took 10.71 seconds to spawn the instance on the hypervisor. [ 678.269873] env[65726]: DEBUG nova.compute.manager [None req-36828e27-b100-49cc-98c2-ad0b25ca6572 tempest-ServerGroupTestJSON-125182767 tempest-ServerGroupTestJSON-125182767-project-member] [instance: c2929727-a7ea-4f63-997d-474c29a305fc] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 678.269873] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c66c1b9e-420f-4e52-abbb-1bdd5cffc22b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.341777] env[65726]: DEBUG oslo_vmware.api [None req-fb0f14b3-fd50-4c8d-a984-877b235dcb32 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Task: {'id': task-5115594, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 678.433183] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a855f66e-90ca-414e-bf72-381766c5402c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.444374] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab2e25d5-1483-4173-86cf-5d5f972dc94a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.478378] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8143d628-04e5-47bc-9963-8d88dcda44f9 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.487564] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23820638-c414-47d6-aa19-63d682968ab1 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.503055] env[65726]: DEBUG nova.compute.provider_tree [None req-bb345b26-9f49-4e52-904f-99ccc71299ac tempest-AttachInterfacesV270Test-133387050 tempest-AttachInterfacesV270Test-133387050-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 678.676683] env[65726]: DEBUG nova.compute.manager [None req-efa812b8-b008-4879-b549-6ccf454f374f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] [instance: ddd36db0-28ad-4582-994b-7eef8b08ed46] Start spawning the instance on the hypervisor. {{(pid=65726) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 678.711027] env[65726]: DEBUG nova.virt.hardware [None req-efa812b8-b008-4879-b549-6ccf454f374f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T19:28:34Z,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1539f06fa3534e90acbc3a60c4b8bd3f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T19:28:35Z,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 678.712360] env[65726]: DEBUG nova.virt.hardware [None req-efa812b8-b008-4879-b549-6ccf454f374f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 678.712360] env[65726]: DEBUG nova.virt.hardware [None req-efa812b8-b008-4879-b549-6ccf454f374f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 678.712360] env[65726]: DEBUG nova.virt.hardware [None req-efa812b8-b008-4879-b549-6ccf454f374f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 678.712360] env[65726]: DEBUG nova.virt.hardware [None req-efa812b8-b008-4879-b549-6ccf454f374f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 678.712360] env[65726]: DEBUG nova.virt.hardware [None req-efa812b8-b008-4879-b549-6ccf454f374f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 678.712913] env[65726]: DEBUG nova.virt.hardware [None req-efa812b8-b008-4879-b549-6ccf454f374f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 678.713459] env[65726]: DEBUG nova.virt.hardware [None req-efa812b8-b008-4879-b549-6ccf454f374f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 678.713714] env[65726]: DEBUG nova.virt.hardware [None req-efa812b8-b008-4879-b549-6ccf454f374f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 678.715454] env[65726]: DEBUG nova.virt.hardware [None req-efa812b8-b008-4879-b549-6ccf454f374f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 678.715615] env[65726]: DEBUG nova.virt.hardware [None req-efa812b8-b008-4879-b549-6ccf454f374f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 678.716570] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ab37007-015d-4157-b41b-c8c3b1fa3923 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.728474] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f86a9a1c-caa4-4c72-8bb1-2caeeae5d218 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.751058] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-efa812b8-b008-4879-b549-6ccf454f374f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] [instance: ddd36db0-28ad-4582-994b-7eef8b08ed46] Instance VIF info [] {{(pid=65726) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 678.755220] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-efa812b8-b008-4879-b549-6ccf454f374f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] Creating folder: Project (3a57ef04f09c44358dcaeccd41aee0f3). Parent ref: group-v995008. {{(pid=65726) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 678.755220] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5ad9f855-850d-47c5-bf69-e4cb43cbaa51 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.771586] env[65726]: INFO nova.virt.vmwareapi.vm_util [None req-efa812b8-b008-4879-b549-6ccf454f374f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] Created folder: Project (3a57ef04f09c44358dcaeccd41aee0f3) in parent group-v995008. [ 678.771800] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-efa812b8-b008-4879-b549-6ccf454f374f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] Creating folder: Instances. Parent ref: group-v995075. {{(pid=65726) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 678.772132] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ff0edcfc-c740-4a1d-bfa5-339f51900547 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.787810] env[65726]: INFO nova.virt.vmwareapi.vm_util [None req-efa812b8-b008-4879-b549-6ccf454f374f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] Created folder: Instances in parent group-v995075. [ 678.788250] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-efa812b8-b008-4879-b549-6ccf454f374f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 678.791253] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ddd36db0-28ad-4582-994b-7eef8b08ed46] Creating VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 678.791572] env[65726]: INFO nova.compute.manager [None req-36828e27-b100-49cc-98c2-ad0b25ca6572 tempest-ServerGroupTestJSON-125182767 tempest-ServerGroupTestJSON-125182767-project-member] [instance: c2929727-a7ea-4f63-997d-474c29a305fc] Took 30.22 seconds to build instance. [ 678.793042] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5ba7dea0-6e65-4b1d-852c-eb2eb0da007c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.817623] env[65726]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 678.817623] env[65726]: value = "task-5115597" [ 678.817623] env[65726]: _type = "Task" [ 678.817623] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 678.835401] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115597, 'name': CreateVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 678.841788] env[65726]: DEBUG oslo_vmware.api [None req-fb0f14b3-fd50-4c8d-a984-877b235dcb32 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Task: {'id': task-5115594, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 679.006033] env[65726]: DEBUG nova.scheduler.client.report [None req-bb345b26-9f49-4e52-904f-99ccc71299ac tempest-AttachInterfacesV270Test-133387050 tempest-AttachInterfacesV270Test-133387050-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 679.308319] env[65726]: DEBUG oslo_concurrency.lockutils [None req-36828e27-b100-49cc-98c2-ad0b25ca6572 tempest-ServerGroupTestJSON-125182767 tempest-ServerGroupTestJSON-125182767-project-member] Lock "c2929727-a7ea-4f63-997d-474c29a305fc" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 47.656s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 679.333929] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115597, 'name': CreateVM_Task, 'duration_secs': 0.471444} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 679.338300] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ddd36db0-28ad-4582-994b-7eef8b08ed46] Created VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 679.339737] env[65726]: DEBUG oslo_concurrency.lockutils [None req-efa812b8-b008-4879-b549-6ccf454f374f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 679.339737] env[65726]: DEBUG oslo_concurrency.lockutils [None req-efa812b8-b008-4879-b549-6ccf454f374f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 679.340571] env[65726]: DEBUG oslo_concurrency.lockutils [None req-efa812b8-b008-4879-b549-6ccf454f374f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 679.341522] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8d8f8e91-6b54-4cce-ad46-d7ad4299fe10 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.347618] env[65726]: DEBUG oslo_vmware.api [None req-fb0f14b3-fd50-4c8d-a984-877b235dcb32 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Task: {'id': task-5115594, 'name': PowerOnVM_Task, 'duration_secs': 1.074745} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 679.348313] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-fb0f14b3-fd50-4c8d-a984-877b235dcb32 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] [instance: d7905c4f-1ba3-43b3-b4b3-aebacd2f9cc3] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 679.348596] env[65726]: INFO nova.compute.manager [None req-fb0f14b3-fd50-4c8d-a984-877b235dcb32 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] [instance: d7905c4f-1ba3-43b3-b4b3-aebacd2f9cc3] Took 10.53 seconds to spawn the instance on the hypervisor. [ 679.348896] env[65726]: DEBUG nova.compute.manager [None req-fb0f14b3-fd50-4c8d-a984-877b235dcb32 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] [instance: d7905c4f-1ba3-43b3-b4b3-aebacd2f9cc3] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 679.349988] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-304d3ae5-9945-4166-91be-7a7e5ad7505c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.354902] env[65726]: DEBUG oslo_vmware.api [None req-efa812b8-b008-4879-b549-6ccf454f374f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] Waiting for the task: (returnval){ [ 679.354902] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52a31a85-0ede-ff3d-eeb6-2dd8105e979f" [ 679.354902] env[65726]: _type = "Task" [ 679.354902] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 679.368846] env[65726]: DEBUG oslo_vmware.api [None req-efa812b8-b008-4879-b549-6ccf454f374f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52a31a85-0ede-ff3d-eeb6-2dd8105e979f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 679.399734] env[65726]: DEBUG nova.compute.manager [req-7d7ed494-2504-47ff-8644-27f0025b83e9 req-7709672e-61a6-41f7-8269-9b76b8f822db service nova] [instance: 7748c23f-9ea9-4172-8ab7-187678272bb3] Received event network-vif-deleted-8bb2d672-e56e-4f01-bfbf-74a279d9da7d {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 679.400038] env[65726]: DEBUG nova.compute.manager [req-7d7ed494-2504-47ff-8644-27f0025b83e9 req-7709672e-61a6-41f7-8269-9b76b8f822db service nova] [instance: fd481728-3ef0-4a96-affd-ab2dd3f596bb] Received event network-changed-7df3f02b-e4e5-4840-bf9f-3498771b9184 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 679.400208] env[65726]: DEBUG nova.compute.manager [req-7d7ed494-2504-47ff-8644-27f0025b83e9 req-7709672e-61a6-41f7-8269-9b76b8f822db service nova] [instance: fd481728-3ef0-4a96-affd-ab2dd3f596bb] Refreshing instance network info cache due to event network-changed-7df3f02b-e4e5-4840-bf9f-3498771b9184. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 679.400419] env[65726]: DEBUG oslo_concurrency.lockutils [req-7d7ed494-2504-47ff-8644-27f0025b83e9 req-7709672e-61a6-41f7-8269-9b76b8f822db service nova] Acquiring lock "refresh_cache-fd481728-3ef0-4a96-affd-ab2dd3f596bb" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 679.400544] env[65726]: DEBUG oslo_concurrency.lockutils [req-7d7ed494-2504-47ff-8644-27f0025b83e9 req-7709672e-61a6-41f7-8269-9b76b8f822db service nova] Acquired lock "refresh_cache-fd481728-3ef0-4a96-affd-ab2dd3f596bb" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 679.400695] env[65726]: DEBUG nova.network.neutron [req-7d7ed494-2504-47ff-8644-27f0025b83e9 req-7709672e-61a6-41f7-8269-9b76b8f822db service nova] [instance: fd481728-3ef0-4a96-affd-ab2dd3f596bb] Refreshing network info cache for port 7df3f02b-e4e5-4840-bf9f-3498771b9184 {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 679.412569] env[65726]: DEBUG oslo_concurrency.lockutils [None req-06465b67-a5cc-4aeb-80ed-00dcc8ab79e5 tempest-ServersTestJSON-1793637688 tempest-ServersTestJSON-1793637688-project-member] Acquiring lock "88fb7235-cd0e-49dc-9d78-3583d5d3f528" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 679.412787] env[65726]: DEBUG oslo_concurrency.lockutils [None req-06465b67-a5cc-4aeb-80ed-00dcc8ab79e5 tempest-ServersTestJSON-1793637688 tempest-ServersTestJSON-1793637688-project-member] Lock "88fb7235-cd0e-49dc-9d78-3583d5d3f528" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 679.412977] env[65726]: DEBUG oslo_concurrency.lockutils [None req-06465b67-a5cc-4aeb-80ed-00dcc8ab79e5 tempest-ServersTestJSON-1793637688 tempest-ServersTestJSON-1793637688-project-member] Acquiring lock "88fb7235-cd0e-49dc-9d78-3583d5d3f528-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 679.413157] env[65726]: DEBUG oslo_concurrency.lockutils [None req-06465b67-a5cc-4aeb-80ed-00dcc8ab79e5 tempest-ServersTestJSON-1793637688 tempest-ServersTestJSON-1793637688-project-member] Lock "88fb7235-cd0e-49dc-9d78-3583d5d3f528-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 679.413314] env[65726]: DEBUG oslo_concurrency.lockutils [None req-06465b67-a5cc-4aeb-80ed-00dcc8ab79e5 tempest-ServersTestJSON-1793637688 tempest-ServersTestJSON-1793637688-project-member] Lock "88fb7235-cd0e-49dc-9d78-3583d5d3f528-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 679.416131] env[65726]: INFO nova.compute.manager [None req-06465b67-a5cc-4aeb-80ed-00dcc8ab79e5 tempest-ServersTestJSON-1793637688 tempest-ServersTestJSON-1793637688-project-member] [instance: 88fb7235-cd0e-49dc-9d78-3583d5d3f528] Terminating instance [ 679.511431] env[65726]: DEBUG oslo_concurrency.lockutils [None req-bb345b26-9f49-4e52-904f-99ccc71299ac tempest-AttachInterfacesV270Test-133387050 tempest-AttachInterfacesV270Test-133387050-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.863s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 679.512038] env[65726]: DEBUG nova.compute.manager [None req-bb345b26-9f49-4e52-904f-99ccc71299ac tempest-AttachInterfacesV270Test-133387050 tempest-AttachInterfacesV270Test-133387050-project-member] [instance: c977257e-dce0-4ecc-b42b-0f6ebab38797] Start building networks asynchronously for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 679.514907] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5fd4688c-49eb-401e-afbd-116789f6a402 tempest-ServerDiagnosticsV248Test-347810772 tempest-ServerDiagnosticsV248Test-347810772-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 18.660s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 679.515080] env[65726]: DEBUG nova.objects.instance [None req-5fd4688c-49eb-401e-afbd-116789f6a402 tempest-ServerDiagnosticsV248Test-347810772 tempest-ServerDiagnosticsV248Test-347810772-project-member] Lazy-loading 'resources' on Instance uuid 4d4dc88b-0d6c-430c-ab4f-c81a56e4657b {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 679.814135] env[65726]: DEBUG nova.compute.manager [None req-c24265b5-a0c6-4266-b4c0-64535a550ddd tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: bd839fe1-8801-4ba1-9c23-288453258b66] Starting instance... {{(pid=65726) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 679.872297] env[65726]: DEBUG oslo_vmware.api [None req-efa812b8-b008-4879-b549-6ccf454f374f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52a31a85-0ede-ff3d-eeb6-2dd8105e979f, 'name': SearchDatastore_Task, 'duration_secs': 0.023418} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 679.872889] env[65726]: DEBUG oslo_concurrency.lockutils [None req-efa812b8-b008-4879-b549-6ccf454f374f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 679.873139] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-efa812b8-b008-4879-b549-6ccf454f374f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] [instance: ddd36db0-28ad-4582-994b-7eef8b08ed46] Processing image b52362a3-ee8a-4cbf-b06f-513b0cc8f95c {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 679.873377] env[65726]: DEBUG oslo_concurrency.lockutils [None req-efa812b8-b008-4879-b549-6ccf454f374f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 679.873532] env[65726]: DEBUG oslo_concurrency.lockutils [None req-efa812b8-b008-4879-b549-6ccf454f374f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 679.873704] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-efa812b8-b008-4879-b549-6ccf454f374f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 679.875624] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-078dd7fa-06b3-4736-905b-2f5e0e210f34 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.877616] env[65726]: INFO nova.compute.manager [None req-fb0f14b3-fd50-4c8d-a984-877b235dcb32 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] [instance: d7905c4f-1ba3-43b3-b4b3-aebacd2f9cc3] Took 30.73 seconds to build instance. [ 679.890089] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-efa812b8-b008-4879-b549-6ccf454f374f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 679.890312] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-efa812b8-b008-4879-b549-6ccf454f374f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65726) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 679.891076] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-69528af7-1ea5-443e-8a3d-171cb9816b0d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.901064] env[65726]: DEBUG oslo_vmware.api [None req-efa812b8-b008-4879-b549-6ccf454f374f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] Waiting for the task: (returnval){ [ 679.901064] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52c14e66-232d-6387-b1e8-fadea69b1212" [ 679.901064] env[65726]: _type = "Task" [ 679.901064] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 679.906254] env[65726]: WARNING openstack [req-7d7ed494-2504-47ff-8644-27f0025b83e9 req-7709672e-61a6-41f7-8269-9b76b8f822db service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 679.907058] env[65726]: WARNING openstack [req-7d7ed494-2504-47ff-8644-27f0025b83e9 req-7709672e-61a6-41f7-8269-9b76b8f822db service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 679.921644] env[65726]: DEBUG nova.compute.manager [None req-06465b67-a5cc-4aeb-80ed-00dcc8ab79e5 tempest-ServersTestJSON-1793637688 tempest-ServersTestJSON-1793637688-project-member] [instance: 88fb7235-cd0e-49dc-9d78-3583d5d3f528] Start destroying the instance on the hypervisor. {{(pid=65726) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 679.921861] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-06465b67-a5cc-4aeb-80ed-00dcc8ab79e5 tempest-ServersTestJSON-1793637688 tempest-ServersTestJSON-1793637688-project-member] [instance: 88fb7235-cd0e-49dc-9d78-3583d5d3f528] Destroying instance {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 679.922464] env[65726]: DEBUG oslo_vmware.api [None req-efa812b8-b008-4879-b549-6ccf454f374f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52c14e66-232d-6387-b1e8-fadea69b1212, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 679.923290] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3fa54912-9d4d-427b-acfa-74d21abac40d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.932082] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-06465b67-a5cc-4aeb-80ed-00dcc8ab79e5 tempest-ServersTestJSON-1793637688 tempest-ServersTestJSON-1793637688-project-member] [instance: 88fb7235-cd0e-49dc-9d78-3583d5d3f528] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 679.932350] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8c02f595-32dc-4248-96c4-1e88145d81a0 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.941910] env[65726]: DEBUG oslo_vmware.api [None req-06465b67-a5cc-4aeb-80ed-00dcc8ab79e5 tempest-ServersTestJSON-1793637688 tempest-ServersTestJSON-1793637688-project-member] Waiting for the task: (returnval){ [ 679.941910] env[65726]: value = "task-5115598" [ 679.941910] env[65726]: _type = "Task" [ 679.941910] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 679.954098] env[65726]: DEBUG oslo_vmware.api [None req-06465b67-a5cc-4aeb-80ed-00dcc8ab79e5 tempest-ServersTestJSON-1793637688 tempest-ServersTestJSON-1793637688-project-member] Task: {'id': task-5115598, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 680.018942] env[65726]: DEBUG nova.compute.utils [None req-bb345b26-9f49-4e52-904f-99ccc71299ac tempest-AttachInterfacesV270Test-133387050 tempest-AttachInterfacesV270Test-133387050-project-member] Using /dev/sd instead of None {{(pid=65726) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 680.024848] env[65726]: DEBUG nova.compute.manager [None req-bb345b26-9f49-4e52-904f-99ccc71299ac tempest-AttachInterfacesV270Test-133387050 tempest-AttachInterfacesV270Test-133387050-project-member] [instance: c977257e-dce0-4ecc-b42b-0f6ebab38797] Allocating IP information in the background. {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 680.024848] env[65726]: DEBUG nova.network.neutron [None req-bb345b26-9f49-4e52-904f-99ccc71299ac tempest-AttachInterfacesV270Test-133387050 tempest-AttachInterfacesV270Test-133387050-project-member] [instance: c977257e-dce0-4ecc-b42b-0f6ebab38797] allocate_for_instance() {{(pid=65726) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 680.024848] env[65726]: WARNING neutronclient.v2_0.client [None req-bb345b26-9f49-4e52-904f-99ccc71299ac tempest-AttachInterfacesV270Test-133387050 tempest-AttachInterfacesV270Test-133387050-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 680.025327] env[65726]: WARNING neutronclient.v2_0.client [None req-bb345b26-9f49-4e52-904f-99ccc71299ac tempest-AttachInterfacesV270Test-133387050 tempest-AttachInterfacesV270Test-133387050-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 680.025975] env[65726]: WARNING openstack [None req-bb345b26-9f49-4e52-904f-99ccc71299ac tempest-AttachInterfacesV270Test-133387050 tempest-AttachInterfacesV270Test-133387050-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 680.026461] env[65726]: WARNING openstack [None req-bb345b26-9f49-4e52-904f-99ccc71299ac tempest-AttachInterfacesV270Test-133387050 tempest-AttachInterfacesV270Test-133387050-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 680.077140] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b479f246-d94b-4634-888f-dee4586c3272 tempest-ServerAddressesNegativeTestJSON-963263679 tempest-ServerAddressesNegativeTestJSON-963263679-project-member] Acquiring lock "00746508-d0d6-4dfa-9026-772398d004ee" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 680.077689] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b479f246-d94b-4634-888f-dee4586c3272 tempest-ServerAddressesNegativeTestJSON-963263679 tempest-ServerAddressesNegativeTestJSON-963263679-project-member] Lock "00746508-d0d6-4dfa-9026-772398d004ee" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 680.083035] env[65726]: DEBUG nova.policy [None req-bb345b26-9f49-4e52-904f-99ccc71299ac tempest-AttachInterfacesV270Test-133387050 tempest-AttachInterfacesV270Test-133387050-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2c36c238ecc6496197f29bff39511747', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '459cd960efb744ef91c113bb584444bd', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65726) authorize /opt/stack/nova/nova/policy.py:192}} [ 680.123662] env[65726]: WARNING openstack [req-7d7ed494-2504-47ff-8644-27f0025b83e9 req-7709672e-61a6-41f7-8269-9b76b8f822db service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 680.123662] env[65726]: WARNING openstack [req-7d7ed494-2504-47ff-8644-27f0025b83e9 req-7709672e-61a6-41f7-8269-9b76b8f822db service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 680.287189] env[65726]: WARNING openstack [req-7d7ed494-2504-47ff-8644-27f0025b83e9 req-7709672e-61a6-41f7-8269-9b76b8f822db service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 680.287563] env[65726]: WARNING openstack [req-7d7ed494-2504-47ff-8644-27f0025b83e9 req-7709672e-61a6-41f7-8269-9b76b8f822db service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 680.345844] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c24265b5-a0c6-4266-b4c0-64535a550ddd tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 680.364287] env[65726]: DEBUG nova.network.neutron [None req-bb345b26-9f49-4e52-904f-99ccc71299ac tempest-AttachInterfacesV270Test-133387050 tempest-AttachInterfacesV270Test-133387050-project-member] [instance: c977257e-dce0-4ecc-b42b-0f6ebab38797] Successfully created port: 881e11d6-dc73-4e2b-9d9b-eec7da46f63f {{(pid=65726) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 680.383880] env[65726]: DEBUG oslo_concurrency.lockutils [None req-fb0f14b3-fd50-4c8d-a984-877b235dcb32 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Lock "d7905c4f-1ba3-43b3-b4b3-aebacd2f9cc3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 42.461s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 680.421632] env[65726]: DEBUG oslo_vmware.api [None req-efa812b8-b008-4879-b549-6ccf454f374f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52c14e66-232d-6387-b1e8-fadea69b1212, 'name': SearchDatastore_Task, 'duration_secs': 0.029975} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 680.422579] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-acdd680f-0066-4c14-9e47-85ee44d70973 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.430021] env[65726]: DEBUG oslo_vmware.api [None req-efa812b8-b008-4879-b549-6ccf454f374f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] Waiting for the task: (returnval){ [ 680.430021] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]524dac32-c837-f08c-dce7-84bfd7490e5a" [ 680.430021] env[65726]: _type = "Task" [ 680.430021] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 680.453697] env[65726]: DEBUG oslo_vmware.api [None req-efa812b8-b008-4879-b549-6ccf454f374f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]524dac32-c837-f08c-dce7-84bfd7490e5a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 680.463240] env[65726]: DEBUG oslo_vmware.api [None req-06465b67-a5cc-4aeb-80ed-00dcc8ab79e5 tempest-ServersTestJSON-1793637688 tempest-ServersTestJSON-1793637688-project-member] Task: {'id': task-5115598, 'name': PowerOffVM_Task, 'duration_secs': 0.297577} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 680.466917] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-06465b67-a5cc-4aeb-80ed-00dcc8ab79e5 tempest-ServersTestJSON-1793637688 tempest-ServersTestJSON-1793637688-project-member] [instance: 88fb7235-cd0e-49dc-9d78-3583d5d3f528] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 680.467049] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-06465b67-a5cc-4aeb-80ed-00dcc8ab79e5 tempest-ServersTestJSON-1793637688 tempest-ServersTestJSON-1793637688-project-member] [instance: 88fb7235-cd0e-49dc-9d78-3583d5d3f528] Unregistering the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 680.468030] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1dfadf10-1ebe-4291-9342-4633accd0393 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.511031] env[65726]: DEBUG nova.network.neutron [req-7d7ed494-2504-47ff-8644-27f0025b83e9 req-7709672e-61a6-41f7-8269-9b76b8f822db service nova] [instance: fd481728-3ef0-4a96-affd-ab2dd3f596bb] Updated VIF entry in instance network info cache for port 7df3f02b-e4e5-4840-bf9f-3498771b9184. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 680.511031] env[65726]: DEBUG nova.network.neutron [req-7d7ed494-2504-47ff-8644-27f0025b83e9 req-7709672e-61a6-41f7-8269-9b76b8f822db service nova] [instance: fd481728-3ef0-4a96-affd-ab2dd3f596bb] Updating instance_info_cache with network_info: [{"id": "7df3f02b-e4e5-4840-bf9f-3498771b9184", "address": "fa:16:3e:6b:c1:e9", "network": {"id": "d37113a1-409c-4c2c-81f9-8f80822b85b9", "bridge": "br-int", "label": "tempest-ServersWithSpecificFlavorTestJSON-471231219-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.139", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "79d3e960f190439fae7c61d693e9356e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "13af9422-d668-4413-b63a-766558d83a3b", "external-id": "nsx-vlan-transportzone-842", "segmentation_id": 842, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7df3f02b-e4", "ovs_interfaceid": "7df3f02b-e4e5-4840-bf9f-3498771b9184", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 680.525316] env[65726]: DEBUG nova.compute.manager [None req-bb345b26-9f49-4e52-904f-99ccc71299ac tempest-AttachInterfacesV270Test-133387050 tempest-AttachInterfacesV270Test-133387050-project-member] [instance: c977257e-dce0-4ecc-b42b-0f6ebab38797] Start building block device mappings for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 680.543677] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-06465b67-a5cc-4aeb-80ed-00dcc8ab79e5 tempest-ServersTestJSON-1793637688 tempest-ServersTestJSON-1793637688-project-member] [instance: 88fb7235-cd0e-49dc-9d78-3583d5d3f528] Unregistered the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 680.544262] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-06465b67-a5cc-4aeb-80ed-00dcc8ab79e5 tempest-ServersTestJSON-1793637688 tempest-ServersTestJSON-1793637688-project-member] [instance: 88fb7235-cd0e-49dc-9d78-3583d5d3f528] Deleting contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 680.544262] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-06465b67-a5cc-4aeb-80ed-00dcc8ab79e5 tempest-ServersTestJSON-1793637688 tempest-ServersTestJSON-1793637688-project-member] Deleting the datastore file [datastore1] 88fb7235-cd0e-49dc-9d78-3583d5d3f528 {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 680.544617] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-763b042b-8ff6-4819-a713-6fb0ee8af621 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.558025] env[65726]: DEBUG oslo_vmware.api [None req-06465b67-a5cc-4aeb-80ed-00dcc8ab79e5 tempest-ServersTestJSON-1793637688 tempest-ServersTestJSON-1793637688-project-member] Waiting for the task: (returnval){ [ 680.558025] env[65726]: value = "task-5115600" [ 680.558025] env[65726]: _type = "Task" [ 680.558025] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 680.569388] env[65726]: DEBUG oslo_vmware.api [None req-06465b67-a5cc-4aeb-80ed-00dcc8ab79e5 tempest-ServersTestJSON-1793637688 tempest-ServersTestJSON-1793637688-project-member] Task: {'id': task-5115600, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 680.733787] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00a1cf03-79de-4eee-86c8-c52f8412e975 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.740944] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6ed5cf6-d95a-443c-b1d0-488c07c3b76d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.774414] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c2f0db1-5d67-43f1-ae90-1d9df45eed6c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.783216] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6de66a85-bdb1-4e68-b8f4-4b62cfedd26f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.798709] env[65726]: DEBUG nova.compute.provider_tree [None req-5fd4688c-49eb-401e-afbd-116789f6a402 tempest-ServerDiagnosticsV248Test-347810772 tempest-ServerDiagnosticsV248Test-347810772-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 680.887592] env[65726]: DEBUG nova.compute.manager [None req-2737dc3f-715d-481c-b9a9-998472382a0d tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] [instance: 83573c2b-9448-456f-8fd2-b19661dd6cc4] Starting instance... {{(pid=65726) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 680.943776] env[65726]: DEBUG oslo_vmware.api [None req-efa812b8-b008-4879-b549-6ccf454f374f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]524dac32-c837-f08c-dce7-84bfd7490e5a, 'name': SearchDatastore_Task, 'duration_secs': 0.025277} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 680.944110] env[65726]: DEBUG oslo_concurrency.lockutils [None req-efa812b8-b008-4879-b549-6ccf454f374f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 680.944381] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-efa812b8-b008-4879-b549-6ccf454f374f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore2] ddd36db0-28ad-4582-994b-7eef8b08ed46/ddd36db0-28ad-4582-994b-7eef8b08ed46.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 680.944645] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-485760e0-878c-4e1e-982f-d74a320a91ca {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.954282] env[65726]: DEBUG oslo_vmware.api [None req-efa812b8-b008-4879-b549-6ccf454f374f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] Waiting for the task: (returnval){ [ 680.954282] env[65726]: value = "task-5115601" [ 680.954282] env[65726]: _type = "Task" [ 680.954282] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 680.964726] env[65726]: DEBUG oslo_vmware.api [None req-efa812b8-b008-4879-b549-6ccf454f374f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] Task: {'id': task-5115601, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 681.014773] env[65726]: DEBUG oslo_concurrency.lockutils [req-7d7ed494-2504-47ff-8644-27f0025b83e9 req-7709672e-61a6-41f7-8269-9b76b8f822db service nova] Releasing lock "refresh_cache-fd481728-3ef0-4a96-affd-ab2dd3f596bb" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 681.071635] env[65726]: DEBUG oslo_vmware.api [None req-06465b67-a5cc-4aeb-80ed-00dcc8ab79e5 tempest-ServersTestJSON-1793637688 tempest-ServersTestJSON-1793637688-project-member] Task: {'id': task-5115600, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.457234} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 681.072252] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-06465b67-a5cc-4aeb-80ed-00dcc8ab79e5 tempest-ServersTestJSON-1793637688 tempest-ServersTestJSON-1793637688-project-member] Deleted the datastore file {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 681.072338] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-06465b67-a5cc-4aeb-80ed-00dcc8ab79e5 tempest-ServersTestJSON-1793637688 tempest-ServersTestJSON-1793637688-project-member] [instance: 88fb7235-cd0e-49dc-9d78-3583d5d3f528] Deleted contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 681.073437] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-06465b67-a5cc-4aeb-80ed-00dcc8ab79e5 tempest-ServersTestJSON-1793637688 tempest-ServersTestJSON-1793637688-project-member] [instance: 88fb7235-cd0e-49dc-9d78-3583d5d3f528] Instance destroyed {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 681.073437] env[65726]: INFO nova.compute.manager [None req-06465b67-a5cc-4aeb-80ed-00dcc8ab79e5 tempest-ServersTestJSON-1793637688 tempest-ServersTestJSON-1793637688-project-member] [instance: 88fb7235-cd0e-49dc-9d78-3583d5d3f528] Took 1.15 seconds to destroy the instance on the hypervisor. [ 681.073437] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-06465b67-a5cc-4aeb-80ed-00dcc8ab79e5 tempest-ServersTestJSON-1793637688 tempest-ServersTestJSON-1793637688-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 681.073710] env[65726]: DEBUG nova.compute.manager [-] [instance: 88fb7235-cd0e-49dc-9d78-3583d5d3f528] Deallocating network for instance {{(pid=65726) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 681.073710] env[65726]: DEBUG nova.network.neutron [-] [instance: 88fb7235-cd0e-49dc-9d78-3583d5d3f528] deallocate_for_instance() {{(pid=65726) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 681.074260] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 681.074658] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 681.118935] env[65726]: DEBUG oslo_concurrency.lockutils [None req-04afb47c-692a-46d9-a640-68cc2c90206f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Acquiring lock "c108f2a5-031b-47ee-9a5b-d62c2f42c26b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 681.118935] env[65726]: DEBUG oslo_concurrency.lockutils [None req-04afb47c-692a-46d9-a640-68cc2c90206f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Lock "c108f2a5-031b-47ee-9a5b-d62c2f42c26b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 681.127568] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 681.127839] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 681.141276] env[65726]: DEBUG oslo_concurrency.lockutils [None req-44e83fc7-afcc-4134-bbbb-6bc576931bb8 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Acquiring lock "895797e4-2941-44cd-aab1-67afa4fac02d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 681.141553] env[65726]: DEBUG oslo_concurrency.lockutils [None req-44e83fc7-afcc-4134-bbbb-6bc576931bb8 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Lock "895797e4-2941-44cd-aab1-67afa4fac02d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 681.302578] env[65726]: DEBUG nova.scheduler.client.report [None req-5fd4688c-49eb-401e-afbd-116789f6a402 tempest-ServerDiagnosticsV248Test-347810772 tempest-ServerDiagnosticsV248Test-347810772-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 681.319032] env[65726]: DEBUG oslo_concurrency.lockutils [None req-2fbf5465-13e9-48fa-a233-c5a8ee2f78d3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Acquiring lock "703fff2d-5aec-4ed1-b987-9e1cc3a25f67" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 681.319032] env[65726]: DEBUG oslo_concurrency.lockutils [None req-2fbf5465-13e9-48fa-a233-c5a8ee2f78d3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Lock "703fff2d-5aec-4ed1-b987-9e1cc3a25f67" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 681.319032] env[65726]: DEBUG nova.compute.manager [None req-2fbf5465-13e9-48fa-a233-c5a8ee2f78d3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] [instance: 703fff2d-5aec-4ed1-b987-9e1cc3a25f67] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 681.319032] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33fc3bdd-0897-42f0-97ba-abfe82e9376d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.328282] env[65726]: DEBUG nova.compute.manager [None req-2fbf5465-13e9-48fa-a233-c5a8ee2f78d3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] [instance: 703fff2d-5aec-4ed1-b987-9e1cc3a25f67] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=65726) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3459}} [ 681.329777] env[65726]: DEBUG nova.objects.instance [None req-2fbf5465-13e9-48fa-a233-c5a8ee2f78d3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Lazy-loading 'flavor' on Instance uuid 703fff2d-5aec-4ed1-b987-9e1cc3a25f67 {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 681.429130] env[65726]: DEBUG oslo_concurrency.lockutils [None req-2737dc3f-715d-481c-b9a9-998472382a0d tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 681.467103] env[65726]: DEBUG oslo_vmware.api [None req-efa812b8-b008-4879-b549-6ccf454f374f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] Task: {'id': task-5115601, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 681.538346] env[65726]: DEBUG nova.compute.manager [None req-bb345b26-9f49-4e52-904f-99ccc71299ac tempest-AttachInterfacesV270Test-133387050 tempest-AttachInterfacesV270Test-133387050-project-member] [instance: c977257e-dce0-4ecc-b42b-0f6ebab38797] Start spawning the instance on the hypervisor. {{(pid=65726) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 681.576414] env[65726]: DEBUG nova.virt.hardware [None req-bb345b26-9f49-4e52-904f-99ccc71299ac tempest-AttachInterfacesV270Test-133387050 tempest-AttachInterfacesV270Test-133387050-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T19:28:34Z,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1539f06fa3534e90acbc3a60c4b8bd3f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T19:28:35Z,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 681.576708] env[65726]: DEBUG nova.virt.hardware [None req-bb345b26-9f49-4e52-904f-99ccc71299ac tempest-AttachInterfacesV270Test-133387050 tempest-AttachInterfacesV270Test-133387050-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 681.577143] env[65726]: DEBUG nova.virt.hardware [None req-bb345b26-9f49-4e52-904f-99ccc71299ac tempest-AttachInterfacesV270Test-133387050 tempest-AttachInterfacesV270Test-133387050-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 681.577415] env[65726]: DEBUG nova.virt.hardware [None req-bb345b26-9f49-4e52-904f-99ccc71299ac tempest-AttachInterfacesV270Test-133387050 tempest-AttachInterfacesV270Test-133387050-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 681.577576] env[65726]: DEBUG nova.virt.hardware [None req-bb345b26-9f49-4e52-904f-99ccc71299ac tempest-AttachInterfacesV270Test-133387050 tempest-AttachInterfacesV270Test-133387050-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 681.577738] env[65726]: DEBUG nova.virt.hardware [None req-bb345b26-9f49-4e52-904f-99ccc71299ac tempest-AttachInterfacesV270Test-133387050 tempest-AttachInterfacesV270Test-133387050-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 681.577974] env[65726]: DEBUG nova.virt.hardware [None req-bb345b26-9f49-4e52-904f-99ccc71299ac tempest-AttachInterfacesV270Test-133387050 tempest-AttachInterfacesV270Test-133387050-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 681.578152] env[65726]: DEBUG nova.virt.hardware [None req-bb345b26-9f49-4e52-904f-99ccc71299ac tempest-AttachInterfacesV270Test-133387050 tempest-AttachInterfacesV270Test-133387050-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 681.578328] env[65726]: DEBUG nova.virt.hardware [None req-bb345b26-9f49-4e52-904f-99ccc71299ac tempest-AttachInterfacesV270Test-133387050 tempest-AttachInterfacesV270Test-133387050-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 681.578492] env[65726]: DEBUG nova.virt.hardware [None req-bb345b26-9f49-4e52-904f-99ccc71299ac tempest-AttachInterfacesV270Test-133387050 tempest-AttachInterfacesV270Test-133387050-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 681.578664] env[65726]: DEBUG nova.virt.hardware [None req-bb345b26-9f49-4e52-904f-99ccc71299ac tempest-AttachInterfacesV270Test-133387050 tempest-AttachInterfacesV270Test-133387050-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 681.579995] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec1f9634-a45d-47ca-8034-014b6972763c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.589994] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2582a92-faed-4666-814f-80b93d4e473e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.765792] env[65726]: DEBUG nova.compute.manager [req-5cf22853-4b8c-4f14-ab78-e2c5fe0514f6 req-6db2a11a-f472-402f-9d05-1dcefc510cca service nova] [instance: 5a8d4357-4b44-4a19-b1da-42d188c38adc] Received event network-changed-9aaacd82-31c1-4cfa-bbce-860b30db74af {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 681.768539] env[65726]: DEBUG nova.compute.manager [req-5cf22853-4b8c-4f14-ab78-e2c5fe0514f6 req-6db2a11a-f472-402f-9d05-1dcefc510cca service nova] [instance: 5a8d4357-4b44-4a19-b1da-42d188c38adc] Refreshing instance network info cache due to event network-changed-9aaacd82-31c1-4cfa-bbce-860b30db74af. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 681.768539] env[65726]: DEBUG oslo_concurrency.lockutils [req-5cf22853-4b8c-4f14-ab78-e2c5fe0514f6 req-6db2a11a-f472-402f-9d05-1dcefc510cca service nova] Acquiring lock "refresh_cache-5a8d4357-4b44-4a19-b1da-42d188c38adc" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 681.768539] env[65726]: DEBUG oslo_concurrency.lockutils [req-5cf22853-4b8c-4f14-ab78-e2c5fe0514f6 req-6db2a11a-f472-402f-9d05-1dcefc510cca service nova] Acquired lock "refresh_cache-5a8d4357-4b44-4a19-b1da-42d188c38adc" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 681.768539] env[65726]: DEBUG nova.network.neutron [req-5cf22853-4b8c-4f14-ab78-e2c5fe0514f6 req-6db2a11a-f472-402f-9d05-1dcefc510cca service nova] [instance: 5a8d4357-4b44-4a19-b1da-42d188c38adc] Refreshing network info cache for port 9aaacd82-31c1-4cfa-bbce-860b30db74af {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 681.813030] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5fd4688c-49eb-401e-afbd-116789f6a402 tempest-ServerDiagnosticsV248Test-347810772 tempest-ServerDiagnosticsV248Test-347810772-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.296s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 681.813792] env[65726]: DEBUG oslo_concurrency.lockutils [None req-f72e4138-ba89-47ec-80a1-52d74fe64f4e tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 20.367s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 681.813994] env[65726]: DEBUG nova.objects.instance [None req-f72e4138-ba89-47ec-80a1-52d74fe64f4e tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Lazy-loading 'resources' on Instance uuid 8e9c27f6-e9fa-4ed0-b6ae-3b5df52d1c44 {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 681.854292] env[65726]: INFO nova.scheduler.client.report [None req-5fd4688c-49eb-401e-afbd-116789f6a402 tempest-ServerDiagnosticsV248Test-347810772 tempest-ServerDiagnosticsV248Test-347810772-project-member] Deleted allocations for instance 4d4dc88b-0d6c-430c-ab4f-c81a56e4657b [ 681.968477] env[65726]: DEBUG oslo_vmware.api [None req-efa812b8-b008-4879-b549-6ccf454f374f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] Task: {'id': task-5115601, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.550509} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 681.969029] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-efa812b8-b008-4879-b549-6ccf454f374f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore2] ddd36db0-28ad-4582-994b-7eef8b08ed46/ddd36db0-28ad-4582-994b-7eef8b08ed46.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 681.969392] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-efa812b8-b008-4879-b549-6ccf454f374f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] [instance: ddd36db0-28ad-4582-994b-7eef8b08ed46] Extending root virtual disk to 1048576 {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 681.969790] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3be8eaa6-b7ff-48d0-bb72-fce255397311 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.978384] env[65726]: DEBUG oslo_vmware.api [None req-efa812b8-b008-4879-b549-6ccf454f374f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] Waiting for the task: (returnval){ [ 681.978384] env[65726]: value = "task-5115602" [ 681.978384] env[65726]: _type = "Task" [ 681.978384] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 681.988272] env[65726]: DEBUG oslo_vmware.api [None req-efa812b8-b008-4879-b549-6ccf454f374f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] Task: {'id': task-5115602, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 682.014139] env[65726]: DEBUG nova.network.neutron [None req-bb345b26-9f49-4e52-904f-99ccc71299ac tempest-AttachInterfacesV270Test-133387050 tempest-AttachInterfacesV270Test-133387050-project-member] [instance: c977257e-dce0-4ecc-b42b-0f6ebab38797] Successfully updated port: 881e11d6-dc73-4e2b-9d9b-eec7da46f63f {{(pid=65726) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 682.206594] env[65726]: DEBUG nova.network.neutron [-] [instance: 88fb7235-cd0e-49dc-9d78-3583d5d3f528] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 682.274273] env[65726]: WARNING openstack [req-5cf22853-4b8c-4f14-ab78-e2c5fe0514f6 req-6db2a11a-f472-402f-9d05-1dcefc510cca service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 682.274273] env[65726]: WARNING openstack [req-5cf22853-4b8c-4f14-ab78-e2c5fe0514f6 req-6db2a11a-f472-402f-9d05-1dcefc510cca service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 682.340512] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-2fbf5465-13e9-48fa-a233-c5a8ee2f78d3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] [instance: 703fff2d-5aec-4ed1-b987-9e1cc3a25f67] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 682.340512] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ca88a32d-7dd5-4ebf-ab17-ccbabc60d3f5 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.353721] env[65726]: DEBUG oslo_vmware.api [None req-2fbf5465-13e9-48fa-a233-c5a8ee2f78d3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Waiting for the task: (returnval){ [ 682.353721] env[65726]: value = "task-5115603" [ 682.353721] env[65726]: _type = "Task" [ 682.353721] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 682.369418] env[65726]: DEBUG oslo_vmware.api [None req-2fbf5465-13e9-48fa-a233-c5a8ee2f78d3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Task: {'id': task-5115603, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 682.370132] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5fd4688c-49eb-401e-afbd-116789f6a402 tempest-ServerDiagnosticsV248Test-347810772 tempest-ServerDiagnosticsV248Test-347810772-project-member] Lock "4d4dc88b-0d6c-430c-ab4f-c81a56e4657b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 27.006s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 682.435452] env[65726]: WARNING openstack [req-5cf22853-4b8c-4f14-ab78-e2c5fe0514f6 req-6db2a11a-f472-402f-9d05-1dcefc510cca service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 682.435839] env[65726]: WARNING openstack [req-5cf22853-4b8c-4f14-ab78-e2c5fe0514f6 req-6db2a11a-f472-402f-9d05-1dcefc510cca service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 682.496042] env[65726]: DEBUG oslo_vmware.api [None req-efa812b8-b008-4879-b549-6ccf454f374f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] Task: {'id': task-5115602, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.119767} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 682.496042] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-efa812b8-b008-4879-b549-6ccf454f374f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] [instance: ddd36db0-28ad-4582-994b-7eef8b08ed46] Extended root virtual disk {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 682.496805] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a7e76c5-cead-4804-b9fe-66f221dfba7a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.525903] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-efa812b8-b008-4879-b549-6ccf454f374f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] [instance: ddd36db0-28ad-4582-994b-7eef8b08ed46] Reconfiguring VM instance instance-00000019 to attach disk [datastore2] ddd36db0-28ad-4582-994b-7eef8b08ed46/ddd36db0-28ad-4582-994b-7eef8b08ed46.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 682.529561] env[65726]: DEBUG oslo_concurrency.lockutils [None req-bb345b26-9f49-4e52-904f-99ccc71299ac tempest-AttachInterfacesV270Test-133387050 tempest-AttachInterfacesV270Test-133387050-project-member] Acquiring lock "refresh_cache-c977257e-dce0-4ecc-b42b-0f6ebab38797" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 682.529765] env[65726]: DEBUG oslo_concurrency.lockutils [None req-bb345b26-9f49-4e52-904f-99ccc71299ac tempest-AttachInterfacesV270Test-133387050 tempest-AttachInterfacesV270Test-133387050-project-member] Acquired lock "refresh_cache-c977257e-dce0-4ecc-b42b-0f6ebab38797" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 682.529941] env[65726]: DEBUG nova.network.neutron [None req-bb345b26-9f49-4e52-904f-99ccc71299ac tempest-AttachInterfacesV270Test-133387050 tempest-AttachInterfacesV270Test-133387050-project-member] [instance: c977257e-dce0-4ecc-b42b-0f6ebab38797] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 682.535372] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5b28c2ce-dc7e-4edd-9c53-70270da0a276 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.554790] env[65726]: WARNING openstack [req-5cf22853-4b8c-4f14-ab78-e2c5fe0514f6 req-6db2a11a-f472-402f-9d05-1dcefc510cca service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 682.554790] env[65726]: WARNING openstack [req-5cf22853-4b8c-4f14-ab78-e2c5fe0514f6 req-6db2a11a-f472-402f-9d05-1dcefc510cca service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 682.571299] env[65726]: DEBUG oslo_vmware.api [None req-efa812b8-b008-4879-b549-6ccf454f374f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] Waiting for the task: (returnval){ [ 682.571299] env[65726]: value = "task-5115604" [ 682.571299] env[65726]: _type = "Task" [ 682.571299] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 682.584012] env[65726]: DEBUG oslo_vmware.api [None req-efa812b8-b008-4879-b549-6ccf454f374f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] Task: {'id': task-5115604, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 682.662206] env[65726]: DEBUG nova.network.neutron [req-5cf22853-4b8c-4f14-ab78-e2c5fe0514f6 req-6db2a11a-f472-402f-9d05-1dcefc510cca service nova] [instance: 5a8d4357-4b44-4a19-b1da-42d188c38adc] Updated VIF entry in instance network info cache for port 9aaacd82-31c1-4cfa-bbce-860b30db74af. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 682.662572] env[65726]: DEBUG nova.network.neutron [req-5cf22853-4b8c-4f14-ab78-e2c5fe0514f6 req-6db2a11a-f472-402f-9d05-1dcefc510cca service nova] [instance: 5a8d4357-4b44-4a19-b1da-42d188c38adc] Updating instance_info_cache with network_info: [{"id": "9aaacd82-31c1-4cfa-bbce-860b30db74af", "address": "fa:16:3e:a2:ff:16", "network": {"id": "bee3bacb-6a14-41ab-a9a3-a82d9ca0b175", "bridge": "br-int", "label": "tempest-ServersTestBootFromVolume-514317700-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.231", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c52b5289c78f45d3942a6c0a4b026207", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dad4f433-bb0b-45c7-8040-972ef2277f75", "external-id": "nsx-vlan-transportzone-451", "segmentation_id": 451, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9aaacd82-31", "ovs_interfaceid": "9aaacd82-31c1-4cfa-bbce-860b30db74af", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 682.713453] env[65726]: INFO nova.compute.manager [-] [instance: 88fb7235-cd0e-49dc-9d78-3583d5d3f528] Took 1.64 seconds to deallocate network for instance. [ 682.863862] env[65726]: DEBUG oslo_vmware.api [None req-2fbf5465-13e9-48fa-a233-c5a8ee2f78d3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Task: {'id': task-5115603, 'name': PowerOffVM_Task, 'duration_secs': 0.246879} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 682.866477] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-2fbf5465-13e9-48fa-a233-c5a8ee2f78d3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] [instance: 703fff2d-5aec-4ed1-b987-9e1cc3a25f67] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 682.866675] env[65726]: DEBUG nova.compute.manager [None req-2fbf5465-13e9-48fa-a233-c5a8ee2f78d3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] [instance: 703fff2d-5aec-4ed1-b987-9e1cc3a25f67] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 682.867713] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f10cd7ae-455f-4dbc-bb90-46d7376f59ef {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.955524] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc4357a3-eaa6-43fe-9c7a-419733c5adae {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.963944] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e501b214-9590-4eb8-b342-b9034712e842 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.999076] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-573c28ed-7336-41b1-9bc1-2aa00d20efb9 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.017507] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7346d2f1-19d1-49d5-84dd-954d66d1472c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.031342] env[65726]: DEBUG nova.compute.provider_tree [None req-f72e4138-ba89-47ec-80a1-52d74fe64f4e tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 683.034303] env[65726]: WARNING openstack [None req-bb345b26-9f49-4e52-904f-99ccc71299ac tempest-AttachInterfacesV270Test-133387050 tempest-AttachInterfacesV270Test-133387050-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 683.034303] env[65726]: WARNING openstack [None req-bb345b26-9f49-4e52-904f-99ccc71299ac tempest-AttachInterfacesV270Test-133387050 tempest-AttachInterfacesV270Test-133387050-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 683.039776] env[65726]: DEBUG nova.network.neutron [None req-bb345b26-9f49-4e52-904f-99ccc71299ac tempest-AttachInterfacesV270Test-133387050 tempest-AttachInterfacesV270Test-133387050-project-member] [instance: c977257e-dce0-4ecc-b42b-0f6ebab38797] Instance cache missing network info. {{(pid=65726) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 683.089142] env[65726]: DEBUG oslo_vmware.api [None req-efa812b8-b008-4879-b549-6ccf454f374f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] Task: {'id': task-5115604, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 683.101690] env[65726]: DEBUG oslo_concurrency.lockutils [None req-af2f0c1d-9bc5-40c3-a712-d408e73bc806 tempest-ServerDiagnosticsNegativeTest-1635132145 tempest-ServerDiagnosticsNegativeTest-1635132145-project-member] Acquiring lock "0d6f14c5-5770-476f-a186-4cf1b2d4fbe8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 683.101690] env[65726]: DEBUG oslo_concurrency.lockutils [None req-af2f0c1d-9bc5-40c3-a712-d408e73bc806 tempest-ServerDiagnosticsNegativeTest-1635132145 tempest-ServerDiagnosticsNegativeTest-1635132145-project-member] Lock "0d6f14c5-5770-476f-a186-4cf1b2d4fbe8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 683.124029] env[65726]: WARNING openstack [None req-bb345b26-9f49-4e52-904f-99ccc71299ac tempest-AttachInterfacesV270Test-133387050 tempest-AttachInterfacesV270Test-133387050-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 683.124443] env[65726]: WARNING openstack [None req-bb345b26-9f49-4e52-904f-99ccc71299ac tempest-AttachInterfacesV270Test-133387050 tempest-AttachInterfacesV270Test-133387050-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 683.166169] env[65726]: DEBUG oslo_concurrency.lockutils [req-5cf22853-4b8c-4f14-ab78-e2c5fe0514f6 req-6db2a11a-f472-402f-9d05-1dcefc510cca service nova] Releasing lock "refresh_cache-5a8d4357-4b44-4a19-b1da-42d188c38adc" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 683.221734] env[65726]: DEBUG oslo_concurrency.lockutils [None req-06465b67-a5cc-4aeb-80ed-00dcc8ab79e5 tempest-ServersTestJSON-1793637688 tempest-ServersTestJSON-1793637688-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 683.351150] env[65726]: WARNING openstack [None req-bb345b26-9f49-4e52-904f-99ccc71299ac tempest-AttachInterfacesV270Test-133387050 tempest-AttachInterfacesV270Test-133387050-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 683.352160] env[65726]: WARNING openstack [None req-bb345b26-9f49-4e52-904f-99ccc71299ac tempest-AttachInterfacesV270Test-133387050 tempest-AttachInterfacesV270Test-133387050-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 683.380573] env[65726]: DEBUG oslo_concurrency.lockutils [None req-2fbf5465-13e9-48fa-a233-c5a8ee2f78d3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Lock "703fff2d-5aec-4ed1-b987-9e1cc3a25f67" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.064s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 683.495301] env[65726]: DEBUG nova.network.neutron [None req-bb345b26-9f49-4e52-904f-99ccc71299ac tempest-AttachInterfacesV270Test-133387050 tempest-AttachInterfacesV270Test-133387050-project-member] [instance: c977257e-dce0-4ecc-b42b-0f6ebab38797] Updating instance_info_cache with network_info: [{"id": "881e11d6-dc73-4e2b-9d9b-eec7da46f63f", "address": "fa:16:3e:9d:66:49", "network": {"id": "70d88d94-31cf-4dcf-b41c-385f1939383d", "bridge": "br-int", "label": "tempest-AttachInterfacesV270Test-103265114-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "459cd960efb744ef91c113bb584444bd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "96d8be6c-b557-4b40-b0f5-838c62a3c904", "external-id": "nsx-vlan-transportzone-144", "segmentation_id": 144, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap881e11d6-dc", "ovs_interfaceid": "881e11d6-dc73-4e2b-9d9b-eec7da46f63f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 683.531734] env[65726]: DEBUG nova.compute.manager [req-16d47591-d7ea-479c-9a3f-89821c2bf429 req-fbf7f49f-aebf-48be-95c3-c1e928dbe1be service nova] [instance: c977257e-dce0-4ecc-b42b-0f6ebab38797] Received event network-vif-plugged-881e11d6-dc73-4e2b-9d9b-eec7da46f63f {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 683.531734] env[65726]: DEBUG oslo_concurrency.lockutils [req-16d47591-d7ea-479c-9a3f-89821c2bf429 req-fbf7f49f-aebf-48be-95c3-c1e928dbe1be service nova] Acquiring lock "c977257e-dce0-4ecc-b42b-0f6ebab38797-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 683.531830] env[65726]: DEBUG oslo_concurrency.lockutils [req-16d47591-d7ea-479c-9a3f-89821c2bf429 req-fbf7f49f-aebf-48be-95c3-c1e928dbe1be service nova] Lock "c977257e-dce0-4ecc-b42b-0f6ebab38797-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 683.532013] env[65726]: DEBUG oslo_concurrency.lockutils [req-16d47591-d7ea-479c-9a3f-89821c2bf429 req-fbf7f49f-aebf-48be-95c3-c1e928dbe1be service nova] Lock "c977257e-dce0-4ecc-b42b-0f6ebab38797-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 683.532292] env[65726]: DEBUG nova.compute.manager [req-16d47591-d7ea-479c-9a3f-89821c2bf429 req-fbf7f49f-aebf-48be-95c3-c1e928dbe1be service nova] [instance: c977257e-dce0-4ecc-b42b-0f6ebab38797] No waiting events found dispatching network-vif-plugged-881e11d6-dc73-4e2b-9d9b-eec7da46f63f {{(pid=65726) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 683.532462] env[65726]: WARNING nova.compute.manager [req-16d47591-d7ea-479c-9a3f-89821c2bf429 req-fbf7f49f-aebf-48be-95c3-c1e928dbe1be service nova] [instance: c977257e-dce0-4ecc-b42b-0f6ebab38797] Received unexpected event network-vif-plugged-881e11d6-dc73-4e2b-9d9b-eec7da46f63f for instance with vm_state building and task_state spawning. [ 683.548019] env[65726]: DEBUG nova.scheduler.client.report [None req-f72e4138-ba89-47ec-80a1-52d74fe64f4e tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 683.585728] env[65726]: DEBUG oslo_vmware.api [None req-efa812b8-b008-4879-b549-6ccf454f374f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] Task: {'id': task-5115604, 'name': ReconfigVM_Task, 'duration_secs': 0.902684} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 683.585728] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-efa812b8-b008-4879-b549-6ccf454f374f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] [instance: ddd36db0-28ad-4582-994b-7eef8b08ed46] Reconfigured VM instance instance-00000019 to attach disk [datastore2] ddd36db0-28ad-4582-994b-7eef8b08ed46/ddd36db0-28ad-4582-994b-7eef8b08ed46.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 683.586775] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e97e0521-7a25-4297-903b-e20555506bc1 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.594435] env[65726]: DEBUG oslo_vmware.api [None req-efa812b8-b008-4879-b549-6ccf454f374f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] Waiting for the task: (returnval){ [ 683.594435] env[65726]: value = "task-5115605" [ 683.594435] env[65726]: _type = "Task" [ 683.594435] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 683.605210] env[65726]: DEBUG oslo_vmware.api [None req-efa812b8-b008-4879-b549-6ccf454f374f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] Task: {'id': task-5115605, 'name': Rename_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 683.709878] env[65726]: INFO nova.compute.manager [None req-ca40f660-02e2-41e4-8187-3f53d529f31e tempest-ServersAdmin275Test-1899430852 tempest-ServersAdmin275Test-1899430852-project-admin] [instance: 3b933684-ac19-44b0-a49d-6af45501e166] Rebuilding instance [ 683.763904] env[65726]: DEBUG nova.compute.manager [None req-ca40f660-02e2-41e4-8187-3f53d529f31e tempest-ServersAdmin275Test-1899430852 tempest-ServersAdmin275Test-1899430852-project-admin] [instance: 3b933684-ac19-44b0-a49d-6af45501e166] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 683.765364] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8018b98-5924-4b19-a5a0-a88d51857949 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.001191] env[65726]: DEBUG oslo_concurrency.lockutils [None req-bb345b26-9f49-4e52-904f-99ccc71299ac tempest-AttachInterfacesV270Test-133387050 tempest-AttachInterfacesV270Test-133387050-project-member] Releasing lock "refresh_cache-c977257e-dce0-4ecc-b42b-0f6ebab38797" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 684.001627] env[65726]: DEBUG nova.compute.manager [None req-bb345b26-9f49-4e52-904f-99ccc71299ac tempest-AttachInterfacesV270Test-133387050 tempest-AttachInterfacesV270Test-133387050-project-member] [instance: c977257e-dce0-4ecc-b42b-0f6ebab38797] Instance network_info: |[{"id": "881e11d6-dc73-4e2b-9d9b-eec7da46f63f", "address": "fa:16:3e:9d:66:49", "network": {"id": "70d88d94-31cf-4dcf-b41c-385f1939383d", "bridge": "br-int", "label": "tempest-AttachInterfacesV270Test-103265114-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "459cd960efb744ef91c113bb584444bd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "96d8be6c-b557-4b40-b0f5-838c62a3c904", "external-id": "nsx-vlan-transportzone-144", "segmentation_id": 144, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap881e11d6-dc", "ovs_interfaceid": "881e11d6-dc73-4e2b-9d9b-eec7da46f63f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 684.002132] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-bb345b26-9f49-4e52-904f-99ccc71299ac tempest-AttachInterfacesV270Test-133387050 tempest-AttachInterfacesV270Test-133387050-project-member] [instance: c977257e-dce0-4ecc-b42b-0f6ebab38797] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:9d:66:49', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '96d8be6c-b557-4b40-b0f5-838c62a3c904', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '881e11d6-dc73-4e2b-9d9b-eec7da46f63f', 'vif_model': 'vmxnet3'}] {{(pid=65726) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 684.013515] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-bb345b26-9f49-4e52-904f-99ccc71299ac tempest-AttachInterfacesV270Test-133387050 tempest-AttachInterfacesV270Test-133387050-project-member] Creating folder: Project (459cd960efb744ef91c113bb584444bd). Parent ref: group-v995008. {{(pid=65726) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 684.013826] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7f032f4b-85a5-46b6-adaf-aa2a98543bb9 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.028102] env[65726]: INFO nova.virt.vmwareapi.vm_util [None req-bb345b26-9f49-4e52-904f-99ccc71299ac tempest-AttachInterfacesV270Test-133387050 tempest-AttachInterfacesV270Test-133387050-project-member] Created folder: Project (459cd960efb744ef91c113bb584444bd) in parent group-v995008. [ 684.028484] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-bb345b26-9f49-4e52-904f-99ccc71299ac tempest-AttachInterfacesV270Test-133387050 tempest-AttachInterfacesV270Test-133387050-project-member] Creating folder: Instances. Parent ref: group-v995078. {{(pid=65726) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 684.029207] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c123fd8f-3626-4914-b44d-3d34aa737b4b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.042051] env[65726]: INFO nova.virt.vmwareapi.vm_util [None req-bb345b26-9f49-4e52-904f-99ccc71299ac tempest-AttachInterfacesV270Test-133387050 tempest-AttachInterfacesV270Test-133387050-project-member] Created folder: Instances in parent group-v995078. [ 684.042211] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-bb345b26-9f49-4e52-904f-99ccc71299ac tempest-AttachInterfacesV270Test-133387050 tempest-AttachInterfacesV270Test-133387050-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 684.042706] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c977257e-dce0-4ecc-b42b-0f6ebab38797] Creating VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 684.042831] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-56c7d8c2-ee7d-4493-926a-383e6a07477c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.073258] env[65726]: DEBUG oslo_concurrency.lockutils [None req-f72e4138-ba89-47ec-80a1-52d74fe64f4e tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.258s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 684.079648] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e9359f40-01a5-4a39-b7b9-47a0a48e2ede tempest-VolumesAdminNegativeTest-2004477850 tempest-VolumesAdminNegativeTest-2004477850-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.163s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 684.081212] env[65726]: INFO nova.compute.claims [None req-e9359f40-01a5-4a39-b7b9-47a0a48e2ede tempest-VolumesAdminNegativeTest-2004477850 tempest-VolumesAdminNegativeTest-2004477850-project-member] [instance: 30b17a17-229c-4ca7-9ae6-c67ccdcd3963] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 684.092501] env[65726]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 684.092501] env[65726]: value = "task-5115608" [ 684.092501] env[65726]: _type = "Task" [ 684.092501] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 684.107310] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115608, 'name': CreateVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 684.110817] env[65726]: DEBUG oslo_vmware.api [None req-efa812b8-b008-4879-b549-6ccf454f374f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] Task: {'id': task-5115605, 'name': Rename_Task, 'duration_secs': 0.251532} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 684.111442] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-efa812b8-b008-4879-b549-6ccf454f374f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] [instance: ddd36db0-28ad-4582-994b-7eef8b08ed46] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 684.111605] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-21a4c8d2-00e3-4a8e-b41a-34f3868f3534 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.119083] env[65726]: INFO nova.scheduler.client.report [None req-f72e4138-ba89-47ec-80a1-52d74fe64f4e tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Deleted allocations for instance 8e9c27f6-e9fa-4ed0-b6ae-3b5df52d1c44 [ 684.121938] env[65726]: DEBUG oslo_vmware.api [None req-efa812b8-b008-4879-b549-6ccf454f374f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] Waiting for the task: (returnval){ [ 684.121938] env[65726]: value = "task-5115609" [ 684.121938] env[65726]: _type = "Task" [ 684.121938] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 684.146143] env[65726]: DEBUG oslo_vmware.api [None req-efa812b8-b008-4879-b549-6ccf454f374f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] Task: {'id': task-5115609, 'name': PowerOnVM_Task} progress is 33%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 684.269204] env[65726]: DEBUG oslo_concurrency.lockutils [None req-78fe5da2-df9b-4290-851e-bd53c863c8dd tempest-ServerGroupTestJSON-125182767 tempest-ServerGroupTestJSON-125182767-project-member] Acquiring lock "c2929727-a7ea-4f63-997d-474c29a305fc" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 684.269560] env[65726]: DEBUG oslo_concurrency.lockutils [None req-78fe5da2-df9b-4290-851e-bd53c863c8dd tempest-ServerGroupTestJSON-125182767 tempest-ServerGroupTestJSON-125182767-project-member] Lock "c2929727-a7ea-4f63-997d-474c29a305fc" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 684.269728] env[65726]: DEBUG oslo_concurrency.lockutils [None req-78fe5da2-df9b-4290-851e-bd53c863c8dd tempest-ServerGroupTestJSON-125182767 tempest-ServerGroupTestJSON-125182767-project-member] Acquiring lock "c2929727-a7ea-4f63-997d-474c29a305fc-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 684.269968] env[65726]: DEBUG oslo_concurrency.lockutils [None req-78fe5da2-df9b-4290-851e-bd53c863c8dd tempest-ServerGroupTestJSON-125182767 tempest-ServerGroupTestJSON-125182767-project-member] Lock "c2929727-a7ea-4f63-997d-474c29a305fc-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 684.270401] env[65726]: DEBUG oslo_concurrency.lockutils [None req-78fe5da2-df9b-4290-851e-bd53c863c8dd tempest-ServerGroupTestJSON-125182767 tempest-ServerGroupTestJSON-125182767-project-member] Lock "c2929727-a7ea-4f63-997d-474c29a305fc-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 684.273131] env[65726]: INFO nova.compute.manager [None req-78fe5da2-df9b-4290-851e-bd53c863c8dd tempest-ServerGroupTestJSON-125182767 tempest-ServerGroupTestJSON-125182767-project-member] [instance: c2929727-a7ea-4f63-997d-474c29a305fc] Terminating instance [ 684.605478] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115608, 'name': CreateVM_Task, 'duration_secs': 0.491225} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 684.605478] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c977257e-dce0-4ecc-b42b-0f6ebab38797] Created VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 684.606800] env[65726]: WARNING openstack [None req-bb345b26-9f49-4e52-904f-99ccc71299ac tempest-AttachInterfacesV270Test-133387050 tempest-AttachInterfacesV270Test-133387050-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 684.607309] env[65726]: WARNING openstack [None req-bb345b26-9f49-4e52-904f-99ccc71299ac tempest-AttachInterfacesV270Test-133387050 tempest-AttachInterfacesV270Test-133387050-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 684.612877] env[65726]: DEBUG oslo_concurrency.lockutils [None req-bb345b26-9f49-4e52-904f-99ccc71299ac tempest-AttachInterfacesV270Test-133387050 tempest-AttachInterfacesV270Test-133387050-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 684.613062] env[65726]: DEBUG oslo_concurrency.lockutils [None req-bb345b26-9f49-4e52-904f-99ccc71299ac tempest-AttachInterfacesV270Test-133387050 tempest-AttachInterfacesV270Test-133387050-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 684.613417] env[65726]: DEBUG oslo_concurrency.lockutils [None req-bb345b26-9f49-4e52-904f-99ccc71299ac tempest-AttachInterfacesV270Test-133387050 tempest-AttachInterfacesV270Test-133387050-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 684.614120] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-83e5924a-0393-4007-a4d1-c84ece2b6cc0 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.621091] env[65726]: DEBUG oslo_vmware.api [None req-bb345b26-9f49-4e52-904f-99ccc71299ac tempest-AttachInterfacesV270Test-133387050 tempest-AttachInterfacesV270Test-133387050-project-member] Waiting for the task: (returnval){ [ 684.621091] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52f48d83-4477-c947-4ee5-ef679fbb67ad" [ 684.621091] env[65726]: _type = "Task" [ 684.621091] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 684.634585] env[65726]: DEBUG oslo_vmware.api [None req-bb345b26-9f49-4e52-904f-99ccc71299ac tempest-AttachInterfacesV270Test-133387050 tempest-AttachInterfacesV270Test-133387050-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52f48d83-4477-c947-4ee5-ef679fbb67ad, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 684.635129] env[65726]: DEBUG oslo_concurrency.lockutils [None req-f72e4138-ba89-47ec-80a1-52d74fe64f4e tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Lock "8e9c27f6-e9fa-4ed0-b6ae-3b5df52d1c44" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 26.722s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 684.646452] env[65726]: DEBUG oslo_vmware.api [None req-efa812b8-b008-4879-b549-6ccf454f374f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] Task: {'id': task-5115609, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 684.777948] env[65726]: DEBUG nova.compute.manager [None req-78fe5da2-df9b-4290-851e-bd53c863c8dd tempest-ServerGroupTestJSON-125182767 tempest-ServerGroupTestJSON-125182767-project-member] [instance: c2929727-a7ea-4f63-997d-474c29a305fc] Start destroying the instance on the hypervisor. {{(pid=65726) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 684.777948] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-78fe5da2-df9b-4290-851e-bd53c863c8dd tempest-ServerGroupTestJSON-125182767 tempest-ServerGroupTestJSON-125182767-project-member] [instance: c2929727-a7ea-4f63-997d-474c29a305fc] Destroying instance {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 684.778705] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c914211-9df5-4c10-89d9-899104d8f0ae {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.782396] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-ca40f660-02e2-41e4-8187-3f53d529f31e tempest-ServersAdmin275Test-1899430852 tempest-ServersAdmin275Test-1899430852-project-admin] [instance: 3b933684-ac19-44b0-a49d-6af45501e166] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 684.782663] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b8008809-bc62-495f-b76a-050077928e1d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.789407] env[65726]: DEBUG nova.compute.manager [req-23f97b8c-7998-4a4d-b0c8-2a3a2f1c40b6 req-e0b2b303-71c3-430a-985c-3d6960366fad service nova] [instance: 88fb7235-cd0e-49dc-9d78-3583d5d3f528] Received event network-vif-deleted-b915a591-ee09-4390-ae01-09778bd3f2c3 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 684.794665] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-78fe5da2-df9b-4290-851e-bd53c863c8dd tempest-ServerGroupTestJSON-125182767 tempest-ServerGroupTestJSON-125182767-project-member] [instance: c2929727-a7ea-4f63-997d-474c29a305fc] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 684.794949] env[65726]: DEBUG oslo_vmware.api [None req-ca40f660-02e2-41e4-8187-3f53d529f31e tempest-ServersAdmin275Test-1899430852 tempest-ServersAdmin275Test-1899430852-project-admin] Waiting for the task: (returnval){ [ 684.794949] env[65726]: value = "task-5115610" [ 684.794949] env[65726]: _type = "Task" [ 684.794949] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 684.795256] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0788f0b5-4540-4a5d-9736-d1c07e355abe {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.807522] env[65726]: DEBUG oslo_vmware.api [None req-ca40f660-02e2-41e4-8187-3f53d529f31e tempest-ServersAdmin275Test-1899430852 tempest-ServersAdmin275Test-1899430852-project-admin] Task: {'id': task-5115610, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 684.809146] env[65726]: DEBUG oslo_vmware.api [None req-78fe5da2-df9b-4290-851e-bd53c863c8dd tempest-ServerGroupTestJSON-125182767 tempest-ServerGroupTestJSON-125182767-project-member] Waiting for the task: (returnval){ [ 684.809146] env[65726]: value = "task-5115611" [ 684.809146] env[65726]: _type = "Task" [ 684.809146] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 684.818827] env[65726]: DEBUG oslo_vmware.api [None req-78fe5da2-df9b-4290-851e-bd53c863c8dd tempest-ServerGroupTestJSON-125182767 tempest-ServerGroupTestJSON-125182767-project-member] Task: {'id': task-5115611, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 685.132248] env[65726]: DEBUG oslo_vmware.api [None req-bb345b26-9f49-4e52-904f-99ccc71299ac tempest-AttachInterfacesV270Test-133387050 tempest-AttachInterfacesV270Test-133387050-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52f48d83-4477-c947-4ee5-ef679fbb67ad, 'name': SearchDatastore_Task, 'duration_secs': 0.020035} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 685.135465] env[65726]: DEBUG oslo_concurrency.lockutils [None req-bb345b26-9f49-4e52-904f-99ccc71299ac tempest-AttachInterfacesV270Test-133387050 tempest-AttachInterfacesV270Test-133387050-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 685.135642] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-bb345b26-9f49-4e52-904f-99ccc71299ac tempest-AttachInterfacesV270Test-133387050 tempest-AttachInterfacesV270Test-133387050-project-member] [instance: c977257e-dce0-4ecc-b42b-0f6ebab38797] Processing image b52362a3-ee8a-4cbf-b06f-513b0cc8f95c {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 685.135829] env[65726]: DEBUG oslo_concurrency.lockutils [None req-bb345b26-9f49-4e52-904f-99ccc71299ac tempest-AttachInterfacesV270Test-133387050 tempest-AttachInterfacesV270Test-133387050-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 685.136653] env[65726]: DEBUG oslo_concurrency.lockutils [None req-bb345b26-9f49-4e52-904f-99ccc71299ac tempest-AttachInterfacesV270Test-133387050 tempest-AttachInterfacesV270Test-133387050-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 685.136653] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-bb345b26-9f49-4e52-904f-99ccc71299ac tempest-AttachInterfacesV270Test-133387050 tempest-AttachInterfacesV270Test-133387050-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 685.139872] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b80ffb28-4055-43a7-a2e2-2a09ed4446a6 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.148409] env[65726]: DEBUG oslo_vmware.api [None req-efa812b8-b008-4879-b549-6ccf454f374f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] Task: {'id': task-5115609, 'name': PowerOnVM_Task, 'duration_secs': 0.634343} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 685.148688] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-efa812b8-b008-4879-b549-6ccf454f374f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] [instance: ddd36db0-28ad-4582-994b-7eef8b08ed46] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 685.148958] env[65726]: INFO nova.compute.manager [None req-efa812b8-b008-4879-b549-6ccf454f374f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] [instance: ddd36db0-28ad-4582-994b-7eef8b08ed46] Took 6.47 seconds to spawn the instance on the hypervisor. [ 685.149130] env[65726]: DEBUG nova.compute.manager [None req-efa812b8-b008-4879-b549-6ccf454f374f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] [instance: ddd36db0-28ad-4582-994b-7eef8b08ed46] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 685.149984] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a74b252c-c7b2-417b-9e91-8963abe35b4c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.153741] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-bb345b26-9f49-4e52-904f-99ccc71299ac tempest-AttachInterfacesV270Test-133387050 tempest-AttachInterfacesV270Test-133387050-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 685.153950] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-bb345b26-9f49-4e52-904f-99ccc71299ac tempest-AttachInterfacesV270Test-133387050 tempest-AttachInterfacesV270Test-133387050-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65726) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 685.157767] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fe23074e-7c15-4cce-9c98-8cc5fdfb48d6 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.170530] env[65726]: DEBUG oslo_vmware.api [None req-bb345b26-9f49-4e52-904f-99ccc71299ac tempest-AttachInterfacesV270Test-133387050 tempest-AttachInterfacesV270Test-133387050-project-member] Waiting for the task: (returnval){ [ 685.170530] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]525ecb35-da87-0ade-c21a-60c6410c1625" [ 685.170530] env[65726]: _type = "Task" [ 685.170530] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 685.179579] env[65726]: DEBUG oslo_vmware.api [None req-bb345b26-9f49-4e52-904f-99ccc71299ac tempest-AttachInterfacesV270Test-133387050 tempest-AttachInterfacesV270Test-133387050-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]525ecb35-da87-0ade-c21a-60c6410c1625, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 685.311685] env[65726]: DEBUG oslo_vmware.api [None req-ca40f660-02e2-41e4-8187-3f53d529f31e tempest-ServersAdmin275Test-1899430852 tempest-ServersAdmin275Test-1899430852-project-admin] Task: {'id': task-5115610, 'name': PowerOffVM_Task, 'duration_secs': 0.211989} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 685.315902] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-ca40f660-02e2-41e4-8187-3f53d529f31e tempest-ServersAdmin275Test-1899430852 tempest-ServersAdmin275Test-1899430852-project-admin] [instance: 3b933684-ac19-44b0-a49d-6af45501e166] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 685.316029] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-ca40f660-02e2-41e4-8187-3f53d529f31e tempest-ServersAdmin275Test-1899430852 tempest-ServersAdmin275Test-1899430852-project-admin] [instance: 3b933684-ac19-44b0-a49d-6af45501e166] Destroying instance {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 685.317173] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb45e3e2-9e93-42e2-acf5-3f6e28763d09 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.328973] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-ca40f660-02e2-41e4-8187-3f53d529f31e tempest-ServersAdmin275Test-1899430852 tempest-ServersAdmin275Test-1899430852-project-admin] [instance: 3b933684-ac19-44b0-a49d-6af45501e166] Unregistering the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 685.329298] env[65726]: DEBUG oslo_vmware.api [None req-78fe5da2-df9b-4290-851e-bd53c863c8dd tempest-ServerGroupTestJSON-125182767 tempest-ServerGroupTestJSON-125182767-project-member] Task: {'id': task-5115611, 'name': PowerOffVM_Task, 'duration_secs': 0.212605} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 685.329511] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ab0cb75e-741a-40d3-bb2b-6e3c151d13f5 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.331278] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-78fe5da2-df9b-4290-851e-bd53c863c8dd tempest-ServerGroupTestJSON-125182767 tempest-ServerGroupTestJSON-125182767-project-member] [instance: c2929727-a7ea-4f63-997d-474c29a305fc] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 685.331468] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-78fe5da2-df9b-4290-851e-bd53c863c8dd tempest-ServerGroupTestJSON-125182767 tempest-ServerGroupTestJSON-125182767-project-member] [instance: c2929727-a7ea-4f63-997d-474c29a305fc] Unregistering the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 685.331716] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-aa36f601-bc6b-4c2b-98f9-65441ef3a1d6 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.365684] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-ca40f660-02e2-41e4-8187-3f53d529f31e tempest-ServersAdmin275Test-1899430852 tempest-ServersAdmin275Test-1899430852-project-admin] [instance: 3b933684-ac19-44b0-a49d-6af45501e166] Unregistered the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 685.365912] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-ca40f660-02e2-41e4-8187-3f53d529f31e tempest-ServersAdmin275Test-1899430852 tempest-ServersAdmin275Test-1899430852-project-admin] [instance: 3b933684-ac19-44b0-a49d-6af45501e166] Deleting contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 685.366144] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-ca40f660-02e2-41e4-8187-3f53d529f31e tempest-ServersAdmin275Test-1899430852 tempest-ServersAdmin275Test-1899430852-project-admin] Deleting the datastore file [datastore1] 3b933684-ac19-44b0-a49d-6af45501e166 {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 685.366407] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f18c9e0d-5592-484e-a1f2-c68c04a39efd {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.378289] env[65726]: DEBUG oslo_vmware.api [None req-ca40f660-02e2-41e4-8187-3f53d529f31e tempest-ServersAdmin275Test-1899430852 tempest-ServersAdmin275Test-1899430852-project-admin] Waiting for the task: (returnval){ [ 685.378289] env[65726]: value = "task-5115614" [ 685.378289] env[65726]: _type = "Task" [ 685.378289] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 685.387641] env[65726]: DEBUG oslo_vmware.api [None req-ca40f660-02e2-41e4-8187-3f53d529f31e tempest-ServersAdmin275Test-1899430852 tempest-ServersAdmin275Test-1899430852-project-admin] Task: {'id': task-5115614, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 685.403387] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-78fe5da2-df9b-4290-851e-bd53c863c8dd tempest-ServerGroupTestJSON-125182767 tempest-ServerGroupTestJSON-125182767-project-member] [instance: c2929727-a7ea-4f63-997d-474c29a305fc] Unregistered the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 685.403719] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-78fe5da2-df9b-4290-851e-bd53c863c8dd tempest-ServerGroupTestJSON-125182767 tempest-ServerGroupTestJSON-125182767-project-member] [instance: c2929727-a7ea-4f63-997d-474c29a305fc] Deleting contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 685.403801] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-78fe5da2-df9b-4290-851e-bd53c863c8dd tempest-ServerGroupTestJSON-125182767 tempest-ServerGroupTestJSON-125182767-project-member] Deleting the datastore file [datastore1] c2929727-a7ea-4f63-997d-474c29a305fc {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 685.404134] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0200c614-080f-4947-8826-d542b11721cc {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.412375] env[65726]: DEBUG oslo_vmware.api [None req-78fe5da2-df9b-4290-851e-bd53c863c8dd tempest-ServerGroupTestJSON-125182767 tempest-ServerGroupTestJSON-125182767-project-member] Waiting for the task: (returnval){ [ 685.412375] env[65726]: value = "task-5115615" [ 685.412375] env[65726]: _type = "Task" [ 685.412375] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 685.424032] env[65726]: DEBUG oslo_vmware.api [None req-78fe5da2-df9b-4290-851e-bd53c863c8dd tempest-ServerGroupTestJSON-125182767 tempest-ServerGroupTestJSON-125182767-project-member] Task: {'id': task-5115615, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 685.659425] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04b81537-1c81-47cc-9a23-213c6248bdcb {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.669267] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09c47538-89d2-4305-a14a-0b2847953baf {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.684627] env[65726]: INFO nova.compute.manager [None req-efa812b8-b008-4879-b549-6ccf454f374f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] [instance: ddd36db0-28ad-4582-994b-7eef8b08ed46] Took 26.26 seconds to build instance. [ 685.730877] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b232d6c0-52b9-473d-8d7c-d304dda0be01 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.733900] env[65726]: DEBUG oslo_vmware.api [None req-bb345b26-9f49-4e52-904f-99ccc71299ac tempest-AttachInterfacesV270Test-133387050 tempest-AttachInterfacesV270Test-133387050-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]525ecb35-da87-0ade-c21a-60c6410c1625, 'name': SearchDatastore_Task, 'duration_secs': 0.0335} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 685.735704] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-beef6be7-1601-42b8-b34b-ea140beda2fd {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.743178] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e377336-c94e-4eb3-8e69-e54a928e4710 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.749469] env[65726]: DEBUG oslo_vmware.api [None req-bb345b26-9f49-4e52-904f-99ccc71299ac tempest-AttachInterfacesV270Test-133387050 tempest-AttachInterfacesV270Test-133387050-project-member] Waiting for the task: (returnval){ [ 685.749469] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]528b8fb2-701b-8d8f-88d4-e99ee768ff3c" [ 685.749469] env[65726]: _type = "Task" [ 685.749469] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 685.761138] env[65726]: DEBUG nova.compute.provider_tree [None req-e9359f40-01a5-4a39-b7b9-47a0a48e2ede tempest-VolumesAdminNegativeTest-2004477850 tempest-VolumesAdminNegativeTest-2004477850-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 685.770068] env[65726]: DEBUG oslo_vmware.api [None req-bb345b26-9f49-4e52-904f-99ccc71299ac tempest-AttachInterfacesV270Test-133387050 tempest-AttachInterfacesV270Test-133387050-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]528b8fb2-701b-8d8f-88d4-e99ee768ff3c, 'name': SearchDatastore_Task, 'duration_secs': 0.018698} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 685.770325] env[65726]: DEBUG oslo_concurrency.lockutils [None req-bb345b26-9f49-4e52-904f-99ccc71299ac tempest-AttachInterfacesV270Test-133387050 tempest-AttachInterfacesV270Test-133387050-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 685.770604] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-bb345b26-9f49-4e52-904f-99ccc71299ac tempest-AttachInterfacesV270Test-133387050 tempest-AttachInterfacesV270Test-133387050-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore2] c977257e-dce0-4ecc-b42b-0f6ebab38797/c977257e-dce0-4ecc-b42b-0f6ebab38797.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 685.770882] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-addf1393-9b5c-4a54-bf94-15555f41d426 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.779565] env[65726]: DEBUG oslo_vmware.api [None req-bb345b26-9f49-4e52-904f-99ccc71299ac tempest-AttachInterfacesV270Test-133387050 tempest-AttachInterfacesV270Test-133387050-project-member] Waiting for the task: (returnval){ [ 685.779565] env[65726]: value = "task-5115616" [ 685.779565] env[65726]: _type = "Task" [ 685.779565] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 685.791059] env[65726]: DEBUG oslo_vmware.api [None req-bb345b26-9f49-4e52-904f-99ccc71299ac tempest-AttachInterfacesV270Test-133387050 tempest-AttachInterfacesV270Test-133387050-project-member] Task: {'id': task-5115616, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 685.889666] env[65726]: DEBUG oslo_vmware.api [None req-ca40f660-02e2-41e4-8187-3f53d529f31e tempest-ServersAdmin275Test-1899430852 tempest-ServersAdmin275Test-1899430852-project-admin] Task: {'id': task-5115614, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.144186} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 685.889982] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-ca40f660-02e2-41e4-8187-3f53d529f31e tempest-ServersAdmin275Test-1899430852 tempest-ServersAdmin275Test-1899430852-project-admin] Deleted the datastore file {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 685.890209] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-ca40f660-02e2-41e4-8187-3f53d529f31e tempest-ServersAdmin275Test-1899430852 tempest-ServersAdmin275Test-1899430852-project-admin] [instance: 3b933684-ac19-44b0-a49d-6af45501e166] Deleted contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 685.890439] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-ca40f660-02e2-41e4-8187-3f53d529f31e tempest-ServersAdmin275Test-1899430852 tempest-ServersAdmin275Test-1899430852-project-admin] [instance: 3b933684-ac19-44b0-a49d-6af45501e166] Instance destroyed {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 685.924909] env[65726]: DEBUG oslo_vmware.api [None req-78fe5da2-df9b-4290-851e-bd53c863c8dd tempest-ServerGroupTestJSON-125182767 tempest-ServerGroupTestJSON-125182767-project-member] Task: {'id': task-5115615, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.306106} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 685.924995] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-78fe5da2-df9b-4290-851e-bd53c863c8dd tempest-ServerGroupTestJSON-125182767 tempest-ServerGroupTestJSON-125182767-project-member] Deleted the datastore file {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 685.925218] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-78fe5da2-df9b-4290-851e-bd53c863c8dd tempest-ServerGroupTestJSON-125182767 tempest-ServerGroupTestJSON-125182767-project-member] [instance: c2929727-a7ea-4f63-997d-474c29a305fc] Deleted contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 685.925908] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-78fe5da2-df9b-4290-851e-bd53c863c8dd tempest-ServerGroupTestJSON-125182767 tempest-ServerGroupTestJSON-125182767-project-member] [instance: c2929727-a7ea-4f63-997d-474c29a305fc] Instance destroyed {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 685.925908] env[65726]: INFO nova.compute.manager [None req-78fe5da2-df9b-4290-851e-bd53c863c8dd tempest-ServerGroupTestJSON-125182767 tempest-ServerGroupTestJSON-125182767-project-member] [instance: c2929727-a7ea-4f63-997d-474c29a305fc] Took 1.15 seconds to destroy the instance on the hypervisor. [ 685.926066] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-78fe5da2-df9b-4290-851e-bd53c863c8dd tempest-ServerGroupTestJSON-125182767 tempest-ServerGroupTestJSON-125182767-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 685.926234] env[65726]: DEBUG nova.compute.manager [-] [instance: c2929727-a7ea-4f63-997d-474c29a305fc] Deallocating network for instance {{(pid=65726) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 685.926332] env[65726]: DEBUG nova.network.neutron [-] [instance: c2929727-a7ea-4f63-997d-474c29a305fc] deallocate_for_instance() {{(pid=65726) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 685.927202] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 685.927304] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 686.084504] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 686.084504] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 686.231185] env[65726]: DEBUG oslo_concurrency.lockutils [None req-efa812b8-b008-4879-b549-6ccf454f374f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] Lock "ddd36db0-28ad-4582-994b-7eef8b08ed46" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 48.067s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 686.267447] env[65726]: DEBUG nova.scheduler.client.report [None req-e9359f40-01a5-4a39-b7b9-47a0a48e2ede tempest-VolumesAdminNegativeTest-2004477850 tempest-VolumesAdminNegativeTest-2004477850-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 686.293867] env[65726]: DEBUG oslo_vmware.api [None req-bb345b26-9f49-4e52-904f-99ccc71299ac tempest-AttachInterfacesV270Test-133387050 tempest-AttachInterfacesV270Test-133387050-project-member] Task: {'id': task-5115616, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 686.737392] env[65726]: DEBUG nova.compute.manager [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] [instance: 02351120-8385-4403-8464-a154f3da9380] Starting instance... {{(pid=65726) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 686.770403] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e9359f40-01a5-4a39-b7b9-47a0a48e2ede tempest-VolumesAdminNegativeTest-2004477850 tempest-VolumesAdminNegativeTest-2004477850-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.691s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 686.773019] env[65726]: DEBUG nova.compute.manager [None req-e9359f40-01a5-4a39-b7b9-47a0a48e2ede tempest-VolumesAdminNegativeTest-2004477850 tempest-VolumesAdminNegativeTest-2004477850-project-member] [instance: 30b17a17-229c-4ca7-9ae6-c67ccdcd3963] Start building networks asynchronously for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 686.773981] env[65726]: DEBUG oslo_concurrency.lockutils [None req-60511481-93fa-4f8b-817d-01a577f85a26 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.792s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 686.776050] env[65726]: INFO nova.compute.claims [None req-60511481-93fa-4f8b-817d-01a577f85a26 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: 49bcab1f-7b64-4999-abff-37771c58a271] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 686.797011] env[65726]: DEBUG oslo_vmware.api [None req-bb345b26-9f49-4e52-904f-99ccc71299ac tempest-AttachInterfacesV270Test-133387050 tempest-AttachInterfacesV270Test-133387050-project-member] Task: {'id': task-5115616, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.816379} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 686.797326] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-bb345b26-9f49-4e52-904f-99ccc71299ac tempest-AttachInterfacesV270Test-133387050 tempest-AttachInterfacesV270Test-133387050-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore2] c977257e-dce0-4ecc-b42b-0f6ebab38797/c977257e-dce0-4ecc-b42b-0f6ebab38797.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 686.797533] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-bb345b26-9f49-4e52-904f-99ccc71299ac tempest-AttachInterfacesV270Test-133387050 tempest-AttachInterfacesV270Test-133387050-project-member] [instance: c977257e-dce0-4ecc-b42b-0f6ebab38797] Extending root virtual disk to 1048576 {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 686.797815] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-cdf125ac-4aa9-4814-a3b2-762ab32b66f4 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.810968] env[65726]: DEBUG oslo_vmware.api [None req-bb345b26-9f49-4e52-904f-99ccc71299ac tempest-AttachInterfacesV270Test-133387050 tempest-AttachInterfacesV270Test-133387050-project-member] Waiting for the task: (returnval){ [ 686.810968] env[65726]: value = "task-5115617" [ 686.810968] env[65726]: _type = "Task" [ 686.810968] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 686.824466] env[65726]: DEBUG oslo_vmware.api [None req-bb345b26-9f49-4e52-904f-99ccc71299ac tempest-AttachInterfacesV270Test-133387050 tempest-AttachInterfacesV270Test-133387050-project-member] Task: {'id': task-5115617, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 686.900933] env[65726]: DEBUG nova.network.neutron [-] [instance: c2929727-a7ea-4f63-997d-474c29a305fc] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 686.932537] env[65726]: DEBUG nova.virt.hardware [None req-ca40f660-02e2-41e4-8187-3f53d529f31e tempest-ServersAdmin275Test-1899430852 tempest-ServersAdmin275Test-1899430852-project-admin] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T19:28:34Z,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1539f06fa3534e90acbc3a60c4b8bd3f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T19:28:35Z,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 686.932895] env[65726]: DEBUG nova.virt.hardware [None req-ca40f660-02e2-41e4-8187-3f53d529f31e tempest-ServersAdmin275Test-1899430852 tempest-ServersAdmin275Test-1899430852-project-admin] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 686.933072] env[65726]: DEBUG nova.virt.hardware [None req-ca40f660-02e2-41e4-8187-3f53d529f31e tempest-ServersAdmin275Test-1899430852 tempest-ServersAdmin275Test-1899430852-project-admin] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 686.933312] env[65726]: DEBUG nova.virt.hardware [None req-ca40f660-02e2-41e4-8187-3f53d529f31e tempest-ServersAdmin275Test-1899430852 tempest-ServersAdmin275Test-1899430852-project-admin] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 686.933461] env[65726]: DEBUG nova.virt.hardware [None req-ca40f660-02e2-41e4-8187-3f53d529f31e tempest-ServersAdmin275Test-1899430852 tempest-ServersAdmin275Test-1899430852-project-admin] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 686.933600] env[65726]: DEBUG nova.virt.hardware [None req-ca40f660-02e2-41e4-8187-3f53d529f31e tempest-ServersAdmin275Test-1899430852 tempest-ServersAdmin275Test-1899430852-project-admin] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 686.933841] env[65726]: DEBUG nova.virt.hardware [None req-ca40f660-02e2-41e4-8187-3f53d529f31e tempest-ServersAdmin275Test-1899430852 tempest-ServersAdmin275Test-1899430852-project-admin] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 686.934049] env[65726]: DEBUG nova.virt.hardware [None req-ca40f660-02e2-41e4-8187-3f53d529f31e tempest-ServersAdmin275Test-1899430852 tempest-ServersAdmin275Test-1899430852-project-admin] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 686.934232] env[65726]: DEBUG nova.virt.hardware [None req-ca40f660-02e2-41e4-8187-3f53d529f31e tempest-ServersAdmin275Test-1899430852 tempest-ServersAdmin275Test-1899430852-project-admin] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 686.934390] env[65726]: DEBUG nova.virt.hardware [None req-ca40f660-02e2-41e4-8187-3f53d529f31e tempest-ServersAdmin275Test-1899430852 tempest-ServersAdmin275Test-1899430852-project-admin] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 686.934557] env[65726]: DEBUG nova.virt.hardware [None req-ca40f660-02e2-41e4-8187-3f53d529f31e tempest-ServersAdmin275Test-1899430852 tempest-ServersAdmin275Test-1899430852-project-admin] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 686.935810] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab93558f-0fec-4ba0-ac0f-fef4e87e9af2 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.946145] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e9e0a9b-dd64-48f6-bbb2-e2ad1bf3e437 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.966012] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-ca40f660-02e2-41e4-8187-3f53d529f31e tempest-ServersAdmin275Test-1899430852 tempest-ServersAdmin275Test-1899430852-project-admin] [instance: 3b933684-ac19-44b0-a49d-6af45501e166] Instance VIF info [] {{(pid=65726) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 686.971878] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-ca40f660-02e2-41e4-8187-3f53d529f31e tempest-ServersAdmin275Test-1899430852 tempest-ServersAdmin275Test-1899430852-project-admin] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 686.972216] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3b933684-ac19-44b0-a49d-6af45501e166] Creating VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 686.972460] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-046fb541-0ce0-4be5-8173-92148b78121b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.990131] env[65726]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 686.990131] env[65726]: value = "task-5115618" [ 686.990131] env[65726]: _type = "Task" [ 686.990131] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 686.998782] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115618, 'name': CreateVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 687.162194] env[65726]: DEBUG nova.compute.manager [req-2de94fe8-7d07-45cb-8f64-c377d87dd2c1 req-07d1f21e-6f78-41b7-be4e-9a7418b493b9 service nova] [instance: c977257e-dce0-4ecc-b42b-0f6ebab38797] Received event network-changed-881e11d6-dc73-4e2b-9d9b-eec7da46f63f {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 687.163184] env[65726]: DEBUG nova.compute.manager [req-2de94fe8-7d07-45cb-8f64-c377d87dd2c1 req-07d1f21e-6f78-41b7-be4e-9a7418b493b9 service nova] [instance: c977257e-dce0-4ecc-b42b-0f6ebab38797] Refreshing instance network info cache due to event network-changed-881e11d6-dc73-4e2b-9d9b-eec7da46f63f. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 687.166464] env[65726]: DEBUG oslo_concurrency.lockutils [req-2de94fe8-7d07-45cb-8f64-c377d87dd2c1 req-07d1f21e-6f78-41b7-be4e-9a7418b493b9 service nova] Acquiring lock "refresh_cache-c977257e-dce0-4ecc-b42b-0f6ebab38797" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 687.166464] env[65726]: DEBUG oslo_concurrency.lockutils [req-2de94fe8-7d07-45cb-8f64-c377d87dd2c1 req-07d1f21e-6f78-41b7-be4e-9a7418b493b9 service nova] Acquired lock "refresh_cache-c977257e-dce0-4ecc-b42b-0f6ebab38797" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 687.166464] env[65726]: DEBUG nova.network.neutron [req-2de94fe8-7d07-45cb-8f64-c377d87dd2c1 req-07d1f21e-6f78-41b7-be4e-9a7418b493b9 service nova] [instance: c977257e-dce0-4ecc-b42b-0f6ebab38797] Refreshing network info cache for port 881e11d6-dc73-4e2b-9d9b-eec7da46f63f {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 687.196171] env[65726]: DEBUG oslo_concurrency.lockutils [None req-48153719-e51d-407a-98ef-93401c6270f2 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Acquiring lock "65aaa2eb-3ca5-4c2b-92f9-b524366c10dd" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 687.196409] env[65726]: DEBUG oslo_concurrency.lockutils [None req-48153719-e51d-407a-98ef-93401c6270f2 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Lock "65aaa2eb-3ca5-4c2b-92f9-b524366c10dd" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 687.268102] env[65726]: DEBUG oslo_concurrency.lockutils [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 687.286742] env[65726]: DEBUG nova.compute.utils [None req-e9359f40-01a5-4a39-b7b9-47a0a48e2ede tempest-VolumesAdminNegativeTest-2004477850 tempest-VolumesAdminNegativeTest-2004477850-project-member] Using /dev/sd instead of None {{(pid=65726) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 687.288577] env[65726]: DEBUG nova.compute.manager [None req-e9359f40-01a5-4a39-b7b9-47a0a48e2ede tempest-VolumesAdminNegativeTest-2004477850 tempest-VolumesAdminNegativeTest-2004477850-project-member] [instance: 30b17a17-229c-4ca7-9ae6-c67ccdcd3963] Allocating IP information in the background. {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 687.289112] env[65726]: DEBUG nova.network.neutron [None req-e9359f40-01a5-4a39-b7b9-47a0a48e2ede tempest-VolumesAdminNegativeTest-2004477850 tempest-VolumesAdminNegativeTest-2004477850-project-member] [instance: 30b17a17-229c-4ca7-9ae6-c67ccdcd3963] allocate_for_instance() {{(pid=65726) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 687.290451] env[65726]: WARNING neutronclient.v2_0.client [None req-e9359f40-01a5-4a39-b7b9-47a0a48e2ede tempest-VolumesAdminNegativeTest-2004477850 tempest-VolumesAdminNegativeTest-2004477850-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 687.290904] env[65726]: WARNING neutronclient.v2_0.client [None req-e9359f40-01a5-4a39-b7b9-47a0a48e2ede tempest-VolumesAdminNegativeTest-2004477850 tempest-VolumesAdminNegativeTest-2004477850-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 687.291756] env[65726]: WARNING openstack [None req-e9359f40-01a5-4a39-b7b9-47a0a48e2ede tempest-VolumesAdminNegativeTest-2004477850 tempest-VolumesAdminNegativeTest-2004477850-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 687.292521] env[65726]: WARNING openstack [None req-e9359f40-01a5-4a39-b7b9-47a0a48e2ede tempest-VolumesAdminNegativeTest-2004477850 tempest-VolumesAdminNegativeTest-2004477850-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 687.329571] env[65726]: DEBUG oslo_vmware.api [None req-bb345b26-9f49-4e52-904f-99ccc71299ac tempest-AttachInterfacesV270Test-133387050 tempest-AttachInterfacesV270Test-133387050-project-member] Task: {'id': task-5115617, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.112504} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 687.330420] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-bb345b26-9f49-4e52-904f-99ccc71299ac tempest-AttachInterfacesV270Test-133387050 tempest-AttachInterfacesV270Test-133387050-project-member] [instance: c977257e-dce0-4ecc-b42b-0f6ebab38797] Extended root virtual disk {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 687.332648] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e049431e-1ccb-4c71-86df-588f5946f4eb {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.381743] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-bb345b26-9f49-4e52-904f-99ccc71299ac tempest-AttachInterfacesV270Test-133387050 tempest-AttachInterfacesV270Test-133387050-project-member] [instance: c977257e-dce0-4ecc-b42b-0f6ebab38797] Reconfiguring VM instance instance-0000001a to attach disk [datastore2] c977257e-dce0-4ecc-b42b-0f6ebab38797/c977257e-dce0-4ecc-b42b-0f6ebab38797.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 687.387510] env[65726]: DEBUG nova.policy [None req-e9359f40-01a5-4a39-b7b9-47a0a48e2ede tempest-VolumesAdminNegativeTest-2004477850 tempest-VolumesAdminNegativeTest-2004477850-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3e8b52b61d684e98a02a512f0390c72e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '33d7ed4a125d4cc2ae93fbe4085f66d7', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65726) authorize /opt/stack/nova/nova/policy.py:192}} [ 687.392821] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-93f9e52f-af01-44c2-aab3-c7bde9064f1a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.420190] env[65726]: INFO nova.compute.manager [-] [instance: c2929727-a7ea-4f63-997d-474c29a305fc] Took 1.49 seconds to deallocate network for instance. [ 687.430789] env[65726]: DEBUG oslo_vmware.api [None req-bb345b26-9f49-4e52-904f-99ccc71299ac tempest-AttachInterfacesV270Test-133387050 tempest-AttachInterfacesV270Test-133387050-project-member] Waiting for the task: (returnval){ [ 687.430789] env[65726]: value = "task-5115619" [ 687.430789] env[65726]: _type = "Task" [ 687.430789] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 687.443882] env[65726]: DEBUG oslo_vmware.api [None req-bb345b26-9f49-4e52-904f-99ccc71299ac tempest-AttachInterfacesV270Test-133387050 tempest-AttachInterfacesV270Test-133387050-project-member] Task: {'id': task-5115619, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 687.501490] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115618, 'name': CreateVM_Task, 'duration_secs': 0.456454} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 687.501697] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3b933684-ac19-44b0-a49d-6af45501e166] Created VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 687.502103] env[65726]: DEBUG oslo_concurrency.lockutils [None req-ca40f660-02e2-41e4-8187-3f53d529f31e tempest-ServersAdmin275Test-1899430852 tempest-ServersAdmin275Test-1899430852-project-admin] Acquiring lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 687.502253] env[65726]: DEBUG oslo_concurrency.lockutils [None req-ca40f660-02e2-41e4-8187-3f53d529f31e tempest-ServersAdmin275Test-1899430852 tempest-ServersAdmin275Test-1899430852-project-admin] Acquired lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 687.502625] env[65726]: DEBUG oslo_concurrency.lockutils [None req-ca40f660-02e2-41e4-8187-3f53d529f31e tempest-ServersAdmin275Test-1899430852 tempest-ServersAdmin275Test-1899430852-project-admin] Acquired external semaphore "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 687.502989] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b38857c0-32f6-418d-86a9-9c410b6e24f8 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.508774] env[65726]: DEBUG oslo_vmware.api [None req-ca40f660-02e2-41e4-8187-3f53d529f31e tempest-ServersAdmin275Test-1899430852 tempest-ServersAdmin275Test-1899430852-project-admin] Waiting for the task: (returnval){ [ 687.508774] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]528aea2d-b9b2-419f-9af5-67ced9d1069d" [ 687.508774] env[65726]: _type = "Task" [ 687.508774] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 687.518102] env[65726]: DEBUG oslo_vmware.api [None req-ca40f660-02e2-41e4-8187-3f53d529f31e tempest-ServersAdmin275Test-1899430852 tempest-ServersAdmin275Test-1899430852-project-admin] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]528aea2d-b9b2-419f-9af5-67ced9d1069d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 687.668481] env[65726]: WARNING openstack [req-2de94fe8-7d07-45cb-8f64-c377d87dd2c1 req-07d1f21e-6f78-41b7-be4e-9a7418b493b9 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 687.669075] env[65726]: WARNING openstack [req-2de94fe8-7d07-45cb-8f64-c377d87dd2c1 req-07d1f21e-6f78-41b7-be4e-9a7418b493b9 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 687.679853] env[65726]: DEBUG nova.network.neutron [None req-e9359f40-01a5-4a39-b7b9-47a0a48e2ede tempest-VolumesAdminNegativeTest-2004477850 tempest-VolumesAdminNegativeTest-2004477850-project-member] [instance: 30b17a17-229c-4ca7-9ae6-c67ccdcd3963] Successfully created port: b1810a58-276a-4ccf-b223-ed03a7c987b1 {{(pid=65726) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 687.803329] env[65726]: DEBUG nova.compute.manager [None req-e9359f40-01a5-4a39-b7b9-47a0a48e2ede tempest-VolumesAdminNegativeTest-2004477850 tempest-VolumesAdminNegativeTest-2004477850-project-member] [instance: 30b17a17-229c-4ca7-9ae6-c67ccdcd3963] Start building block device mappings for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 687.825911] env[65726]: WARNING openstack [req-2de94fe8-7d07-45cb-8f64-c377d87dd2c1 req-07d1f21e-6f78-41b7-be4e-9a7418b493b9 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 687.826523] env[65726]: WARNING openstack [req-2de94fe8-7d07-45cb-8f64-c377d87dd2c1 req-07d1f21e-6f78-41b7-be4e-9a7418b493b9 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 687.929718] env[65726]: DEBUG oslo_concurrency.lockutils [None req-78fe5da2-df9b-4290-851e-bd53c863c8dd tempest-ServerGroupTestJSON-125182767 tempest-ServerGroupTestJSON-125182767-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 687.944213] env[65726]: DEBUG oslo_vmware.api [None req-bb345b26-9f49-4e52-904f-99ccc71299ac tempest-AttachInterfacesV270Test-133387050 tempest-AttachInterfacesV270Test-133387050-project-member] Task: {'id': task-5115619, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 687.955160] env[65726]: WARNING openstack [req-2de94fe8-7d07-45cb-8f64-c377d87dd2c1 req-07d1f21e-6f78-41b7-be4e-9a7418b493b9 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 687.955160] env[65726]: WARNING openstack [req-2de94fe8-7d07-45cb-8f64-c377d87dd2c1 req-07d1f21e-6f78-41b7-be4e-9a7418b493b9 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 688.021555] env[65726]: DEBUG oslo_vmware.api [None req-ca40f660-02e2-41e4-8187-3f53d529f31e tempest-ServersAdmin275Test-1899430852 tempest-ServersAdmin275Test-1899430852-project-admin] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]528aea2d-b9b2-419f-9af5-67ced9d1069d, 'name': SearchDatastore_Task, 'duration_secs': 0.03424} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 688.023504] env[65726]: DEBUG oslo_concurrency.lockutils [None req-ca40f660-02e2-41e4-8187-3f53d529f31e tempest-ServersAdmin275Test-1899430852 tempest-ServersAdmin275Test-1899430852-project-admin] Releasing lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 688.023504] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-ca40f660-02e2-41e4-8187-3f53d529f31e tempest-ServersAdmin275Test-1899430852 tempest-ServersAdmin275Test-1899430852-project-admin] [instance: 3b933684-ac19-44b0-a49d-6af45501e166] Processing image b52362a3-ee8a-4cbf-b06f-513b0cc8f95c {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 688.023504] env[65726]: DEBUG oslo_concurrency.lockutils [None req-ca40f660-02e2-41e4-8187-3f53d529f31e tempest-ServersAdmin275Test-1899430852 tempest-ServersAdmin275Test-1899430852-project-admin] Acquiring lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 688.023504] env[65726]: DEBUG oslo_concurrency.lockutils [None req-ca40f660-02e2-41e4-8187-3f53d529f31e tempest-ServersAdmin275Test-1899430852 tempest-ServersAdmin275Test-1899430852-project-admin] Acquired lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 688.024083] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-ca40f660-02e2-41e4-8187-3f53d529f31e tempest-ServersAdmin275Test-1899430852 tempest-ServersAdmin275Test-1899430852-project-admin] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 688.024083] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-07809b1d-a4ae-4846-b834-9644028d5781 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.037892] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-ca40f660-02e2-41e4-8187-3f53d529f31e tempest-ServersAdmin275Test-1899430852 tempest-ServersAdmin275Test-1899430852-project-admin] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 688.038644] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-ca40f660-02e2-41e4-8187-3f53d529f31e tempest-ServersAdmin275Test-1899430852 tempest-ServersAdmin275Test-1899430852-project-admin] Folder [datastore1] devstack-image-cache_base created. {{(pid=65726) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 688.039443] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-218b23f2-f7f4-4f09-92a7-1df49a1e3a50 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.053759] env[65726]: DEBUG oslo_vmware.api [None req-ca40f660-02e2-41e4-8187-3f53d529f31e tempest-ServersAdmin275Test-1899430852 tempest-ServersAdmin275Test-1899430852-project-admin] Waiting for the task: (returnval){ [ 688.053759] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52c4235b-201c-7120-4873-97e40dbcb28d" [ 688.053759] env[65726]: _type = "Task" [ 688.053759] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 688.067599] env[65726]: DEBUG oslo_vmware.api [None req-ca40f660-02e2-41e4-8187-3f53d529f31e tempest-ServersAdmin275Test-1899430852 tempest-ServersAdmin275Test-1899430852-project-admin] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52c4235b-201c-7120-4873-97e40dbcb28d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 688.106157] env[65726]: DEBUG nova.network.neutron [req-2de94fe8-7d07-45cb-8f64-c377d87dd2c1 req-07d1f21e-6f78-41b7-be4e-9a7418b493b9 service nova] [instance: c977257e-dce0-4ecc-b42b-0f6ebab38797] Updated VIF entry in instance network info cache for port 881e11d6-dc73-4e2b-9d9b-eec7da46f63f. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 688.106527] env[65726]: DEBUG nova.network.neutron [req-2de94fe8-7d07-45cb-8f64-c377d87dd2c1 req-07d1f21e-6f78-41b7-be4e-9a7418b493b9 service nova] [instance: c977257e-dce0-4ecc-b42b-0f6ebab38797] Updating instance_info_cache with network_info: [{"id": "881e11d6-dc73-4e2b-9d9b-eec7da46f63f", "address": "fa:16:3e:9d:66:49", "network": {"id": "70d88d94-31cf-4dcf-b41c-385f1939383d", "bridge": "br-int", "label": "tempest-AttachInterfacesV270Test-103265114-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "459cd960efb744ef91c113bb584444bd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "96d8be6c-b557-4b40-b0f5-838c62a3c904", "external-id": "nsx-vlan-transportzone-144", "segmentation_id": 144, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap881e11d6-dc", "ovs_interfaceid": "881e11d6-dc73-4e2b-9d9b-eec7da46f63f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 688.450233] env[65726]: DEBUG oslo_vmware.api [None req-bb345b26-9f49-4e52-904f-99ccc71299ac tempest-AttachInterfacesV270Test-133387050 tempest-AttachInterfacesV270Test-133387050-project-member] Task: {'id': task-5115619, 'name': ReconfigVM_Task, 'duration_secs': 0.835906} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 688.450543] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-bb345b26-9f49-4e52-904f-99ccc71299ac tempest-AttachInterfacesV270Test-133387050 tempest-AttachInterfacesV270Test-133387050-project-member] [instance: c977257e-dce0-4ecc-b42b-0f6ebab38797] Reconfigured VM instance instance-0000001a to attach disk [datastore2] c977257e-dce0-4ecc-b42b-0f6ebab38797/c977257e-dce0-4ecc-b42b-0f6ebab38797.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 688.451204] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-35e706b1-100d-468e-8d75-598d9a1c2680 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.464502] env[65726]: DEBUG oslo_vmware.api [None req-bb345b26-9f49-4e52-904f-99ccc71299ac tempest-AttachInterfacesV270Test-133387050 tempest-AttachInterfacesV270Test-133387050-project-member] Waiting for the task: (returnval){ [ 688.464502] env[65726]: value = "task-5115620" [ 688.464502] env[65726]: _type = "Task" [ 688.464502] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 688.476065] env[65726]: DEBUG oslo_vmware.api [None req-bb345b26-9f49-4e52-904f-99ccc71299ac tempest-AttachInterfacesV270Test-133387050 tempest-AttachInterfacesV270Test-133387050-project-member] Task: {'id': task-5115620, 'name': Rename_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 688.477686] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8b7138d-f978-465c-a19e-ee32e1e4002d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.486662] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6c61353-43ed-439b-b22a-870e40d8d7b3 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.522046] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-899b19dd-e7d8-483d-a224-cd1e6744bf5c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.530109] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3d87898-267f-473a-ad53-05a666205088 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.546287] env[65726]: DEBUG nova.compute.provider_tree [None req-60511481-93fa-4f8b-817d-01a577f85a26 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 688.565691] env[65726]: DEBUG oslo_vmware.api [None req-ca40f660-02e2-41e4-8187-3f53d529f31e tempest-ServersAdmin275Test-1899430852 tempest-ServersAdmin275Test-1899430852-project-admin] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52c4235b-201c-7120-4873-97e40dbcb28d, 'name': SearchDatastore_Task, 'duration_secs': 0.02193} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 688.568376] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-eccdd1fd-21f4-41eb-aa77-5ea9eeba5907 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.574778] env[65726]: DEBUG oslo_vmware.api [None req-ca40f660-02e2-41e4-8187-3f53d529f31e tempest-ServersAdmin275Test-1899430852 tempest-ServersAdmin275Test-1899430852-project-admin] Waiting for the task: (returnval){ [ 688.574778] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5212820c-f5c3-aacb-3ffd-b031295bf635" [ 688.574778] env[65726]: _type = "Task" [ 688.574778] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 688.587844] env[65726]: DEBUG oslo_vmware.api [None req-ca40f660-02e2-41e4-8187-3f53d529f31e tempest-ServersAdmin275Test-1899430852 tempest-ServersAdmin275Test-1899430852-project-admin] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5212820c-f5c3-aacb-3ffd-b031295bf635, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 688.609932] env[65726]: DEBUG oslo_concurrency.lockutils [req-2de94fe8-7d07-45cb-8f64-c377d87dd2c1 req-07d1f21e-6f78-41b7-be4e-9a7418b493b9 service nova] Releasing lock "refresh_cache-c977257e-dce0-4ecc-b42b-0f6ebab38797" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 688.818810] env[65726]: DEBUG nova.compute.manager [None req-e9359f40-01a5-4a39-b7b9-47a0a48e2ede tempest-VolumesAdminNegativeTest-2004477850 tempest-VolumesAdminNegativeTest-2004477850-project-member] [instance: 30b17a17-229c-4ca7-9ae6-c67ccdcd3963] Start spawning the instance on the hypervisor. {{(pid=65726) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 688.842747] env[65726]: DEBUG nova.virt.hardware [None req-e9359f40-01a5-4a39-b7b9-47a0a48e2ede tempest-VolumesAdminNegativeTest-2004477850 tempest-VolumesAdminNegativeTest-2004477850-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T19:28:34Z,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1539f06fa3534e90acbc3a60c4b8bd3f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T19:28:35Z,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 688.842747] env[65726]: DEBUG nova.virt.hardware [None req-e9359f40-01a5-4a39-b7b9-47a0a48e2ede tempest-VolumesAdminNegativeTest-2004477850 tempest-VolumesAdminNegativeTest-2004477850-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 688.842747] env[65726]: DEBUG nova.virt.hardware [None req-e9359f40-01a5-4a39-b7b9-47a0a48e2ede tempest-VolumesAdminNegativeTest-2004477850 tempest-VolumesAdminNegativeTest-2004477850-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 688.842990] env[65726]: DEBUG nova.virt.hardware [None req-e9359f40-01a5-4a39-b7b9-47a0a48e2ede tempest-VolumesAdminNegativeTest-2004477850 tempest-VolumesAdminNegativeTest-2004477850-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 688.842990] env[65726]: DEBUG nova.virt.hardware [None req-e9359f40-01a5-4a39-b7b9-47a0a48e2ede tempest-VolumesAdminNegativeTest-2004477850 tempest-VolumesAdminNegativeTest-2004477850-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 688.842990] env[65726]: DEBUG nova.virt.hardware [None req-e9359f40-01a5-4a39-b7b9-47a0a48e2ede tempest-VolumesAdminNegativeTest-2004477850 tempest-VolumesAdminNegativeTest-2004477850-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 688.842990] env[65726]: DEBUG nova.virt.hardware [None req-e9359f40-01a5-4a39-b7b9-47a0a48e2ede tempest-VolumesAdminNegativeTest-2004477850 tempest-VolumesAdminNegativeTest-2004477850-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 688.842990] env[65726]: DEBUG nova.virt.hardware [None req-e9359f40-01a5-4a39-b7b9-47a0a48e2ede tempest-VolumesAdminNegativeTest-2004477850 tempest-VolumesAdminNegativeTest-2004477850-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 688.843187] env[65726]: DEBUG nova.virt.hardware [None req-e9359f40-01a5-4a39-b7b9-47a0a48e2ede tempest-VolumesAdminNegativeTest-2004477850 tempest-VolumesAdminNegativeTest-2004477850-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 688.843367] env[65726]: DEBUG nova.virt.hardware [None req-e9359f40-01a5-4a39-b7b9-47a0a48e2ede tempest-VolumesAdminNegativeTest-2004477850 tempest-VolumesAdminNegativeTest-2004477850-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 688.843541] env[65726]: DEBUG nova.virt.hardware [None req-e9359f40-01a5-4a39-b7b9-47a0a48e2ede tempest-VolumesAdminNegativeTest-2004477850 tempest-VolumesAdminNegativeTest-2004477850-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 688.844420] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-048b6e3c-7004-4347-a022-9a73dc369e3c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.853992] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-324a0e75-683f-44a1-8ea4-79aa8f821500 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.859155] env[65726]: DEBUG nova.objects.instance [None req-d234de18-f385-4dcc-bc17-774bb6452dfb tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Lazy-loading 'flavor' on Instance uuid 703fff2d-5aec-4ed1-b987-9e1cc3a25f67 {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 688.979975] env[65726]: DEBUG oslo_vmware.api [None req-bb345b26-9f49-4e52-904f-99ccc71299ac tempest-AttachInterfacesV270Test-133387050 tempest-AttachInterfacesV270Test-133387050-project-member] Task: {'id': task-5115620, 'name': Rename_Task, 'duration_secs': 0.17112} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 688.980407] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-bb345b26-9f49-4e52-904f-99ccc71299ac tempest-AttachInterfacesV270Test-133387050 tempest-AttachInterfacesV270Test-133387050-project-member] [instance: c977257e-dce0-4ecc-b42b-0f6ebab38797] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 688.980675] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d2435b33-f74c-4fcf-a9c2-6cead9511e46 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.987977] env[65726]: DEBUG oslo_vmware.api [None req-bb345b26-9f49-4e52-904f-99ccc71299ac tempest-AttachInterfacesV270Test-133387050 tempest-AttachInterfacesV270Test-133387050-project-member] Waiting for the task: (returnval){ [ 688.987977] env[65726]: value = "task-5115621" [ 688.987977] env[65726]: _type = "Task" [ 688.987977] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 688.997922] env[65726]: DEBUG oslo_vmware.api [None req-bb345b26-9f49-4e52-904f-99ccc71299ac tempest-AttachInterfacesV270Test-133387050 tempest-AttachInterfacesV270Test-133387050-project-member] Task: {'id': task-5115621, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 689.048193] env[65726]: DEBUG nova.scheduler.client.report [None req-60511481-93fa-4f8b-817d-01a577f85a26 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 689.093132] env[65726]: DEBUG oslo_vmware.api [None req-ca40f660-02e2-41e4-8187-3f53d529f31e tempest-ServersAdmin275Test-1899430852 tempest-ServersAdmin275Test-1899430852-project-admin] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5212820c-f5c3-aacb-3ffd-b031295bf635, 'name': SearchDatastore_Task, 'duration_secs': 0.023597} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 689.093132] env[65726]: DEBUG oslo_concurrency.lockutils [None req-ca40f660-02e2-41e4-8187-3f53d529f31e tempest-ServersAdmin275Test-1899430852 tempest-ServersAdmin275Test-1899430852-project-admin] Releasing lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 689.093132] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-ca40f660-02e2-41e4-8187-3f53d529f31e tempest-ServersAdmin275Test-1899430852 tempest-ServersAdmin275Test-1899430852-project-admin] Copying Virtual Disk [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] 3b933684-ac19-44b0-a49d-6af45501e166/3b933684-ac19-44b0-a49d-6af45501e166.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 689.093132] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-51223199-5c8c-4acb-8e45-58c9cb38d67a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.100751] env[65726]: DEBUG oslo_vmware.api [None req-ca40f660-02e2-41e4-8187-3f53d529f31e tempest-ServersAdmin275Test-1899430852 tempest-ServersAdmin275Test-1899430852-project-admin] Waiting for the task: (returnval){ [ 689.100751] env[65726]: value = "task-5115622" [ 689.100751] env[65726]: _type = "Task" [ 689.100751] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 689.112742] env[65726]: DEBUG oslo_vmware.api [None req-ca40f660-02e2-41e4-8187-3f53d529f31e tempest-ServersAdmin275Test-1899430852 tempest-ServersAdmin275Test-1899430852-project-admin] Task: {'id': task-5115622, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 689.365369] env[65726]: DEBUG oslo_concurrency.lockutils [None req-d234de18-f385-4dcc-bc17-774bb6452dfb tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Acquiring lock "refresh_cache-703fff2d-5aec-4ed1-b987-9e1cc3a25f67" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 689.365555] env[65726]: DEBUG oslo_concurrency.lockutils [None req-d234de18-f385-4dcc-bc17-774bb6452dfb tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Acquired lock "refresh_cache-703fff2d-5aec-4ed1-b987-9e1cc3a25f67" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 689.365840] env[65726]: DEBUG nova.network.neutron [None req-d234de18-f385-4dcc-bc17-774bb6452dfb tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] [instance: 703fff2d-5aec-4ed1-b987-9e1cc3a25f67] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 689.366102] env[65726]: DEBUG nova.objects.instance [None req-d234de18-f385-4dcc-bc17-774bb6452dfb tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Lazy-loading 'info_cache' on Instance uuid 703fff2d-5aec-4ed1-b987-9e1cc3a25f67 {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 689.413325] env[65726]: DEBUG nova.network.neutron [None req-e9359f40-01a5-4a39-b7b9-47a0a48e2ede tempest-VolumesAdminNegativeTest-2004477850 tempest-VolumesAdminNegativeTest-2004477850-project-member] [instance: 30b17a17-229c-4ca7-9ae6-c67ccdcd3963] Successfully updated port: b1810a58-276a-4ccf-b223-ed03a7c987b1 {{(pid=65726) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 689.502396] env[65726]: DEBUG oslo_vmware.api [None req-bb345b26-9f49-4e52-904f-99ccc71299ac tempest-AttachInterfacesV270Test-133387050 tempest-AttachInterfacesV270Test-133387050-project-member] Task: {'id': task-5115621, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 689.554741] env[65726]: DEBUG oslo_concurrency.lockutils [None req-60511481-93fa-4f8b-817d-01a577f85a26 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.781s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 689.555438] env[65726]: DEBUG nova.compute.manager [None req-60511481-93fa-4f8b-817d-01a577f85a26 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: 49bcab1f-7b64-4999-abff-37771c58a271] Start building networks asynchronously for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 689.559547] env[65726]: DEBUG oslo_concurrency.lockutils [None req-f3878371-4599-418c-a420-84e54e420a24 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.494s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 689.561591] env[65726]: INFO nova.compute.claims [None req-f3878371-4599-418c-a420-84e54e420a24 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] [instance: 423af936-a383-4652-8887-9c8b8f507909] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 689.612266] env[65726]: DEBUG oslo_vmware.api [None req-ca40f660-02e2-41e4-8187-3f53d529f31e tempest-ServersAdmin275Test-1899430852 tempest-ServersAdmin275Test-1899430852-project-admin] Task: {'id': task-5115622, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 689.866351] env[65726]: DEBUG nova.compute.manager [req-5c9f8f4a-b998-4469-bcb6-75a83a9e60cb req-663d0498-5468-4958-9211-4b97de47ab44 service nova] [instance: c2929727-a7ea-4f63-997d-474c29a305fc] Received event network-vif-deleted-27cddbbc-b8ac-4cf0-8baf-d03643c3de8d {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 689.869974] env[65726]: DEBUG nova.objects.base [None req-d234de18-f385-4dcc-bc17-774bb6452dfb tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Object Instance<703fff2d-5aec-4ed1-b987-9e1cc3a25f67> lazy-loaded attributes: flavor,info_cache {{(pid=65726) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 689.923083] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e9359f40-01a5-4a39-b7b9-47a0a48e2ede tempest-VolumesAdminNegativeTest-2004477850 tempest-VolumesAdminNegativeTest-2004477850-project-member] Acquiring lock "refresh_cache-30b17a17-229c-4ca7-9ae6-c67ccdcd3963" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 689.923083] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e9359f40-01a5-4a39-b7b9-47a0a48e2ede tempest-VolumesAdminNegativeTest-2004477850 tempest-VolumesAdminNegativeTest-2004477850-project-member] Acquired lock "refresh_cache-30b17a17-229c-4ca7-9ae6-c67ccdcd3963" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 689.923431] env[65726]: DEBUG nova.network.neutron [None req-e9359f40-01a5-4a39-b7b9-47a0a48e2ede tempest-VolumesAdminNegativeTest-2004477850 tempest-VolumesAdminNegativeTest-2004477850-project-member] [instance: 30b17a17-229c-4ca7-9ae6-c67ccdcd3963] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 689.995297] env[65726]: DEBUG nova.compute.manager [req-585b8189-1772-4196-9ce2-e3bc8fa5e2b4 req-93bce518-fb1b-4317-b3b7-5ce16d59d910 service nova] [instance: 30b17a17-229c-4ca7-9ae6-c67ccdcd3963] Received event network-vif-plugged-b1810a58-276a-4ccf-b223-ed03a7c987b1 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 689.995506] env[65726]: DEBUG oslo_concurrency.lockutils [req-585b8189-1772-4196-9ce2-e3bc8fa5e2b4 req-93bce518-fb1b-4317-b3b7-5ce16d59d910 service nova] Acquiring lock "30b17a17-229c-4ca7-9ae6-c67ccdcd3963-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 689.995755] env[65726]: DEBUG oslo_concurrency.lockutils [req-585b8189-1772-4196-9ce2-e3bc8fa5e2b4 req-93bce518-fb1b-4317-b3b7-5ce16d59d910 service nova] Lock "30b17a17-229c-4ca7-9ae6-c67ccdcd3963-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 689.995922] env[65726]: DEBUG oslo_concurrency.lockutils [req-585b8189-1772-4196-9ce2-e3bc8fa5e2b4 req-93bce518-fb1b-4317-b3b7-5ce16d59d910 service nova] Lock "30b17a17-229c-4ca7-9ae6-c67ccdcd3963-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 689.996098] env[65726]: DEBUG nova.compute.manager [req-585b8189-1772-4196-9ce2-e3bc8fa5e2b4 req-93bce518-fb1b-4317-b3b7-5ce16d59d910 service nova] [instance: 30b17a17-229c-4ca7-9ae6-c67ccdcd3963] No waiting events found dispatching network-vif-plugged-b1810a58-276a-4ccf-b223-ed03a7c987b1 {{(pid=65726) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 689.996332] env[65726]: WARNING nova.compute.manager [req-585b8189-1772-4196-9ce2-e3bc8fa5e2b4 req-93bce518-fb1b-4317-b3b7-5ce16d59d910 service nova] [instance: 30b17a17-229c-4ca7-9ae6-c67ccdcd3963] Received unexpected event network-vif-plugged-b1810a58-276a-4ccf-b223-ed03a7c987b1 for instance with vm_state building and task_state spawning. [ 690.003237] env[65726]: DEBUG oslo_vmware.api [None req-bb345b26-9f49-4e52-904f-99ccc71299ac tempest-AttachInterfacesV270Test-133387050 tempest-AttachInterfacesV270Test-133387050-project-member] Task: {'id': task-5115621, 'name': PowerOnVM_Task, 'duration_secs': 0.550425} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 690.003574] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-bb345b26-9f49-4e52-904f-99ccc71299ac tempest-AttachInterfacesV270Test-133387050 tempest-AttachInterfacesV270Test-133387050-project-member] [instance: c977257e-dce0-4ecc-b42b-0f6ebab38797] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 690.004040] env[65726]: INFO nova.compute.manager [None req-bb345b26-9f49-4e52-904f-99ccc71299ac tempest-AttachInterfacesV270Test-133387050 tempest-AttachInterfacesV270Test-133387050-project-member] [instance: c977257e-dce0-4ecc-b42b-0f6ebab38797] Took 8.47 seconds to spawn the instance on the hypervisor. [ 690.004108] env[65726]: DEBUG nova.compute.manager [None req-bb345b26-9f49-4e52-904f-99ccc71299ac tempest-AttachInterfacesV270Test-133387050 tempest-AttachInterfacesV270Test-133387050-project-member] [instance: c977257e-dce0-4ecc-b42b-0f6ebab38797] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 690.006086] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6cd2c994-2032-4042-b8a1-778d9da58837 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.068082] env[65726]: DEBUG nova.compute.utils [None req-60511481-93fa-4f8b-817d-01a577f85a26 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Using /dev/sd instead of None {{(pid=65726) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 690.072753] env[65726]: DEBUG nova.compute.manager [None req-60511481-93fa-4f8b-817d-01a577f85a26 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: 49bcab1f-7b64-4999-abff-37771c58a271] Allocating IP information in the background. {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 690.073105] env[65726]: DEBUG nova.network.neutron [None req-60511481-93fa-4f8b-817d-01a577f85a26 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: 49bcab1f-7b64-4999-abff-37771c58a271] allocate_for_instance() {{(pid=65726) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 690.073473] env[65726]: WARNING neutronclient.v2_0.client [None req-60511481-93fa-4f8b-817d-01a577f85a26 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 690.073890] env[65726]: WARNING neutronclient.v2_0.client [None req-60511481-93fa-4f8b-817d-01a577f85a26 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 690.074784] env[65726]: WARNING openstack [None req-60511481-93fa-4f8b-817d-01a577f85a26 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 690.075507] env[65726]: WARNING openstack [None req-60511481-93fa-4f8b-817d-01a577f85a26 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 690.113073] env[65726]: DEBUG oslo_vmware.api [None req-ca40f660-02e2-41e4-8187-3f53d529f31e tempest-ServersAdmin275Test-1899430852 tempest-ServersAdmin275Test-1899430852-project-admin] Task: {'id': task-5115622, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.637929} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 690.114104] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-ca40f660-02e2-41e4-8187-3f53d529f31e tempest-ServersAdmin275Test-1899430852 tempest-ServersAdmin275Test-1899430852-project-admin] Copied Virtual Disk [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] 3b933684-ac19-44b0-a49d-6af45501e166/3b933684-ac19-44b0-a49d-6af45501e166.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 690.114403] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-ca40f660-02e2-41e4-8187-3f53d529f31e tempest-ServersAdmin275Test-1899430852 tempest-ServersAdmin275Test-1899430852-project-admin] [instance: 3b933684-ac19-44b0-a49d-6af45501e166] Extending root virtual disk to 1048576 {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 690.117442] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-975b85ca-bbd6-4f77-aa8e-6d18a33e384d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.127279] env[65726]: DEBUG oslo_vmware.api [None req-ca40f660-02e2-41e4-8187-3f53d529f31e tempest-ServersAdmin275Test-1899430852 tempest-ServersAdmin275Test-1899430852-project-admin] Waiting for the task: (returnval){ [ 690.127279] env[65726]: value = "task-5115623" [ 690.127279] env[65726]: _type = "Task" [ 690.127279] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 690.140018] env[65726]: DEBUG oslo_vmware.api [None req-ca40f660-02e2-41e4-8187-3f53d529f31e tempest-ServersAdmin275Test-1899430852 tempest-ServersAdmin275Test-1899430852-project-admin] Task: {'id': task-5115623, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 690.149849] env[65726]: DEBUG nova.policy [None req-60511481-93fa-4f8b-817d-01a577f85a26 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9b65b9194933469c908c7cc04478444b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd46ccb00794f458b85da4a93879139ee', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65726) authorize /opt/stack/nova/nova/policy.py:192}} [ 690.373562] env[65726]: WARNING openstack [None req-d234de18-f385-4dcc-bc17-774bb6452dfb tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 690.374804] env[65726]: WARNING openstack [None req-d234de18-f385-4dcc-bc17-774bb6452dfb tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 690.426109] env[65726]: WARNING openstack [None req-e9359f40-01a5-4a39-b7b9-47a0a48e2ede tempest-VolumesAdminNegativeTest-2004477850 tempest-VolumesAdminNegativeTest-2004477850-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 690.426396] env[65726]: WARNING openstack [None req-e9359f40-01a5-4a39-b7b9-47a0a48e2ede tempest-VolumesAdminNegativeTest-2004477850 tempest-VolumesAdminNegativeTest-2004477850-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 690.434610] env[65726]: DEBUG nova.network.neutron [None req-e9359f40-01a5-4a39-b7b9-47a0a48e2ede tempest-VolumesAdminNegativeTest-2004477850 tempest-VolumesAdminNegativeTest-2004477850-project-member] [instance: 30b17a17-229c-4ca7-9ae6-c67ccdcd3963] Instance cache missing network info. {{(pid=65726) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 690.447859] env[65726]: DEBUG nova.network.neutron [None req-60511481-93fa-4f8b-817d-01a577f85a26 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: 49bcab1f-7b64-4999-abff-37771c58a271] Successfully created port: 45612b7e-c052-4346-a779-5f94f9adc084 {{(pid=65726) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 690.524424] env[65726]: INFO nova.compute.manager [None req-bb345b26-9f49-4e52-904f-99ccc71299ac tempest-AttachInterfacesV270Test-133387050 tempest-AttachInterfacesV270Test-133387050-project-member] [instance: c977257e-dce0-4ecc-b42b-0f6ebab38797] Took 30.42 seconds to build instance. [ 690.590306] env[65726]: DEBUG nova.compute.manager [None req-60511481-93fa-4f8b-817d-01a577f85a26 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: 49bcab1f-7b64-4999-abff-37771c58a271] Start building block device mappings for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 690.640851] env[65726]: DEBUG oslo_vmware.api [None req-ca40f660-02e2-41e4-8187-3f53d529f31e tempest-ServersAdmin275Test-1899430852 tempest-ServersAdmin275Test-1899430852-project-admin] Task: {'id': task-5115623, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.12954} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 690.647376] env[65726]: WARNING openstack [None req-e9359f40-01a5-4a39-b7b9-47a0a48e2ede tempest-VolumesAdminNegativeTest-2004477850 tempest-VolumesAdminNegativeTest-2004477850-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 690.647446] env[65726]: WARNING openstack [None req-e9359f40-01a5-4a39-b7b9-47a0a48e2ede tempest-VolumesAdminNegativeTest-2004477850 tempest-VolumesAdminNegativeTest-2004477850-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 690.656451] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-ca40f660-02e2-41e4-8187-3f53d529f31e tempest-ServersAdmin275Test-1899430852 tempest-ServersAdmin275Test-1899430852-project-admin] [instance: 3b933684-ac19-44b0-a49d-6af45501e166] Extended root virtual disk {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 690.657282] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-782534ae-723e-471a-8dae-a52afc4d0f2c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.673562] env[65726]: WARNING openstack [None req-d234de18-f385-4dcc-bc17-774bb6452dfb tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 690.674096] env[65726]: WARNING openstack [None req-d234de18-f385-4dcc-bc17-774bb6452dfb tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 690.693173] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-ca40f660-02e2-41e4-8187-3f53d529f31e tempest-ServersAdmin275Test-1899430852 tempest-ServersAdmin275Test-1899430852-project-admin] [instance: 3b933684-ac19-44b0-a49d-6af45501e166] Reconfiguring VM instance instance-00000013 to attach disk [datastore1] 3b933684-ac19-44b0-a49d-6af45501e166/3b933684-ac19-44b0-a49d-6af45501e166.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 690.697767] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f9edb8dd-9ef4-4918-a09d-076be8290c61 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.720206] env[65726]: DEBUG oslo_vmware.api [None req-ca40f660-02e2-41e4-8187-3f53d529f31e tempest-ServersAdmin275Test-1899430852 tempest-ServersAdmin275Test-1899430852-project-admin] Waiting for the task: (returnval){ [ 690.720206] env[65726]: value = "task-5115624" [ 690.720206] env[65726]: _type = "Task" [ 690.720206] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 690.732263] env[65726]: DEBUG oslo_vmware.api [None req-ca40f660-02e2-41e4-8187-3f53d529f31e tempest-ServersAdmin275Test-1899430852 tempest-ServersAdmin275Test-1899430852-project-admin] Task: {'id': task-5115624, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 690.784957] env[65726]: WARNING openstack [None req-e9359f40-01a5-4a39-b7b9-47a0a48e2ede tempest-VolumesAdminNegativeTest-2004477850 tempest-VolumesAdminNegativeTest-2004477850-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 690.785095] env[65726]: WARNING openstack [None req-e9359f40-01a5-4a39-b7b9-47a0a48e2ede tempest-VolumesAdminNegativeTest-2004477850 tempest-VolumesAdminNegativeTest-2004477850-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 690.808866] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-883ef344-8557-4b26-a1e0-a2229d2e4c2c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.819815] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9558092-9da9-4bbb-aa43-abb28e895b4b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.858497] env[65726]: WARNING openstack [None req-d234de18-f385-4dcc-bc17-774bb6452dfb tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 690.858907] env[65726]: WARNING openstack [None req-d234de18-f385-4dcc-bc17-774bb6452dfb tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 690.867614] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-128057b6-31a0-4db5-96f0-5a08d1ce91bf {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.881719] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7468bab7-7e86-4a39-a7fa-23181d3fef89 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.897597] env[65726]: DEBUG nova.compute.provider_tree [None req-f3878371-4599-418c-a420-84e54e420a24 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 691.027250] env[65726]: DEBUG oslo_concurrency.lockutils [None req-bb345b26-9f49-4e52-904f-99ccc71299ac tempest-AttachInterfacesV270Test-133387050 tempest-AttachInterfacesV270Test-133387050-project-member] Lock "c977257e-dce0-4ecc-b42b-0f6ebab38797" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 50.421s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 691.182163] env[65726]: DEBUG nova.network.neutron [None req-e9359f40-01a5-4a39-b7b9-47a0a48e2ede tempest-VolumesAdminNegativeTest-2004477850 tempest-VolumesAdminNegativeTest-2004477850-project-member] [instance: 30b17a17-229c-4ca7-9ae6-c67ccdcd3963] Updating instance_info_cache with network_info: [{"id": "b1810a58-276a-4ccf-b223-ed03a7c987b1", "address": "fa:16:3e:b4:fe:bf", "network": {"id": "f9d993c3-5487-4b97-89c3-5f30cdf9c387", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-133871862-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "33d7ed4a125d4cc2ae93fbe4085f66d7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "aa8c2f93-f287-41b3-adb6-4942a7ea2a0b", "external-id": "nsx-vlan-transportzone-363", "segmentation_id": 363, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb1810a58-27", "ovs_interfaceid": "b1810a58-276a-4ccf-b223-ed03a7c987b1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 691.232868] env[65726]: DEBUG oslo_vmware.api [None req-ca40f660-02e2-41e4-8187-3f53d529f31e tempest-ServersAdmin275Test-1899430852 tempest-ServersAdmin275Test-1899430852-project-admin] Task: {'id': task-5115624, 'name': ReconfigVM_Task, 'duration_secs': 0.438299} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 691.232868] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-ca40f660-02e2-41e4-8187-3f53d529f31e tempest-ServersAdmin275Test-1899430852 tempest-ServersAdmin275Test-1899430852-project-admin] [instance: 3b933684-ac19-44b0-a49d-6af45501e166] Reconfigured VM instance instance-00000013 to attach disk [datastore1] 3b933684-ac19-44b0-a49d-6af45501e166/3b933684-ac19-44b0-a49d-6af45501e166.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 691.233189] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ad02f529-5742-4073-a50f-7bd1716d2fba {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.241222] env[65726]: INFO nova.compute.manager [None req-e30832c9-8006-4ee6-8d81-96e2fa3ed74f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] [instance: ddd36db0-28ad-4582-994b-7eef8b08ed46] Rebuilding instance [ 691.243569] env[65726]: DEBUG oslo_vmware.api [None req-ca40f660-02e2-41e4-8187-3f53d529f31e tempest-ServersAdmin275Test-1899430852 tempest-ServersAdmin275Test-1899430852-project-admin] Waiting for the task: (returnval){ [ 691.243569] env[65726]: value = "task-5115625" [ 691.243569] env[65726]: _type = "Task" [ 691.243569] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 691.258626] env[65726]: DEBUG oslo_vmware.api [None req-ca40f660-02e2-41e4-8187-3f53d529f31e tempest-ServersAdmin275Test-1899430852 tempest-ServersAdmin275Test-1899430852-project-admin] Task: {'id': task-5115625, 'name': Rename_Task} progress is 6%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 691.294011] env[65726]: DEBUG nova.compute.manager [None req-e30832c9-8006-4ee6-8d81-96e2fa3ed74f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] [instance: ddd36db0-28ad-4582-994b-7eef8b08ed46] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 691.295057] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b5d96bb-0aea-4265-94ba-50247bd9c3fa {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.358645] env[65726]: DEBUG nova.network.neutron [None req-d234de18-f385-4dcc-bc17-774bb6452dfb tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] [instance: 703fff2d-5aec-4ed1-b987-9e1cc3a25f67] Updating instance_info_cache with network_info: [{"id": "29330b5b-6171-4de8-bf77-6c57f6a8bd58", "address": "fa:16:3e:1d:eb:7e", "network": {"id": "f639fd12-fb47-4b3f-9895-6facebc89726", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-2112264538-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.2", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.3"}}], "meta": {"injected": false, "tenant_id": "3a680eae0e734ea8b2274b504d0a46d8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "257e5ea7-8b80-4301-9900-a754f1fe2031", "external-id": "nsx-vlan-transportzone-682", "segmentation_id": 682, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap29330b5b-61", "ovs_interfaceid": "29330b5b-6171-4de8-bf77-6c57f6a8bd58", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 691.400833] env[65726]: DEBUG nova.scheduler.client.report [None req-f3878371-4599-418c-a420-84e54e420a24 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 691.534460] env[65726]: DEBUG nova.compute.manager [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] [instance: ca67d1a4-d304-4d06-a436-f257b60ca080] Starting instance... {{(pid=65726) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 691.602131] env[65726]: DEBUG nova.compute.manager [None req-60511481-93fa-4f8b-817d-01a577f85a26 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: 49bcab1f-7b64-4999-abff-37771c58a271] Start spawning the instance on the hypervisor. {{(pid=65726) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 691.628919] env[65726]: DEBUG nova.virt.hardware [None req-60511481-93fa-4f8b-817d-01a577f85a26 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T19:28:34Z,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1539f06fa3534e90acbc3a60c4b8bd3f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T19:28:35Z,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 691.629183] env[65726]: DEBUG nova.virt.hardware [None req-60511481-93fa-4f8b-817d-01a577f85a26 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 691.629335] env[65726]: DEBUG nova.virt.hardware [None req-60511481-93fa-4f8b-817d-01a577f85a26 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 691.629624] env[65726]: DEBUG nova.virt.hardware [None req-60511481-93fa-4f8b-817d-01a577f85a26 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 691.629773] env[65726]: DEBUG nova.virt.hardware [None req-60511481-93fa-4f8b-817d-01a577f85a26 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 691.629912] env[65726]: DEBUG nova.virt.hardware [None req-60511481-93fa-4f8b-817d-01a577f85a26 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 691.630248] env[65726]: DEBUG nova.virt.hardware [None req-60511481-93fa-4f8b-817d-01a577f85a26 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 691.630454] env[65726]: DEBUG nova.virt.hardware [None req-60511481-93fa-4f8b-817d-01a577f85a26 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 691.630628] env[65726]: DEBUG nova.virt.hardware [None req-60511481-93fa-4f8b-817d-01a577f85a26 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 691.630786] env[65726]: DEBUG nova.virt.hardware [None req-60511481-93fa-4f8b-817d-01a577f85a26 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 691.630954] env[65726]: DEBUG nova.virt.hardware [None req-60511481-93fa-4f8b-817d-01a577f85a26 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 691.631877] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5f01c6b-cb92-4af3-bc9a-35648f57fe8b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.642409] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-edf1ed93-88ba-4783-958a-5731ec8b23aa {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.687028] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e9359f40-01a5-4a39-b7b9-47a0a48e2ede tempest-VolumesAdminNegativeTest-2004477850 tempest-VolumesAdminNegativeTest-2004477850-project-member] Releasing lock "refresh_cache-30b17a17-229c-4ca7-9ae6-c67ccdcd3963" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 691.687028] env[65726]: DEBUG nova.compute.manager [None req-e9359f40-01a5-4a39-b7b9-47a0a48e2ede tempest-VolumesAdminNegativeTest-2004477850 tempest-VolumesAdminNegativeTest-2004477850-project-member] [instance: 30b17a17-229c-4ca7-9ae6-c67ccdcd3963] Instance network_info: |[{"id": "b1810a58-276a-4ccf-b223-ed03a7c987b1", "address": "fa:16:3e:b4:fe:bf", "network": {"id": "f9d993c3-5487-4b97-89c3-5f30cdf9c387", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-133871862-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "33d7ed4a125d4cc2ae93fbe4085f66d7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "aa8c2f93-f287-41b3-adb6-4942a7ea2a0b", "external-id": "nsx-vlan-transportzone-363", "segmentation_id": 363, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb1810a58-27", "ovs_interfaceid": "b1810a58-276a-4ccf-b223-ed03a7c987b1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 691.687357] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-e9359f40-01a5-4a39-b7b9-47a0a48e2ede tempest-VolumesAdminNegativeTest-2004477850 tempest-VolumesAdminNegativeTest-2004477850-project-member] [instance: 30b17a17-229c-4ca7-9ae6-c67ccdcd3963] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b4:fe:bf', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'aa8c2f93-f287-41b3-adb6-4942a7ea2a0b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b1810a58-276a-4ccf-b223-ed03a7c987b1', 'vif_model': 'vmxnet3'}] {{(pid=65726) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 691.693370] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-e9359f40-01a5-4a39-b7b9-47a0a48e2ede tempest-VolumesAdminNegativeTest-2004477850 tempest-VolumesAdminNegativeTest-2004477850-project-member] Creating folder: Project (33d7ed4a125d4cc2ae93fbe4085f66d7). Parent ref: group-v995008. {{(pid=65726) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 691.693815] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-dce36f72-3265-430c-a61f-3574609cb0c9 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.707022] env[65726]: INFO nova.virt.vmwareapi.vm_util [None req-e9359f40-01a5-4a39-b7b9-47a0a48e2ede tempest-VolumesAdminNegativeTest-2004477850 tempest-VolumesAdminNegativeTest-2004477850-project-member] Created folder: Project (33d7ed4a125d4cc2ae93fbe4085f66d7) in parent group-v995008. [ 691.707022] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-e9359f40-01a5-4a39-b7b9-47a0a48e2ede tempest-VolumesAdminNegativeTest-2004477850 tempest-VolumesAdminNegativeTest-2004477850-project-member] Creating folder: Instances. Parent ref: group-v995082. {{(pid=65726) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 691.707022] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0ed7fe12-efc4-496d-8eda-b61acf0cf6b2 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.720500] env[65726]: INFO nova.virt.vmwareapi.vm_util [None req-e9359f40-01a5-4a39-b7b9-47a0a48e2ede tempest-VolumesAdminNegativeTest-2004477850 tempest-VolumesAdminNegativeTest-2004477850-project-member] Created folder: Instances in parent group-v995082. [ 691.722135] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-e9359f40-01a5-4a39-b7b9-47a0a48e2ede tempest-VolumesAdminNegativeTest-2004477850 tempest-VolumesAdminNegativeTest-2004477850-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 691.722135] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 30b17a17-229c-4ca7-9ae6-c67ccdcd3963] Creating VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 691.722135] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-fa1026c3-35dc-4497-ab41-677c18ae8936 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.749229] env[65726]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 691.749229] env[65726]: value = "task-5115628" [ 691.749229] env[65726]: _type = "Task" [ 691.749229] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 691.756029] env[65726]: DEBUG oslo_vmware.api [None req-ca40f660-02e2-41e4-8187-3f53d529f31e tempest-ServersAdmin275Test-1899430852 tempest-ServersAdmin275Test-1899430852-project-admin] Task: {'id': task-5115625, 'name': Rename_Task, 'duration_secs': 0.156835} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 691.756860] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-ca40f660-02e2-41e4-8187-3f53d529f31e tempest-ServersAdmin275Test-1899430852 tempest-ServersAdmin275Test-1899430852-project-admin] [instance: 3b933684-ac19-44b0-a49d-6af45501e166] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 691.757226] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-cd313556-3863-420b-a1f8-1d0a11a6b169 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.762723] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115628, 'name': CreateVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 691.768811] env[65726]: DEBUG oslo_vmware.api [None req-ca40f660-02e2-41e4-8187-3f53d529f31e tempest-ServersAdmin275Test-1899430852 tempest-ServersAdmin275Test-1899430852-project-admin] Waiting for the task: (returnval){ [ 691.768811] env[65726]: value = "task-5115629" [ 691.768811] env[65726]: _type = "Task" [ 691.768811] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 691.778359] env[65726]: DEBUG oslo_vmware.api [None req-ca40f660-02e2-41e4-8187-3f53d529f31e tempest-ServersAdmin275Test-1899430852 tempest-ServersAdmin275Test-1899430852-project-admin] Task: {'id': task-5115629, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 691.863178] env[65726]: DEBUG oslo_concurrency.lockutils [None req-d234de18-f385-4dcc-bc17-774bb6452dfb tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Releasing lock "refresh_cache-703fff2d-5aec-4ed1-b987-9e1cc3a25f67" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 691.906811] env[65726]: DEBUG oslo_concurrency.lockutils [None req-f3878371-4599-418c-a420-84e54e420a24 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.347s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 691.907109] env[65726]: DEBUG nova.compute.manager [None req-f3878371-4599-418c-a420-84e54e420a24 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] [instance: 423af936-a383-4652-8887-9c8b8f507909] Start building networks asynchronously for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 691.909990] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 24.094s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 691.909990] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 691.909990] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=65726) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:937}} [ 691.910275] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9b3c93a2-43e7-4858-b887-85486b2f5f2a tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 23.801s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 691.912355] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d4f8563-8cd2-47a6-be4d-b20c9dd98c6d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.923936] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3cfc72eb-0f43-45bb-baa2-053971bee5aa {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.943619] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dedf30e6-efe4-49d7-bcec-434368d76543 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.952716] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7cd9ff8-f10d-4442-acac-5ae7583d1d51 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.991269] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=179066MB free_disk=95GB free_vcpus=48 pci_devices=None {{(pid=65726) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1136}} [ 691.991439] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 692.074144] env[65726]: DEBUG oslo_concurrency.lockutils [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 692.199304] env[65726]: DEBUG nova.network.neutron [None req-60511481-93fa-4f8b-817d-01a577f85a26 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: 49bcab1f-7b64-4999-abff-37771c58a271] Successfully updated port: 45612b7e-c052-4346-a779-5f94f9adc084 {{(pid=65726) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 692.265509] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115628, 'name': CreateVM_Task, 'duration_secs': 0.430467} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 692.265864] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 30b17a17-229c-4ca7-9ae6-c67ccdcd3963] Created VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 692.266960] env[65726]: WARNING openstack [None req-e9359f40-01a5-4a39-b7b9-47a0a48e2ede tempest-VolumesAdminNegativeTest-2004477850 tempest-VolumesAdminNegativeTest-2004477850-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 692.267344] env[65726]: WARNING openstack [None req-e9359f40-01a5-4a39-b7b9-47a0a48e2ede tempest-VolumesAdminNegativeTest-2004477850 tempest-VolumesAdminNegativeTest-2004477850-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 692.276459] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e9359f40-01a5-4a39-b7b9-47a0a48e2ede tempest-VolumesAdminNegativeTest-2004477850 tempest-VolumesAdminNegativeTest-2004477850-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 692.276647] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e9359f40-01a5-4a39-b7b9-47a0a48e2ede tempest-VolumesAdminNegativeTest-2004477850 tempest-VolumesAdminNegativeTest-2004477850-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 692.277206] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e9359f40-01a5-4a39-b7b9-47a0a48e2ede tempest-VolumesAdminNegativeTest-2004477850 tempest-VolumesAdminNegativeTest-2004477850-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 692.277797] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d933d778-5a76-4313-91c1-abd728096b04 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.287927] env[65726]: DEBUG oslo_vmware.api [None req-e9359f40-01a5-4a39-b7b9-47a0a48e2ede tempest-VolumesAdminNegativeTest-2004477850 tempest-VolumesAdminNegativeTest-2004477850-project-member] Waiting for the task: (returnval){ [ 692.287927] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5220e2b8-319b-53ad-6305-ce675335aeb8" [ 692.287927] env[65726]: _type = "Task" [ 692.287927] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 692.292247] env[65726]: DEBUG oslo_vmware.api [None req-ca40f660-02e2-41e4-8187-3f53d529f31e tempest-ServersAdmin275Test-1899430852 tempest-ServersAdmin275Test-1899430852-project-admin] Task: {'id': task-5115629, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 692.302715] env[65726]: DEBUG oslo_vmware.api [None req-e9359f40-01a5-4a39-b7b9-47a0a48e2ede tempest-VolumesAdminNegativeTest-2004477850 tempest-VolumesAdminNegativeTest-2004477850-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5220e2b8-319b-53ad-6305-ce675335aeb8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 692.312590] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-e30832c9-8006-4ee6-8d81-96e2fa3ed74f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] [instance: ddd36db0-28ad-4582-994b-7eef8b08ed46] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 692.313014] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-20300897-f005-4178-9438-27a234630898 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.322118] env[65726]: DEBUG oslo_vmware.api [None req-e30832c9-8006-4ee6-8d81-96e2fa3ed74f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] Waiting for the task: (returnval){ [ 692.322118] env[65726]: value = "task-5115630" [ 692.322118] env[65726]: _type = "Task" [ 692.322118] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 692.333987] env[65726]: DEBUG oslo_vmware.api [None req-e30832c9-8006-4ee6-8d81-96e2fa3ed74f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] Task: {'id': task-5115630, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 692.414117] env[65726]: DEBUG nova.compute.utils [None req-f3878371-4599-418c-a420-84e54e420a24 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Using /dev/sd instead of None {{(pid=65726) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 692.415746] env[65726]: DEBUG nova.compute.manager [None req-f3878371-4599-418c-a420-84e54e420a24 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] [instance: 423af936-a383-4652-8887-9c8b8f507909] Allocating IP information in the background. {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 692.416337] env[65726]: DEBUG nova.network.neutron [None req-f3878371-4599-418c-a420-84e54e420a24 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] [instance: 423af936-a383-4652-8887-9c8b8f507909] allocate_for_instance() {{(pid=65726) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 692.416337] env[65726]: WARNING neutronclient.v2_0.client [None req-f3878371-4599-418c-a420-84e54e420a24 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 692.416629] env[65726]: WARNING neutronclient.v2_0.client [None req-f3878371-4599-418c-a420-84e54e420a24 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 692.419510] env[65726]: WARNING openstack [None req-f3878371-4599-418c-a420-84e54e420a24 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 692.419878] env[65726]: WARNING openstack [None req-f3878371-4599-418c-a420-84e54e420a24 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 692.432093] env[65726]: DEBUG nova.compute.manager [None req-f3878371-4599-418c-a420-84e54e420a24 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] [instance: 423af936-a383-4652-8887-9c8b8f507909] Start building block device mappings for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 692.572648] env[65726]: DEBUG nova.policy [None req-f3878371-4599-418c-a420-84e54e420a24 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e13d7ce17e914f49ac5f47aba8de2fc1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '27fdee694977439d88ed2b08abb28405', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65726) authorize /opt/stack/nova/nova/policy.py:192}} [ 692.703616] env[65726]: DEBUG oslo_concurrency.lockutils [None req-60511481-93fa-4f8b-817d-01a577f85a26 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Acquiring lock "refresh_cache-49bcab1f-7b64-4999-abff-37771c58a271" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 692.703616] env[65726]: DEBUG oslo_concurrency.lockutils [None req-60511481-93fa-4f8b-817d-01a577f85a26 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Acquired lock "refresh_cache-49bcab1f-7b64-4999-abff-37771c58a271" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 692.703775] env[65726]: DEBUG nova.network.neutron [None req-60511481-93fa-4f8b-817d-01a577f85a26 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: 49bcab1f-7b64-4999-abff-37771c58a271] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 692.794498] env[65726]: DEBUG oslo_vmware.api [None req-ca40f660-02e2-41e4-8187-3f53d529f31e tempest-ServersAdmin275Test-1899430852 tempest-ServersAdmin275Test-1899430852-project-admin] Task: {'id': task-5115629, 'name': PowerOnVM_Task, 'duration_secs': 0.549313} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 692.795159] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-ca40f660-02e2-41e4-8187-3f53d529f31e tempest-ServersAdmin275Test-1899430852 tempest-ServersAdmin275Test-1899430852-project-admin] [instance: 3b933684-ac19-44b0-a49d-6af45501e166] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 692.795439] env[65726]: DEBUG nova.compute.manager [None req-ca40f660-02e2-41e4-8187-3f53d529f31e tempest-ServersAdmin275Test-1899430852 tempest-ServersAdmin275Test-1899430852-project-admin] [instance: 3b933684-ac19-44b0-a49d-6af45501e166] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 692.799602] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86906a80-b909-485a-8578-d753be313cb7 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.813912] env[65726]: DEBUG oslo_vmware.api [None req-e9359f40-01a5-4a39-b7b9-47a0a48e2ede tempest-VolumesAdminNegativeTest-2004477850 tempest-VolumesAdminNegativeTest-2004477850-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5220e2b8-319b-53ad-6305-ce675335aeb8, 'name': SearchDatastore_Task, 'duration_secs': 0.013725} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 692.816744] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e9359f40-01a5-4a39-b7b9-47a0a48e2ede tempest-VolumesAdminNegativeTest-2004477850 tempest-VolumesAdminNegativeTest-2004477850-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 692.817045] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-e9359f40-01a5-4a39-b7b9-47a0a48e2ede tempest-VolumesAdminNegativeTest-2004477850 tempest-VolumesAdminNegativeTest-2004477850-project-member] [instance: 30b17a17-229c-4ca7-9ae6-c67ccdcd3963] Processing image b52362a3-ee8a-4cbf-b06f-513b0cc8f95c {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 692.817281] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e9359f40-01a5-4a39-b7b9-47a0a48e2ede tempest-VolumesAdminNegativeTest-2004477850 tempest-VolumesAdminNegativeTest-2004477850-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 692.817422] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e9359f40-01a5-4a39-b7b9-47a0a48e2ede tempest-VolumesAdminNegativeTest-2004477850 tempest-VolumesAdminNegativeTest-2004477850-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 692.817600] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-e9359f40-01a5-4a39-b7b9-47a0a48e2ede tempest-VolumesAdminNegativeTest-2004477850 tempest-VolumesAdminNegativeTest-2004477850-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 692.818131] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4c20b2c0-ea22-40a1-b39a-fa9adf429d7b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.835195] env[65726]: DEBUG oslo_vmware.api [None req-e30832c9-8006-4ee6-8d81-96e2fa3ed74f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] Task: {'id': task-5115630, 'name': PowerOffVM_Task, 'duration_secs': 0.179693} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 692.835195] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-e9359f40-01a5-4a39-b7b9-47a0a48e2ede tempest-VolumesAdminNegativeTest-2004477850 tempest-VolumesAdminNegativeTest-2004477850-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 692.835195] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-e9359f40-01a5-4a39-b7b9-47a0a48e2ede tempest-VolumesAdminNegativeTest-2004477850 tempest-VolumesAdminNegativeTest-2004477850-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65726) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 692.836158] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-e30832c9-8006-4ee6-8d81-96e2fa3ed74f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] [instance: ddd36db0-28ad-4582-994b-7eef8b08ed46] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 692.836913] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-e30832c9-8006-4ee6-8d81-96e2fa3ed74f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] [instance: ddd36db0-28ad-4582-994b-7eef8b08ed46] Destroying instance {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 692.836913] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-279201b6-3268-45f6-aeae-bb3838cdf17f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.839910] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d812af07-2ac9-4c12-9959-4d570343cff6 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.846834] env[65726]: DEBUG oslo_vmware.api [None req-e9359f40-01a5-4a39-b7b9-47a0a48e2ede tempest-VolumesAdminNegativeTest-2004477850 tempest-VolumesAdminNegativeTest-2004477850-project-member] Waiting for the task: (returnval){ [ 692.846834] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5264c17c-2626-a64d-02e4-7a2a01986787" [ 692.846834] env[65726]: _type = "Task" [ 692.846834] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 692.853743] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-e30832c9-8006-4ee6-8d81-96e2fa3ed74f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] [instance: ddd36db0-28ad-4582-994b-7eef8b08ed46] Unregistering the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 692.859858] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ababd445-426a-4a89-a7bb-f3c16fdc124e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.869336] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-d234de18-f385-4dcc-bc17-774bb6452dfb tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] [instance: 703fff2d-5aec-4ed1-b987-9e1cc3a25f67] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 692.869565] env[65726]: DEBUG oslo_vmware.api [None req-e9359f40-01a5-4a39-b7b9-47a0a48e2ede tempest-VolumesAdminNegativeTest-2004477850 tempest-VolumesAdminNegativeTest-2004477850-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5264c17c-2626-a64d-02e4-7a2a01986787, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 692.869735] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-079cea28-e39b-4bdf-bd7b-56101a51ed17 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.881645] env[65726]: DEBUG oslo_vmware.api [None req-d234de18-f385-4dcc-bc17-774bb6452dfb tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Waiting for the task: (returnval){ [ 692.881645] env[65726]: value = "task-5115632" [ 692.881645] env[65726]: _type = "Task" [ 692.881645] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 692.892418] env[65726]: DEBUG nova.network.neutron [None req-f3878371-4599-418c-a420-84e54e420a24 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] [instance: 423af936-a383-4652-8887-9c8b8f507909] Successfully created port: 11b095b5-d149-4b9c-a363-5e8acbc8262d {{(pid=65726) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 692.905758] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-e30832c9-8006-4ee6-8d81-96e2fa3ed74f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] [instance: ddd36db0-28ad-4582-994b-7eef8b08ed46] Unregistered the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 692.905758] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-e30832c9-8006-4ee6-8d81-96e2fa3ed74f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] [instance: ddd36db0-28ad-4582-994b-7eef8b08ed46] Deleting contents of the VM from datastore datastore2 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 692.905758] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-e30832c9-8006-4ee6-8d81-96e2fa3ed74f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] Deleting the datastore file [datastore2] ddd36db0-28ad-4582-994b-7eef8b08ed46 {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 692.905758] env[65726]: DEBUG oslo_vmware.api [None req-d234de18-f385-4dcc-bc17-774bb6452dfb tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Task: {'id': task-5115632, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 692.906039] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3068a06e-e629-4b98-bd6b-14a87b8535ec {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.917312] env[65726]: DEBUG oslo_vmware.api [None req-e30832c9-8006-4ee6-8d81-96e2fa3ed74f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] Waiting for the task: (returnval){ [ 692.917312] env[65726]: value = "task-5115633" [ 692.917312] env[65726]: _type = "Task" [ 692.917312] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 692.932374] env[65726]: DEBUG oslo_vmware.api [None req-e30832c9-8006-4ee6-8d81-96e2fa3ed74f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] Task: {'id': task-5115633, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 693.128926] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b90e193c-43d8-4df3-bab1-667eba5008c0 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.139251] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-606e0556-da5e-4d25-a7a1-eaf835387283 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.177412] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7845cf14-73d3-45cd-af3c-ed290a0e9411 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.187221] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40403bb9-5c13-4166-969f-2f3774c173bf {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.202304] env[65726]: DEBUG nova.compute.provider_tree [None req-9b3c93a2-43e7-4858-b887-85486b2f5f2a tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 693.209257] env[65726]: WARNING openstack [None req-60511481-93fa-4f8b-817d-01a577f85a26 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 693.209504] env[65726]: WARNING openstack [None req-60511481-93fa-4f8b-817d-01a577f85a26 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 693.214706] env[65726]: DEBUG nova.network.neutron [None req-60511481-93fa-4f8b-817d-01a577f85a26 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: 49bcab1f-7b64-4999-abff-37771c58a271] Instance cache missing network info. {{(pid=65726) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 693.328355] env[65726]: DEBUG oslo_concurrency.lockutils [None req-ca40f660-02e2-41e4-8187-3f53d529f31e tempest-ServersAdmin275Test-1899430852 tempest-ServersAdmin275Test-1899430852-project-admin] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 693.364139] env[65726]: DEBUG oslo_vmware.api [None req-e9359f40-01a5-4a39-b7b9-47a0a48e2ede tempest-VolumesAdminNegativeTest-2004477850 tempest-VolumesAdminNegativeTest-2004477850-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5264c17c-2626-a64d-02e4-7a2a01986787, 'name': SearchDatastore_Task, 'duration_secs': 0.053577} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 693.364942] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cc3fcae7-1c09-43e7-b9bf-b9b50ba96ca4 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.370816] env[65726]: DEBUG oslo_vmware.api [None req-e9359f40-01a5-4a39-b7b9-47a0a48e2ede tempest-VolumesAdminNegativeTest-2004477850 tempest-VolumesAdminNegativeTest-2004477850-project-member] Waiting for the task: (returnval){ [ 693.370816] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52f6585e-4234-22ef-76b0-0c45d17ca356" [ 693.370816] env[65726]: _type = "Task" [ 693.370816] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 693.379274] env[65726]: DEBUG oslo_vmware.api [None req-e9359f40-01a5-4a39-b7b9-47a0a48e2ede tempest-VolumesAdminNegativeTest-2004477850 tempest-VolumesAdminNegativeTest-2004477850-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52f6585e-4234-22ef-76b0-0c45d17ca356, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 693.391339] env[65726]: DEBUG oslo_vmware.api [None req-d234de18-f385-4dcc-bc17-774bb6452dfb tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Task: {'id': task-5115632, 'name': PowerOnVM_Task} progress is 94%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 693.431266] env[65726]: DEBUG oslo_vmware.api [None req-e30832c9-8006-4ee6-8d81-96e2fa3ed74f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] Task: {'id': task-5115633, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.199722} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 693.431535] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-e30832c9-8006-4ee6-8d81-96e2fa3ed74f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] Deleted the datastore file {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 693.431713] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-e30832c9-8006-4ee6-8d81-96e2fa3ed74f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] [instance: ddd36db0-28ad-4582-994b-7eef8b08ed46] Deleted contents of the VM from datastore datastore2 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 693.431887] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-e30832c9-8006-4ee6-8d81-96e2fa3ed74f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] [instance: ddd36db0-28ad-4582-994b-7eef8b08ed46] Instance destroyed {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 693.446671] env[65726]: DEBUG nova.compute.manager [None req-f3878371-4599-418c-a420-84e54e420a24 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] [instance: 423af936-a383-4652-8887-9c8b8f507909] Start spawning the instance on the hypervisor. {{(pid=65726) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 693.478196] env[65726]: DEBUG nova.virt.hardware [None req-f3878371-4599-418c-a420-84e54e420a24 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T19:28:34Z,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1539f06fa3534e90acbc3a60c4b8bd3f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T19:28:35Z,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 693.479306] env[65726]: DEBUG nova.virt.hardware [None req-f3878371-4599-418c-a420-84e54e420a24 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 693.479306] env[65726]: DEBUG nova.virt.hardware [None req-f3878371-4599-418c-a420-84e54e420a24 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 693.479306] env[65726]: DEBUG nova.virt.hardware [None req-f3878371-4599-418c-a420-84e54e420a24 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 693.479306] env[65726]: DEBUG nova.virt.hardware [None req-f3878371-4599-418c-a420-84e54e420a24 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 693.479306] env[65726]: DEBUG nova.virt.hardware [None req-f3878371-4599-418c-a420-84e54e420a24 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 693.479519] env[65726]: DEBUG nova.virt.hardware [None req-f3878371-4599-418c-a420-84e54e420a24 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 693.479519] env[65726]: DEBUG nova.virt.hardware [None req-f3878371-4599-418c-a420-84e54e420a24 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 693.479582] env[65726]: DEBUG nova.virt.hardware [None req-f3878371-4599-418c-a420-84e54e420a24 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 693.479771] env[65726]: DEBUG nova.virt.hardware [None req-f3878371-4599-418c-a420-84e54e420a24 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 693.479997] env[65726]: DEBUG nova.virt.hardware [None req-f3878371-4599-418c-a420-84e54e420a24 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 693.484021] env[65726]: DEBUG oslo_concurrency.lockutils [None req-7818044c-0961-4cd4-9c09-6f80f02faa41 tempest-AttachInterfacesV270Test-133387050 tempest-AttachInterfacesV270Test-133387050-project-member] Acquiring lock "interface-c977257e-dce0-4ecc-b42b-0f6ebab38797-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 693.484021] env[65726]: DEBUG oslo_concurrency.lockutils [None req-7818044c-0961-4cd4-9c09-6f80f02faa41 tempest-AttachInterfacesV270Test-133387050 tempest-AttachInterfacesV270Test-133387050-project-member] Lock "interface-c977257e-dce0-4ecc-b42b-0f6ebab38797-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 693.484021] env[65726]: DEBUG nova.objects.instance [None req-7818044c-0961-4cd4-9c09-6f80f02faa41 tempest-AttachInterfacesV270Test-133387050 tempest-AttachInterfacesV270Test-133387050-project-member] Lazy-loading 'flavor' on Instance uuid c977257e-dce0-4ecc-b42b-0f6ebab38797 {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 693.484021] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad5c0cef-1374-4b20-a68f-65b9bf6923d0 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.495422] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9892807d-65ca-4bce-b68a-f426eafdf8aa {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.684152] env[65726]: DEBUG nova.compute.manager [req-87a3175c-30ea-4b4e-977c-d5c322b876cf req-cba48a18-f565-4a73-9f39-47b8ca32f8a3 service nova] [instance: 30b17a17-229c-4ca7-9ae6-c67ccdcd3963] Received event network-changed-b1810a58-276a-4ccf-b223-ed03a7c987b1 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 693.684152] env[65726]: DEBUG nova.compute.manager [req-87a3175c-30ea-4b4e-977c-d5c322b876cf req-cba48a18-f565-4a73-9f39-47b8ca32f8a3 service nova] [instance: 30b17a17-229c-4ca7-9ae6-c67ccdcd3963] Refreshing instance network info cache due to event network-changed-b1810a58-276a-4ccf-b223-ed03a7c987b1. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 693.684152] env[65726]: DEBUG oslo_concurrency.lockutils [req-87a3175c-30ea-4b4e-977c-d5c322b876cf req-cba48a18-f565-4a73-9f39-47b8ca32f8a3 service nova] Acquiring lock "refresh_cache-30b17a17-229c-4ca7-9ae6-c67ccdcd3963" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 693.684152] env[65726]: DEBUG oslo_concurrency.lockutils [req-87a3175c-30ea-4b4e-977c-d5c322b876cf req-cba48a18-f565-4a73-9f39-47b8ca32f8a3 service nova] Acquired lock "refresh_cache-30b17a17-229c-4ca7-9ae6-c67ccdcd3963" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 693.684152] env[65726]: DEBUG nova.network.neutron [req-87a3175c-30ea-4b4e-977c-d5c322b876cf req-cba48a18-f565-4a73-9f39-47b8ca32f8a3 service nova] [instance: 30b17a17-229c-4ca7-9ae6-c67ccdcd3963] Refreshing network info cache for port b1810a58-276a-4ccf-b223-ed03a7c987b1 {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 693.710256] env[65726]: DEBUG nova.scheduler.client.report [None req-9b3c93a2-43e7-4858-b887-85486b2f5f2a tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 693.884625] env[65726]: DEBUG oslo_vmware.api [None req-e9359f40-01a5-4a39-b7b9-47a0a48e2ede tempest-VolumesAdminNegativeTest-2004477850 tempest-VolumesAdminNegativeTest-2004477850-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52f6585e-4234-22ef-76b0-0c45d17ca356, 'name': SearchDatastore_Task, 'duration_secs': 0.021704} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 693.886931] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e9359f40-01a5-4a39-b7b9-47a0a48e2ede tempest-VolumesAdminNegativeTest-2004477850 tempest-VolumesAdminNegativeTest-2004477850-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 693.887233] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-e9359f40-01a5-4a39-b7b9-47a0a48e2ede tempest-VolumesAdminNegativeTest-2004477850 tempest-VolumesAdminNegativeTest-2004477850-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] 30b17a17-229c-4ca7-9ae6-c67ccdcd3963/30b17a17-229c-4ca7-9ae6-c67ccdcd3963.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 693.887528] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6784e185-a3f9-4dcf-9afb-541fd4a1072a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.895946] env[65726]: DEBUG oslo_vmware.api [None req-d234de18-f385-4dcc-bc17-774bb6452dfb tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Task: {'id': task-5115632, 'name': PowerOnVM_Task, 'duration_secs': 0.62354} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 693.897450] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-d234de18-f385-4dcc-bc17-774bb6452dfb tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] [instance: 703fff2d-5aec-4ed1-b987-9e1cc3a25f67] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 693.897671] env[65726]: DEBUG nova.compute.manager [None req-d234de18-f385-4dcc-bc17-774bb6452dfb tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] [instance: 703fff2d-5aec-4ed1-b987-9e1cc3a25f67] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 693.898091] env[65726]: DEBUG oslo_vmware.api [None req-e9359f40-01a5-4a39-b7b9-47a0a48e2ede tempest-VolumesAdminNegativeTest-2004477850 tempest-VolumesAdminNegativeTest-2004477850-project-member] Waiting for the task: (returnval){ [ 693.898091] env[65726]: value = "task-5115634" [ 693.898091] env[65726]: _type = "Task" [ 693.898091] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 693.898906] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e82d9597-deba-4c5d-95da-7370bf35a4ef {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.918555] env[65726]: DEBUG oslo_vmware.api [None req-e9359f40-01a5-4a39-b7b9-47a0a48e2ede tempest-VolumesAdminNegativeTest-2004477850 tempest-VolumesAdminNegativeTest-2004477850-project-member] Task: {'id': task-5115634, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 693.989142] env[65726]: WARNING openstack [None req-7818044c-0961-4cd4-9c09-6f80f02faa41 tempest-AttachInterfacesV270Test-133387050 tempest-AttachInterfacesV270Test-133387050-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 693.990600] env[65726]: WARNING openstack [None req-7818044c-0961-4cd4-9c09-6f80f02faa41 tempest-AttachInterfacesV270Test-133387050 tempest-AttachInterfacesV270Test-133387050-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 693.996297] env[65726]: DEBUG nova.objects.instance [None req-7818044c-0961-4cd4-9c09-6f80f02faa41 tempest-AttachInterfacesV270Test-133387050 tempest-AttachInterfacesV270Test-133387050-project-member] Lazy-loading 'pci_requests' on Instance uuid c977257e-dce0-4ecc-b42b-0f6ebab38797 {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 694.193684] env[65726]: WARNING openstack [req-87a3175c-30ea-4b4e-977c-d5c322b876cf req-cba48a18-f565-4a73-9f39-47b8ca32f8a3 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 694.194158] env[65726]: WARNING openstack [req-87a3175c-30ea-4b4e-977c-d5c322b876cf req-cba48a18-f565-4a73-9f39-47b8ca32f8a3 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 694.413860] env[65726]: DEBUG oslo_vmware.api [None req-e9359f40-01a5-4a39-b7b9-47a0a48e2ede tempest-VolumesAdminNegativeTest-2004477850 tempest-VolumesAdminNegativeTest-2004477850-project-member] Task: {'id': task-5115634, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 694.474651] env[65726]: DEBUG nova.virt.hardware [None req-e30832c9-8006-4ee6-8d81-96e2fa3ed74f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T19:28:34Z,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1539f06fa3534e90acbc3a60c4b8bd3f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T19:28:35Z,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 694.474837] env[65726]: DEBUG nova.virt.hardware [None req-e30832c9-8006-4ee6-8d81-96e2fa3ed74f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 694.474985] env[65726]: DEBUG nova.virt.hardware [None req-e30832c9-8006-4ee6-8d81-96e2fa3ed74f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 694.475185] env[65726]: DEBUG nova.virt.hardware [None req-e30832c9-8006-4ee6-8d81-96e2fa3ed74f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 694.475322] env[65726]: DEBUG nova.virt.hardware [None req-e30832c9-8006-4ee6-8d81-96e2fa3ed74f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 694.475460] env[65726]: DEBUG nova.virt.hardware [None req-e30832c9-8006-4ee6-8d81-96e2fa3ed74f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 694.475724] env[65726]: DEBUG nova.virt.hardware [None req-e30832c9-8006-4ee6-8d81-96e2fa3ed74f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 694.475879] env[65726]: DEBUG nova.virt.hardware [None req-e30832c9-8006-4ee6-8d81-96e2fa3ed74f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 694.476100] env[65726]: DEBUG nova.virt.hardware [None req-e30832c9-8006-4ee6-8d81-96e2fa3ed74f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 694.476314] env[65726]: DEBUG nova.virt.hardware [None req-e30832c9-8006-4ee6-8d81-96e2fa3ed74f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 694.476487] env[65726]: DEBUG nova.virt.hardware [None req-e30832c9-8006-4ee6-8d81-96e2fa3ed74f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 694.477642] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6241cac-13be-4490-8dab-cdfe182bdc4a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.488119] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16c8d403-5884-4a01-8965-6647388db026 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.505692] env[65726]: DEBUG nova.objects.base [None req-7818044c-0961-4cd4-9c09-6f80f02faa41 tempest-AttachInterfacesV270Test-133387050 tempest-AttachInterfacesV270Test-133387050-project-member] Object Instance lazy-loaded attributes: flavor,pci_requests {{(pid=65726) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 694.506083] env[65726]: DEBUG nova.network.neutron [None req-7818044c-0961-4cd4-9c09-6f80f02faa41 tempest-AttachInterfacesV270Test-133387050 tempest-AttachInterfacesV270Test-133387050-project-member] [instance: c977257e-dce0-4ecc-b42b-0f6ebab38797] allocate_for_instance() {{(pid=65726) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 694.506439] env[65726]: WARNING neutronclient.v2_0.client [None req-7818044c-0961-4cd4-9c09-6f80f02faa41 tempest-AttachInterfacesV270Test-133387050 tempest-AttachInterfacesV270Test-133387050-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 694.506816] env[65726]: WARNING neutronclient.v2_0.client [None req-7818044c-0961-4cd4-9c09-6f80f02faa41 tempest-AttachInterfacesV270Test-133387050 tempest-AttachInterfacesV270Test-133387050-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 694.507443] env[65726]: WARNING openstack [None req-7818044c-0961-4cd4-9c09-6f80f02faa41 tempest-AttachInterfacesV270Test-133387050 tempest-AttachInterfacesV270Test-133387050-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 694.507794] env[65726]: WARNING openstack [None req-7818044c-0961-4cd4-9c09-6f80f02faa41 tempest-AttachInterfacesV270Test-133387050 tempest-AttachInterfacesV270Test-133387050-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 694.517164] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-e30832c9-8006-4ee6-8d81-96e2fa3ed74f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] [instance: ddd36db0-28ad-4582-994b-7eef8b08ed46] Instance VIF info [] {{(pid=65726) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 694.523608] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-e30832c9-8006-4ee6-8d81-96e2fa3ed74f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 694.524871] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ddd36db0-28ad-4582-994b-7eef8b08ed46] Creating VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 694.525150] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d0fb8867-1a9e-42f4-845f-dcfc91d696e5 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.546579] env[65726]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 694.546579] env[65726]: value = "task-5115635" [ 694.546579] env[65726]: _type = "Task" [ 694.546579] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 694.556766] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115635, 'name': CreateVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 694.558611] env[65726]: WARNING openstack [None req-60511481-93fa-4f8b-817d-01a577f85a26 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 694.559160] env[65726]: WARNING openstack [None req-60511481-93fa-4f8b-817d-01a577f85a26 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 694.586048] env[65726]: DEBUG nova.network.neutron [None req-f3878371-4599-418c-a420-84e54e420a24 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] [instance: 423af936-a383-4652-8887-9c8b8f507909] Successfully updated port: 11b095b5-d149-4b9c-a363-5e8acbc8262d {{(pid=65726) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 694.721842] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9b3c93a2-43e7-4858-b887-85486b2f5f2a tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.811s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 694.724853] env[65726]: DEBUG oslo_concurrency.lockutils [None req-34fc70a6-ee2b-4816-b933-3db54eee74ae tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 21.870s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 694.725266] env[65726]: DEBUG nova.objects.instance [None req-34fc70a6-ee2b-4816-b933-3db54eee74ae tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Lazy-loading 'resources' on Instance uuid 811c1bbd-259c-4f5e-9a61-581bfba0ec4d {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 694.740941] env[65726]: WARNING openstack [None req-60511481-93fa-4f8b-817d-01a577f85a26 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 694.741379] env[65726]: WARNING openstack [None req-60511481-93fa-4f8b-817d-01a577f85a26 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 694.752548] env[65726]: DEBUG oslo_concurrency.lockutils [None req-7818044c-0961-4cd4-9c09-6f80f02faa41 tempest-AttachInterfacesV270Test-133387050 tempest-AttachInterfacesV270Test-133387050-project-member] Lock "interface-c977257e-dce0-4ecc-b42b-0f6ebab38797-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 1.271s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 694.756301] env[65726]: WARNING openstack [req-87a3175c-30ea-4b4e-977c-d5c322b876cf req-cba48a18-f565-4a73-9f39-47b8ca32f8a3 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 694.756708] env[65726]: WARNING openstack [req-87a3175c-30ea-4b4e-977c-d5c322b876cf req-cba48a18-f565-4a73-9f39-47b8ca32f8a3 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 694.846911] env[65726]: DEBUG nova.network.neutron [None req-60511481-93fa-4f8b-817d-01a577f85a26 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: 49bcab1f-7b64-4999-abff-37771c58a271] Updating instance_info_cache with network_info: [{"id": "45612b7e-c052-4346-a779-5f94f9adc084", "address": "fa:16:3e:9d:95:60", "network": {"id": "51a66448-0f72-4266-8868-856b48be839f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-155658961-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d46ccb00794f458b85da4a93879139ee", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b1f3e6c3-5584-4852-9017-476ab8ac4946", "external-id": "nsx-vlan-transportzone-304", "segmentation_id": 304, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap45612b7e-c0", "ovs_interfaceid": "45612b7e-c052-4346-a779-5f94f9adc084", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 694.851837] env[65726]: WARNING openstack [req-87a3175c-30ea-4b4e-977c-d5c322b876cf req-cba48a18-f565-4a73-9f39-47b8ca32f8a3 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 694.851837] env[65726]: WARNING openstack [req-87a3175c-30ea-4b4e-977c-d5c322b876cf req-cba48a18-f565-4a73-9f39-47b8ca32f8a3 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 694.915065] env[65726]: DEBUG oslo_vmware.api [None req-e9359f40-01a5-4a39-b7b9-47a0a48e2ede tempest-VolumesAdminNegativeTest-2004477850 tempest-VolumesAdminNegativeTest-2004477850-project-member] Task: {'id': task-5115634, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.772967} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 694.915356] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-e9359f40-01a5-4a39-b7b9-47a0a48e2ede tempest-VolumesAdminNegativeTest-2004477850 tempest-VolumesAdminNegativeTest-2004477850-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] 30b17a17-229c-4ca7-9ae6-c67ccdcd3963/30b17a17-229c-4ca7-9ae6-c67ccdcd3963.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 694.915564] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-e9359f40-01a5-4a39-b7b9-47a0a48e2ede tempest-VolumesAdminNegativeTest-2004477850 tempest-VolumesAdminNegativeTest-2004477850-project-member] [instance: 30b17a17-229c-4ca7-9ae6-c67ccdcd3963] Extending root virtual disk to 1048576 {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 694.915820] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-672d3a1e-c758-4a39-ac22-32d3956d0218 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.924152] env[65726]: DEBUG oslo_vmware.api [None req-e9359f40-01a5-4a39-b7b9-47a0a48e2ede tempest-VolumesAdminNegativeTest-2004477850 tempest-VolumesAdminNegativeTest-2004477850-project-member] Waiting for the task: (returnval){ [ 694.924152] env[65726]: value = "task-5115636" [ 694.924152] env[65726]: _type = "Task" [ 694.924152] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 694.937027] env[65726]: DEBUG oslo_vmware.api [None req-e9359f40-01a5-4a39-b7b9-47a0a48e2ede tempest-VolumesAdminNegativeTest-2004477850 tempest-VolumesAdminNegativeTest-2004477850-project-member] Task: {'id': task-5115636, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 694.953051] env[65726]: DEBUG nova.network.neutron [req-87a3175c-30ea-4b4e-977c-d5c322b876cf req-cba48a18-f565-4a73-9f39-47b8ca32f8a3 service nova] [instance: 30b17a17-229c-4ca7-9ae6-c67ccdcd3963] Updated VIF entry in instance network info cache for port b1810a58-276a-4ccf-b223-ed03a7c987b1. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 694.954761] env[65726]: DEBUG nova.network.neutron [req-87a3175c-30ea-4b4e-977c-d5c322b876cf req-cba48a18-f565-4a73-9f39-47b8ca32f8a3 service nova] [instance: 30b17a17-229c-4ca7-9ae6-c67ccdcd3963] Updating instance_info_cache with network_info: [{"id": "b1810a58-276a-4ccf-b223-ed03a7c987b1", "address": "fa:16:3e:b4:fe:bf", "network": {"id": "f9d993c3-5487-4b97-89c3-5f30cdf9c387", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-133871862-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "33d7ed4a125d4cc2ae93fbe4085f66d7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "aa8c2f93-f287-41b3-adb6-4942a7ea2a0b", "external-id": "nsx-vlan-transportzone-363", "segmentation_id": 363, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb1810a58-27", "ovs_interfaceid": "b1810a58-276a-4ccf-b223-ed03a7c987b1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 695.058265] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115635, 'name': CreateVM_Task, 'duration_secs': 0.449116} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 695.058541] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ddd36db0-28ad-4582-994b-7eef8b08ed46] Created VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 695.058881] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e30832c9-8006-4ee6-8d81-96e2fa3ed74f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 695.059058] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e30832c9-8006-4ee6-8d81-96e2fa3ed74f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 695.060241] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e30832c9-8006-4ee6-8d81-96e2fa3ed74f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 695.060241] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c9d20252-9d3d-4406-b43e-4ddaae65d88b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.065703] env[65726]: DEBUG oslo_vmware.api [None req-e30832c9-8006-4ee6-8d81-96e2fa3ed74f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] Waiting for the task: (returnval){ [ 695.065703] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5231f2cc-750b-bcd8-296e-11fe9b40a580" [ 695.065703] env[65726]: _type = "Task" [ 695.065703] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 695.074139] env[65726]: DEBUG oslo_vmware.api [None req-e30832c9-8006-4ee6-8d81-96e2fa3ed74f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5231f2cc-750b-bcd8-296e-11fe9b40a580, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 695.089315] env[65726]: DEBUG oslo_concurrency.lockutils [None req-f3878371-4599-418c-a420-84e54e420a24 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Acquiring lock "refresh_cache-423af936-a383-4652-8887-9c8b8f507909" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 695.089509] env[65726]: DEBUG oslo_concurrency.lockutils [None req-f3878371-4599-418c-a420-84e54e420a24 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Acquired lock "refresh_cache-423af936-a383-4652-8887-9c8b8f507909" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 695.090234] env[65726]: DEBUG nova.network.neutron [None req-f3878371-4599-418c-a420-84e54e420a24 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] [instance: 423af936-a383-4652-8887-9c8b8f507909] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 695.287618] env[65726]: INFO nova.scheduler.client.report [None req-9b3c93a2-43e7-4858-b887-85486b2f5f2a tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Deleted allocation for migration 8f64b022-cb5e-47d0-89a3-b1455f9404a6 [ 695.364390] env[65726]: DEBUG oslo_concurrency.lockutils [None req-60511481-93fa-4f8b-817d-01a577f85a26 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Releasing lock "refresh_cache-49bcab1f-7b64-4999-abff-37771c58a271" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 695.364772] env[65726]: DEBUG nova.compute.manager [None req-60511481-93fa-4f8b-817d-01a577f85a26 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: 49bcab1f-7b64-4999-abff-37771c58a271] Instance network_info: |[{"id": "45612b7e-c052-4346-a779-5f94f9adc084", "address": "fa:16:3e:9d:95:60", "network": {"id": "51a66448-0f72-4266-8868-856b48be839f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-155658961-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d46ccb00794f458b85da4a93879139ee", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b1f3e6c3-5584-4852-9017-476ab8ac4946", "external-id": "nsx-vlan-transportzone-304", "segmentation_id": 304, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap45612b7e-c0", "ovs_interfaceid": "45612b7e-c052-4346-a779-5f94f9adc084", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 695.365562] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-60511481-93fa-4f8b-817d-01a577f85a26 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: 49bcab1f-7b64-4999-abff-37771c58a271] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:9d:95:60', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'b1f3e6c3-5584-4852-9017-476ab8ac4946', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '45612b7e-c052-4346-a779-5f94f9adc084', 'vif_model': 'vmxnet3'}] {{(pid=65726) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 695.374096] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-60511481-93fa-4f8b-817d-01a577f85a26 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Creating folder: Project (d46ccb00794f458b85da4a93879139ee). Parent ref: group-v995008. {{(pid=65726) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 695.374407] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b12cfa89-adfe-4e1f-9688-3d4ac380cc54 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.388810] env[65726]: INFO nova.virt.vmwareapi.vm_util [None req-60511481-93fa-4f8b-817d-01a577f85a26 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Created folder: Project (d46ccb00794f458b85da4a93879139ee) in parent group-v995008. [ 695.389336] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-60511481-93fa-4f8b-817d-01a577f85a26 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Creating folder: Instances. Parent ref: group-v995086. {{(pid=65726) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 695.389396] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-89b02509-884a-4ce5-9b1d-faa19dfce351 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.402836] env[65726]: INFO nova.virt.vmwareapi.vm_util [None req-60511481-93fa-4f8b-817d-01a577f85a26 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Created folder: Instances in parent group-v995086. [ 695.403604] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-60511481-93fa-4f8b-817d-01a577f85a26 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 695.403604] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 49bcab1f-7b64-4999-abff-37771c58a271] Creating VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 695.403604] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e5e71f3d-79e8-469f-b03e-e29089f0f3b2 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.429555] env[65726]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 695.429555] env[65726]: value = "task-5115639" [ 695.429555] env[65726]: _type = "Task" [ 695.429555] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 695.438659] env[65726]: DEBUG oslo_vmware.api [None req-e9359f40-01a5-4a39-b7b9-47a0a48e2ede tempest-VolumesAdminNegativeTest-2004477850 tempest-VolumesAdminNegativeTest-2004477850-project-member] Task: {'id': task-5115636, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.115206} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 695.439941] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-e9359f40-01a5-4a39-b7b9-47a0a48e2ede tempest-VolumesAdminNegativeTest-2004477850 tempest-VolumesAdminNegativeTest-2004477850-project-member] [instance: 30b17a17-229c-4ca7-9ae6-c67ccdcd3963] Extended root virtual disk {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 695.440558] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e9c490a-3282-42b6-b260-2c366be5ceec {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.446860] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115639, 'name': CreateVM_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 695.463187] env[65726]: DEBUG oslo_concurrency.lockutils [req-87a3175c-30ea-4b4e-977c-d5c322b876cf req-cba48a18-f565-4a73-9f39-47b8ca32f8a3 service nova] Releasing lock "refresh_cache-30b17a17-229c-4ca7-9ae6-c67ccdcd3963" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 695.473093] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-e9359f40-01a5-4a39-b7b9-47a0a48e2ede tempest-VolumesAdminNegativeTest-2004477850 tempest-VolumesAdminNegativeTest-2004477850-project-member] [instance: 30b17a17-229c-4ca7-9ae6-c67ccdcd3963] Reconfiguring VM instance instance-0000001b to attach disk [datastore1] 30b17a17-229c-4ca7-9ae6-c67ccdcd3963/30b17a17-229c-4ca7-9ae6-c67ccdcd3963.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 695.476357] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b3e49e7a-3322-402e-8396-038cd694a98c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.499928] env[65726]: DEBUG oslo_vmware.api [None req-e9359f40-01a5-4a39-b7b9-47a0a48e2ede tempest-VolumesAdminNegativeTest-2004477850 tempest-VolumesAdminNegativeTest-2004477850-project-member] Waiting for the task: (returnval){ [ 695.499928] env[65726]: value = "task-5115640" [ 695.499928] env[65726]: _type = "Task" [ 695.499928] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 695.510233] env[65726]: DEBUG oslo_vmware.api [None req-e9359f40-01a5-4a39-b7b9-47a0a48e2ede tempest-VolumesAdminNegativeTest-2004477850 tempest-VolumesAdminNegativeTest-2004477850-project-member] Task: {'id': task-5115640, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 695.581694] env[65726]: DEBUG oslo_vmware.api [None req-e30832c9-8006-4ee6-8d81-96e2fa3ed74f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5231f2cc-750b-bcd8-296e-11fe9b40a580, 'name': SearchDatastore_Task, 'duration_secs': 0.011742} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 695.584943] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e30832c9-8006-4ee6-8d81-96e2fa3ed74f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 695.585260] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-e30832c9-8006-4ee6-8d81-96e2fa3ed74f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] [instance: ddd36db0-28ad-4582-994b-7eef8b08ed46] Processing image b52362a3-ee8a-4cbf-b06f-513b0cc8f95c {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 695.585538] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e30832c9-8006-4ee6-8d81-96e2fa3ed74f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 695.585681] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e30832c9-8006-4ee6-8d81-96e2fa3ed74f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 695.585855] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-e30832c9-8006-4ee6-8d81-96e2fa3ed74f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 695.586389] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a359be2e-9cef-4ead-9a19-22be0d945a6b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.593030] env[65726]: WARNING openstack [None req-f3878371-4599-418c-a420-84e54e420a24 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 695.593228] env[65726]: WARNING openstack [None req-f3878371-4599-418c-a420-84e54e420a24 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 695.599308] env[65726]: DEBUG nova.network.neutron [None req-f3878371-4599-418c-a420-84e54e420a24 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] [instance: 423af936-a383-4652-8887-9c8b8f507909] Instance cache missing network info. {{(pid=65726) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 695.606772] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-e30832c9-8006-4ee6-8d81-96e2fa3ed74f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 695.607052] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-e30832c9-8006-4ee6-8d81-96e2fa3ed74f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65726) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 695.608561] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4d4cac88-ca35-4ab3-bbe0-079a86c1d33e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.616526] env[65726]: DEBUG oslo_vmware.api [None req-e30832c9-8006-4ee6-8d81-96e2fa3ed74f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] Waiting for the task: (returnval){ [ 695.616526] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52332237-d003-03b4-6e43-18cca1a29f38" [ 695.616526] env[65726]: _type = "Task" [ 695.616526] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 695.629714] env[65726]: DEBUG oslo_vmware.api [None req-e30832c9-8006-4ee6-8d81-96e2fa3ed74f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52332237-d003-03b4-6e43-18cca1a29f38, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 695.673289] env[65726]: WARNING openstack [None req-f3878371-4599-418c-a420-84e54e420a24 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 695.674014] env[65726]: WARNING openstack [None req-f3878371-4599-418c-a420-84e54e420a24 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 695.763553] env[65726]: WARNING openstack [None req-f3878371-4599-418c-a420-84e54e420a24 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 695.764125] env[65726]: WARNING openstack [None req-f3878371-4599-418c-a420-84e54e420a24 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 695.797762] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9b3c93a2-43e7-4858-b887-85486b2f5f2a tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Lock "0e064341-4e4a-407b-8c26-3eb04b409029" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 31.895s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 695.882158] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e8f76de-8b28-4d26-a18c-eb3ae6e4965d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.890897] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef3ed850-871b-4aae-aa7d-0a043aae198c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.924043] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31c3a66b-1363-4007-925b-18dda91b57ea {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.935810] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1748d3ba-8321-4194-a7ba-259319403723 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.947170] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115639, 'name': CreateVM_Task, 'duration_secs': 0.404787} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 695.954947] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 49bcab1f-7b64-4999-abff-37771c58a271] Created VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 695.955673] env[65726]: DEBUG nova.compute.provider_tree [None req-34fc70a6-ee2b-4816-b933-3db54eee74ae tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 695.957574] env[65726]: WARNING openstack [None req-60511481-93fa-4f8b-817d-01a577f85a26 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 695.959109] env[65726]: WARNING openstack [None req-60511481-93fa-4f8b-817d-01a577f85a26 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 695.962993] env[65726]: DEBUG oslo_concurrency.lockutils [None req-60511481-93fa-4f8b-817d-01a577f85a26 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 695.963298] env[65726]: DEBUG oslo_concurrency.lockutils [None req-60511481-93fa-4f8b-817d-01a577f85a26 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 695.963748] env[65726]: DEBUG oslo_concurrency.lockutils [None req-60511481-93fa-4f8b-817d-01a577f85a26 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 695.964347] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e16ee887-2183-4f71-9ba1-39608a9b4cd9 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.970668] env[65726]: DEBUG oslo_vmware.api [None req-60511481-93fa-4f8b-817d-01a577f85a26 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Waiting for the task: (returnval){ [ 695.970668] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]520a732f-1d90-bbd2-7c04-8c27939fec0a" [ 695.970668] env[65726]: _type = "Task" [ 695.970668] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 695.981758] env[65726]: DEBUG oslo_vmware.api [None req-60511481-93fa-4f8b-817d-01a577f85a26 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]520a732f-1d90-bbd2-7c04-8c27939fec0a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 696.011608] env[65726]: DEBUG oslo_vmware.api [None req-e9359f40-01a5-4a39-b7b9-47a0a48e2ede tempest-VolumesAdminNegativeTest-2004477850 tempest-VolumesAdminNegativeTest-2004477850-project-member] Task: {'id': task-5115640, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 696.128804] env[65726]: DEBUG oslo_vmware.api [None req-e30832c9-8006-4ee6-8d81-96e2fa3ed74f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52332237-d003-03b4-6e43-18cca1a29f38, 'name': SearchDatastore_Task, 'duration_secs': 0.01766} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 696.128804] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8edc1b35-2a32-4812-8f71-c523707c109a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.134394] env[65726]: DEBUG oslo_vmware.api [None req-e30832c9-8006-4ee6-8d81-96e2fa3ed74f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] Waiting for the task: (returnval){ [ 696.134394] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]529cc391-6204-9749-5b88-c2bd7d46814e" [ 696.134394] env[65726]: _type = "Task" [ 696.134394] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 696.145026] env[65726]: DEBUG oslo_vmware.api [None req-e30832c9-8006-4ee6-8d81-96e2fa3ed74f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]529cc391-6204-9749-5b88-c2bd7d46814e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 696.251934] env[65726]: DEBUG nova.network.neutron [None req-f3878371-4599-418c-a420-84e54e420a24 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] [instance: 423af936-a383-4652-8887-9c8b8f507909] Updating instance_info_cache with network_info: [{"id": "11b095b5-d149-4b9c-a363-5e8acbc8262d", "address": "fa:16:3e:4d:ab:f3", "network": {"id": "55d629b8-a541-4720-9289-927bbc6e6f22", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.128", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "1539f06fa3534e90acbc3a60c4b8bd3f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "35342bcb-8b06-472e-b3c0-43fd3d6c4b30", "external-id": "nsx-vlan-transportzone-524", "segmentation_id": 524, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap11b095b5-d1", "ovs_interfaceid": "11b095b5-d149-4b9c-a363-5e8acbc8262d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 696.369461] env[65726]: DEBUG nova.compute.manager [req-5f56fa1c-0a6d-4dd1-970c-4d739de98a98 req-9b58a54e-5b33-424e-a647-e9ff428baa0e service nova] [instance: 423af936-a383-4652-8887-9c8b8f507909] Received event network-vif-plugged-11b095b5-d149-4b9c-a363-5e8acbc8262d {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 696.369790] env[65726]: DEBUG oslo_concurrency.lockutils [req-5f56fa1c-0a6d-4dd1-970c-4d739de98a98 req-9b58a54e-5b33-424e-a647-e9ff428baa0e service nova] Acquiring lock "423af936-a383-4652-8887-9c8b8f507909-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 696.370043] env[65726]: DEBUG oslo_concurrency.lockutils [req-5f56fa1c-0a6d-4dd1-970c-4d739de98a98 req-9b58a54e-5b33-424e-a647-e9ff428baa0e service nova] Lock "423af936-a383-4652-8887-9c8b8f507909-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 696.370207] env[65726]: DEBUG oslo_concurrency.lockutils [req-5f56fa1c-0a6d-4dd1-970c-4d739de98a98 req-9b58a54e-5b33-424e-a647-e9ff428baa0e service nova] Lock "423af936-a383-4652-8887-9c8b8f507909-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 696.370584] env[65726]: DEBUG nova.compute.manager [req-5f56fa1c-0a6d-4dd1-970c-4d739de98a98 req-9b58a54e-5b33-424e-a647-e9ff428baa0e service nova] [instance: 423af936-a383-4652-8887-9c8b8f507909] No waiting events found dispatching network-vif-plugged-11b095b5-d149-4b9c-a363-5e8acbc8262d {{(pid=65726) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 696.370584] env[65726]: WARNING nova.compute.manager [req-5f56fa1c-0a6d-4dd1-970c-4d739de98a98 req-9b58a54e-5b33-424e-a647-e9ff428baa0e service nova] [instance: 423af936-a383-4652-8887-9c8b8f507909] Received unexpected event network-vif-plugged-11b095b5-d149-4b9c-a363-5e8acbc8262d for instance with vm_state building and task_state spawning. [ 696.466290] env[65726]: DEBUG nova.scheduler.client.report [None req-34fc70a6-ee2b-4816-b933-3db54eee74ae tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 696.488425] env[65726]: DEBUG oslo_vmware.api [None req-60511481-93fa-4f8b-817d-01a577f85a26 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]520a732f-1d90-bbd2-7c04-8c27939fec0a, 'name': SearchDatastore_Task, 'duration_secs': 0.023362} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 696.488952] env[65726]: DEBUG oslo_concurrency.lockutils [None req-60511481-93fa-4f8b-817d-01a577f85a26 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 696.489320] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-60511481-93fa-4f8b-817d-01a577f85a26 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: 49bcab1f-7b64-4999-abff-37771c58a271] Processing image b52362a3-ee8a-4cbf-b06f-513b0cc8f95c {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 696.489559] env[65726]: DEBUG oslo_concurrency.lockutils [None req-60511481-93fa-4f8b-817d-01a577f85a26 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 696.489700] env[65726]: DEBUG oslo_concurrency.lockutils [None req-60511481-93fa-4f8b-817d-01a577f85a26 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 696.489876] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-60511481-93fa-4f8b-817d-01a577f85a26 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 696.490726] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-69158f64-2b92-426f-9b49-89b176652213 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.508076] env[65726]: DEBUG nova.compute.manager [req-892bee8d-2a7e-43c4-99a8-93139b745e96 req-a64b0cbe-c2d6-473f-9f86-13d77adb055e service nova] [instance: 49bcab1f-7b64-4999-abff-37771c58a271] Received event network-vif-plugged-45612b7e-c052-4346-a779-5f94f9adc084 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 696.508188] env[65726]: DEBUG oslo_concurrency.lockutils [req-892bee8d-2a7e-43c4-99a8-93139b745e96 req-a64b0cbe-c2d6-473f-9f86-13d77adb055e service nova] Acquiring lock "49bcab1f-7b64-4999-abff-37771c58a271-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 696.508384] env[65726]: DEBUG oslo_concurrency.lockutils [req-892bee8d-2a7e-43c4-99a8-93139b745e96 req-a64b0cbe-c2d6-473f-9f86-13d77adb055e service nova] Lock "49bcab1f-7b64-4999-abff-37771c58a271-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 696.508541] env[65726]: DEBUG oslo_concurrency.lockutils [req-892bee8d-2a7e-43c4-99a8-93139b745e96 req-a64b0cbe-c2d6-473f-9f86-13d77adb055e service nova] Lock "49bcab1f-7b64-4999-abff-37771c58a271-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 696.508697] env[65726]: DEBUG nova.compute.manager [req-892bee8d-2a7e-43c4-99a8-93139b745e96 req-a64b0cbe-c2d6-473f-9f86-13d77adb055e service nova] [instance: 49bcab1f-7b64-4999-abff-37771c58a271] No waiting events found dispatching network-vif-plugged-45612b7e-c052-4346-a779-5f94f9adc084 {{(pid=65726) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 696.509078] env[65726]: WARNING nova.compute.manager [req-892bee8d-2a7e-43c4-99a8-93139b745e96 req-a64b0cbe-c2d6-473f-9f86-13d77adb055e service nova] [instance: 49bcab1f-7b64-4999-abff-37771c58a271] Received unexpected event network-vif-plugged-45612b7e-c052-4346-a779-5f94f9adc084 for instance with vm_state building and task_state spawning. [ 696.509372] env[65726]: DEBUG nova.compute.manager [req-892bee8d-2a7e-43c4-99a8-93139b745e96 req-a64b0cbe-c2d6-473f-9f86-13d77adb055e service nova] [instance: 49bcab1f-7b64-4999-abff-37771c58a271] Received event network-changed-45612b7e-c052-4346-a779-5f94f9adc084 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 696.509536] env[65726]: DEBUG nova.compute.manager [req-892bee8d-2a7e-43c4-99a8-93139b745e96 req-a64b0cbe-c2d6-473f-9f86-13d77adb055e service nova] [instance: 49bcab1f-7b64-4999-abff-37771c58a271] Refreshing instance network info cache due to event network-changed-45612b7e-c052-4346-a779-5f94f9adc084. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 696.509718] env[65726]: DEBUG oslo_concurrency.lockutils [req-892bee8d-2a7e-43c4-99a8-93139b745e96 req-a64b0cbe-c2d6-473f-9f86-13d77adb055e service nova] Acquiring lock "refresh_cache-49bcab1f-7b64-4999-abff-37771c58a271" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 696.509890] env[65726]: DEBUG oslo_concurrency.lockutils [req-892bee8d-2a7e-43c4-99a8-93139b745e96 req-a64b0cbe-c2d6-473f-9f86-13d77adb055e service nova] Acquired lock "refresh_cache-49bcab1f-7b64-4999-abff-37771c58a271" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 696.510150] env[65726]: DEBUG nova.network.neutron [req-892bee8d-2a7e-43c4-99a8-93139b745e96 req-a64b0cbe-c2d6-473f-9f86-13d77adb055e service nova] [instance: 49bcab1f-7b64-4999-abff-37771c58a271] Refreshing network info cache for port 45612b7e-c052-4346-a779-5f94f9adc084 {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 696.513531] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-60511481-93fa-4f8b-817d-01a577f85a26 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 696.513672] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-60511481-93fa-4f8b-817d-01a577f85a26 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65726) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 696.515312] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-971e8525-ebec-47d2-87ba-ca167825ec16 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.523625] env[65726]: DEBUG oslo_vmware.api [None req-e9359f40-01a5-4a39-b7b9-47a0a48e2ede tempest-VolumesAdminNegativeTest-2004477850 tempest-VolumesAdminNegativeTest-2004477850-project-member] Task: {'id': task-5115640, 'name': ReconfigVM_Task, 'duration_secs': 0.703468} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 696.524292] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-e9359f40-01a5-4a39-b7b9-47a0a48e2ede tempest-VolumesAdminNegativeTest-2004477850 tempest-VolumesAdminNegativeTest-2004477850-project-member] [instance: 30b17a17-229c-4ca7-9ae6-c67ccdcd3963] Reconfigured VM instance instance-0000001b to attach disk [datastore1] 30b17a17-229c-4ca7-9ae6-c67ccdcd3963/30b17a17-229c-4ca7-9ae6-c67ccdcd3963.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 696.525010] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-cdf423cb-a091-4840-9267-7cf6c4a7b55a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.528714] env[65726]: DEBUG oslo_vmware.api [None req-60511481-93fa-4f8b-817d-01a577f85a26 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Waiting for the task: (returnval){ [ 696.528714] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5291c029-5bf7-d17d-5ceb-a49b605f9daf" [ 696.528714] env[65726]: _type = "Task" [ 696.528714] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 696.535020] env[65726]: DEBUG oslo_vmware.api [None req-e9359f40-01a5-4a39-b7b9-47a0a48e2ede tempest-VolumesAdminNegativeTest-2004477850 tempest-VolumesAdminNegativeTest-2004477850-project-member] Waiting for the task: (returnval){ [ 696.535020] env[65726]: value = "task-5115641" [ 696.535020] env[65726]: _type = "Task" [ 696.535020] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 696.543713] env[65726]: DEBUG oslo_vmware.api [None req-60511481-93fa-4f8b-817d-01a577f85a26 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5291c029-5bf7-d17d-5ceb-a49b605f9daf, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 696.550663] env[65726]: DEBUG oslo_vmware.api [None req-e9359f40-01a5-4a39-b7b9-47a0a48e2ede tempest-VolumesAdminNegativeTest-2004477850 tempest-VolumesAdminNegativeTest-2004477850-project-member] Task: {'id': task-5115641, 'name': Rename_Task} progress is 6%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 696.647072] env[65726]: DEBUG oslo_vmware.api [None req-e30832c9-8006-4ee6-8d81-96e2fa3ed74f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]529cc391-6204-9749-5b88-c2bd7d46814e, 'name': SearchDatastore_Task, 'duration_secs': 0.01827} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 696.647850] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e30832c9-8006-4ee6-8d81-96e2fa3ed74f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 696.648925] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-e30832c9-8006-4ee6-8d81-96e2fa3ed74f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore2] ddd36db0-28ad-4582-994b-7eef8b08ed46/ddd36db0-28ad-4582-994b-7eef8b08ed46.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 696.649176] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ce8472a0-8b3b-4e55-997d-757a2f41dd4e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.658865] env[65726]: DEBUG oslo_vmware.api [None req-e30832c9-8006-4ee6-8d81-96e2fa3ed74f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] Waiting for the task: (returnval){ [ 696.658865] env[65726]: value = "task-5115642" [ 696.658865] env[65726]: _type = "Task" [ 696.658865] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 696.668935] env[65726]: DEBUG oslo_vmware.api [None req-e30832c9-8006-4ee6-8d81-96e2fa3ed74f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] Task: {'id': task-5115642, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 696.755861] env[65726]: DEBUG oslo_concurrency.lockutils [None req-f3878371-4599-418c-a420-84e54e420a24 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Releasing lock "refresh_cache-423af936-a383-4652-8887-9c8b8f507909" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 696.756402] env[65726]: DEBUG nova.compute.manager [None req-f3878371-4599-418c-a420-84e54e420a24 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] [instance: 423af936-a383-4652-8887-9c8b8f507909] Instance network_info: |[{"id": "11b095b5-d149-4b9c-a363-5e8acbc8262d", "address": "fa:16:3e:4d:ab:f3", "network": {"id": "55d629b8-a541-4720-9289-927bbc6e6f22", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.128", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "1539f06fa3534e90acbc3a60c4b8bd3f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "35342bcb-8b06-472e-b3c0-43fd3d6c4b30", "external-id": "nsx-vlan-transportzone-524", "segmentation_id": 524, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap11b095b5-d1", "ovs_interfaceid": "11b095b5-d149-4b9c-a363-5e8acbc8262d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 696.757062] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-f3878371-4599-418c-a420-84e54e420a24 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] [instance: 423af936-a383-4652-8887-9c8b8f507909] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:4d:ab:f3', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '35342bcb-8b06-472e-b3c0-43fd3d6c4b30', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '11b095b5-d149-4b9c-a363-5e8acbc8262d', 'vif_model': 'vmxnet3'}] {{(pid=65726) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 696.770574] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-f3878371-4599-418c-a420-84e54e420a24 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 696.771577] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 423af936-a383-4652-8887-9c8b8f507909] Creating VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 696.771577] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7841bde4-aefe-4814-b317-fd87ca1eb6c0 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.797625] env[65726]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 696.797625] env[65726]: value = "task-5115643" [ 696.797625] env[65726]: _type = "Task" [ 696.797625] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 696.810594] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115643, 'name': CreateVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 696.973216] env[65726]: DEBUG oslo_concurrency.lockutils [None req-34fc70a6-ee2b-4816-b933-3db54eee74ae tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.248s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 696.979515] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5dd7a34f-2002-4a12-9099-b0b630d2774d tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.945s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 696.980210] env[65726]: INFO nova.compute.claims [None req-5dd7a34f-2002-4a12-9099-b0b630d2774d tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] [instance: 815f056b-2d88-40dd-9ef7-a891ec8d92cf] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 697.010840] env[65726]: INFO nova.scheduler.client.report [None req-34fc70a6-ee2b-4816-b933-3db54eee74ae tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Deleted allocations for instance 811c1bbd-259c-4f5e-9a61-581bfba0ec4d [ 697.016222] env[65726]: WARNING openstack [req-892bee8d-2a7e-43c4-99a8-93139b745e96 req-a64b0cbe-c2d6-473f-9f86-13d77adb055e service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 697.016725] env[65726]: WARNING openstack [req-892bee8d-2a7e-43c4-99a8-93139b745e96 req-a64b0cbe-c2d6-473f-9f86-13d77adb055e service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 697.050476] env[65726]: DEBUG oslo_vmware.api [None req-60511481-93fa-4f8b-817d-01a577f85a26 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5291c029-5bf7-d17d-5ceb-a49b605f9daf, 'name': SearchDatastore_Task, 'duration_secs': 0.022663} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 697.061045] env[65726]: DEBUG oslo_vmware.api [None req-e9359f40-01a5-4a39-b7b9-47a0a48e2ede tempest-VolumesAdminNegativeTest-2004477850 tempest-VolumesAdminNegativeTest-2004477850-project-member] Task: {'id': task-5115641, 'name': Rename_Task, 'duration_secs': 0.345946} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 697.061045] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-104752e7-5b0b-40d4-838f-7214634dd3b6 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.067815] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-e9359f40-01a5-4a39-b7b9-47a0a48e2ede tempest-VolumesAdminNegativeTest-2004477850 tempest-VolumesAdminNegativeTest-2004477850-project-member] [instance: 30b17a17-229c-4ca7-9ae6-c67ccdcd3963] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 697.070444] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-73903ea3-c9b9-40dc-903f-f788a4214b9e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.086120] env[65726]: DEBUG oslo_vmware.api [None req-60511481-93fa-4f8b-817d-01a577f85a26 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Waiting for the task: (returnval){ [ 697.086120] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52051289-dbbe-552c-6104-e9173610bd3f" [ 697.086120] env[65726]: _type = "Task" [ 697.086120] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 697.094214] env[65726]: DEBUG oslo_vmware.api [None req-e9359f40-01a5-4a39-b7b9-47a0a48e2ede tempest-VolumesAdminNegativeTest-2004477850 tempest-VolumesAdminNegativeTest-2004477850-project-member] Waiting for the task: (returnval){ [ 697.094214] env[65726]: value = "task-5115644" [ 697.094214] env[65726]: _type = "Task" [ 697.094214] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 697.100623] env[65726]: DEBUG oslo_vmware.api [None req-60511481-93fa-4f8b-817d-01a577f85a26 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52051289-dbbe-552c-6104-e9173610bd3f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 697.109301] env[65726]: DEBUG oslo_vmware.api [None req-e9359f40-01a5-4a39-b7b9-47a0a48e2ede tempest-VolumesAdminNegativeTest-2004477850 tempest-VolumesAdminNegativeTest-2004477850-project-member] Task: {'id': task-5115644, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 697.171395] env[65726]: DEBUG oslo_vmware.api [None req-e30832c9-8006-4ee6-8d81-96e2fa3ed74f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] Task: {'id': task-5115642, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 697.201681] env[65726]: WARNING openstack [req-892bee8d-2a7e-43c4-99a8-93139b745e96 req-a64b0cbe-c2d6-473f-9f86-13d77adb055e service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 697.202127] env[65726]: WARNING openstack [req-892bee8d-2a7e-43c4-99a8-93139b745e96 req-a64b0cbe-c2d6-473f-9f86-13d77adb055e service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 697.312017] env[65726]: WARNING openstack [req-892bee8d-2a7e-43c4-99a8-93139b745e96 req-a64b0cbe-c2d6-473f-9f86-13d77adb055e service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 697.312972] env[65726]: WARNING openstack [req-892bee8d-2a7e-43c4-99a8-93139b745e96 req-a64b0cbe-c2d6-473f-9f86-13d77adb055e service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 697.336362] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115643, 'name': CreateVM_Task} progress is 25%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 697.434418] env[65726]: DEBUG nova.network.neutron [req-892bee8d-2a7e-43c4-99a8-93139b745e96 req-a64b0cbe-c2d6-473f-9f86-13d77adb055e service nova] [instance: 49bcab1f-7b64-4999-abff-37771c58a271] Updated VIF entry in instance network info cache for port 45612b7e-c052-4346-a779-5f94f9adc084. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 697.435164] env[65726]: DEBUG nova.network.neutron [req-892bee8d-2a7e-43c4-99a8-93139b745e96 req-a64b0cbe-c2d6-473f-9f86-13d77adb055e service nova] [instance: 49bcab1f-7b64-4999-abff-37771c58a271] Updating instance_info_cache with network_info: [{"id": "45612b7e-c052-4346-a779-5f94f9adc084", "address": "fa:16:3e:9d:95:60", "network": {"id": "51a66448-0f72-4266-8868-856b48be839f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-155658961-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d46ccb00794f458b85da4a93879139ee", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b1f3e6c3-5584-4852-9017-476ab8ac4946", "external-id": "nsx-vlan-transportzone-304", "segmentation_id": 304, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap45612b7e-c0", "ovs_interfaceid": "45612b7e-c052-4346-a779-5f94f9adc084", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 697.490877] env[65726]: DEBUG oslo_concurrency.lockutils [None req-f61c0733-7339-4ad4-9f78-b4a5ed0d0ff1 tempest-AttachInterfacesV270Test-133387050 tempest-AttachInterfacesV270Test-133387050-project-member] Acquiring lock "c977257e-dce0-4ecc-b42b-0f6ebab38797" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 697.491327] env[65726]: DEBUG oslo_concurrency.lockutils [None req-f61c0733-7339-4ad4-9f78-b4a5ed0d0ff1 tempest-AttachInterfacesV270Test-133387050 tempest-AttachInterfacesV270Test-133387050-project-member] Lock "c977257e-dce0-4ecc-b42b-0f6ebab38797" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 697.491526] env[65726]: DEBUG oslo_concurrency.lockutils [None req-f61c0733-7339-4ad4-9f78-b4a5ed0d0ff1 tempest-AttachInterfacesV270Test-133387050 tempest-AttachInterfacesV270Test-133387050-project-member] Acquiring lock "c977257e-dce0-4ecc-b42b-0f6ebab38797-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 697.491802] env[65726]: DEBUG oslo_concurrency.lockutils [None req-f61c0733-7339-4ad4-9f78-b4a5ed0d0ff1 tempest-AttachInterfacesV270Test-133387050 tempest-AttachInterfacesV270Test-133387050-project-member] Lock "c977257e-dce0-4ecc-b42b-0f6ebab38797-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 697.491977] env[65726]: DEBUG oslo_concurrency.lockutils [None req-f61c0733-7339-4ad4-9f78-b4a5ed0d0ff1 tempest-AttachInterfacesV270Test-133387050 tempest-AttachInterfacesV270Test-133387050-project-member] Lock "c977257e-dce0-4ecc-b42b-0f6ebab38797-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 697.495801] env[65726]: INFO nova.compute.manager [None req-f61c0733-7339-4ad4-9f78-b4a5ed0d0ff1 tempest-AttachInterfacesV270Test-133387050 tempest-AttachInterfacesV270Test-133387050-project-member] [instance: c977257e-dce0-4ecc-b42b-0f6ebab38797] Terminating instance [ 697.529132] env[65726]: DEBUG oslo_concurrency.lockutils [None req-34fc70a6-ee2b-4816-b933-3db54eee74ae tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Lock "811c1bbd-259c-4f5e-9a61-581bfba0ec4d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 28.547s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 697.602135] env[65726]: DEBUG oslo_vmware.api [None req-60511481-93fa-4f8b-817d-01a577f85a26 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52051289-dbbe-552c-6104-e9173610bd3f, 'name': SearchDatastore_Task, 'duration_secs': 0.012693} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 697.602774] env[65726]: DEBUG oslo_concurrency.lockutils [None req-60511481-93fa-4f8b-817d-01a577f85a26 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 697.603400] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-60511481-93fa-4f8b-817d-01a577f85a26 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] 49bcab1f-7b64-4999-abff-37771c58a271/49bcab1f-7b64-4999-abff-37771c58a271.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 697.603400] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e73c8f8f-66ee-4ed5-b8ef-91d8d73f5e10 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.609798] env[65726]: DEBUG oslo_vmware.api [None req-e9359f40-01a5-4a39-b7b9-47a0a48e2ede tempest-VolumesAdminNegativeTest-2004477850 tempest-VolumesAdminNegativeTest-2004477850-project-member] Task: {'id': task-5115644, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 697.615905] env[65726]: DEBUG oslo_vmware.api [None req-60511481-93fa-4f8b-817d-01a577f85a26 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Waiting for the task: (returnval){ [ 697.615905] env[65726]: value = "task-5115645" [ 697.615905] env[65726]: _type = "Task" [ 697.615905] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 697.624829] env[65726]: DEBUG oslo_vmware.api [None req-60511481-93fa-4f8b-817d-01a577f85a26 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': task-5115645, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 697.670213] env[65726]: DEBUG oslo_vmware.api [None req-e30832c9-8006-4ee6-8d81-96e2fa3ed74f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] Task: {'id': task-5115642, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.843286} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 697.670506] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-e30832c9-8006-4ee6-8d81-96e2fa3ed74f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore2] ddd36db0-28ad-4582-994b-7eef8b08ed46/ddd36db0-28ad-4582-994b-7eef8b08ed46.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 697.670718] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-e30832c9-8006-4ee6-8d81-96e2fa3ed74f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] [instance: ddd36db0-28ad-4582-994b-7eef8b08ed46] Extending root virtual disk to 1048576 {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 697.670977] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-081fa9d9-136d-4b48-9ede-f6e822beb67a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.679184] env[65726]: DEBUG oslo_vmware.api [None req-e30832c9-8006-4ee6-8d81-96e2fa3ed74f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] Waiting for the task: (returnval){ [ 697.679184] env[65726]: value = "task-5115646" [ 697.679184] env[65726]: _type = "Task" [ 697.679184] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 697.687956] env[65726]: DEBUG oslo_vmware.api [None req-e30832c9-8006-4ee6-8d81-96e2fa3ed74f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] Task: {'id': task-5115646, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 697.809938] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115643, 'name': CreateVM_Task, 'duration_secs': 0.794766} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 697.810090] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 423af936-a383-4652-8887-9c8b8f507909] Created VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 697.810857] env[65726]: WARNING openstack [None req-f3878371-4599-418c-a420-84e54e420a24 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 697.811400] env[65726]: WARNING openstack [None req-f3878371-4599-418c-a420-84e54e420a24 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 697.816482] env[65726]: DEBUG oslo_concurrency.lockutils [None req-f3878371-4599-418c-a420-84e54e420a24 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 697.816649] env[65726]: DEBUG oslo_concurrency.lockutils [None req-f3878371-4599-418c-a420-84e54e420a24 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 697.817964] env[65726]: DEBUG oslo_concurrency.lockutils [None req-f3878371-4599-418c-a420-84e54e420a24 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 697.817964] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bb472056-9ccd-416a-b81b-662c6058306b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.827029] env[65726]: DEBUG oslo_vmware.api [None req-f3878371-4599-418c-a420-84e54e420a24 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Waiting for the task: (returnval){ [ 697.827029] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52796eae-e080-2438-6d20-0f2354e24163" [ 697.827029] env[65726]: _type = "Task" [ 697.827029] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 697.833141] env[65726]: DEBUG oslo_vmware.api [None req-f3878371-4599-418c-a420-84e54e420a24 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52796eae-e080-2438-6d20-0f2354e24163, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 697.922201] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c19da911-fb8f-4bec-a53e-39a31fa2d36e tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] Acquiring lock "3b933684-ac19-44b0-a49d-6af45501e166" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 697.922467] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c19da911-fb8f-4bec-a53e-39a31fa2d36e tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] Lock "3b933684-ac19-44b0-a49d-6af45501e166" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 697.922671] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c19da911-fb8f-4bec-a53e-39a31fa2d36e tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] Acquiring lock "3b933684-ac19-44b0-a49d-6af45501e166-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 697.922868] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c19da911-fb8f-4bec-a53e-39a31fa2d36e tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] Lock "3b933684-ac19-44b0-a49d-6af45501e166-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 697.923017] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c19da911-fb8f-4bec-a53e-39a31fa2d36e tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] Lock "3b933684-ac19-44b0-a49d-6af45501e166-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 697.926765] env[65726]: INFO nova.compute.manager [None req-c19da911-fb8f-4bec-a53e-39a31fa2d36e tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] [instance: 3b933684-ac19-44b0-a49d-6af45501e166] Terminating instance [ 697.938020] env[65726]: DEBUG oslo_concurrency.lockutils [req-892bee8d-2a7e-43c4-99a8-93139b745e96 req-a64b0cbe-c2d6-473f-9f86-13d77adb055e service nova] Releasing lock "refresh_cache-49bcab1f-7b64-4999-abff-37771c58a271" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 698.002847] env[65726]: DEBUG nova.compute.manager [None req-f61c0733-7339-4ad4-9f78-b4a5ed0d0ff1 tempest-AttachInterfacesV270Test-133387050 tempest-AttachInterfacesV270Test-133387050-project-member] [instance: c977257e-dce0-4ecc-b42b-0f6ebab38797] Start destroying the instance on the hypervisor. {{(pid=65726) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 698.003272] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-f61c0733-7339-4ad4-9f78-b4a5ed0d0ff1 tempest-AttachInterfacesV270Test-133387050 tempest-AttachInterfacesV270Test-133387050-project-member] [instance: c977257e-dce0-4ecc-b42b-0f6ebab38797] Destroying instance {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 698.004218] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3753f8a-b9f3-416c-a0b0-65e2e8a1a29a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.014701] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-f61c0733-7339-4ad4-9f78-b4a5ed0d0ff1 tempest-AttachInterfacesV270Test-133387050 tempest-AttachInterfacesV270Test-133387050-project-member] [instance: c977257e-dce0-4ecc-b42b-0f6ebab38797] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 698.016907] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-cb6f4b32-8418-4193-9963-482abfdca19f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.025440] env[65726]: DEBUG oslo_concurrency.lockutils [None req-6c979061-2d03-4525-8b87-afa801f4542e tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Acquiring lock "6d2b86c2-0a44-433f-8b3f-2e9c7f693baa" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 698.025676] env[65726]: DEBUG oslo_concurrency.lockutils [None req-6c979061-2d03-4525-8b87-afa801f4542e tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Lock "6d2b86c2-0a44-433f-8b3f-2e9c7f693baa" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 698.034265] env[65726]: DEBUG oslo_vmware.api [None req-f61c0733-7339-4ad4-9f78-b4a5ed0d0ff1 tempest-AttachInterfacesV270Test-133387050 tempest-AttachInterfacesV270Test-133387050-project-member] Waiting for the task: (returnval){ [ 698.034265] env[65726]: value = "task-5115647" [ 698.034265] env[65726]: _type = "Task" [ 698.034265] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 698.047191] env[65726]: DEBUG oslo_vmware.api [None req-f61c0733-7339-4ad4-9f78-b4a5ed0d0ff1 tempest-AttachInterfacesV270Test-133387050 tempest-AttachInterfacesV270Test-133387050-project-member] Task: {'id': task-5115647, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 698.106514] env[65726]: DEBUG oslo_vmware.api [None req-e9359f40-01a5-4a39-b7b9-47a0a48e2ede tempest-VolumesAdminNegativeTest-2004477850 tempest-VolumesAdminNegativeTest-2004477850-project-member] Task: {'id': task-5115644, 'name': PowerOnVM_Task} progress is 88%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 698.132721] env[65726]: DEBUG oslo_vmware.api [None req-60511481-93fa-4f8b-817d-01a577f85a26 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': task-5115645, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 698.196980] env[65726]: DEBUG oslo_vmware.api [None req-e30832c9-8006-4ee6-8d81-96e2fa3ed74f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] Task: {'id': task-5115646, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.076359} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 698.197406] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-e30832c9-8006-4ee6-8d81-96e2fa3ed74f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] [instance: ddd36db0-28ad-4582-994b-7eef8b08ed46] Extended root virtual disk {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 698.198247] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbe8daa8-4824-4171-bb0e-b511cf03b509 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.228909] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-e30832c9-8006-4ee6-8d81-96e2fa3ed74f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] [instance: ddd36db0-28ad-4582-994b-7eef8b08ed46] Reconfiguring VM instance instance-00000019 to attach disk [datastore2] ddd36db0-28ad-4582-994b-7eef8b08ed46/ddd36db0-28ad-4582-994b-7eef8b08ed46.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 698.231963] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-819f7b2b-58d3-4a00-97ed-c4694e8953fa {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.256234] env[65726]: DEBUG oslo_vmware.api [None req-e30832c9-8006-4ee6-8d81-96e2fa3ed74f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] Waiting for the task: (returnval){ [ 698.256234] env[65726]: value = "task-5115648" [ 698.256234] env[65726]: _type = "Task" [ 698.256234] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 698.269601] env[65726]: DEBUG oslo_vmware.api [None req-e30832c9-8006-4ee6-8d81-96e2fa3ed74f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] Task: {'id': task-5115648, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 698.337690] env[65726]: DEBUG oslo_vmware.api [None req-f3878371-4599-418c-a420-84e54e420a24 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52796eae-e080-2438-6d20-0f2354e24163, 'name': SearchDatastore_Task, 'duration_secs': 0.015155} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 698.341010] env[65726]: DEBUG oslo_concurrency.lockutils [None req-f3878371-4599-418c-a420-84e54e420a24 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 698.342494] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-f3878371-4599-418c-a420-84e54e420a24 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] [instance: 423af936-a383-4652-8887-9c8b8f507909] Processing image b52362a3-ee8a-4cbf-b06f-513b0cc8f95c {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 698.342494] env[65726]: DEBUG oslo_concurrency.lockutils [None req-f3878371-4599-418c-a420-84e54e420a24 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 698.342494] env[65726]: DEBUG oslo_concurrency.lockutils [None req-f3878371-4599-418c-a420-84e54e420a24 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 698.342494] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-f3878371-4599-418c-a420-84e54e420a24 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 698.342826] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e8a608da-9c4a-472c-bc9d-c277a350921e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.362618] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-f3878371-4599-418c-a420-84e54e420a24 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 698.362618] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-f3878371-4599-418c-a420-84e54e420a24 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65726) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 698.366128] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c19733f8-681a-4479-ada9-89f50d413d6b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.373976] env[65726]: DEBUG oslo_vmware.api [None req-f3878371-4599-418c-a420-84e54e420a24 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Waiting for the task: (returnval){ [ 698.373976] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52d6bd01-6bba-c45a-8825-fac6835d6b5b" [ 698.373976] env[65726]: _type = "Task" [ 698.373976] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 698.384950] env[65726]: DEBUG oslo_vmware.api [None req-f3878371-4599-418c-a420-84e54e420a24 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52d6bd01-6bba-c45a-8825-fac6835d6b5b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 698.430038] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c19da911-fb8f-4bec-a53e-39a31fa2d36e tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] Acquiring lock "refresh_cache-3b933684-ac19-44b0-a49d-6af45501e166" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 698.430251] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c19da911-fb8f-4bec-a53e-39a31fa2d36e tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] Acquired lock "refresh_cache-3b933684-ac19-44b0-a49d-6af45501e166" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 698.430433] env[65726]: DEBUG nova.network.neutron [None req-c19da911-fb8f-4bec-a53e-39a31fa2d36e tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] [instance: 3b933684-ac19-44b0-a49d-6af45501e166] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 698.547749] env[65726]: DEBUG oslo_vmware.api [None req-f61c0733-7339-4ad4-9f78-b4a5ed0d0ff1 tempest-AttachInterfacesV270Test-133387050 tempest-AttachInterfacesV270Test-133387050-project-member] Task: {'id': task-5115647, 'name': PowerOffVM_Task, 'duration_secs': 0.372189} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 698.552604] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-f61c0733-7339-4ad4-9f78-b4a5ed0d0ff1 tempest-AttachInterfacesV270Test-133387050 tempest-AttachInterfacesV270Test-133387050-project-member] [instance: c977257e-dce0-4ecc-b42b-0f6ebab38797] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 698.552791] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-f61c0733-7339-4ad4-9f78-b4a5ed0d0ff1 tempest-AttachInterfacesV270Test-133387050 tempest-AttachInterfacesV270Test-133387050-project-member] [instance: c977257e-dce0-4ecc-b42b-0f6ebab38797] Unregistering the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 698.553476] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ebc09b9f-776a-445a-bba3-5875ea40d13f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.606746] env[65726]: DEBUG oslo_vmware.api [None req-e9359f40-01a5-4a39-b7b9-47a0a48e2ede tempest-VolumesAdminNegativeTest-2004477850 tempest-VolumesAdminNegativeTest-2004477850-project-member] Task: {'id': task-5115644, 'name': PowerOnVM_Task, 'duration_secs': 1.272008} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 698.610120] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-e9359f40-01a5-4a39-b7b9-47a0a48e2ede tempest-VolumesAdminNegativeTest-2004477850 tempest-VolumesAdminNegativeTest-2004477850-project-member] [instance: 30b17a17-229c-4ca7-9ae6-c67ccdcd3963] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 698.610501] env[65726]: INFO nova.compute.manager [None req-e9359f40-01a5-4a39-b7b9-47a0a48e2ede tempest-VolumesAdminNegativeTest-2004477850 tempest-VolumesAdminNegativeTest-2004477850-project-member] [instance: 30b17a17-229c-4ca7-9ae6-c67ccdcd3963] Took 9.79 seconds to spawn the instance on the hypervisor. [ 698.610730] env[65726]: DEBUG nova.compute.manager [None req-e9359f40-01a5-4a39-b7b9-47a0a48e2ede tempest-VolumesAdminNegativeTest-2004477850 tempest-VolumesAdminNegativeTest-2004477850-project-member] [instance: 30b17a17-229c-4ca7-9ae6-c67ccdcd3963] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 698.612038] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d072f520-18ad-4103-a448-c5b40d24e0a4 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.637485] env[65726]: DEBUG oslo_vmware.api [None req-60511481-93fa-4f8b-817d-01a577f85a26 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': task-5115645, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.787929} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 698.639170] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-60511481-93fa-4f8b-817d-01a577f85a26 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] 49bcab1f-7b64-4999-abff-37771c58a271/49bcab1f-7b64-4999-abff-37771c58a271.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 698.639610] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-60511481-93fa-4f8b-817d-01a577f85a26 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: 49bcab1f-7b64-4999-abff-37771c58a271] Extending root virtual disk to 1048576 {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 698.639821] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-f61c0733-7339-4ad4-9f78-b4a5ed0d0ff1 tempest-AttachInterfacesV270Test-133387050 tempest-AttachInterfacesV270Test-133387050-project-member] [instance: c977257e-dce0-4ecc-b42b-0f6ebab38797] Unregistered the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 698.640011] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-f61c0733-7339-4ad4-9f78-b4a5ed0d0ff1 tempest-AttachInterfacesV270Test-133387050 tempest-AttachInterfacesV270Test-133387050-project-member] [instance: c977257e-dce0-4ecc-b42b-0f6ebab38797] Deleting contents of the VM from datastore datastore2 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 698.640199] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-f61c0733-7339-4ad4-9f78-b4a5ed0d0ff1 tempest-AttachInterfacesV270Test-133387050 tempest-AttachInterfacesV270Test-133387050-project-member] Deleting the datastore file [datastore2] c977257e-dce0-4ecc-b42b-0f6ebab38797 {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 698.640470] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-069b10de-4c3f-4a98-9496-b4c5769d98c0 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.642765] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4c281e50-9b4f-4bfb-a012-04619cac2c1d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.650829] env[65726]: DEBUG oslo_vmware.api [None req-60511481-93fa-4f8b-817d-01a577f85a26 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Waiting for the task: (returnval){ [ 698.650829] env[65726]: value = "task-5115650" [ 698.650829] env[65726]: _type = "Task" [ 698.650829] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 698.652198] env[65726]: DEBUG oslo_vmware.api [None req-f61c0733-7339-4ad4-9f78-b4a5ed0d0ff1 tempest-AttachInterfacesV270Test-133387050 tempest-AttachInterfacesV270Test-133387050-project-member] Waiting for the task: (returnval){ [ 698.652198] env[65726]: value = "task-5115651" [ 698.652198] env[65726]: _type = "Task" [ 698.652198] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 698.664833] env[65726]: DEBUG oslo_vmware.api [None req-60511481-93fa-4f8b-817d-01a577f85a26 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': task-5115650, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 698.670728] env[65726]: DEBUG oslo_vmware.api [None req-f61c0733-7339-4ad4-9f78-b4a5ed0d0ff1 tempest-AttachInterfacesV270Test-133387050 tempest-AttachInterfacesV270Test-133387050-project-member] Task: {'id': task-5115651, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 698.713270] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ca014d8-91e6-4df7-8440-d050029dc102 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.722760] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63899906-9681-4a85-aa88-947c4e29e79a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.767035] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00786acd-4345-456c-8008-e1d49c45555b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.779186] env[65726]: DEBUG oslo_vmware.api [None req-e30832c9-8006-4ee6-8d81-96e2fa3ed74f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] Task: {'id': task-5115648, 'name': ReconfigVM_Task, 'duration_secs': 0.366372} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 698.780555] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41a63926-b495-4821-9e5b-a65212c573b1 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.784884] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-e30832c9-8006-4ee6-8d81-96e2fa3ed74f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] [instance: ddd36db0-28ad-4582-994b-7eef8b08ed46] Reconfigured VM instance instance-00000019 to attach disk [datastore2] ddd36db0-28ad-4582-994b-7eef8b08ed46/ddd36db0-28ad-4582-994b-7eef8b08ed46.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 698.785536] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-45af46d1-9b40-4656-a0e8-e40b098e0a13 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.800011] env[65726]: DEBUG nova.compute.provider_tree [None req-5dd7a34f-2002-4a12-9099-b0b630d2774d tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Updating inventory in ProviderTree for provider 07c4692f-bdb4-4058-9173-ff9664830295 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 698.803365] env[65726]: DEBUG oslo_vmware.api [None req-e30832c9-8006-4ee6-8d81-96e2fa3ed74f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] Waiting for the task: (returnval){ [ 698.803365] env[65726]: value = "task-5115652" [ 698.803365] env[65726]: _type = "Task" [ 698.803365] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 698.817689] env[65726]: DEBUG oslo_vmware.api [None req-e30832c9-8006-4ee6-8d81-96e2fa3ed74f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] Task: {'id': task-5115652, 'name': Rename_Task} progress is 10%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 698.885139] env[65726]: DEBUG oslo_vmware.api [None req-f3878371-4599-418c-a420-84e54e420a24 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52d6bd01-6bba-c45a-8825-fac6835d6b5b, 'name': SearchDatastore_Task, 'duration_secs': 0.013702} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 698.887565] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dcc5ca1c-4766-4fe8-aabd-eb20db973ce6 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.897187] env[65726]: DEBUG oslo_vmware.api [None req-f3878371-4599-418c-a420-84e54e420a24 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Waiting for the task: (returnval){ [ 698.897187] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52cd877e-c730-3ea8-e660-0afae8d16608" [ 698.897187] env[65726]: _type = "Task" [ 698.897187] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 698.907097] env[65726]: DEBUG oslo_vmware.api [None req-f3878371-4599-418c-a420-84e54e420a24 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52cd877e-c730-3ea8-e660-0afae8d16608, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 698.933410] env[65726]: WARNING openstack [None req-c19da911-fb8f-4bec-a53e-39a31fa2d36e tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 698.933805] env[65726]: WARNING openstack [None req-c19da911-fb8f-4bec-a53e-39a31fa2d36e tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 698.939181] env[65726]: DEBUG nova.network.neutron [None req-c19da911-fb8f-4bec-a53e-39a31fa2d36e tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] [instance: 3b933684-ac19-44b0-a49d-6af45501e166] Instance cache missing network info. {{(pid=65726) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 699.018406] env[65726]: DEBUG nova.compute.manager [req-ffb345b8-d69d-4740-8c54-5df420a64f4c req-010a4998-ac15-44dd-8ecb-b7fb94d85657 service nova] [instance: 423af936-a383-4652-8887-9c8b8f507909] Received event network-changed-11b095b5-d149-4b9c-a363-5e8acbc8262d {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 699.018669] env[65726]: DEBUG nova.compute.manager [req-ffb345b8-d69d-4740-8c54-5df420a64f4c req-010a4998-ac15-44dd-8ecb-b7fb94d85657 service nova] [instance: 423af936-a383-4652-8887-9c8b8f507909] Refreshing instance network info cache due to event network-changed-11b095b5-d149-4b9c-a363-5e8acbc8262d. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 699.018814] env[65726]: DEBUG oslo_concurrency.lockutils [req-ffb345b8-d69d-4740-8c54-5df420a64f4c req-010a4998-ac15-44dd-8ecb-b7fb94d85657 service nova] Acquiring lock "refresh_cache-423af936-a383-4652-8887-9c8b8f507909" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 699.018951] env[65726]: DEBUG oslo_concurrency.lockutils [req-ffb345b8-d69d-4740-8c54-5df420a64f4c req-010a4998-ac15-44dd-8ecb-b7fb94d85657 service nova] Acquired lock "refresh_cache-423af936-a383-4652-8887-9c8b8f507909" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 699.019150] env[65726]: DEBUG nova.network.neutron [req-ffb345b8-d69d-4740-8c54-5df420a64f4c req-010a4998-ac15-44dd-8ecb-b7fb94d85657 service nova] [instance: 423af936-a383-4652-8887-9c8b8f507909] Refreshing network info cache for port 11b095b5-d149-4b9c-a363-5e8acbc8262d {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 699.050619] env[65726]: DEBUG nova.network.neutron [None req-c19da911-fb8f-4bec-a53e-39a31fa2d36e tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] [instance: 3b933684-ac19-44b0-a49d-6af45501e166] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 699.140337] env[65726]: INFO nova.compute.manager [None req-e9359f40-01a5-4a39-b7b9-47a0a48e2ede tempest-VolumesAdminNegativeTest-2004477850 tempest-VolumesAdminNegativeTest-2004477850-project-member] [instance: 30b17a17-229c-4ca7-9ae6-c67ccdcd3963] Took 36.25 seconds to build instance. [ 699.167442] env[65726]: DEBUG oslo_vmware.api [None req-60511481-93fa-4f8b-817d-01a577f85a26 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': task-5115650, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.071002} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 699.170746] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-60511481-93fa-4f8b-817d-01a577f85a26 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: 49bcab1f-7b64-4999-abff-37771c58a271] Extended root virtual disk {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 699.171160] env[65726]: DEBUG oslo_vmware.api [None req-f61c0733-7339-4ad4-9f78-b4a5ed0d0ff1 tempest-AttachInterfacesV270Test-133387050 tempest-AttachInterfacesV270Test-133387050-project-member] Task: {'id': task-5115651, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 699.172211] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bf9d22b-c3ec-4d84-8374-41d2a3735f08 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.195253] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-60511481-93fa-4f8b-817d-01a577f85a26 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: 49bcab1f-7b64-4999-abff-37771c58a271] Reconfiguring VM instance instance-0000001c to attach disk [datastore1] 49bcab1f-7b64-4999-abff-37771c58a271/49bcab1f-7b64-4999-abff-37771c58a271.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 699.195571] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ba6ef258-f25f-4452-8251-55b1274f5bb0 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.218557] env[65726]: DEBUG oslo_vmware.api [None req-60511481-93fa-4f8b-817d-01a577f85a26 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Waiting for the task: (returnval){ [ 699.218557] env[65726]: value = "task-5115653" [ 699.218557] env[65726]: _type = "Task" [ 699.218557] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 699.228511] env[65726]: DEBUG oslo_vmware.api [None req-60511481-93fa-4f8b-817d-01a577f85a26 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': task-5115653, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 699.318875] env[65726]: DEBUG oslo_vmware.api [None req-e30832c9-8006-4ee6-8d81-96e2fa3ed74f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] Task: {'id': task-5115652, 'name': Rename_Task, 'duration_secs': 0.155395} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 699.319425] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-e30832c9-8006-4ee6-8d81-96e2fa3ed74f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] [instance: ddd36db0-28ad-4582-994b-7eef8b08ed46] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 699.319734] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-891e74f1-b221-4456-ab73-ccc06771d490 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.327950] env[65726]: DEBUG oslo_vmware.api [None req-e30832c9-8006-4ee6-8d81-96e2fa3ed74f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] Waiting for the task: (returnval){ [ 699.327950] env[65726]: value = "task-5115654" [ 699.327950] env[65726]: _type = "Task" [ 699.327950] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 699.333015] env[65726]: ERROR nova.scheduler.client.report [None req-5dd7a34f-2002-4a12-9099-b0b630d2774d tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] [req-4c9a430e-00f1-4ddc-bb16-3704be0a856c] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 07c4692f-bdb4-4058-9173-ff9664830295. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-4c9a430e-00f1-4ddc-bb16-3704be0a856c"}]} [ 699.341576] env[65726]: DEBUG oslo_vmware.api [None req-e30832c9-8006-4ee6-8d81-96e2fa3ed74f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] Task: {'id': task-5115654, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 699.356157] env[65726]: DEBUG nova.scheduler.client.report [None req-5dd7a34f-2002-4a12-9099-b0b630d2774d tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Refreshing inventories for resource provider 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 699.372935] env[65726]: DEBUG nova.scheduler.client.report [None req-5dd7a34f-2002-4a12-9099-b0b630d2774d tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Updating ProviderTree inventory for provider 07c4692f-bdb4-4058-9173-ff9664830295 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 699.373255] env[65726]: DEBUG nova.compute.provider_tree [None req-5dd7a34f-2002-4a12-9099-b0b630d2774d tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Updating inventory in ProviderTree for provider 07c4692f-bdb4-4058-9173-ff9664830295 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 699.389924] env[65726]: DEBUG nova.scheduler.client.report [None req-5dd7a34f-2002-4a12-9099-b0b630d2774d tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Refreshing aggregate associations for resource provider 07c4692f-bdb4-4058-9173-ff9664830295, aggregates: None {{(pid=65726) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 699.408619] env[65726]: DEBUG oslo_vmware.api [None req-f3878371-4599-418c-a420-84e54e420a24 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52cd877e-c730-3ea8-e660-0afae8d16608, 'name': SearchDatastore_Task, 'duration_secs': 0.033289} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 699.408936] env[65726]: DEBUG oslo_concurrency.lockutils [None req-f3878371-4599-418c-a420-84e54e420a24 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 699.409252] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-f3878371-4599-418c-a420-84e54e420a24 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore2] 423af936-a383-4652-8887-9c8b8f507909/423af936-a383-4652-8887-9c8b8f507909.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 699.409503] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-cbd10bc4-b266-4c7b-a45a-4d32fdcad718 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.412478] env[65726]: DEBUG nova.scheduler.client.report [None req-5dd7a34f-2002-4a12-9099-b0b630d2774d tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Refreshing trait associations for resource provider 07c4692f-bdb4-4058-9173-ff9664830295, traits: COMPUTE_IMAGE_TYPE_ISO,HW_ARCH_X86_64,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=65726) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 699.421429] env[65726]: DEBUG oslo_vmware.api [None req-f3878371-4599-418c-a420-84e54e420a24 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Waiting for the task: (returnval){ [ 699.421429] env[65726]: value = "task-5115655" [ 699.421429] env[65726]: _type = "Task" [ 699.421429] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 699.432092] env[65726]: DEBUG oslo_vmware.api [None req-f3878371-4599-418c-a420-84e54e420a24 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Task: {'id': task-5115655, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 699.524089] env[65726]: WARNING openstack [req-ffb345b8-d69d-4740-8c54-5df420a64f4c req-010a4998-ac15-44dd-8ecb-b7fb94d85657 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 699.524389] env[65726]: WARNING openstack [req-ffb345b8-d69d-4740-8c54-5df420a64f4c req-010a4998-ac15-44dd-8ecb-b7fb94d85657 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 699.555877] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c19da911-fb8f-4bec-a53e-39a31fa2d36e tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] Releasing lock "refresh_cache-3b933684-ac19-44b0-a49d-6af45501e166" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 699.556352] env[65726]: DEBUG nova.compute.manager [None req-c19da911-fb8f-4bec-a53e-39a31fa2d36e tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] [instance: 3b933684-ac19-44b0-a49d-6af45501e166] Start destroying the instance on the hypervisor. {{(pid=65726) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 699.557044] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-c19da911-fb8f-4bec-a53e-39a31fa2d36e tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] [instance: 3b933684-ac19-44b0-a49d-6af45501e166] Destroying instance {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 699.558071] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63f3a66f-5d07-4a85-bdc0-5eec626052f8 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.566881] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-c19da911-fb8f-4bec-a53e-39a31fa2d36e tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] [instance: 3b933684-ac19-44b0-a49d-6af45501e166] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 699.571577] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a534cee8-4d45-4ff7-bc11-4c6901adce76 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.581276] env[65726]: DEBUG oslo_vmware.api [None req-c19da911-fb8f-4bec-a53e-39a31fa2d36e tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] Waiting for the task: (returnval){ [ 699.581276] env[65726]: value = "task-5115656" [ 699.581276] env[65726]: _type = "Task" [ 699.581276] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 699.590785] env[65726]: DEBUG oslo_vmware.api [None req-c19da911-fb8f-4bec-a53e-39a31fa2d36e tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] Task: {'id': task-5115656, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 699.644441] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e9359f40-01a5-4a39-b7b9-47a0a48e2ede tempest-VolumesAdminNegativeTest-2004477850 tempest-VolumesAdminNegativeTest-2004477850-project-member] Lock "30b17a17-229c-4ca7-9ae6-c67ccdcd3963" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 56.780s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 699.677922] env[65726]: DEBUG oslo_vmware.api [None req-f61c0733-7339-4ad4-9f78-b4a5ed0d0ff1 tempest-AttachInterfacesV270Test-133387050 tempest-AttachInterfacesV270Test-133387050-project-member] Task: {'id': task-5115651, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.581599} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 699.681519] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-f61c0733-7339-4ad4-9f78-b4a5ed0d0ff1 tempest-AttachInterfacesV270Test-133387050 tempest-AttachInterfacesV270Test-133387050-project-member] Deleted the datastore file {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 699.683333] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-f61c0733-7339-4ad4-9f78-b4a5ed0d0ff1 tempest-AttachInterfacesV270Test-133387050 tempest-AttachInterfacesV270Test-133387050-project-member] [instance: c977257e-dce0-4ecc-b42b-0f6ebab38797] Deleted contents of the VM from datastore datastore2 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 699.683333] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-f61c0733-7339-4ad4-9f78-b4a5ed0d0ff1 tempest-AttachInterfacesV270Test-133387050 tempest-AttachInterfacesV270Test-133387050-project-member] [instance: c977257e-dce0-4ecc-b42b-0f6ebab38797] Instance destroyed {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 699.683333] env[65726]: INFO nova.compute.manager [None req-f61c0733-7339-4ad4-9f78-b4a5ed0d0ff1 tempest-AttachInterfacesV270Test-133387050 tempest-AttachInterfacesV270Test-133387050-project-member] [instance: c977257e-dce0-4ecc-b42b-0f6ebab38797] Took 1.68 seconds to destroy the instance on the hypervisor. [ 699.683333] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-f61c0733-7339-4ad4-9f78-b4a5ed0d0ff1 tempest-AttachInterfacesV270Test-133387050 tempest-AttachInterfacesV270Test-133387050-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 699.683333] env[65726]: DEBUG nova.compute.manager [-] [instance: c977257e-dce0-4ecc-b42b-0f6ebab38797] Deallocating network for instance {{(pid=65726) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 699.683333] env[65726]: DEBUG nova.network.neutron [-] [instance: c977257e-dce0-4ecc-b42b-0f6ebab38797] deallocate_for_instance() {{(pid=65726) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 699.683701] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 699.683809] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 699.694884] env[65726]: WARNING openstack [req-ffb345b8-d69d-4740-8c54-5df420a64f4c req-010a4998-ac15-44dd-8ecb-b7fb94d85657 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 699.695408] env[65726]: WARNING openstack [req-ffb345b8-d69d-4740-8c54-5df420a64f4c req-010a4998-ac15-44dd-8ecb-b7fb94d85657 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 699.738090] env[65726]: DEBUG oslo_vmware.api [None req-60511481-93fa-4f8b-817d-01a577f85a26 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': task-5115653, 'name': ReconfigVM_Task, 'duration_secs': 0.418865} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 699.743253] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 699.744038] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 699.756281] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-60511481-93fa-4f8b-817d-01a577f85a26 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: 49bcab1f-7b64-4999-abff-37771c58a271] Reconfigured VM instance instance-0000001c to attach disk [datastore1] 49bcab1f-7b64-4999-abff-37771c58a271/49bcab1f-7b64-4999-abff-37771c58a271.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 699.760826] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4f40ea49-6799-4d0d-95b3-86683b022b9f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.785047] env[65726]: DEBUG oslo_vmware.api [None req-60511481-93fa-4f8b-817d-01a577f85a26 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Waiting for the task: (returnval){ [ 699.785047] env[65726]: value = "task-5115657" [ 699.785047] env[65726]: _type = "Task" [ 699.785047] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 699.798418] env[65726]: DEBUG oslo_vmware.api [None req-60511481-93fa-4f8b-817d-01a577f85a26 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': task-5115657, 'name': Rename_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 699.829868] env[65726]: WARNING openstack [req-ffb345b8-d69d-4740-8c54-5df420a64f4c req-010a4998-ac15-44dd-8ecb-b7fb94d85657 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 699.830648] env[65726]: WARNING openstack [req-ffb345b8-d69d-4740-8c54-5df420a64f4c req-010a4998-ac15-44dd-8ecb-b7fb94d85657 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 699.867072] env[65726]: DEBUG oslo_vmware.api [None req-e30832c9-8006-4ee6-8d81-96e2fa3ed74f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] Task: {'id': task-5115654, 'name': PowerOnVM_Task, 'duration_secs': 0.498137} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 699.867072] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-e30832c9-8006-4ee6-8d81-96e2fa3ed74f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] [instance: ddd36db0-28ad-4582-994b-7eef8b08ed46] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 699.867282] env[65726]: DEBUG nova.compute.manager [None req-e30832c9-8006-4ee6-8d81-96e2fa3ed74f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] [instance: ddd36db0-28ad-4582-994b-7eef8b08ed46] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 699.868148] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-074b1d9d-870a-4261-9231-dcd453578ea8 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.941611] env[65726]: DEBUG oslo_vmware.api [None req-f3878371-4599-418c-a420-84e54e420a24 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Task: {'id': task-5115655, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 700.010180] env[65726]: DEBUG nova.network.neutron [req-ffb345b8-d69d-4740-8c54-5df420a64f4c req-010a4998-ac15-44dd-8ecb-b7fb94d85657 service nova] [instance: 423af936-a383-4652-8887-9c8b8f507909] Updated VIF entry in instance network info cache for port 11b095b5-d149-4b9c-a363-5e8acbc8262d. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 700.010770] env[65726]: DEBUG nova.network.neutron [req-ffb345b8-d69d-4740-8c54-5df420a64f4c req-010a4998-ac15-44dd-8ecb-b7fb94d85657 service nova] [instance: 423af936-a383-4652-8887-9c8b8f507909] Updating instance_info_cache with network_info: [{"id": "11b095b5-d149-4b9c-a363-5e8acbc8262d", "address": "fa:16:3e:4d:ab:f3", "network": {"id": "55d629b8-a541-4720-9289-927bbc6e6f22", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.128", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "1539f06fa3534e90acbc3a60c4b8bd3f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "35342bcb-8b06-472e-b3c0-43fd3d6c4b30", "external-id": "nsx-vlan-transportzone-524", "segmentation_id": 524, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap11b095b5-d1", "ovs_interfaceid": "11b095b5-d149-4b9c-a363-5e8acbc8262d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 700.090300] env[65726]: DEBUG oslo_vmware.api [None req-c19da911-fb8f-4bec-a53e-39a31fa2d36e tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] Task: {'id': task-5115656, 'name': PowerOffVM_Task, 'duration_secs': 0.152238} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 700.090729] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-c19da911-fb8f-4bec-a53e-39a31fa2d36e tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] [instance: 3b933684-ac19-44b0-a49d-6af45501e166] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 700.090967] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-c19da911-fb8f-4bec-a53e-39a31fa2d36e tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] [instance: 3b933684-ac19-44b0-a49d-6af45501e166] Unregistering the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 700.091365] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-cd7ef4d7-758b-47d2-ab26-eda4265298e1 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.119598] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-c19da911-fb8f-4bec-a53e-39a31fa2d36e tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] [instance: 3b933684-ac19-44b0-a49d-6af45501e166] Unregistered the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 700.120350] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-c19da911-fb8f-4bec-a53e-39a31fa2d36e tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] [instance: 3b933684-ac19-44b0-a49d-6af45501e166] Deleting contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 700.120350] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-c19da911-fb8f-4bec-a53e-39a31fa2d36e tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] Deleting the datastore file [datastore1] 3b933684-ac19-44b0-a49d-6af45501e166 {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 700.121463] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5bf95456-bf7e-4d86-850a-24f53c2a9d59 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.126991] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe24b36c-f3fb-473b-b44f-db4a1b73626c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.131943] env[65726]: DEBUG oslo_vmware.api [None req-c19da911-fb8f-4bec-a53e-39a31fa2d36e tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] Waiting for the task: (returnval){ [ 700.131943] env[65726]: value = "task-5115659" [ 700.131943] env[65726]: _type = "Task" [ 700.131943] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 700.139454] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-796ceb63-2991-492e-adae-1c7ab6dccdd6 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.146826] env[65726]: DEBUG oslo_vmware.api [None req-c19da911-fb8f-4bec-a53e-39a31fa2d36e tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] Task: {'id': task-5115659, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 700.147739] env[65726]: DEBUG nova.compute.manager [None req-b0adde83-84f2-4fa3-84bf-8f387eee1293 tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] [instance: 3827eed1-b7a1-4922-a4de-f0c038cb8566] Starting instance... {{(pid=65726) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 700.179106] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-954a556e-8cb3-4e08-b22a-ab9eb330b66f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.187973] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-581c0116-51c9-4137-91f5-fb61093f3802 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.203519] env[65726]: DEBUG nova.compute.provider_tree [None req-5dd7a34f-2002-4a12-9099-b0b630d2774d tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Updating inventory in ProviderTree for provider 07c4692f-bdb4-4058-9173-ff9664830295 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 700.293748] env[65726]: DEBUG oslo_vmware.api [None req-60511481-93fa-4f8b-817d-01a577f85a26 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': task-5115657, 'name': Rename_Task, 'duration_secs': 0.230607} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 700.293853] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-60511481-93fa-4f8b-817d-01a577f85a26 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: 49bcab1f-7b64-4999-abff-37771c58a271] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 700.294065] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a76da629-0167-4214-b7d7-916fd4a8fcbf {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.302820] env[65726]: DEBUG oslo_vmware.api [None req-60511481-93fa-4f8b-817d-01a577f85a26 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Waiting for the task: (returnval){ [ 700.302820] env[65726]: value = "task-5115660" [ 700.302820] env[65726]: _type = "Task" [ 700.302820] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 700.315389] env[65726]: DEBUG oslo_vmware.api [None req-60511481-93fa-4f8b-817d-01a577f85a26 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': task-5115660, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 700.393830] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e30832c9-8006-4ee6-8d81-96e2fa3ed74f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 700.437602] env[65726]: DEBUG oslo_vmware.api [None req-f3878371-4599-418c-a420-84e54e420a24 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Task: {'id': task-5115655, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.636153} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 700.438592] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-f3878371-4599-418c-a420-84e54e420a24 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore2] 423af936-a383-4652-8887-9c8b8f507909/423af936-a383-4652-8887-9c8b8f507909.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 700.439100] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-f3878371-4599-418c-a420-84e54e420a24 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] [instance: 423af936-a383-4652-8887-9c8b8f507909] Extending root virtual disk to 1048576 {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 700.442184] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-501ee243-e593-4602-9a65-0d91c6126a1a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.448837] env[65726]: DEBUG oslo_vmware.api [None req-f3878371-4599-418c-a420-84e54e420a24 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Waiting for the task: (returnval){ [ 700.448837] env[65726]: value = "task-5115661" [ 700.448837] env[65726]: _type = "Task" [ 700.448837] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 700.459943] env[65726]: DEBUG oslo_vmware.api [None req-f3878371-4599-418c-a420-84e54e420a24 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Task: {'id': task-5115661, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 700.486974] env[65726]: DEBUG oslo_concurrency.lockutils [None req-d51d38e8-71be-4889-a206-d52403737076 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Acquiring lock "9fc0c9ee-9dc7-4059-b8d2-a5db965e00c2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 700.487316] env[65726]: DEBUG oslo_concurrency.lockutils [None req-d51d38e8-71be-4889-a206-d52403737076 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Lock "9fc0c9ee-9dc7-4059-b8d2-a5db965e00c2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 700.514275] env[65726]: DEBUG oslo_concurrency.lockutils [req-ffb345b8-d69d-4740-8c54-5df420a64f4c req-010a4998-ac15-44dd-8ecb-b7fb94d85657 service nova] Releasing lock "refresh_cache-423af936-a383-4652-8887-9c8b8f507909" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 700.612191] env[65726]: DEBUG nova.network.neutron [-] [instance: c977257e-dce0-4ecc-b42b-0f6ebab38797] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 700.645648] env[65726]: DEBUG oslo_vmware.api [None req-c19da911-fb8f-4bec-a53e-39a31fa2d36e tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] Task: {'id': task-5115659, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.223469} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 700.647121] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-c19da911-fb8f-4bec-a53e-39a31fa2d36e tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] Deleted the datastore file {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 700.647121] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-c19da911-fb8f-4bec-a53e-39a31fa2d36e tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] [instance: 3b933684-ac19-44b0-a49d-6af45501e166] Deleted contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 700.647121] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-c19da911-fb8f-4bec-a53e-39a31fa2d36e tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] [instance: 3b933684-ac19-44b0-a49d-6af45501e166] Instance destroyed {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 700.647266] env[65726]: INFO nova.compute.manager [None req-c19da911-fb8f-4bec-a53e-39a31fa2d36e tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] [instance: 3b933684-ac19-44b0-a49d-6af45501e166] Took 1.09 seconds to destroy the instance on the hypervisor. [ 700.647600] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-c19da911-fb8f-4bec-a53e-39a31fa2d36e tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 700.647806] env[65726]: DEBUG nova.compute.manager [-] [instance: 3b933684-ac19-44b0-a49d-6af45501e166] Deallocating network for instance {{(pid=65726) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 700.647896] env[65726]: DEBUG nova.network.neutron [-] [instance: 3b933684-ac19-44b0-a49d-6af45501e166] deallocate_for_instance() {{(pid=65726) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 700.648410] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 700.648655] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 700.702099] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b0adde83-84f2-4fa3-84bf-8f387eee1293 tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 700.710560] env[65726]: DEBUG nova.network.neutron [-] [instance: 3b933684-ac19-44b0-a49d-6af45501e166] Instance cache missing network info. {{(pid=65726) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 700.711118] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 700.711574] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 700.741690] env[65726]: ERROR nova.scheduler.client.report [None req-5dd7a34f-2002-4a12-9099-b0b630d2774d tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] [req-f491c6dd-6c51-4376-aca7-f8089695b43c] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 07c4692f-bdb4-4058-9173-ff9664830295. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-f491c6dd-6c51-4376-aca7-f8089695b43c"}]} [ 700.772037] env[65726]: DEBUG nova.scheduler.client.report [None req-5dd7a34f-2002-4a12-9099-b0b630d2774d tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Refreshing inventories for resource provider 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 700.788153] env[65726]: DEBUG nova.scheduler.client.report [None req-5dd7a34f-2002-4a12-9099-b0b630d2774d tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Updating ProviderTree inventory for provider 07c4692f-bdb4-4058-9173-ff9664830295 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 700.789059] env[65726]: DEBUG nova.compute.provider_tree [None req-5dd7a34f-2002-4a12-9099-b0b630d2774d tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Updating inventory in ProviderTree for provider 07c4692f-bdb4-4058-9173-ff9664830295 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 700.820170] env[65726]: DEBUG nova.scheduler.client.report [None req-5dd7a34f-2002-4a12-9099-b0b630d2774d tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Refreshing aggregate associations for resource provider 07c4692f-bdb4-4058-9173-ff9664830295, aggregates: None {{(pid=65726) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 700.824315] env[65726]: DEBUG oslo_vmware.api [None req-60511481-93fa-4f8b-817d-01a577f85a26 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': task-5115660, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 700.851368] env[65726]: DEBUG nova.scheduler.client.report [None req-5dd7a34f-2002-4a12-9099-b0b630d2774d tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Refreshing trait associations for resource provider 07c4692f-bdb4-4058-9173-ff9664830295, traits: COMPUTE_IMAGE_TYPE_ISO,HW_ARCH_X86_64,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=65726) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 700.964961] env[65726]: DEBUG oslo_vmware.api [None req-f3878371-4599-418c-a420-84e54e420a24 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Task: {'id': task-5115661, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.071235} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 700.968073] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-f3878371-4599-418c-a420-84e54e420a24 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] [instance: 423af936-a383-4652-8887-9c8b8f507909] Extended root virtual disk {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 700.969708] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1519e9d9-2abd-4924-b944-72f6df563e75 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.994734] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-f3878371-4599-418c-a420-84e54e420a24 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] [instance: 423af936-a383-4652-8887-9c8b8f507909] Reconfiguring VM instance instance-0000001d to attach disk [datastore2] 423af936-a383-4652-8887-9c8b8f507909/423af936-a383-4652-8887-9c8b8f507909.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 700.997722] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1e001d44-2b6a-4111-a02b-5ca16f66f19e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.019856] env[65726]: DEBUG oslo_vmware.api [None req-f3878371-4599-418c-a420-84e54e420a24 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Waiting for the task: (returnval){ [ 701.019856] env[65726]: value = "task-5115662" [ 701.019856] env[65726]: _type = "Task" [ 701.019856] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 701.035849] env[65726]: DEBUG oslo_vmware.api [None req-f3878371-4599-418c-a420-84e54e420a24 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Task: {'id': task-5115662, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 701.117473] env[65726]: INFO nova.compute.manager [-] [instance: c977257e-dce0-4ecc-b42b-0f6ebab38797] Took 1.43 seconds to deallocate network for instance. [ 701.197108] env[65726]: DEBUG nova.compute.manager [req-d3490cba-4326-4ab2-b745-374600705b6e req-4ccd59e8-4537-4b0b-a46f-dc59e29ce8eb service nova] [instance: c977257e-dce0-4ecc-b42b-0f6ebab38797] Received event network-vif-deleted-881e11d6-dc73-4e2b-9d9b-eec7da46f63f {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 701.218965] env[65726]: DEBUG nova.network.neutron [-] [instance: 3b933684-ac19-44b0-a49d-6af45501e166] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 701.317928] env[65726]: DEBUG oslo_vmware.api [None req-60511481-93fa-4f8b-817d-01a577f85a26 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': task-5115660, 'name': PowerOnVM_Task} progress is 88%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 701.505196] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-719800f9-8111-44cd-8b0c-a7853ee69f3f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.513781] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71d38d34-140b-4efb-8a55-439979317b76 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.555194] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e700824-46c7-46a3-ad2e-e4a4fddb002e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.563569] env[65726]: DEBUG oslo_vmware.api [None req-f3878371-4599-418c-a420-84e54e420a24 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Task: {'id': task-5115662, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 701.569570] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ac7b8af-02ce-4846-aa83-7d9976009b16 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.584748] env[65726]: DEBUG nova.compute.provider_tree [None req-5dd7a34f-2002-4a12-9099-b0b630d2774d tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Updating inventory in ProviderTree for provider 07c4692f-bdb4-4058-9173-ff9664830295 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 701.627931] env[65726]: DEBUG oslo_concurrency.lockutils [None req-f61c0733-7339-4ad4-9f78-b4a5ed0d0ff1 tempest-AttachInterfacesV270Test-133387050 tempest-AttachInterfacesV270Test-133387050-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 701.722019] env[65726]: INFO nova.compute.manager [-] [instance: 3b933684-ac19-44b0-a49d-6af45501e166] Took 1.07 seconds to deallocate network for instance. [ 701.819163] env[65726]: DEBUG oslo_vmware.api [None req-60511481-93fa-4f8b-817d-01a577f85a26 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': task-5115660, 'name': PowerOnVM_Task} progress is 88%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 702.033973] env[65726]: DEBUG oslo_vmware.api [None req-f3878371-4599-418c-a420-84e54e420a24 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Task: {'id': task-5115662, 'name': ReconfigVM_Task, 'duration_secs': 0.67024} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 702.035014] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-f3878371-4599-418c-a420-84e54e420a24 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] [instance: 423af936-a383-4652-8887-9c8b8f507909] Reconfigured VM instance instance-0000001d to attach disk [datastore2] 423af936-a383-4652-8887-9c8b8f507909/423af936-a383-4652-8887-9c8b8f507909.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 702.035822] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7e43b733-642b-4789-9d7b-061b1929beb4 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.043727] env[65726]: DEBUG oslo_vmware.api [None req-f3878371-4599-418c-a420-84e54e420a24 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Waiting for the task: (returnval){ [ 702.043727] env[65726]: value = "task-5115663" [ 702.043727] env[65726]: _type = "Task" [ 702.043727] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 702.067029] env[65726]: DEBUG oslo_vmware.api [None req-f3878371-4599-418c-a420-84e54e420a24 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Task: {'id': task-5115663, 'name': Rename_Task} progress is 6%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 702.111132] env[65726]: ERROR nova.scheduler.client.report [None req-5dd7a34f-2002-4a12-9099-b0b630d2774d tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] [req-9739dbca-24ec-42f6-a61f-724a20a4bd42] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 07c4692f-bdb4-4058-9173-ff9664830295. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-9739dbca-24ec-42f6-a61f-724a20a4bd42"}]} [ 702.133154] env[65726]: DEBUG nova.scheduler.client.report [None req-5dd7a34f-2002-4a12-9099-b0b630d2774d tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Refreshing inventories for resource provider 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 702.149560] env[65726]: DEBUG nova.scheduler.client.report [None req-5dd7a34f-2002-4a12-9099-b0b630d2774d tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Updating ProviderTree inventory for provider 07c4692f-bdb4-4058-9173-ff9664830295 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 702.149768] env[65726]: DEBUG nova.compute.provider_tree [None req-5dd7a34f-2002-4a12-9099-b0b630d2774d tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Updating inventory in ProviderTree for provider 07c4692f-bdb4-4058-9173-ff9664830295 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 702.166017] env[65726]: DEBUG nova.scheduler.client.report [None req-5dd7a34f-2002-4a12-9099-b0b630d2774d tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Refreshing aggregate associations for resource provider 07c4692f-bdb4-4058-9173-ff9664830295, aggregates: None {{(pid=65726) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 702.185301] env[65726]: DEBUG nova.scheduler.client.report [None req-5dd7a34f-2002-4a12-9099-b0b630d2774d tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Refreshing trait associations for resource provider 07c4692f-bdb4-4058-9173-ff9664830295, traits: COMPUTE_IMAGE_TYPE_ISO,HW_ARCH_X86_64,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=65726) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 702.234819] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c19da911-fb8f-4bec-a53e-39a31fa2d36e tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 702.320573] env[65726]: DEBUG oslo_vmware.api [None req-60511481-93fa-4f8b-817d-01a577f85a26 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': task-5115660, 'name': PowerOnVM_Task, 'duration_secs': 1.741283} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 702.320736] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-60511481-93fa-4f8b-817d-01a577f85a26 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: 49bcab1f-7b64-4999-abff-37771c58a271] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 702.320940] env[65726]: INFO nova.compute.manager [None req-60511481-93fa-4f8b-817d-01a577f85a26 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: 49bcab1f-7b64-4999-abff-37771c58a271] Took 10.72 seconds to spawn the instance on the hypervisor. [ 702.321146] env[65726]: DEBUG nova.compute.manager [None req-60511481-93fa-4f8b-817d-01a577f85a26 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: 49bcab1f-7b64-4999-abff-37771c58a271] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 702.324817] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4a3ab78-10bc-4648-82ed-75d3074b8108 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.559385] env[65726]: DEBUG oslo_vmware.api [None req-f3878371-4599-418c-a420-84e54e420a24 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Task: {'id': task-5115663, 'name': Rename_Task, 'duration_secs': 0.270864} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 702.559385] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-f3878371-4599-418c-a420-84e54e420a24 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] [instance: 423af936-a383-4652-8887-9c8b8f507909] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 702.559385] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-48652849-1b86-406d-a63d-16ad81b19a1b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.565364] env[65726]: DEBUG oslo_concurrency.lockutils [None req-462b38ac-85e6-4728-81f8-4b729f9b1c26 tempest-ServersAaction247Test-712120018 tempest-ServersAaction247Test-712120018-project-member] Acquiring lock "ad48cadd-9bb6-4191-a68a-5c8fdaeebd44" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 702.565642] env[65726]: DEBUG oslo_concurrency.lockutils [None req-462b38ac-85e6-4728-81f8-4b729f9b1c26 tempest-ServersAaction247Test-712120018 tempest-ServersAaction247Test-712120018-project-member] Lock "ad48cadd-9bb6-4191-a68a-5c8fdaeebd44" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 702.574628] env[65726]: DEBUG oslo_vmware.api [None req-f3878371-4599-418c-a420-84e54e420a24 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Waiting for the task: (returnval){ [ 702.574628] env[65726]: value = "task-5115664" [ 702.574628] env[65726]: _type = "Task" [ 702.574628] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 702.585214] env[65726]: DEBUG oslo_vmware.api [None req-f3878371-4599-418c-a420-84e54e420a24 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Task: {'id': task-5115664, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 702.849515] env[65726]: INFO nova.compute.manager [None req-60511481-93fa-4f8b-817d-01a577f85a26 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: 49bcab1f-7b64-4999-abff-37771c58a271] Took 39.88 seconds to build instance. [ 702.856233] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5e5bfa8e-de8f-41bf-8724-f77b78155db2 tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] Acquiring lock "ddd36db0-28ad-4582-994b-7eef8b08ed46" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 702.856765] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5e5bfa8e-de8f-41bf-8724-f77b78155db2 tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] Lock "ddd36db0-28ad-4582-994b-7eef8b08ed46" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 702.860773] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5e5bfa8e-de8f-41bf-8724-f77b78155db2 tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] Acquiring lock "ddd36db0-28ad-4582-994b-7eef8b08ed46-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 702.860773] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5e5bfa8e-de8f-41bf-8724-f77b78155db2 tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] Lock "ddd36db0-28ad-4582-994b-7eef8b08ed46-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 702.860773] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5e5bfa8e-de8f-41bf-8724-f77b78155db2 tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] Lock "ddd36db0-28ad-4582-994b-7eef8b08ed46-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 702.860773] env[65726]: INFO nova.compute.manager [None req-5e5bfa8e-de8f-41bf-8724-f77b78155db2 tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] [instance: ddd36db0-28ad-4582-994b-7eef8b08ed46] Terminating instance [ 702.886915] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea4c928b-a24c-4b62-bc96-33d8f52438a2 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.896065] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a691887a-0bf8-46f0-8ce4-8cafe16e4caf {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.931060] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-edc91a34-c4da-437f-afa7-201ea0975f6a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.942977] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0f21154-0147-49cf-8d17-03908ed96057 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.960960] env[65726]: DEBUG nova.compute.provider_tree [None req-5dd7a34f-2002-4a12-9099-b0b630d2774d tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Updating inventory in ProviderTree for provider 07c4692f-bdb4-4058-9173-ff9664830295 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 703.088456] env[65726]: DEBUG oslo_vmware.api [None req-f3878371-4599-418c-a420-84e54e420a24 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Task: {'id': task-5115664, 'name': PowerOnVM_Task} progress is 87%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 703.352642] env[65726]: DEBUG oslo_concurrency.lockutils [None req-60511481-93fa-4f8b-817d-01a577f85a26 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Lock "49bcab1f-7b64-4999-abff-37771c58a271" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 60.455s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 703.371920] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5e5bfa8e-de8f-41bf-8724-f77b78155db2 tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] Acquiring lock "refresh_cache-ddd36db0-28ad-4582-994b-7eef8b08ed46" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 703.372167] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5e5bfa8e-de8f-41bf-8724-f77b78155db2 tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] Acquired lock "refresh_cache-ddd36db0-28ad-4582-994b-7eef8b08ed46" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 703.372354] env[65726]: DEBUG nova.network.neutron [None req-5e5bfa8e-de8f-41bf-8724-f77b78155db2 tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] [instance: ddd36db0-28ad-4582-994b-7eef8b08ed46] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 703.500332] env[65726]: DEBUG nova.scheduler.client.report [None req-5dd7a34f-2002-4a12-9099-b0b630d2774d tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Updated inventory for provider 07c4692f-bdb4-4058-9173-ff9664830295 with generation 55 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 703.500539] env[65726]: DEBUG nova.compute.provider_tree [None req-5dd7a34f-2002-4a12-9099-b0b630d2774d tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Updating resource provider 07c4692f-bdb4-4058-9173-ff9664830295 generation from 55 to 56 during operation: update_inventory {{(pid=65726) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 703.500604] env[65726]: DEBUG nova.compute.provider_tree [None req-5dd7a34f-2002-4a12-9099-b0b630d2774d tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Updating inventory in ProviderTree for provider 07c4692f-bdb4-4058-9173-ff9664830295 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 703.585467] env[65726]: DEBUG oslo_vmware.api [None req-f3878371-4599-418c-a420-84e54e420a24 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Task: {'id': task-5115664, 'name': PowerOnVM_Task, 'duration_secs': 0.6997} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 703.585857] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-f3878371-4599-418c-a420-84e54e420a24 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] [instance: 423af936-a383-4652-8887-9c8b8f507909] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 703.586090] env[65726]: INFO nova.compute.manager [None req-f3878371-4599-418c-a420-84e54e420a24 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] [instance: 423af936-a383-4652-8887-9c8b8f507909] Took 10.14 seconds to spawn the instance on the hypervisor. [ 703.586280] env[65726]: DEBUG nova.compute.manager [None req-f3878371-4599-418c-a420-84e54e420a24 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] [instance: 423af936-a383-4652-8887-9c8b8f507909] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 703.587139] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6c0ced2-5aca-40da-be8a-12fc5f51dc4d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.856559] env[65726]: DEBUG nova.compute.manager [None req-b479f246-d94b-4634-888f-dee4586c3272 tempest-ServerAddressesNegativeTestJSON-963263679 tempest-ServerAddressesNegativeTestJSON-963263679-project-member] [instance: 00746508-d0d6-4dfa-9026-772398d004ee] Starting instance... {{(pid=65726) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 703.875912] env[65726]: WARNING openstack [None req-5e5bfa8e-de8f-41bf-8724-f77b78155db2 tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 703.876331] env[65726]: WARNING openstack [None req-5e5bfa8e-de8f-41bf-8724-f77b78155db2 tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 703.890850] env[65726]: DEBUG nova.network.neutron [None req-5e5bfa8e-de8f-41bf-8724-f77b78155db2 tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] [instance: ddd36db0-28ad-4582-994b-7eef8b08ed46] Instance cache missing network info. {{(pid=65726) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 703.998929] env[65726]: DEBUG nova.network.neutron [None req-5e5bfa8e-de8f-41bf-8724-f77b78155db2 tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] [instance: ddd36db0-28ad-4582-994b-7eef8b08ed46] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 704.005879] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5dd7a34f-2002-4a12-9099-b0b630d2774d tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 7.028s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 704.006059] env[65726]: DEBUG nova.compute.manager [None req-5dd7a34f-2002-4a12-9099-b0b630d2774d tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] [instance: 815f056b-2d88-40dd-9ef7-a891ec8d92cf] Start building networks asynchronously for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 704.012023] env[65726]: DEBUG oslo_concurrency.lockutils [None req-f3171f49-8536-4923-b5cf-31399a616a07 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 29.360s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 704.012023] env[65726]: INFO nova.compute.claims [None req-f3171f49-8536-4923-b5cf-31399a616a07 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 608cc0ea-3f6b-4b4f-83c1-01aa50999d63] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 704.110812] env[65726]: INFO nova.compute.manager [None req-f3878371-4599-418c-a420-84e54e420a24 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] [instance: 423af936-a383-4652-8887-9c8b8f507909] Took 38.09 seconds to build instance. [ 704.393450] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b479f246-d94b-4634-888f-dee4586c3272 tempest-ServerAddressesNegativeTestJSON-963263679 tempest-ServerAddressesNegativeTestJSON-963263679-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 704.504360] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5e5bfa8e-de8f-41bf-8724-f77b78155db2 tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] Releasing lock "refresh_cache-ddd36db0-28ad-4582-994b-7eef8b08ed46" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 704.504660] env[65726]: DEBUG nova.compute.manager [None req-5e5bfa8e-de8f-41bf-8724-f77b78155db2 tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] [instance: ddd36db0-28ad-4582-994b-7eef8b08ed46] Start destroying the instance on the hypervisor. {{(pid=65726) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 704.504901] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-5e5bfa8e-de8f-41bf-8724-f77b78155db2 tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] [instance: ddd36db0-28ad-4582-994b-7eef8b08ed46] Destroying instance {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 704.505839] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-996c3496-d19d-4cc1-b248-ee7459d0bff9 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.518148] env[65726]: DEBUG nova.compute.utils [None req-5dd7a34f-2002-4a12-9099-b0b630d2774d tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Using /dev/sd instead of None {{(pid=65726) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 704.525412] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-5e5bfa8e-de8f-41bf-8724-f77b78155db2 tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] [instance: ddd36db0-28ad-4582-994b-7eef8b08ed46] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 704.525530] env[65726]: DEBUG nova.compute.manager [None req-5dd7a34f-2002-4a12-9099-b0b630d2774d tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] [instance: 815f056b-2d88-40dd-9ef7-a891ec8d92cf] Allocating IP information in the background. {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 704.525740] env[65726]: DEBUG nova.network.neutron [None req-5dd7a34f-2002-4a12-9099-b0b630d2774d tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] [instance: 815f056b-2d88-40dd-9ef7-a891ec8d92cf] allocate_for_instance() {{(pid=65726) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 704.526278] env[65726]: WARNING neutronclient.v2_0.client [None req-5dd7a34f-2002-4a12-9099-b0b630d2774d tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 704.526554] env[65726]: WARNING neutronclient.v2_0.client [None req-5dd7a34f-2002-4a12-9099-b0b630d2774d tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 704.527270] env[65726]: WARNING openstack [None req-5dd7a34f-2002-4a12-9099-b0b630d2774d tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 704.527635] env[65726]: WARNING openstack [None req-5dd7a34f-2002-4a12-9099-b0b630d2774d tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 704.535399] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-779df0a5-6e0b-45ad-8e7d-237b49e5240e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.545683] env[65726]: DEBUG oslo_vmware.api [None req-5e5bfa8e-de8f-41bf-8724-f77b78155db2 tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] Waiting for the task: (returnval){ [ 704.545683] env[65726]: value = "task-5115665" [ 704.545683] env[65726]: _type = "Task" [ 704.545683] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 704.555103] env[65726]: DEBUG oslo_vmware.api [None req-5e5bfa8e-de8f-41bf-8724-f77b78155db2 tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] Task: {'id': task-5115665, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 704.617512] env[65726]: DEBUG nova.policy [None req-5dd7a34f-2002-4a12-9099-b0b630d2774d tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6ce8616097d940ffa52ef44c14b84622', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '06ad412f524c40538c835d9d664cd85e', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65726) authorize /opt/stack/nova/nova/policy.py:192}} [ 704.618857] env[65726]: DEBUG oslo_concurrency.lockutils [None req-f3878371-4599-418c-a420-84e54e420a24 tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Lock "423af936-a383-4652-8887-9c8b8f507909" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 60.292s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 704.904034] env[65726]: DEBUG nova.network.neutron [None req-5dd7a34f-2002-4a12-9099-b0b630d2774d tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] [instance: 815f056b-2d88-40dd-9ef7-a891ec8d92cf] Successfully created port: 1dfb06a2-5081-48fc-a9ef-23b637910832 {{(pid=65726) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 705.021088] env[65726]: DEBUG nova.compute.manager [None req-5dd7a34f-2002-4a12-9099-b0b630d2774d tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] [instance: 815f056b-2d88-40dd-9ef7-a891ec8d92cf] Start building block device mappings for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 705.056587] env[65726]: DEBUG oslo_vmware.api [None req-5e5bfa8e-de8f-41bf-8724-f77b78155db2 tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] Task: {'id': task-5115665, 'name': PowerOffVM_Task, 'duration_secs': 0.205405} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 705.056967] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-5e5bfa8e-de8f-41bf-8724-f77b78155db2 tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] [instance: ddd36db0-28ad-4582-994b-7eef8b08ed46] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 705.057061] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-5e5bfa8e-de8f-41bf-8724-f77b78155db2 tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] [instance: ddd36db0-28ad-4582-994b-7eef8b08ed46] Unregistering the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 705.057803] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b0118b8a-d37f-4d67-8a69-8ab70edf0675 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.087673] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-5e5bfa8e-de8f-41bf-8724-f77b78155db2 tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] [instance: ddd36db0-28ad-4582-994b-7eef8b08ed46] Unregistered the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 705.087673] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-5e5bfa8e-de8f-41bf-8724-f77b78155db2 tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] [instance: ddd36db0-28ad-4582-994b-7eef8b08ed46] Deleting contents of the VM from datastore datastore2 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 705.087673] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-5e5bfa8e-de8f-41bf-8724-f77b78155db2 tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] Deleting the datastore file [datastore2] ddd36db0-28ad-4582-994b-7eef8b08ed46 {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 705.087673] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-037d2b45-b5ab-4083-98d2-71b7f2cbf163 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.105239] env[65726]: DEBUG oslo_vmware.api [None req-5e5bfa8e-de8f-41bf-8724-f77b78155db2 tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] Waiting for the task: (returnval){ [ 705.105239] env[65726]: value = "task-5115667" [ 705.105239] env[65726]: _type = "Task" [ 705.105239] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 705.122829] env[65726]: DEBUG oslo_concurrency.lockutils [None req-8b18678f-63d3-424e-98c8-800d5081389c tempest-VolumesAdminNegativeTest-2004477850 tempest-VolumesAdminNegativeTest-2004477850-project-member] Acquiring lock "30b17a17-229c-4ca7-9ae6-c67ccdcd3963" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 705.124785] env[65726]: DEBUG oslo_concurrency.lockutils [None req-8b18678f-63d3-424e-98c8-800d5081389c tempest-VolumesAdminNegativeTest-2004477850 tempest-VolumesAdminNegativeTest-2004477850-project-member] Lock "30b17a17-229c-4ca7-9ae6-c67ccdcd3963" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 705.126048] env[65726]: DEBUG oslo_concurrency.lockutils [None req-8b18678f-63d3-424e-98c8-800d5081389c tempest-VolumesAdminNegativeTest-2004477850 tempest-VolumesAdminNegativeTest-2004477850-project-member] Acquiring lock "30b17a17-229c-4ca7-9ae6-c67ccdcd3963-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 705.126048] env[65726]: DEBUG oslo_concurrency.lockutils [None req-8b18678f-63d3-424e-98c8-800d5081389c tempest-VolumesAdminNegativeTest-2004477850 tempest-VolumesAdminNegativeTest-2004477850-project-member] Lock "30b17a17-229c-4ca7-9ae6-c67ccdcd3963-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 705.126048] env[65726]: DEBUG oslo_concurrency.lockutils [None req-8b18678f-63d3-424e-98c8-800d5081389c tempest-VolumesAdminNegativeTest-2004477850 tempest-VolumesAdminNegativeTest-2004477850-project-member] Lock "30b17a17-229c-4ca7-9ae6-c67ccdcd3963-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 705.128239] env[65726]: DEBUG nova.compute.manager [None req-04afb47c-692a-46d9-a640-68cc2c90206f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: c108f2a5-031b-47ee-9a5b-d62c2f42c26b] Starting instance... {{(pid=65726) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 705.131845] env[65726]: DEBUG oslo_vmware.api [None req-5e5bfa8e-de8f-41bf-8724-f77b78155db2 tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] Task: {'id': task-5115667, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 705.133410] env[65726]: INFO nova.compute.manager [None req-8b18678f-63d3-424e-98c8-800d5081389c tempest-VolumesAdminNegativeTest-2004477850 tempest-VolumesAdminNegativeTest-2004477850-project-member] [instance: 30b17a17-229c-4ca7-9ae6-c67ccdcd3963] Terminating instance [ 705.622480] env[65726]: DEBUG oslo_vmware.api [None req-5e5bfa8e-de8f-41bf-8724-f77b78155db2 tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] Task: {'id': task-5115667, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.12355} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 705.622802] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-5e5bfa8e-de8f-41bf-8724-f77b78155db2 tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] Deleted the datastore file {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 705.623074] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-5e5bfa8e-de8f-41bf-8724-f77b78155db2 tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] [instance: ddd36db0-28ad-4582-994b-7eef8b08ed46] Deleted contents of the VM from datastore datastore2 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 705.623334] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-5e5bfa8e-de8f-41bf-8724-f77b78155db2 tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] [instance: ddd36db0-28ad-4582-994b-7eef8b08ed46] Instance destroyed {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 705.623656] env[65726]: INFO nova.compute.manager [None req-5e5bfa8e-de8f-41bf-8724-f77b78155db2 tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] [instance: ddd36db0-28ad-4582-994b-7eef8b08ed46] Took 1.12 seconds to destroy the instance on the hypervisor. [ 705.624059] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-5e5bfa8e-de8f-41bf-8724-f77b78155db2 tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 705.624503] env[65726]: DEBUG nova.compute.manager [-] [instance: ddd36db0-28ad-4582-994b-7eef8b08ed46] Deallocating network for instance {{(pid=65726) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 705.624503] env[65726]: DEBUG nova.network.neutron [-] [instance: ddd36db0-28ad-4582-994b-7eef8b08ed46] deallocate_for_instance() {{(pid=65726) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 705.625223] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 705.625599] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 705.642508] env[65726]: DEBUG nova.compute.manager [None req-8b18678f-63d3-424e-98c8-800d5081389c tempest-VolumesAdminNegativeTest-2004477850 tempest-VolumesAdminNegativeTest-2004477850-project-member] [instance: 30b17a17-229c-4ca7-9ae6-c67ccdcd3963] Start destroying the instance on the hypervisor. {{(pid=65726) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 705.642792] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-8b18678f-63d3-424e-98c8-800d5081389c tempest-VolumesAdminNegativeTest-2004477850 tempest-VolumesAdminNegativeTest-2004477850-project-member] [instance: 30b17a17-229c-4ca7-9ae6-c67ccdcd3963] Destroying instance {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 705.646770] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4be4a804-7040-4076-9896-70c1a418d965 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.658388] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-8b18678f-63d3-424e-98c8-800d5081389c tempest-VolumesAdminNegativeTest-2004477850 tempest-VolumesAdminNegativeTest-2004477850-project-member] [instance: 30b17a17-229c-4ca7-9ae6-c67ccdcd3963] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 705.662759] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-64d50ecc-da2e-4dd9-bd25-158040265def {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.666012] env[65726]: DEBUG oslo_concurrency.lockutils [None req-04afb47c-692a-46d9-a640-68cc2c90206f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 705.672740] env[65726]: DEBUG oslo_vmware.api [None req-8b18678f-63d3-424e-98c8-800d5081389c tempest-VolumesAdminNegativeTest-2004477850 tempest-VolumesAdminNegativeTest-2004477850-project-member] Waiting for the task: (returnval){ [ 705.672740] env[65726]: value = "task-5115668" [ 705.672740] env[65726]: _type = "Task" [ 705.672740] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 705.677559] env[65726]: DEBUG nova.network.neutron [-] [instance: ddd36db0-28ad-4582-994b-7eef8b08ed46] Instance cache missing network info. {{(pid=65726) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 705.677559] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 705.677559] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 705.693907] env[65726]: DEBUG oslo_vmware.api [None req-8b18678f-63d3-424e-98c8-800d5081389c tempest-VolumesAdminNegativeTest-2004477850 tempest-VolumesAdminNegativeTest-2004477850-project-member] Task: {'id': task-5115668, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 705.700770] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39db54ce-0b2c-4810-a689-be5db8358379 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.708501] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59f355f6-e20d-47c5-9c1a-f5541e78e862 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.741430] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b801032e-e14f-4544-8c1b-21caebfc964b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.756056] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4cb82ff-58f5-476c-8085-2af4ecd4cc84 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.775536] env[65726]: DEBUG nova.compute.provider_tree [None req-f3171f49-8536-4923-b5cf-31399a616a07 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 705.779188] env[65726]: DEBUG nova.compute.manager [req-f7e2c231-c58d-4adb-8685-2fedd2d29013 req-e2b76d0e-7323-45bb-9d29-caad6e9450f2 service nova] [instance: 49bcab1f-7b64-4999-abff-37771c58a271] Received event network-changed-45612b7e-c052-4346-a779-5f94f9adc084 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 705.779437] env[65726]: DEBUG nova.compute.manager [req-f7e2c231-c58d-4adb-8685-2fedd2d29013 req-e2b76d0e-7323-45bb-9d29-caad6e9450f2 service nova] [instance: 49bcab1f-7b64-4999-abff-37771c58a271] Refreshing instance network info cache due to event network-changed-45612b7e-c052-4346-a779-5f94f9adc084. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 705.779655] env[65726]: DEBUG oslo_concurrency.lockutils [req-f7e2c231-c58d-4adb-8685-2fedd2d29013 req-e2b76d0e-7323-45bb-9d29-caad6e9450f2 service nova] Acquiring lock "refresh_cache-49bcab1f-7b64-4999-abff-37771c58a271" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 705.779807] env[65726]: DEBUG oslo_concurrency.lockutils [req-f7e2c231-c58d-4adb-8685-2fedd2d29013 req-e2b76d0e-7323-45bb-9d29-caad6e9450f2 service nova] Acquired lock "refresh_cache-49bcab1f-7b64-4999-abff-37771c58a271" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 705.779964] env[65726]: DEBUG nova.network.neutron [req-f7e2c231-c58d-4adb-8685-2fedd2d29013 req-e2b76d0e-7323-45bb-9d29-caad6e9450f2 service nova] [instance: 49bcab1f-7b64-4999-abff-37771c58a271] Refreshing network info cache for port 45612b7e-c052-4346-a779-5f94f9adc084 {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 705.842516] env[65726]: DEBUG oslo_concurrency.lockutils [None req-300cacab-e765-49c9-a8a1-8e47a20dc6a2 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Acquiring lock "76249623-6f83-46a3-b8c5-c001111aa698" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 705.842516] env[65726]: DEBUG oslo_concurrency.lockutils [None req-300cacab-e765-49c9-a8a1-8e47a20dc6a2 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Lock "76249623-6f83-46a3-b8c5-c001111aa698" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 706.039516] env[65726]: DEBUG nova.compute.manager [None req-5dd7a34f-2002-4a12-9099-b0b630d2774d tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] [instance: 815f056b-2d88-40dd-9ef7-a891ec8d92cf] Start spawning the instance on the hypervisor. {{(pid=65726) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 706.067963] env[65726]: DEBUG nova.virt.hardware [None req-5dd7a34f-2002-4a12-9099-b0b630d2774d tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T19:28:34Z,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1539f06fa3534e90acbc3a60c4b8bd3f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T19:28:35Z,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 706.068262] env[65726]: DEBUG nova.virt.hardware [None req-5dd7a34f-2002-4a12-9099-b0b630d2774d tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 706.068712] env[65726]: DEBUG nova.virt.hardware [None req-5dd7a34f-2002-4a12-9099-b0b630d2774d tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 706.068901] env[65726]: DEBUG nova.virt.hardware [None req-5dd7a34f-2002-4a12-9099-b0b630d2774d tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 706.069050] env[65726]: DEBUG nova.virt.hardware [None req-5dd7a34f-2002-4a12-9099-b0b630d2774d tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 706.069257] env[65726]: DEBUG nova.virt.hardware [None req-5dd7a34f-2002-4a12-9099-b0b630d2774d tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 706.069898] env[65726]: DEBUG nova.virt.hardware [None req-5dd7a34f-2002-4a12-9099-b0b630d2774d tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 706.070196] env[65726]: DEBUG nova.virt.hardware [None req-5dd7a34f-2002-4a12-9099-b0b630d2774d tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 706.070497] env[65726]: DEBUG nova.virt.hardware [None req-5dd7a34f-2002-4a12-9099-b0b630d2774d tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 706.070976] env[65726]: DEBUG nova.virt.hardware [None req-5dd7a34f-2002-4a12-9099-b0b630d2774d tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 706.071398] env[65726]: DEBUG nova.virt.hardware [None req-5dd7a34f-2002-4a12-9099-b0b630d2774d tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 706.072863] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-817fb2e8-2628-4d80-91eb-1a85c9407a8b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.083357] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b12f63b3-8ae3-4d21-afd0-77575df0aa3b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.183361] env[65726]: DEBUG oslo_vmware.api [None req-8b18678f-63d3-424e-98c8-800d5081389c tempest-VolumesAdminNegativeTest-2004477850 tempest-VolumesAdminNegativeTest-2004477850-project-member] Task: {'id': task-5115668, 'name': PowerOffVM_Task, 'duration_secs': 0.332995} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 706.183703] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-8b18678f-63d3-424e-98c8-800d5081389c tempest-VolumesAdminNegativeTest-2004477850 tempest-VolumesAdminNegativeTest-2004477850-project-member] [instance: 30b17a17-229c-4ca7-9ae6-c67ccdcd3963] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 706.183891] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-8b18678f-63d3-424e-98c8-800d5081389c tempest-VolumesAdminNegativeTest-2004477850 tempest-VolumesAdminNegativeTest-2004477850-project-member] [instance: 30b17a17-229c-4ca7-9ae6-c67ccdcd3963] Unregistering the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 706.184165] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-93a97059-caee-4117-8f6c-a7c8bf6820a1 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.187374] env[65726]: DEBUG nova.network.neutron [-] [instance: ddd36db0-28ad-4582-994b-7eef8b08ed46] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 706.257656] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-8b18678f-63d3-424e-98c8-800d5081389c tempest-VolumesAdminNegativeTest-2004477850 tempest-VolumesAdminNegativeTest-2004477850-project-member] [instance: 30b17a17-229c-4ca7-9ae6-c67ccdcd3963] Unregistered the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 706.257815] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-8b18678f-63d3-424e-98c8-800d5081389c tempest-VolumesAdminNegativeTest-2004477850 tempest-VolumesAdminNegativeTest-2004477850-project-member] [instance: 30b17a17-229c-4ca7-9ae6-c67ccdcd3963] Deleting contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 706.257986] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-8b18678f-63d3-424e-98c8-800d5081389c tempest-VolumesAdminNegativeTest-2004477850 tempest-VolumesAdminNegativeTest-2004477850-project-member] Deleting the datastore file [datastore1] 30b17a17-229c-4ca7-9ae6-c67ccdcd3963 {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 706.258296] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5c8aab3a-22ec-447e-8435-398cbd4ab19a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.266594] env[65726]: DEBUG oslo_vmware.api [None req-8b18678f-63d3-424e-98c8-800d5081389c tempest-VolumesAdminNegativeTest-2004477850 tempest-VolumesAdminNegativeTest-2004477850-project-member] Waiting for the task: (returnval){ [ 706.266594] env[65726]: value = "task-5115670" [ 706.266594] env[65726]: _type = "Task" [ 706.266594] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 706.275706] env[65726]: DEBUG oslo_vmware.api [None req-8b18678f-63d3-424e-98c8-800d5081389c tempest-VolumesAdminNegativeTest-2004477850 tempest-VolumesAdminNegativeTest-2004477850-project-member] Task: {'id': task-5115670, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 706.282783] env[65726]: DEBUG nova.scheduler.client.report [None req-f3171f49-8536-4923-b5cf-31399a616a07 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 706.287010] env[65726]: WARNING openstack [req-f7e2c231-c58d-4adb-8685-2fedd2d29013 req-e2b76d0e-7323-45bb-9d29-caad6e9450f2 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 706.288051] env[65726]: WARNING openstack [req-f7e2c231-c58d-4adb-8685-2fedd2d29013 req-e2b76d0e-7323-45bb-9d29-caad6e9450f2 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 706.536771] env[65726]: DEBUG nova.network.neutron [None req-5dd7a34f-2002-4a12-9099-b0b630d2774d tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] [instance: 815f056b-2d88-40dd-9ef7-a891ec8d92cf] Successfully updated port: 1dfb06a2-5081-48fc-a9ef-23b637910832 {{(pid=65726) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 706.622775] env[65726]: WARNING openstack [req-f7e2c231-c58d-4adb-8685-2fedd2d29013 req-e2b76d0e-7323-45bb-9d29-caad6e9450f2 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 706.623156] env[65726]: WARNING openstack [req-f7e2c231-c58d-4adb-8685-2fedd2d29013 req-e2b76d0e-7323-45bb-9d29-caad6e9450f2 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 706.692187] env[65726]: WARNING openstack [req-f7e2c231-c58d-4adb-8685-2fedd2d29013 req-e2b76d0e-7323-45bb-9d29-caad6e9450f2 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 706.692591] env[65726]: WARNING openstack [req-f7e2c231-c58d-4adb-8685-2fedd2d29013 req-e2b76d0e-7323-45bb-9d29-caad6e9450f2 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 706.700388] env[65726]: INFO nova.compute.manager [-] [instance: ddd36db0-28ad-4582-994b-7eef8b08ed46] Took 1.08 seconds to deallocate network for instance. [ 706.777934] env[65726]: DEBUG oslo_vmware.api [None req-8b18678f-63d3-424e-98c8-800d5081389c tempest-VolumesAdminNegativeTest-2004477850 tempest-VolumesAdminNegativeTest-2004477850-project-member] Task: {'id': task-5115670, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 706.785124] env[65726]: DEBUG nova.network.neutron [req-f7e2c231-c58d-4adb-8685-2fedd2d29013 req-e2b76d0e-7323-45bb-9d29-caad6e9450f2 service nova] [instance: 49bcab1f-7b64-4999-abff-37771c58a271] Updated VIF entry in instance network info cache for port 45612b7e-c052-4346-a779-5f94f9adc084. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 706.785124] env[65726]: DEBUG nova.network.neutron [req-f7e2c231-c58d-4adb-8685-2fedd2d29013 req-e2b76d0e-7323-45bb-9d29-caad6e9450f2 service nova] [instance: 49bcab1f-7b64-4999-abff-37771c58a271] Updating instance_info_cache with network_info: [{"id": "45612b7e-c052-4346-a779-5f94f9adc084", "address": "fa:16:3e:9d:95:60", "network": {"id": "51a66448-0f72-4266-8868-856b48be839f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-155658961-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.161", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d46ccb00794f458b85da4a93879139ee", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b1f3e6c3-5584-4852-9017-476ab8ac4946", "external-id": "nsx-vlan-transportzone-304", "segmentation_id": 304, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap45612b7e-c0", "ovs_interfaceid": "45612b7e-c052-4346-a779-5f94f9adc084", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 706.795460] env[65726]: DEBUG oslo_concurrency.lockutils [None req-f3171f49-8536-4923-b5cf-31399a616a07 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.787s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 706.796274] env[65726]: DEBUG nova.compute.manager [None req-f3171f49-8536-4923-b5cf-31399a616a07 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 608cc0ea-3f6b-4b4f-83c1-01aa50999d63] Start building networks asynchronously for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 706.802535] env[65726]: DEBUG oslo_concurrency.lockutils [None req-fd996099-26ef-437f-83c3-98cb140e9de7 tempest-TenantUsagesTestJSON-327211524 tempest-TenantUsagesTestJSON-327211524-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 30.688s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 706.802535] env[65726]: DEBUG nova.objects.instance [None req-fd996099-26ef-437f-83c3-98cb140e9de7 tempest-TenantUsagesTestJSON-327211524 tempest-TenantUsagesTestJSON-327211524-project-member] Lazy-loading 'resources' on Instance uuid 7748c23f-9ea9-4172-8ab7-187678272bb3 {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 707.045163] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5dd7a34f-2002-4a12-9099-b0b630d2774d tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Acquiring lock "refresh_cache-815f056b-2d88-40dd-9ef7-a891ec8d92cf" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 707.045476] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5dd7a34f-2002-4a12-9099-b0b630d2774d tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Acquired lock "refresh_cache-815f056b-2d88-40dd-9ef7-a891ec8d92cf" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 707.046194] env[65726]: DEBUG nova.network.neutron [None req-5dd7a34f-2002-4a12-9099-b0b630d2774d tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] [instance: 815f056b-2d88-40dd-9ef7-a891ec8d92cf] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 707.210282] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5e5bfa8e-de8f-41bf-8724-f77b78155db2 tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 707.280109] env[65726]: DEBUG oslo_vmware.api [None req-8b18678f-63d3-424e-98c8-800d5081389c tempest-VolumesAdminNegativeTest-2004477850 tempest-VolumesAdminNegativeTest-2004477850-project-member] Task: {'id': task-5115670, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.560376} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 707.280109] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-8b18678f-63d3-424e-98c8-800d5081389c tempest-VolumesAdminNegativeTest-2004477850 tempest-VolumesAdminNegativeTest-2004477850-project-member] Deleted the datastore file {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 707.280109] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-8b18678f-63d3-424e-98c8-800d5081389c tempest-VolumesAdminNegativeTest-2004477850 tempest-VolumesAdminNegativeTest-2004477850-project-member] [instance: 30b17a17-229c-4ca7-9ae6-c67ccdcd3963] Deleted contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 707.280336] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-8b18678f-63d3-424e-98c8-800d5081389c tempest-VolumesAdminNegativeTest-2004477850 tempest-VolumesAdminNegativeTest-2004477850-project-member] [instance: 30b17a17-229c-4ca7-9ae6-c67ccdcd3963] Instance destroyed {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 707.280599] env[65726]: INFO nova.compute.manager [None req-8b18678f-63d3-424e-98c8-800d5081389c tempest-VolumesAdminNegativeTest-2004477850 tempest-VolumesAdminNegativeTest-2004477850-project-member] [instance: 30b17a17-229c-4ca7-9ae6-c67ccdcd3963] Took 1.64 seconds to destroy the instance on the hypervisor. [ 707.280762] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-8b18678f-63d3-424e-98c8-800d5081389c tempest-VolumesAdminNegativeTest-2004477850 tempest-VolumesAdminNegativeTest-2004477850-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 707.280960] env[65726]: DEBUG nova.compute.manager [-] [instance: 30b17a17-229c-4ca7-9ae6-c67ccdcd3963] Deallocating network for instance {{(pid=65726) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 707.281073] env[65726]: DEBUG nova.network.neutron [-] [instance: 30b17a17-229c-4ca7-9ae6-c67ccdcd3963] deallocate_for_instance() {{(pid=65726) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 707.281586] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 707.281840] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 707.288570] env[65726]: DEBUG oslo_concurrency.lockutils [req-f7e2c231-c58d-4adb-8685-2fedd2d29013 req-e2b76d0e-7323-45bb-9d29-caad6e9450f2 service nova] Releasing lock "refresh_cache-49bcab1f-7b64-4999-abff-37771c58a271" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 707.305426] env[65726]: DEBUG nova.compute.utils [None req-f3171f49-8536-4923-b5cf-31399a616a07 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Using /dev/sd instead of None {{(pid=65726) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 707.310074] env[65726]: DEBUG nova.compute.manager [None req-f3171f49-8536-4923-b5cf-31399a616a07 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 608cc0ea-3f6b-4b4f-83c1-01aa50999d63] Allocating IP information in the background. {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 707.310277] env[65726]: DEBUG nova.network.neutron [None req-f3171f49-8536-4923-b5cf-31399a616a07 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 608cc0ea-3f6b-4b4f-83c1-01aa50999d63] allocate_for_instance() {{(pid=65726) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 707.310748] env[65726]: WARNING neutronclient.v2_0.client [None req-f3171f49-8536-4923-b5cf-31399a616a07 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 707.311345] env[65726]: WARNING neutronclient.v2_0.client [None req-f3171f49-8536-4923-b5cf-31399a616a07 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 707.312070] env[65726]: WARNING openstack [None req-f3171f49-8536-4923-b5cf-31399a616a07 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 707.312181] env[65726]: WARNING openstack [None req-f3171f49-8536-4923-b5cf-31399a616a07 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 707.367192] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 707.367671] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 707.402462] env[65726]: DEBUG nova.policy [None req-f3171f49-8536-4923-b5cf-31399a616a07 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b251fcbd04044fcda192bc3febac1a90', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4eece77569624f90bf64e5c51974173f', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65726) authorize /opt/stack/nova/nova/policy.py:192}} [ 707.551332] env[65726]: WARNING openstack [None req-5dd7a34f-2002-4a12-9099-b0b630d2774d tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 707.552027] env[65726]: WARNING openstack [None req-5dd7a34f-2002-4a12-9099-b0b630d2774d tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 707.556818] env[65726]: DEBUG nova.network.neutron [None req-5dd7a34f-2002-4a12-9099-b0b630d2774d tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] [instance: 815f056b-2d88-40dd-9ef7-a891ec8d92cf] Instance cache missing network info. {{(pid=65726) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 707.578840] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79b3021a-9c43-41b6-8199-065492a0a829 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.587834] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-85e035ac-36aa-4f5f-93c0-1a4ac0d2c453 tempest-ServersAdminNegativeTestJSON-113649088 tempest-ServersAdminNegativeTestJSON-113649088-project-admin] [instance: 423af936-a383-4652-8887-9c8b8f507909] Suspending the VM {{(pid=65726) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1162}} [ 707.588502] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-83550ab7-de20-4670-ac90-abb7f939a636 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.596788] env[65726]: DEBUG oslo_vmware.api [None req-85e035ac-36aa-4f5f-93c0-1a4ac0d2c453 tempest-ServersAdminNegativeTestJSON-113649088 tempest-ServersAdminNegativeTestJSON-113649088-project-admin] Waiting for the task: (returnval){ [ 707.596788] env[65726]: value = "task-5115671" [ 707.596788] env[65726]: _type = "Task" [ 707.596788] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 707.606327] env[65726]: DEBUG oslo_vmware.api [None req-85e035ac-36aa-4f5f-93c0-1a4ac0d2c453 tempest-ServersAdminNegativeTestJSON-113649088 tempest-ServersAdminNegativeTestJSON-113649088-project-admin] Task: {'id': task-5115671, 'name': SuspendVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 707.687259] env[65726]: WARNING openstack [None req-5dd7a34f-2002-4a12-9099-b0b630d2774d tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 707.687545] env[65726]: WARNING openstack [None req-5dd7a34f-2002-4a12-9099-b0b630d2774d tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 707.780910] env[65726]: WARNING openstack [None req-5dd7a34f-2002-4a12-9099-b0b630d2774d tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 707.781060] env[65726]: WARNING openstack [None req-5dd7a34f-2002-4a12-9099-b0b630d2774d tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 707.811258] env[65726]: DEBUG nova.compute.manager [None req-f3171f49-8536-4923-b5cf-31399a616a07 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 608cc0ea-3f6b-4b4f-83c1-01aa50999d63] Start building block device mappings for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 707.817215] env[65726]: DEBUG nova.network.neutron [None req-f3171f49-8536-4923-b5cf-31399a616a07 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 608cc0ea-3f6b-4b4f-83c1-01aa50999d63] Successfully created port: e9f7132a-8d5a-47fe-afc6-d45bf3687ea4 {{(pid=65726) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 707.883715] env[65726]: DEBUG nova.network.neutron [None req-5dd7a34f-2002-4a12-9099-b0b630d2774d tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] [instance: 815f056b-2d88-40dd-9ef7-a891ec8d92cf] Updating instance_info_cache with network_info: [{"id": "1dfb06a2-5081-48fc-a9ef-23b637910832", "address": "fa:16:3e:84:38:5a", "network": {"id": "55d629b8-a541-4720-9289-927bbc6e6f22", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.118", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "1539f06fa3534e90acbc3a60c4b8bd3f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "35342bcb-8b06-472e-b3c0-43fd3d6c4b30", "external-id": "nsx-vlan-transportzone-524", "segmentation_id": 524, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1dfb06a2-50", "ovs_interfaceid": "1dfb06a2-5081-48fc-a9ef-23b637910832", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 707.939496] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fb724c5-9dee-49c1-b7fd-73ad4bf88b4b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.948393] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5728a13-4148-4617-87e2-46900ec49b4f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.983715] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-adb4f706-f9e2-45fd-aa64-d76d512c15fc {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.992271] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5c068a1-7178-4c6f-8646-71d096adef1f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.010332] env[65726]: DEBUG nova.compute.provider_tree [None req-fd996099-26ef-437f-83c3-98cb140e9de7 tempest-TenantUsagesTestJSON-327211524 tempest-TenantUsagesTestJSON-327211524-project-member] Updating inventory in ProviderTree for provider 07c4692f-bdb4-4058-9173-ff9664830295 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 708.108761] env[65726]: DEBUG oslo_vmware.api [None req-85e035ac-36aa-4f5f-93c0-1a4ac0d2c453 tempest-ServersAdminNegativeTestJSON-113649088 tempest-ServersAdminNegativeTestJSON-113649088-project-admin] Task: {'id': task-5115671, 'name': SuspendVM_Task} progress is 58%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 708.131713] env[65726]: DEBUG nova.network.neutron [-] [instance: 30b17a17-229c-4ca7-9ae6-c67ccdcd3963] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 708.390892] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5dd7a34f-2002-4a12-9099-b0b630d2774d tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Releasing lock "refresh_cache-815f056b-2d88-40dd-9ef7-a891ec8d92cf" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 708.390892] env[65726]: DEBUG nova.compute.manager [None req-5dd7a34f-2002-4a12-9099-b0b630d2774d tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] [instance: 815f056b-2d88-40dd-9ef7-a891ec8d92cf] Instance network_info: |[{"id": "1dfb06a2-5081-48fc-a9ef-23b637910832", "address": "fa:16:3e:84:38:5a", "network": {"id": "55d629b8-a541-4720-9289-927bbc6e6f22", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.118", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "1539f06fa3534e90acbc3a60c4b8bd3f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "35342bcb-8b06-472e-b3c0-43fd3d6c4b30", "external-id": "nsx-vlan-transportzone-524", "segmentation_id": 524, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1dfb06a2-50", "ovs_interfaceid": "1dfb06a2-5081-48fc-a9ef-23b637910832", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 708.391124] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-5dd7a34f-2002-4a12-9099-b0b630d2774d tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] [instance: 815f056b-2d88-40dd-9ef7-a891ec8d92cf] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:84:38:5a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '35342bcb-8b06-472e-b3c0-43fd3d6c4b30', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '1dfb06a2-5081-48fc-a9ef-23b637910832', 'vif_model': 'vmxnet3'}] {{(pid=65726) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 708.399657] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-5dd7a34f-2002-4a12-9099-b0b630d2774d tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 708.399657] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 815f056b-2d88-40dd-9ef7-a891ec8d92cf] Creating VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 708.399776] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-014a2ce5-a877-4a14-8ebe-7e67796776ea {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.422170] env[65726]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 708.422170] env[65726]: value = "task-5115672" [ 708.422170] env[65726]: _type = "Task" [ 708.422170] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 708.431413] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115672, 'name': CreateVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 708.540172] env[65726]: ERROR nova.scheduler.client.report [None req-fd996099-26ef-437f-83c3-98cb140e9de7 tempest-TenantUsagesTestJSON-327211524 tempest-TenantUsagesTestJSON-327211524-project-member] [req-cc71076b-0aa8-402c-bd57-2a4b26ded976] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 07c4692f-bdb4-4058-9173-ff9664830295. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-cc71076b-0aa8-402c-bd57-2a4b26ded976"}]} [ 708.559122] env[65726]: DEBUG nova.scheduler.client.report [None req-fd996099-26ef-437f-83c3-98cb140e9de7 tempest-TenantUsagesTestJSON-327211524 tempest-TenantUsagesTestJSON-327211524-project-member] Refreshing inventories for resource provider 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 708.579087] env[65726]: DEBUG nova.scheduler.client.report [None req-fd996099-26ef-437f-83c3-98cb140e9de7 tempest-TenantUsagesTestJSON-327211524 tempest-TenantUsagesTestJSON-327211524-project-member] Updating ProviderTree inventory for provider 07c4692f-bdb4-4058-9173-ff9664830295 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 708.579087] env[65726]: DEBUG nova.compute.provider_tree [None req-fd996099-26ef-437f-83c3-98cb140e9de7 tempest-TenantUsagesTestJSON-327211524 tempest-TenantUsagesTestJSON-327211524-project-member] Updating inventory in ProviderTree for provider 07c4692f-bdb4-4058-9173-ff9664830295 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 708.590137] env[65726]: DEBUG nova.scheduler.client.report [None req-fd996099-26ef-437f-83c3-98cb140e9de7 tempest-TenantUsagesTestJSON-327211524 tempest-TenantUsagesTestJSON-327211524-project-member] Refreshing aggregate associations for resource provider 07c4692f-bdb4-4058-9173-ff9664830295, aggregates: None {{(pid=65726) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 708.609571] env[65726]: DEBUG oslo_vmware.api [None req-85e035ac-36aa-4f5f-93c0-1a4ac0d2c453 tempest-ServersAdminNegativeTestJSON-113649088 tempest-ServersAdminNegativeTestJSON-113649088-project-admin] Task: {'id': task-5115671, 'name': SuspendVM_Task, 'duration_secs': 0.681487} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 708.609571] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-85e035ac-36aa-4f5f-93c0-1a4ac0d2c453 tempest-ServersAdminNegativeTestJSON-113649088 tempest-ServersAdminNegativeTestJSON-113649088-project-admin] [instance: 423af936-a383-4652-8887-9c8b8f507909] Suspended the VM {{(pid=65726) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1166}} [ 708.610690] env[65726]: DEBUG nova.compute.manager [None req-85e035ac-36aa-4f5f-93c0-1a4ac0d2c453 tempest-ServersAdminNegativeTestJSON-113649088 tempest-ServersAdminNegativeTestJSON-113649088-project-admin] [instance: 423af936-a383-4652-8887-9c8b8f507909] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 708.610915] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f6def12-96a7-48ce-b684-a979d6e6cd89 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.614461] env[65726]: DEBUG nova.scheduler.client.report [None req-fd996099-26ef-437f-83c3-98cb140e9de7 tempest-TenantUsagesTestJSON-327211524 tempest-TenantUsagesTestJSON-327211524-project-member] Refreshing trait associations for resource provider 07c4692f-bdb4-4058-9173-ff9664830295, traits: COMPUTE_IMAGE_TYPE_ISO,HW_ARCH_X86_64,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=65726) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 708.633921] env[65726]: INFO nova.compute.manager [-] [instance: 30b17a17-229c-4ca7-9ae6-c67ccdcd3963] Took 1.35 seconds to deallocate network for instance. [ 708.821304] env[65726]: DEBUG nova.compute.manager [None req-f3171f49-8536-4923-b5cf-31399a616a07 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 608cc0ea-3f6b-4b4f-83c1-01aa50999d63] Start spawning the instance on the hypervisor. {{(pid=65726) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 708.862041] env[65726]: DEBUG nova.virt.hardware [None req-f3171f49-8536-4923-b5cf-31399a616a07 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T19:28:34Z,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1539f06fa3534e90acbc3a60c4b8bd3f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T19:28:35Z,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 708.862287] env[65726]: DEBUG nova.virt.hardware [None req-f3171f49-8536-4923-b5cf-31399a616a07 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 708.862488] env[65726]: DEBUG nova.virt.hardware [None req-f3171f49-8536-4923-b5cf-31399a616a07 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 708.862736] env[65726]: DEBUG nova.virt.hardware [None req-f3171f49-8536-4923-b5cf-31399a616a07 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 708.862937] env[65726]: DEBUG nova.virt.hardware [None req-f3171f49-8536-4923-b5cf-31399a616a07 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 708.863156] env[65726]: DEBUG nova.virt.hardware [None req-f3171f49-8536-4923-b5cf-31399a616a07 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 708.863452] env[65726]: DEBUG nova.virt.hardware [None req-f3171f49-8536-4923-b5cf-31399a616a07 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 708.863664] env[65726]: DEBUG nova.virt.hardware [None req-f3171f49-8536-4923-b5cf-31399a616a07 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 708.863893] env[65726]: DEBUG nova.virt.hardware [None req-f3171f49-8536-4923-b5cf-31399a616a07 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 708.864122] env[65726]: DEBUG nova.virt.hardware [None req-f3171f49-8536-4923-b5cf-31399a616a07 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 708.864376] env[65726]: DEBUG nova.virt.hardware [None req-f3171f49-8536-4923-b5cf-31399a616a07 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 708.865548] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83888dfc-3cbc-4de6-8c0a-84796f6d32e8 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.875336] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63b441c9-6cb4-497f-9e66-187be4fa1279 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.935729] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115672, 'name': CreateVM_Task} progress is 25%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 708.940888] env[65726]: DEBUG nova.compute.manager [req-7a3dda68-e3b0-4e03-93ab-db4f879544ae req-63925900-8002-40a4-9e9d-2f02e2cf25ad service nova] [instance: 30b17a17-229c-4ca7-9ae6-c67ccdcd3963] Received event network-vif-deleted-b1810a58-276a-4ccf-b223-ed03a7c987b1 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 709.143754] env[65726]: DEBUG oslo_concurrency.lockutils [None req-8b18678f-63d3-424e-98c8-800d5081389c tempest-VolumesAdminNegativeTest-2004477850 tempest-VolumesAdminNegativeTest-2004477850-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 709.191877] env[65726]: DEBUG nova.compute.manager [req-dd3ee866-f867-4595-9c5f-3b40b17e5d4f req-ff89f0f3-92a5-4e01-9822-9b0a457879bd service nova] [instance: 815f056b-2d88-40dd-9ef7-a891ec8d92cf] Received event network-vif-plugged-1dfb06a2-5081-48fc-a9ef-23b637910832 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 709.191983] env[65726]: DEBUG oslo_concurrency.lockutils [req-dd3ee866-f867-4595-9c5f-3b40b17e5d4f req-ff89f0f3-92a5-4e01-9822-9b0a457879bd service nova] Acquiring lock "815f056b-2d88-40dd-9ef7-a891ec8d92cf-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 709.192206] env[65726]: DEBUG oslo_concurrency.lockutils [req-dd3ee866-f867-4595-9c5f-3b40b17e5d4f req-ff89f0f3-92a5-4e01-9822-9b0a457879bd service nova] Lock "815f056b-2d88-40dd-9ef7-a891ec8d92cf-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 709.192376] env[65726]: DEBUG oslo_concurrency.lockutils [req-dd3ee866-f867-4595-9c5f-3b40b17e5d4f req-ff89f0f3-92a5-4e01-9822-9b0a457879bd service nova] Lock "815f056b-2d88-40dd-9ef7-a891ec8d92cf-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 709.192545] env[65726]: DEBUG nova.compute.manager [req-dd3ee866-f867-4595-9c5f-3b40b17e5d4f req-ff89f0f3-92a5-4e01-9822-9b0a457879bd service nova] [instance: 815f056b-2d88-40dd-9ef7-a891ec8d92cf] No waiting events found dispatching network-vif-plugged-1dfb06a2-5081-48fc-a9ef-23b637910832 {{(pid=65726) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 709.192729] env[65726]: WARNING nova.compute.manager [req-dd3ee866-f867-4595-9c5f-3b40b17e5d4f req-ff89f0f3-92a5-4e01-9822-9b0a457879bd service nova] [instance: 815f056b-2d88-40dd-9ef7-a891ec8d92cf] Received unexpected event network-vif-plugged-1dfb06a2-5081-48fc-a9ef-23b637910832 for instance with vm_state building and task_state spawning. [ 709.192962] env[65726]: DEBUG nova.compute.manager [req-dd3ee866-f867-4595-9c5f-3b40b17e5d4f req-ff89f0f3-92a5-4e01-9822-9b0a457879bd service nova] [instance: 815f056b-2d88-40dd-9ef7-a891ec8d92cf] Received event network-changed-1dfb06a2-5081-48fc-a9ef-23b637910832 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 709.193134] env[65726]: DEBUG nova.compute.manager [req-dd3ee866-f867-4595-9c5f-3b40b17e5d4f req-ff89f0f3-92a5-4e01-9822-9b0a457879bd service nova] [instance: 815f056b-2d88-40dd-9ef7-a891ec8d92cf] Refreshing instance network info cache due to event network-changed-1dfb06a2-5081-48fc-a9ef-23b637910832. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 709.193350] env[65726]: DEBUG oslo_concurrency.lockutils [req-dd3ee866-f867-4595-9c5f-3b40b17e5d4f req-ff89f0f3-92a5-4e01-9822-9b0a457879bd service nova] Acquiring lock "refresh_cache-815f056b-2d88-40dd-9ef7-a891ec8d92cf" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 709.193514] env[65726]: DEBUG oslo_concurrency.lockutils [req-dd3ee866-f867-4595-9c5f-3b40b17e5d4f req-ff89f0f3-92a5-4e01-9822-9b0a457879bd service nova] Acquired lock "refresh_cache-815f056b-2d88-40dd-9ef7-a891ec8d92cf" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 709.193670] env[65726]: DEBUG nova.network.neutron [req-dd3ee866-f867-4595-9c5f-3b40b17e5d4f req-ff89f0f3-92a5-4e01-9822-9b0a457879bd service nova] [instance: 815f056b-2d88-40dd-9ef7-a891ec8d92cf] Refreshing network info cache for port 1dfb06a2-5081-48fc-a9ef-23b637910832 {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 709.218200] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14255055-283f-4126-8a9c-71d1453e22ab {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.226891] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-706e8b8f-9325-42ea-be88-cdbd9110fac2 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.258900] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fd8c1a1-72c0-470e-b9ab-d9d7006ea205 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.268187] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5f0870f-bb68-4d09-ad86-a4da8032725b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.283251] env[65726]: DEBUG nova.compute.provider_tree [None req-fd996099-26ef-437f-83c3-98cb140e9de7 tempest-TenantUsagesTestJSON-327211524 tempest-TenantUsagesTestJSON-327211524-project-member] Updating inventory in ProviderTree for provider 07c4692f-bdb4-4058-9173-ff9664830295 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 709.413219] env[65726]: DEBUG nova.network.neutron [None req-f3171f49-8536-4923-b5cf-31399a616a07 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 608cc0ea-3f6b-4b4f-83c1-01aa50999d63] Successfully updated port: e9f7132a-8d5a-47fe-afc6-d45bf3687ea4 {{(pid=65726) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 709.436386] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115672, 'name': CreateVM_Task, 'duration_secs': 0.851515} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 709.436386] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 815f056b-2d88-40dd-9ef7-a891ec8d92cf] Created VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 709.437374] env[65726]: WARNING openstack [None req-5dd7a34f-2002-4a12-9099-b0b630d2774d tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 709.437717] env[65726]: WARNING openstack [None req-5dd7a34f-2002-4a12-9099-b0b630d2774d tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 709.442788] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5dd7a34f-2002-4a12-9099-b0b630d2774d tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 709.442940] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5dd7a34f-2002-4a12-9099-b0b630d2774d tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 709.443263] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5dd7a34f-2002-4a12-9099-b0b630d2774d tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 709.443536] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4f3da7ce-0374-49ff-bf22-bc7246a15345 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.449899] env[65726]: DEBUG oslo_vmware.api [None req-5dd7a34f-2002-4a12-9099-b0b630d2774d tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Waiting for the task: (returnval){ [ 709.449899] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52fe8e95-3e09-0fda-d3aa-c9aef96ac8de" [ 709.449899] env[65726]: _type = "Task" [ 709.449899] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 709.458593] env[65726]: DEBUG oslo_vmware.api [None req-5dd7a34f-2002-4a12-9099-b0b630d2774d tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52fe8e95-3e09-0fda-d3aa-c9aef96ac8de, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 709.702878] env[65726]: WARNING openstack [req-dd3ee866-f867-4595-9c5f-3b40b17e5d4f req-ff89f0f3-92a5-4e01-9822-9b0a457879bd service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 709.702878] env[65726]: WARNING openstack [req-dd3ee866-f867-4595-9c5f-3b40b17e5d4f req-ff89f0f3-92a5-4e01-9822-9b0a457879bd service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 709.817087] env[65726]: WARNING openstack [req-dd3ee866-f867-4595-9c5f-3b40b17e5d4f req-ff89f0f3-92a5-4e01-9822-9b0a457879bd service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 709.817202] env[65726]: WARNING openstack [req-dd3ee866-f867-4595-9c5f-3b40b17e5d4f req-ff89f0f3-92a5-4e01-9822-9b0a457879bd service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 709.824507] env[65726]: DEBUG nova.scheduler.client.report [None req-fd996099-26ef-437f-83c3-98cb140e9de7 tempest-TenantUsagesTestJSON-327211524 tempest-TenantUsagesTestJSON-327211524-project-member] Updated inventory for provider 07c4692f-bdb4-4058-9173-ff9664830295 with generation 57 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 709.824852] env[65726]: DEBUG nova.compute.provider_tree [None req-fd996099-26ef-437f-83c3-98cb140e9de7 tempest-TenantUsagesTestJSON-327211524 tempest-TenantUsagesTestJSON-327211524-project-member] Updating resource provider 07c4692f-bdb4-4058-9173-ff9664830295 generation from 57 to 58 during operation: update_inventory {{(pid=65726) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 709.824918] env[65726]: DEBUG nova.compute.provider_tree [None req-fd996099-26ef-437f-83c3-98cb140e9de7 tempest-TenantUsagesTestJSON-327211524 tempest-TenantUsagesTestJSON-327211524-project-member] Updating inventory in ProviderTree for provider 07c4692f-bdb4-4058-9173-ff9664830295 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 709.895161] env[65726]: WARNING openstack [req-dd3ee866-f867-4595-9c5f-3b40b17e5d4f req-ff89f0f3-92a5-4e01-9822-9b0a457879bd service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 709.895161] env[65726]: WARNING openstack [req-dd3ee866-f867-4595-9c5f-3b40b17e5d4f req-ff89f0f3-92a5-4e01-9822-9b0a457879bd service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 709.917737] env[65726]: DEBUG oslo_concurrency.lockutils [None req-f3171f49-8536-4923-b5cf-31399a616a07 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Acquiring lock "refresh_cache-608cc0ea-3f6b-4b4f-83c1-01aa50999d63" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 709.917737] env[65726]: DEBUG oslo_concurrency.lockutils [None req-f3171f49-8536-4923-b5cf-31399a616a07 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Acquired lock "refresh_cache-608cc0ea-3f6b-4b4f-83c1-01aa50999d63" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 709.917891] env[65726]: DEBUG nova.network.neutron [None req-f3171f49-8536-4923-b5cf-31399a616a07 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 608cc0ea-3f6b-4b4f-83c1-01aa50999d63] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 709.960455] env[65726]: DEBUG oslo_vmware.api [None req-5dd7a34f-2002-4a12-9099-b0b630d2774d tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52fe8e95-3e09-0fda-d3aa-c9aef96ac8de, 'name': SearchDatastore_Task, 'duration_secs': 0.011883} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 709.961163] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5dd7a34f-2002-4a12-9099-b0b630d2774d tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 709.961411] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-5dd7a34f-2002-4a12-9099-b0b630d2774d tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] [instance: 815f056b-2d88-40dd-9ef7-a891ec8d92cf] Processing image b52362a3-ee8a-4cbf-b06f-513b0cc8f95c {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 709.961675] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5dd7a34f-2002-4a12-9099-b0b630d2774d tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 709.961818] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5dd7a34f-2002-4a12-9099-b0b630d2774d tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 709.961993] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-5dd7a34f-2002-4a12-9099-b0b630d2774d tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 709.962322] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d49f3936-96c9-4e95-a673-4d02b33a4b74 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.972546] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-5dd7a34f-2002-4a12-9099-b0b630d2774d tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 709.972764] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-5dd7a34f-2002-4a12-9099-b0b630d2774d tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65726) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 709.973571] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-37c7ca4a-2828-4612-9d42-7fa241b35b19 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.980522] env[65726]: DEBUG oslo_vmware.api [None req-5dd7a34f-2002-4a12-9099-b0b630d2774d tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Waiting for the task: (returnval){ [ 709.980522] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]524aa715-441b-1278-727c-ba259be097a7" [ 709.980522] env[65726]: _type = "Task" [ 709.980522] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 709.988085] env[65726]: DEBUG nova.network.neutron [req-dd3ee866-f867-4595-9c5f-3b40b17e5d4f req-ff89f0f3-92a5-4e01-9822-9b0a457879bd service nova] [instance: 815f056b-2d88-40dd-9ef7-a891ec8d92cf] Updated VIF entry in instance network info cache for port 1dfb06a2-5081-48fc-a9ef-23b637910832. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 709.988482] env[65726]: DEBUG nova.network.neutron [req-dd3ee866-f867-4595-9c5f-3b40b17e5d4f req-ff89f0f3-92a5-4e01-9822-9b0a457879bd service nova] [instance: 815f056b-2d88-40dd-9ef7-a891ec8d92cf] Updating instance_info_cache with network_info: [{"id": "1dfb06a2-5081-48fc-a9ef-23b637910832", "address": "fa:16:3e:84:38:5a", "network": {"id": "55d629b8-a541-4720-9289-927bbc6e6f22", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.118", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "1539f06fa3534e90acbc3a60c4b8bd3f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "35342bcb-8b06-472e-b3c0-43fd3d6c4b30", "external-id": "nsx-vlan-transportzone-524", "segmentation_id": 524, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1dfb06a2-50", "ovs_interfaceid": "1dfb06a2-5081-48fc-a9ef-23b637910832", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 709.998025] env[65726]: DEBUG oslo_vmware.api [None req-5dd7a34f-2002-4a12-9099-b0b630d2774d tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]524aa715-441b-1278-727c-ba259be097a7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 710.330948] env[65726]: DEBUG oslo_concurrency.lockutils [None req-fd996099-26ef-437f-83c3-98cb140e9de7 tempest-TenantUsagesTestJSON-327211524 tempest-TenantUsagesTestJSON-327211524-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 3.530s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 710.333429] env[65726]: DEBUG oslo_concurrency.lockutils [None req-296fd45e-ee7a-4e44-912a-f3b5a22cbbf0 tempest-ServerExternalEventsTest-1845125219 tempest-ServerExternalEventsTest-1845125219-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 34.194s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 710.334942] env[65726]: INFO nova.compute.claims [None req-296fd45e-ee7a-4e44-912a-f3b5a22cbbf0 tempest-ServerExternalEventsTest-1845125219 tempest-ServerExternalEventsTest-1845125219-project-member] [instance: 96acc5f8-b411-408e-929d-d1035cfd50db] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 710.357287] env[65726]: INFO nova.scheduler.client.report [None req-fd996099-26ef-437f-83c3-98cb140e9de7 tempest-TenantUsagesTestJSON-327211524 tempest-TenantUsagesTestJSON-327211524-project-member] Deleted allocations for instance 7748c23f-9ea9-4172-8ab7-187678272bb3 [ 710.420920] env[65726]: WARNING openstack [None req-f3171f49-8536-4923-b5cf-31399a616a07 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 710.421312] env[65726]: WARNING openstack [None req-f3171f49-8536-4923-b5cf-31399a616a07 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 710.426515] env[65726]: DEBUG nova.network.neutron [None req-f3171f49-8536-4923-b5cf-31399a616a07 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 608cc0ea-3f6b-4b4f-83c1-01aa50999d63] Instance cache missing network info. {{(pid=65726) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 710.488465] env[65726]: WARNING openstack [None req-f3171f49-8536-4923-b5cf-31399a616a07 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 710.488845] env[65726]: WARNING openstack [None req-f3171f49-8536-4923-b5cf-31399a616a07 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 710.500716] env[65726]: DEBUG oslo_concurrency.lockutils [req-dd3ee866-f867-4595-9c5f-3b40b17e5d4f req-ff89f0f3-92a5-4e01-9822-9b0a457879bd service nova] Releasing lock "refresh_cache-815f056b-2d88-40dd-9ef7-a891ec8d92cf" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 710.501155] env[65726]: DEBUG oslo_vmware.api [None req-5dd7a34f-2002-4a12-9099-b0b630d2774d tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]524aa715-441b-1278-727c-ba259be097a7, 'name': SearchDatastore_Task, 'duration_secs': 0.010988} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 710.502211] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-05711a0d-bc66-4a16-94d7-3019b264cf9b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.508926] env[65726]: DEBUG oslo_vmware.api [None req-5dd7a34f-2002-4a12-9099-b0b630d2774d tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Waiting for the task: (returnval){ [ 710.508926] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]522318f7-85e4-a8e1-7d96-4ad713a21368" [ 710.508926] env[65726]: _type = "Task" [ 710.508926] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 710.518057] env[65726]: DEBUG oslo_vmware.api [None req-5dd7a34f-2002-4a12-9099-b0b630d2774d tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]522318f7-85e4-a8e1-7d96-4ad713a21368, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 710.570706] env[65726]: WARNING openstack [None req-f3171f49-8536-4923-b5cf-31399a616a07 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 710.571112] env[65726]: WARNING openstack [None req-f3171f49-8536-4923-b5cf-31399a616a07 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 710.665165] env[65726]: DEBUG nova.network.neutron [None req-f3171f49-8536-4923-b5cf-31399a616a07 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 608cc0ea-3f6b-4b4f-83c1-01aa50999d63] Updating instance_info_cache with network_info: [{"id": "e9f7132a-8d5a-47fe-afc6-d45bf3687ea4", "address": "fa:16:3e:fe:4d:df", "network": {"id": "0d0f561d-ac41-4ef5-9926-c707409799f7", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1134321217-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4eece77569624f90bf64e5c51974173f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2ce62383-8e84-4e26-955b-74c11392f4c9", "external-id": "nsx-vlan-transportzone-215", "segmentation_id": 215, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape9f7132a-8d", "ovs_interfaceid": "e9f7132a-8d5a-47fe-afc6-d45bf3687ea4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 710.735395] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c00c2da9-ebd0-4704-a276-dac36258bf53 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Acquiring lock "f5151062-57b1-4e4e-93f4-aab0e7f504d4" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 710.735395] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c00c2da9-ebd0-4704-a276-dac36258bf53 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Lock "f5151062-57b1-4e4e-93f4-aab0e7f504d4" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 710.735696] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c00c2da9-ebd0-4704-a276-dac36258bf53 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Acquiring lock "f5151062-57b1-4e4e-93f4-aab0e7f504d4-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 710.736745] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c00c2da9-ebd0-4704-a276-dac36258bf53 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Lock "f5151062-57b1-4e4e-93f4-aab0e7f504d4-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 710.736745] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c00c2da9-ebd0-4704-a276-dac36258bf53 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Lock "f5151062-57b1-4e4e-93f4-aab0e7f504d4-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 710.738540] env[65726]: INFO nova.compute.manager [None req-c00c2da9-ebd0-4704-a276-dac36258bf53 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] [instance: f5151062-57b1-4e4e-93f4-aab0e7f504d4] Terminating instance [ 710.866522] env[65726]: DEBUG oslo_concurrency.lockutils [None req-fd996099-26ef-437f-83c3-98cb140e9de7 tempest-TenantUsagesTestJSON-327211524 tempest-TenantUsagesTestJSON-327211524-project-member] Lock "7748c23f-9ea9-4172-8ab7-187678272bb3" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 38.276s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 711.020501] env[65726]: DEBUG oslo_vmware.api [None req-5dd7a34f-2002-4a12-9099-b0b630d2774d tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]522318f7-85e4-a8e1-7d96-4ad713a21368, 'name': SearchDatastore_Task, 'duration_secs': 0.011426} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 711.020696] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5dd7a34f-2002-4a12-9099-b0b630d2774d tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 711.021105] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-5dd7a34f-2002-4a12-9099-b0b630d2774d tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] 815f056b-2d88-40dd-9ef7-a891ec8d92cf/815f056b-2d88-40dd-9ef7-a891ec8d92cf.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 711.021446] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a1195f21-0779-40b0-afbd-38846034b737 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.029811] env[65726]: DEBUG oslo_vmware.api [None req-5dd7a34f-2002-4a12-9099-b0b630d2774d tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Waiting for the task: (returnval){ [ 711.029811] env[65726]: value = "task-5115673" [ 711.029811] env[65726]: _type = "Task" [ 711.029811] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 711.043326] env[65726]: DEBUG oslo_vmware.api [None req-5dd7a34f-2002-4a12-9099-b0b630d2774d tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Task: {'id': task-5115673, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 711.168180] env[65726]: DEBUG oslo_concurrency.lockutils [None req-f3171f49-8536-4923-b5cf-31399a616a07 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Releasing lock "refresh_cache-608cc0ea-3f6b-4b4f-83c1-01aa50999d63" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 711.168606] env[65726]: DEBUG nova.compute.manager [None req-f3171f49-8536-4923-b5cf-31399a616a07 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 608cc0ea-3f6b-4b4f-83c1-01aa50999d63] Instance network_info: |[{"id": "e9f7132a-8d5a-47fe-afc6-d45bf3687ea4", "address": "fa:16:3e:fe:4d:df", "network": {"id": "0d0f561d-ac41-4ef5-9926-c707409799f7", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1134321217-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4eece77569624f90bf64e5c51974173f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2ce62383-8e84-4e26-955b-74c11392f4c9", "external-id": "nsx-vlan-transportzone-215", "segmentation_id": 215, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape9f7132a-8d", "ovs_interfaceid": "e9f7132a-8d5a-47fe-afc6-d45bf3687ea4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 711.169149] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-f3171f49-8536-4923-b5cf-31399a616a07 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 608cc0ea-3f6b-4b4f-83c1-01aa50999d63] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:fe:4d:df', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '2ce62383-8e84-4e26-955b-74c11392f4c9', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e9f7132a-8d5a-47fe-afc6-d45bf3687ea4', 'vif_model': 'vmxnet3'}] {{(pid=65726) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 711.176848] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-f3171f49-8536-4923-b5cf-31399a616a07 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Creating folder: Project (4eece77569624f90bf64e5c51974173f). Parent ref: group-v995008. {{(pid=65726) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 711.177274] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-6d74f65b-2e28-435c-b0c3-2c266a32e076 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.192390] env[65726]: INFO nova.virt.vmwareapi.vm_util [None req-f3171f49-8536-4923-b5cf-31399a616a07 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Created folder: Project (4eece77569624f90bf64e5c51974173f) in parent group-v995008. [ 711.192572] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-f3171f49-8536-4923-b5cf-31399a616a07 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Creating folder: Instances. Parent ref: group-v995091. {{(pid=65726) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 711.193785] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-1e45bbb6-cc14-411b-8267-c505b06fc6c8 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.206844] env[65726]: INFO nova.virt.vmwareapi.vm_util [None req-f3171f49-8536-4923-b5cf-31399a616a07 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Created folder: Instances in parent group-v995091. [ 711.206844] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-f3171f49-8536-4923-b5cf-31399a616a07 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 711.206844] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 608cc0ea-3f6b-4b4f-83c1-01aa50999d63] Creating VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 711.206844] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8cb4bbd1-3919-4658-a510-1a8e32044010 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.235058] env[65726]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 711.235058] env[65726]: value = "task-5115676" [ 711.235058] env[65726]: _type = "Task" [ 711.235058] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 711.244114] env[65726]: DEBUG nova.compute.manager [None req-c00c2da9-ebd0-4704-a276-dac36258bf53 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] [instance: f5151062-57b1-4e4e-93f4-aab0e7f504d4] Start destroying the instance on the hypervisor. {{(pid=65726) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 711.244385] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-c00c2da9-ebd0-4704-a276-dac36258bf53 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] [instance: f5151062-57b1-4e4e-93f4-aab0e7f504d4] Destroying instance {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 711.248614] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4dc36fb5-1a75-40a3-95af-f74455129e2b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.252302] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115676, 'name': CreateVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 711.259456] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-c00c2da9-ebd0-4704-a276-dac36258bf53 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] [instance: f5151062-57b1-4e4e-93f4-aab0e7f504d4] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 711.260043] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-77c048f1-10c3-4e49-b8da-fe36f09e85ab {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.270796] env[65726]: DEBUG oslo_vmware.api [None req-c00c2da9-ebd0-4704-a276-dac36258bf53 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Waiting for the task: (returnval){ [ 711.270796] env[65726]: value = "task-5115677" [ 711.270796] env[65726]: _type = "Task" [ 711.270796] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 711.283519] env[65726]: DEBUG oslo_vmware.api [None req-c00c2da9-ebd0-4704-a276-dac36258bf53 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Task: {'id': task-5115677, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 711.544312] env[65726]: DEBUG oslo_vmware.api [None req-5dd7a34f-2002-4a12-9099-b0b630d2774d tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Task: {'id': task-5115673, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 711.565936] env[65726]: DEBUG nova.compute.manager [req-abe14589-e2ed-4063-a1f6-da86e9a6123a req-d55c367e-0b7d-43e3-907c-06f37f06725a service nova] [instance: 608cc0ea-3f6b-4b4f-83c1-01aa50999d63] Received event network-vif-plugged-e9f7132a-8d5a-47fe-afc6-d45bf3687ea4 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 711.566089] env[65726]: DEBUG oslo_concurrency.lockutils [req-abe14589-e2ed-4063-a1f6-da86e9a6123a req-d55c367e-0b7d-43e3-907c-06f37f06725a service nova] Acquiring lock "608cc0ea-3f6b-4b4f-83c1-01aa50999d63-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 711.566735] env[65726]: DEBUG oslo_concurrency.lockutils [req-abe14589-e2ed-4063-a1f6-da86e9a6123a req-d55c367e-0b7d-43e3-907c-06f37f06725a service nova] Lock "608cc0ea-3f6b-4b4f-83c1-01aa50999d63-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 711.566735] env[65726]: DEBUG oslo_concurrency.lockutils [req-abe14589-e2ed-4063-a1f6-da86e9a6123a req-d55c367e-0b7d-43e3-907c-06f37f06725a service nova] Lock "608cc0ea-3f6b-4b4f-83c1-01aa50999d63-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 711.566735] env[65726]: DEBUG nova.compute.manager [req-abe14589-e2ed-4063-a1f6-da86e9a6123a req-d55c367e-0b7d-43e3-907c-06f37f06725a service nova] [instance: 608cc0ea-3f6b-4b4f-83c1-01aa50999d63] No waiting events found dispatching network-vif-plugged-e9f7132a-8d5a-47fe-afc6-d45bf3687ea4 {{(pid=65726) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 711.567382] env[65726]: WARNING nova.compute.manager [req-abe14589-e2ed-4063-a1f6-da86e9a6123a req-d55c367e-0b7d-43e3-907c-06f37f06725a service nova] [instance: 608cc0ea-3f6b-4b4f-83c1-01aa50999d63] Received unexpected event network-vif-plugged-e9f7132a-8d5a-47fe-afc6-d45bf3687ea4 for instance with vm_state building and task_state spawning. [ 711.567382] env[65726]: DEBUG nova.compute.manager [req-abe14589-e2ed-4063-a1f6-da86e9a6123a req-d55c367e-0b7d-43e3-907c-06f37f06725a service nova] [instance: 608cc0ea-3f6b-4b4f-83c1-01aa50999d63] Received event network-changed-e9f7132a-8d5a-47fe-afc6-d45bf3687ea4 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 711.567382] env[65726]: DEBUG nova.compute.manager [req-abe14589-e2ed-4063-a1f6-da86e9a6123a req-d55c367e-0b7d-43e3-907c-06f37f06725a service nova] [instance: 608cc0ea-3f6b-4b4f-83c1-01aa50999d63] Refreshing instance network info cache due to event network-changed-e9f7132a-8d5a-47fe-afc6-d45bf3687ea4. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 711.567382] env[65726]: DEBUG oslo_concurrency.lockutils [req-abe14589-e2ed-4063-a1f6-da86e9a6123a req-d55c367e-0b7d-43e3-907c-06f37f06725a service nova] Acquiring lock "refresh_cache-608cc0ea-3f6b-4b4f-83c1-01aa50999d63" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 711.567382] env[65726]: DEBUG oslo_concurrency.lockutils [req-abe14589-e2ed-4063-a1f6-da86e9a6123a req-d55c367e-0b7d-43e3-907c-06f37f06725a service nova] Acquired lock "refresh_cache-608cc0ea-3f6b-4b4f-83c1-01aa50999d63" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 711.567581] env[65726]: DEBUG nova.network.neutron [req-abe14589-e2ed-4063-a1f6-da86e9a6123a req-d55c367e-0b7d-43e3-907c-06f37f06725a service nova] [instance: 608cc0ea-3f6b-4b4f-83c1-01aa50999d63] Refreshing network info cache for port e9f7132a-8d5a-47fe-afc6-d45bf3687ea4 {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 711.730033] env[65726]: DEBUG oslo_concurrency.lockutils [None req-1023a64a-e0ea-45c1-b4cf-b0858cc7b3e9 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Acquiring lock "3a60b719-9db7-436e-9908-25f6a233c465" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 711.730322] env[65726]: DEBUG oslo_concurrency.lockutils [None req-1023a64a-e0ea-45c1-b4cf-b0858cc7b3e9 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Lock "3a60b719-9db7-436e-9908-25f6a233c465" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 711.730533] env[65726]: DEBUG oslo_concurrency.lockutils [None req-1023a64a-e0ea-45c1-b4cf-b0858cc7b3e9 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Acquiring lock "3a60b719-9db7-436e-9908-25f6a233c465-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 711.730711] env[65726]: DEBUG oslo_concurrency.lockutils [None req-1023a64a-e0ea-45c1-b4cf-b0858cc7b3e9 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Lock "3a60b719-9db7-436e-9908-25f6a233c465-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 711.730874] env[65726]: DEBUG oslo_concurrency.lockutils [None req-1023a64a-e0ea-45c1-b4cf-b0858cc7b3e9 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Lock "3a60b719-9db7-436e-9908-25f6a233c465-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 711.736835] env[65726]: INFO nova.compute.manager [None req-1023a64a-e0ea-45c1-b4cf-b0858cc7b3e9 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] [instance: 3a60b719-9db7-436e-9908-25f6a233c465] Terminating instance [ 711.753550] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115676, 'name': CreateVM_Task, 'duration_secs': 0.431088} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 711.755158] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 608cc0ea-3f6b-4b4f-83c1-01aa50999d63] Created VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 711.755347] env[65726]: WARNING openstack [None req-f3171f49-8536-4923-b5cf-31399a616a07 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 711.755612] env[65726]: WARNING openstack [None req-f3171f49-8536-4923-b5cf-31399a616a07 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 711.761621] env[65726]: DEBUG oslo_concurrency.lockutils [None req-f3171f49-8536-4923-b5cf-31399a616a07 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 711.761621] env[65726]: DEBUG oslo_concurrency.lockutils [None req-f3171f49-8536-4923-b5cf-31399a616a07 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 711.761621] env[65726]: DEBUG oslo_concurrency.lockutils [None req-f3171f49-8536-4923-b5cf-31399a616a07 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 711.764281] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-576dadc5-aae4-4387-8ceb-97ef98db6076 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.771423] env[65726]: DEBUG oslo_vmware.api [None req-f3171f49-8536-4923-b5cf-31399a616a07 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Waiting for the task: (returnval){ [ 711.771423] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52ac5fa6-fd64-85f4-5e22-9206a54d4b04" [ 711.771423] env[65726]: _type = "Task" [ 711.771423] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 711.787856] env[65726]: DEBUG oslo_vmware.api [None req-c00c2da9-ebd0-4704-a276-dac36258bf53 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Task: {'id': task-5115677, 'name': PowerOffVM_Task, 'duration_secs': 0.314214} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 711.791379] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-c00c2da9-ebd0-4704-a276-dac36258bf53 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] [instance: f5151062-57b1-4e4e-93f4-aab0e7f504d4] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 711.791560] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-c00c2da9-ebd0-4704-a276-dac36258bf53 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] [instance: f5151062-57b1-4e4e-93f4-aab0e7f504d4] Unregistering the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 711.791817] env[65726]: DEBUG oslo_vmware.api [None req-f3171f49-8536-4923-b5cf-31399a616a07 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52ac5fa6-fd64-85f4-5e22-9206a54d4b04, 'name': SearchDatastore_Task, 'duration_secs': 0.010035} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 711.794276] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-85fb013f-c77a-4893-93ce-c46fb6bbdc52 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.795819] env[65726]: DEBUG oslo_concurrency.lockutils [None req-f3171f49-8536-4923-b5cf-31399a616a07 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 711.796051] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-f3171f49-8536-4923-b5cf-31399a616a07 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 608cc0ea-3f6b-4b4f-83c1-01aa50999d63] Processing image b52362a3-ee8a-4cbf-b06f-513b0cc8f95c {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 711.796293] env[65726]: DEBUG oslo_concurrency.lockutils [None req-f3171f49-8536-4923-b5cf-31399a616a07 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 711.796434] env[65726]: DEBUG oslo_concurrency.lockutils [None req-f3171f49-8536-4923-b5cf-31399a616a07 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 711.796606] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-f3171f49-8536-4923-b5cf-31399a616a07 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 711.797349] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-09a2a0bd-d448-48cb-80db-133873357c4f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.808154] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-f3171f49-8536-4923-b5cf-31399a616a07 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 711.808348] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-f3171f49-8536-4923-b5cf-31399a616a07 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65726) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 711.809202] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9b98b3d3-58b3-424f-929b-4072581d209b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.819286] env[65726]: DEBUG oslo_vmware.api [None req-f3171f49-8536-4923-b5cf-31399a616a07 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Waiting for the task: (returnval){ [ 711.819286] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52f75404-5eed-170c-48a1-4917f0dba11e" [ 711.819286] env[65726]: _type = "Task" [ 711.819286] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 711.831021] env[65726]: DEBUG oslo_concurrency.lockutils [None req-681fb0ed-e41c-4d7e-b603-04df1dff2263 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Acquiring lock "04ffdf80-c54c-4587-9bb4-d520dc440501" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 711.831321] env[65726]: DEBUG oslo_concurrency.lockutils [None req-681fb0ed-e41c-4d7e-b603-04df1dff2263 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Lock "04ffdf80-c54c-4587-9bb4-d520dc440501" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 711.831519] env[65726]: DEBUG oslo_concurrency.lockutils [None req-681fb0ed-e41c-4d7e-b603-04df1dff2263 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Acquiring lock "04ffdf80-c54c-4587-9bb4-d520dc440501-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 711.831695] env[65726]: DEBUG oslo_concurrency.lockutils [None req-681fb0ed-e41c-4d7e-b603-04df1dff2263 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Lock "04ffdf80-c54c-4587-9bb4-d520dc440501-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 711.831862] env[65726]: DEBUG oslo_concurrency.lockutils [None req-681fb0ed-e41c-4d7e-b603-04df1dff2263 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Lock "04ffdf80-c54c-4587-9bb4-d520dc440501-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 711.833706] env[65726]: DEBUG oslo_vmware.api [None req-f3171f49-8536-4923-b5cf-31399a616a07 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52f75404-5eed-170c-48a1-4917f0dba11e, 'name': SearchDatastore_Task, 'duration_secs': 0.011094} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 711.838471] env[65726]: INFO nova.compute.manager [None req-681fb0ed-e41c-4d7e-b603-04df1dff2263 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] [instance: 04ffdf80-c54c-4587-9bb4-d520dc440501] Terminating instance [ 711.841109] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-58ecfe6a-0300-4c0e-b61a-41167443437d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.849455] env[65726]: DEBUG oslo_vmware.api [None req-f3171f49-8536-4923-b5cf-31399a616a07 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Waiting for the task: (returnval){ [ 711.849455] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52e3c87a-8a95-0cbc-a192-32a3aadd8599" [ 711.849455] env[65726]: _type = "Task" [ 711.849455] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 711.860959] env[65726]: DEBUG oslo_vmware.api [None req-f3171f49-8536-4923-b5cf-31399a616a07 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52e3c87a-8a95-0cbc-a192-32a3aadd8599, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 711.867218] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-c00c2da9-ebd0-4704-a276-dac36258bf53 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] [instance: f5151062-57b1-4e4e-93f4-aab0e7f504d4] Unregistered the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 711.867457] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-c00c2da9-ebd0-4704-a276-dac36258bf53 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] [instance: f5151062-57b1-4e4e-93f4-aab0e7f504d4] Deleting contents of the VM from datastore datastore2 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 711.867457] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-c00c2da9-ebd0-4704-a276-dac36258bf53 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Deleting the datastore file [datastore2] f5151062-57b1-4e4e-93f4-aab0e7f504d4 {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 711.868511] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e57ab761-79bd-4e0d-9166-11fb21328a5b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.875817] env[65726]: DEBUG oslo_vmware.api [None req-c00c2da9-ebd0-4704-a276-dac36258bf53 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Waiting for the task: (returnval){ [ 711.875817] env[65726]: value = "task-5115679" [ 711.875817] env[65726]: _type = "Task" [ 711.875817] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 711.886351] env[65726]: DEBUG oslo_vmware.api [None req-c00c2da9-ebd0-4704-a276-dac36258bf53 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Task: {'id': task-5115679, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 712.027725] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3f099b5-1dcb-4c39-a5ef-df977430058e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.040248] env[65726]: DEBUG oslo_vmware.api [None req-5dd7a34f-2002-4a12-9099-b0b630d2774d tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Task: {'id': task-5115673, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.537515} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 712.042466] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-5dd7a34f-2002-4a12-9099-b0b630d2774d tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] 815f056b-2d88-40dd-9ef7-a891ec8d92cf/815f056b-2d88-40dd-9ef7-a891ec8d92cf.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 712.042713] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-5dd7a34f-2002-4a12-9099-b0b630d2774d tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] [instance: 815f056b-2d88-40dd-9ef7-a891ec8d92cf] Extending root virtual disk to 1048576 {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 712.043052] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-37a5a84a-dad9-4f21-aac1-b64e45c28985 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.046177] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2612658-8f4b-4c9e-86ce-23398d2d9b55 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.081509] env[65726]: WARNING openstack [req-abe14589-e2ed-4063-a1f6-da86e9a6123a req-d55c367e-0b7d-43e3-907c-06f37f06725a service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 712.082534] env[65726]: WARNING openstack [req-abe14589-e2ed-4063-a1f6-da86e9a6123a req-d55c367e-0b7d-43e3-907c-06f37f06725a service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 712.092939] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5364cd4-0ed8-48c1-8c09-7afb8a9af0cc {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.096367] env[65726]: DEBUG oslo_vmware.api [None req-5dd7a34f-2002-4a12-9099-b0b630d2774d tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Waiting for the task: (returnval){ [ 712.096367] env[65726]: value = "task-5115680" [ 712.096367] env[65726]: _type = "Task" [ 712.096367] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 712.105959] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57fc385f-9c69-4396-bb46-e51c6c47e9b9 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.114758] env[65726]: DEBUG oslo_vmware.api [None req-5dd7a34f-2002-4a12-9099-b0b630d2774d tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Task: {'id': task-5115680, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 712.127644] env[65726]: DEBUG nova.compute.provider_tree [None req-296fd45e-ee7a-4e44-912a-f3b5a22cbbf0 tempest-ServerExternalEventsTest-1845125219 tempest-ServerExternalEventsTest-1845125219-project-member] Updating inventory in ProviderTree for provider 07c4692f-bdb4-4058-9173-ff9664830295 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 712.248872] env[65726]: DEBUG nova.compute.manager [None req-1023a64a-e0ea-45c1-b4cf-b0858cc7b3e9 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] [instance: 3a60b719-9db7-436e-9908-25f6a233c465] Start destroying the instance on the hypervisor. {{(pid=65726) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 712.249164] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-1023a64a-e0ea-45c1-b4cf-b0858cc7b3e9 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] [instance: 3a60b719-9db7-436e-9908-25f6a233c465] Destroying instance {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 712.250616] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a566ca64-f44c-4e75-bd9f-7ec24f54891d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.260316] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-1023a64a-e0ea-45c1-b4cf-b0858cc7b3e9 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] [instance: 3a60b719-9db7-436e-9908-25f6a233c465] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 712.260638] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-85fd6101-78f9-4221-bc72-9595c41a15c0 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.269661] env[65726]: DEBUG oslo_vmware.api [None req-1023a64a-e0ea-45c1-b4cf-b0858cc7b3e9 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Waiting for the task: (returnval){ [ 712.269661] env[65726]: value = "task-5115681" [ 712.269661] env[65726]: _type = "Task" [ 712.269661] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 712.280459] env[65726]: DEBUG oslo_vmware.api [None req-1023a64a-e0ea-45c1-b4cf-b0858cc7b3e9 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Task: {'id': task-5115681, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 712.314109] env[65726]: WARNING openstack [req-abe14589-e2ed-4063-a1f6-da86e9a6123a req-d55c367e-0b7d-43e3-907c-06f37f06725a service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 712.314548] env[65726]: WARNING openstack [req-abe14589-e2ed-4063-a1f6-da86e9a6123a req-d55c367e-0b7d-43e3-907c-06f37f06725a service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 712.347148] env[65726]: DEBUG nova.compute.manager [None req-681fb0ed-e41c-4d7e-b603-04df1dff2263 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] [instance: 04ffdf80-c54c-4587-9bb4-d520dc440501] Start destroying the instance on the hypervisor. {{(pid=65726) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 712.347148] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-681fb0ed-e41c-4d7e-b603-04df1dff2263 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] [instance: 04ffdf80-c54c-4587-9bb4-d520dc440501] Destroying instance {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 712.348461] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d97b5faf-a8cd-4de2-9b2a-8cb9a610cecc {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.368387] env[65726]: DEBUG oslo_vmware.api [None req-f3171f49-8536-4923-b5cf-31399a616a07 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52e3c87a-8a95-0cbc-a192-32a3aadd8599, 'name': SearchDatastore_Task, 'duration_secs': 0.011062} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 712.379101] env[65726]: DEBUG oslo_concurrency.lockutils [None req-f3171f49-8536-4923-b5cf-31399a616a07 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 712.379620] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-f3171f49-8536-4923-b5cf-31399a616a07 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] 608cc0ea-3f6b-4b4f-83c1-01aa50999d63/608cc0ea-3f6b-4b4f-83c1-01aa50999d63.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 712.380134] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-681fb0ed-e41c-4d7e-b603-04df1dff2263 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] [instance: 04ffdf80-c54c-4587-9bb4-d520dc440501] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 712.381183] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9b6ff86c-dbe8-43cb-96d4-f1bee9c06ccd {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.384698] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c3bd48f5-f5c7-462a-8c56-9584e94ba21f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.400948] env[65726]: DEBUG oslo_vmware.api [None req-c00c2da9-ebd0-4704-a276-dac36258bf53 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Task: {'id': task-5115679, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.14151} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 712.405212] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-c00c2da9-ebd0-4704-a276-dac36258bf53 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Deleted the datastore file {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 712.406237] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-c00c2da9-ebd0-4704-a276-dac36258bf53 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] [instance: f5151062-57b1-4e4e-93f4-aab0e7f504d4] Deleted contents of the VM from datastore datastore2 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 712.406237] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-c00c2da9-ebd0-4704-a276-dac36258bf53 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] [instance: f5151062-57b1-4e4e-93f4-aab0e7f504d4] Instance destroyed {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 712.406237] env[65726]: INFO nova.compute.manager [None req-c00c2da9-ebd0-4704-a276-dac36258bf53 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] [instance: f5151062-57b1-4e4e-93f4-aab0e7f504d4] Took 1.16 seconds to destroy the instance on the hypervisor. [ 712.406485] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-c00c2da9-ebd0-4704-a276-dac36258bf53 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 712.406914] env[65726]: DEBUG oslo_vmware.api [None req-681fb0ed-e41c-4d7e-b603-04df1dff2263 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Waiting for the task: (returnval){ [ 712.406914] env[65726]: value = "task-5115683" [ 712.406914] env[65726]: _type = "Task" [ 712.406914] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 712.408540] env[65726]: DEBUG oslo_vmware.api [None req-f3171f49-8536-4923-b5cf-31399a616a07 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Waiting for the task: (returnval){ [ 712.408540] env[65726]: value = "task-5115682" [ 712.408540] env[65726]: _type = "Task" [ 712.408540] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 712.408540] env[65726]: DEBUG nova.compute.manager [-] [instance: f5151062-57b1-4e4e-93f4-aab0e7f504d4] Deallocating network for instance {{(pid=65726) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 712.408902] env[65726]: DEBUG nova.network.neutron [-] [instance: f5151062-57b1-4e4e-93f4-aab0e7f504d4] deallocate_for_instance() {{(pid=65726) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 712.409531] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 712.410381] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 712.432943] env[65726]: DEBUG oslo_vmware.api [None req-681fb0ed-e41c-4d7e-b603-04df1dff2263 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Task: {'id': task-5115683, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 712.436843] env[65726]: WARNING openstack [req-abe14589-e2ed-4063-a1f6-da86e9a6123a req-d55c367e-0b7d-43e3-907c-06f37f06725a service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 712.437128] env[65726]: WARNING openstack [req-abe14589-e2ed-4063-a1f6-da86e9a6123a req-d55c367e-0b7d-43e3-907c-06f37f06725a service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 712.449318] env[65726]: DEBUG oslo_vmware.api [None req-f3171f49-8536-4923-b5cf-31399a616a07 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5115682, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 712.485769] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 712.486020] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 712.608820] env[65726]: DEBUG oslo_vmware.api [None req-5dd7a34f-2002-4a12-9099-b0b630d2774d tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Task: {'id': task-5115680, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.126475} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 712.608977] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-5dd7a34f-2002-4a12-9099-b0b630d2774d tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] [instance: 815f056b-2d88-40dd-9ef7-a891ec8d92cf] Extended root virtual disk {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 712.610548] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ccccf16-9f58-4da2-88ac-bcaa03e85992 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.635803] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-5dd7a34f-2002-4a12-9099-b0b630d2774d tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] [instance: 815f056b-2d88-40dd-9ef7-a891ec8d92cf] Reconfiguring VM instance instance-0000001e to attach disk [datastore1] 815f056b-2d88-40dd-9ef7-a891ec8d92cf/815f056b-2d88-40dd-9ef7-a891ec8d92cf.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 712.638941] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ea8a265c-9d07-4adb-84d3-85e2c00c84cd {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.662038] env[65726]: DEBUG oslo_vmware.api [None req-5dd7a34f-2002-4a12-9099-b0b630d2774d tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Waiting for the task: (returnval){ [ 712.662038] env[65726]: value = "task-5115684" [ 712.662038] env[65726]: _type = "Task" [ 712.662038] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 712.676052] env[65726]: DEBUG oslo_vmware.api [None req-5dd7a34f-2002-4a12-9099-b0b630d2774d tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Task: {'id': task-5115684, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 712.680030] env[65726]: DEBUG nova.network.neutron [req-abe14589-e2ed-4063-a1f6-da86e9a6123a req-d55c367e-0b7d-43e3-907c-06f37f06725a service nova] [instance: 608cc0ea-3f6b-4b4f-83c1-01aa50999d63] Updated VIF entry in instance network info cache for port e9f7132a-8d5a-47fe-afc6-d45bf3687ea4. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 712.680553] env[65726]: DEBUG nova.network.neutron [req-abe14589-e2ed-4063-a1f6-da86e9a6123a req-d55c367e-0b7d-43e3-907c-06f37f06725a service nova] [instance: 608cc0ea-3f6b-4b4f-83c1-01aa50999d63] Updating instance_info_cache with network_info: [{"id": "e9f7132a-8d5a-47fe-afc6-d45bf3687ea4", "address": "fa:16:3e:fe:4d:df", "network": {"id": "0d0f561d-ac41-4ef5-9926-c707409799f7", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1134321217-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4eece77569624f90bf64e5c51974173f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2ce62383-8e84-4e26-955b-74c11392f4c9", "external-id": "nsx-vlan-transportzone-215", "segmentation_id": 215, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape9f7132a-8d", "ovs_interfaceid": "e9f7132a-8d5a-47fe-afc6-d45bf3687ea4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 712.712483] env[65726]: DEBUG nova.scheduler.client.report [None req-296fd45e-ee7a-4e44-912a-f3b5a22cbbf0 tempest-ServerExternalEventsTest-1845125219 tempest-ServerExternalEventsTest-1845125219-project-member] Updated inventory for provider 07c4692f-bdb4-4058-9173-ff9664830295 with generation 58 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 712.712636] env[65726]: DEBUG nova.compute.provider_tree [None req-296fd45e-ee7a-4e44-912a-f3b5a22cbbf0 tempest-ServerExternalEventsTest-1845125219 tempest-ServerExternalEventsTest-1845125219-project-member] Updating resource provider 07c4692f-bdb4-4058-9173-ff9664830295 generation from 58 to 59 during operation: update_inventory {{(pid=65726) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 712.712811] env[65726]: DEBUG nova.compute.provider_tree [None req-296fd45e-ee7a-4e44-912a-f3b5a22cbbf0 tempest-ServerExternalEventsTest-1845125219 tempest-ServerExternalEventsTest-1845125219-project-member] Updating inventory in ProviderTree for provider 07c4692f-bdb4-4058-9173-ff9664830295 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 712.782975] env[65726]: DEBUG oslo_vmware.api [None req-1023a64a-e0ea-45c1-b4cf-b0858cc7b3e9 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Task: {'id': task-5115681, 'name': PowerOffVM_Task, 'duration_secs': 0.213722} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 712.782975] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-1023a64a-e0ea-45c1-b4cf-b0858cc7b3e9 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] [instance: 3a60b719-9db7-436e-9908-25f6a233c465] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 712.782975] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-1023a64a-e0ea-45c1-b4cf-b0858cc7b3e9 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] [instance: 3a60b719-9db7-436e-9908-25f6a233c465] Unregistering the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 712.782975] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d66fdb8b-2277-4dcf-b52d-aaeb64adbbe2 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.858694] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-1023a64a-e0ea-45c1-b4cf-b0858cc7b3e9 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] [instance: 3a60b719-9db7-436e-9908-25f6a233c465] Unregistered the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 712.859178] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-1023a64a-e0ea-45c1-b4cf-b0858cc7b3e9 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] [instance: 3a60b719-9db7-436e-9908-25f6a233c465] Deleting contents of the VM from datastore datastore2 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 712.859480] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-1023a64a-e0ea-45c1-b4cf-b0858cc7b3e9 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Deleting the datastore file [datastore2] 3a60b719-9db7-436e-9908-25f6a233c465 {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 712.859857] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9e65f347-ad07-45c5-9598-18a44bdc515d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.869114] env[65726]: DEBUG oslo_vmware.api [None req-1023a64a-e0ea-45c1-b4cf-b0858cc7b3e9 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Waiting for the task: (returnval){ [ 712.869114] env[65726]: value = "task-5115686" [ 712.869114] env[65726]: _type = "Task" [ 712.869114] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 712.881076] env[65726]: DEBUG oslo_vmware.api [None req-1023a64a-e0ea-45c1-b4cf-b0858cc7b3e9 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Task: {'id': task-5115686, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 712.934759] env[65726]: DEBUG oslo_vmware.api [None req-681fb0ed-e41c-4d7e-b603-04df1dff2263 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Task: {'id': task-5115683, 'name': PowerOffVM_Task, 'duration_secs': 0.220049} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 712.938759] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-681fb0ed-e41c-4d7e-b603-04df1dff2263 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] [instance: 04ffdf80-c54c-4587-9bb4-d520dc440501] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 712.938759] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-681fb0ed-e41c-4d7e-b603-04df1dff2263 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] [instance: 04ffdf80-c54c-4587-9bb4-d520dc440501] Unregistering the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 712.940859] env[65726]: DEBUG oslo_vmware.api [None req-f3171f49-8536-4923-b5cf-31399a616a07 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5115682, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 712.940859] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3f239e95-f531-45f8-92dc-fa549cc2be6d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.002935] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-681fb0ed-e41c-4d7e-b603-04df1dff2263 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] [instance: 04ffdf80-c54c-4587-9bb4-d520dc440501] Unregistered the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 713.003183] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-681fb0ed-e41c-4d7e-b603-04df1dff2263 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] [instance: 04ffdf80-c54c-4587-9bb4-d520dc440501] Deleting contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 713.003470] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-681fb0ed-e41c-4d7e-b603-04df1dff2263 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Deleting the datastore file [datastore1] 04ffdf80-c54c-4587-9bb4-d520dc440501 {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 713.003713] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d557936f-8eda-428f-bba4-cadbc8f55996 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.013550] env[65726]: DEBUG oslo_vmware.api [None req-681fb0ed-e41c-4d7e-b603-04df1dff2263 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Waiting for the task: (returnval){ [ 713.013550] env[65726]: value = "task-5115688" [ 713.013550] env[65726]: _type = "Task" [ 713.013550] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 713.023827] env[65726]: DEBUG oslo_vmware.api [None req-681fb0ed-e41c-4d7e-b603-04df1dff2263 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Task: {'id': task-5115688, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 713.177989] env[65726]: DEBUG oslo_vmware.api [None req-5dd7a34f-2002-4a12-9099-b0b630d2774d tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Task: {'id': task-5115684, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 713.184217] env[65726]: DEBUG oslo_concurrency.lockutils [req-abe14589-e2ed-4063-a1f6-da86e9a6123a req-d55c367e-0b7d-43e3-907c-06f37f06725a service nova] Releasing lock "refresh_cache-608cc0ea-3f6b-4b4f-83c1-01aa50999d63" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 713.224096] env[65726]: DEBUG oslo_concurrency.lockutils [None req-296fd45e-ee7a-4e44-912a-f3b5a22cbbf0 tempest-ServerExternalEventsTest-1845125219 tempest-ServerExternalEventsTest-1845125219-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.888s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 713.224096] env[65726]: DEBUG nova.compute.manager [None req-296fd45e-ee7a-4e44-912a-f3b5a22cbbf0 tempest-ServerExternalEventsTest-1845125219 tempest-ServerExternalEventsTest-1845125219-project-member] [instance: 96acc5f8-b411-408e-929d-d1035cfd50db] Start building networks asynchronously for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 713.224872] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3d632edf-0ed8-46ec-b4fd-7092ea7f050e tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 35.012s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 713.224872] env[65726]: DEBUG nova.objects.instance [None req-3d632edf-0ed8-46ec-b4fd-7092ea7f050e tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] [instance: 3b933684-ac19-44b0-a49d-6af45501e166] Trying to apply a migration context that does not seem to be set for this instance {{(pid=65726) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 713.266198] env[65726]: DEBUG nova.network.neutron [-] [instance: f5151062-57b1-4e4e-93f4-aab0e7f504d4] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 713.380945] env[65726]: DEBUG oslo_vmware.api [None req-1023a64a-e0ea-45c1-b4cf-b0858cc7b3e9 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Task: {'id': task-5115686, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.256185} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 713.381206] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-1023a64a-e0ea-45c1-b4cf-b0858cc7b3e9 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Deleted the datastore file {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 713.381387] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-1023a64a-e0ea-45c1-b4cf-b0858cc7b3e9 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] [instance: 3a60b719-9db7-436e-9908-25f6a233c465] Deleted contents of the VM from datastore datastore2 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 713.381586] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-1023a64a-e0ea-45c1-b4cf-b0858cc7b3e9 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] [instance: 3a60b719-9db7-436e-9908-25f6a233c465] Instance destroyed {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 713.381768] env[65726]: INFO nova.compute.manager [None req-1023a64a-e0ea-45c1-b4cf-b0858cc7b3e9 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] [instance: 3a60b719-9db7-436e-9908-25f6a233c465] Took 1.13 seconds to destroy the instance on the hypervisor. [ 713.382033] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-1023a64a-e0ea-45c1-b4cf-b0858cc7b3e9 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 713.382241] env[65726]: DEBUG nova.compute.manager [-] [instance: 3a60b719-9db7-436e-9908-25f6a233c465] Deallocating network for instance {{(pid=65726) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 713.382343] env[65726]: DEBUG nova.network.neutron [-] [instance: 3a60b719-9db7-436e-9908-25f6a233c465] deallocate_for_instance() {{(pid=65726) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 713.383451] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 713.383451] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 713.432476] env[65726]: DEBUG oslo_vmware.api [None req-f3171f49-8536-4923-b5cf-31399a616a07 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5115682, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.560758} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 713.432793] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-f3171f49-8536-4923-b5cf-31399a616a07 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] 608cc0ea-3f6b-4b4f-83c1-01aa50999d63/608cc0ea-3f6b-4b4f-83c1-01aa50999d63.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 713.433018] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-f3171f49-8536-4923-b5cf-31399a616a07 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 608cc0ea-3f6b-4b4f-83c1-01aa50999d63] Extending root virtual disk to 1048576 {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 713.433823] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5329c421-0f41-4578-ac36-dae63df5d82a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.442170] env[65726]: DEBUG oslo_vmware.api [None req-f3171f49-8536-4923-b5cf-31399a616a07 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Waiting for the task: (returnval){ [ 713.442170] env[65726]: value = "task-5115689" [ 713.442170] env[65726]: _type = "Task" [ 713.442170] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 713.452999] env[65726]: DEBUG oslo_vmware.api [None req-f3171f49-8536-4923-b5cf-31399a616a07 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5115689, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 713.465217] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 713.465217] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 713.529251] env[65726]: DEBUG oslo_vmware.api [None req-681fb0ed-e41c-4d7e-b603-04df1dff2263 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Task: {'id': task-5115688, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.175745} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 713.529251] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-681fb0ed-e41c-4d7e-b603-04df1dff2263 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Deleted the datastore file {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 713.529251] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-681fb0ed-e41c-4d7e-b603-04df1dff2263 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] [instance: 04ffdf80-c54c-4587-9bb4-d520dc440501] Deleted contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 713.529251] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-681fb0ed-e41c-4d7e-b603-04df1dff2263 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] [instance: 04ffdf80-c54c-4587-9bb4-d520dc440501] Instance destroyed {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 713.529251] env[65726]: INFO nova.compute.manager [None req-681fb0ed-e41c-4d7e-b603-04df1dff2263 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] [instance: 04ffdf80-c54c-4587-9bb4-d520dc440501] Took 1.18 seconds to destroy the instance on the hypervisor. [ 713.529615] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-681fb0ed-e41c-4d7e-b603-04df1dff2263 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 713.529924] env[65726]: DEBUG nova.compute.manager [-] [instance: 04ffdf80-c54c-4587-9bb4-d520dc440501] Deallocating network for instance {{(pid=65726) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 713.530139] env[65726]: DEBUG nova.network.neutron [-] [instance: 04ffdf80-c54c-4587-9bb4-d520dc440501] deallocate_for_instance() {{(pid=65726) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 713.530825] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 713.532580] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 713.677440] env[65726]: DEBUG oslo_vmware.api [None req-5dd7a34f-2002-4a12-9099-b0b630d2774d tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Task: {'id': task-5115684, 'name': ReconfigVM_Task, 'duration_secs': 0.556908} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 713.679571] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-5dd7a34f-2002-4a12-9099-b0b630d2774d tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] [instance: 815f056b-2d88-40dd-9ef7-a891ec8d92cf] Reconfigured VM instance instance-0000001e to attach disk [datastore1] 815f056b-2d88-40dd-9ef7-a891ec8d92cf/815f056b-2d88-40dd-9ef7-a891ec8d92cf.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 713.682103] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0879964f-cbfd-41e5-8ed4-261140c23f5a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.694039] env[65726]: DEBUG oslo_vmware.api [None req-5dd7a34f-2002-4a12-9099-b0b630d2774d tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Waiting for the task: (returnval){ [ 713.694039] env[65726]: value = "task-5115690" [ 713.694039] env[65726]: _type = "Task" [ 713.694039] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 713.703475] env[65726]: DEBUG oslo_vmware.api [None req-5dd7a34f-2002-4a12-9099-b0b630d2774d tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Task: {'id': task-5115690, 'name': Rename_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 713.731196] env[65726]: DEBUG nova.compute.utils [None req-296fd45e-ee7a-4e44-912a-f3b5a22cbbf0 tempest-ServerExternalEventsTest-1845125219 tempest-ServerExternalEventsTest-1845125219-project-member] Using /dev/sd instead of None {{(pid=65726) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 713.739155] env[65726]: DEBUG nova.compute.manager [None req-296fd45e-ee7a-4e44-912a-f3b5a22cbbf0 tempest-ServerExternalEventsTest-1845125219 tempest-ServerExternalEventsTest-1845125219-project-member] [instance: 96acc5f8-b411-408e-929d-d1035cfd50db] Allocating IP information in the background. {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 713.739815] env[65726]: DEBUG nova.network.neutron [None req-296fd45e-ee7a-4e44-912a-f3b5a22cbbf0 tempest-ServerExternalEventsTest-1845125219 tempest-ServerExternalEventsTest-1845125219-project-member] [instance: 96acc5f8-b411-408e-929d-d1035cfd50db] allocate_for_instance() {{(pid=65726) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 713.740389] env[65726]: WARNING neutronclient.v2_0.client [None req-296fd45e-ee7a-4e44-912a-f3b5a22cbbf0 tempest-ServerExternalEventsTest-1845125219 tempest-ServerExternalEventsTest-1845125219-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 713.741205] env[65726]: WARNING neutronclient.v2_0.client [None req-296fd45e-ee7a-4e44-912a-f3b5a22cbbf0 tempest-ServerExternalEventsTest-1845125219 tempest-ServerExternalEventsTest-1845125219-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 713.741936] env[65726]: WARNING openstack [None req-296fd45e-ee7a-4e44-912a-f3b5a22cbbf0 tempest-ServerExternalEventsTest-1845125219 tempest-ServerExternalEventsTest-1845125219-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 713.742502] env[65726]: WARNING openstack [None req-296fd45e-ee7a-4e44-912a-f3b5a22cbbf0 tempest-ServerExternalEventsTest-1845125219 tempest-ServerExternalEventsTest-1845125219-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 713.753639] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3d632edf-0ed8-46ec-b4fd-7092ea7f050e tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 0.529s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 713.755478] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c24265b5-a0c6-4266-b4c0-64535a550ddd tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 33.411s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 713.760200] env[65726]: INFO nova.compute.claims [None req-c24265b5-a0c6-4266-b4c0-64535a550ddd tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: bd839fe1-8801-4ba1-9c23-288453258b66] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 713.769711] env[65726]: INFO nova.compute.manager [-] [instance: f5151062-57b1-4e4e-93f4-aab0e7f504d4] Took 1.36 seconds to deallocate network for instance. [ 713.800755] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 713.801888] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 713.957406] env[65726]: DEBUG oslo_vmware.api [None req-f3171f49-8536-4923-b5cf-31399a616a07 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5115689, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.363666} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 713.957848] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-f3171f49-8536-4923-b5cf-31399a616a07 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 608cc0ea-3f6b-4b4f-83c1-01aa50999d63] Extended root virtual disk {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 713.958576] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a223fcac-da3e-43b9-ba9f-9dabf7aa53f8 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.993808] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-f3171f49-8536-4923-b5cf-31399a616a07 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 608cc0ea-3f6b-4b4f-83c1-01aa50999d63] Reconfiguring VM instance instance-0000001f to attach disk [datastore1] 608cc0ea-3f6b-4b4f-83c1-01aa50999d63/608cc0ea-3f6b-4b4f-83c1-01aa50999d63.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 713.994689] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-59d9cddd-247c-4cc7-be59-f75aeac76b2e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.017310] env[65726]: DEBUG oslo_vmware.api [None req-f3171f49-8536-4923-b5cf-31399a616a07 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Waiting for the task: (returnval){ [ 714.017310] env[65726]: value = "task-5115691" [ 714.017310] env[65726]: _type = "Task" [ 714.017310] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 714.030998] env[65726]: DEBUG oslo_vmware.api [None req-f3171f49-8536-4923-b5cf-31399a616a07 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5115691, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 714.176272] env[65726]: DEBUG nova.policy [None req-296fd45e-ee7a-4e44-912a-f3b5a22cbbf0 tempest-ServerExternalEventsTest-1845125219 tempest-ServerExternalEventsTest-1845125219-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '350d7f349eff4bcebe467567649678e0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'fddc635807d54dc0a5b4f2f732ba6962', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65726) authorize /opt/stack/nova/nova/policy.py:192}} [ 714.214636] env[65726]: DEBUG oslo_vmware.api [None req-5dd7a34f-2002-4a12-9099-b0b630d2774d tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Task: {'id': task-5115690, 'name': Rename_Task, 'duration_secs': 0.450814} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 714.215607] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-5dd7a34f-2002-4a12-9099-b0b630d2774d tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] [instance: 815f056b-2d88-40dd-9ef7-a891ec8d92cf] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 714.219799] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-50bcb597-f313-401c-8298-5f0ec88b9eb1 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.226950] env[65726]: DEBUG oslo_vmware.api [None req-5dd7a34f-2002-4a12-9099-b0b630d2774d tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Waiting for the task: (returnval){ [ 714.226950] env[65726]: value = "task-5115692" [ 714.226950] env[65726]: _type = "Task" [ 714.226950] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 714.236435] env[65726]: DEBUG nova.compute.manager [None req-296fd45e-ee7a-4e44-912a-f3b5a22cbbf0 tempest-ServerExternalEventsTest-1845125219 tempest-ServerExternalEventsTest-1845125219-project-member] [instance: 96acc5f8-b411-408e-929d-d1035cfd50db] Start building block device mappings for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 714.249057] env[65726]: DEBUG oslo_vmware.api [None req-5dd7a34f-2002-4a12-9099-b0b630d2774d tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Task: {'id': task-5115692, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 714.252261] env[65726]: DEBUG nova.network.neutron [-] [instance: 3a60b719-9db7-436e-9908-25f6a233c465] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 714.279098] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c00c2da9-ebd0-4704-a276-dac36258bf53 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 714.439058] env[65726]: DEBUG nova.compute.manager [req-85c3ade2-71c4-45bf-9058-7486b76bee2f req-4e7080ce-b532-44af-bcce-4dfd03dd5e9f service nova] [instance: f5151062-57b1-4e4e-93f4-aab0e7f504d4] Received event network-vif-deleted-83750abb-1d26-4e23-a675-61ef785b858b {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 714.530140] env[65726]: DEBUG oslo_vmware.api [None req-f3171f49-8536-4923-b5cf-31399a616a07 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5115691, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 714.583855] env[65726]: DEBUG nova.network.neutron [-] [instance: 04ffdf80-c54c-4587-9bb4-d520dc440501] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 714.631363] env[65726]: DEBUG nova.network.neutron [None req-296fd45e-ee7a-4e44-912a-f3b5a22cbbf0 tempest-ServerExternalEventsTest-1845125219 tempest-ServerExternalEventsTest-1845125219-project-member] [instance: 96acc5f8-b411-408e-929d-d1035cfd50db] Successfully created port: c90c9d5c-3ed3-4f46-bd35-88e6fd91c396 {{(pid=65726) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 714.741476] env[65726]: DEBUG oslo_vmware.api [None req-5dd7a34f-2002-4a12-9099-b0b630d2774d tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Task: {'id': task-5115692, 'name': PowerOnVM_Task} progress is 88%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 714.758082] env[65726]: INFO nova.compute.manager [-] [instance: 3a60b719-9db7-436e-9908-25f6a233c465] Took 1.38 seconds to deallocate network for instance. [ 715.029914] env[65726]: DEBUG oslo_vmware.api [None req-f3171f49-8536-4923-b5cf-31399a616a07 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5115691, 'name': ReconfigVM_Task, 'duration_secs': 0.551665} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 715.033077] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-f3171f49-8536-4923-b5cf-31399a616a07 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 608cc0ea-3f6b-4b4f-83c1-01aa50999d63] Reconfigured VM instance instance-0000001f to attach disk [datastore1] 608cc0ea-3f6b-4b4f-83c1-01aa50999d63/608cc0ea-3f6b-4b4f-83c1-01aa50999d63.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 715.036626] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2081a06a-4ba0-47f3-b75f-a44287cf0e23 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.042265] env[65726]: DEBUG oslo_vmware.api [None req-f3171f49-8536-4923-b5cf-31399a616a07 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Waiting for the task: (returnval){ [ 715.042265] env[65726]: value = "task-5115693" [ 715.042265] env[65726]: _type = "Task" [ 715.042265] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 715.060228] env[65726]: DEBUG oslo_vmware.api [None req-f3171f49-8536-4923-b5cf-31399a616a07 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5115693, 'name': Rename_Task} progress is 6%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 715.090495] env[65726]: INFO nova.compute.manager [-] [instance: 04ffdf80-c54c-4587-9bb4-d520dc440501] Took 1.56 seconds to deallocate network for instance. [ 715.240798] env[65726]: DEBUG oslo_vmware.api [None req-5dd7a34f-2002-4a12-9099-b0b630d2774d tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Task: {'id': task-5115692, 'name': PowerOnVM_Task, 'duration_secs': 0.753394} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 715.240798] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-5dd7a34f-2002-4a12-9099-b0b630d2774d tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] [instance: 815f056b-2d88-40dd-9ef7-a891ec8d92cf] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 715.240798] env[65726]: INFO nova.compute.manager [None req-5dd7a34f-2002-4a12-9099-b0b630d2774d tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] [instance: 815f056b-2d88-40dd-9ef7-a891ec8d92cf] Took 9.20 seconds to spawn the instance on the hypervisor. [ 715.240798] env[65726]: DEBUG nova.compute.manager [None req-5dd7a34f-2002-4a12-9099-b0b630d2774d tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] [instance: 815f056b-2d88-40dd-9ef7-a891ec8d92cf] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 715.240798] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72985bec-a60e-48d6-870b-2d4e9bf0c319 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.260163] env[65726]: DEBUG nova.compute.manager [None req-296fd45e-ee7a-4e44-912a-f3b5a22cbbf0 tempest-ServerExternalEventsTest-1845125219 tempest-ServerExternalEventsTest-1845125219-project-member] [instance: 96acc5f8-b411-408e-929d-d1035cfd50db] Start spawning the instance on the hypervisor. {{(pid=65726) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 715.269535] env[65726]: DEBUG oslo_concurrency.lockutils [None req-1023a64a-e0ea-45c1-b4cf-b0858cc7b3e9 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 715.296722] env[65726]: DEBUG nova.virt.hardware [None req-296fd45e-ee7a-4e44-912a-f3b5a22cbbf0 tempest-ServerExternalEventsTest-1845125219 tempest-ServerExternalEventsTest-1845125219-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T19:28:34Z,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1539f06fa3534e90acbc3a60c4b8bd3f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T19:28:35Z,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 715.297469] env[65726]: DEBUG nova.virt.hardware [None req-296fd45e-ee7a-4e44-912a-f3b5a22cbbf0 tempest-ServerExternalEventsTest-1845125219 tempest-ServerExternalEventsTest-1845125219-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 715.297682] env[65726]: DEBUG nova.virt.hardware [None req-296fd45e-ee7a-4e44-912a-f3b5a22cbbf0 tempest-ServerExternalEventsTest-1845125219 tempest-ServerExternalEventsTest-1845125219-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 715.298183] env[65726]: DEBUG nova.virt.hardware [None req-296fd45e-ee7a-4e44-912a-f3b5a22cbbf0 tempest-ServerExternalEventsTest-1845125219 tempest-ServerExternalEventsTest-1845125219-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 715.298254] env[65726]: DEBUG nova.virt.hardware [None req-296fd45e-ee7a-4e44-912a-f3b5a22cbbf0 tempest-ServerExternalEventsTest-1845125219 tempest-ServerExternalEventsTest-1845125219-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 715.298886] env[65726]: DEBUG nova.virt.hardware [None req-296fd45e-ee7a-4e44-912a-f3b5a22cbbf0 tempest-ServerExternalEventsTest-1845125219 tempest-ServerExternalEventsTest-1845125219-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 715.299233] env[65726]: DEBUG nova.virt.hardware [None req-296fd45e-ee7a-4e44-912a-f3b5a22cbbf0 tempest-ServerExternalEventsTest-1845125219 tempest-ServerExternalEventsTest-1845125219-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 715.299525] env[65726]: DEBUG nova.virt.hardware [None req-296fd45e-ee7a-4e44-912a-f3b5a22cbbf0 tempest-ServerExternalEventsTest-1845125219 tempest-ServerExternalEventsTest-1845125219-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 715.299761] env[65726]: DEBUG nova.virt.hardware [None req-296fd45e-ee7a-4e44-912a-f3b5a22cbbf0 tempest-ServerExternalEventsTest-1845125219 tempest-ServerExternalEventsTest-1845125219-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 715.300044] env[65726]: DEBUG nova.virt.hardware [None req-296fd45e-ee7a-4e44-912a-f3b5a22cbbf0 tempest-ServerExternalEventsTest-1845125219 tempest-ServerExternalEventsTest-1845125219-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 715.300298] env[65726]: DEBUG nova.virt.hardware [None req-296fd45e-ee7a-4e44-912a-f3b5a22cbbf0 tempest-ServerExternalEventsTest-1845125219 tempest-ServerExternalEventsTest-1845125219-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 715.301892] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98fd7600-dcc1-448d-bb96-c80329c8d475 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.315029] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f9e6b17-b745-4d1f-8d03-efa0e158caa0 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.455357] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9d255a4-c02d-4c2b-9011-ac8172868446 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.463942] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a479103-d793-4f7b-96b6-883133321d66 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.497736] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8cb5c6d-c1c4-4a2d-ba19-9a7a65d9317f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.506566] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e21a06da-4c7f-4201-887e-21ef7c7b5fe7 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.520990] env[65726]: DEBUG nova.compute.provider_tree [None req-c24265b5-a0c6-4266-b4c0-64535a550ddd tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Updating inventory in ProviderTree for provider 07c4692f-bdb4-4058-9173-ff9664830295 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 715.553590] env[65726]: DEBUG oslo_vmware.api [None req-f3171f49-8536-4923-b5cf-31399a616a07 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5115693, 'name': Rename_Task, 'duration_secs': 0.187755} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 715.553847] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-f3171f49-8536-4923-b5cf-31399a616a07 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 608cc0ea-3f6b-4b4f-83c1-01aa50999d63] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 715.554129] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-48f292e3-037a-40a9-abbb-64e57b176aad {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.561627] env[65726]: DEBUG oslo_vmware.api [None req-f3171f49-8536-4923-b5cf-31399a616a07 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Waiting for the task: (returnval){ [ 715.561627] env[65726]: value = "task-5115694" [ 715.561627] env[65726]: _type = "Task" [ 715.561627] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 715.570847] env[65726]: DEBUG oslo_vmware.api [None req-f3171f49-8536-4923-b5cf-31399a616a07 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5115694, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 715.602422] env[65726]: DEBUG oslo_concurrency.lockutils [None req-681fb0ed-e41c-4d7e-b603-04df1dff2263 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 715.766015] env[65726]: INFO nova.compute.manager [None req-5dd7a34f-2002-4a12-9099-b0b630d2774d tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] [instance: 815f056b-2d88-40dd-9ef7-a891ec8d92cf] Took 42.76 seconds to build instance. [ 716.070574] env[65726]: DEBUG nova.scheduler.client.report [None req-c24265b5-a0c6-4266-b4c0-64535a550ddd tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Updated inventory for provider 07c4692f-bdb4-4058-9173-ff9664830295 with generation 59 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 716.071086] env[65726]: DEBUG nova.compute.provider_tree [None req-c24265b5-a0c6-4266-b4c0-64535a550ddd tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Updating resource provider 07c4692f-bdb4-4058-9173-ff9664830295 generation from 59 to 60 during operation: update_inventory {{(pid=65726) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 716.071086] env[65726]: DEBUG nova.compute.provider_tree [None req-c24265b5-a0c6-4266-b4c0-64535a550ddd tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Updating inventory in ProviderTree for provider 07c4692f-bdb4-4058-9173-ff9664830295 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 716.086628] env[65726]: DEBUG oslo_vmware.api [None req-f3171f49-8536-4923-b5cf-31399a616a07 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5115694, 'name': PowerOnVM_Task, 'duration_secs': 0.487749} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 716.086959] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-f3171f49-8536-4923-b5cf-31399a616a07 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 608cc0ea-3f6b-4b4f-83c1-01aa50999d63] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 716.087250] env[65726]: INFO nova.compute.manager [None req-f3171f49-8536-4923-b5cf-31399a616a07 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 608cc0ea-3f6b-4b4f-83c1-01aa50999d63] Took 7.27 seconds to spawn the instance on the hypervisor. [ 716.087451] env[65726]: DEBUG nova.compute.manager [None req-f3171f49-8536-4923-b5cf-31399a616a07 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 608cc0ea-3f6b-4b4f-83c1-01aa50999d63] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 716.088685] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ceb32760-9364-472f-b183-226e9fed0098 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.267974] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5dd7a34f-2002-4a12-9099-b0b630d2774d tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Lock "815f056b-2d88-40dd-9ef7-a891ec8d92cf" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 70.693s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 716.325698] env[65726]: DEBUG nova.network.neutron [None req-296fd45e-ee7a-4e44-912a-f3b5a22cbbf0 tempest-ServerExternalEventsTest-1845125219 tempest-ServerExternalEventsTest-1845125219-project-member] [instance: 96acc5f8-b411-408e-929d-d1035cfd50db] Successfully updated port: c90c9d5c-3ed3-4f46-bd35-88e6fd91c396 {{(pid=65726) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 716.581357] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c24265b5-a0c6-4266-b4c0-64535a550ddd tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.826s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 716.582899] env[65726]: DEBUG nova.compute.manager [None req-c24265b5-a0c6-4266-b4c0-64535a550ddd tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: bd839fe1-8801-4ba1-9c23-288453258b66] Start building networks asynchronously for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 716.586844] env[65726]: DEBUG oslo_concurrency.lockutils [None req-2737dc3f-715d-481c-b9a9-998472382a0d tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 35.159s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 716.588886] env[65726]: INFO nova.compute.claims [None req-2737dc3f-715d-481c-b9a9-998472382a0d tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] [instance: 83573c2b-9448-456f-8fd2-b19661dd6cc4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 716.623173] env[65726]: INFO nova.compute.manager [None req-f3171f49-8536-4923-b5cf-31399a616a07 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 608cc0ea-3f6b-4b4f-83c1-01aa50999d63] Took 42.01 seconds to build instance. [ 716.771168] env[65726]: DEBUG nova.compute.manager [None req-44e83fc7-afcc-4134-bbbb-6bc576931bb8 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: 895797e4-2941-44cd-aab1-67afa4fac02d] Starting instance... {{(pid=65726) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 716.829020] env[65726]: DEBUG oslo_concurrency.lockutils [None req-296fd45e-ee7a-4e44-912a-f3b5a22cbbf0 tempest-ServerExternalEventsTest-1845125219 tempest-ServerExternalEventsTest-1845125219-project-member] Acquiring lock "refresh_cache-96acc5f8-b411-408e-929d-d1035cfd50db" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 716.829286] env[65726]: DEBUG oslo_concurrency.lockutils [None req-296fd45e-ee7a-4e44-912a-f3b5a22cbbf0 tempest-ServerExternalEventsTest-1845125219 tempest-ServerExternalEventsTest-1845125219-project-member] Acquired lock "refresh_cache-96acc5f8-b411-408e-929d-d1035cfd50db" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 716.829650] env[65726]: DEBUG nova.network.neutron [None req-296fd45e-ee7a-4e44-912a-f3b5a22cbbf0 tempest-ServerExternalEventsTest-1845125219 tempest-ServerExternalEventsTest-1845125219-project-member] [instance: 96acc5f8-b411-408e-929d-d1035cfd50db] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 717.094250] env[65726]: DEBUG nova.compute.utils [None req-c24265b5-a0c6-4266-b4c0-64535a550ddd tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Using /dev/sd instead of None {{(pid=65726) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 717.104182] env[65726]: DEBUG nova.compute.manager [None req-c24265b5-a0c6-4266-b4c0-64535a550ddd tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: bd839fe1-8801-4ba1-9c23-288453258b66] Allocating IP information in the background. {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 717.104290] env[65726]: DEBUG nova.network.neutron [None req-c24265b5-a0c6-4266-b4c0-64535a550ddd tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: bd839fe1-8801-4ba1-9c23-288453258b66] allocate_for_instance() {{(pid=65726) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 717.104990] env[65726]: WARNING neutronclient.v2_0.client [None req-c24265b5-a0c6-4266-b4c0-64535a550ddd tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 717.104990] env[65726]: WARNING neutronclient.v2_0.client [None req-c24265b5-a0c6-4266-b4c0-64535a550ddd tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 717.105729] env[65726]: WARNING openstack [None req-c24265b5-a0c6-4266-b4c0-64535a550ddd tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 717.106236] env[65726]: WARNING openstack [None req-c24265b5-a0c6-4266-b4c0-64535a550ddd tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 717.126086] env[65726]: DEBUG oslo_concurrency.lockutils [None req-f3171f49-8536-4923-b5cf-31399a616a07 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Lock "608cc0ea-3f6b-4b4f-83c1-01aa50999d63" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 63.779s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 717.199488] env[65726]: DEBUG nova.policy [None req-c24265b5-a0c6-4266-b4c0-64535a550ddd tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9b6087cfa55d41a4bad9f6cee89865d5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '39067912d091464192516c08eb576696', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65726) authorize /opt/stack/nova/nova/policy.py:192}} [ 717.312311] env[65726]: DEBUG oslo_concurrency.lockutils [None req-44e83fc7-afcc-4134-bbbb-6bc576931bb8 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 717.337529] env[65726]: WARNING openstack [None req-296fd45e-ee7a-4e44-912a-f3b5a22cbbf0 tempest-ServerExternalEventsTest-1845125219 tempest-ServerExternalEventsTest-1845125219-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 717.337913] env[65726]: WARNING openstack [None req-296fd45e-ee7a-4e44-912a-f3b5a22cbbf0 tempest-ServerExternalEventsTest-1845125219 tempest-ServerExternalEventsTest-1845125219-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 717.348441] env[65726]: DEBUG nova.network.neutron [None req-296fd45e-ee7a-4e44-912a-f3b5a22cbbf0 tempest-ServerExternalEventsTest-1845125219 tempest-ServerExternalEventsTest-1845125219-project-member] [instance: 96acc5f8-b411-408e-929d-d1035cfd50db] Instance cache missing network info. {{(pid=65726) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 717.434187] env[65726]: WARNING openstack [None req-296fd45e-ee7a-4e44-912a-f3b5a22cbbf0 tempest-ServerExternalEventsTest-1845125219 tempest-ServerExternalEventsTest-1845125219-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 717.434620] env[65726]: WARNING openstack [None req-296fd45e-ee7a-4e44-912a-f3b5a22cbbf0 tempest-ServerExternalEventsTest-1845125219 tempest-ServerExternalEventsTest-1845125219-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 717.505134] env[65726]: WARNING openstack [None req-296fd45e-ee7a-4e44-912a-f3b5a22cbbf0 tempest-ServerExternalEventsTest-1845125219 tempest-ServerExternalEventsTest-1845125219-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 717.505723] env[65726]: WARNING openstack [None req-296fd45e-ee7a-4e44-912a-f3b5a22cbbf0 tempest-ServerExternalEventsTest-1845125219 tempest-ServerExternalEventsTest-1845125219-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 717.526606] env[65726]: DEBUG nova.network.neutron [None req-c24265b5-a0c6-4266-b4c0-64535a550ddd tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: bd839fe1-8801-4ba1-9c23-288453258b66] Successfully created port: 32d9913d-f98b-4e80-af5a-5f88083281b3 {{(pid=65726) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 717.622035] env[65726]: DEBUG nova.compute.manager [None req-c24265b5-a0c6-4266-b4c0-64535a550ddd tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: bd839fe1-8801-4ba1-9c23-288453258b66] Start building block device mappings for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 717.630283] env[65726]: DEBUG nova.compute.manager [None req-af2f0c1d-9bc5-40c3-a712-d408e73bc806 tempest-ServerDiagnosticsNegativeTest-1635132145 tempest-ServerDiagnosticsNegativeTest-1635132145-project-member] [instance: 0d6f14c5-5770-476f-a186-4cf1b2d4fbe8] Starting instance... {{(pid=65726) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 717.634534] env[65726]: DEBUG nova.network.neutron [None req-296fd45e-ee7a-4e44-912a-f3b5a22cbbf0 tempest-ServerExternalEventsTest-1845125219 tempest-ServerExternalEventsTest-1845125219-project-member] [instance: 96acc5f8-b411-408e-929d-d1035cfd50db] Updating instance_info_cache with network_info: [{"id": "c90c9d5c-3ed3-4f46-bd35-88e6fd91c396", "address": "fa:16:3e:ff:e6:2d", "network": {"id": "55d629b8-a541-4720-9289-927bbc6e6f22", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.205", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "1539f06fa3534e90acbc3a60c4b8bd3f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "35342bcb-8b06-472e-b3c0-43fd3d6c4b30", "external-id": "nsx-vlan-transportzone-524", "segmentation_id": 524, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc90c9d5c-3e", "ovs_interfaceid": "c90c9d5c-3ed3-4f46-bd35-88e6fd91c396", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 717.804367] env[65726]: DEBUG nova.compute.manager [req-8fda48b1-1989-4e1f-951e-bdde06887dee req-cb811575-fe42-45cd-b600-0f13aad6c22e service nova] [instance: 96acc5f8-b411-408e-929d-d1035cfd50db] Received event network-vif-plugged-c90c9d5c-3ed3-4f46-bd35-88e6fd91c396 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 717.804623] env[65726]: DEBUG oslo_concurrency.lockutils [req-8fda48b1-1989-4e1f-951e-bdde06887dee req-cb811575-fe42-45cd-b600-0f13aad6c22e service nova] Acquiring lock "96acc5f8-b411-408e-929d-d1035cfd50db-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 717.805038] env[65726]: DEBUG oslo_concurrency.lockutils [req-8fda48b1-1989-4e1f-951e-bdde06887dee req-cb811575-fe42-45cd-b600-0f13aad6c22e service nova] Lock "96acc5f8-b411-408e-929d-d1035cfd50db-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 717.805038] env[65726]: DEBUG oslo_concurrency.lockutils [req-8fda48b1-1989-4e1f-951e-bdde06887dee req-cb811575-fe42-45cd-b600-0f13aad6c22e service nova] Lock "96acc5f8-b411-408e-929d-d1035cfd50db-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 717.805223] env[65726]: DEBUG nova.compute.manager [req-8fda48b1-1989-4e1f-951e-bdde06887dee req-cb811575-fe42-45cd-b600-0f13aad6c22e service nova] [instance: 96acc5f8-b411-408e-929d-d1035cfd50db] No waiting events found dispatching network-vif-plugged-c90c9d5c-3ed3-4f46-bd35-88e6fd91c396 {{(pid=65726) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 717.805341] env[65726]: WARNING nova.compute.manager [req-8fda48b1-1989-4e1f-951e-bdde06887dee req-cb811575-fe42-45cd-b600-0f13aad6c22e service nova] [instance: 96acc5f8-b411-408e-929d-d1035cfd50db] Received unexpected event network-vif-plugged-c90c9d5c-3ed3-4f46-bd35-88e6fd91c396 for instance with vm_state building and task_state spawning. [ 717.860121] env[65726]: DEBUG nova.compute.manager [req-635d5c0e-746a-4978-9b31-81580fd7f0cb req-16e1518d-ea61-4d0f-9a10-6b19a410519c service nova] [instance: 3a60b719-9db7-436e-9908-25f6a233c465] Received event network-vif-deleted-5c88141b-7089-49fb-afaa-a38f2cc47aaa {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 717.860455] env[65726]: DEBUG nova.compute.manager [req-635d5c0e-746a-4978-9b31-81580fd7f0cb req-16e1518d-ea61-4d0f-9a10-6b19a410519c service nova] [instance: 04ffdf80-c54c-4587-9bb4-d520dc440501] Received event network-vif-deleted-ece7a121-1068-4cb7-b500-9b6a1e8c4a1c {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 718.139084] env[65726]: DEBUG oslo_concurrency.lockutils [None req-296fd45e-ee7a-4e44-912a-f3b5a22cbbf0 tempest-ServerExternalEventsTest-1845125219 tempest-ServerExternalEventsTest-1845125219-project-member] Releasing lock "refresh_cache-96acc5f8-b411-408e-929d-d1035cfd50db" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 718.139643] env[65726]: DEBUG nova.compute.manager [None req-296fd45e-ee7a-4e44-912a-f3b5a22cbbf0 tempest-ServerExternalEventsTest-1845125219 tempest-ServerExternalEventsTest-1845125219-project-member] [instance: 96acc5f8-b411-408e-929d-d1035cfd50db] Instance network_info: |[{"id": "c90c9d5c-3ed3-4f46-bd35-88e6fd91c396", "address": "fa:16:3e:ff:e6:2d", "network": {"id": "55d629b8-a541-4720-9289-927bbc6e6f22", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.205", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "1539f06fa3534e90acbc3a60c4b8bd3f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "35342bcb-8b06-472e-b3c0-43fd3d6c4b30", "external-id": "nsx-vlan-transportzone-524", "segmentation_id": 524, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc90c9d5c-3e", "ovs_interfaceid": "c90c9d5c-3ed3-4f46-bd35-88e6fd91c396", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 718.142452] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-296fd45e-ee7a-4e44-912a-f3b5a22cbbf0 tempest-ServerExternalEventsTest-1845125219 tempest-ServerExternalEventsTest-1845125219-project-member] [instance: 96acc5f8-b411-408e-929d-d1035cfd50db] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ff:e6:2d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '35342bcb-8b06-472e-b3c0-43fd3d6c4b30', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c90c9d5c-3ed3-4f46-bd35-88e6fd91c396', 'vif_model': 'vmxnet3'}] {{(pid=65726) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 718.151738] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-296fd45e-ee7a-4e44-912a-f3b5a22cbbf0 tempest-ServerExternalEventsTest-1845125219 tempest-ServerExternalEventsTest-1845125219-project-member] Creating folder: Project (fddc635807d54dc0a5b4f2f732ba6962). Parent ref: group-v995008. {{(pid=65726) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 718.155948] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-55d84378-51a6-462e-bc5a-bd67502e96d2 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.163125] env[65726]: DEBUG oslo_concurrency.lockutils [None req-af2f0c1d-9bc5-40c3-a712-d408e73bc806 tempest-ServerDiagnosticsNegativeTest-1635132145 tempest-ServerDiagnosticsNegativeTest-1635132145-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 718.169511] env[65726]: INFO nova.virt.vmwareapi.vm_util [None req-296fd45e-ee7a-4e44-912a-f3b5a22cbbf0 tempest-ServerExternalEventsTest-1845125219 tempest-ServerExternalEventsTest-1845125219-project-member] Created folder: Project (fddc635807d54dc0a5b4f2f732ba6962) in parent group-v995008. [ 718.169611] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-296fd45e-ee7a-4e44-912a-f3b5a22cbbf0 tempest-ServerExternalEventsTest-1845125219 tempest-ServerExternalEventsTest-1845125219-project-member] Creating folder: Instances. Parent ref: group-v995094. {{(pid=65726) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 718.169888] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-114f8203-902e-4668-92cc-f844a8a54e1b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.184039] env[65726]: INFO nova.virt.vmwareapi.vm_util [None req-296fd45e-ee7a-4e44-912a-f3b5a22cbbf0 tempest-ServerExternalEventsTest-1845125219 tempest-ServerExternalEventsTest-1845125219-project-member] Created folder: Instances in parent group-v995094. [ 718.184321] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-296fd45e-ee7a-4e44-912a-f3b5a22cbbf0 tempest-ServerExternalEventsTest-1845125219 tempest-ServerExternalEventsTest-1845125219-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 718.186938] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 96acc5f8-b411-408e-929d-d1035cfd50db] Creating VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 718.187645] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5d4f2b24-f32c-4e10-a0c7-e47dfee68aff {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.210623] env[65726]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 718.210623] env[65726]: value = "task-5115697" [ 718.210623] env[65726]: _type = "Task" [ 718.210623] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 718.217279] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12275abc-3a57-4764-a954-5a205d5a9b28 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.223331] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115697, 'name': CreateVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 718.228344] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7d132d7-993b-4189-a276-cf4b4f34bf68 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.261738] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0865c44-697c-4f85-aec5-812b3f50a129 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.270733] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89262363-c54b-4089-8778-f319cd3e729d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.285704] env[65726]: DEBUG nova.compute.provider_tree [None req-2737dc3f-715d-481c-b9a9-998472382a0d tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 718.415346] env[65726]: DEBUG oslo_concurrency.lockutils [None req-af1f7990-88d5-411a-be3e-a2d4ed6b5b5b tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Acquiring lock "621b899a-e9ec-4fc0-a574-4a08cecff6d9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 718.415625] env[65726]: DEBUG oslo_concurrency.lockutils [None req-af1f7990-88d5-411a-be3e-a2d4ed6b5b5b tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Lock "621b899a-e9ec-4fc0-a574-4a08cecff6d9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 718.632977] env[65726]: DEBUG nova.compute.manager [None req-c24265b5-a0c6-4266-b4c0-64535a550ddd tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: bd839fe1-8801-4ba1-9c23-288453258b66] Start spawning the instance on the hypervisor. {{(pid=65726) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 718.665057] env[65726]: DEBUG nova.virt.hardware [None req-c24265b5-a0c6-4266-b4c0-64535a550ddd tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T19:28:34Z,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1539f06fa3534e90acbc3a60c4b8bd3f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T19:28:35Z,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 718.665057] env[65726]: DEBUG nova.virt.hardware [None req-c24265b5-a0c6-4266-b4c0-64535a550ddd tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 718.665057] env[65726]: DEBUG nova.virt.hardware [None req-c24265b5-a0c6-4266-b4c0-64535a550ddd tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 718.665057] env[65726]: DEBUG nova.virt.hardware [None req-c24265b5-a0c6-4266-b4c0-64535a550ddd tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 718.665284] env[65726]: DEBUG nova.virt.hardware [None req-c24265b5-a0c6-4266-b4c0-64535a550ddd tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 718.665284] env[65726]: DEBUG nova.virt.hardware [None req-c24265b5-a0c6-4266-b4c0-64535a550ddd tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 718.665429] env[65726]: DEBUG nova.virt.hardware [None req-c24265b5-a0c6-4266-b4c0-64535a550ddd tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 718.665601] env[65726]: DEBUG nova.virt.hardware [None req-c24265b5-a0c6-4266-b4c0-64535a550ddd tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 718.665773] env[65726]: DEBUG nova.virt.hardware [None req-c24265b5-a0c6-4266-b4c0-64535a550ddd tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 718.665949] env[65726]: DEBUG nova.virt.hardware [None req-c24265b5-a0c6-4266-b4c0-64535a550ddd tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 718.666130] env[65726]: DEBUG nova.virt.hardware [None req-c24265b5-a0c6-4266-b4c0-64535a550ddd tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 718.667133] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5df5a28-28b1-410b-8522-f66393113495 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.677395] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf4cdeda-c972-4c18-a035-9889b272d752 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.720933] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115697, 'name': CreateVM_Task} progress is 99%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 718.791535] env[65726]: DEBUG nova.scheduler.client.report [None req-2737dc3f-715d-481c-b9a9-998472382a0d tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 719.147226] env[65726]: DEBUG nova.network.neutron [None req-c24265b5-a0c6-4266-b4c0-64535a550ddd tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: bd839fe1-8801-4ba1-9c23-288453258b66] Successfully updated port: 32d9913d-f98b-4e80-af5a-5f88083281b3 {{(pid=65726) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 719.221566] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115697, 'name': CreateVM_Task, 'duration_secs': 0.559285} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 719.221775] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 96acc5f8-b411-408e-929d-d1035cfd50db] Created VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 719.222616] env[65726]: WARNING openstack [None req-296fd45e-ee7a-4e44-912a-f3b5a22cbbf0 tempest-ServerExternalEventsTest-1845125219 tempest-ServerExternalEventsTest-1845125219-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 719.222980] env[65726]: WARNING openstack [None req-296fd45e-ee7a-4e44-912a-f3b5a22cbbf0 tempest-ServerExternalEventsTest-1845125219 tempest-ServerExternalEventsTest-1845125219-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 719.230039] env[65726]: DEBUG oslo_concurrency.lockutils [None req-296fd45e-ee7a-4e44-912a-f3b5a22cbbf0 tempest-ServerExternalEventsTest-1845125219 tempest-ServerExternalEventsTest-1845125219-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 719.230666] env[65726]: DEBUG oslo_concurrency.lockutils [None req-296fd45e-ee7a-4e44-912a-f3b5a22cbbf0 tempest-ServerExternalEventsTest-1845125219 tempest-ServerExternalEventsTest-1845125219-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 719.231240] env[65726]: DEBUG oslo_concurrency.lockutils [None req-296fd45e-ee7a-4e44-912a-f3b5a22cbbf0 tempest-ServerExternalEventsTest-1845125219 tempest-ServerExternalEventsTest-1845125219-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 719.234603] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3263cbc3-2863-463f-98a9-a659346c1df8 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.240411] env[65726]: DEBUG oslo_vmware.api [None req-296fd45e-ee7a-4e44-912a-f3b5a22cbbf0 tempest-ServerExternalEventsTest-1845125219 tempest-ServerExternalEventsTest-1845125219-project-member] Waiting for the task: (returnval){ [ 719.240411] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5271169e-7ddb-bad4-f0c5-d33f3b8eff45" [ 719.240411] env[65726]: _type = "Task" [ 719.240411] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 719.250758] env[65726]: DEBUG oslo_vmware.api [None req-296fd45e-ee7a-4e44-912a-f3b5a22cbbf0 tempest-ServerExternalEventsTest-1845125219 tempest-ServerExternalEventsTest-1845125219-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5271169e-7ddb-bad4-f0c5-d33f3b8eff45, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 719.296547] env[65726]: DEBUG oslo_concurrency.lockutils [None req-2737dc3f-715d-481c-b9a9-998472382a0d tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.710s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 719.301027] env[65726]: DEBUG nova.compute.manager [None req-2737dc3f-715d-481c-b9a9-998472382a0d tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] [instance: 83573c2b-9448-456f-8fd2-b19661dd6cc4] Start building networks asynchronously for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 719.303051] env[65726]: DEBUG oslo_concurrency.lockutils [None req-06465b67-a5cc-4aeb-80ed-00dcc8ab79e5 tempest-ServersTestJSON-1793637688 tempest-ServersTestJSON-1793637688-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 36.081s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 719.303287] env[65726]: DEBUG nova.objects.instance [None req-06465b67-a5cc-4aeb-80ed-00dcc8ab79e5 tempest-ServersTestJSON-1793637688 tempest-ServersTestJSON-1793637688-project-member] Lazy-loading 'resources' on Instance uuid 88fb7235-cd0e-49dc-9d78-3583d5d3f528 {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 719.650378] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c24265b5-a0c6-4266-b4c0-64535a550ddd tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Acquiring lock "refresh_cache-bd839fe1-8801-4ba1-9c23-288453258b66" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 719.650598] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c24265b5-a0c6-4266-b4c0-64535a550ddd tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Acquired lock "refresh_cache-bd839fe1-8801-4ba1-9c23-288453258b66" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 719.650775] env[65726]: DEBUG nova.network.neutron [None req-c24265b5-a0c6-4266-b4c0-64535a550ddd tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: bd839fe1-8801-4ba1-9c23-288453258b66] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 719.752061] env[65726]: DEBUG oslo_vmware.api [None req-296fd45e-ee7a-4e44-912a-f3b5a22cbbf0 tempest-ServerExternalEventsTest-1845125219 tempest-ServerExternalEventsTest-1845125219-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5271169e-7ddb-bad4-f0c5-d33f3b8eff45, 'name': SearchDatastore_Task, 'duration_secs': 0.034231} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 719.752845] env[65726]: DEBUG oslo_concurrency.lockutils [None req-296fd45e-ee7a-4e44-912a-f3b5a22cbbf0 tempest-ServerExternalEventsTest-1845125219 tempest-ServerExternalEventsTest-1845125219-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 719.753853] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-296fd45e-ee7a-4e44-912a-f3b5a22cbbf0 tempest-ServerExternalEventsTest-1845125219 tempest-ServerExternalEventsTest-1845125219-project-member] [instance: 96acc5f8-b411-408e-929d-d1035cfd50db] Processing image b52362a3-ee8a-4cbf-b06f-513b0cc8f95c {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 719.753853] env[65726]: DEBUG oslo_concurrency.lockutils [None req-296fd45e-ee7a-4e44-912a-f3b5a22cbbf0 tempest-ServerExternalEventsTest-1845125219 tempest-ServerExternalEventsTest-1845125219-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 719.753853] env[65726]: DEBUG oslo_concurrency.lockutils [None req-296fd45e-ee7a-4e44-912a-f3b5a22cbbf0 tempest-ServerExternalEventsTest-1845125219 tempest-ServerExternalEventsTest-1845125219-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 719.753853] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-296fd45e-ee7a-4e44-912a-f3b5a22cbbf0 tempest-ServerExternalEventsTest-1845125219 tempest-ServerExternalEventsTest-1845125219-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 719.754162] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b05abed7-3106-4cee-9c5c-5602c341fa34 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.768045] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-296fd45e-ee7a-4e44-912a-f3b5a22cbbf0 tempest-ServerExternalEventsTest-1845125219 tempest-ServerExternalEventsTest-1845125219-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 719.768045] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-296fd45e-ee7a-4e44-912a-f3b5a22cbbf0 tempest-ServerExternalEventsTest-1845125219 tempest-ServerExternalEventsTest-1845125219-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65726) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 719.768045] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-85438263-aca2-42bf-9e60-78c14464d0f7 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.771977] env[65726]: DEBUG oslo_vmware.api [None req-296fd45e-ee7a-4e44-912a-f3b5a22cbbf0 tempest-ServerExternalEventsTest-1845125219 tempest-ServerExternalEventsTest-1845125219-project-member] Waiting for the task: (returnval){ [ 719.771977] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5210ae31-d756-5b7f-41b1-6d68a8e164ee" [ 719.771977] env[65726]: _type = "Task" [ 719.771977] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 719.781026] env[65726]: DEBUG oslo_vmware.api [None req-296fd45e-ee7a-4e44-912a-f3b5a22cbbf0 tempest-ServerExternalEventsTest-1845125219 tempest-ServerExternalEventsTest-1845125219-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5210ae31-d756-5b7f-41b1-6d68a8e164ee, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 719.809752] env[65726]: DEBUG nova.compute.utils [None req-2737dc3f-715d-481c-b9a9-998472382a0d tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] Using /dev/sd instead of None {{(pid=65726) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 719.811780] env[65726]: DEBUG nova.compute.manager [None req-2737dc3f-715d-481c-b9a9-998472382a0d tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] [instance: 83573c2b-9448-456f-8fd2-b19661dd6cc4] Not allocating networking since 'none' was specified. {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2015}} [ 719.967215] env[65726]: DEBUG nova.compute.manager [req-2d361622-38c8-4d30-9e3a-bf1999f11d50 req-0283ff69-1ce8-4ee0-ab10-b3e3788dfc02 service nova] [instance: 96acc5f8-b411-408e-929d-d1035cfd50db] Received event network-changed-c90c9d5c-3ed3-4f46-bd35-88e6fd91c396 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 719.967411] env[65726]: DEBUG nova.compute.manager [req-2d361622-38c8-4d30-9e3a-bf1999f11d50 req-0283ff69-1ce8-4ee0-ab10-b3e3788dfc02 service nova] [instance: 96acc5f8-b411-408e-929d-d1035cfd50db] Refreshing instance network info cache due to event network-changed-c90c9d5c-3ed3-4f46-bd35-88e6fd91c396. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 719.967627] env[65726]: DEBUG oslo_concurrency.lockutils [req-2d361622-38c8-4d30-9e3a-bf1999f11d50 req-0283ff69-1ce8-4ee0-ab10-b3e3788dfc02 service nova] Acquiring lock "refresh_cache-96acc5f8-b411-408e-929d-d1035cfd50db" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 719.967762] env[65726]: DEBUG oslo_concurrency.lockutils [req-2d361622-38c8-4d30-9e3a-bf1999f11d50 req-0283ff69-1ce8-4ee0-ab10-b3e3788dfc02 service nova] Acquired lock "refresh_cache-96acc5f8-b411-408e-929d-d1035cfd50db" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 719.967914] env[65726]: DEBUG nova.network.neutron [req-2d361622-38c8-4d30-9e3a-bf1999f11d50 req-0283ff69-1ce8-4ee0-ab10-b3e3788dfc02 service nova] [instance: 96acc5f8-b411-408e-929d-d1035cfd50db] Refreshing network info cache for port c90c9d5c-3ed3-4f46-bd35-88e6fd91c396 {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 720.048076] env[65726]: DEBUG nova.compute.manager [req-381a4cbc-a8b2-4387-a340-363cc9279fe4 req-d12fd3a7-f5cf-43e4-929b-53c5524490f7 service nova] [instance: bd839fe1-8801-4ba1-9c23-288453258b66] Received event network-vif-plugged-32d9913d-f98b-4e80-af5a-5f88083281b3 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 720.048321] env[65726]: DEBUG oslo_concurrency.lockutils [req-381a4cbc-a8b2-4387-a340-363cc9279fe4 req-d12fd3a7-f5cf-43e4-929b-53c5524490f7 service nova] Acquiring lock "bd839fe1-8801-4ba1-9c23-288453258b66-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 720.048524] env[65726]: DEBUG oslo_concurrency.lockutils [req-381a4cbc-a8b2-4387-a340-363cc9279fe4 req-d12fd3a7-f5cf-43e4-929b-53c5524490f7 service nova] Lock "bd839fe1-8801-4ba1-9c23-288453258b66-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 720.048683] env[65726]: DEBUG oslo_concurrency.lockutils [req-381a4cbc-a8b2-4387-a340-363cc9279fe4 req-d12fd3a7-f5cf-43e4-929b-53c5524490f7 service nova] Lock "bd839fe1-8801-4ba1-9c23-288453258b66-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 720.048844] env[65726]: DEBUG nova.compute.manager [req-381a4cbc-a8b2-4387-a340-363cc9279fe4 req-d12fd3a7-f5cf-43e4-929b-53c5524490f7 service nova] [instance: bd839fe1-8801-4ba1-9c23-288453258b66] No waiting events found dispatching network-vif-plugged-32d9913d-f98b-4e80-af5a-5f88083281b3 {{(pid=65726) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 720.048997] env[65726]: WARNING nova.compute.manager [req-381a4cbc-a8b2-4387-a340-363cc9279fe4 req-d12fd3a7-f5cf-43e4-929b-53c5524490f7 service nova] [instance: bd839fe1-8801-4ba1-9c23-288453258b66] Received unexpected event network-vif-plugged-32d9913d-f98b-4e80-af5a-5f88083281b3 for instance with vm_state building and task_state spawning. [ 720.049165] env[65726]: DEBUG nova.compute.manager [req-381a4cbc-a8b2-4387-a340-363cc9279fe4 req-d12fd3a7-f5cf-43e4-929b-53c5524490f7 service nova] [instance: bd839fe1-8801-4ba1-9c23-288453258b66] Received event network-changed-32d9913d-f98b-4e80-af5a-5f88083281b3 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 720.049324] env[65726]: DEBUG nova.compute.manager [req-381a4cbc-a8b2-4387-a340-363cc9279fe4 req-d12fd3a7-f5cf-43e4-929b-53c5524490f7 service nova] [instance: bd839fe1-8801-4ba1-9c23-288453258b66] Refreshing instance network info cache due to event network-changed-32d9913d-f98b-4e80-af5a-5f88083281b3. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 720.049512] env[65726]: DEBUG oslo_concurrency.lockutils [req-381a4cbc-a8b2-4387-a340-363cc9279fe4 req-d12fd3a7-f5cf-43e4-929b-53c5524490f7 service nova] Acquiring lock "refresh_cache-bd839fe1-8801-4ba1-9c23-288453258b66" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 720.122502] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5acbf047-f1f8-4297-909c-a3d7a9b0fdd6 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Acquiring lock "815f056b-2d88-40dd-9ef7-a891ec8d92cf" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 720.122779] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5acbf047-f1f8-4297-909c-a3d7a9b0fdd6 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Lock "815f056b-2d88-40dd-9ef7-a891ec8d92cf" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 720.122992] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5acbf047-f1f8-4297-909c-a3d7a9b0fdd6 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Acquiring lock "815f056b-2d88-40dd-9ef7-a891ec8d92cf-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 720.123201] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5acbf047-f1f8-4297-909c-a3d7a9b0fdd6 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Lock "815f056b-2d88-40dd-9ef7-a891ec8d92cf-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 720.123633] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5acbf047-f1f8-4297-909c-a3d7a9b0fdd6 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Lock "815f056b-2d88-40dd-9ef7-a891ec8d92cf-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 720.125944] env[65726]: INFO nova.compute.manager [None req-5acbf047-f1f8-4297-909c-a3d7a9b0fdd6 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] [instance: 815f056b-2d88-40dd-9ef7-a891ec8d92cf] Terminating instance [ 720.155451] env[65726]: WARNING openstack [None req-c24265b5-a0c6-4266-b4c0-64535a550ddd tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 720.156581] env[65726]: WARNING openstack [None req-c24265b5-a0c6-4266-b4c0-64535a550ddd tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 720.162567] env[65726]: DEBUG nova.network.neutron [None req-c24265b5-a0c6-4266-b4c0-64535a550ddd tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: bd839fe1-8801-4ba1-9c23-288453258b66] Instance cache missing network info. {{(pid=65726) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 720.226030] env[65726]: WARNING openstack [None req-c24265b5-a0c6-4266-b4c0-64535a550ddd tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 720.226425] env[65726]: WARNING openstack [None req-c24265b5-a0c6-4266-b4c0-64535a550ddd tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 720.284501] env[65726]: DEBUG oslo_vmware.api [None req-296fd45e-ee7a-4e44-912a-f3b5a22cbbf0 tempest-ServerExternalEventsTest-1845125219 tempest-ServerExternalEventsTest-1845125219-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5210ae31-d756-5b7f-41b1-6d68a8e164ee, 'name': SearchDatastore_Task, 'duration_secs': 0.010399} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 720.289071] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-beb8ae2f-fd26-467d-9b06-83b8272b3ff7 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.295451] env[65726]: DEBUG oslo_vmware.api [None req-296fd45e-ee7a-4e44-912a-f3b5a22cbbf0 tempest-ServerExternalEventsTest-1845125219 tempest-ServerExternalEventsTest-1845125219-project-member] Waiting for the task: (returnval){ [ 720.295451] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52a3959b-ee6f-ea82-1d48-3de3f6319efa" [ 720.295451] env[65726]: _type = "Task" [ 720.295451] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 720.305479] env[65726]: WARNING openstack [None req-c24265b5-a0c6-4266-b4c0-64535a550ddd tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 720.305927] env[65726]: WARNING openstack [None req-c24265b5-a0c6-4266-b4c0-64535a550ddd tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 720.315057] env[65726]: DEBUG nova.compute.manager [None req-2737dc3f-715d-481c-b9a9-998472382a0d tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] [instance: 83573c2b-9448-456f-8fd2-b19661dd6cc4] Start building block device mappings for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 720.322958] env[65726]: DEBUG oslo_vmware.api [None req-296fd45e-ee7a-4e44-912a-f3b5a22cbbf0 tempest-ServerExternalEventsTest-1845125219 tempest-ServerExternalEventsTest-1845125219-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52a3959b-ee6f-ea82-1d48-3de3f6319efa, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 720.396586] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f49d0d9-754c-448e-8d64-74d7268931a2 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.404761] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f011d3a2-6d3c-4f20-ba60-3e567591b279 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.415018] env[65726]: DEBUG nova.network.neutron [None req-c24265b5-a0c6-4266-b4c0-64535a550ddd tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: bd839fe1-8801-4ba1-9c23-288453258b66] Updating instance_info_cache with network_info: [{"id": "32d9913d-f98b-4e80-af5a-5f88083281b3", "address": "fa:16:3e:89:aa:99", "network": {"id": "e206b5b9-da36-4e6f-926f-914dec38fac2", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1053201655-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "39067912d091464192516c08eb576696", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7c80243e-93a7-4a95-bc8d-e9534bacd66e", "external-id": "nsx-vlan-transportzone-306", "segmentation_id": 306, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap32d9913d-f9", "ovs_interfaceid": "32d9913d-f98b-4e80-af5a-5f88083281b3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 720.444506] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ae0cb1d-e49d-45ba-be78-9777cabc0635 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.453746] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f965b17-93e5-46a5-a7d3-48f41c93035b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.468942] env[65726]: DEBUG nova.compute.provider_tree [None req-06465b67-a5cc-4aeb-80ed-00dcc8ab79e5 tempest-ServersTestJSON-1793637688 tempest-ServersTestJSON-1793637688-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 720.470854] env[65726]: WARNING openstack [req-2d361622-38c8-4d30-9e3a-bf1999f11d50 req-0283ff69-1ce8-4ee0-ab10-b3e3788dfc02 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 720.471386] env[65726]: WARNING openstack [req-2d361622-38c8-4d30-9e3a-bf1999f11d50 req-0283ff69-1ce8-4ee0-ab10-b3e3788dfc02 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 720.632801] env[65726]: DEBUG nova.compute.manager [None req-5acbf047-f1f8-4297-909c-a3d7a9b0fdd6 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] [instance: 815f056b-2d88-40dd-9ef7-a891ec8d92cf] Start destroying the instance on the hypervisor. {{(pid=65726) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 720.633064] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-5acbf047-f1f8-4297-909c-a3d7a9b0fdd6 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] [instance: 815f056b-2d88-40dd-9ef7-a891ec8d92cf] Destroying instance {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 720.633957] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3155bf02-e997-4f1c-9ce3-3ce0164cfc58 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.643502] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-5acbf047-f1f8-4297-909c-a3d7a9b0fdd6 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] [instance: 815f056b-2d88-40dd-9ef7-a891ec8d92cf] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 720.643888] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e6a38790-36a8-49ba-a225-2c6bf2a0c95d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.651506] env[65726]: DEBUG oslo_vmware.api [None req-5acbf047-f1f8-4297-909c-a3d7a9b0fdd6 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Waiting for the task: (returnval){ [ 720.651506] env[65726]: value = "task-5115698" [ 720.651506] env[65726]: _type = "Task" [ 720.651506] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 720.661361] env[65726]: DEBUG oslo_vmware.api [None req-5acbf047-f1f8-4297-909c-a3d7a9b0fdd6 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Task: {'id': task-5115698, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 720.806447] env[65726]: DEBUG oslo_vmware.api [None req-296fd45e-ee7a-4e44-912a-f3b5a22cbbf0 tempest-ServerExternalEventsTest-1845125219 tempest-ServerExternalEventsTest-1845125219-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52a3959b-ee6f-ea82-1d48-3de3f6319efa, 'name': SearchDatastore_Task, 'duration_secs': 0.020075} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 720.806726] env[65726]: DEBUG oslo_concurrency.lockutils [None req-296fd45e-ee7a-4e44-912a-f3b5a22cbbf0 tempest-ServerExternalEventsTest-1845125219 tempest-ServerExternalEventsTest-1845125219-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 720.807075] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-296fd45e-ee7a-4e44-912a-f3b5a22cbbf0 tempest-ServerExternalEventsTest-1845125219 tempest-ServerExternalEventsTest-1845125219-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore2] 96acc5f8-b411-408e-929d-d1035cfd50db/96acc5f8-b411-408e-929d-d1035cfd50db.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 720.807438] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-af8056e9-383c-41f0-8186-7f511988c868 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.815951] env[65726]: DEBUG oslo_vmware.api [None req-296fd45e-ee7a-4e44-912a-f3b5a22cbbf0 tempest-ServerExternalEventsTest-1845125219 tempest-ServerExternalEventsTest-1845125219-project-member] Waiting for the task: (returnval){ [ 720.815951] env[65726]: value = "task-5115699" [ 720.815951] env[65726]: _type = "Task" [ 720.815951] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 720.830993] env[65726]: DEBUG oslo_vmware.api [None req-296fd45e-ee7a-4e44-912a-f3b5a22cbbf0 tempest-ServerExternalEventsTest-1845125219 tempest-ServerExternalEventsTest-1845125219-project-member] Task: {'id': task-5115699, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 720.893364] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b1ce1b66-ca58-46d0-b624-7c893de1adee tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Acquiring lock "423af936-a383-4652-8887-9c8b8f507909" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 720.893364] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b1ce1b66-ca58-46d0-b624-7c893de1adee tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Lock "423af936-a383-4652-8887-9c8b8f507909" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 720.893364] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b1ce1b66-ca58-46d0-b624-7c893de1adee tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Acquiring lock "423af936-a383-4652-8887-9c8b8f507909-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 720.893364] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b1ce1b66-ca58-46d0-b624-7c893de1adee tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Lock "423af936-a383-4652-8887-9c8b8f507909-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 720.893656] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b1ce1b66-ca58-46d0-b624-7c893de1adee tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Lock "423af936-a383-4652-8887-9c8b8f507909-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 720.898019] env[65726]: INFO nova.compute.manager [None req-b1ce1b66-ca58-46d0-b624-7c893de1adee tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] [instance: 423af936-a383-4652-8887-9c8b8f507909] Terminating instance [ 720.919545] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c24265b5-a0c6-4266-b4c0-64535a550ddd tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Releasing lock "refresh_cache-bd839fe1-8801-4ba1-9c23-288453258b66" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 720.919545] env[65726]: DEBUG nova.compute.manager [None req-c24265b5-a0c6-4266-b4c0-64535a550ddd tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: bd839fe1-8801-4ba1-9c23-288453258b66] Instance network_info: |[{"id": "32d9913d-f98b-4e80-af5a-5f88083281b3", "address": "fa:16:3e:89:aa:99", "network": {"id": "e206b5b9-da36-4e6f-926f-914dec38fac2", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1053201655-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "39067912d091464192516c08eb576696", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7c80243e-93a7-4a95-bc8d-e9534bacd66e", "external-id": "nsx-vlan-transportzone-306", "segmentation_id": 306, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap32d9913d-f9", "ovs_interfaceid": "32d9913d-f98b-4e80-af5a-5f88083281b3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 720.919772] env[65726]: DEBUG oslo_concurrency.lockutils [req-381a4cbc-a8b2-4387-a340-363cc9279fe4 req-d12fd3a7-f5cf-43e4-929b-53c5524490f7 service nova] Acquired lock "refresh_cache-bd839fe1-8801-4ba1-9c23-288453258b66" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 720.919772] env[65726]: DEBUG nova.network.neutron [req-381a4cbc-a8b2-4387-a340-363cc9279fe4 req-d12fd3a7-f5cf-43e4-929b-53c5524490f7 service nova] [instance: bd839fe1-8801-4ba1-9c23-288453258b66] Refreshing network info cache for port 32d9913d-f98b-4e80-af5a-5f88083281b3 {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 720.919834] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-c24265b5-a0c6-4266-b4c0-64535a550ddd tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: bd839fe1-8801-4ba1-9c23-288453258b66] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:89:aa:99', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '7c80243e-93a7-4a95-bc8d-e9534bacd66e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '32d9913d-f98b-4e80-af5a-5f88083281b3', 'vif_model': 'vmxnet3'}] {{(pid=65726) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 720.928727] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-c24265b5-a0c6-4266-b4c0-64535a550ddd tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 720.929567] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bd839fe1-8801-4ba1-9c23-288453258b66] Creating VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 720.929927] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-eb2bb10d-e33c-4400-9dcb-9feb84c44f65 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.954264] env[65726]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 720.954264] env[65726]: value = "task-5115700" [ 720.954264] env[65726]: _type = "Task" [ 720.954264] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 720.964898] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115700, 'name': CreateVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 720.979138] env[65726]: WARNING openstack [req-2d361622-38c8-4d30-9e3a-bf1999f11d50 req-0283ff69-1ce8-4ee0-ab10-b3e3788dfc02 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 720.979559] env[65726]: WARNING openstack [req-2d361622-38c8-4d30-9e3a-bf1999f11d50 req-0283ff69-1ce8-4ee0-ab10-b3e3788dfc02 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 720.988121] env[65726]: DEBUG nova.scheduler.client.report [None req-06465b67-a5cc-4aeb-80ed-00dcc8ab79e5 tempest-ServersTestJSON-1793637688 tempest-ServersTestJSON-1793637688-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 721.072220] env[65726]: WARNING openstack [req-2d361622-38c8-4d30-9e3a-bf1999f11d50 req-0283ff69-1ce8-4ee0-ab10-b3e3788dfc02 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 721.072629] env[65726]: WARNING openstack [req-2d361622-38c8-4d30-9e3a-bf1999f11d50 req-0283ff69-1ce8-4ee0-ab10-b3e3788dfc02 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 721.163420] env[65726]: DEBUG oslo_vmware.api [None req-5acbf047-f1f8-4297-909c-a3d7a9b0fdd6 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Task: {'id': task-5115698, 'name': PowerOffVM_Task, 'duration_secs': 0.192315} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 721.163825] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-5acbf047-f1f8-4297-909c-a3d7a9b0fdd6 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] [instance: 815f056b-2d88-40dd-9ef7-a891ec8d92cf] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 721.165673] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-5acbf047-f1f8-4297-909c-a3d7a9b0fdd6 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] [instance: 815f056b-2d88-40dd-9ef7-a891ec8d92cf] Unregistering the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 721.165673] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0ea2bef3-0285-403a-9705-8bb57b3811c4 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.196882] env[65726]: DEBUG nova.network.neutron [req-2d361622-38c8-4d30-9e3a-bf1999f11d50 req-0283ff69-1ce8-4ee0-ab10-b3e3788dfc02 service nova] [instance: 96acc5f8-b411-408e-929d-d1035cfd50db] Updated VIF entry in instance network info cache for port c90c9d5c-3ed3-4f46-bd35-88e6fd91c396. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 721.197382] env[65726]: DEBUG nova.network.neutron [req-2d361622-38c8-4d30-9e3a-bf1999f11d50 req-0283ff69-1ce8-4ee0-ab10-b3e3788dfc02 service nova] [instance: 96acc5f8-b411-408e-929d-d1035cfd50db] Updating instance_info_cache with network_info: [{"id": "c90c9d5c-3ed3-4f46-bd35-88e6fd91c396", "address": "fa:16:3e:ff:e6:2d", "network": {"id": "55d629b8-a541-4720-9289-927bbc6e6f22", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.205", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "1539f06fa3534e90acbc3a60c4b8bd3f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "35342bcb-8b06-472e-b3c0-43fd3d6c4b30", "external-id": "nsx-vlan-transportzone-524", "segmentation_id": 524, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc90c9d5c-3e", "ovs_interfaceid": "c90c9d5c-3ed3-4f46-bd35-88e6fd91c396", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 721.242785] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-5acbf047-f1f8-4297-909c-a3d7a9b0fdd6 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] [instance: 815f056b-2d88-40dd-9ef7-a891ec8d92cf] Unregistered the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 721.243070] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-5acbf047-f1f8-4297-909c-a3d7a9b0fdd6 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] [instance: 815f056b-2d88-40dd-9ef7-a891ec8d92cf] Deleting contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 721.243267] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-5acbf047-f1f8-4297-909c-a3d7a9b0fdd6 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Deleting the datastore file [datastore1] 815f056b-2d88-40dd-9ef7-a891ec8d92cf {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 721.243708] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-43a91109-4288-4997-9129-ddc67db0b009 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.256185] env[65726]: DEBUG oslo_vmware.api [None req-5acbf047-f1f8-4297-909c-a3d7a9b0fdd6 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Waiting for the task: (returnval){ [ 721.256185] env[65726]: value = "task-5115702" [ 721.256185] env[65726]: _type = "Task" [ 721.256185] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 721.272728] env[65726]: DEBUG oslo_vmware.api [None req-5acbf047-f1f8-4297-909c-a3d7a9b0fdd6 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Task: {'id': task-5115702, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 721.328411] env[65726]: DEBUG oslo_vmware.api [None req-296fd45e-ee7a-4e44-912a-f3b5a22cbbf0 tempest-ServerExternalEventsTest-1845125219 tempest-ServerExternalEventsTest-1845125219-project-member] Task: {'id': task-5115699, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 721.335093] env[65726]: DEBUG nova.compute.manager [None req-2737dc3f-715d-481c-b9a9-998472382a0d tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] [instance: 83573c2b-9448-456f-8fd2-b19661dd6cc4] Start spawning the instance on the hypervisor. {{(pid=65726) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 721.365104] env[65726]: DEBUG nova.virt.hardware [None req-2737dc3f-715d-481c-b9a9-998472382a0d tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T19:28:34Z,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1539f06fa3534e90acbc3a60c4b8bd3f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T19:28:35Z,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 721.365348] env[65726]: DEBUG nova.virt.hardware [None req-2737dc3f-715d-481c-b9a9-998472382a0d tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 721.365500] env[65726]: DEBUG nova.virt.hardware [None req-2737dc3f-715d-481c-b9a9-998472382a0d tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 721.365673] env[65726]: DEBUG nova.virt.hardware [None req-2737dc3f-715d-481c-b9a9-998472382a0d tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 721.365811] env[65726]: DEBUG nova.virt.hardware [None req-2737dc3f-715d-481c-b9a9-998472382a0d tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 721.365950] env[65726]: DEBUG nova.virt.hardware [None req-2737dc3f-715d-481c-b9a9-998472382a0d tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 721.366370] env[65726]: DEBUG nova.virt.hardware [None req-2737dc3f-715d-481c-b9a9-998472382a0d tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 721.366594] env[65726]: DEBUG nova.virt.hardware [None req-2737dc3f-715d-481c-b9a9-998472382a0d tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 721.366827] env[65726]: DEBUG nova.virt.hardware [None req-2737dc3f-715d-481c-b9a9-998472382a0d tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 721.367047] env[65726]: DEBUG nova.virt.hardware [None req-2737dc3f-715d-481c-b9a9-998472382a0d tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 721.367267] env[65726]: DEBUG nova.virt.hardware [None req-2737dc3f-715d-481c-b9a9-998472382a0d tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 721.368219] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47edae68-ecd1-495b-a70e-ca634a7756f1 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.377499] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7630ea61-064b-4df1-82cb-8ed0e4d8eb24 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.393919] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-2737dc3f-715d-481c-b9a9-998472382a0d tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] [instance: 83573c2b-9448-456f-8fd2-b19661dd6cc4] Instance VIF info [] {{(pid=65726) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 721.399612] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-2737dc3f-715d-481c-b9a9-998472382a0d tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] Creating folder: Project (1f3ef84831b3441c8532cdce35f4d245). Parent ref: group-v995008. {{(pid=65726) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 721.399966] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-8473021f-67c9-4c11-8867-6adcbca5cc1d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.404500] env[65726]: DEBUG nova.compute.manager [None req-b1ce1b66-ca58-46d0-b624-7c893de1adee tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] [instance: 423af936-a383-4652-8887-9c8b8f507909] Start destroying the instance on the hypervisor. {{(pid=65726) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 721.404711] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-b1ce1b66-ca58-46d0-b624-7c893de1adee tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] [instance: 423af936-a383-4652-8887-9c8b8f507909] Destroying instance {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 721.405660] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d7d0fb3-b567-495a-942b-2dc6962d6590 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.414138] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-b1ce1b66-ca58-46d0-b624-7c893de1adee tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] [instance: 423af936-a383-4652-8887-9c8b8f507909] Unregistering the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 721.414543] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d1a9fa45-72b1-4971-9584-07d3daa4d42b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.416496] env[65726]: INFO nova.virt.vmwareapi.vm_util [None req-2737dc3f-715d-481c-b9a9-998472382a0d tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] Created folder: Project (1f3ef84831b3441c8532cdce35f4d245) in parent group-v995008. [ 721.416671] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-2737dc3f-715d-481c-b9a9-998472382a0d tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] Creating folder: Instances. Parent ref: group-v995098. {{(pid=65726) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 721.416894] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-8610ae09-db74-4acd-b464-735d03269e44 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.429360] env[65726]: INFO nova.virt.vmwareapi.vm_util [None req-2737dc3f-715d-481c-b9a9-998472382a0d tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] Created folder: Instances in parent group-v995098. [ 721.429694] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-2737dc3f-715d-481c-b9a9-998472382a0d tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 721.432256] env[65726]: WARNING openstack [req-381a4cbc-a8b2-4387-a340-363cc9279fe4 req-d12fd3a7-f5cf-43e4-929b-53c5524490f7 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 721.432256] env[65726]: WARNING openstack [req-381a4cbc-a8b2-4387-a340-363cc9279fe4 req-d12fd3a7-f5cf-43e4-929b-53c5524490f7 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 721.440116] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 83573c2b-9448-456f-8fd2-b19661dd6cc4] Creating VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 721.440612] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-64e5ddd0-a0dd-4efd-8d1e-8584343cb4b8 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.463243] env[65726]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 721.463243] env[65726]: value = "task-5115706" [ 721.463243] env[65726]: _type = "Task" [ 721.463243] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 721.472257] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115700, 'name': CreateVM_Task, 'duration_secs': 0.403615} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 721.473994] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bd839fe1-8801-4ba1-9c23-288453258b66] Created VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 721.473994] env[65726]: WARNING openstack [None req-c24265b5-a0c6-4266-b4c0-64535a550ddd tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 721.474259] env[65726]: WARNING openstack [None req-c24265b5-a0c6-4266-b4c0-64535a550ddd tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 721.480256] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c24265b5-a0c6-4266-b4c0-64535a550ddd tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 721.480256] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c24265b5-a0c6-4266-b4c0-64535a550ddd tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 721.480256] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c24265b5-a0c6-4266-b4c0-64535a550ddd tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 721.488984] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-821c63e4-26d2-4806-9de4-997996679013 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.491068] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115706, 'name': CreateVM_Task} progress is 6%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 721.492674] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-b1ce1b66-ca58-46d0-b624-7c893de1adee tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] [instance: 423af936-a383-4652-8887-9c8b8f507909] Unregistered the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 721.493847] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-b1ce1b66-ca58-46d0-b624-7c893de1adee tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] [instance: 423af936-a383-4652-8887-9c8b8f507909] Deleting contents of the VM from datastore datastore2 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 721.493847] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-b1ce1b66-ca58-46d0-b624-7c893de1adee tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Deleting the datastore file [datastore2] 423af936-a383-4652-8887-9c8b8f507909 {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 721.493847] env[65726]: DEBUG oslo_concurrency.lockutils [None req-06465b67-a5cc-4aeb-80ed-00dcc8ab79e5 tempest-ServersTestJSON-1793637688 tempest-ServersTestJSON-1793637688-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.191s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 721.496967] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f6a9da9e-ced8-4a0c-8e94-f778e6cbfcf3 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.499820] env[65726]: DEBUG oslo_concurrency.lockutils [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 34.232s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 721.501679] env[65726]: INFO nova.compute.claims [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] [instance: 02351120-8385-4403-8464-a154f3da9380] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 721.506985] env[65726]: DEBUG oslo_vmware.api [None req-c24265b5-a0c6-4266-b4c0-64535a550ddd tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Waiting for the task: (returnval){ [ 721.506985] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]523b5e05-1773-a061-306b-977bcc8a3aca" [ 721.506985] env[65726]: _type = "Task" [ 721.506985] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 721.514111] env[65726]: DEBUG oslo_vmware.api [None req-b1ce1b66-ca58-46d0-b624-7c893de1adee tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Waiting for the task: (returnval){ [ 721.514111] env[65726]: value = "task-5115707" [ 721.514111] env[65726]: _type = "Task" [ 721.514111] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 721.521319] env[65726]: DEBUG oslo_vmware.api [None req-c24265b5-a0c6-4266-b4c0-64535a550ddd tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]523b5e05-1773-a061-306b-977bcc8a3aca, 'name': SearchDatastore_Task, 'duration_secs': 0.016653} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 721.522752] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c24265b5-a0c6-4266-b4c0-64535a550ddd tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 721.523091] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-c24265b5-a0c6-4266-b4c0-64535a550ddd tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: bd839fe1-8801-4ba1-9c23-288453258b66] Processing image b52362a3-ee8a-4cbf-b06f-513b0cc8f95c {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 721.523407] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c24265b5-a0c6-4266-b4c0-64535a550ddd tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 721.523595] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c24265b5-a0c6-4266-b4c0-64535a550ddd tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 721.523793] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-c24265b5-a0c6-4266-b4c0-64535a550ddd tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 721.524084] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-359e5865-96e0-4f9e-a756-c62cb86e076d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.527642] env[65726]: INFO nova.scheduler.client.report [None req-06465b67-a5cc-4aeb-80ed-00dcc8ab79e5 tempest-ServersTestJSON-1793637688 tempest-ServersTestJSON-1793637688-project-member] Deleted allocations for instance 88fb7235-cd0e-49dc-9d78-3583d5d3f528 [ 721.536303] env[65726]: DEBUG oslo_vmware.api [None req-b1ce1b66-ca58-46d0-b624-7c893de1adee tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Task: {'id': task-5115707, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 721.547871] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-c24265b5-a0c6-4266-b4c0-64535a550ddd tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 721.548862] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-c24265b5-a0c6-4266-b4c0-64535a550ddd tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65726) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 721.549324] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6e99a670-ecb6-4789-9c31-fbacada51b52 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.558136] env[65726]: DEBUG oslo_vmware.api [None req-c24265b5-a0c6-4266-b4c0-64535a550ddd tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Waiting for the task: (returnval){ [ 721.558136] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]529720a6-6eb1-af3e-7275-ce35b24dc5fc" [ 721.558136] env[65726]: _type = "Task" [ 721.558136] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 721.573570] env[65726]: DEBUG oslo_vmware.api [None req-c24265b5-a0c6-4266-b4c0-64535a550ddd tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]529720a6-6eb1-af3e-7275-ce35b24dc5fc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 721.661740] env[65726]: WARNING openstack [req-381a4cbc-a8b2-4387-a340-363cc9279fe4 req-d12fd3a7-f5cf-43e4-929b-53c5524490f7 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 721.662217] env[65726]: WARNING openstack [req-381a4cbc-a8b2-4387-a340-363cc9279fe4 req-d12fd3a7-f5cf-43e4-929b-53c5524490f7 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 721.703385] env[65726]: DEBUG oslo_concurrency.lockutils [req-2d361622-38c8-4d30-9e3a-bf1999f11d50 req-0283ff69-1ce8-4ee0-ab10-b3e3788dfc02 service nova] Releasing lock "refresh_cache-96acc5f8-b411-408e-929d-d1035cfd50db" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 721.703385] env[65726]: DEBUG nova.compute.manager [req-2d361622-38c8-4d30-9e3a-bf1999f11d50 req-0283ff69-1ce8-4ee0-ab10-b3e3788dfc02 service nova] [instance: 608cc0ea-3f6b-4b4f-83c1-01aa50999d63] Received event network-changed-e9f7132a-8d5a-47fe-afc6-d45bf3687ea4 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 721.703385] env[65726]: DEBUG nova.compute.manager [req-2d361622-38c8-4d30-9e3a-bf1999f11d50 req-0283ff69-1ce8-4ee0-ab10-b3e3788dfc02 service nova] [instance: 608cc0ea-3f6b-4b4f-83c1-01aa50999d63] Refreshing instance network info cache due to event network-changed-e9f7132a-8d5a-47fe-afc6-d45bf3687ea4. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 721.703385] env[65726]: DEBUG oslo_concurrency.lockutils [req-2d361622-38c8-4d30-9e3a-bf1999f11d50 req-0283ff69-1ce8-4ee0-ab10-b3e3788dfc02 service nova] Acquiring lock "refresh_cache-608cc0ea-3f6b-4b4f-83c1-01aa50999d63" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 721.703734] env[65726]: DEBUG oslo_concurrency.lockutils [req-2d361622-38c8-4d30-9e3a-bf1999f11d50 req-0283ff69-1ce8-4ee0-ab10-b3e3788dfc02 service nova] Acquired lock "refresh_cache-608cc0ea-3f6b-4b4f-83c1-01aa50999d63" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 721.703734] env[65726]: DEBUG nova.network.neutron [req-2d361622-38c8-4d30-9e3a-bf1999f11d50 req-0283ff69-1ce8-4ee0-ab10-b3e3788dfc02 service nova] [instance: 608cc0ea-3f6b-4b4f-83c1-01aa50999d63] Refreshing network info cache for port e9f7132a-8d5a-47fe-afc6-d45bf3687ea4 {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 721.750682] env[65726]: WARNING openstack [req-381a4cbc-a8b2-4387-a340-363cc9279fe4 req-d12fd3a7-f5cf-43e4-929b-53c5524490f7 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 721.750682] env[65726]: WARNING openstack [req-381a4cbc-a8b2-4387-a340-363cc9279fe4 req-d12fd3a7-f5cf-43e4-929b-53c5524490f7 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 721.773222] env[65726]: DEBUG oslo_vmware.api [None req-5acbf047-f1f8-4297-909c-a3d7a9b0fdd6 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Task: {'id': task-5115702, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.263933} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 721.773477] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-5acbf047-f1f8-4297-909c-a3d7a9b0fdd6 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Deleted the datastore file {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 721.773636] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-5acbf047-f1f8-4297-909c-a3d7a9b0fdd6 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] [instance: 815f056b-2d88-40dd-9ef7-a891ec8d92cf] Deleted contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 721.773770] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-5acbf047-f1f8-4297-909c-a3d7a9b0fdd6 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] [instance: 815f056b-2d88-40dd-9ef7-a891ec8d92cf] Instance destroyed {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 721.774019] env[65726]: INFO nova.compute.manager [None req-5acbf047-f1f8-4297-909c-a3d7a9b0fdd6 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] [instance: 815f056b-2d88-40dd-9ef7-a891ec8d92cf] Took 1.14 seconds to destroy the instance on the hypervisor. [ 721.774295] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-5acbf047-f1f8-4297-909c-a3d7a9b0fdd6 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 721.774500] env[65726]: DEBUG nova.compute.manager [-] [instance: 815f056b-2d88-40dd-9ef7-a891ec8d92cf] Deallocating network for instance {{(pid=65726) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 721.774607] env[65726]: DEBUG nova.network.neutron [-] [instance: 815f056b-2d88-40dd-9ef7-a891ec8d92cf] deallocate_for_instance() {{(pid=65726) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 721.775159] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 721.775442] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 721.827918] env[65726]: DEBUG oslo_vmware.api [None req-296fd45e-ee7a-4e44-912a-f3b5a22cbbf0 tempest-ServerExternalEventsTest-1845125219 tempest-ServerExternalEventsTest-1845125219-project-member] Task: {'id': task-5115699, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.516681} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 721.834603] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-296fd45e-ee7a-4e44-912a-f3b5a22cbbf0 tempest-ServerExternalEventsTest-1845125219 tempest-ServerExternalEventsTest-1845125219-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore2] 96acc5f8-b411-408e-929d-d1035cfd50db/96acc5f8-b411-408e-929d-d1035cfd50db.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 721.834603] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-296fd45e-ee7a-4e44-912a-f3b5a22cbbf0 tempest-ServerExternalEventsTest-1845125219 tempest-ServerExternalEventsTest-1845125219-project-member] [instance: 96acc5f8-b411-408e-929d-d1035cfd50db] Extending root virtual disk to 1048576 {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 721.834603] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-aed0a33d-1bc8-4d5a-8147-8e60c6d66c4a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.843402] env[65726]: DEBUG oslo_vmware.api [None req-296fd45e-ee7a-4e44-912a-f3b5a22cbbf0 tempest-ServerExternalEventsTest-1845125219 tempest-ServerExternalEventsTest-1845125219-project-member] Waiting for the task: (returnval){ [ 721.843402] env[65726]: value = "task-5115708" [ 721.843402] env[65726]: _type = "Task" [ 721.843402] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 721.851291] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 721.851555] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 721.864342] env[65726]: DEBUG oslo_vmware.api [None req-296fd45e-ee7a-4e44-912a-f3b5a22cbbf0 tempest-ServerExternalEventsTest-1845125219 tempest-ServerExternalEventsTest-1845125219-project-member] Task: {'id': task-5115708, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 721.896023] env[65726]: DEBUG nova.network.neutron [req-381a4cbc-a8b2-4387-a340-363cc9279fe4 req-d12fd3a7-f5cf-43e4-929b-53c5524490f7 service nova] [instance: bd839fe1-8801-4ba1-9c23-288453258b66] Updated VIF entry in instance network info cache for port 32d9913d-f98b-4e80-af5a-5f88083281b3. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 721.896023] env[65726]: DEBUG nova.network.neutron [req-381a4cbc-a8b2-4387-a340-363cc9279fe4 req-d12fd3a7-f5cf-43e4-929b-53c5524490f7 service nova] [instance: bd839fe1-8801-4ba1-9c23-288453258b66] Updating instance_info_cache with network_info: [{"id": "32d9913d-f98b-4e80-af5a-5f88083281b3", "address": "fa:16:3e:89:aa:99", "network": {"id": "e206b5b9-da36-4e6f-926f-914dec38fac2", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1053201655-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "39067912d091464192516c08eb576696", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7c80243e-93a7-4a95-bc8d-e9534bacd66e", "external-id": "nsx-vlan-transportzone-306", "segmentation_id": 306, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap32d9913d-f9", "ovs_interfaceid": "32d9913d-f98b-4e80-af5a-5f88083281b3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 721.974690] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115706, 'name': CreateVM_Task, 'duration_secs': 0.336598} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 721.974786] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 83573c2b-9448-456f-8fd2-b19661dd6cc4] Created VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 721.975268] env[65726]: DEBUG oslo_concurrency.lockutils [None req-2737dc3f-715d-481c-b9a9-998472382a0d tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 721.975426] env[65726]: DEBUG oslo_concurrency.lockutils [None req-2737dc3f-715d-481c-b9a9-998472382a0d tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 721.975808] env[65726]: DEBUG oslo_concurrency.lockutils [None req-2737dc3f-715d-481c-b9a9-998472382a0d tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 721.976715] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ca69ab48-99d4-4ab5-9100-aba1464adf9f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.985809] env[65726]: DEBUG oslo_vmware.api [None req-2737dc3f-715d-481c-b9a9-998472382a0d tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] Waiting for the task: (returnval){ [ 721.985809] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]521c9f32-f82a-5902-8ee8-9fc9d9755a60" [ 721.985809] env[65726]: _type = "Task" [ 721.985809] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 721.994573] env[65726]: DEBUG oslo_vmware.api [None req-2737dc3f-715d-481c-b9a9-998472382a0d tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]521c9f32-f82a-5902-8ee8-9fc9d9755a60, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 722.023430] env[65726]: DEBUG oslo_vmware.api [None req-b1ce1b66-ca58-46d0-b624-7c893de1adee tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Task: {'id': task-5115707, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.187666} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 722.024335] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-b1ce1b66-ca58-46d0-b624-7c893de1adee tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Deleted the datastore file {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 722.024535] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-b1ce1b66-ca58-46d0-b624-7c893de1adee tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] [instance: 423af936-a383-4652-8887-9c8b8f507909] Deleted contents of the VM from datastore datastore2 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 722.024677] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-b1ce1b66-ca58-46d0-b624-7c893de1adee tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] [instance: 423af936-a383-4652-8887-9c8b8f507909] Instance destroyed {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 722.024836] env[65726]: INFO nova.compute.manager [None req-b1ce1b66-ca58-46d0-b624-7c893de1adee tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] [instance: 423af936-a383-4652-8887-9c8b8f507909] Took 0.62 seconds to destroy the instance on the hypervisor. [ 722.025083] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-b1ce1b66-ca58-46d0-b624-7c893de1adee tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 722.026768] env[65726]: DEBUG nova.compute.manager [-] [instance: 423af936-a383-4652-8887-9c8b8f507909] Deallocating network for instance {{(pid=65726) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 722.026768] env[65726]: DEBUG nova.network.neutron [-] [instance: 423af936-a383-4652-8887-9c8b8f507909] deallocate_for_instance() {{(pid=65726) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 722.026768] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 722.026768] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 722.040178] env[65726]: DEBUG oslo_concurrency.lockutils [None req-06465b67-a5cc-4aeb-80ed-00dcc8ab79e5 tempest-ServersTestJSON-1793637688 tempest-ServersTestJSON-1793637688-project-member] Lock "88fb7235-cd0e-49dc-9d78-3583d5d3f528" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 42.627s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 722.073761] env[65726]: DEBUG oslo_vmware.api [None req-c24265b5-a0c6-4266-b4c0-64535a550ddd tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]529720a6-6eb1-af3e-7275-ce35b24dc5fc, 'name': SearchDatastore_Task, 'duration_secs': 0.01315} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 722.074595] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-12153534-26d4-4132-aa72-6a7c6a110589 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.082967] env[65726]: DEBUG oslo_vmware.api [None req-c24265b5-a0c6-4266-b4c0-64535a550ddd tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Waiting for the task: (returnval){ [ 722.082967] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52b42912-64b9-8b8f-5604-33c44848918d" [ 722.082967] env[65726]: _type = "Task" [ 722.082967] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 722.097527] env[65726]: DEBUG oslo_vmware.api [None req-c24265b5-a0c6-4266-b4c0-64535a550ddd tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52b42912-64b9-8b8f-5604-33c44848918d, 'name': SearchDatastore_Task, 'duration_secs': 0.011153} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 722.097839] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c24265b5-a0c6-4266-b4c0-64535a550ddd tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 722.098141] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-c24265b5-a0c6-4266-b4c0-64535a550ddd tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore2] bd839fe1-8801-4ba1-9c23-288453258b66/bd839fe1-8801-4ba1-9c23-288453258b66.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 722.098439] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a719c9cb-cc41-4d4a-9b4e-79b6dc9571de {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.107113] env[65726]: DEBUG oslo_vmware.api [None req-c24265b5-a0c6-4266-b4c0-64535a550ddd tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Waiting for the task: (returnval){ [ 722.107113] env[65726]: value = "task-5115709" [ 722.107113] env[65726]: _type = "Task" [ 722.107113] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 722.110236] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 722.110599] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 722.129678] env[65726]: DEBUG oslo_vmware.api [None req-c24265b5-a0c6-4266-b4c0-64535a550ddd tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': task-5115709, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 722.207261] env[65726]: WARNING openstack [req-2d361622-38c8-4d30-9e3a-bf1999f11d50 req-0283ff69-1ce8-4ee0-ab10-b3e3788dfc02 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 722.207696] env[65726]: WARNING openstack [req-2d361622-38c8-4d30-9e3a-bf1999f11d50 req-0283ff69-1ce8-4ee0-ab10-b3e3788dfc02 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 722.355243] env[65726]: DEBUG oslo_vmware.api [None req-296fd45e-ee7a-4e44-912a-f3b5a22cbbf0 tempest-ServerExternalEventsTest-1845125219 tempest-ServerExternalEventsTest-1845125219-project-member] Task: {'id': task-5115708, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069796} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 722.355637] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-296fd45e-ee7a-4e44-912a-f3b5a22cbbf0 tempest-ServerExternalEventsTest-1845125219 tempest-ServerExternalEventsTest-1845125219-project-member] [instance: 96acc5f8-b411-408e-929d-d1035cfd50db] Extended root virtual disk {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 722.358242] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab87db60-3567-4be5-8704-f581e6177a1e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.384866] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-296fd45e-ee7a-4e44-912a-f3b5a22cbbf0 tempest-ServerExternalEventsTest-1845125219 tempest-ServerExternalEventsTest-1845125219-project-member] [instance: 96acc5f8-b411-408e-929d-d1035cfd50db] Reconfiguring VM instance instance-00000020 to attach disk [datastore2] 96acc5f8-b411-408e-929d-d1035cfd50db/96acc5f8-b411-408e-929d-d1035cfd50db.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 722.386164] env[65726]: DEBUG nova.network.neutron [-] [instance: 423af936-a383-4652-8887-9c8b8f507909] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 722.389042] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ea92b93b-51ef-46f3-a2f0-5c1ae4d82715 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.404347] env[65726]: INFO nova.compute.manager [-] [instance: 423af936-a383-4652-8887-9c8b8f507909] Took 0.38 seconds to deallocate network for instance. [ 722.404855] env[65726]: DEBUG oslo_concurrency.lockutils [req-381a4cbc-a8b2-4387-a340-363cc9279fe4 req-d12fd3a7-f5cf-43e4-929b-53c5524490f7 service nova] Releasing lock "refresh_cache-bd839fe1-8801-4ba1-9c23-288453258b66" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 722.419600] env[65726]: DEBUG oslo_vmware.api [None req-296fd45e-ee7a-4e44-912a-f3b5a22cbbf0 tempest-ServerExternalEventsTest-1845125219 tempest-ServerExternalEventsTest-1845125219-project-member] Waiting for the task: (returnval){ [ 722.419600] env[65726]: value = "task-5115710" [ 722.419600] env[65726]: _type = "Task" [ 722.419600] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 722.431533] env[65726]: DEBUG oslo_vmware.api [None req-296fd45e-ee7a-4e44-912a-f3b5a22cbbf0 tempest-ServerExternalEventsTest-1845125219 tempest-ServerExternalEventsTest-1845125219-project-member] Task: {'id': task-5115710, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 722.502068] env[65726]: DEBUG oslo_vmware.api [None req-2737dc3f-715d-481c-b9a9-998472382a0d tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]521c9f32-f82a-5902-8ee8-9fc9d9755a60, 'name': SearchDatastore_Task, 'duration_secs': 0.010298} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 722.502068] env[65726]: DEBUG oslo_concurrency.lockutils [None req-2737dc3f-715d-481c-b9a9-998472382a0d tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 722.502068] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-2737dc3f-715d-481c-b9a9-998472382a0d tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] [instance: 83573c2b-9448-456f-8fd2-b19661dd6cc4] Processing image b52362a3-ee8a-4cbf-b06f-513b0cc8f95c {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 722.502068] env[65726]: DEBUG oslo_concurrency.lockutils [None req-2737dc3f-715d-481c-b9a9-998472382a0d tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 722.503930] env[65726]: DEBUG oslo_concurrency.lockutils [None req-2737dc3f-715d-481c-b9a9-998472382a0d tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 722.503930] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-2737dc3f-715d-481c-b9a9-998472382a0d tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 722.503930] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a19506b5-cd57-4c3d-9116-3777a9710f22 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.519171] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-2737dc3f-715d-481c-b9a9-998472382a0d tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 722.519959] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-2737dc3f-715d-481c-b9a9-998472382a0d tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65726) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 722.520904] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-30e4ac58-d343-4a88-9c77-ace6d3c410b8 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.527485] env[65726]: DEBUG oslo_vmware.api [None req-2737dc3f-715d-481c-b9a9-998472382a0d tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] Waiting for the task: (returnval){ [ 722.527485] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52357cc6-fd8d-fc4d-47a3-c7a67fef706b" [ 722.527485] env[65726]: _type = "Task" [ 722.527485] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 722.538544] env[65726]: DEBUG oslo_vmware.api [None req-2737dc3f-715d-481c-b9a9-998472382a0d tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52357cc6-fd8d-fc4d-47a3-c7a67fef706b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 722.553170] env[65726]: DEBUG nova.network.neutron [-] [instance: 815f056b-2d88-40dd-9ef7-a891ec8d92cf] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 722.634321] env[65726]: DEBUG oslo_vmware.api [None req-c24265b5-a0c6-4266-b4c0-64535a550ddd tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': task-5115709, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.475075} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 722.634606] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-c24265b5-a0c6-4266-b4c0-64535a550ddd tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore2] bd839fe1-8801-4ba1-9c23-288453258b66/bd839fe1-8801-4ba1-9c23-288453258b66.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 722.635248] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-c24265b5-a0c6-4266-b4c0-64535a550ddd tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: bd839fe1-8801-4ba1-9c23-288453258b66] Extending root virtual disk to 1048576 {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 722.635519] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-251cb697-0201-4f7b-9587-fcdefa95f43b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.643951] env[65726]: DEBUG oslo_vmware.api [None req-c24265b5-a0c6-4266-b4c0-64535a550ddd tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Waiting for the task: (returnval){ [ 722.643951] env[65726]: value = "task-5115711" [ 722.643951] env[65726]: _type = "Task" [ 722.643951] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 722.656855] env[65726]: DEBUG oslo_vmware.api [None req-c24265b5-a0c6-4266-b4c0-64535a550ddd tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': task-5115711, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 722.796272] env[65726]: DEBUG nova.compute.manager [req-cf89f264-84b1-4d5e-a432-f7da3d31d415 req-bcfd61c8-efb3-4d55-b27b-2f7a8a00d343 service nova] [instance: 815f056b-2d88-40dd-9ef7-a891ec8d92cf] Received event network-vif-deleted-1dfb06a2-5081-48fc-a9ef-23b637910832 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 722.917446] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b1ce1b66-ca58-46d0-b624-7c893de1adee tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 722.932047] env[65726]: DEBUG oslo_vmware.api [None req-296fd45e-ee7a-4e44-912a-f3b5a22cbbf0 tempest-ServerExternalEventsTest-1845125219 tempest-ServerExternalEventsTest-1845125219-project-member] Task: {'id': task-5115710, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 722.975270] env[65726]: WARNING openstack [req-2d361622-38c8-4d30-9e3a-bf1999f11d50 req-0283ff69-1ce8-4ee0-ab10-b3e3788dfc02 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 722.975641] env[65726]: WARNING openstack [req-2d361622-38c8-4d30-9e3a-bf1999f11d50 req-0283ff69-1ce8-4ee0-ab10-b3e3788dfc02 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 723.038228] env[65726]: DEBUG oslo_vmware.api [None req-2737dc3f-715d-481c-b9a9-998472382a0d tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52357cc6-fd8d-fc4d-47a3-c7a67fef706b, 'name': SearchDatastore_Task, 'duration_secs': 0.040272} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 723.039036] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6313c2f4-a8c5-40e3-b0e5-c9371a07c415 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.045665] env[65726]: DEBUG oslo_vmware.api [None req-2737dc3f-715d-481c-b9a9-998472382a0d tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] Waiting for the task: (returnval){ [ 723.045665] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52fadd57-8490-d211-a493-da86787a2978" [ 723.045665] env[65726]: _type = "Task" [ 723.045665] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 723.058410] env[65726]: INFO nova.compute.manager [-] [instance: 815f056b-2d88-40dd-9ef7-a891ec8d92cf] Took 1.28 seconds to deallocate network for instance. [ 723.058734] env[65726]: DEBUG oslo_vmware.api [None req-2737dc3f-715d-481c-b9a9-998472382a0d tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52fadd57-8490-d211-a493-da86787a2978, 'name': SearchDatastore_Task, 'duration_secs': 0.01031} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 723.063094] env[65726]: DEBUG oslo_concurrency.lockutils [None req-2737dc3f-715d-481c-b9a9-998472382a0d tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 723.063227] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-2737dc3f-715d-481c-b9a9-998472382a0d tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore2] 83573c2b-9448-456f-8fd2-b19661dd6cc4/83573c2b-9448-456f-8fd2-b19661dd6cc4.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 723.066152] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a9cccc2b-ed2c-45d0-b456-74934f79d37e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.074858] env[65726]: DEBUG oslo_vmware.api [None req-2737dc3f-715d-481c-b9a9-998472382a0d tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] Waiting for the task: (returnval){ [ 723.074858] env[65726]: value = "task-5115712" [ 723.074858] env[65726]: _type = "Task" [ 723.074858] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 723.090031] env[65726]: DEBUG oslo_vmware.api [None req-2737dc3f-715d-481c-b9a9-998472382a0d tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] Task: {'id': task-5115712, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 723.119755] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c64af09a-6774-420a-9224-d6c7fa8c39f1 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.129886] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce0cfc42-c68a-4d56-850c-db59430e2a90 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.175114] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44b76139-fe74-43d3-a8c5-da2717065d4f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.191982] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75b7e856-bd78-4dcf-a4c5-2a319f70d5ea {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.196659] env[65726]: DEBUG oslo_vmware.api [None req-c24265b5-a0c6-4266-b4c0-64535a550ddd tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': task-5115711, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.072177} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 723.196988] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-c24265b5-a0c6-4266-b4c0-64535a550ddd tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: bd839fe1-8801-4ba1-9c23-288453258b66] Extended root virtual disk {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 723.198525] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4522ae5-c99d-4aa7-8800-915f2ed93709 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.213213] env[65726]: DEBUG nova.compute.provider_tree [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 723.244367] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-c24265b5-a0c6-4266-b4c0-64535a550ddd tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: bd839fe1-8801-4ba1-9c23-288453258b66] Reconfiguring VM instance instance-00000021 to attach disk [datastore2] bd839fe1-8801-4ba1-9c23-288453258b66/bd839fe1-8801-4ba1-9c23-288453258b66.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 723.244760] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-73d10845-078b-4228-aaaa-ce1b6b5837b1 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.280390] env[65726]: DEBUG oslo_vmware.api [None req-c24265b5-a0c6-4266-b4c0-64535a550ddd tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Waiting for the task: (returnval){ [ 723.280390] env[65726]: value = "task-5115713" [ 723.280390] env[65726]: _type = "Task" [ 723.280390] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 723.293092] env[65726]: DEBUG oslo_vmware.api [None req-c24265b5-a0c6-4266-b4c0-64535a550ddd tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': task-5115713, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 723.434364] env[65726]: DEBUG oslo_vmware.api [None req-296fd45e-ee7a-4e44-912a-f3b5a22cbbf0 tempest-ServerExternalEventsTest-1845125219 tempest-ServerExternalEventsTest-1845125219-project-member] Task: {'id': task-5115710, 'name': ReconfigVM_Task, 'duration_secs': 0.866387} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 723.434732] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-296fd45e-ee7a-4e44-912a-f3b5a22cbbf0 tempest-ServerExternalEventsTest-1845125219 tempest-ServerExternalEventsTest-1845125219-project-member] [instance: 96acc5f8-b411-408e-929d-d1035cfd50db] Reconfigured VM instance instance-00000020 to attach disk [datastore2] 96acc5f8-b411-408e-929d-d1035cfd50db/96acc5f8-b411-408e-929d-d1035cfd50db.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 723.435479] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b05fe36e-332d-4e31-b209-c25fee005074 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.444468] env[65726]: DEBUG oslo_vmware.api [None req-296fd45e-ee7a-4e44-912a-f3b5a22cbbf0 tempest-ServerExternalEventsTest-1845125219 tempest-ServerExternalEventsTest-1845125219-project-member] Waiting for the task: (returnval){ [ 723.444468] env[65726]: value = "task-5115714" [ 723.444468] env[65726]: _type = "Task" [ 723.444468] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 723.456155] env[65726]: DEBUG oslo_vmware.api [None req-296fd45e-ee7a-4e44-912a-f3b5a22cbbf0 tempest-ServerExternalEventsTest-1845125219 tempest-ServerExternalEventsTest-1845125219-project-member] Task: {'id': task-5115714, 'name': Rename_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 723.570708] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5acbf047-f1f8-4297-909c-a3d7a9b0fdd6 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 723.587404] env[65726]: DEBUG oslo_vmware.api [None req-2737dc3f-715d-481c-b9a9-998472382a0d tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] Task: {'id': task-5115712, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.458796} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 723.587404] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-2737dc3f-715d-481c-b9a9-998472382a0d tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore2] 83573c2b-9448-456f-8fd2-b19661dd6cc4/83573c2b-9448-456f-8fd2-b19661dd6cc4.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 723.587404] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-2737dc3f-715d-481c-b9a9-998472382a0d tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] [instance: 83573c2b-9448-456f-8fd2-b19661dd6cc4] Extending root virtual disk to 1048576 {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 723.587404] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-6ee9bbad-8884-45db-9114-55c8d8b0a564 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.596710] env[65726]: DEBUG oslo_vmware.api [None req-2737dc3f-715d-481c-b9a9-998472382a0d tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] Waiting for the task: (returnval){ [ 723.596710] env[65726]: value = "task-5115715" [ 723.596710] env[65726]: _type = "Task" [ 723.596710] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 723.607167] env[65726]: DEBUG oslo_vmware.api [None req-2737dc3f-715d-481c-b9a9-998472382a0d tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] Task: {'id': task-5115715, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 723.689964] env[65726]: WARNING openstack [req-2d361622-38c8-4d30-9e3a-bf1999f11d50 req-0283ff69-1ce8-4ee0-ab10-b3e3788dfc02 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 723.689964] env[65726]: WARNING openstack [req-2d361622-38c8-4d30-9e3a-bf1999f11d50 req-0283ff69-1ce8-4ee0-ab10-b3e3788dfc02 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 723.721658] env[65726]: DEBUG nova.scheduler.client.report [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 723.772191] env[65726]: DEBUG nova.compute.manager [req-cc6b3baa-84dd-4eb4-a8dc-132d1aed04d3 req-281c9a9d-51b2-4d26-8299-05be7373811a service nova] [instance: 423af936-a383-4652-8887-9c8b8f507909] Received event network-vif-deleted-11b095b5-d149-4b9c-a363-5e8acbc8262d {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 723.793720] env[65726]: DEBUG oslo_vmware.api [None req-c24265b5-a0c6-4266-b4c0-64535a550ddd tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': task-5115713, 'name': ReconfigVM_Task, 'duration_secs': 0.371489} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 723.794084] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-c24265b5-a0c6-4266-b4c0-64535a550ddd tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: bd839fe1-8801-4ba1-9c23-288453258b66] Reconfigured VM instance instance-00000021 to attach disk [datastore2] bd839fe1-8801-4ba1-9c23-288453258b66/bd839fe1-8801-4ba1-9c23-288453258b66.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 723.794776] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d6c3d5c5-2b1f-4e13-b7bd-d8786a1c3500 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.803947] env[65726]: DEBUG oslo_vmware.api [None req-c24265b5-a0c6-4266-b4c0-64535a550ddd tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Waiting for the task: (returnval){ [ 723.803947] env[65726]: value = "task-5115716" [ 723.803947] env[65726]: _type = "Task" [ 723.803947] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 723.817919] env[65726]: DEBUG oslo_vmware.api [None req-c24265b5-a0c6-4266-b4c0-64535a550ddd tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': task-5115716, 'name': Rename_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 723.961619] env[65726]: DEBUG oslo_vmware.api [None req-296fd45e-ee7a-4e44-912a-f3b5a22cbbf0 tempest-ServerExternalEventsTest-1845125219 tempest-ServerExternalEventsTest-1845125219-project-member] Task: {'id': task-5115714, 'name': Rename_Task, 'duration_secs': 0.306899} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 723.961619] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-296fd45e-ee7a-4e44-912a-f3b5a22cbbf0 tempest-ServerExternalEventsTest-1845125219 tempest-ServerExternalEventsTest-1845125219-project-member] [instance: 96acc5f8-b411-408e-929d-d1035cfd50db] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 723.961619] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-33189ade-9485-42f7-99e4-a293a34eaa71 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.970109] env[65726]: DEBUG oslo_vmware.api [None req-296fd45e-ee7a-4e44-912a-f3b5a22cbbf0 tempest-ServerExternalEventsTest-1845125219 tempest-ServerExternalEventsTest-1845125219-project-member] Waiting for the task: (returnval){ [ 723.970109] env[65726]: value = "task-5115717" [ 723.970109] env[65726]: _type = "Task" [ 723.970109] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 723.981301] env[65726]: DEBUG oslo_vmware.api [None req-296fd45e-ee7a-4e44-912a-f3b5a22cbbf0 tempest-ServerExternalEventsTest-1845125219 tempest-ServerExternalEventsTest-1845125219-project-member] Task: {'id': task-5115717, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 724.110558] env[65726]: DEBUG oslo_vmware.api [None req-2737dc3f-715d-481c-b9a9-998472382a0d tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] Task: {'id': task-5115715, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.073823} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 724.111341] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-2737dc3f-715d-481c-b9a9-998472382a0d tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] [instance: 83573c2b-9448-456f-8fd2-b19661dd6cc4] Extended root virtual disk {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 724.113058] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab8a710b-0f85-42f7-9fa4-f483aaff25a7 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.144328] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-2737dc3f-715d-481c-b9a9-998472382a0d tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] [instance: 83573c2b-9448-456f-8fd2-b19661dd6cc4] Reconfiguring VM instance instance-00000022 to attach disk [datastore2] 83573c2b-9448-456f-8fd2-b19661dd6cc4/83573c2b-9448-456f-8fd2-b19661dd6cc4.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 724.144328] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e5d2ad55-c2b3-49e9-b7d4-83bd80f6f3cf {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.163671] env[65726]: DEBUG oslo_vmware.api [None req-2737dc3f-715d-481c-b9a9-998472382a0d tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] Waiting for the task: (returnval){ [ 724.163671] env[65726]: value = "task-5115718" [ 724.163671] env[65726]: _type = "Task" [ 724.163671] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 724.177615] env[65726]: DEBUG oslo_vmware.api [None req-2737dc3f-715d-481c-b9a9-998472382a0d tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] Task: {'id': task-5115718, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 724.224882] env[65726]: DEBUG oslo_concurrency.lockutils [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.725s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 724.225484] env[65726]: DEBUG nova.compute.manager [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] [instance: 02351120-8385-4403-8464-a154f3da9380] Start building networks asynchronously for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 724.228126] env[65726]: DEBUG oslo_concurrency.lockutils [None req-78fe5da2-df9b-4290-851e-bd53c863c8dd tempest-ServerGroupTestJSON-125182767 tempest-ServerGroupTestJSON-125182767-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 36.299s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 724.228360] env[65726]: DEBUG nova.objects.instance [None req-78fe5da2-df9b-4290-851e-bd53c863c8dd tempest-ServerGroupTestJSON-125182767 tempest-ServerGroupTestJSON-125182767-project-member] Lazy-loading 'resources' on Instance uuid c2929727-a7ea-4f63-997d-474c29a305fc {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 724.318598] env[65726]: DEBUG oslo_vmware.api [None req-c24265b5-a0c6-4266-b4c0-64535a550ddd tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': task-5115716, 'name': Rename_Task, 'duration_secs': 0.272274} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 724.318979] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-c24265b5-a0c6-4266-b4c0-64535a550ddd tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: bd839fe1-8801-4ba1-9c23-288453258b66] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 724.319321] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3a7c9e64-335b-485a-ac9b-5245bd7809c7 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.326825] env[65726]: DEBUG oslo_vmware.api [None req-c24265b5-a0c6-4266-b4c0-64535a550ddd tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Waiting for the task: (returnval){ [ 724.326825] env[65726]: value = "task-5115719" [ 724.326825] env[65726]: _type = "Task" [ 724.326825] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 724.337322] env[65726]: DEBUG oslo_vmware.api [None req-c24265b5-a0c6-4266-b4c0-64535a550ddd tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': task-5115719, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 724.378768] env[65726]: DEBUG nova.network.neutron [req-2d361622-38c8-4d30-9e3a-bf1999f11d50 req-0283ff69-1ce8-4ee0-ab10-b3e3788dfc02 service nova] [instance: 608cc0ea-3f6b-4b4f-83c1-01aa50999d63] Updated VIF entry in instance network info cache for port e9f7132a-8d5a-47fe-afc6-d45bf3687ea4. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 724.379295] env[65726]: DEBUG nova.network.neutron [req-2d361622-38c8-4d30-9e3a-bf1999f11d50 req-0283ff69-1ce8-4ee0-ab10-b3e3788dfc02 service nova] [instance: 608cc0ea-3f6b-4b4f-83c1-01aa50999d63] Updating instance_info_cache with network_info: [{"id": "e9f7132a-8d5a-47fe-afc6-d45bf3687ea4", "address": "fa:16:3e:fe:4d:df", "network": {"id": "0d0f561d-ac41-4ef5-9926-c707409799f7", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1134321217-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.160", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4eece77569624f90bf64e5c51974173f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2ce62383-8e84-4e26-955b-74c11392f4c9", "external-id": "nsx-vlan-transportzone-215", "segmentation_id": 215, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape9f7132a-8d", "ovs_interfaceid": "e9f7132a-8d5a-47fe-afc6-d45bf3687ea4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 724.485416] env[65726]: DEBUG oslo_vmware.api [None req-296fd45e-ee7a-4e44-912a-f3b5a22cbbf0 tempest-ServerExternalEventsTest-1845125219 tempest-ServerExternalEventsTest-1845125219-project-member] Task: {'id': task-5115717, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 724.682094] env[65726]: DEBUG oslo_vmware.api [None req-2737dc3f-715d-481c-b9a9-998472382a0d tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] Task: {'id': task-5115718, 'name': ReconfigVM_Task, 'duration_secs': 0.440212} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 724.682461] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-2737dc3f-715d-481c-b9a9-998472382a0d tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] [instance: 83573c2b-9448-456f-8fd2-b19661dd6cc4] Reconfigured VM instance instance-00000022 to attach disk [datastore2] 83573c2b-9448-456f-8fd2-b19661dd6cc4/83573c2b-9448-456f-8fd2-b19661dd6cc4.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 724.683172] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-19c1942b-2a5e-4900-9cf8-3e0ccd78ce05 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.693555] env[65726]: DEBUG oslo_vmware.api [None req-2737dc3f-715d-481c-b9a9-998472382a0d tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] Waiting for the task: (returnval){ [ 724.693555] env[65726]: value = "task-5115720" [ 724.693555] env[65726]: _type = "Task" [ 724.693555] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 724.703392] env[65726]: DEBUG oslo_vmware.api [None req-2737dc3f-715d-481c-b9a9-998472382a0d tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] Task: {'id': task-5115720, 'name': Rename_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 724.740085] env[65726]: DEBUG nova.compute.utils [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Using /dev/sd instead of None {{(pid=65726) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 724.742371] env[65726]: DEBUG nova.compute.manager [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] [instance: 02351120-8385-4403-8464-a154f3da9380] Allocating IP information in the background. {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 724.742633] env[65726]: DEBUG nova.network.neutron [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] [instance: 02351120-8385-4403-8464-a154f3da9380] allocate_for_instance() {{(pid=65726) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 724.743912] env[65726]: WARNING neutronclient.v2_0.client [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 724.743912] env[65726]: WARNING neutronclient.v2_0.client [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 724.744076] env[65726]: WARNING openstack [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 724.744505] env[65726]: WARNING openstack [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 724.755625] env[65726]: DEBUG nova.compute.manager [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] [instance: 02351120-8385-4403-8464-a154f3da9380] Start building block device mappings for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 724.816039] env[65726]: DEBUG nova.policy [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '838543f31e974acf83678c7f5c27b102', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7c3641c4a34b4bf8b09403220e153e65', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65726) authorize /opt/stack/nova/nova/policy.py:192}} [ 724.840561] env[65726]: DEBUG oslo_vmware.api [None req-c24265b5-a0c6-4266-b4c0-64535a550ddd tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': task-5115719, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 724.885185] env[65726]: DEBUG oslo_concurrency.lockutils [req-2d361622-38c8-4d30-9e3a-bf1999f11d50 req-0283ff69-1ce8-4ee0-ab10-b3e3788dfc02 service nova] Releasing lock "refresh_cache-608cc0ea-3f6b-4b4f-83c1-01aa50999d63" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 724.988310] env[65726]: DEBUG oslo_vmware.api [None req-296fd45e-ee7a-4e44-912a-f3b5a22cbbf0 tempest-ServerExternalEventsTest-1845125219 tempest-ServerExternalEventsTest-1845125219-project-member] Task: {'id': task-5115717, 'name': PowerOnVM_Task, 'duration_secs': 0.645136} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 724.988569] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-296fd45e-ee7a-4e44-912a-f3b5a22cbbf0 tempest-ServerExternalEventsTest-1845125219 tempest-ServerExternalEventsTest-1845125219-project-member] [instance: 96acc5f8-b411-408e-929d-d1035cfd50db] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 724.988867] env[65726]: INFO nova.compute.manager [None req-296fd45e-ee7a-4e44-912a-f3b5a22cbbf0 tempest-ServerExternalEventsTest-1845125219 tempest-ServerExternalEventsTest-1845125219-project-member] [instance: 96acc5f8-b411-408e-929d-d1035cfd50db] Took 9.73 seconds to spawn the instance on the hypervisor. [ 724.989113] env[65726]: DEBUG nova.compute.manager [None req-296fd45e-ee7a-4e44-912a-f3b5a22cbbf0 tempest-ServerExternalEventsTest-1845125219 tempest-ServerExternalEventsTest-1845125219-project-member] [instance: 96acc5f8-b411-408e-929d-d1035cfd50db] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 724.989998] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6478bbff-0320-4747-afd4-01241c47ca9e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.147306] env[65726]: DEBUG nova.network.neutron [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] [instance: 02351120-8385-4403-8464-a154f3da9380] Successfully created port: f9ef8504-78db-4da2-abc7-94b9a07a014e {{(pid=65726) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 725.204655] env[65726]: DEBUG oslo_vmware.api [None req-2737dc3f-715d-481c-b9a9-998472382a0d tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] Task: {'id': task-5115720, 'name': Rename_Task, 'duration_secs': 0.354548} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 725.207366] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-2737dc3f-715d-481c-b9a9-998472382a0d tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] [instance: 83573c2b-9448-456f-8fd2-b19661dd6cc4] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 725.207827] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-15e0caa4-f6dd-4263-8576-43664aa26319 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.215337] env[65726]: DEBUG oslo_vmware.api [None req-2737dc3f-715d-481c-b9a9-998472382a0d tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] Waiting for the task: (returnval){ [ 725.215337] env[65726]: value = "task-5115721" [ 725.215337] env[65726]: _type = "Task" [ 725.215337] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 725.228484] env[65726]: DEBUG oslo_vmware.api [None req-2737dc3f-715d-481c-b9a9-998472382a0d tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] Task: {'id': task-5115721, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 725.342727] env[65726]: DEBUG oslo_vmware.api [None req-c24265b5-a0c6-4266-b4c0-64535a550ddd tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': task-5115719, 'name': PowerOnVM_Task, 'duration_secs': 0.541023} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 725.343080] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-c24265b5-a0c6-4266-b4c0-64535a550ddd tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: bd839fe1-8801-4ba1-9c23-288453258b66] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 725.343315] env[65726]: INFO nova.compute.manager [None req-c24265b5-a0c6-4266-b4c0-64535a550ddd tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: bd839fe1-8801-4ba1-9c23-288453258b66] Took 6.71 seconds to spawn the instance on the hypervisor. [ 725.343515] env[65726]: DEBUG nova.compute.manager [None req-c24265b5-a0c6-4266-b4c0-64535a550ddd tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: bd839fe1-8801-4ba1-9c23-288453258b66] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 725.344345] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7dcc2940-d7a4-48a7-83a1-2dbbae7a0a25 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.486239] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a214487-75a8-47f3-b72a-22d32c83942c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.496117] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0b78a55-40a0-4919-8a03-f34be294eb7f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.538070] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbfbf9f2-9e4a-46ed-850a-85c4a7c10dbd {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.541298] env[65726]: INFO nova.compute.manager [None req-296fd45e-ee7a-4e44-912a-f3b5a22cbbf0 tempest-ServerExternalEventsTest-1845125219 tempest-ServerExternalEventsTest-1845125219-project-member] [instance: 96acc5f8-b411-408e-929d-d1035cfd50db] Took 49.44 seconds to build instance. [ 725.548315] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c394dac1-2f93-403d-ae4b-87f42235091b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.564015] env[65726]: DEBUG nova.compute.provider_tree [None req-78fe5da2-df9b-4290-851e-bd53c863c8dd tempest-ServerGroupTestJSON-125182767 tempest-ServerGroupTestJSON-125182767-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 725.729508] env[65726]: DEBUG oslo_vmware.api [None req-2737dc3f-715d-481c-b9a9-998472382a0d tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] Task: {'id': task-5115721, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 725.766986] env[65726]: DEBUG nova.compute.manager [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] [instance: 02351120-8385-4403-8464-a154f3da9380] Start spawning the instance on the hypervisor. {{(pid=65726) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 725.795245] env[65726]: DEBUG nova.virt.hardware [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T19:28:34Z,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1539f06fa3534e90acbc3a60c4b8bd3f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T19:28:35Z,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 725.795514] env[65726]: DEBUG nova.virt.hardware [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 725.795675] env[65726]: DEBUG nova.virt.hardware [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 725.795861] env[65726]: DEBUG nova.virt.hardware [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 725.796018] env[65726]: DEBUG nova.virt.hardware [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 725.796165] env[65726]: DEBUG nova.virt.hardware [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 725.796377] env[65726]: DEBUG nova.virt.hardware [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 725.796569] env[65726]: DEBUG nova.virt.hardware [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 725.796819] env[65726]: DEBUG nova.virt.hardware [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 725.797010] env[65726]: DEBUG nova.virt.hardware [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 725.797194] env[65726]: DEBUG nova.virt.hardware [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 725.798104] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c64bbc68-93bb-498b-afe6-bd6bd6ce0544 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.807440] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cf0eadd-8c8e-4c82-ab2f-367d15e569de {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.874583] env[65726]: INFO nova.compute.manager [None req-c24265b5-a0c6-4266-b4c0-64535a550ddd tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: bd839fe1-8801-4ba1-9c23-288453258b66] Took 45.55 seconds to build instance. [ 726.043473] env[65726]: DEBUG oslo_concurrency.lockutils [None req-296fd45e-ee7a-4e44-912a-f3b5a22cbbf0 tempest-ServerExternalEventsTest-1845125219 tempest-ServerExternalEventsTest-1845125219-project-member] Lock "96acc5f8-b411-408e-929d-d1035cfd50db" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 71.515s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 726.068235] env[65726]: DEBUG nova.scheduler.client.report [None req-78fe5da2-df9b-4290-851e-bd53c863c8dd tempest-ServerGroupTestJSON-125182767 tempest-ServerGroupTestJSON-125182767-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 726.225955] env[65726]: DEBUG oslo_vmware.api [None req-2737dc3f-715d-481c-b9a9-998472382a0d tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] Task: {'id': task-5115721, 'name': PowerOnVM_Task, 'duration_secs': 1.009908} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 726.226286] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-2737dc3f-715d-481c-b9a9-998472382a0d tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] [instance: 83573c2b-9448-456f-8fd2-b19661dd6cc4] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 726.226552] env[65726]: INFO nova.compute.manager [None req-2737dc3f-715d-481c-b9a9-998472382a0d tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] [instance: 83573c2b-9448-456f-8fd2-b19661dd6cc4] Took 4.89 seconds to spawn the instance on the hypervisor. [ 726.226761] env[65726]: DEBUG nova.compute.manager [None req-2737dc3f-715d-481c-b9a9-998472382a0d tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] [instance: 83573c2b-9448-456f-8fd2-b19661dd6cc4] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 726.227613] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf311b30-240c-482a-b3a0-950e60b40913 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.376678] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c24265b5-a0c6-4266-b4c0-64535a550ddd tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Lock "bd839fe1-8801-4ba1-9c23-288453258b66" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 61.727s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 726.546715] env[65726]: DEBUG nova.compute.manager [None req-48153719-e51d-407a-98ef-93401c6270f2 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] [instance: 65aaa2eb-3ca5-4c2b-92f9-b524366c10dd] Starting instance... {{(pid=65726) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 726.573571] env[65726]: DEBUG oslo_concurrency.lockutils [None req-78fe5da2-df9b-4290-851e-bd53c863c8dd tempest-ServerGroupTestJSON-125182767 tempest-ServerGroupTestJSON-125182767-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.345s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 726.576104] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 34.585s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 726.608215] env[65726]: INFO nova.scheduler.client.report [None req-78fe5da2-df9b-4290-851e-bd53c863c8dd tempest-ServerGroupTestJSON-125182767 tempest-ServerGroupTestJSON-125182767-project-member] Deleted allocations for instance c2929727-a7ea-4f63-997d-474c29a305fc [ 726.747885] env[65726]: INFO nova.compute.manager [None req-2737dc3f-715d-481c-b9a9-998472382a0d tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] [instance: 83573c2b-9448-456f-8fd2-b19661dd6cc4] Took 45.35 seconds to build instance. [ 726.813019] env[65726]: DEBUG nova.network.neutron [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] [instance: 02351120-8385-4403-8464-a154f3da9380] Successfully updated port: f9ef8504-78db-4da2-abc7-94b9a07a014e {{(pid=65726) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 726.880065] env[65726]: DEBUG nova.compute.manager [None req-6c979061-2d03-4525-8b87-afa801f4542e tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: 6d2b86c2-0a44-433f-8b3f-2e9c7f693baa] Starting instance... {{(pid=65726) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 727.067932] env[65726]: DEBUG oslo_concurrency.lockutils [None req-48153719-e51d-407a-98ef-93401c6270f2 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 727.120377] env[65726]: DEBUG oslo_concurrency.lockutils [None req-78fe5da2-df9b-4290-851e-bd53c863c8dd tempest-ServerGroupTestJSON-125182767 tempest-ServerGroupTestJSON-125182767-project-member] Lock "c2929727-a7ea-4f63-997d-474c29a305fc" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 42.851s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 727.251154] env[65726]: DEBUG oslo_concurrency.lockutils [None req-2737dc3f-715d-481c-b9a9-998472382a0d tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] Lock "83573c2b-9448-456f-8fd2-b19661dd6cc4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 60.639s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 727.318027] env[65726]: DEBUG oslo_concurrency.lockutils [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Acquiring lock "refresh_cache-02351120-8385-4403-8464-a154f3da9380" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 727.318526] env[65726]: DEBUG oslo_concurrency.lockutils [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Acquired lock "refresh_cache-02351120-8385-4403-8464-a154f3da9380" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 727.318928] env[65726]: DEBUG nova.network.neutron [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] [instance: 02351120-8385-4403-8464-a154f3da9380] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 727.418540] env[65726]: DEBUG oslo_concurrency.lockutils [None req-6c979061-2d03-4525-8b87-afa801f4542e tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 727.634147] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Instance 703fff2d-5aec-4ed1-b987-9e1cc3a25f67 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65726) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 727.636141] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Instance ee676e3b-3326-46a3-940d-ebbb8b108991 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65726) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 727.636141] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Instance cd539d8b-921b-4947-954c-8573380dbdc8 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65726) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 727.636141] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Instance aa2716a9-2cf8-4e51-800d-1c8f98b2cf05 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65726) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 727.636141] env[65726]: WARNING nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Instance 04ffdf80-c54c-4587-9bb4-d520dc440501 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 727.636302] env[65726]: WARNING nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Instance 3a60b719-9db7-436e-9908-25f6a233c465 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 727.636302] env[65726]: WARNING nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Instance f5151062-57b1-4e4e-93f4-aab0e7f504d4 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 727.636302] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Instance f2d9090c-988f-43f4-9c81-7aa718a3438a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65726) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 727.636302] env[65726]: WARNING nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Instance 3b933684-ac19-44b0-a49d-6af45501e166 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 727.636302] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Instance 274ab469-61a9-4b7e-852c-074c871e3abf actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65726) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 727.636518] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Instance 5a8d4357-4b44-4a19-b1da-42d188c38adc actively managed on this compute host and has allocations in placement: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65726) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 727.636518] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Instance fd481728-3ef0-4a96-affd-ab2dd3f596bb actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65726) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 727.636518] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Instance 0e064341-4e4a-407b-8c26-3eb04b409029 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=65726) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 727.636518] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Instance d7905c4f-1ba3-43b3-b4b3-aebacd2f9cc3 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65726) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 727.636955] env[65726]: WARNING nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Instance ddd36db0-28ad-4582-994b-7eef8b08ed46 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 727.637225] env[65726]: WARNING nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Instance c977257e-dce0-4ecc-b42b-0f6ebab38797 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 727.637445] env[65726]: WARNING nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Instance 30b17a17-229c-4ca7-9ae6-c67ccdcd3963 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 727.637676] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Instance 49bcab1f-7b64-4999-abff-37771c58a271 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65726) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 727.637893] env[65726]: WARNING nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Instance 423af936-a383-4652-8887-9c8b8f507909 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 727.638137] env[65726]: WARNING nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Instance 815f056b-2d88-40dd-9ef7-a891ec8d92cf is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 727.638564] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Instance 608cc0ea-3f6b-4b4f-83c1-01aa50999d63 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65726) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 727.639151] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Instance 96acc5f8-b411-408e-929d-d1035cfd50db actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65726) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 727.639151] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Instance bd839fe1-8801-4ba1-9c23-288453258b66 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65726) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 727.639151] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Instance 83573c2b-9448-456f-8fd2-b19661dd6cc4 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65726) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 727.639151] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Instance 02351120-8385-4403-8464-a154f3da9380 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65726) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 727.753517] env[65726]: DEBUG nova.compute.manager [None req-d51d38e8-71be-4889-a206-d52403737076 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: 9fc0c9ee-9dc7-4059-b8d2-a5db965e00c2] Starting instance... {{(pid=65726) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 727.783300] env[65726]: DEBUG nova.compute.manager [req-b29a62c6-95ec-4e5f-8743-220ad3f87c37 req-7396d989-874e-401c-98ef-e70d561973f7 service nova] [instance: 02351120-8385-4403-8464-a154f3da9380] Received event network-vif-plugged-f9ef8504-78db-4da2-abc7-94b9a07a014e {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 727.783561] env[65726]: DEBUG oslo_concurrency.lockutils [req-b29a62c6-95ec-4e5f-8743-220ad3f87c37 req-7396d989-874e-401c-98ef-e70d561973f7 service nova] Acquiring lock "02351120-8385-4403-8464-a154f3da9380-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 727.783727] env[65726]: DEBUG oslo_concurrency.lockutils [req-b29a62c6-95ec-4e5f-8743-220ad3f87c37 req-7396d989-874e-401c-98ef-e70d561973f7 service nova] Lock "02351120-8385-4403-8464-a154f3da9380-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 727.783887] env[65726]: DEBUG oslo_concurrency.lockutils [req-b29a62c6-95ec-4e5f-8743-220ad3f87c37 req-7396d989-874e-401c-98ef-e70d561973f7 service nova] Lock "02351120-8385-4403-8464-a154f3da9380-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 727.784282] env[65726]: DEBUG nova.compute.manager [req-b29a62c6-95ec-4e5f-8743-220ad3f87c37 req-7396d989-874e-401c-98ef-e70d561973f7 service nova] [instance: 02351120-8385-4403-8464-a154f3da9380] No waiting events found dispatching network-vif-plugged-f9ef8504-78db-4da2-abc7-94b9a07a014e {{(pid=65726) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 727.784512] env[65726]: WARNING nova.compute.manager [req-b29a62c6-95ec-4e5f-8743-220ad3f87c37 req-7396d989-874e-401c-98ef-e70d561973f7 service nova] [instance: 02351120-8385-4403-8464-a154f3da9380] Received unexpected event network-vif-plugged-f9ef8504-78db-4da2-abc7-94b9a07a014e for instance with vm_state building and task_state spawning. [ 727.823278] env[65726]: WARNING openstack [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 727.823662] env[65726]: WARNING openstack [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 727.829182] env[65726]: DEBUG nova.network.neutron [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] [instance: 02351120-8385-4403-8464-a154f3da9380] Instance cache missing network info. {{(pid=65726) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 727.887486] env[65726]: WARNING openstack [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 727.887870] env[65726]: WARNING openstack [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 727.957164] env[65726]: WARNING openstack [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 727.957715] env[65726]: WARNING openstack [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 727.993854] env[65726]: DEBUG nova.compute.manager [None req-3b5c8bda-39bb-468a-bdce-9c81be5debce tempest-ServerExternalEventsTest-1019349750 tempest-ServerExternalEventsTest-1019349750-project] [instance: 96acc5f8-b411-408e-929d-d1035cfd50db] Received event network-changed {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 727.994090] env[65726]: DEBUG nova.compute.manager [None req-3b5c8bda-39bb-468a-bdce-9c81be5debce tempest-ServerExternalEventsTest-1019349750 tempest-ServerExternalEventsTest-1019349750-project] [instance: 96acc5f8-b411-408e-929d-d1035cfd50db] Refreshing instance network info cache due to event network-changed. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 727.995038] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3b5c8bda-39bb-468a-bdce-9c81be5debce tempest-ServerExternalEventsTest-1019349750 tempest-ServerExternalEventsTest-1019349750-project] Acquiring lock "refresh_cache-96acc5f8-b411-408e-929d-d1035cfd50db" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 727.995038] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3b5c8bda-39bb-468a-bdce-9c81be5debce tempest-ServerExternalEventsTest-1019349750 tempest-ServerExternalEventsTest-1019349750-project] Acquired lock "refresh_cache-96acc5f8-b411-408e-929d-d1035cfd50db" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 727.996574] env[65726]: DEBUG nova.network.neutron [None req-3b5c8bda-39bb-468a-bdce-9c81be5debce tempest-ServerExternalEventsTest-1019349750 tempest-ServerExternalEventsTest-1019349750-project] [instance: 96acc5f8-b411-408e-929d-d1035cfd50db] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 728.070874] env[65726]: DEBUG nova.network.neutron [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] [instance: 02351120-8385-4403-8464-a154f3da9380] Updating instance_info_cache with network_info: [{"id": "f9ef8504-78db-4da2-abc7-94b9a07a014e", "address": "fa:16:3e:ce:88:da", "network": {"id": "c68b8bcc-840e-459b-bee0-76a9ffdc08cb", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-2061916615-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7c3641c4a34b4bf8b09403220e153e65", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "aa0d7fd6-4fd3-4451-9208-8a3cf25e30e4", "external-id": "nsx-vlan-transportzone-545", "segmentation_id": 545, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf9ef8504-78", "ovs_interfaceid": "f9ef8504-78db-4da2-abc7-94b9a07a014e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 728.141922] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Instance ca67d1a4-d304-4d06-a436-f257b60ca080 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65726) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 728.282554] env[65726]: DEBUG oslo_concurrency.lockutils [None req-d51d38e8-71be-4889-a206-d52403737076 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 728.403480] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32fa6e49-9925-4095-8a1e-8ccec0efc3ac {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.411116] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-36ef78ec-0b05-4841-bd7b-2242493adf78 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: bd839fe1-8801-4ba1-9c23-288453258b66] Suspending the VM {{(pid=65726) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1162}} [ 728.411386] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-9ab30be4-a432-4e52-adeb-8a57555efd17 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.417688] env[65726]: DEBUG oslo_vmware.api [None req-36ef78ec-0b05-4841-bd7b-2242493adf78 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Waiting for the task: (returnval){ [ 728.417688] env[65726]: value = "task-5115722" [ 728.417688] env[65726]: _type = "Task" [ 728.417688] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 728.427173] env[65726]: DEBUG oslo_vmware.api [None req-36ef78ec-0b05-4841-bd7b-2242493adf78 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': task-5115722, 'name': SuspendVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 728.503044] env[65726]: WARNING openstack [None req-3b5c8bda-39bb-468a-bdce-9c81be5debce tempest-ServerExternalEventsTest-1019349750 tempest-ServerExternalEventsTest-1019349750-project] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 728.503044] env[65726]: WARNING openstack [None req-3b5c8bda-39bb-468a-bdce-9c81be5debce tempest-ServerExternalEventsTest-1019349750 tempest-ServerExternalEventsTest-1019349750-project] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 728.577689] env[65726]: DEBUG oslo_concurrency.lockutils [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Releasing lock "refresh_cache-02351120-8385-4403-8464-a154f3da9380" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 728.577689] env[65726]: DEBUG nova.compute.manager [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] [instance: 02351120-8385-4403-8464-a154f3da9380] Instance network_info: |[{"id": "f9ef8504-78db-4da2-abc7-94b9a07a014e", "address": "fa:16:3e:ce:88:da", "network": {"id": "c68b8bcc-840e-459b-bee0-76a9ffdc08cb", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-2061916615-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7c3641c4a34b4bf8b09403220e153e65", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "aa0d7fd6-4fd3-4451-9208-8a3cf25e30e4", "external-id": "nsx-vlan-transportzone-545", "segmentation_id": 545, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf9ef8504-78", "ovs_interfaceid": "f9ef8504-78db-4da2-abc7-94b9a07a014e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 728.580069] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] [instance: 02351120-8385-4403-8464-a154f3da9380] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ce:88:da', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'aa0d7fd6-4fd3-4451-9208-8a3cf25e30e4', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f9ef8504-78db-4da2-abc7-94b9a07a014e', 'vif_model': 'vmxnet3'}] {{(pid=65726) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 728.590301] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 728.590602] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 02351120-8385-4403-8464-a154f3da9380] Creating VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 728.591511] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2da3e9a1-ebd7-4787-aefa-e3140668e660 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.623020] env[65726]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 728.623020] env[65726]: value = "task-5115723" [ 728.623020] env[65726]: _type = "Task" [ 728.623020] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 728.633294] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115723, 'name': CreateVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 728.639972] env[65726]: WARNING openstack [None req-3b5c8bda-39bb-468a-bdce-9c81be5debce tempest-ServerExternalEventsTest-1019349750 tempest-ServerExternalEventsTest-1019349750-project] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 728.640766] env[65726]: WARNING openstack [None req-3b5c8bda-39bb-468a-bdce-9c81be5debce tempest-ServerExternalEventsTest-1019349750 tempest-ServerExternalEventsTest-1019349750-project] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 728.654231] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Instance 3827eed1-b7a1-4922-a4de-f0c038cb8566 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65726) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 728.762500] env[65726]: WARNING openstack [None req-3b5c8bda-39bb-468a-bdce-9c81be5debce tempest-ServerExternalEventsTest-1019349750 tempest-ServerExternalEventsTest-1019349750-project] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 728.762500] env[65726]: WARNING openstack [None req-3b5c8bda-39bb-468a-bdce-9c81be5debce tempest-ServerExternalEventsTest-1019349750 tempest-ServerExternalEventsTest-1019349750-project] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 728.902990] env[65726]: DEBUG nova.network.neutron [None req-3b5c8bda-39bb-468a-bdce-9c81be5debce tempest-ServerExternalEventsTest-1019349750 tempest-ServerExternalEventsTest-1019349750-project] [instance: 96acc5f8-b411-408e-929d-d1035cfd50db] Updating instance_info_cache with network_info: [{"id": "c90c9d5c-3ed3-4f46-bd35-88e6fd91c396", "address": "fa:16:3e:ff:e6:2d", "network": {"id": "55d629b8-a541-4720-9289-927bbc6e6f22", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.205", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "1539f06fa3534e90acbc3a60c4b8bd3f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "35342bcb-8b06-472e-b3c0-43fd3d6c4b30", "external-id": "nsx-vlan-transportzone-524", "segmentation_id": 524, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc90c9d5c-3e", "ovs_interfaceid": "c90c9d5c-3ed3-4f46-bd35-88e6fd91c396", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 728.931260] env[65726]: DEBUG oslo_vmware.api [None req-36ef78ec-0b05-4841-bd7b-2242493adf78 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': task-5115722, 'name': SuspendVM_Task} progress is 62%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 729.133027] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115723, 'name': CreateVM_Task, 'duration_secs': 0.393129} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 729.133457] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 02351120-8385-4403-8464-a154f3da9380] Created VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 729.134310] env[65726]: WARNING openstack [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 729.134752] env[65726]: WARNING openstack [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 729.140449] env[65726]: DEBUG oslo_concurrency.lockutils [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 729.140449] env[65726]: DEBUG oslo_concurrency.lockutils [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 729.140753] env[65726]: DEBUG oslo_concurrency.lockutils [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 729.141368] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2ba89093-bd04-42c9-8b96-f930500a7452 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.146872] env[65726]: DEBUG oslo_vmware.api [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Waiting for the task: (returnval){ [ 729.146872] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52a31095-ad87-59dd-f118-bb518a5c5ba1" [ 729.146872] env[65726]: _type = "Task" [ 729.146872] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 729.157682] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Instance 00746508-d0d6-4dfa-9026-772398d004ee has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65726) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 729.159438] env[65726]: DEBUG oslo_vmware.api [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52a31095-ad87-59dd-f118-bb518a5c5ba1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 729.405928] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3b5c8bda-39bb-468a-bdce-9c81be5debce tempest-ServerExternalEventsTest-1019349750 tempest-ServerExternalEventsTest-1019349750-project] Releasing lock "refresh_cache-96acc5f8-b411-408e-929d-d1035cfd50db" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 729.428710] env[65726]: DEBUG oslo_vmware.api [None req-36ef78ec-0b05-4841-bd7b-2242493adf78 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': task-5115722, 'name': SuspendVM_Task, 'duration_secs': 0.680198} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 729.429055] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-36ef78ec-0b05-4841-bd7b-2242493adf78 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: bd839fe1-8801-4ba1-9c23-288453258b66] Suspended the VM {{(pid=65726) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1166}} [ 729.429416] env[65726]: DEBUG nova.compute.manager [None req-36ef78ec-0b05-4841-bd7b-2242493adf78 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: bd839fe1-8801-4ba1-9c23-288453258b66] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 729.430310] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f1bee91-9a8b-44e5-aad8-cf6142c21e85 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.495740] env[65726]: DEBUG oslo_concurrency.lockutils [None req-cda5c6a3-fc21-4556-afc4-6ebd218aa8a5 tempest-ServerExternalEventsTest-1845125219 tempest-ServerExternalEventsTest-1845125219-project-member] Acquiring lock "96acc5f8-b411-408e-929d-d1035cfd50db" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 729.496045] env[65726]: DEBUG oslo_concurrency.lockutils [None req-cda5c6a3-fc21-4556-afc4-6ebd218aa8a5 tempest-ServerExternalEventsTest-1845125219 tempest-ServerExternalEventsTest-1845125219-project-member] Lock "96acc5f8-b411-408e-929d-d1035cfd50db" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 729.496258] env[65726]: DEBUG oslo_concurrency.lockutils [None req-cda5c6a3-fc21-4556-afc4-6ebd218aa8a5 tempest-ServerExternalEventsTest-1845125219 tempest-ServerExternalEventsTest-1845125219-project-member] Acquiring lock "96acc5f8-b411-408e-929d-d1035cfd50db-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 729.496433] env[65726]: DEBUG oslo_concurrency.lockutils [None req-cda5c6a3-fc21-4556-afc4-6ebd218aa8a5 tempest-ServerExternalEventsTest-1845125219 tempest-ServerExternalEventsTest-1845125219-project-member] Lock "96acc5f8-b411-408e-929d-d1035cfd50db-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 729.496593] env[65726]: DEBUG oslo_concurrency.lockutils [None req-cda5c6a3-fc21-4556-afc4-6ebd218aa8a5 tempest-ServerExternalEventsTest-1845125219 tempest-ServerExternalEventsTest-1845125219-project-member] Lock "96acc5f8-b411-408e-929d-d1035cfd50db-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 729.498927] env[65726]: INFO nova.compute.manager [None req-cda5c6a3-fc21-4556-afc4-6ebd218aa8a5 tempest-ServerExternalEventsTest-1845125219 tempest-ServerExternalEventsTest-1845125219-project-member] [instance: 96acc5f8-b411-408e-929d-d1035cfd50db] Terminating instance [ 729.581621] env[65726]: DEBUG oslo_concurrency.lockutils [None req-04d3888f-ac16-4533-b7db-d3c669b17255 tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Acquiring lock "bc2c12e0-0d06-432f-b42f-be468e3b6ee4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 729.581883] env[65726]: DEBUG oslo_concurrency.lockutils [None req-04d3888f-ac16-4533-b7db-d3c669b17255 tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Lock "bc2c12e0-0d06-432f-b42f-be468e3b6ee4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 729.657867] env[65726]: DEBUG oslo_vmware.api [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52a31095-ad87-59dd-f118-bb518a5c5ba1, 'name': SearchDatastore_Task, 'duration_secs': 0.01128} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 729.658174] env[65726]: DEBUG oslo_concurrency.lockutils [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 729.658413] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] [instance: 02351120-8385-4403-8464-a154f3da9380] Processing image b52362a3-ee8a-4cbf-b06f-513b0cc8f95c {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 729.658634] env[65726]: DEBUG oslo_concurrency.lockutils [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 729.658768] env[65726]: DEBUG oslo_concurrency.lockutils [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 729.658937] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 729.659233] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0edab132-b24d-4e90-b897-c7e6faa3e9e2 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.663339] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Instance c108f2a5-031b-47ee-9a5b-d62c2f42c26b has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65726) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 729.672702] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 729.673076] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65726) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 729.674070] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5ce0a353-dfaf-4fdf-be42-f6e5f3f4e3f5 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.682377] env[65726]: DEBUG oslo_vmware.api [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Waiting for the task: (returnval){ [ 729.682377] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]523c3031-a75f-b0a7-23b2-18d49ee6d929" [ 729.682377] env[65726]: _type = "Task" [ 729.682377] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 729.693683] env[65726]: DEBUG oslo_vmware.api [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]523c3031-a75f-b0a7-23b2-18d49ee6d929, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 730.004627] env[65726]: DEBUG nova.compute.manager [None req-cda5c6a3-fc21-4556-afc4-6ebd218aa8a5 tempest-ServerExternalEventsTest-1845125219 tempest-ServerExternalEventsTest-1845125219-project-member] [instance: 96acc5f8-b411-408e-929d-d1035cfd50db] Start destroying the instance on the hypervisor. {{(pid=65726) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 730.005087] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-cda5c6a3-fc21-4556-afc4-6ebd218aa8a5 tempest-ServerExternalEventsTest-1845125219 tempest-ServerExternalEventsTest-1845125219-project-member] [instance: 96acc5f8-b411-408e-929d-d1035cfd50db] Destroying instance {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 730.006071] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2375c02-7533-4646-956a-03875f5680bd {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.015477] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-cda5c6a3-fc21-4556-afc4-6ebd218aa8a5 tempest-ServerExternalEventsTest-1845125219 tempest-ServerExternalEventsTest-1845125219-project-member] [instance: 96acc5f8-b411-408e-929d-d1035cfd50db] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 730.015761] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1d0adf41-2c8a-43de-8f43-5bbcaf95361a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.023182] env[65726]: DEBUG oslo_vmware.api [None req-cda5c6a3-fc21-4556-afc4-6ebd218aa8a5 tempest-ServerExternalEventsTest-1845125219 tempest-ServerExternalEventsTest-1845125219-project-member] Waiting for the task: (returnval){ [ 730.023182] env[65726]: value = "task-5115724" [ 730.023182] env[65726]: _type = "Task" [ 730.023182] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 730.034695] env[65726]: DEBUG oslo_vmware.api [None req-cda5c6a3-fc21-4556-afc4-6ebd218aa8a5 tempest-ServerExternalEventsTest-1845125219 tempest-ServerExternalEventsTest-1845125219-project-member] Task: {'id': task-5115724, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 730.166566] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Instance 895797e4-2941-44cd-aab1-67afa4fac02d has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65726) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 730.194983] env[65726]: DEBUG oslo_vmware.api [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]523c3031-a75f-b0a7-23b2-18d49ee6d929, 'name': SearchDatastore_Task, 'duration_secs': 0.010729} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 730.195838] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-00acc4a2-9e1b-4d1c-a635-b6c0bfdf8316 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.202136] env[65726]: DEBUG oslo_vmware.api [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Waiting for the task: (returnval){ [ 730.202136] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]527fd843-3f38-7a90-6da1-301e630974ec" [ 730.202136] env[65726]: _type = "Task" [ 730.202136] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 730.212536] env[65726]: DEBUG oslo_vmware.api [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]527fd843-3f38-7a90-6da1-301e630974ec, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 730.281113] env[65726]: DEBUG nova.compute.manager [req-2df5918b-11a7-4563-b6e2-34adc9d1b54d req-fba67f16-6685-4ce2-8191-fdc0289c9ff9 service nova] [instance: 02351120-8385-4403-8464-a154f3da9380] Received event network-changed-f9ef8504-78db-4da2-abc7-94b9a07a014e {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 730.281113] env[65726]: DEBUG nova.compute.manager [req-2df5918b-11a7-4563-b6e2-34adc9d1b54d req-fba67f16-6685-4ce2-8191-fdc0289c9ff9 service nova] [instance: 02351120-8385-4403-8464-a154f3da9380] Refreshing instance network info cache due to event network-changed-f9ef8504-78db-4da2-abc7-94b9a07a014e. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 730.281243] env[65726]: DEBUG oslo_concurrency.lockutils [req-2df5918b-11a7-4563-b6e2-34adc9d1b54d req-fba67f16-6685-4ce2-8191-fdc0289c9ff9 service nova] Acquiring lock "refresh_cache-02351120-8385-4403-8464-a154f3da9380" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 730.281342] env[65726]: DEBUG oslo_concurrency.lockutils [req-2df5918b-11a7-4563-b6e2-34adc9d1b54d req-fba67f16-6685-4ce2-8191-fdc0289c9ff9 service nova] Acquired lock "refresh_cache-02351120-8385-4403-8464-a154f3da9380" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 730.281494] env[65726]: DEBUG nova.network.neutron [req-2df5918b-11a7-4563-b6e2-34adc9d1b54d req-fba67f16-6685-4ce2-8191-fdc0289c9ff9 service nova] [instance: 02351120-8385-4403-8464-a154f3da9380] Refreshing network info cache for port f9ef8504-78db-4da2-abc7-94b9a07a014e {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 730.487415] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a76cfabd-af7f-4569-b968-716d45bcf7f1 tempest-InstanceActionsNegativeTestJSON-844675881 tempest-InstanceActionsNegativeTestJSON-844675881-project-member] Acquiring lock "1bd26aef-995e-43b6-af9f-077e878c8c44" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 730.487781] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a76cfabd-af7f-4569-b968-716d45bcf7f1 tempest-InstanceActionsNegativeTestJSON-844675881 tempest-InstanceActionsNegativeTestJSON-844675881-project-member] Lock "1bd26aef-995e-43b6-af9f-077e878c8c44" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 730.533358] env[65726]: DEBUG oslo_vmware.api [None req-cda5c6a3-fc21-4556-afc4-6ebd218aa8a5 tempest-ServerExternalEventsTest-1845125219 tempest-ServerExternalEventsTest-1845125219-project-member] Task: {'id': task-5115724, 'name': PowerOffVM_Task, 'duration_secs': 0.304572} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 730.533660] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-cda5c6a3-fc21-4556-afc4-6ebd218aa8a5 tempest-ServerExternalEventsTest-1845125219 tempest-ServerExternalEventsTest-1845125219-project-member] [instance: 96acc5f8-b411-408e-929d-d1035cfd50db] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 730.533815] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-cda5c6a3-fc21-4556-afc4-6ebd218aa8a5 tempest-ServerExternalEventsTest-1845125219 tempest-ServerExternalEventsTest-1845125219-project-member] [instance: 96acc5f8-b411-408e-929d-d1035cfd50db] Unregistering the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 730.534088] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-44105c9c-27db-4476-b85b-95f84d0d04cc {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.613036] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-cda5c6a3-fc21-4556-afc4-6ebd218aa8a5 tempest-ServerExternalEventsTest-1845125219 tempest-ServerExternalEventsTest-1845125219-project-member] [instance: 96acc5f8-b411-408e-929d-d1035cfd50db] Unregistered the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 730.613237] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-cda5c6a3-fc21-4556-afc4-6ebd218aa8a5 tempest-ServerExternalEventsTest-1845125219 tempest-ServerExternalEventsTest-1845125219-project-member] [instance: 96acc5f8-b411-408e-929d-d1035cfd50db] Deleting contents of the VM from datastore datastore2 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 730.613473] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-cda5c6a3-fc21-4556-afc4-6ebd218aa8a5 tempest-ServerExternalEventsTest-1845125219 tempest-ServerExternalEventsTest-1845125219-project-member] Deleting the datastore file [datastore2] 96acc5f8-b411-408e-929d-d1035cfd50db {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 730.613793] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-793a1599-e036-44de-b3f1-2734bf35d961 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.622283] env[65726]: DEBUG oslo_vmware.api [None req-cda5c6a3-fc21-4556-afc4-6ebd218aa8a5 tempest-ServerExternalEventsTest-1845125219 tempest-ServerExternalEventsTest-1845125219-project-member] Waiting for the task: (returnval){ [ 730.622283] env[65726]: value = "task-5115726" [ 730.622283] env[65726]: _type = "Task" [ 730.622283] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 730.631472] env[65726]: DEBUG oslo_vmware.api [None req-cda5c6a3-fc21-4556-afc4-6ebd218aa8a5 tempest-ServerExternalEventsTest-1845125219 tempest-ServerExternalEventsTest-1845125219-project-member] Task: {'id': task-5115726, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 730.669418] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Instance 0d6f14c5-5770-476f-a186-4cf1b2d4fbe8 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65726) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 730.714808] env[65726]: DEBUG oslo_vmware.api [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]527fd843-3f38-7a90-6da1-301e630974ec, 'name': SearchDatastore_Task, 'duration_secs': 0.034429} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 730.715175] env[65726]: DEBUG oslo_concurrency.lockutils [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 730.715478] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] 02351120-8385-4403-8464-a154f3da9380/02351120-8385-4403-8464-a154f3da9380.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 730.715922] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a6575995-59ed-4ea6-8b56-5eb3ea2cdd77 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.724044] env[65726]: DEBUG oslo_vmware.api [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Waiting for the task: (returnval){ [ 730.724044] env[65726]: value = "task-5115727" [ 730.724044] env[65726]: _type = "Task" [ 730.724044] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 730.735720] env[65726]: DEBUG oslo_vmware.api [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Task: {'id': task-5115727, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 730.784455] env[65726]: WARNING openstack [req-2df5918b-11a7-4563-b6e2-34adc9d1b54d req-fba67f16-6685-4ce2-8191-fdc0289c9ff9 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 730.784867] env[65726]: WARNING openstack [req-2df5918b-11a7-4563-b6e2-34adc9d1b54d req-fba67f16-6685-4ce2-8191-fdc0289c9ff9 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 730.898577] env[65726]: WARNING openstack [req-2df5918b-11a7-4563-b6e2-34adc9d1b54d req-fba67f16-6685-4ce2-8191-fdc0289c9ff9 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 730.899077] env[65726]: WARNING openstack [req-2df5918b-11a7-4563-b6e2-34adc9d1b54d req-fba67f16-6685-4ce2-8191-fdc0289c9ff9 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 730.983258] env[65726]: WARNING openstack [req-2df5918b-11a7-4563-b6e2-34adc9d1b54d req-fba67f16-6685-4ce2-8191-fdc0289c9ff9 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 730.983712] env[65726]: WARNING openstack [req-2df5918b-11a7-4563-b6e2-34adc9d1b54d req-fba67f16-6685-4ce2-8191-fdc0289c9ff9 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 731.078689] env[65726]: DEBUG nova.network.neutron [req-2df5918b-11a7-4563-b6e2-34adc9d1b54d req-fba67f16-6685-4ce2-8191-fdc0289c9ff9 service nova] [instance: 02351120-8385-4403-8464-a154f3da9380] Updated VIF entry in instance network info cache for port f9ef8504-78db-4da2-abc7-94b9a07a014e. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 731.079064] env[65726]: DEBUG nova.network.neutron [req-2df5918b-11a7-4563-b6e2-34adc9d1b54d req-fba67f16-6685-4ce2-8191-fdc0289c9ff9 service nova] [instance: 02351120-8385-4403-8464-a154f3da9380] Updating instance_info_cache with network_info: [{"id": "f9ef8504-78db-4da2-abc7-94b9a07a014e", "address": "fa:16:3e:ce:88:da", "network": {"id": "c68b8bcc-840e-459b-bee0-76a9ffdc08cb", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-2061916615-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7c3641c4a34b4bf8b09403220e153e65", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "aa0d7fd6-4fd3-4451-9208-8a3cf25e30e4", "external-id": "nsx-vlan-transportzone-545", "segmentation_id": 545, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf9ef8504-78", "ovs_interfaceid": "f9ef8504-78db-4da2-abc7-94b9a07a014e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 731.135361] env[65726]: DEBUG oslo_vmware.api [None req-cda5c6a3-fc21-4556-afc4-6ebd218aa8a5 tempest-ServerExternalEventsTest-1845125219 tempest-ServerExternalEventsTest-1845125219-project-member] Task: {'id': task-5115726, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.13691} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 731.135730] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-cda5c6a3-fc21-4556-afc4-6ebd218aa8a5 tempest-ServerExternalEventsTest-1845125219 tempest-ServerExternalEventsTest-1845125219-project-member] Deleted the datastore file {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 731.135882] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-cda5c6a3-fc21-4556-afc4-6ebd218aa8a5 tempest-ServerExternalEventsTest-1845125219 tempest-ServerExternalEventsTest-1845125219-project-member] [instance: 96acc5f8-b411-408e-929d-d1035cfd50db] Deleted contents of the VM from datastore datastore2 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 731.136104] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-cda5c6a3-fc21-4556-afc4-6ebd218aa8a5 tempest-ServerExternalEventsTest-1845125219 tempest-ServerExternalEventsTest-1845125219-project-member] [instance: 96acc5f8-b411-408e-929d-d1035cfd50db] Instance destroyed {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 731.136360] env[65726]: INFO nova.compute.manager [None req-cda5c6a3-fc21-4556-afc4-6ebd218aa8a5 tempest-ServerExternalEventsTest-1845125219 tempest-ServerExternalEventsTest-1845125219-project-member] [instance: 96acc5f8-b411-408e-929d-d1035cfd50db] Took 1.13 seconds to destroy the instance on the hypervisor. [ 731.136673] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-cda5c6a3-fc21-4556-afc4-6ebd218aa8a5 tempest-ServerExternalEventsTest-1845125219 tempest-ServerExternalEventsTest-1845125219-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 731.136975] env[65726]: DEBUG nova.compute.manager [-] [instance: 96acc5f8-b411-408e-929d-d1035cfd50db] Deallocating network for instance {{(pid=65726) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 731.137077] env[65726]: DEBUG nova.network.neutron [-] [instance: 96acc5f8-b411-408e-929d-d1035cfd50db] deallocate_for_instance() {{(pid=65726) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 731.137678] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 731.138036] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 731.155144] env[65726]: INFO nova.compute.manager [None req-11e4b8a8-60d9-4939-a6a9-6a357b669b13 tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] [instance: 83573c2b-9448-456f-8fd2-b19661dd6cc4] Rebuilding instance [ 731.175136] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Instance 65aaa2eb-3ca5-4c2b-92f9-b524366c10dd has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65726) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 731.209325] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 731.209685] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 731.221278] env[65726]: DEBUG nova.compute.manager [None req-11e4b8a8-60d9-4939-a6a9-6a357b669b13 tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] [instance: 83573c2b-9448-456f-8fd2-b19661dd6cc4] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 731.222306] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36266c56-32be-4f3c-844e-175039980292 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.235043] env[65726]: DEBUG oslo_vmware.api [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Task: {'id': task-5115727, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 731.583034] env[65726]: DEBUG oslo_concurrency.lockutils [req-2df5918b-11a7-4563-b6e2-34adc9d1b54d req-fba67f16-6685-4ce2-8191-fdc0289c9ff9 service nova] Releasing lock "refresh_cache-02351120-8385-4403-8464-a154f3da9380" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 731.678922] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Instance 6d2b86c2-0a44-433f-8b3f-2e9c7f693baa has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65726) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 731.720233] env[65726]: DEBUG nova.compute.manager [req-defec175-0b9b-4013-9a85-c3fed7d74b2e req-713c04e9-aca1-423f-a2f3-7b413415f51d service nova] [instance: 96acc5f8-b411-408e-929d-d1035cfd50db] Received event network-vif-deleted-c90c9d5c-3ed3-4f46-bd35-88e6fd91c396 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 731.720233] env[65726]: INFO nova.compute.manager [req-defec175-0b9b-4013-9a85-c3fed7d74b2e req-713c04e9-aca1-423f-a2f3-7b413415f51d service nova] [instance: 96acc5f8-b411-408e-929d-d1035cfd50db] Neutron deleted interface c90c9d5c-3ed3-4f46-bd35-88e6fd91c396; detaching it from the instance and deleting it from the info cache [ 731.720233] env[65726]: DEBUG nova.network.neutron [req-defec175-0b9b-4013-9a85-c3fed7d74b2e req-713c04e9-aca1-423f-a2f3-7b413415f51d service nova] [instance: 96acc5f8-b411-408e-929d-d1035cfd50db] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 731.740503] env[65726]: DEBUG oslo_vmware.api [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Task: {'id': task-5115727, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.519332} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 731.741461] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] 02351120-8385-4403-8464-a154f3da9380/02351120-8385-4403-8464-a154f3da9380.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 731.741699] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] [instance: 02351120-8385-4403-8464-a154f3da9380] Extending root virtual disk to 1048576 {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 731.744716] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4988bdd1-b695-44fd-9ea7-567315898885 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.753161] env[65726]: DEBUG oslo_vmware.api [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Waiting for the task: (returnval){ [ 731.753161] env[65726]: value = "task-5115728" [ 731.753161] env[65726]: _type = "Task" [ 731.753161] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 731.764120] env[65726]: DEBUG oslo_vmware.api [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Task: {'id': task-5115728, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 731.831263] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5ae556a9-54a3-4f16-bd92-05c390a03c24 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Acquiring lock "fd481728-3ef0-4a96-affd-ab2dd3f596bb" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 731.831530] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5ae556a9-54a3-4f16-bd92-05c390a03c24 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Lock "fd481728-3ef0-4a96-affd-ab2dd3f596bb" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 731.832428] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5ae556a9-54a3-4f16-bd92-05c390a03c24 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Acquiring lock "fd481728-3ef0-4a96-affd-ab2dd3f596bb-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 731.832513] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5ae556a9-54a3-4f16-bd92-05c390a03c24 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Lock "fd481728-3ef0-4a96-affd-ab2dd3f596bb-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 731.833061] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5ae556a9-54a3-4f16-bd92-05c390a03c24 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Lock "fd481728-3ef0-4a96-affd-ab2dd3f596bb-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 731.835631] env[65726]: INFO nova.compute.manager [None req-5ae556a9-54a3-4f16-bd92-05c390a03c24 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] [instance: fd481728-3ef0-4a96-affd-ab2dd3f596bb] Terminating instance [ 731.971554] env[65726]: DEBUG nova.network.neutron [-] [instance: 96acc5f8-b411-408e-929d-d1035cfd50db] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 731.982243] env[65726]: DEBUG nova.compute.manager [None req-61a6387d-c5ed-4ab7-853e-9ba63af33205 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: bd839fe1-8801-4ba1-9c23-288453258b66] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 731.983651] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b3028d3-2ac7-4ae6-8b59-93700f1d5170 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.186058] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Instance 9fc0c9ee-9dc7-4059-b8d2-a5db965e00c2 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65726) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 732.223539] env[65726]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8f3bc40d-1627-44d1-9630-0135040dfe19 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.235020] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76783d90-e197-42b1-a965-77aa439089db {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.247095] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-11e4b8a8-60d9-4939-a6a9-6a357b669b13 tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] [instance: 83573c2b-9448-456f-8fd2-b19661dd6cc4] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 732.247970] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-bc3762e0-564e-401a-b086-cb32aca6cf0d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.256602] env[65726]: DEBUG oslo_vmware.api [None req-11e4b8a8-60d9-4939-a6a9-6a357b669b13 tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] Waiting for the task: (returnval){ [ 732.256602] env[65726]: value = "task-5115729" [ 732.256602] env[65726]: _type = "Task" [ 732.256602] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 732.278610] env[65726]: DEBUG nova.compute.manager [req-defec175-0b9b-4013-9a85-c3fed7d74b2e req-713c04e9-aca1-423f-a2f3-7b413415f51d service nova] [instance: 96acc5f8-b411-408e-929d-d1035cfd50db] Detach interface failed, port_id=c90c9d5c-3ed3-4f46-bd35-88e6fd91c396, reason: Instance 96acc5f8-b411-408e-929d-d1035cfd50db could not be found. {{(pid=65726) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11669}} [ 732.284862] env[65726]: DEBUG oslo_vmware.api [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Task: {'id': task-5115728, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.070152} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 732.289822] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] [instance: 02351120-8385-4403-8464-a154f3da9380] Extended root virtual disk {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 732.289822] env[65726]: DEBUG oslo_vmware.api [None req-11e4b8a8-60d9-4939-a6a9-6a357b669b13 tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] Task: {'id': task-5115729, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 732.289822] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd596fba-98d2-41ee-85ca-7cfb14944b66 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.317500] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] [instance: 02351120-8385-4403-8464-a154f3da9380] Reconfiguring VM instance instance-00000023 to attach disk [datastore1] 02351120-8385-4403-8464-a154f3da9380/02351120-8385-4403-8464-a154f3da9380.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 732.317982] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-33b215fa-d709-49fb-ae29-8c62815eb7f5 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.339638] env[65726]: DEBUG nova.compute.manager [None req-5ae556a9-54a3-4f16-bd92-05c390a03c24 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] [instance: fd481728-3ef0-4a96-affd-ab2dd3f596bb] Start destroying the instance on the hypervisor. {{(pid=65726) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 732.341987] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-5ae556a9-54a3-4f16-bd92-05c390a03c24 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] [instance: fd481728-3ef0-4a96-affd-ab2dd3f596bb] Destroying instance {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 732.341987] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6346231-dcfb-4ad6-a0c0-536122db65b1 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.347245] env[65726]: DEBUG oslo_vmware.api [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Waiting for the task: (returnval){ [ 732.347245] env[65726]: value = "task-5115730" [ 732.347245] env[65726]: _type = "Task" [ 732.347245] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 732.354178] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-5ae556a9-54a3-4f16-bd92-05c390a03c24 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] [instance: fd481728-3ef0-4a96-affd-ab2dd3f596bb] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 732.355019] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5268b421-91f2-4591-89d8-185b8cd93b1c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.362773] env[65726]: DEBUG oslo_vmware.api [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Task: {'id': task-5115730, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 732.364518] env[65726]: DEBUG oslo_vmware.api [None req-5ae556a9-54a3-4f16-bd92-05c390a03c24 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Waiting for the task: (returnval){ [ 732.364518] env[65726]: value = "task-5115731" [ 732.364518] env[65726]: _type = "Task" [ 732.364518] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 732.376242] env[65726]: DEBUG oslo_vmware.api [None req-5ae556a9-54a3-4f16-bd92-05c390a03c24 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Task: {'id': task-5115731, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 732.475100] env[65726]: INFO nova.compute.manager [-] [instance: 96acc5f8-b411-408e-929d-d1035cfd50db] Took 1.34 seconds to deallocate network for instance. [ 732.496374] env[65726]: INFO nova.compute.manager [None req-61a6387d-c5ed-4ab7-853e-9ba63af33205 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: bd839fe1-8801-4ba1-9c23-288453258b66] instance snapshotting [ 732.496598] env[65726]: WARNING nova.compute.manager [None req-61a6387d-c5ed-4ab7-853e-9ba63af33205 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: bd839fe1-8801-4ba1-9c23-288453258b66] trying to snapshot a non-running instance: (state: 7 expected: 1) [ 732.500627] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ce534c4-0724-4abd-8864-d8184e1da383 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.524924] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e36893b-4291-4e5d-beb1-4c1aa4d3952b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.688889] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Instance ad48cadd-9bb6-4191-a68a-5c8fdaeebd44 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65726) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 732.771292] env[65726]: DEBUG oslo_vmware.api [None req-11e4b8a8-60d9-4939-a6a9-6a357b669b13 tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] Task: {'id': task-5115729, 'name': PowerOffVM_Task, 'duration_secs': 0.134903} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 732.771613] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-11e4b8a8-60d9-4939-a6a9-6a357b669b13 tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] [instance: 83573c2b-9448-456f-8fd2-b19661dd6cc4] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 732.772506] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-11e4b8a8-60d9-4939-a6a9-6a357b669b13 tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] [instance: 83573c2b-9448-456f-8fd2-b19661dd6cc4] Destroying instance {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 732.773688] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27e48e14-7261-4595-bf3f-3291d7e74c19 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.781959] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-11e4b8a8-60d9-4939-a6a9-6a357b669b13 tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] [instance: 83573c2b-9448-456f-8fd2-b19661dd6cc4] Unregistering the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 732.782256] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f8eaaf94-5bf6-4f49-8977-cce681944845 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.812377] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-11e4b8a8-60d9-4939-a6a9-6a357b669b13 tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] [instance: 83573c2b-9448-456f-8fd2-b19661dd6cc4] Unregistered the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 732.812603] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-11e4b8a8-60d9-4939-a6a9-6a357b669b13 tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] [instance: 83573c2b-9448-456f-8fd2-b19661dd6cc4] Deleting contents of the VM from datastore datastore2 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 732.813115] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-11e4b8a8-60d9-4939-a6a9-6a357b669b13 tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] Deleting the datastore file [datastore2] 83573c2b-9448-456f-8fd2-b19661dd6cc4 {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 732.815127] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-aa7dc52c-54e4-492f-bb73-a908f03c9ab3 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.825219] env[65726]: DEBUG oslo_vmware.api [None req-11e4b8a8-60d9-4939-a6a9-6a357b669b13 tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] Waiting for the task: (returnval){ [ 732.825219] env[65726]: value = "task-5115733" [ 732.825219] env[65726]: _type = "Task" [ 732.825219] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 732.836363] env[65726]: DEBUG oslo_vmware.api [None req-11e4b8a8-60d9-4939-a6a9-6a357b669b13 tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] Task: {'id': task-5115733, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 732.859109] env[65726]: DEBUG oslo_vmware.api [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Task: {'id': task-5115730, 'name': ReconfigVM_Task, 'duration_secs': 0.342819} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 732.859171] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] [instance: 02351120-8385-4403-8464-a154f3da9380] Reconfigured VM instance instance-00000023 to attach disk [datastore1] 02351120-8385-4403-8464-a154f3da9380/02351120-8385-4403-8464-a154f3da9380.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 732.859951] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-eeb1bf37-7132-4f7a-8fdd-d0c1a1dd2330 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.870687] env[65726]: DEBUG oslo_vmware.api [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Waiting for the task: (returnval){ [ 732.870687] env[65726]: value = "task-5115734" [ 732.870687] env[65726]: _type = "Task" [ 732.870687] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 732.878283] env[65726]: DEBUG oslo_vmware.api [None req-5ae556a9-54a3-4f16-bd92-05c390a03c24 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Task: {'id': task-5115731, 'name': PowerOffVM_Task, 'duration_secs': 0.203463} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 732.884022] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-5ae556a9-54a3-4f16-bd92-05c390a03c24 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] [instance: fd481728-3ef0-4a96-affd-ab2dd3f596bb] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 732.884022] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-5ae556a9-54a3-4f16-bd92-05c390a03c24 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] [instance: fd481728-3ef0-4a96-affd-ab2dd3f596bb] Unregistering the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 732.884022] env[65726]: DEBUG oslo_vmware.api [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Task: {'id': task-5115734, 'name': Rename_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 732.884022] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4165c24a-3fea-4b51-a7a3-a8f31ba3a6aa {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.956381] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-5ae556a9-54a3-4f16-bd92-05c390a03c24 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] [instance: fd481728-3ef0-4a96-affd-ab2dd3f596bb] Unregistered the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 732.956834] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-5ae556a9-54a3-4f16-bd92-05c390a03c24 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] [instance: fd481728-3ef0-4a96-affd-ab2dd3f596bb] Deleting contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 732.956944] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-5ae556a9-54a3-4f16-bd92-05c390a03c24 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Deleting the datastore file [datastore1] fd481728-3ef0-4a96-affd-ab2dd3f596bb {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 732.957165] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-50644f38-8004-42c6-a1fe-478b5d7854cb {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.965016] env[65726]: DEBUG oslo_vmware.api [None req-5ae556a9-54a3-4f16-bd92-05c390a03c24 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Waiting for the task: (returnval){ [ 732.965016] env[65726]: value = "task-5115736" [ 732.965016] env[65726]: _type = "Task" [ 732.965016] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 732.978047] env[65726]: DEBUG oslo_vmware.api [None req-5ae556a9-54a3-4f16-bd92-05c390a03c24 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Task: {'id': task-5115736, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 732.983737] env[65726]: DEBUG oslo_concurrency.lockutils [None req-cda5c6a3-fc21-4556-afc4-6ebd218aa8a5 tempest-ServerExternalEventsTest-1845125219 tempest-ServerExternalEventsTest-1845125219-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 733.041454] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-61a6387d-c5ed-4ab7-853e-9ba63af33205 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: bd839fe1-8801-4ba1-9c23-288453258b66] Creating Snapshot of the VM instance {{(pid=65726) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 733.041454] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-4d74c947-ad00-4bf8-bb6a-6c382e1f9ee2 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.048862] env[65726]: DEBUG oslo_vmware.api [None req-61a6387d-c5ed-4ab7-853e-9ba63af33205 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Waiting for the task: (returnval){ [ 733.048862] env[65726]: value = "task-5115737" [ 733.048862] env[65726]: _type = "Task" [ 733.048862] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 733.059348] env[65726]: DEBUG oslo_vmware.api [None req-61a6387d-c5ed-4ab7-853e-9ba63af33205 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': task-5115737, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 733.194770] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Instance 76249623-6f83-46a3-b8c5-c001111aa698 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65726) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 733.323724] env[65726]: DEBUG oslo_concurrency.lockutils [None req-855996fc-ecea-49d8-a71b-8b4307731732 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Acquiring lock "6c8db442-daaa-4eca-ae24-b9d4afaf3a71" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 733.324127] env[65726]: DEBUG oslo_concurrency.lockutils [None req-855996fc-ecea-49d8-a71b-8b4307731732 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Lock "6c8db442-daaa-4eca-ae24-b9d4afaf3a71" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 733.335913] env[65726]: DEBUG oslo_vmware.api [None req-11e4b8a8-60d9-4939-a6a9-6a357b669b13 tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] Task: {'id': task-5115733, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.159784} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 733.335913] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-11e4b8a8-60d9-4939-a6a9-6a357b669b13 tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] Deleted the datastore file {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 733.335913] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-11e4b8a8-60d9-4939-a6a9-6a357b669b13 tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] [instance: 83573c2b-9448-456f-8fd2-b19661dd6cc4] Deleted contents of the VM from datastore datastore2 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 733.335913] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-11e4b8a8-60d9-4939-a6a9-6a357b669b13 tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] [instance: 83573c2b-9448-456f-8fd2-b19661dd6cc4] Instance destroyed {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 733.381522] env[65726]: DEBUG oslo_vmware.api [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Task: {'id': task-5115734, 'name': Rename_Task, 'duration_secs': 0.175928} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 733.382769] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] [instance: 02351120-8385-4403-8464-a154f3da9380] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 733.383075] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-60f17627-dedd-4b42-804a-27229e724954 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.391994] env[65726]: DEBUG oslo_vmware.api [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Waiting for the task: (returnval){ [ 733.391994] env[65726]: value = "task-5115738" [ 733.391994] env[65726]: _type = "Task" [ 733.391994] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 733.405188] env[65726]: DEBUG oslo_vmware.api [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Task: {'id': task-5115738, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 733.476770] env[65726]: DEBUG oslo_vmware.api [None req-5ae556a9-54a3-4f16-bd92-05c390a03c24 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Task: {'id': task-5115736, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.165036} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 733.476996] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-5ae556a9-54a3-4f16-bd92-05c390a03c24 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Deleted the datastore file {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 733.478301] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-5ae556a9-54a3-4f16-bd92-05c390a03c24 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] [instance: fd481728-3ef0-4a96-affd-ab2dd3f596bb] Deleted contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 733.478507] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-5ae556a9-54a3-4f16-bd92-05c390a03c24 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] [instance: fd481728-3ef0-4a96-affd-ab2dd3f596bb] Instance destroyed {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 733.478725] env[65726]: INFO nova.compute.manager [None req-5ae556a9-54a3-4f16-bd92-05c390a03c24 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] [instance: fd481728-3ef0-4a96-affd-ab2dd3f596bb] Took 1.14 seconds to destroy the instance on the hypervisor. [ 733.479381] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-5ae556a9-54a3-4f16-bd92-05c390a03c24 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 733.479381] env[65726]: DEBUG nova.compute.manager [-] [instance: fd481728-3ef0-4a96-affd-ab2dd3f596bb] Deallocating network for instance {{(pid=65726) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 733.479381] env[65726]: DEBUG nova.network.neutron [-] [instance: fd481728-3ef0-4a96-affd-ab2dd3f596bb] deallocate_for_instance() {{(pid=65726) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 733.479969] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 733.480791] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 733.533563] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 733.534528] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 733.561120] env[65726]: DEBUG oslo_vmware.api [None req-61a6387d-c5ed-4ab7-853e-9ba63af33205 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': task-5115737, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 733.700081] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Instance 621b899a-e9ec-4fc0-a574-4a08cecff6d9 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65726) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 733.700081] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Total usable vcpus: 48, total allocated vcpus: 16 {{(pid=65726) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1159}} [ 733.700619] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=3648MB phys_disk=100GB used_disk=15GB total_vcpus=48 used_vcpus=16 pci_stats=[] stats={'failed_builds': '0', 'num_instances': '16', 'num_vm_active': '15', 'num_task_None': '15', 'num_os_type_None': '16', 'num_proj_fae9ef0198a2468a9793374b76378abf': '1', 'io_workload': '1', 'num_proj_3a680eae0e734ea8b2274b504d0a46d8': '2', 'num_proj_27fdee694977439d88ed2b08abb28405': '1', 'num_proj_b01c77c483364e23b4a48b1e1c9c9d1c': '1', 'num_proj_b794c99309e94c1181fc9d8eeeb84702': '2', 'num_proj_c52b5289c78f45d3942a6c0a4b026207': '1', 'num_proj_79d3e960f190439fae7c61d693e9356e': '1', 'num_proj_aac5fe6c99f740ada3747088b2b0ad89': '1', 'num_proj_d46ccb00794f458b85da4a93879139ee': '1', 'num_proj_4eece77569624f90bf64e5c51974173f': '1', 'num_proj_fddc635807d54dc0a5b4f2f732ba6962': '1', 'num_proj_39067912d091464192516c08eb576696': '1', 'num_proj_1f3ef84831b3441c8532cdce35f4d245': '1', 'num_vm_building': '1', 'num_task_spawning': '1', 'num_proj_7c3641c4a34b4bf8b09403220e153e65': '1'} {{(pid=65726) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1168}} [ 733.907076] env[65726]: DEBUG oslo_vmware.api [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Task: {'id': task-5115738, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 734.074012] env[65726]: DEBUG oslo_vmware.api [None req-61a6387d-c5ed-4ab7-853e-9ba63af33205 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': task-5115737, 'name': CreateSnapshot_Task, 'duration_secs': 0.764566} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 734.074012] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-61a6387d-c5ed-4ab7-853e-9ba63af33205 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: bd839fe1-8801-4ba1-9c23-288453258b66] Created Snapshot of the VM instance {{(pid=65726) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 734.074012] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c39d88e-9dcd-4726-b59f-a32b5b0a3762 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.262984] env[65726]: DEBUG nova.compute.manager [req-1e62ac29-5e61-40e8-bb7e-93be09da5276 req-99874e86-f804-4a3f-b477-06c00bcdf29d service nova] [instance: fd481728-3ef0-4a96-affd-ab2dd3f596bb] Received event network-vif-deleted-7df3f02b-e4e5-4840-bf9f-3498771b9184 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 734.263159] env[65726]: INFO nova.compute.manager [req-1e62ac29-5e61-40e8-bb7e-93be09da5276 req-99874e86-f804-4a3f-b477-06c00bcdf29d service nova] [instance: fd481728-3ef0-4a96-affd-ab2dd3f596bb] Neutron deleted interface 7df3f02b-e4e5-4840-bf9f-3498771b9184; detaching it from the instance and deleting it from the info cache [ 734.263287] env[65726]: DEBUG nova.network.neutron [req-1e62ac29-5e61-40e8-bb7e-93be09da5276 req-99874e86-f804-4a3f-b477-06c00bcdf29d service nova] [instance: fd481728-3ef0-4a96-affd-ab2dd3f596bb] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 734.391036] env[65726]: DEBUG nova.virt.hardware [None req-11e4b8a8-60d9-4939-a6a9-6a357b669b13 tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T19:28:34Z,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1539f06fa3534e90acbc3a60c4b8bd3f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T19:28:35Z,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 734.392213] env[65726]: DEBUG nova.virt.hardware [None req-11e4b8a8-60d9-4939-a6a9-6a357b669b13 tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 734.392393] env[65726]: DEBUG nova.virt.hardware [None req-11e4b8a8-60d9-4939-a6a9-6a357b669b13 tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 734.392577] env[65726]: DEBUG nova.virt.hardware [None req-11e4b8a8-60d9-4939-a6a9-6a357b669b13 tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 734.392719] env[65726]: DEBUG nova.virt.hardware [None req-11e4b8a8-60d9-4939-a6a9-6a357b669b13 tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 734.392857] env[65726]: DEBUG nova.virt.hardware [None req-11e4b8a8-60d9-4939-a6a9-6a357b669b13 tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 734.393622] env[65726]: DEBUG nova.virt.hardware [None req-11e4b8a8-60d9-4939-a6a9-6a357b669b13 tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 734.393813] env[65726]: DEBUG nova.virt.hardware [None req-11e4b8a8-60d9-4939-a6a9-6a357b669b13 tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 734.394031] env[65726]: DEBUG nova.virt.hardware [None req-11e4b8a8-60d9-4939-a6a9-6a357b669b13 tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 734.394176] env[65726]: DEBUG nova.virt.hardware [None req-11e4b8a8-60d9-4939-a6a9-6a357b669b13 tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 734.394323] env[65726]: DEBUG nova.virt.hardware [None req-11e4b8a8-60d9-4939-a6a9-6a357b669b13 tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 734.395301] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd5e8492-f799-45da-be45-ee298e96d8ab {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.416923] env[65726]: DEBUG oslo_vmware.api [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Task: {'id': task-5115738, 'name': PowerOnVM_Task, 'duration_secs': 0.538602} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 734.417808] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] [instance: 02351120-8385-4403-8464-a154f3da9380] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 734.418051] env[65726]: INFO nova.compute.manager [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] [instance: 02351120-8385-4403-8464-a154f3da9380] Took 8.65 seconds to spawn the instance on the hypervisor. [ 734.418227] env[65726]: DEBUG nova.compute.manager [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] [instance: 02351120-8385-4403-8464-a154f3da9380] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 734.419562] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a6ab12e-9b6a-4012-ac37-57d849ad16ce {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.424806] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3093544a-301b-4114-afa4-a663471af692 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.440821] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-11e4b8a8-60d9-4939-a6a9-6a357b669b13 tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] [instance: 83573c2b-9448-456f-8fd2-b19661dd6cc4] Instance VIF info [] {{(pid=65726) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 734.447048] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-11e4b8a8-60d9-4939-a6a9-6a357b669b13 tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 734.450017] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 83573c2b-9448-456f-8fd2-b19661dd6cc4] Creating VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 734.450748] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2251f08b-dfab-4438-a647-8ebb0844c5bc {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.457411] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c7860a22-9ca6-4529-9750-44fa530ab4f7 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.476923] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e8c93fc-5065-4a07-9c2a-8731569a2e91 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.481067] env[65726]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 734.481067] env[65726]: value = "task-5115739" [ 734.481067] env[65726]: _type = "Task" [ 734.481067] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 734.517930] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f789d494-f869-4b5c-a4f7-9e695e59d2df {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.526871] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115739, 'name': CreateVM_Task} progress is 15%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 734.540134] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d40010d-e094-4d7f-88cb-b1ae015cfa00 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.553882] env[65726]: DEBUG nova.compute.provider_tree [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Updating inventory in ProviderTree for provider 07c4692f-bdb4-4058-9173-ff9664830295 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 734.605245] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-61a6387d-c5ed-4ab7-853e-9ba63af33205 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: bd839fe1-8801-4ba1-9c23-288453258b66] Creating linked-clone VM from snapshot {{(pid=65726) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 734.605936] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-4d7198ab-072d-41f5-870c-27e907c97515 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.609582] env[65726]: DEBUG nova.network.neutron [-] [instance: fd481728-3ef0-4a96-affd-ab2dd3f596bb] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 734.618198] env[65726]: DEBUG oslo_vmware.api [None req-61a6387d-c5ed-4ab7-853e-9ba63af33205 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Waiting for the task: (returnval){ [ 734.618198] env[65726]: value = "task-5115740" [ 734.618198] env[65726]: _type = "Task" [ 734.618198] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 734.627335] env[65726]: DEBUG oslo_vmware.api [None req-61a6387d-c5ed-4ab7-853e-9ba63af33205 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': task-5115740, 'name': CloneVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 734.771986] env[65726]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-fcc20c41-8735-4049-b366-244ff6c8a640 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.784409] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aef02239-bbef-439a-9f50-b8a1290212bd {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.826445] env[65726]: DEBUG nova.compute.manager [req-1e62ac29-5e61-40e8-bb7e-93be09da5276 req-99874e86-f804-4a3f-b477-06c00bcdf29d service nova] [instance: fd481728-3ef0-4a96-affd-ab2dd3f596bb] Detach interface failed, port_id=7df3f02b-e4e5-4840-bf9f-3498771b9184, reason: Instance fd481728-3ef0-4a96-affd-ab2dd3f596bb could not be found. {{(pid=65726) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11669}} [ 734.980913] env[65726]: INFO nova.compute.manager [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] [instance: 02351120-8385-4403-8464-a154f3da9380] Took 47.74 seconds to build instance. [ 734.992026] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115739, 'name': CreateVM_Task} progress is 99%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 735.080897] env[65726]: ERROR nova.scheduler.client.report [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] [req-2cec20c5-f722-4236-95bd-c924604be4ed] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 07c4692f-bdb4-4058-9173-ff9664830295. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-2cec20c5-f722-4236-95bd-c924604be4ed"}]} [ 735.106624] env[65726]: DEBUG nova.scheduler.client.report [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Refreshing inventories for resource provider 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 735.112158] env[65726]: INFO nova.compute.manager [-] [instance: fd481728-3ef0-4a96-affd-ab2dd3f596bb] Took 1.63 seconds to deallocate network for instance. [ 735.123257] env[65726]: DEBUG nova.scheduler.client.report [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Updating ProviderTree inventory for provider 07c4692f-bdb4-4058-9173-ff9664830295 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 735.123533] env[65726]: DEBUG nova.compute.provider_tree [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Updating inventory in ProviderTree for provider 07c4692f-bdb4-4058-9173-ff9664830295 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 735.129303] env[65726]: DEBUG oslo_vmware.api [None req-61a6387d-c5ed-4ab7-853e-9ba63af33205 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': task-5115740, 'name': CloneVM_Task} progress is 94%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 735.142206] env[65726]: DEBUG nova.scheduler.client.report [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Refreshing aggregate associations for resource provider 07c4692f-bdb4-4058-9173-ff9664830295, aggregates: None {{(pid=65726) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 735.165170] env[65726]: DEBUG nova.scheduler.client.report [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Refreshing trait associations for resource provider 07c4692f-bdb4-4058-9173-ff9664830295, traits: COMPUTE_IMAGE_TYPE_ISO,HW_ARCH_X86_64,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=65726) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 735.486506] env[65726]: DEBUG oslo_concurrency.lockutils [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Lock "02351120-8385-4403-8464-a154f3da9380" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 57.747s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 735.494182] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115739, 'name': CreateVM_Task, 'duration_secs': 0.606167} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 735.496494] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 83573c2b-9448-456f-8fd2-b19661dd6cc4] Created VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 735.497638] env[65726]: DEBUG oslo_concurrency.lockutils [None req-11e4b8a8-60d9-4939-a6a9-6a357b669b13 tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 735.497638] env[65726]: DEBUG oslo_concurrency.lockutils [None req-11e4b8a8-60d9-4939-a6a9-6a357b669b13 tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 735.497638] env[65726]: DEBUG oslo_concurrency.lockutils [None req-11e4b8a8-60d9-4939-a6a9-6a357b669b13 tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 735.497848] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f262bc27-5613-44ca-be98-7a038ef2bf9a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.503629] env[65726]: DEBUG oslo_vmware.api [None req-11e4b8a8-60d9-4939-a6a9-6a357b669b13 tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] Waiting for the task: (returnval){ [ 735.503629] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52472526-d68f-4057-aeff-754c782621ac" [ 735.503629] env[65726]: _type = "Task" [ 735.503629] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 735.515423] env[65726]: DEBUG oslo_vmware.api [None req-11e4b8a8-60d9-4939-a6a9-6a357b669b13 tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52472526-d68f-4057-aeff-754c782621ac, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 735.620308] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5ae556a9-54a3-4f16-bd92-05c390a03c24 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 735.631743] env[65726]: DEBUG oslo_vmware.api [None req-61a6387d-c5ed-4ab7-853e-9ba63af33205 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': task-5115740, 'name': CloneVM_Task} progress is 94%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 735.734853] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89c38617-fbf7-4fe6-8975-6aaecf6e5ec7 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.742818] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d228124f-d353-47bd-9231-1bfe77e96ef9 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.774708] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80201d17-3aac-4eb3-885f-59c4f045a09e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.783109] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3b7486e-a38b-4464-a2fc-2449c30e0543 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.801020] env[65726]: DEBUG nova.compute.provider_tree [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Updating inventory in ProviderTree for provider 07c4692f-bdb4-4058-9173-ff9664830295 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 735.989325] env[65726]: DEBUG nova.compute.manager [None req-462b38ac-85e6-4728-81f8-4b729f9b1c26 tempest-ServersAaction247Test-712120018 tempest-ServersAaction247Test-712120018-project-member] [instance: ad48cadd-9bb6-4191-a68a-5c8fdaeebd44] Starting instance... {{(pid=65726) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 736.019903] env[65726]: DEBUG oslo_vmware.api [None req-11e4b8a8-60d9-4939-a6a9-6a357b669b13 tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52472526-d68f-4057-aeff-754c782621ac, 'name': SearchDatastore_Task, 'duration_secs': 0.012863} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 736.019903] env[65726]: DEBUG oslo_concurrency.lockutils [None req-11e4b8a8-60d9-4939-a6a9-6a357b669b13 tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 736.019903] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-11e4b8a8-60d9-4939-a6a9-6a357b669b13 tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] [instance: 83573c2b-9448-456f-8fd2-b19661dd6cc4] Processing image b52362a3-ee8a-4cbf-b06f-513b0cc8f95c {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 736.019903] env[65726]: DEBUG oslo_concurrency.lockutils [None req-11e4b8a8-60d9-4939-a6a9-6a357b669b13 tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 736.020068] env[65726]: DEBUG oslo_concurrency.lockutils [None req-11e4b8a8-60d9-4939-a6a9-6a357b669b13 tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 736.020151] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-11e4b8a8-60d9-4939-a6a9-6a357b669b13 tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 736.020449] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-802f4954-b13c-488d-829c-b5dd53720ecb {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.034492] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-11e4b8a8-60d9-4939-a6a9-6a357b669b13 tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 736.034492] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-11e4b8a8-60d9-4939-a6a9-6a357b669b13 tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65726) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 736.034492] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ac3b2ecd-6a43-4f11-8e21-0c61d2f0834d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.043780] env[65726]: DEBUG oslo_vmware.api [None req-11e4b8a8-60d9-4939-a6a9-6a357b669b13 tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] Waiting for the task: (returnval){ [ 736.043780] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52d2d653-1b75-28e2-14ed-d16e95722c1c" [ 736.043780] env[65726]: _type = "Task" [ 736.043780] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 736.054626] env[65726]: DEBUG oslo_vmware.api [None req-11e4b8a8-60d9-4939-a6a9-6a357b669b13 tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52d2d653-1b75-28e2-14ed-d16e95722c1c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 736.132621] env[65726]: DEBUG oslo_vmware.api [None req-61a6387d-c5ed-4ab7-853e-9ba63af33205 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': task-5115740, 'name': CloneVM_Task, 'duration_secs': 1.466318} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 736.132876] env[65726]: INFO nova.virt.vmwareapi.vmops [None req-61a6387d-c5ed-4ab7-853e-9ba63af33205 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: bd839fe1-8801-4ba1-9c23-288453258b66] Created linked-clone VM from snapshot [ 736.133670] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4125186d-076b-4fc4-8262-d40fcf7de5e0 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.141790] env[65726]: DEBUG nova.virt.vmwareapi.images [None req-61a6387d-c5ed-4ab7-853e-9ba63af33205 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: bd839fe1-8801-4ba1-9c23-288453258b66] Uploading image 249aae9c-fa0a-456d-bd7e-ad65ba07d625 {{(pid=65726) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 736.166359] env[65726]: DEBUG oslo_vmware.rw_handles [None req-61a6387d-c5ed-4ab7-853e-9ba63af33205 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 736.166359] env[65726]: value = "vm-995104" [ 736.166359] env[65726]: _type = "VirtualMachine" [ 736.166359] env[65726]: }. {{(pid=65726) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 736.166681] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-5bec79af-e8d3-498c-9375-4fbcb46dc619 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.174820] env[65726]: DEBUG oslo_vmware.rw_handles [None req-61a6387d-c5ed-4ab7-853e-9ba63af33205 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Lease: (returnval){ [ 736.174820] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]520c6feb-fcc2-bf6c-ee01-306905f68f9d" [ 736.174820] env[65726]: _type = "HttpNfcLease" [ 736.174820] env[65726]: } obtained for exporting VM: (result){ [ 736.174820] env[65726]: value = "vm-995104" [ 736.174820] env[65726]: _type = "VirtualMachine" [ 736.174820] env[65726]: }. {{(pid=65726) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 736.175068] env[65726]: DEBUG oslo_vmware.api [None req-61a6387d-c5ed-4ab7-853e-9ba63af33205 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Waiting for the lease: (returnval){ [ 736.175068] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]520c6feb-fcc2-bf6c-ee01-306905f68f9d" [ 736.175068] env[65726]: _type = "HttpNfcLease" [ 736.175068] env[65726]: } to be ready. {{(pid=65726) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 736.182659] env[65726]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 736.182659] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]520c6feb-fcc2-bf6c-ee01-306905f68f9d" [ 736.182659] env[65726]: _type = "HttpNfcLease" [ 736.182659] env[65726]: } is initializing. {{(pid=65726) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 736.333795] env[65726]: DEBUG nova.scheduler.client.report [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Updated inventory for provider 07c4692f-bdb4-4058-9173-ff9664830295 with generation 64 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 736.334127] env[65726]: DEBUG nova.compute.provider_tree [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Updating resource provider 07c4692f-bdb4-4058-9173-ff9664830295 generation from 64 to 65 during operation: update_inventory {{(pid=65726) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 736.334367] env[65726]: DEBUG nova.compute.provider_tree [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Updating inventory in ProviderTree for provider 07c4692f-bdb4-4058-9173-ff9664830295 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 736.534299] env[65726]: DEBUG oslo_concurrency.lockutils [None req-462b38ac-85e6-4728-81f8-4b729f9b1c26 tempest-ServersAaction247Test-712120018 tempest-ServersAaction247Test-712120018-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 736.561691] env[65726]: DEBUG oslo_vmware.api [None req-11e4b8a8-60d9-4939-a6a9-6a357b669b13 tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52d2d653-1b75-28e2-14ed-d16e95722c1c, 'name': SearchDatastore_Task, 'duration_secs': 0.019339} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 736.562878] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f5ecce19-247f-4827-bbf0-89f690c1a5cf {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.570199] env[65726]: DEBUG oslo_vmware.api [None req-11e4b8a8-60d9-4939-a6a9-6a357b669b13 tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] Waiting for the task: (returnval){ [ 736.570199] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52ff2c0c-138d-df41-9ba9-8d28733aef07" [ 736.570199] env[65726]: _type = "Task" [ 736.570199] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 736.583042] env[65726]: DEBUG oslo_vmware.api [None req-11e4b8a8-60d9-4939-a6a9-6a357b669b13 tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52ff2c0c-138d-df41-9ba9-8d28733aef07, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 736.687670] env[65726]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 736.687670] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]520c6feb-fcc2-bf6c-ee01-306905f68f9d" [ 736.687670] env[65726]: _type = "HttpNfcLease" [ 736.687670] env[65726]: } is ready. {{(pid=65726) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 736.687991] env[65726]: DEBUG oslo_vmware.rw_handles [None req-61a6387d-c5ed-4ab7-853e-9ba63af33205 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 736.687991] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]520c6feb-fcc2-bf6c-ee01-306905f68f9d" [ 736.687991] env[65726]: _type = "HttpNfcLease" [ 736.687991] env[65726]: }. {{(pid=65726) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 736.688836] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28f42f05-3195-4470-b968-c3921e2f1d4b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.700251] env[65726]: DEBUG oslo_vmware.rw_handles [None req-61a6387d-c5ed-4ab7-853e-9ba63af33205 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/526f27e0-8c7b-4c32-362a-63165cd7aeea/disk-0.vmdk from lease info. {{(pid=65726) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 736.700528] env[65726]: DEBUG oslo_vmware.rw_handles [None req-61a6387d-c5ed-4ab7-853e-9ba63af33205 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/526f27e0-8c7b-4c32-362a-63165cd7aeea/disk-0.vmdk for reading. {{(pid=65726) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 736.802357] env[65726]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-87ce0ff7-e3ae-4f01-ba5d-0c15af21f5b3 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.843354] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=65726) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1097}} [ 736.843354] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 10.267s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 736.843883] env[65726]: DEBUG oslo_concurrency.lockutils [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 44.770s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 736.845773] env[65726]: INFO nova.compute.claims [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] [instance: ca67d1a4-d304-4d06-a436-f257b60ca080] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 737.083477] env[65726]: DEBUG oslo_vmware.api [None req-11e4b8a8-60d9-4939-a6a9-6a357b669b13 tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52ff2c0c-138d-df41-9ba9-8d28733aef07, 'name': SearchDatastore_Task, 'duration_secs': 0.012705} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 737.083992] env[65726]: DEBUG oslo_concurrency.lockutils [None req-11e4b8a8-60d9-4939-a6a9-6a357b669b13 tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 737.084528] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-11e4b8a8-60d9-4939-a6a9-6a357b669b13 tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] 83573c2b-9448-456f-8fd2-b19661dd6cc4/83573c2b-9448-456f-8fd2-b19661dd6cc4.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 737.084704] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2ddd42f0-80c1-4ab7-9951-4d9946d42b1c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.093011] env[65726]: DEBUG oslo_vmware.api [None req-11e4b8a8-60d9-4939-a6a9-6a357b669b13 tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] Waiting for the task: (returnval){ [ 737.093011] env[65726]: value = "task-5115742" [ 737.093011] env[65726]: _type = "Task" [ 737.093011] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 737.109956] env[65726]: DEBUG oslo_vmware.api [None req-11e4b8a8-60d9-4939-a6a9-6a357b669b13 tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] Task: {'id': task-5115742, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 737.148312] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5489bce2-093c-4c74-b145-6df2a52e95cf tempest-ServersTestBootFromVolume-762458528 tempest-ServersTestBootFromVolume-762458528-project-member] Acquiring lock "5a8d4357-4b44-4a19-b1da-42d188c38adc" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 737.148789] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5489bce2-093c-4c74-b145-6df2a52e95cf tempest-ServersTestBootFromVolume-762458528 tempest-ServersTestBootFromVolume-762458528-project-member] Lock "5a8d4357-4b44-4a19-b1da-42d188c38adc" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 737.150642] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5489bce2-093c-4c74-b145-6df2a52e95cf tempest-ServersTestBootFromVolume-762458528 tempest-ServersTestBootFromVolume-762458528-project-member] Acquiring lock "5a8d4357-4b44-4a19-b1da-42d188c38adc-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 737.150642] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5489bce2-093c-4c74-b145-6df2a52e95cf tempest-ServersTestBootFromVolume-762458528 tempest-ServersTestBootFromVolume-762458528-project-member] Lock "5a8d4357-4b44-4a19-b1da-42d188c38adc-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 737.150642] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5489bce2-093c-4c74-b145-6df2a52e95cf tempest-ServersTestBootFromVolume-762458528 tempest-ServersTestBootFromVolume-762458528-project-member] Lock "5a8d4357-4b44-4a19-b1da-42d188c38adc-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 737.153240] env[65726]: INFO nova.compute.manager [None req-5489bce2-093c-4c74-b145-6df2a52e95cf tempest-ServersTestBootFromVolume-762458528 tempest-ServersTestBootFromVolume-762458528-project-member] [instance: 5a8d4357-4b44-4a19-b1da-42d188c38adc] Terminating instance [ 737.612681] env[65726]: DEBUG oslo_vmware.api [None req-11e4b8a8-60d9-4939-a6a9-6a357b669b13 tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] Task: {'id': task-5115742, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 737.659967] env[65726]: DEBUG nova.compute.manager [None req-5489bce2-093c-4c74-b145-6df2a52e95cf tempest-ServersTestBootFromVolume-762458528 tempest-ServersTestBootFromVolume-762458528-project-member] [instance: 5a8d4357-4b44-4a19-b1da-42d188c38adc] Start destroying the instance on the hypervisor. {{(pid=65726) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 737.660282] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-5489bce2-093c-4c74-b145-6df2a52e95cf tempest-ServersTestBootFromVolume-762458528 tempest-ServersTestBootFromVolume-762458528-project-member] [instance: 5a8d4357-4b44-4a19-b1da-42d188c38adc] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 737.663115] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a3fb200c-4f6f-4cbc-9126-1970916a0a78 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.672958] env[65726]: DEBUG oslo_vmware.api [None req-5489bce2-093c-4c74-b145-6df2a52e95cf tempest-ServersTestBootFromVolume-762458528 tempest-ServersTestBootFromVolume-762458528-project-member] Waiting for the task: (returnval){ [ 737.672958] env[65726]: value = "task-5115743" [ 737.672958] env[65726]: _type = "Task" [ 737.672958] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 737.687197] env[65726]: DEBUG oslo_vmware.api [None req-5489bce2-093c-4c74-b145-6df2a52e95cf tempest-ServersTestBootFromVolume-762458528 tempest-ServersTestBootFromVolume-762458528-project-member] Task: {'id': task-5115743, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 738.109908] env[65726]: DEBUG oslo_vmware.api [None req-11e4b8a8-60d9-4939-a6a9-6a357b669b13 tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] Task: {'id': task-5115742, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.645884} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 738.110431] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-11e4b8a8-60d9-4939-a6a9-6a357b669b13 tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] 83573c2b-9448-456f-8fd2-b19661dd6cc4/83573c2b-9448-456f-8fd2-b19661dd6cc4.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 738.110742] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-11e4b8a8-60d9-4939-a6a9-6a357b669b13 tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] [instance: 83573c2b-9448-456f-8fd2-b19661dd6cc4] Extending root virtual disk to 1048576 {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 738.111890] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-834cdf70-3c5c-4f76-a5c2-5604adb3c64d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.120762] env[65726]: DEBUG oslo_vmware.api [None req-11e4b8a8-60d9-4939-a6a9-6a357b669b13 tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] Waiting for the task: (returnval){ [ 738.120762] env[65726]: value = "task-5115744" [ 738.120762] env[65726]: _type = "Task" [ 738.120762] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 738.136253] env[65726]: DEBUG oslo_vmware.api [None req-11e4b8a8-60d9-4939-a6a9-6a357b669b13 tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] Task: {'id': task-5115744, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 738.187464] env[65726]: DEBUG oslo_vmware.api [None req-5489bce2-093c-4c74-b145-6df2a52e95cf tempest-ServersTestBootFromVolume-762458528 tempest-ServersTestBootFromVolume-762458528-project-member] Task: {'id': task-5115743, 'name': PowerOffVM_Task, 'duration_secs': 0.304393} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 738.187825] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-5489bce2-093c-4c74-b145-6df2a52e95cf tempest-ServersTestBootFromVolume-762458528 tempest-ServersTestBootFromVolume-762458528-project-member] [instance: 5a8d4357-4b44-4a19-b1da-42d188c38adc] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 738.188081] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-5489bce2-093c-4c74-b145-6df2a52e95cf tempest-ServersTestBootFromVolume-762458528 tempest-ServersTestBootFromVolume-762458528-project-member] [instance: 5a8d4357-4b44-4a19-b1da-42d188c38adc] Volume detach. Driver type: vmdk {{(pid=65726) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 738.188415] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-5489bce2-093c-4c74-b145-6df2a52e95cf tempest-ServersTestBootFromVolume-762458528 tempest-ServersTestBootFromVolume-762458528-project-member] [instance: 5a8d4357-4b44-4a19-b1da-42d188c38adc] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-995020', 'volume_id': 'b11cb57f-92f1-416d-b59b-70cffe1353ff', 'name': 'volume-b11cb57f-92f1-416d-b59b-70cffe1353ff', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '5a8d4357-4b44-4a19-b1da-42d188c38adc', 'attached_at': '', 'detached_at': '', 'volume_id': 'b11cb57f-92f1-416d-b59b-70cffe1353ff', 'serial': 'b11cb57f-92f1-416d-b59b-70cffe1353ff'} {{(pid=65726) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 738.189235] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ac5813b-7086-4a32-baf3-90afe1000f7d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.216766] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-366d44b0-6e48-4dea-b17c-2de352b730af {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.227018] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25d820dc-235f-48e1-993b-3b3e0a33ef10 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.251023] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d25d73b-dc61-4470-a82c-b780142ec9ef {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.267605] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-5489bce2-093c-4c74-b145-6df2a52e95cf tempest-ServersTestBootFromVolume-762458528 tempest-ServersTestBootFromVolume-762458528-project-member] The volume has not been displaced from its original location: [datastore1] volume-b11cb57f-92f1-416d-b59b-70cffe1353ff/volume-b11cb57f-92f1-416d-b59b-70cffe1353ff.vmdk. No consolidation needed. {{(pid=65726) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 738.273290] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-5489bce2-093c-4c74-b145-6df2a52e95cf tempest-ServersTestBootFromVolume-762458528 tempest-ServersTestBootFromVolume-762458528-project-member] [instance: 5a8d4357-4b44-4a19-b1da-42d188c38adc] Reconfiguring VM instance instance-00000015 to detach disk 2000 {{(pid=65726) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 738.276346] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-46fb223c-7680-41d6-ac27-f79a89489d65 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.301020] env[65726]: DEBUG oslo_vmware.api [None req-5489bce2-093c-4c74-b145-6df2a52e95cf tempest-ServersTestBootFromVolume-762458528 tempest-ServersTestBootFromVolume-762458528-project-member] Waiting for the task: (returnval){ [ 738.301020] env[65726]: value = "task-5115745" [ 738.301020] env[65726]: _type = "Task" [ 738.301020] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 738.312695] env[65726]: DEBUG oslo_vmware.api [None req-5489bce2-093c-4c74-b145-6df2a52e95cf tempest-ServersTestBootFromVolume-762458528 tempest-ServersTestBootFromVolume-762458528-project-member] Task: {'id': task-5115745, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 738.635701] env[65726]: DEBUG oslo_vmware.api [None req-11e4b8a8-60d9-4939-a6a9-6a357b669b13 tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] Task: {'id': task-5115744, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.087399} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 738.636041] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-11e4b8a8-60d9-4939-a6a9-6a357b669b13 tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] [instance: 83573c2b-9448-456f-8fd2-b19661dd6cc4] Extended root virtual disk {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 738.636889] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3798e03d-4f69-49fb-82cb-4fc46f9ecc0b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.643237] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-edc835e8-92a1-444d-8297-e8cb94a3ec33 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.669541] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-11e4b8a8-60d9-4939-a6a9-6a357b669b13 tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] [instance: 83573c2b-9448-456f-8fd2-b19661dd6cc4] Reconfiguring VM instance instance-00000022 to attach disk [datastore1] 83573c2b-9448-456f-8fd2-b19661dd6cc4/83573c2b-9448-456f-8fd2-b19661dd6cc4.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 738.670463] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3e55ad68-a8a0-4284-b9bd-a8822e076169 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.697169] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0822585c-5afc-413e-b9f9-cf3c48348225 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.702066] env[65726]: DEBUG oslo_vmware.api [None req-11e4b8a8-60d9-4939-a6a9-6a357b669b13 tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] Waiting for the task: (returnval){ [ 738.702066] env[65726]: value = "task-5115746" [ 738.702066] env[65726]: _type = "Task" [ 738.702066] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 738.736166] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de3e4752-f5fc-4717-b4e2-77bf29dad209 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.738907] env[65726]: DEBUG oslo_vmware.api [None req-11e4b8a8-60d9-4939-a6a9-6a357b669b13 tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] Task: {'id': task-5115746, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 738.746311] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6ef0837-eade-49f9-9cb3-44b6b8b04cae {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.761638] env[65726]: DEBUG nova.compute.provider_tree [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 738.810147] env[65726]: DEBUG oslo_vmware.api [None req-5489bce2-093c-4c74-b145-6df2a52e95cf tempest-ServersTestBootFromVolume-762458528 tempest-ServersTestBootFromVolume-762458528-project-member] Task: {'id': task-5115745, 'name': ReconfigVM_Task, 'duration_secs': 0.271883} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 738.810447] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-5489bce2-093c-4c74-b145-6df2a52e95cf tempest-ServersTestBootFromVolume-762458528 tempest-ServersTestBootFromVolume-762458528-project-member] [instance: 5a8d4357-4b44-4a19-b1da-42d188c38adc] Reconfigured VM instance instance-00000015 to detach disk 2000 {{(pid=65726) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 738.815276] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-dcf664ae-0e49-47b4-ab03-715beef3046e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.832696] env[65726]: DEBUG oslo_vmware.api [None req-5489bce2-093c-4c74-b145-6df2a52e95cf tempest-ServersTestBootFromVolume-762458528 tempest-ServersTestBootFromVolume-762458528-project-member] Waiting for the task: (returnval){ [ 738.832696] env[65726]: value = "task-5115747" [ 738.832696] env[65726]: _type = "Task" [ 738.832696] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 738.843193] env[65726]: DEBUG oslo_vmware.api [None req-5489bce2-093c-4c74-b145-6df2a52e95cf tempest-ServersTestBootFromVolume-762458528 tempest-ServersTestBootFromVolume-762458528-project-member] Task: {'id': task-5115747, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 739.215076] env[65726]: DEBUG oslo_vmware.api [None req-11e4b8a8-60d9-4939-a6a9-6a357b669b13 tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] Task: {'id': task-5115746, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 739.265502] env[65726]: DEBUG nova.scheduler.client.report [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 739.343316] env[65726]: DEBUG oslo_vmware.api [None req-5489bce2-093c-4c74-b145-6df2a52e95cf tempest-ServersTestBootFromVolume-762458528 tempest-ServersTestBootFromVolume-762458528-project-member] Task: {'id': task-5115747, 'name': ReconfigVM_Task, 'duration_secs': 0.275065} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 739.343643] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-5489bce2-093c-4c74-b145-6df2a52e95cf tempest-ServersTestBootFromVolume-762458528 tempest-ServersTestBootFromVolume-762458528-project-member] [instance: 5a8d4357-4b44-4a19-b1da-42d188c38adc] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-995020', 'volume_id': 'b11cb57f-92f1-416d-b59b-70cffe1353ff', 'name': 'volume-b11cb57f-92f1-416d-b59b-70cffe1353ff', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '5a8d4357-4b44-4a19-b1da-42d188c38adc', 'attached_at': '', 'detached_at': '', 'volume_id': 'b11cb57f-92f1-416d-b59b-70cffe1353ff', 'serial': 'b11cb57f-92f1-416d-b59b-70cffe1353ff'} {{(pid=65726) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 739.343899] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-5489bce2-093c-4c74-b145-6df2a52e95cf tempest-ServersTestBootFromVolume-762458528 tempest-ServersTestBootFromVolume-762458528-project-member] [instance: 5a8d4357-4b44-4a19-b1da-42d188c38adc] Destroying instance {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 739.344756] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebb1da5f-ddba-485c-8a2f-aabe844d9a03 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.352454] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-5489bce2-093c-4c74-b145-6df2a52e95cf tempest-ServersTestBootFromVolume-762458528 tempest-ServersTestBootFromVolume-762458528-project-member] [instance: 5a8d4357-4b44-4a19-b1da-42d188c38adc] Unregistering the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 739.352752] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-db892190-c977-44ca-8bff-843ee9a670fd {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.472715] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-5489bce2-093c-4c74-b145-6df2a52e95cf tempest-ServersTestBootFromVolume-762458528 tempest-ServersTestBootFromVolume-762458528-project-member] [instance: 5a8d4357-4b44-4a19-b1da-42d188c38adc] Unregistered the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 739.473077] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-5489bce2-093c-4c74-b145-6df2a52e95cf tempest-ServersTestBootFromVolume-762458528 tempest-ServersTestBootFromVolume-762458528-project-member] [instance: 5a8d4357-4b44-4a19-b1da-42d188c38adc] Deleting contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 739.473299] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-5489bce2-093c-4c74-b145-6df2a52e95cf tempest-ServersTestBootFromVolume-762458528 tempest-ServersTestBootFromVolume-762458528-project-member] Deleting the datastore file [datastore1] 5a8d4357-4b44-4a19-b1da-42d188c38adc {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 739.473610] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f520a01a-fd71-422b-971a-87fbfc198a6c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.482736] env[65726]: DEBUG oslo_vmware.api [None req-5489bce2-093c-4c74-b145-6df2a52e95cf tempest-ServersTestBootFromVolume-762458528 tempest-ServersTestBootFromVolume-762458528-project-member] Waiting for the task: (returnval){ [ 739.482736] env[65726]: value = "task-5115749" [ 739.482736] env[65726]: _type = "Task" [ 739.482736] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 739.491932] env[65726]: DEBUG oslo_vmware.api [None req-5489bce2-093c-4c74-b145-6df2a52e95cf tempest-ServersTestBootFromVolume-762458528 tempest-ServersTestBootFromVolume-762458528-project-member] Task: {'id': task-5115749, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 739.714559] env[65726]: DEBUG oslo_vmware.api [None req-11e4b8a8-60d9-4939-a6a9-6a357b669b13 tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] Task: {'id': task-5115746, 'name': ReconfigVM_Task, 'duration_secs': 0.777557} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 739.714889] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-11e4b8a8-60d9-4939-a6a9-6a357b669b13 tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] [instance: 83573c2b-9448-456f-8fd2-b19661dd6cc4] Reconfigured VM instance instance-00000022 to attach disk [datastore1] 83573c2b-9448-456f-8fd2-b19661dd6cc4/83573c2b-9448-456f-8fd2-b19661dd6cc4.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 739.715545] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8be4c665-840e-427a-844a-df677406ac34 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.723592] env[65726]: DEBUG oslo_vmware.api [None req-11e4b8a8-60d9-4939-a6a9-6a357b669b13 tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] Waiting for the task: (returnval){ [ 739.723592] env[65726]: value = "task-5115750" [ 739.723592] env[65726]: _type = "Task" [ 739.723592] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 739.737736] env[65726]: DEBUG oslo_vmware.api [None req-11e4b8a8-60d9-4939-a6a9-6a357b669b13 tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] Task: {'id': task-5115750, 'name': Rename_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 739.770959] env[65726]: DEBUG oslo_concurrency.lockutils [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.927s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 739.771577] env[65726]: DEBUG nova.compute.manager [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] [instance: ca67d1a4-d304-4d06-a436-f257b60ca080] Start building networks asynchronously for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 739.774792] env[65726]: DEBUG oslo_concurrency.lockutils [None req-ca40f660-02e2-41e4-8187-3f53d529f31e tempest-ServersAdmin275Test-1899430852 tempest-ServersAdmin275Test-1899430852-project-admin] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 46.447s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 739.778086] env[65726]: DEBUG nova.objects.instance [None req-ca40f660-02e2-41e4-8187-3f53d529f31e tempest-ServersAdmin275Test-1899430852 tempest-ServersAdmin275Test-1899430852-project-admin] [instance: 3b933684-ac19-44b0-a49d-6af45501e166] Trying to apply a migration context that does not seem to be set for this instance {{(pid=65726) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 739.997032] env[65726]: DEBUG oslo_vmware.api [None req-5489bce2-093c-4c74-b145-6df2a52e95cf tempest-ServersTestBootFromVolume-762458528 tempest-ServersTestBootFromVolume-762458528-project-member] Task: {'id': task-5115749, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.137942} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 739.997032] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-5489bce2-093c-4c74-b145-6df2a52e95cf tempest-ServersTestBootFromVolume-762458528 tempest-ServersTestBootFromVolume-762458528-project-member] Deleted the datastore file {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 739.997032] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-5489bce2-093c-4c74-b145-6df2a52e95cf tempest-ServersTestBootFromVolume-762458528 tempest-ServersTestBootFromVolume-762458528-project-member] [instance: 5a8d4357-4b44-4a19-b1da-42d188c38adc] Deleted contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 739.997032] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-5489bce2-093c-4c74-b145-6df2a52e95cf tempest-ServersTestBootFromVolume-762458528 tempest-ServersTestBootFromVolume-762458528-project-member] [instance: 5a8d4357-4b44-4a19-b1da-42d188c38adc] Instance destroyed {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 739.997032] env[65726]: INFO nova.compute.manager [None req-5489bce2-093c-4c74-b145-6df2a52e95cf tempest-ServersTestBootFromVolume-762458528 tempest-ServersTestBootFromVolume-762458528-project-member] [instance: 5a8d4357-4b44-4a19-b1da-42d188c38adc] Took 2.34 seconds to destroy the instance on the hypervisor. [ 739.997366] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-5489bce2-093c-4c74-b145-6df2a52e95cf tempest-ServersTestBootFromVolume-762458528 tempest-ServersTestBootFromVolume-762458528-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 739.997366] env[65726]: DEBUG nova.compute.manager [-] [instance: 5a8d4357-4b44-4a19-b1da-42d188c38adc] Deallocating network for instance {{(pid=65726) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 739.997366] env[65726]: DEBUG nova.network.neutron [-] [instance: 5a8d4357-4b44-4a19-b1da-42d188c38adc] deallocate_for_instance() {{(pid=65726) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 739.997366] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 739.997366] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 740.036459] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 740.037088] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 740.236197] env[65726]: DEBUG oslo_vmware.api [None req-11e4b8a8-60d9-4939-a6a9-6a357b669b13 tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] Task: {'id': task-5115750, 'name': Rename_Task, 'duration_secs': 0.213351} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 740.236635] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-11e4b8a8-60d9-4939-a6a9-6a357b669b13 tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] [instance: 83573c2b-9448-456f-8fd2-b19661dd6cc4] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 740.236931] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2eff5a9b-7988-4b5c-aafe-ddf201d97665 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.252214] env[65726]: DEBUG oslo_vmware.api [None req-11e4b8a8-60d9-4939-a6a9-6a357b669b13 tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] Waiting for the task: (returnval){ [ 740.252214] env[65726]: value = "task-5115751" [ 740.252214] env[65726]: _type = "Task" [ 740.252214] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 740.265489] env[65726]: DEBUG oslo_vmware.api [None req-11e4b8a8-60d9-4939-a6a9-6a357b669b13 tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] Task: {'id': task-5115751, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 740.281135] env[65726]: DEBUG nova.compute.utils [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Using /dev/sd instead of None {{(pid=65726) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 740.286366] env[65726]: DEBUG nova.compute.manager [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] [instance: ca67d1a4-d304-4d06-a436-f257b60ca080] Allocating IP information in the background. {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 740.289412] env[65726]: DEBUG nova.network.neutron [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] [instance: ca67d1a4-d304-4d06-a436-f257b60ca080] allocate_for_instance() {{(pid=65726) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 740.289841] env[65726]: WARNING neutronclient.v2_0.client [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 740.290160] env[65726]: WARNING neutronclient.v2_0.client [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 740.290817] env[65726]: WARNING openstack [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 740.291170] env[65726]: WARNING openstack [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 740.303598] env[65726]: DEBUG oslo_concurrency.lockutils [None req-ca40f660-02e2-41e4-8187-3f53d529f31e tempest-ServersAdmin275Test-1899430852 tempest-ServersAdmin275Test-1899430852-project-admin] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 0.529s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 740.304685] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e30832c9-8006-4ee6-8d81-96e2fa3ed74f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 39.911s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 740.304865] env[65726]: DEBUG nova.objects.instance [None req-e30832c9-8006-4ee6-8d81-96e2fa3ed74f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] [instance: ddd36db0-28ad-4582-994b-7eef8b08ed46] Trying to apply a migration context that does not seem to be set for this instance {{(pid=65726) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 740.356352] env[65726]: DEBUG nova.policy [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '838543f31e974acf83678c7f5c27b102', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7c3641c4a34b4bf8b09403220e153e65', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65726) authorize /opt/stack/nova/nova/policy.py:192}} [ 740.674204] env[65726]: DEBUG nova.network.neutron [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] [instance: ca67d1a4-d304-4d06-a436-f257b60ca080] Successfully created port: 8789d2e9-5db3-4e5d-bdb2-14160bd43a1d {{(pid=65726) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 740.769136] env[65726]: DEBUG oslo_vmware.api [None req-11e4b8a8-60d9-4939-a6a9-6a357b669b13 tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] Task: {'id': task-5115751, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 740.787019] env[65726]: DEBUG nova.compute.manager [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] [instance: ca67d1a4-d304-4d06-a436-f257b60ca080] Start building block device mappings for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 741.030640] env[65726]: DEBUG nova.network.neutron [-] [instance: 5a8d4357-4b44-4a19-b1da-42d188c38adc] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 741.048103] env[65726]: DEBUG oslo_concurrency.lockutils [None req-ae076e22-617f-47ad-a346-b93664495097 tempest-ServersNegativeTestMultiTenantJSON-944098855 tempest-ServersNegativeTestMultiTenantJSON-944098855-project-member] Acquiring lock "018c8ac5-66c4-4a9a-ab45-85e84f92b4fe" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 741.048604] env[65726]: DEBUG oslo_concurrency.lockutils [None req-ae076e22-617f-47ad-a346-b93664495097 tempest-ServersNegativeTestMultiTenantJSON-944098855 tempest-ServersNegativeTestMultiTenantJSON-944098855-project-member] Lock "018c8ac5-66c4-4a9a-ab45-85e84f92b4fe" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 741.102528] env[65726]: DEBUG nova.compute.manager [req-b9d1e408-4a82-4036-8015-b5fe177f375a req-e845791d-7a1b-4167-960c-08d2a27ac83d service nova] [instance: 5a8d4357-4b44-4a19-b1da-42d188c38adc] Received event network-vif-deleted-9aaacd82-31c1-4cfa-bbce-860b30db74af {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 741.266928] env[65726]: DEBUG oslo_vmware.api [None req-11e4b8a8-60d9-4939-a6a9-6a357b669b13 tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] Task: {'id': task-5115751, 'name': PowerOnVM_Task, 'duration_secs': 0.629641} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 741.268678] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-11e4b8a8-60d9-4939-a6a9-6a357b669b13 tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] [instance: 83573c2b-9448-456f-8fd2-b19661dd6cc4] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 741.268678] env[65726]: DEBUG nova.compute.manager [None req-11e4b8a8-60d9-4939-a6a9-6a357b669b13 tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] [instance: 83573c2b-9448-456f-8fd2-b19661dd6cc4] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 741.268678] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f440fd60-c13b-47dd-9484-443d54fc9bd0 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.318861] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e30832c9-8006-4ee6-8d81-96e2fa3ed74f tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.014s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 741.320320] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b0adde83-84f2-4fa3-84bf-8f387eee1293 tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 40.619s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 741.322532] env[65726]: INFO nova.compute.claims [None req-b0adde83-84f2-4fa3-84bf-8f387eee1293 tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] [instance: 3827eed1-b7a1-4922-a4de-f0c038cb8566] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 741.533340] env[65726]: INFO nova.compute.manager [-] [instance: 5a8d4357-4b44-4a19-b1da-42d188c38adc] Took 1.54 seconds to deallocate network for instance. [ 741.790104] env[65726]: DEBUG oslo_concurrency.lockutils [None req-11e4b8a8-60d9-4939-a6a9-6a357b669b13 tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 741.796858] env[65726]: DEBUG nova.compute.manager [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] [instance: ca67d1a4-d304-4d06-a436-f257b60ca080] Start spawning the instance on the hypervisor. {{(pid=65726) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 741.827739] env[65726]: DEBUG nova.virt.hardware [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T19:28:34Z,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1539f06fa3534e90acbc3a60c4b8bd3f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T19:28:35Z,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 741.828212] env[65726]: DEBUG nova.virt.hardware [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 741.828413] env[65726]: DEBUG nova.virt.hardware [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 741.828604] env[65726]: DEBUG nova.virt.hardware [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 741.828809] env[65726]: DEBUG nova.virt.hardware [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 741.829084] env[65726]: DEBUG nova.virt.hardware [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 741.830061] env[65726]: DEBUG nova.virt.hardware [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 741.830061] env[65726]: DEBUG nova.virt.hardware [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 741.830061] env[65726]: DEBUG nova.virt.hardware [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 741.830206] env[65726]: DEBUG nova.virt.hardware [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 741.830345] env[65726]: DEBUG nova.virt.hardware [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 741.834509] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76fe53d7-23ec-43d8-96d9-92bbc9f9598a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.845948] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-417bb717-343a-4ff9-b680-d9602cfc4abb {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.111669] env[65726]: INFO nova.compute.manager [None req-5489bce2-093c-4c74-b145-6df2a52e95cf tempest-ServersTestBootFromVolume-762458528 tempest-ServersTestBootFromVolume-762458528-project-member] [instance: 5a8d4357-4b44-4a19-b1da-42d188c38adc] Took 0.58 seconds to detach 1 volumes for instance. [ 742.113979] env[65726]: DEBUG nova.compute.manager [None req-5489bce2-093c-4c74-b145-6df2a52e95cf tempest-ServersTestBootFromVolume-762458528 tempest-ServersTestBootFromVolume-762458528-project-member] [instance: 5a8d4357-4b44-4a19-b1da-42d188c38adc] Deleting volume: b11cb57f-92f1-416d-b59b-70cffe1353ff {{(pid=65726) _cleanup_volumes /opt/stack/nova/nova/compute/manager.py:3319}} [ 742.293323] env[65726]: DEBUG nova.network.neutron [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] [instance: ca67d1a4-d304-4d06-a436-f257b60ca080] Successfully updated port: 8789d2e9-5db3-4e5d-bdb2-14160bd43a1d {{(pid=65726) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 742.668030] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5489bce2-093c-4c74-b145-6df2a52e95cf tempest-ServersTestBootFromVolume-762458528 tempest-ServersTestBootFromVolume-762458528-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 742.796972] env[65726]: DEBUG oslo_concurrency.lockutils [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Acquiring lock "refresh_cache-ca67d1a4-d304-4d06-a436-f257b60ca080" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 742.796972] env[65726]: DEBUG oslo_concurrency.lockutils [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Acquired lock "refresh_cache-ca67d1a4-d304-4d06-a436-f257b60ca080" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 742.796972] env[65726]: DEBUG nova.network.neutron [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] [instance: ca67d1a4-d304-4d06-a436-f257b60ca080] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 742.851159] env[65726]: DEBUG oslo_concurrency.lockutils [None req-fd5c9ab4-2b38-4729-88f4-0f437c7dacb5 tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] Acquiring lock "83573c2b-9448-456f-8fd2-b19661dd6cc4" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 742.851414] env[65726]: DEBUG oslo_concurrency.lockutils [None req-fd5c9ab4-2b38-4729-88f4-0f437c7dacb5 tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] Lock "83573c2b-9448-456f-8fd2-b19661dd6cc4" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 742.851621] env[65726]: DEBUG oslo_concurrency.lockutils [None req-fd5c9ab4-2b38-4729-88f4-0f437c7dacb5 tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] Acquiring lock "83573c2b-9448-456f-8fd2-b19661dd6cc4-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 742.851809] env[65726]: DEBUG oslo_concurrency.lockutils [None req-fd5c9ab4-2b38-4729-88f4-0f437c7dacb5 tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] Lock "83573c2b-9448-456f-8fd2-b19661dd6cc4-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 742.851961] env[65726]: DEBUG oslo_concurrency.lockutils [None req-fd5c9ab4-2b38-4729-88f4-0f437c7dacb5 tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] Lock "83573c2b-9448-456f-8fd2-b19661dd6cc4-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 742.854892] env[65726]: INFO nova.compute.manager [None req-fd5c9ab4-2b38-4729-88f4-0f437c7dacb5 tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] [instance: 83573c2b-9448-456f-8fd2-b19661dd6cc4] Terminating instance [ 742.899088] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a432c28-f371-44f7-b649-f6f4ef9cde98 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.908399] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8b289ab-aedc-48ad-b483-195bfe328912 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.940854] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee009faa-1534-487b-9ec4-0ab6d4369ab3 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.950652] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bbeb019-ff8f-408c-98a0-2bc2b2d2d6f9 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.966809] env[65726]: DEBUG nova.compute.provider_tree [None req-b0adde83-84f2-4fa3-84bf-8f387eee1293 tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 743.142562] env[65726]: DEBUG nova.compute.manager [req-610d135b-bb7d-4d70-8050-fa6d18dea786 req-54ca8ff5-4779-4684-9641-5a85dee41fe5 service nova] [instance: ca67d1a4-d304-4d06-a436-f257b60ca080] Received event network-vif-plugged-8789d2e9-5db3-4e5d-bdb2-14160bd43a1d {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 743.142562] env[65726]: DEBUG oslo_concurrency.lockutils [req-610d135b-bb7d-4d70-8050-fa6d18dea786 req-54ca8ff5-4779-4684-9641-5a85dee41fe5 service nova] Acquiring lock "ca67d1a4-d304-4d06-a436-f257b60ca080-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 743.142809] env[65726]: DEBUG oslo_concurrency.lockutils [req-610d135b-bb7d-4d70-8050-fa6d18dea786 req-54ca8ff5-4779-4684-9641-5a85dee41fe5 service nova] Lock "ca67d1a4-d304-4d06-a436-f257b60ca080-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 743.142960] env[65726]: DEBUG oslo_concurrency.lockutils [req-610d135b-bb7d-4d70-8050-fa6d18dea786 req-54ca8ff5-4779-4684-9641-5a85dee41fe5 service nova] Lock "ca67d1a4-d304-4d06-a436-f257b60ca080-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 743.143156] env[65726]: DEBUG nova.compute.manager [req-610d135b-bb7d-4d70-8050-fa6d18dea786 req-54ca8ff5-4779-4684-9641-5a85dee41fe5 service nova] [instance: ca67d1a4-d304-4d06-a436-f257b60ca080] No waiting events found dispatching network-vif-plugged-8789d2e9-5db3-4e5d-bdb2-14160bd43a1d {{(pid=65726) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 743.143329] env[65726]: WARNING nova.compute.manager [req-610d135b-bb7d-4d70-8050-fa6d18dea786 req-54ca8ff5-4779-4684-9641-5a85dee41fe5 service nova] [instance: ca67d1a4-d304-4d06-a436-f257b60ca080] Received unexpected event network-vif-plugged-8789d2e9-5db3-4e5d-bdb2-14160bd43a1d for instance with vm_state building and task_state spawning. [ 743.143498] env[65726]: DEBUG nova.compute.manager [req-610d135b-bb7d-4d70-8050-fa6d18dea786 req-54ca8ff5-4779-4684-9641-5a85dee41fe5 service nova] [instance: ca67d1a4-d304-4d06-a436-f257b60ca080] Received event network-changed-8789d2e9-5db3-4e5d-bdb2-14160bd43a1d {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 743.143644] env[65726]: DEBUG nova.compute.manager [req-610d135b-bb7d-4d70-8050-fa6d18dea786 req-54ca8ff5-4779-4684-9641-5a85dee41fe5 service nova] [instance: ca67d1a4-d304-4d06-a436-f257b60ca080] Refreshing instance network info cache due to event network-changed-8789d2e9-5db3-4e5d-bdb2-14160bd43a1d. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 743.144099] env[65726]: DEBUG oslo_concurrency.lockutils [req-610d135b-bb7d-4d70-8050-fa6d18dea786 req-54ca8ff5-4779-4684-9641-5a85dee41fe5 service nova] Acquiring lock "refresh_cache-ca67d1a4-d304-4d06-a436-f257b60ca080" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 743.300551] env[65726]: WARNING openstack [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 743.300991] env[65726]: WARNING openstack [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 743.306550] env[65726]: DEBUG nova.network.neutron [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] [instance: ca67d1a4-d304-4d06-a436-f257b60ca080] Instance cache missing network info. {{(pid=65726) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 743.360415] env[65726]: WARNING openstack [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 743.360877] env[65726]: WARNING openstack [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 743.369025] env[65726]: DEBUG oslo_concurrency.lockutils [None req-fd5c9ab4-2b38-4729-88f4-0f437c7dacb5 tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] Acquiring lock "refresh_cache-83573c2b-9448-456f-8fd2-b19661dd6cc4" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 743.369025] env[65726]: DEBUG oslo_concurrency.lockutils [None req-fd5c9ab4-2b38-4729-88f4-0f437c7dacb5 tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] Acquired lock "refresh_cache-83573c2b-9448-456f-8fd2-b19661dd6cc4" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 743.369025] env[65726]: DEBUG nova.network.neutron [None req-fd5c9ab4-2b38-4729-88f4-0f437c7dacb5 tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] [instance: 83573c2b-9448-456f-8fd2-b19661dd6cc4] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 743.436177] env[65726]: WARNING openstack [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 743.436695] env[65726]: WARNING openstack [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 743.470135] env[65726]: DEBUG nova.scheduler.client.report [None req-b0adde83-84f2-4fa3-84bf-8f387eee1293 tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 743.539965] env[65726]: DEBUG nova.network.neutron [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] [instance: ca67d1a4-d304-4d06-a436-f257b60ca080] Updating instance_info_cache with network_info: [{"id": "8789d2e9-5db3-4e5d-bdb2-14160bd43a1d", "address": "fa:16:3e:cc:40:7f", "network": {"id": "c68b8bcc-840e-459b-bee0-76a9ffdc08cb", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-2061916615-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7c3641c4a34b4bf8b09403220e153e65", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "aa0d7fd6-4fd3-4451-9208-8a3cf25e30e4", "external-id": "nsx-vlan-transportzone-545", "segmentation_id": 545, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8789d2e9-5d", "ovs_interfaceid": "8789d2e9-5db3-4e5d-bdb2-14160bd43a1d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 743.871136] env[65726]: WARNING openstack [None req-fd5c9ab4-2b38-4729-88f4-0f437c7dacb5 tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 743.871316] env[65726]: WARNING openstack [None req-fd5c9ab4-2b38-4729-88f4-0f437c7dacb5 tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 743.876565] env[65726]: DEBUG nova.network.neutron [None req-fd5c9ab4-2b38-4729-88f4-0f437c7dacb5 tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] [instance: 83573c2b-9448-456f-8fd2-b19661dd6cc4] Instance cache missing network info. {{(pid=65726) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 743.951840] env[65726]: DEBUG nova.network.neutron [None req-fd5c9ab4-2b38-4729-88f4-0f437c7dacb5 tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] [instance: 83573c2b-9448-456f-8fd2-b19661dd6cc4] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 743.975396] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b0adde83-84f2-4fa3-84bf-8f387eee1293 tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.655s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 743.975962] env[65726]: DEBUG nova.compute.manager [None req-b0adde83-84f2-4fa3-84bf-8f387eee1293 tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] [instance: 3827eed1-b7a1-4922-a4de-f0c038cb8566] Start building networks asynchronously for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 743.978785] env[65726]: DEBUG oslo_concurrency.lockutils [None req-f61c0733-7339-4ad4-9f78-b4a5ed0d0ff1 tempest-AttachInterfacesV270Test-133387050 tempest-AttachInterfacesV270Test-133387050-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 42.351s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 743.979084] env[65726]: DEBUG oslo_concurrency.lockutils [None req-f61c0733-7339-4ad4-9f78-b4a5ed0d0ff1 tempest-AttachInterfacesV270Test-133387050 tempest-AttachInterfacesV270Test-133387050-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 743.981923] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c19da911-fb8f-4bec-a53e-39a31fa2d36e tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 41.748s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 743.982138] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c19da911-fb8f-4bec-a53e-39a31fa2d36e tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 743.984098] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b479f246-d94b-4634-888f-dee4586c3272 tempest-ServerAddressesNegativeTestJSON-963263679 tempest-ServerAddressesNegativeTestJSON-963263679-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 39.591s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 743.985623] env[65726]: INFO nova.compute.claims [None req-b479f246-d94b-4634-888f-dee4586c3272 tempest-ServerAddressesNegativeTestJSON-963263679 tempest-ServerAddressesNegativeTestJSON-963263679-project-member] [instance: 00746508-d0d6-4dfa-9026-772398d004ee] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 744.021068] env[65726]: INFO nova.scheduler.client.report [None req-c19da911-fb8f-4bec-a53e-39a31fa2d36e tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] Deleted allocations for instance 3b933684-ac19-44b0-a49d-6af45501e166 [ 744.027728] env[65726]: INFO nova.scheduler.client.report [None req-f61c0733-7339-4ad4-9f78-b4a5ed0d0ff1 tempest-AttachInterfacesV270Test-133387050 tempest-AttachInterfacesV270Test-133387050-project-member] Deleted allocations for instance c977257e-dce0-4ecc-b42b-0f6ebab38797 [ 744.043212] env[65726]: DEBUG oslo_concurrency.lockutils [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Releasing lock "refresh_cache-ca67d1a4-d304-4d06-a436-f257b60ca080" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 744.043571] env[65726]: DEBUG nova.compute.manager [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] [instance: ca67d1a4-d304-4d06-a436-f257b60ca080] Instance network_info: |[{"id": "8789d2e9-5db3-4e5d-bdb2-14160bd43a1d", "address": "fa:16:3e:cc:40:7f", "network": {"id": "c68b8bcc-840e-459b-bee0-76a9ffdc08cb", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-2061916615-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7c3641c4a34b4bf8b09403220e153e65", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "aa0d7fd6-4fd3-4451-9208-8a3cf25e30e4", "external-id": "nsx-vlan-transportzone-545", "segmentation_id": 545, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8789d2e9-5d", "ovs_interfaceid": "8789d2e9-5db3-4e5d-bdb2-14160bd43a1d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 744.043889] env[65726]: DEBUG oslo_concurrency.lockutils [req-610d135b-bb7d-4d70-8050-fa6d18dea786 req-54ca8ff5-4779-4684-9641-5a85dee41fe5 service nova] Acquired lock "refresh_cache-ca67d1a4-d304-4d06-a436-f257b60ca080" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 744.044080] env[65726]: DEBUG nova.network.neutron [req-610d135b-bb7d-4d70-8050-fa6d18dea786 req-54ca8ff5-4779-4684-9641-5a85dee41fe5 service nova] [instance: ca67d1a4-d304-4d06-a436-f257b60ca080] Refreshing network info cache for port 8789d2e9-5db3-4e5d-bdb2-14160bd43a1d {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 744.046382] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] [instance: ca67d1a4-d304-4d06-a436-f257b60ca080] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:cc:40:7f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'aa0d7fd6-4fd3-4451-9208-8a3cf25e30e4', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '8789d2e9-5db3-4e5d-bdb2-14160bd43a1d', 'vif_model': 'vmxnet3'}] {{(pid=65726) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 744.059534] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 744.059534] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ca67d1a4-d304-4d06-a436-f257b60ca080] Creating VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 744.059534] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d5511a04-9d9e-4e16-a406-788e1e42bf8f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.082877] env[65726]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 744.082877] env[65726]: value = "task-5115753" [ 744.082877] env[65726]: _type = "Task" [ 744.082877] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 744.093506] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115753, 'name': CreateVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 744.455742] env[65726]: DEBUG oslo_concurrency.lockutils [None req-fd5c9ab4-2b38-4729-88f4-0f437c7dacb5 tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] Releasing lock "refresh_cache-83573c2b-9448-456f-8fd2-b19661dd6cc4" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 744.456428] env[65726]: DEBUG nova.compute.manager [None req-fd5c9ab4-2b38-4729-88f4-0f437c7dacb5 tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] [instance: 83573c2b-9448-456f-8fd2-b19661dd6cc4] Start destroying the instance on the hypervisor. {{(pid=65726) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 744.456800] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-fd5c9ab4-2b38-4729-88f4-0f437c7dacb5 tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] [instance: 83573c2b-9448-456f-8fd2-b19661dd6cc4] Destroying instance {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 744.457881] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb7e3116-5e4d-4458-981d-b2a790d057f7 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.466710] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd5c9ab4-2b38-4729-88f4-0f437c7dacb5 tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] [instance: 83573c2b-9448-456f-8fd2-b19661dd6cc4] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 744.466995] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6929b089-b376-4436-ac9e-905f01d3f195 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.474713] env[65726]: DEBUG oslo_vmware.api [None req-fd5c9ab4-2b38-4729-88f4-0f437c7dacb5 tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] Waiting for the task: (returnval){ [ 744.474713] env[65726]: value = "task-5115754" [ 744.474713] env[65726]: _type = "Task" [ 744.474713] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 744.485638] env[65726]: DEBUG oslo_vmware.api [None req-fd5c9ab4-2b38-4729-88f4-0f437c7dacb5 tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] Task: {'id': task-5115754, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 744.493339] env[65726]: DEBUG nova.compute.utils [None req-b0adde83-84f2-4fa3-84bf-8f387eee1293 tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] Using /dev/sd instead of None {{(pid=65726) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 744.498644] env[65726]: DEBUG nova.compute.manager [None req-b0adde83-84f2-4fa3-84bf-8f387eee1293 tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] [instance: 3827eed1-b7a1-4922-a4de-f0c038cb8566] Not allocating networking since 'none' was specified. {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2015}} [ 744.528965] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c19da911-fb8f-4bec-a53e-39a31fa2d36e tempest-ServersAdmin275Test-1865878431 tempest-ServersAdmin275Test-1865878431-project-member] Lock "3b933684-ac19-44b0-a49d-6af45501e166" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 46.606s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 744.537205] env[65726]: DEBUG oslo_concurrency.lockutils [None req-f61c0733-7339-4ad4-9f78-b4a5ed0d0ff1 tempest-AttachInterfacesV270Test-133387050 tempest-AttachInterfacesV270Test-133387050-project-member] Lock "c977257e-dce0-4ecc-b42b-0f6ebab38797" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 47.045s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 744.548221] env[65726]: WARNING openstack [req-610d135b-bb7d-4d70-8050-fa6d18dea786 req-54ca8ff5-4779-4684-9641-5a85dee41fe5 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 744.548221] env[65726]: WARNING openstack [req-610d135b-bb7d-4d70-8050-fa6d18dea786 req-54ca8ff5-4779-4684-9641-5a85dee41fe5 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 744.595175] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115753, 'name': CreateVM_Task, 'duration_secs': 0.473489} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 744.595259] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ca67d1a4-d304-4d06-a436-f257b60ca080] Created VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 744.596811] env[65726]: WARNING openstack [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 744.596811] env[65726]: WARNING openstack [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 744.602185] env[65726]: DEBUG oslo_concurrency.lockutils [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 744.602185] env[65726]: DEBUG oslo_concurrency.lockutils [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 744.602267] env[65726]: DEBUG oslo_concurrency.lockutils [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 744.602513] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-da7e300c-3876-4bee-8f20-ef1f4756588a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.608920] env[65726]: DEBUG oslo_vmware.api [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Waiting for the task: (returnval){ [ 744.608920] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52af0e4f-8f5e-f732-6767-48b28ad096f6" [ 744.608920] env[65726]: _type = "Task" [ 744.608920] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 744.619807] env[65726]: DEBUG oslo_vmware.api [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52af0e4f-8f5e-f732-6767-48b28ad096f6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 744.726379] env[65726]: WARNING openstack [req-610d135b-bb7d-4d70-8050-fa6d18dea786 req-54ca8ff5-4779-4684-9641-5a85dee41fe5 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 744.726809] env[65726]: WARNING openstack [req-610d135b-bb7d-4d70-8050-fa6d18dea786 req-54ca8ff5-4779-4684-9641-5a85dee41fe5 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 744.801732] env[65726]: WARNING openstack [req-610d135b-bb7d-4d70-8050-fa6d18dea786 req-54ca8ff5-4779-4684-9641-5a85dee41fe5 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 744.802181] env[65726]: WARNING openstack [req-610d135b-bb7d-4d70-8050-fa6d18dea786 req-54ca8ff5-4779-4684-9641-5a85dee41fe5 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 744.913659] env[65726]: DEBUG nova.network.neutron [req-610d135b-bb7d-4d70-8050-fa6d18dea786 req-54ca8ff5-4779-4684-9641-5a85dee41fe5 service nova] [instance: ca67d1a4-d304-4d06-a436-f257b60ca080] Updated VIF entry in instance network info cache for port 8789d2e9-5db3-4e5d-bdb2-14160bd43a1d. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 744.914201] env[65726]: DEBUG nova.network.neutron [req-610d135b-bb7d-4d70-8050-fa6d18dea786 req-54ca8ff5-4779-4684-9641-5a85dee41fe5 service nova] [instance: ca67d1a4-d304-4d06-a436-f257b60ca080] Updating instance_info_cache with network_info: [{"id": "8789d2e9-5db3-4e5d-bdb2-14160bd43a1d", "address": "fa:16:3e:cc:40:7f", "network": {"id": "c68b8bcc-840e-459b-bee0-76a9ffdc08cb", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-2061916615-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7c3641c4a34b4bf8b09403220e153e65", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "aa0d7fd6-4fd3-4451-9208-8a3cf25e30e4", "external-id": "nsx-vlan-transportzone-545", "segmentation_id": 545, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8789d2e9-5d", "ovs_interfaceid": "8789d2e9-5db3-4e5d-bdb2-14160bd43a1d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 744.988928] env[65726]: DEBUG oslo_vmware.api [None req-fd5c9ab4-2b38-4729-88f4-0f437c7dacb5 tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] Task: {'id': task-5115754, 'name': PowerOffVM_Task, 'duration_secs': 0.206578} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 744.989801] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd5c9ab4-2b38-4729-88f4-0f437c7dacb5 tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] [instance: 83573c2b-9448-456f-8fd2-b19661dd6cc4] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 744.992027] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-fd5c9ab4-2b38-4729-88f4-0f437c7dacb5 tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] [instance: 83573c2b-9448-456f-8fd2-b19661dd6cc4] Unregistering the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 744.992027] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-cf9b3fea-350e-419a-8b26-d994851307c7 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.002048] env[65726]: DEBUG nova.compute.manager [None req-b0adde83-84f2-4fa3-84bf-8f387eee1293 tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] [instance: 3827eed1-b7a1-4922-a4de-f0c038cb8566] Start building block device mappings for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 745.022247] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-fd5c9ab4-2b38-4729-88f4-0f437c7dacb5 tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] [instance: 83573c2b-9448-456f-8fd2-b19661dd6cc4] Unregistered the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 745.023644] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-fd5c9ab4-2b38-4729-88f4-0f437c7dacb5 tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] [instance: 83573c2b-9448-456f-8fd2-b19661dd6cc4] Deleting contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 745.023644] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-fd5c9ab4-2b38-4729-88f4-0f437c7dacb5 tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] Deleting the datastore file [datastore1] 83573c2b-9448-456f-8fd2-b19661dd6cc4 {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 745.023644] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f6504ade-932f-4dfe-a534-d1db5efbf2a8 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.032342] env[65726]: DEBUG oslo_vmware.api [None req-fd5c9ab4-2b38-4729-88f4-0f437c7dacb5 tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] Waiting for the task: (returnval){ [ 745.032342] env[65726]: value = "task-5115756" [ 745.032342] env[65726]: _type = "Task" [ 745.032342] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 745.052231] env[65726]: DEBUG oslo_vmware.api [None req-fd5c9ab4-2b38-4729-88f4-0f437c7dacb5 tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] Task: {'id': task-5115756, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 745.121728] env[65726]: DEBUG oslo_vmware.api [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52af0e4f-8f5e-f732-6767-48b28ad096f6, 'name': SearchDatastore_Task, 'duration_secs': 0.015529} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 745.128804] env[65726]: DEBUG oslo_concurrency.lockutils [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 745.129245] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] [instance: ca67d1a4-d304-4d06-a436-f257b60ca080] Processing image b52362a3-ee8a-4cbf-b06f-513b0cc8f95c {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 745.129658] env[65726]: DEBUG oslo_concurrency.lockutils [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 745.129843] env[65726]: DEBUG oslo_concurrency.lockutils [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 745.130102] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 745.131736] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b740d44b-9537-46aa-b7e5-d2ba01497968 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.146283] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 745.146534] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65726) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 745.148192] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-20eb9d23-c038-4cee-bfa1-a4d7a0cb8041 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.160293] env[65726]: DEBUG oslo_vmware.api [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Waiting for the task: (returnval){ [ 745.160293] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]524652d9-4be5-f820-6642-2995ec3d2e7a" [ 745.160293] env[65726]: _type = "Task" [ 745.160293] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 745.168850] env[65726]: DEBUG oslo_vmware.api [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]524652d9-4be5-f820-6642-2995ec3d2e7a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 745.419160] env[65726]: DEBUG oslo_concurrency.lockutils [req-610d135b-bb7d-4d70-8050-fa6d18dea786 req-54ca8ff5-4779-4684-9641-5a85dee41fe5 service nova] Releasing lock "refresh_cache-ca67d1a4-d304-4d06-a436-f257b60ca080" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 745.556017] env[65726]: DEBUG oslo_vmware.api [None req-fd5c9ab4-2b38-4729-88f4-0f437c7dacb5 tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] Task: {'id': task-5115756, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.111685} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 745.556563] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-fd5c9ab4-2b38-4729-88f4-0f437c7dacb5 tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] Deleted the datastore file {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 745.556609] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-fd5c9ab4-2b38-4729-88f4-0f437c7dacb5 tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] [instance: 83573c2b-9448-456f-8fd2-b19661dd6cc4] Deleted contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 745.556848] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-fd5c9ab4-2b38-4729-88f4-0f437c7dacb5 tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] [instance: 83573c2b-9448-456f-8fd2-b19661dd6cc4] Instance destroyed {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 745.559915] env[65726]: INFO nova.compute.manager [None req-fd5c9ab4-2b38-4729-88f4-0f437c7dacb5 tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] [instance: 83573c2b-9448-456f-8fd2-b19661dd6cc4] Took 1.10 seconds to destroy the instance on the hypervisor. [ 745.559915] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-fd5c9ab4-2b38-4729-88f4-0f437c7dacb5 tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 745.559915] env[65726]: DEBUG nova.compute.manager [-] [instance: 83573c2b-9448-456f-8fd2-b19661dd6cc4] Deallocating network for instance {{(pid=65726) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 745.559915] env[65726]: DEBUG nova.network.neutron [-] [instance: 83573c2b-9448-456f-8fd2-b19661dd6cc4] deallocate_for_instance() {{(pid=65726) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 745.559915] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 745.559915] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 745.602012] env[65726]: DEBUG nova.network.neutron [-] [instance: 83573c2b-9448-456f-8fd2-b19661dd6cc4] Instance cache missing network info. {{(pid=65726) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 745.602799] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 745.603210] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 745.681607] env[65726]: DEBUG oslo_vmware.api [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]524652d9-4be5-f820-6642-2995ec3d2e7a, 'name': SearchDatastore_Task, 'duration_secs': 0.010478} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 745.682941] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-651de3b2-f349-47c2-b44f-bc8ce052055b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.695462] env[65726]: DEBUG oslo_vmware.api [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Waiting for the task: (returnval){ [ 745.695462] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52a43449-1f35-6834-0904-d25b5fa5aa22" [ 745.695462] env[65726]: _type = "Task" [ 745.695462] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 745.713754] env[65726]: DEBUG oslo_vmware.api [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52a43449-1f35-6834-0904-d25b5fa5aa22, 'name': SearchDatastore_Task, 'duration_secs': 0.012112} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 745.714164] env[65726]: DEBUG oslo_concurrency.lockutils [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 745.714559] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore2] ca67d1a4-d304-4d06-a436-f257b60ca080/ca67d1a4-d304-4d06-a436-f257b60ca080.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 745.718019] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a23013d7-4f9c-45d2-8de6-71bb26eac0f4 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.725759] env[65726]: DEBUG oslo_vmware.api [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Waiting for the task: (returnval){ [ 745.725759] env[65726]: value = "task-5115757" [ 745.725759] env[65726]: _type = "Task" [ 745.725759] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 745.744946] env[65726]: DEBUG oslo_vmware.api [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Task: {'id': task-5115757, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 745.750302] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6ace785-7465-45c3-873b-cae0056a1e53 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.760020] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d439b3c4-5989-43cc-bbc8-ca6924a94505 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.795054] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4ad7e06-fb65-47c9-9ef1-c2f0079352e9 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.804886] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11e588a4-59f2-4d1c-8429-9b4bc1077285 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.824433] env[65726]: DEBUG nova.compute.provider_tree [None req-b479f246-d94b-4634-888f-dee4586c3272 tempest-ServerAddressesNegativeTestJSON-963263679 tempest-ServerAddressesNegativeTestJSON-963263679-project-member] Updating inventory in ProviderTree for provider 07c4692f-bdb4-4058-9173-ff9664830295 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 746.022670] env[65726]: DEBUG nova.compute.manager [None req-b0adde83-84f2-4fa3-84bf-8f387eee1293 tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] [instance: 3827eed1-b7a1-4922-a4de-f0c038cb8566] Start spawning the instance on the hypervisor. {{(pid=65726) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 746.055966] env[65726]: DEBUG nova.virt.hardware [None req-b0adde83-84f2-4fa3-84bf-8f387eee1293 tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T19:28:34Z,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1539f06fa3534e90acbc3a60c4b8bd3f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T19:28:35Z,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 746.056317] env[65726]: DEBUG nova.virt.hardware [None req-b0adde83-84f2-4fa3-84bf-8f387eee1293 tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 746.056439] env[65726]: DEBUG nova.virt.hardware [None req-b0adde83-84f2-4fa3-84bf-8f387eee1293 tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 746.056723] env[65726]: DEBUG nova.virt.hardware [None req-b0adde83-84f2-4fa3-84bf-8f387eee1293 tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 746.056895] env[65726]: DEBUG nova.virt.hardware [None req-b0adde83-84f2-4fa3-84bf-8f387eee1293 tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 746.057049] env[65726]: DEBUG nova.virt.hardware [None req-b0adde83-84f2-4fa3-84bf-8f387eee1293 tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 746.057267] env[65726]: DEBUG nova.virt.hardware [None req-b0adde83-84f2-4fa3-84bf-8f387eee1293 tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 746.057440] env[65726]: DEBUG nova.virt.hardware [None req-b0adde83-84f2-4fa3-84bf-8f387eee1293 tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 746.057693] env[65726]: DEBUG nova.virt.hardware [None req-b0adde83-84f2-4fa3-84bf-8f387eee1293 tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 746.057875] env[65726]: DEBUG nova.virt.hardware [None req-b0adde83-84f2-4fa3-84bf-8f387eee1293 tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 746.058059] env[65726]: DEBUG nova.virt.hardware [None req-b0adde83-84f2-4fa3-84bf-8f387eee1293 tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 746.059013] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c13abf7-88b2-4735-b1dc-5fb3817b0127 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.069899] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eeb89fe2-f214-4d3d-8995-192d42b06705 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.087344] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-b0adde83-84f2-4fa3-84bf-8f387eee1293 tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] [instance: 3827eed1-b7a1-4922-a4de-f0c038cb8566] Instance VIF info [] {{(pid=65726) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 746.094024] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-b0adde83-84f2-4fa3-84bf-8f387eee1293 tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] Creating folder: Project (a26c136f028d44409abc385cba383c31). Parent ref: group-v995008. {{(pid=65726) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 746.094504] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9acd48db-069b-4818-b0b0-4df67d0fba6b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.109742] env[65726]: INFO nova.virt.vmwareapi.vm_util [None req-b0adde83-84f2-4fa3-84bf-8f387eee1293 tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] Created folder: Project (a26c136f028d44409abc385cba383c31) in parent group-v995008. [ 746.110060] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-b0adde83-84f2-4fa3-84bf-8f387eee1293 tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] Creating folder: Instances. Parent ref: group-v995106. {{(pid=65726) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 746.110350] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-14477e80-4df9-4af5-a0e3-7a7a78f60128 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.123472] env[65726]: INFO nova.virt.vmwareapi.vm_util [None req-b0adde83-84f2-4fa3-84bf-8f387eee1293 tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] Created folder: Instances in parent group-v995106. [ 746.127029] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-b0adde83-84f2-4fa3-84bf-8f387eee1293 tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 746.127029] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3827eed1-b7a1-4922-a4de-f0c038cb8566] Creating VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 746.127029] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9e690963-7bca-4063-abb5-1ea1a9394798 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.142182] env[65726]: DEBUG nova.network.neutron [-] [instance: 83573c2b-9448-456f-8fd2-b19661dd6cc4] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 746.150737] env[65726]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 746.150737] env[65726]: value = "task-5115760" [ 746.150737] env[65726]: _type = "Task" [ 746.150737] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 746.163465] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115760, 'name': CreateVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 746.243103] env[65726]: DEBUG oslo_vmware.api [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Task: {'id': task-5115757, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 746.353119] env[65726]: ERROR nova.scheduler.client.report [None req-b479f246-d94b-4634-888f-dee4586c3272 tempest-ServerAddressesNegativeTestJSON-963263679 tempest-ServerAddressesNegativeTestJSON-963263679-project-member] [req-e6adbfbb-0e4d-43c2-aae7-f6162705a826] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 07c4692f-bdb4-4058-9173-ff9664830295. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-e6adbfbb-0e4d-43c2-aae7-f6162705a826"}]} [ 746.373033] env[65726]: DEBUG nova.scheduler.client.report [None req-b479f246-d94b-4634-888f-dee4586c3272 tempest-ServerAddressesNegativeTestJSON-963263679 tempest-ServerAddressesNegativeTestJSON-963263679-project-member] Refreshing inventories for resource provider 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 746.395652] env[65726]: DEBUG nova.scheduler.client.report [None req-b479f246-d94b-4634-888f-dee4586c3272 tempest-ServerAddressesNegativeTestJSON-963263679 tempest-ServerAddressesNegativeTestJSON-963263679-project-member] Updating ProviderTree inventory for provider 07c4692f-bdb4-4058-9173-ff9664830295 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 746.395652] env[65726]: DEBUG nova.compute.provider_tree [None req-b479f246-d94b-4634-888f-dee4586c3272 tempest-ServerAddressesNegativeTestJSON-963263679 tempest-ServerAddressesNegativeTestJSON-963263679-project-member] Updating inventory in ProviderTree for provider 07c4692f-bdb4-4058-9173-ff9664830295 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 746.411515] env[65726]: DEBUG nova.scheduler.client.report [None req-b479f246-d94b-4634-888f-dee4586c3272 tempest-ServerAddressesNegativeTestJSON-963263679 tempest-ServerAddressesNegativeTestJSON-963263679-project-member] Refreshing aggregate associations for resource provider 07c4692f-bdb4-4058-9173-ff9664830295, aggregates: None {{(pid=65726) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 746.437480] env[65726]: DEBUG nova.scheduler.client.report [None req-b479f246-d94b-4634-888f-dee4586c3272 tempest-ServerAddressesNegativeTestJSON-963263679 tempest-ServerAddressesNegativeTestJSON-963263679-project-member] Refreshing trait associations for resource provider 07c4692f-bdb4-4058-9173-ff9664830295, traits: COMPUTE_IMAGE_TYPE_ISO,HW_ARCH_X86_64,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=65726) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 746.562144] env[65726]: DEBUG oslo_vmware.rw_handles [None req-61a6387d-c5ed-4ab7-853e-9ba63af33205 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/526f27e0-8c7b-4c32-362a-63165cd7aeea/disk-0.vmdk. {{(pid=65726) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 746.563353] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb1f48c8-3035-4fd0-a22c-216b9124637f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.574750] env[65726]: DEBUG oslo_vmware.rw_handles [None req-61a6387d-c5ed-4ab7-853e-9ba63af33205 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/526f27e0-8c7b-4c32-362a-63165cd7aeea/disk-0.vmdk is in state: ready. {{(pid=65726) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 746.575698] env[65726]: ERROR oslo_vmware.rw_handles [None req-61a6387d-c5ed-4ab7-853e-9ba63af33205 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/526f27e0-8c7b-4c32-362a-63165cd7aeea/disk-0.vmdk due to incomplete transfer. [ 746.575698] env[65726]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-7a9e5c66-f3d0-4b28-bff9-69d6e8e2b94d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.585538] env[65726]: DEBUG oslo_vmware.rw_handles [None req-61a6387d-c5ed-4ab7-853e-9ba63af33205 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/526f27e0-8c7b-4c32-362a-63165cd7aeea/disk-0.vmdk. {{(pid=65726) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 746.585734] env[65726]: DEBUG nova.virt.vmwareapi.images [None req-61a6387d-c5ed-4ab7-853e-9ba63af33205 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: bd839fe1-8801-4ba1-9c23-288453258b66] Uploaded image 249aae9c-fa0a-456d-bd7e-ad65ba07d625 to the Glance image server {{(pid=65726) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 746.588482] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-61a6387d-c5ed-4ab7-853e-9ba63af33205 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: bd839fe1-8801-4ba1-9c23-288453258b66] Destroying the VM {{(pid=65726) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 746.588482] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-418d22c0-b818-4db8-8484-271e79902372 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.596264] env[65726]: DEBUG oslo_vmware.api [None req-61a6387d-c5ed-4ab7-853e-9ba63af33205 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Waiting for the task: (returnval){ [ 746.596264] env[65726]: value = "task-5115761" [ 746.596264] env[65726]: _type = "Task" [ 746.596264] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 746.609069] env[65726]: DEBUG oslo_vmware.api [None req-61a6387d-c5ed-4ab7-853e-9ba63af33205 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': task-5115761, 'name': Destroy_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 746.645542] env[65726]: INFO nova.compute.manager [-] [instance: 83573c2b-9448-456f-8fd2-b19661dd6cc4] Took 1.09 seconds to deallocate network for instance. [ 746.664473] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115760, 'name': CreateVM_Task, 'duration_secs': 0.348493} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 746.664924] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3827eed1-b7a1-4922-a4de-f0c038cb8566] Created VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 746.668285] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b0adde83-84f2-4fa3-84bf-8f387eee1293 tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 746.668358] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b0adde83-84f2-4fa3-84bf-8f387eee1293 tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 746.668646] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b0adde83-84f2-4fa3-84bf-8f387eee1293 tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 746.669248] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d54ecdc6-00d4-4255-a350-5a89f986bfba {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.674735] env[65726]: DEBUG oslo_vmware.api [None req-b0adde83-84f2-4fa3-84bf-8f387eee1293 tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] Waiting for the task: (returnval){ [ 746.674735] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52087388-8ab3-9187-a32d-224f20cab4f5" [ 746.674735] env[65726]: _type = "Task" [ 746.674735] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 746.689131] env[65726]: DEBUG oslo_vmware.api [None req-b0adde83-84f2-4fa3-84bf-8f387eee1293 tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52087388-8ab3-9187-a32d-224f20cab4f5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 746.747788] env[65726]: DEBUG oslo_vmware.api [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Task: {'id': task-5115757, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.536278} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 746.748116] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore2] ca67d1a4-d304-4d06-a436-f257b60ca080/ca67d1a4-d304-4d06-a436-f257b60ca080.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 746.748333] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] [instance: ca67d1a4-d304-4d06-a436-f257b60ca080] Extending root virtual disk to 1048576 {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 746.748609] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d752aed4-0cdb-449f-b05b-85d8db0d9499 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.758265] env[65726]: DEBUG oslo_vmware.api [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Waiting for the task: (returnval){ [ 746.758265] env[65726]: value = "task-5115762" [ 746.758265] env[65726]: _type = "Task" [ 746.758265] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 746.769638] env[65726]: DEBUG oslo_vmware.api [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Task: {'id': task-5115762, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 747.088292] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebbde56b-9b6e-4e2a-8530-6dd98ebe774a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.103083] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b07835c-27e0-4cc3-a91e-be7651ad308d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.115569] env[65726]: DEBUG oslo_vmware.api [None req-61a6387d-c5ed-4ab7-853e-9ba63af33205 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': task-5115761, 'name': Destroy_Task, 'duration_secs': 0.355317} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 747.116469] env[65726]: INFO nova.virt.vmwareapi.vm_util [None req-61a6387d-c5ed-4ab7-853e-9ba63af33205 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: bd839fe1-8801-4ba1-9c23-288453258b66] Destroyed the VM [ 747.117366] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-61a6387d-c5ed-4ab7-853e-9ba63af33205 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: bd839fe1-8801-4ba1-9c23-288453258b66] Deleting Snapshot of the VM instance {{(pid=65726) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 747.143184] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-db9fb994-0898-4cc9-9846-f58be2c5511a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.146235] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ac8c70a-f550-45c6-87e4-90b4b87348b9 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.154982] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0e7c53e-458a-412b-b30c-8690f0471f7f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.161905] env[65726]: DEBUG oslo_concurrency.lockutils [None req-fd5c9ab4-2b38-4729-88f4-0f437c7dacb5 tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 747.162273] env[65726]: DEBUG oslo_vmware.api [None req-61a6387d-c5ed-4ab7-853e-9ba63af33205 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Waiting for the task: (returnval){ [ 747.162273] env[65726]: value = "task-5115763" [ 747.162273] env[65726]: _type = "Task" [ 747.162273] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 747.175075] env[65726]: DEBUG nova.compute.provider_tree [None req-b479f246-d94b-4634-888f-dee4586c3272 tempest-ServerAddressesNegativeTestJSON-963263679 tempest-ServerAddressesNegativeTestJSON-963263679-project-member] Updating inventory in ProviderTree for provider 07c4692f-bdb4-4058-9173-ff9664830295 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 747.184136] env[65726]: DEBUG oslo_vmware.api [None req-61a6387d-c5ed-4ab7-853e-9ba63af33205 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': task-5115763, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 747.190861] env[65726]: DEBUG oslo_vmware.api [None req-b0adde83-84f2-4fa3-84bf-8f387eee1293 tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52087388-8ab3-9187-a32d-224f20cab4f5, 'name': SearchDatastore_Task, 'duration_secs': 0.011331} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 747.191080] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b0adde83-84f2-4fa3-84bf-8f387eee1293 tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 747.191312] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-b0adde83-84f2-4fa3-84bf-8f387eee1293 tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] [instance: 3827eed1-b7a1-4922-a4de-f0c038cb8566] Processing image b52362a3-ee8a-4cbf-b06f-513b0cc8f95c {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 747.191542] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b0adde83-84f2-4fa3-84bf-8f387eee1293 tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 747.191693] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b0adde83-84f2-4fa3-84bf-8f387eee1293 tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 747.191955] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-b0adde83-84f2-4fa3-84bf-8f387eee1293 tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 747.192256] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-925353da-8546-4d3d-ad32-1ea7ee7a1aae {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.202334] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-b0adde83-84f2-4fa3-84bf-8f387eee1293 tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 747.202519] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-b0adde83-84f2-4fa3-84bf-8f387eee1293 tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65726) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 747.203313] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-00332637-e3c7-4831-85b0-31342e950d33 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.210241] env[65726]: DEBUG oslo_vmware.api [None req-b0adde83-84f2-4fa3-84bf-8f387eee1293 tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] Waiting for the task: (returnval){ [ 747.210241] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52bd7a6f-dd60-d555-1bb8-e00100d7c046" [ 747.210241] env[65726]: _type = "Task" [ 747.210241] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 747.219477] env[65726]: DEBUG oslo_vmware.api [None req-b0adde83-84f2-4fa3-84bf-8f387eee1293 tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52bd7a6f-dd60-d555-1bb8-e00100d7c046, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 747.269812] env[65726]: DEBUG oslo_vmware.api [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Task: {'id': task-5115762, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.072954} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 747.270287] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] [instance: ca67d1a4-d304-4d06-a436-f257b60ca080] Extended root virtual disk {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 747.271299] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01deedd8-3906-430c-bcf1-a947c7be5743 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.296267] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] [instance: ca67d1a4-d304-4d06-a436-f257b60ca080] Reconfiguring VM instance instance-00000024 to attach disk [datastore2] ca67d1a4-d304-4d06-a436-f257b60ca080/ca67d1a4-d304-4d06-a436-f257b60ca080.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 747.296539] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7fea5a9b-fea8-4e04-a3f9-ed83ad1f1ff8 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.319182] env[65726]: DEBUG oslo_vmware.api [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Waiting for the task: (returnval){ [ 747.319182] env[65726]: value = "task-5115764" [ 747.319182] env[65726]: _type = "Task" [ 747.319182] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 747.330621] env[65726]: DEBUG oslo_vmware.api [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Task: {'id': task-5115764, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 747.675258] env[65726]: DEBUG oslo_vmware.api [None req-61a6387d-c5ed-4ab7-853e-9ba63af33205 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': task-5115763, 'name': RemoveSnapshot_Task} progress is 100%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 747.711899] env[65726]: DEBUG nova.scheduler.client.report [None req-b479f246-d94b-4634-888f-dee4586c3272 tempest-ServerAddressesNegativeTestJSON-963263679 tempest-ServerAddressesNegativeTestJSON-963263679-project-member] Updated inventory for provider 07c4692f-bdb4-4058-9173-ff9664830295 with generation 66 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 747.712321] env[65726]: DEBUG nova.compute.provider_tree [None req-b479f246-d94b-4634-888f-dee4586c3272 tempest-ServerAddressesNegativeTestJSON-963263679 tempest-ServerAddressesNegativeTestJSON-963263679-project-member] Updating resource provider 07c4692f-bdb4-4058-9173-ff9664830295 generation from 66 to 67 during operation: update_inventory {{(pid=65726) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 747.712431] env[65726]: DEBUG nova.compute.provider_tree [None req-b479f246-d94b-4634-888f-dee4586c3272 tempest-ServerAddressesNegativeTestJSON-963263679 tempest-ServerAddressesNegativeTestJSON-963263679-project-member] Updating inventory in ProviderTree for provider 07c4692f-bdb4-4058-9173-ff9664830295 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 747.727062] env[65726]: DEBUG oslo_vmware.api [None req-b0adde83-84f2-4fa3-84bf-8f387eee1293 tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52bd7a6f-dd60-d555-1bb8-e00100d7c046, 'name': SearchDatastore_Task, 'duration_secs': 0.009945} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 747.728137] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-48f4d34e-bd94-483f-9c5b-112dc6c1610d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.735294] env[65726]: DEBUG oslo_vmware.api [None req-b0adde83-84f2-4fa3-84bf-8f387eee1293 tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] Waiting for the task: (returnval){ [ 747.735294] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52edd842-31ad-c9fd-9198-564bc589d73a" [ 747.735294] env[65726]: _type = "Task" [ 747.735294] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 747.744443] env[65726]: DEBUG oslo_vmware.api [None req-b0adde83-84f2-4fa3-84bf-8f387eee1293 tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52edd842-31ad-c9fd-9198-564bc589d73a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 747.834798] env[65726]: DEBUG oslo_vmware.api [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Task: {'id': task-5115764, 'name': ReconfigVM_Task, 'duration_secs': 0.308545} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 747.835135] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] [instance: ca67d1a4-d304-4d06-a436-f257b60ca080] Reconfigured VM instance instance-00000024 to attach disk [datastore2] ca67d1a4-d304-4d06-a436-f257b60ca080/ca67d1a4-d304-4d06-a436-f257b60ca080.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 747.836164] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4dc5be4d-96d0-4276-932a-1ae27951a0a9 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.845516] env[65726]: DEBUG oslo_vmware.api [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Waiting for the task: (returnval){ [ 747.845516] env[65726]: value = "task-5115765" [ 747.845516] env[65726]: _type = "Task" [ 747.845516] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 747.856543] env[65726]: DEBUG oslo_vmware.api [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Task: {'id': task-5115765, 'name': Rename_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 748.179656] env[65726]: DEBUG oslo_vmware.api [None req-61a6387d-c5ed-4ab7-853e-9ba63af33205 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': task-5115763, 'name': RemoveSnapshot_Task, 'duration_secs': 0.544397} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 748.180227] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-61a6387d-c5ed-4ab7-853e-9ba63af33205 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: bd839fe1-8801-4ba1-9c23-288453258b66] Deleted Snapshot of the VM instance {{(pid=65726) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 748.180637] env[65726]: INFO nova.compute.manager [None req-61a6387d-c5ed-4ab7-853e-9ba63af33205 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: bd839fe1-8801-4ba1-9c23-288453258b66] Took 15.68 seconds to snapshot the instance on the hypervisor. [ 748.221044] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b479f246-d94b-4634-888f-dee4586c3272 tempest-ServerAddressesNegativeTestJSON-963263679 tempest-ServerAddressesNegativeTestJSON-963263679-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 4.237s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 748.221653] env[65726]: DEBUG nova.compute.manager [None req-b479f246-d94b-4634-888f-dee4586c3272 tempest-ServerAddressesNegativeTestJSON-963263679 tempest-ServerAddressesNegativeTestJSON-963263679-project-member] [instance: 00746508-d0d6-4dfa-9026-772398d004ee] Start building networks asynchronously for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 748.224421] env[65726]: DEBUG oslo_concurrency.lockutils [None req-04afb47c-692a-46d9-a640-68cc2c90206f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 42.559s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 748.226029] env[65726]: INFO nova.compute.claims [None req-04afb47c-692a-46d9-a640-68cc2c90206f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: c108f2a5-031b-47ee-9a5b-d62c2f42c26b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 748.253627] env[65726]: DEBUG oslo_vmware.api [None req-b0adde83-84f2-4fa3-84bf-8f387eee1293 tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52edd842-31ad-c9fd-9198-564bc589d73a, 'name': SearchDatastore_Task, 'duration_secs': 0.010264} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 748.253840] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b0adde83-84f2-4fa3-84bf-8f387eee1293 tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 748.254554] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-b0adde83-84f2-4fa3-84bf-8f387eee1293 tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] 3827eed1-b7a1-4922-a4de-f0c038cb8566/3827eed1-b7a1-4922-a4de-f0c038cb8566.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 748.255045] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-002b9698-27c4-417a-ba59-5bd52f5641fb {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.265025] env[65726]: DEBUG oslo_vmware.api [None req-b0adde83-84f2-4fa3-84bf-8f387eee1293 tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] Waiting for the task: (returnval){ [ 748.265025] env[65726]: value = "task-5115766" [ 748.265025] env[65726]: _type = "Task" [ 748.265025] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 748.276015] env[65726]: DEBUG oslo_vmware.api [None req-b0adde83-84f2-4fa3-84bf-8f387eee1293 tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] Task: {'id': task-5115766, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 748.361041] env[65726]: DEBUG oslo_vmware.api [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Task: {'id': task-5115765, 'name': Rename_Task, 'duration_secs': 0.157468} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 748.362120] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] [instance: ca67d1a4-d304-4d06-a436-f257b60ca080] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 748.362120] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1a463794-0323-4ce3-8efc-31e56609bed1 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.370989] env[65726]: DEBUG oslo_vmware.api [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Waiting for the task: (returnval){ [ 748.370989] env[65726]: value = "task-5115767" [ 748.370989] env[65726]: _type = "Task" [ 748.370989] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 748.379764] env[65726]: DEBUG oslo_vmware.api [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Task: {'id': task-5115767, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 748.733317] env[65726]: DEBUG nova.compute.utils [None req-b479f246-d94b-4634-888f-dee4586c3272 tempest-ServerAddressesNegativeTestJSON-963263679 tempest-ServerAddressesNegativeTestJSON-963263679-project-member] Using /dev/sd instead of None {{(pid=65726) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 748.745565] env[65726]: DEBUG nova.compute.manager [None req-b479f246-d94b-4634-888f-dee4586c3272 tempest-ServerAddressesNegativeTestJSON-963263679 tempest-ServerAddressesNegativeTestJSON-963263679-project-member] [instance: 00746508-d0d6-4dfa-9026-772398d004ee] Allocating IP information in the background. {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 748.745937] env[65726]: DEBUG nova.network.neutron [None req-b479f246-d94b-4634-888f-dee4586c3272 tempest-ServerAddressesNegativeTestJSON-963263679 tempest-ServerAddressesNegativeTestJSON-963263679-project-member] [instance: 00746508-d0d6-4dfa-9026-772398d004ee] allocate_for_instance() {{(pid=65726) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 748.747092] env[65726]: WARNING neutronclient.v2_0.client [None req-b479f246-d94b-4634-888f-dee4586c3272 tempest-ServerAddressesNegativeTestJSON-963263679 tempest-ServerAddressesNegativeTestJSON-963263679-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 748.747599] env[65726]: WARNING neutronclient.v2_0.client [None req-b479f246-d94b-4634-888f-dee4586c3272 tempest-ServerAddressesNegativeTestJSON-963263679 tempest-ServerAddressesNegativeTestJSON-963263679-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 748.748263] env[65726]: WARNING openstack [None req-b479f246-d94b-4634-888f-dee4586c3272 tempest-ServerAddressesNegativeTestJSON-963263679 tempest-ServerAddressesNegativeTestJSON-963263679-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 748.748752] env[65726]: WARNING openstack [None req-b479f246-d94b-4634-888f-dee4586c3272 tempest-ServerAddressesNegativeTestJSON-963263679 tempest-ServerAddressesNegativeTestJSON-963263679-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 748.783567] env[65726]: DEBUG oslo_vmware.api [None req-b0adde83-84f2-4fa3-84bf-8f387eee1293 tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] Task: {'id': task-5115766, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 748.838802] env[65726]: DEBUG nova.policy [None req-b479f246-d94b-4634-888f-dee4586c3272 tempest-ServerAddressesNegativeTestJSON-963263679 tempest-ServerAddressesNegativeTestJSON-963263679-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a92866f437ba473c8822c381913be5ef', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0d2e9e9fe170484896239f5f9cc9dc09', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65726) authorize /opt/stack/nova/nova/policy.py:192}} [ 748.884481] env[65726]: DEBUG oslo_vmware.api [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Task: {'id': task-5115767, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 749.145307] env[65726]: DEBUG nova.network.neutron [None req-b479f246-d94b-4634-888f-dee4586c3272 tempest-ServerAddressesNegativeTestJSON-963263679 tempest-ServerAddressesNegativeTestJSON-963263679-project-member] [instance: 00746508-d0d6-4dfa-9026-772398d004ee] Successfully created port: 1151551e-12f0-4e85-a385-f1ebf4e9bc72 {{(pid=65726) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 749.236334] env[65726]: DEBUG nova.compute.manager [None req-b479f246-d94b-4634-888f-dee4586c3272 tempest-ServerAddressesNegativeTestJSON-963263679 tempest-ServerAddressesNegativeTestJSON-963263679-project-member] [instance: 00746508-d0d6-4dfa-9026-772398d004ee] Start building block device mappings for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 749.277200] env[65726]: DEBUG oslo_vmware.api [None req-b0adde83-84f2-4fa3-84bf-8f387eee1293 tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] Task: {'id': task-5115766, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.535979} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 749.277592] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-b0adde83-84f2-4fa3-84bf-8f387eee1293 tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] 3827eed1-b7a1-4922-a4de-f0c038cb8566/3827eed1-b7a1-4922-a4de-f0c038cb8566.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 749.277727] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-b0adde83-84f2-4fa3-84bf-8f387eee1293 tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] [instance: 3827eed1-b7a1-4922-a4de-f0c038cb8566] Extending root virtual disk to 1048576 {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 749.278023] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-de7b0b9b-7d8d-45e5-8609-e6a9d2221984 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.286678] env[65726]: DEBUG oslo_vmware.api [None req-b0adde83-84f2-4fa3-84bf-8f387eee1293 tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] Waiting for the task: (returnval){ [ 749.286678] env[65726]: value = "task-5115768" [ 749.286678] env[65726]: _type = "Task" [ 749.286678] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 749.296872] env[65726]: DEBUG oslo_vmware.api [None req-b0adde83-84f2-4fa3-84bf-8f387eee1293 tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] Task: {'id': task-5115768, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 749.385558] env[65726]: DEBUG oslo_vmware.api [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Task: {'id': task-5115767, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 749.800930] env[65726]: DEBUG oslo_vmware.api [None req-b0adde83-84f2-4fa3-84bf-8f387eee1293 tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] Task: {'id': task-5115768, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.371927} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 749.804438] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-b0adde83-84f2-4fa3-84bf-8f387eee1293 tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] [instance: 3827eed1-b7a1-4922-a4de-f0c038cb8566] Extended root virtual disk {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 749.806029] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e22dea51-4b17-43a1-ae55-6eb0a533838c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.833476] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-b0adde83-84f2-4fa3-84bf-8f387eee1293 tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] [instance: 3827eed1-b7a1-4922-a4de-f0c038cb8566] Reconfiguring VM instance instance-00000025 to attach disk [datastore1] 3827eed1-b7a1-4922-a4de-f0c038cb8566/3827eed1-b7a1-4922-a4de-f0c038cb8566.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 749.836791] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4fed81b8-ee4a-46c7-bd79-75b790042b38 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.859421] env[65726]: DEBUG oslo_vmware.api [None req-b0adde83-84f2-4fa3-84bf-8f387eee1293 tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] Waiting for the task: (returnval){ [ 749.859421] env[65726]: value = "task-5115769" [ 749.859421] env[65726]: _type = "Task" [ 749.859421] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 749.871275] env[65726]: DEBUG oslo_vmware.api [None req-b0adde83-84f2-4fa3-84bf-8f387eee1293 tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] Task: {'id': task-5115769, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 749.884491] env[65726]: DEBUG oslo_vmware.api [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Task: {'id': task-5115767, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 749.989785] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da0ed7f7-2905-4376-a147-ba5b5e2af41c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.999241] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ccb19517-af02-4461-83ca-3734f3a26c8f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.033350] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ad0b6c6-1387-4e3b-9d30-3ed491f9eb0d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.043203] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92faec5b-eb6d-4cf2-8844-ccdc0ad3d60c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.060895] env[65726]: DEBUG nova.compute.provider_tree [None req-04afb47c-692a-46d9-a640-68cc2c90206f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Updating inventory in ProviderTree for provider 07c4692f-bdb4-4058-9173-ff9664830295 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 750.246021] env[65726]: DEBUG nova.compute.manager [None req-b479f246-d94b-4634-888f-dee4586c3272 tempest-ServerAddressesNegativeTestJSON-963263679 tempest-ServerAddressesNegativeTestJSON-963263679-project-member] [instance: 00746508-d0d6-4dfa-9026-772398d004ee] Start spawning the instance on the hypervisor. {{(pid=65726) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 750.277162] env[65726]: DEBUG nova.virt.hardware [None req-b479f246-d94b-4634-888f-dee4586c3272 tempest-ServerAddressesNegativeTestJSON-963263679 tempest-ServerAddressesNegativeTestJSON-963263679-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T19:28:34Z,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1539f06fa3534e90acbc3a60c4b8bd3f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T19:28:35Z,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 750.277420] env[65726]: DEBUG nova.virt.hardware [None req-b479f246-d94b-4634-888f-dee4586c3272 tempest-ServerAddressesNegativeTestJSON-963263679 tempest-ServerAddressesNegativeTestJSON-963263679-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 750.277574] env[65726]: DEBUG nova.virt.hardware [None req-b479f246-d94b-4634-888f-dee4586c3272 tempest-ServerAddressesNegativeTestJSON-963263679 tempest-ServerAddressesNegativeTestJSON-963263679-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 750.277747] env[65726]: DEBUG nova.virt.hardware [None req-b479f246-d94b-4634-888f-dee4586c3272 tempest-ServerAddressesNegativeTestJSON-963263679 tempest-ServerAddressesNegativeTestJSON-963263679-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 750.277878] env[65726]: DEBUG nova.virt.hardware [None req-b479f246-d94b-4634-888f-dee4586c3272 tempest-ServerAddressesNegativeTestJSON-963263679 tempest-ServerAddressesNegativeTestJSON-963263679-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 750.278088] env[65726]: DEBUG nova.virt.hardware [None req-b479f246-d94b-4634-888f-dee4586c3272 tempest-ServerAddressesNegativeTestJSON-963263679 tempest-ServerAddressesNegativeTestJSON-963263679-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 750.278354] env[65726]: DEBUG nova.virt.hardware [None req-b479f246-d94b-4634-888f-dee4586c3272 tempest-ServerAddressesNegativeTestJSON-963263679 tempest-ServerAddressesNegativeTestJSON-963263679-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 750.278548] env[65726]: DEBUG nova.virt.hardware [None req-b479f246-d94b-4634-888f-dee4586c3272 tempest-ServerAddressesNegativeTestJSON-963263679 tempest-ServerAddressesNegativeTestJSON-963263679-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 750.278764] env[65726]: DEBUG nova.virt.hardware [None req-b479f246-d94b-4634-888f-dee4586c3272 tempest-ServerAddressesNegativeTestJSON-963263679 tempest-ServerAddressesNegativeTestJSON-963263679-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 750.278997] env[65726]: DEBUG nova.virt.hardware [None req-b479f246-d94b-4634-888f-dee4586c3272 tempest-ServerAddressesNegativeTestJSON-963263679 tempest-ServerAddressesNegativeTestJSON-963263679-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 750.279227] env[65726]: DEBUG nova.virt.hardware [None req-b479f246-d94b-4634-888f-dee4586c3272 tempest-ServerAddressesNegativeTestJSON-963263679 tempest-ServerAddressesNegativeTestJSON-963263679-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 750.282903] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23327744-ca26-4e5d-9101-eec57e1a1662 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.289414] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-915691aa-1b0c-480e-8ecd-871c0165c5b6 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.370641] env[65726]: DEBUG oslo_vmware.api [None req-b0adde83-84f2-4fa3-84bf-8f387eee1293 tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] Task: {'id': task-5115769, 'name': ReconfigVM_Task, 'duration_secs': 0.326922} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 750.370641] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-b0adde83-84f2-4fa3-84bf-8f387eee1293 tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] [instance: 3827eed1-b7a1-4922-a4de-f0c038cb8566] Reconfigured VM instance instance-00000025 to attach disk [datastore1] 3827eed1-b7a1-4922-a4de-f0c038cb8566/3827eed1-b7a1-4922-a4de-f0c038cb8566.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 750.371086] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-5ca8a158-79be-4241-8d17-3a6fe9259e97 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.382370] env[65726]: DEBUG oslo_vmware.api [None req-b0adde83-84f2-4fa3-84bf-8f387eee1293 tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] Waiting for the task: (returnval){ [ 750.382370] env[65726]: value = "task-5115770" [ 750.382370] env[65726]: _type = "Task" [ 750.382370] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 750.389700] env[65726]: DEBUG oslo_vmware.api [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Task: {'id': task-5115767, 'name': PowerOnVM_Task, 'duration_secs': 1.542937} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 750.390488] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] [instance: ca67d1a4-d304-4d06-a436-f257b60ca080] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 750.390777] env[65726]: INFO nova.compute.manager [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] [instance: ca67d1a4-d304-4d06-a436-f257b60ca080] Took 8.59 seconds to spawn the instance on the hypervisor. [ 750.391035] env[65726]: DEBUG nova.compute.manager [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] [instance: ca67d1a4-d304-4d06-a436-f257b60ca080] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 750.391907] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4cf2d30e-6399-45ce-832c-b4a6427b92d1 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.398600] env[65726]: DEBUG oslo_vmware.api [None req-b0adde83-84f2-4fa3-84bf-8f387eee1293 tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] Task: {'id': task-5115770, 'name': Rename_Task} progress is 10%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 750.603742] env[65726]: DEBUG nova.scheduler.client.report [None req-04afb47c-692a-46d9-a640-68cc2c90206f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Updated inventory for provider 07c4692f-bdb4-4058-9173-ff9664830295 with generation 67 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 750.604088] env[65726]: DEBUG nova.compute.provider_tree [None req-04afb47c-692a-46d9-a640-68cc2c90206f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Updating resource provider 07c4692f-bdb4-4058-9173-ff9664830295 generation from 67 to 68 during operation: update_inventory {{(pid=65726) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 750.604274] env[65726]: DEBUG nova.compute.provider_tree [None req-04afb47c-692a-46d9-a640-68cc2c90206f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Updating inventory in ProviderTree for provider 07c4692f-bdb4-4058-9173-ff9664830295 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 750.762482] env[65726]: DEBUG nova.network.neutron [None req-b479f246-d94b-4634-888f-dee4586c3272 tempest-ServerAddressesNegativeTestJSON-963263679 tempest-ServerAddressesNegativeTestJSON-963263679-project-member] [instance: 00746508-d0d6-4dfa-9026-772398d004ee] Successfully updated port: 1151551e-12f0-4e85-a385-f1ebf4e9bc72 {{(pid=65726) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 750.894659] env[65726]: DEBUG oslo_vmware.api [None req-b0adde83-84f2-4fa3-84bf-8f387eee1293 tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] Task: {'id': task-5115770, 'name': Rename_Task} progress is 99%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 750.915693] env[65726]: INFO nova.compute.manager [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] [instance: ca67d1a4-d304-4d06-a436-f257b60ca080] Took 58.87 seconds to build instance. [ 751.110616] env[65726]: DEBUG oslo_concurrency.lockutils [None req-04afb47c-692a-46d9-a640-68cc2c90206f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.886s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 751.111285] env[65726]: DEBUG nova.compute.manager [None req-04afb47c-692a-46d9-a640-68cc2c90206f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: c108f2a5-031b-47ee-9a5b-d62c2f42c26b] Start building networks asynchronously for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 751.117974] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5e5bfa8e-de8f-41bf-8724-f77b78155db2 tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 43.908s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 751.118221] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5e5bfa8e-de8f-41bf-8724-f77b78155db2 tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 751.121210] env[65726]: DEBUG oslo_concurrency.lockutils [None req-8b18678f-63d3-424e-98c8-800d5081389c tempest-VolumesAdminNegativeTest-2004477850 tempest-VolumesAdminNegativeTest-2004477850-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 41.977s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 751.121210] env[65726]: DEBUG oslo_concurrency.lockutils [None req-8b18678f-63d3-424e-98c8-800d5081389c tempest-VolumesAdminNegativeTest-2004477850 tempest-VolumesAdminNegativeTest-2004477850-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 751.123199] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c00c2da9-ebd0-4704-a276-dac36258bf53 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 36.844s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 751.123394] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c00c2da9-ebd0-4704-a276-dac36258bf53 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 751.128569] env[65726]: DEBUG oslo_concurrency.lockutils [None req-1023a64a-e0ea-45c1-b4cf-b0858cc7b3e9 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 35.857s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 751.128569] env[65726]: DEBUG oslo_concurrency.lockutils [None req-1023a64a-e0ea-45c1-b4cf-b0858cc7b3e9 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 751.128569] env[65726]: DEBUG oslo_concurrency.lockutils [None req-681fb0ed-e41c-4d7e-b603-04df1dff2263 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 35.526s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 751.128569] env[65726]: DEBUG oslo_concurrency.lockutils [None req-681fb0ed-e41c-4d7e-b603-04df1dff2263 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 751.130172] env[65726]: DEBUG oslo_concurrency.lockutils [None req-44e83fc7-afcc-4134-bbbb-6bc576931bb8 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 33.818s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 751.132778] env[65726]: INFO nova.compute.claims [None req-44e83fc7-afcc-4134-bbbb-6bc576931bb8 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: 895797e4-2941-44cd-aab1-67afa4fac02d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 751.171381] env[65726]: INFO nova.scheduler.client.report [None req-8b18678f-63d3-424e-98c8-800d5081389c tempest-VolumesAdminNegativeTest-2004477850 tempest-VolumesAdminNegativeTest-2004477850-project-member] Deleted allocations for instance 30b17a17-229c-4ca7-9ae6-c67ccdcd3963 [ 751.176741] env[65726]: INFO nova.scheduler.client.report [None req-1023a64a-e0ea-45c1-b4cf-b0858cc7b3e9 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Deleted allocations for instance 3a60b719-9db7-436e-9908-25f6a233c465 [ 751.192877] env[65726]: INFO nova.scheduler.client.report [None req-5e5bfa8e-de8f-41bf-8724-f77b78155db2 tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] Deleted allocations for instance ddd36db0-28ad-4582-994b-7eef8b08ed46 [ 751.194592] env[65726]: INFO nova.scheduler.client.report [None req-c00c2da9-ebd0-4704-a276-dac36258bf53 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Deleted allocations for instance f5151062-57b1-4e4e-93f4-aab0e7f504d4 [ 751.220233] env[65726]: INFO nova.scheduler.client.report [None req-681fb0ed-e41c-4d7e-b603-04df1dff2263 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Deleted allocations for instance 04ffdf80-c54c-4587-9bb4-d520dc440501 [ 751.267557] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b479f246-d94b-4634-888f-dee4586c3272 tempest-ServerAddressesNegativeTestJSON-963263679 tempest-ServerAddressesNegativeTestJSON-963263679-project-member] Acquiring lock "refresh_cache-00746508-d0d6-4dfa-9026-772398d004ee" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 751.267830] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b479f246-d94b-4634-888f-dee4586c3272 tempest-ServerAddressesNegativeTestJSON-963263679 tempest-ServerAddressesNegativeTestJSON-963263679-project-member] Acquired lock "refresh_cache-00746508-d0d6-4dfa-9026-772398d004ee" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 751.267982] env[65726]: DEBUG nova.network.neutron [None req-b479f246-d94b-4634-888f-dee4586c3272 tempest-ServerAddressesNegativeTestJSON-963263679 tempest-ServerAddressesNegativeTestJSON-963263679-project-member] [instance: 00746508-d0d6-4dfa-9026-772398d004ee] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 751.396744] env[65726]: DEBUG oslo_vmware.api [None req-b0adde83-84f2-4fa3-84bf-8f387eee1293 tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] Task: {'id': task-5115770, 'name': Rename_Task} progress is 99%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 751.419817] env[65726]: DEBUG oslo_concurrency.lockutils [None req-4a3033b7-02c6-495e-930c-e113b9bee69a tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Lock "ca67d1a4-d304-4d06-a436-f257b60ca080" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 73.610s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 751.536274] env[65726]: DEBUG nova.compute.manager [req-b64a30ff-d200-4fb5-a1c4-d3e35fb7c3fe req-e00c783e-8845-43cf-a7dd-d710cd553d69 service nova] [instance: 00746508-d0d6-4dfa-9026-772398d004ee] Received event network-vif-plugged-1151551e-12f0-4e85-a385-f1ebf4e9bc72 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 751.536659] env[65726]: DEBUG oslo_concurrency.lockutils [req-b64a30ff-d200-4fb5-a1c4-d3e35fb7c3fe req-e00c783e-8845-43cf-a7dd-d710cd553d69 service nova] Acquiring lock "00746508-d0d6-4dfa-9026-772398d004ee-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 751.536993] env[65726]: DEBUG oslo_concurrency.lockutils [req-b64a30ff-d200-4fb5-a1c4-d3e35fb7c3fe req-e00c783e-8845-43cf-a7dd-d710cd553d69 service nova] Lock "00746508-d0d6-4dfa-9026-772398d004ee-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 751.537220] env[65726]: DEBUG oslo_concurrency.lockutils [req-b64a30ff-d200-4fb5-a1c4-d3e35fb7c3fe req-e00c783e-8845-43cf-a7dd-d710cd553d69 service nova] Lock "00746508-d0d6-4dfa-9026-772398d004ee-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 751.537426] env[65726]: DEBUG nova.compute.manager [req-b64a30ff-d200-4fb5-a1c4-d3e35fb7c3fe req-e00c783e-8845-43cf-a7dd-d710cd553d69 service nova] [instance: 00746508-d0d6-4dfa-9026-772398d004ee] No waiting events found dispatching network-vif-plugged-1151551e-12f0-4e85-a385-f1ebf4e9bc72 {{(pid=65726) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 751.537659] env[65726]: WARNING nova.compute.manager [req-b64a30ff-d200-4fb5-a1c4-d3e35fb7c3fe req-e00c783e-8845-43cf-a7dd-d710cd553d69 service nova] [instance: 00746508-d0d6-4dfa-9026-772398d004ee] Received unexpected event network-vif-plugged-1151551e-12f0-4e85-a385-f1ebf4e9bc72 for instance with vm_state building and task_state spawning. [ 751.645547] env[65726]: DEBUG nova.compute.utils [None req-04afb47c-692a-46d9-a640-68cc2c90206f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Using /dev/sd instead of None {{(pid=65726) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 751.647047] env[65726]: DEBUG nova.compute.manager [None req-04afb47c-692a-46d9-a640-68cc2c90206f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: c108f2a5-031b-47ee-9a5b-d62c2f42c26b] Allocating IP information in the background. {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 751.647259] env[65726]: DEBUG nova.network.neutron [None req-04afb47c-692a-46d9-a640-68cc2c90206f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: c108f2a5-031b-47ee-9a5b-d62c2f42c26b] allocate_for_instance() {{(pid=65726) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 751.647595] env[65726]: WARNING neutronclient.v2_0.client [None req-04afb47c-692a-46d9-a640-68cc2c90206f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 751.647942] env[65726]: WARNING neutronclient.v2_0.client [None req-04afb47c-692a-46d9-a640-68cc2c90206f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 751.649882] env[65726]: WARNING openstack [None req-04afb47c-692a-46d9-a640-68cc2c90206f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 751.653354] env[65726]: WARNING openstack [None req-04afb47c-692a-46d9-a640-68cc2c90206f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 751.688385] env[65726]: DEBUG oslo_concurrency.lockutils [None req-8b18678f-63d3-424e-98c8-800d5081389c tempest-VolumesAdminNegativeTest-2004477850 tempest-VolumesAdminNegativeTest-2004477850-project-member] Lock "30b17a17-229c-4ca7-9ae6-c67ccdcd3963" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 46.564s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 751.698353] env[65726]: DEBUG oslo_concurrency.lockutils [None req-81594943-9858-4765-a371-42f9ef0cbec9 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] Acquiring lock "3cbbe146-be3a-43d5-867d-4d669884758c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 751.698599] env[65726]: DEBUG oslo_concurrency.lockutils [None req-81594943-9858-4765-a371-42f9ef0cbec9 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] Lock "3cbbe146-be3a-43d5-867d-4d669884758c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 751.700698] env[65726]: DEBUG oslo_concurrency.lockutils [None req-1023a64a-e0ea-45c1-b4cf-b0858cc7b3e9 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Lock "3a60b719-9db7-436e-9908-25f6a233c465" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 39.970s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 751.712991] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c00c2da9-ebd0-4704-a276-dac36258bf53 tempest-ListServersNegativeTestJSON-1603812164 tempest-ListServersNegativeTestJSON-1603812164-project-member] Lock "f5151062-57b1-4e4e-93f4-aab0e7f504d4" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 40.977s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 751.714757] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5e5bfa8e-de8f-41bf-8724-f77b78155db2 tempest-ServersListShow298Test-1149015439 tempest-ServersListShow298Test-1149015439-project-member] Lock "ddd36db0-28ad-4582-994b-7eef8b08ed46" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 48.858s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 751.724743] env[65726]: DEBUG nova.policy [None req-04afb47c-692a-46d9-a640-68cc2c90206f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4b6c9cc4ac3c487693240e375a83ea19', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b794c99309e94c1181fc9d8eeeb84702', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65726) authorize /opt/stack/nova/nova/policy.py:192}} [ 751.734249] env[65726]: DEBUG oslo_concurrency.lockutils [None req-681fb0ed-e41c-4d7e-b603-04df1dff2263 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Lock "04ffdf80-c54c-4587-9bb4-d520dc440501" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 39.902s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 751.771551] env[65726]: WARNING openstack [None req-b479f246-d94b-4634-888f-dee4586c3272 tempest-ServerAddressesNegativeTestJSON-963263679 tempest-ServerAddressesNegativeTestJSON-963263679-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 751.772016] env[65726]: WARNING openstack [None req-b479f246-d94b-4634-888f-dee4586c3272 tempest-ServerAddressesNegativeTestJSON-963263679 tempest-ServerAddressesNegativeTestJSON-963263679-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 751.781424] env[65726]: DEBUG nova.network.neutron [None req-b479f246-d94b-4634-888f-dee4586c3272 tempest-ServerAddressesNegativeTestJSON-963263679 tempest-ServerAddressesNegativeTestJSON-963263679-project-member] [instance: 00746508-d0d6-4dfa-9026-772398d004ee] Instance cache missing network info. {{(pid=65726) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 751.895741] env[65726]: DEBUG oslo_vmware.api [None req-b0adde83-84f2-4fa3-84bf-8f387eee1293 tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] Task: {'id': task-5115770, 'name': Rename_Task, 'duration_secs': 1.153532} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 751.896238] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-b0adde83-84f2-4fa3-84bf-8f387eee1293 tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] [instance: 3827eed1-b7a1-4922-a4de-f0c038cb8566] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 751.896426] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e2d77f22-2867-403e-8475-6c3fe065b297 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.903459] env[65726]: DEBUG oslo_vmware.api [None req-b0adde83-84f2-4fa3-84bf-8f387eee1293 tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] Waiting for the task: (returnval){ [ 751.903459] env[65726]: value = "task-5115771" [ 751.903459] env[65726]: _type = "Task" [ 751.903459] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 751.913401] env[65726]: DEBUG oslo_vmware.api [None req-b0adde83-84f2-4fa3-84bf-8f387eee1293 tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] Task: {'id': task-5115771, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 751.924984] env[65726]: DEBUG nova.compute.manager [None req-300cacab-e765-49c9-a8a1-8e47a20dc6a2 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] [instance: 76249623-6f83-46a3-b8c5-c001111aa698] Starting instance... {{(pid=65726) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 751.987065] env[65726]: DEBUG nova.network.neutron [None req-04afb47c-692a-46d9-a640-68cc2c90206f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: c108f2a5-031b-47ee-9a5b-d62c2f42c26b] Successfully created port: 981741c6-69f0-42d4-87be-e1c8f62d0e17 {{(pid=65726) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 752.160290] env[65726]: DEBUG nova.compute.manager [None req-04afb47c-692a-46d9-a640-68cc2c90206f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: c108f2a5-031b-47ee-9a5b-d62c2f42c26b] Start building block device mappings for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 752.420320] env[65726]: DEBUG oslo_vmware.api [None req-b0adde83-84f2-4fa3-84bf-8f387eee1293 tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] Task: {'id': task-5115771, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 752.430257] env[65726]: WARNING openstack [None req-b479f246-d94b-4634-888f-dee4586c3272 tempest-ServerAddressesNegativeTestJSON-963263679 tempest-ServerAddressesNegativeTestJSON-963263679-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 752.430257] env[65726]: WARNING openstack [None req-b479f246-d94b-4634-888f-dee4586c3272 tempest-ServerAddressesNegativeTestJSON-963263679 tempest-ServerAddressesNegativeTestJSON-963263679-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 752.463332] env[65726]: DEBUG oslo_concurrency.lockutils [None req-300cacab-e765-49c9-a8a1-8e47a20dc6a2 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 752.740415] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a660a90-85c9-4a3c-9e7e-20652c8d676b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.750052] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f5657d9-3e10-4c47-ad15-316d32ea2f55 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.783491] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99a4e2d7-e341-40c0-a70b-5c6250e031fb {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.792931] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13336c91-c529-47e3-83f7-877da5a1395e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.807535] env[65726]: DEBUG nova.compute.provider_tree [None req-44e83fc7-afcc-4134-bbbb-6bc576931bb8 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 752.915134] env[65726]: DEBUG oslo_vmware.api [None req-b0adde83-84f2-4fa3-84bf-8f387eee1293 tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] Task: {'id': task-5115771, 'name': PowerOnVM_Task, 'duration_secs': 0.620163} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 752.916209] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-b0adde83-84f2-4fa3-84bf-8f387eee1293 tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] [instance: 3827eed1-b7a1-4922-a4de-f0c038cb8566] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 752.916209] env[65726]: INFO nova.compute.manager [None req-b0adde83-84f2-4fa3-84bf-8f387eee1293 tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] [instance: 3827eed1-b7a1-4922-a4de-f0c038cb8566] Took 6.89 seconds to spawn the instance on the hypervisor. [ 752.916209] env[65726]: DEBUG nova.compute.manager [None req-b0adde83-84f2-4fa3-84bf-8f387eee1293 tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] [instance: 3827eed1-b7a1-4922-a4de-f0c038cb8566] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 752.916611] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a8dbf97-9e9a-486b-a521-4bd8fc779a37 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.173624] env[65726]: DEBUG nova.compute.manager [None req-04afb47c-692a-46d9-a640-68cc2c90206f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: c108f2a5-031b-47ee-9a5b-d62c2f42c26b] Start spawning the instance on the hypervisor. {{(pid=65726) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 753.199067] env[65726]: WARNING openstack [None req-b479f246-d94b-4634-888f-dee4586c3272 tempest-ServerAddressesNegativeTestJSON-963263679 tempest-ServerAddressesNegativeTestJSON-963263679-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 753.199931] env[65726]: WARNING openstack [None req-b479f246-d94b-4634-888f-dee4586c3272 tempest-ServerAddressesNegativeTestJSON-963263679 tempest-ServerAddressesNegativeTestJSON-963263679-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 753.210978] env[65726]: DEBUG nova.virt.hardware [None req-04afb47c-692a-46d9-a640-68cc2c90206f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T19:28:34Z,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1539f06fa3534e90acbc3a60c4b8bd3f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T19:28:35Z,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 753.211342] env[65726]: DEBUG nova.virt.hardware [None req-04afb47c-692a-46d9-a640-68cc2c90206f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 753.211413] env[65726]: DEBUG nova.virt.hardware [None req-04afb47c-692a-46d9-a640-68cc2c90206f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 753.211582] env[65726]: DEBUG nova.virt.hardware [None req-04afb47c-692a-46d9-a640-68cc2c90206f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 753.211987] env[65726]: DEBUG nova.virt.hardware [None req-04afb47c-692a-46d9-a640-68cc2c90206f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 753.211987] env[65726]: DEBUG nova.virt.hardware [None req-04afb47c-692a-46d9-a640-68cc2c90206f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 753.212124] env[65726]: DEBUG nova.virt.hardware [None req-04afb47c-692a-46d9-a640-68cc2c90206f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 753.212308] env[65726]: DEBUG nova.virt.hardware [None req-04afb47c-692a-46d9-a640-68cc2c90206f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 753.212431] env[65726]: DEBUG nova.virt.hardware [None req-04afb47c-692a-46d9-a640-68cc2c90206f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 753.212591] env[65726]: DEBUG nova.virt.hardware [None req-04afb47c-692a-46d9-a640-68cc2c90206f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 753.212759] env[65726]: DEBUG nova.virt.hardware [None req-04afb47c-692a-46d9-a640-68cc2c90206f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 753.213966] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e706bdc6-833a-447f-b83f-60ef29755281 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.224153] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27e9de7b-da4f-47c3-9ba6-fce55ebb59a0 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.310620] env[65726]: DEBUG nova.scheduler.client.report [None req-44e83fc7-afcc-4134-bbbb-6bc576931bb8 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 753.434191] env[65726]: INFO nova.compute.manager [None req-b0adde83-84f2-4fa3-84bf-8f387eee1293 tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] [instance: 3827eed1-b7a1-4922-a4de-f0c038cb8566] Took 52.75 seconds to build instance. [ 753.608853] env[65726]: DEBUG oslo_concurrency.lockutils [None req-1640464d-fd11-4b5f-b0e0-f0de74fa607d tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Acquiring lock "bd839fe1-8801-4ba1-9c23-288453258b66" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 753.608853] env[65726]: DEBUG oslo_concurrency.lockutils [None req-1640464d-fd11-4b5f-b0e0-f0de74fa607d tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Lock "bd839fe1-8801-4ba1-9c23-288453258b66" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 753.608853] env[65726]: DEBUG oslo_concurrency.lockutils [None req-1640464d-fd11-4b5f-b0e0-f0de74fa607d tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Acquiring lock "bd839fe1-8801-4ba1-9c23-288453258b66-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 753.608853] env[65726]: DEBUG oslo_concurrency.lockutils [None req-1640464d-fd11-4b5f-b0e0-f0de74fa607d tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Lock "bd839fe1-8801-4ba1-9c23-288453258b66-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 753.610164] env[65726]: DEBUG oslo_concurrency.lockutils [None req-1640464d-fd11-4b5f-b0e0-f0de74fa607d tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Lock "bd839fe1-8801-4ba1-9c23-288453258b66-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 753.611910] env[65726]: INFO nova.compute.manager [None req-1640464d-fd11-4b5f-b0e0-f0de74fa607d tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: bd839fe1-8801-4ba1-9c23-288453258b66] Terminating instance [ 753.815153] env[65726]: DEBUG nova.network.neutron [None req-04afb47c-692a-46d9-a640-68cc2c90206f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: c108f2a5-031b-47ee-9a5b-d62c2f42c26b] Successfully updated port: 981741c6-69f0-42d4-87be-e1c8f62d0e17 {{(pid=65726) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 753.816947] env[65726]: DEBUG oslo_concurrency.lockutils [None req-44e83fc7-afcc-4134-bbbb-6bc576931bb8 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.687s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 753.817506] env[65726]: DEBUG nova.compute.manager [None req-44e83fc7-afcc-4134-bbbb-6bc576931bb8 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: 895797e4-2941-44cd-aab1-67afa4fac02d] Start building networks asynchronously for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 753.823558] env[65726]: DEBUG oslo_concurrency.lockutils [None req-af2f0c1d-9bc5-40c3-a712-d408e73bc806 tempest-ServerDiagnosticsNegativeTest-1635132145 tempest-ServerDiagnosticsNegativeTest-1635132145-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 35.660s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 753.827105] env[65726]: INFO nova.compute.claims [None req-af2f0c1d-9bc5-40c3-a712-d408e73bc806 tempest-ServerDiagnosticsNegativeTest-1635132145 tempest-ServerDiagnosticsNegativeTest-1635132145-project-member] [instance: 0d6f14c5-5770-476f-a186-4cf1b2d4fbe8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 753.852924] env[65726]: DEBUG nova.network.neutron [None req-b479f246-d94b-4634-888f-dee4586c3272 tempest-ServerAddressesNegativeTestJSON-963263679 tempest-ServerAddressesNegativeTestJSON-963263679-project-member] [instance: 00746508-d0d6-4dfa-9026-772398d004ee] Updating instance_info_cache with network_info: [{"id": "1151551e-12f0-4e85-a385-f1ebf4e9bc72", "address": "fa:16:3e:eb:da:97", "network": {"id": "833c6007-53da-4e80-95f8-cd5f400ba120", "bridge": "br-int", "label": "tempest-ServerAddressesNegativeTestJSON-1292797458-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0d2e9e9fe170484896239f5f9cc9dc09", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69f65356-c85e-4b7f-ad28-7c7b5e8cf50c", "external-id": "nsx-vlan-transportzone-281", "segmentation_id": 281, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1151551e-12", "ovs_interfaceid": "1151551e-12f0-4e85-a385-f1ebf4e9bc72", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 753.937687] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b0adde83-84f2-4fa3-84bf-8f387eee1293 tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] Lock "3827eed1-b7a1-4922-a4de-f0c038cb8566" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 75.793s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 754.116935] env[65726]: DEBUG nova.compute.manager [None req-1640464d-fd11-4b5f-b0e0-f0de74fa607d tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: bd839fe1-8801-4ba1-9c23-288453258b66] Start destroying the instance on the hypervisor. {{(pid=65726) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 754.117284] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-1640464d-fd11-4b5f-b0e0-f0de74fa607d tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: bd839fe1-8801-4ba1-9c23-288453258b66] Destroying instance {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 754.118271] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9908f995-d060-4eaf-ba23-632592e594b3 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.126827] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-1640464d-fd11-4b5f-b0e0-f0de74fa607d tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: bd839fe1-8801-4ba1-9c23-288453258b66] Unregistering the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 754.127113] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d33e15e0-d9d1-4b7f-9689-30175c0f7929 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.203705] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-1640464d-fd11-4b5f-b0e0-f0de74fa607d tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: bd839fe1-8801-4ba1-9c23-288453258b66] Unregistered the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 754.203942] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-1640464d-fd11-4b5f-b0e0-f0de74fa607d tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: bd839fe1-8801-4ba1-9c23-288453258b66] Deleting contents of the VM from datastore datastore2 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 754.204146] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-1640464d-fd11-4b5f-b0e0-f0de74fa607d tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Deleting the datastore file [datastore2] bd839fe1-8801-4ba1-9c23-288453258b66 {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 754.204419] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-14804e42-0ebd-4c5f-a417-a67929151dba {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.212114] env[65726]: DEBUG oslo_vmware.api [None req-1640464d-fd11-4b5f-b0e0-f0de74fa607d tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Waiting for the task: (returnval){ [ 754.212114] env[65726]: value = "task-5115773" [ 754.212114] env[65726]: _type = "Task" [ 754.212114] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 754.222376] env[65726]: DEBUG oslo_vmware.api [None req-1640464d-fd11-4b5f-b0e0-f0de74fa607d tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': task-5115773, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 754.324497] env[65726]: DEBUG oslo_concurrency.lockutils [None req-04afb47c-692a-46d9-a640-68cc2c90206f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Acquiring lock "refresh_cache-c108f2a5-031b-47ee-9a5b-d62c2f42c26b" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 754.324769] env[65726]: DEBUG oslo_concurrency.lockutils [None req-04afb47c-692a-46d9-a640-68cc2c90206f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Acquired lock "refresh_cache-c108f2a5-031b-47ee-9a5b-d62c2f42c26b" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 754.325354] env[65726]: DEBUG nova.network.neutron [None req-04afb47c-692a-46d9-a640-68cc2c90206f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: c108f2a5-031b-47ee-9a5b-d62c2f42c26b] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 754.331335] env[65726]: DEBUG nova.compute.utils [None req-44e83fc7-afcc-4134-bbbb-6bc576931bb8 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Using /dev/sd instead of None {{(pid=65726) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 754.335858] env[65726]: DEBUG nova.compute.manager [None req-44e83fc7-afcc-4134-bbbb-6bc576931bb8 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: 895797e4-2941-44cd-aab1-67afa4fac02d] Allocating IP information in the background. {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 754.335858] env[65726]: DEBUG nova.network.neutron [None req-44e83fc7-afcc-4134-bbbb-6bc576931bb8 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: 895797e4-2941-44cd-aab1-67afa4fac02d] allocate_for_instance() {{(pid=65726) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 754.337351] env[65726]: WARNING neutronclient.v2_0.client [None req-44e83fc7-afcc-4134-bbbb-6bc576931bb8 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 754.337351] env[65726]: WARNING neutronclient.v2_0.client [None req-44e83fc7-afcc-4134-bbbb-6bc576931bb8 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 754.337351] env[65726]: WARNING openstack [None req-44e83fc7-afcc-4134-bbbb-6bc576931bb8 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 754.337762] env[65726]: WARNING openstack [None req-44e83fc7-afcc-4134-bbbb-6bc576931bb8 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 754.355804] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b479f246-d94b-4634-888f-dee4586c3272 tempest-ServerAddressesNegativeTestJSON-963263679 tempest-ServerAddressesNegativeTestJSON-963263679-project-member] Releasing lock "refresh_cache-00746508-d0d6-4dfa-9026-772398d004ee" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 754.356242] env[65726]: DEBUG nova.compute.manager [None req-b479f246-d94b-4634-888f-dee4586c3272 tempest-ServerAddressesNegativeTestJSON-963263679 tempest-ServerAddressesNegativeTestJSON-963263679-project-member] [instance: 00746508-d0d6-4dfa-9026-772398d004ee] Instance network_info: |[{"id": "1151551e-12f0-4e85-a385-f1ebf4e9bc72", "address": "fa:16:3e:eb:da:97", "network": {"id": "833c6007-53da-4e80-95f8-cd5f400ba120", "bridge": "br-int", "label": "tempest-ServerAddressesNegativeTestJSON-1292797458-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0d2e9e9fe170484896239f5f9cc9dc09", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69f65356-c85e-4b7f-ad28-7c7b5e8cf50c", "external-id": "nsx-vlan-transportzone-281", "segmentation_id": 281, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1151551e-12", "ovs_interfaceid": "1151551e-12f0-4e85-a385-f1ebf4e9bc72", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 754.356722] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-b479f246-d94b-4634-888f-dee4586c3272 tempest-ServerAddressesNegativeTestJSON-963263679 tempest-ServerAddressesNegativeTestJSON-963263679-project-member] [instance: 00746508-d0d6-4dfa-9026-772398d004ee] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:eb:da:97', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '69f65356-c85e-4b7f-ad28-7c7b5e8cf50c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '1151551e-12f0-4e85-a385-f1ebf4e9bc72', 'vif_model': 'vmxnet3'}] {{(pid=65726) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 754.366874] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-b479f246-d94b-4634-888f-dee4586c3272 tempest-ServerAddressesNegativeTestJSON-963263679 tempest-ServerAddressesNegativeTestJSON-963263679-project-member] Creating folder: Project (0d2e9e9fe170484896239f5f9cc9dc09). Parent ref: group-v995008. {{(pid=65726) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 754.367672] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-2471e7ab-2e06-49b2-a36f-f35f7c6c01b2 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.381018] env[65726]: INFO nova.virt.vmwareapi.vm_util [None req-b479f246-d94b-4634-888f-dee4586c3272 tempest-ServerAddressesNegativeTestJSON-963263679 tempest-ServerAddressesNegativeTestJSON-963263679-project-member] Created folder: Project (0d2e9e9fe170484896239f5f9cc9dc09) in parent group-v995008. [ 754.381077] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-b479f246-d94b-4634-888f-dee4586c3272 tempest-ServerAddressesNegativeTestJSON-963263679 tempest-ServerAddressesNegativeTestJSON-963263679-project-member] Creating folder: Instances. Parent ref: group-v995109. {{(pid=65726) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 754.381522] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-3f3c1c1a-72b3-4869-b5a2-6a2408d00511 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.398662] env[65726]: INFO nova.virt.vmwareapi.vm_util [None req-b479f246-d94b-4634-888f-dee4586c3272 tempest-ServerAddressesNegativeTestJSON-963263679 tempest-ServerAddressesNegativeTestJSON-963263679-project-member] Created folder: Instances in parent group-v995109. [ 754.398662] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-b479f246-d94b-4634-888f-dee4586c3272 tempest-ServerAddressesNegativeTestJSON-963263679 tempest-ServerAddressesNegativeTestJSON-963263679-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 754.398994] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 00746508-d0d6-4dfa-9026-772398d004ee] Creating VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 754.399650] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1f087b3b-49d6-4c9e-859f-4800c095213a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.426060] env[65726]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 754.426060] env[65726]: value = "task-5115776" [ 754.426060] env[65726]: _type = "Task" [ 754.426060] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 754.435829] env[65726]: DEBUG nova.policy [None req-44e83fc7-afcc-4134-bbbb-6bc576931bb8 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f3050294b8f34a0f97d10b038b048779', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a090d553766847e58e3231b966c92565', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65726) authorize /opt/stack/nova/nova/policy.py:192}} [ 754.442961] env[65726]: DEBUG nova.compute.manager [None req-af1f7990-88d5-411a-be3e-a2d4ed6b5b5b tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: 621b899a-e9ec-4fc0-a574-4a08cecff6d9] Starting instance... {{(pid=65726) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 754.450048] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115776, 'name': CreateVM_Task} progress is 6%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 754.726932] env[65726]: DEBUG oslo_vmware.api [None req-1640464d-fd11-4b5f-b0e0-f0de74fa607d tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': task-5115773, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.191943} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 754.727518] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-1640464d-fd11-4b5f-b0e0-f0de74fa607d tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Deleted the datastore file {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 754.727808] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-1640464d-fd11-4b5f-b0e0-f0de74fa607d tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: bd839fe1-8801-4ba1-9c23-288453258b66] Deleted contents of the VM from datastore datastore2 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 754.728090] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-1640464d-fd11-4b5f-b0e0-f0de74fa607d tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: bd839fe1-8801-4ba1-9c23-288453258b66] Instance destroyed {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 754.728297] env[65726]: INFO nova.compute.manager [None req-1640464d-fd11-4b5f-b0e0-f0de74fa607d tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: bd839fe1-8801-4ba1-9c23-288453258b66] Took 0.61 seconds to destroy the instance on the hypervisor. [ 754.728675] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-1640464d-fd11-4b5f-b0e0-f0de74fa607d tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 754.728913] env[65726]: DEBUG nova.compute.manager [-] [instance: bd839fe1-8801-4ba1-9c23-288453258b66] Deallocating network for instance {{(pid=65726) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 754.729060] env[65726]: DEBUG nova.network.neutron [-] [instance: bd839fe1-8801-4ba1-9c23-288453258b66] deallocate_for_instance() {{(pid=65726) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 754.729720] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 754.729976] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 754.782914] env[65726]: DEBUG nova.network.neutron [None req-44e83fc7-afcc-4134-bbbb-6bc576931bb8 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: 895797e4-2941-44cd-aab1-67afa4fac02d] Successfully created port: a1dec2db-0e1e-45dd-8587-a84e4a5e2ef4 {{(pid=65726) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 754.830266] env[65726]: WARNING openstack [None req-04afb47c-692a-46d9-a640-68cc2c90206f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 754.830591] env[65726]: WARNING openstack [None req-04afb47c-692a-46d9-a640-68cc2c90206f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 754.837266] env[65726]: DEBUG nova.network.neutron [None req-04afb47c-692a-46d9-a640-68cc2c90206f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: c108f2a5-031b-47ee-9a5b-d62c2f42c26b] Instance cache missing network info. {{(pid=65726) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 754.846121] env[65726]: DEBUG nova.compute.manager [None req-44e83fc7-afcc-4134-bbbb-6bc576931bb8 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: 895797e4-2941-44cd-aab1-67afa4fac02d] Start building block device mappings for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 754.851260] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 754.851815] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 754.940527] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115776, 'name': CreateVM_Task, 'duration_secs': 0.371724} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 754.941411] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 00746508-d0d6-4dfa-9026-772398d004ee] Created VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 754.942133] env[65726]: WARNING openstack [None req-b479f246-d94b-4634-888f-dee4586c3272 tempest-ServerAddressesNegativeTestJSON-963263679 tempest-ServerAddressesNegativeTestJSON-963263679-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 754.942380] env[65726]: WARNING openstack [None req-b479f246-d94b-4634-888f-dee4586c3272 tempest-ServerAddressesNegativeTestJSON-963263679 tempest-ServerAddressesNegativeTestJSON-963263679-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 754.947739] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b479f246-d94b-4634-888f-dee4586c3272 tempest-ServerAddressesNegativeTestJSON-963263679 tempest-ServerAddressesNegativeTestJSON-963263679-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 754.947969] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b479f246-d94b-4634-888f-dee4586c3272 tempest-ServerAddressesNegativeTestJSON-963263679 tempest-ServerAddressesNegativeTestJSON-963263679-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 754.948218] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b479f246-d94b-4634-888f-dee4586c3272 tempest-ServerAddressesNegativeTestJSON-963263679 tempest-ServerAddressesNegativeTestJSON-963263679-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 754.950938] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c9d1d76a-4979-457b-987a-3b181a5d4a2f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.959197] env[65726]: DEBUG oslo_vmware.api [None req-b479f246-d94b-4634-888f-dee4586c3272 tempest-ServerAddressesNegativeTestJSON-963263679 tempest-ServerAddressesNegativeTestJSON-963263679-project-member] Waiting for the task: (returnval){ [ 754.959197] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]520f214c-6959-bcf9-e2a6-2cc3b1430f72" [ 754.959197] env[65726]: _type = "Task" [ 754.959197] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 754.968511] env[65726]: DEBUG oslo_vmware.api [None req-b479f246-d94b-4634-888f-dee4586c3272 tempest-ServerAddressesNegativeTestJSON-963263679 tempest-ServerAddressesNegativeTestJSON-963263679-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]520f214c-6959-bcf9-e2a6-2cc3b1430f72, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 754.973027] env[65726]: DEBUG oslo_concurrency.lockutils [None req-af1f7990-88d5-411a-be3e-a2d4ed6b5b5b tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 755.010946] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-796dbafb-e7e8-4f01-b7c6-4c13ce63ea86 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.022513] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d4b929a-b17d-4e4c-a9cb-48057d7303ee {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.061243] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e31d4fe-6184-4e6c-b102-ab59310c25ea {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.069698] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42cfeb5e-f752-4b34-9f15-d11dd5a14274 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.085033] env[65726]: DEBUG nova.compute.provider_tree [None req-af2f0c1d-9bc5-40c3-a712-d408e73bc806 tempest-ServerDiagnosticsNegativeTest-1635132145 tempest-ServerDiagnosticsNegativeTest-1635132145-project-member] Updating inventory in ProviderTree for provider 07c4692f-bdb4-4058-9173-ff9664830295 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 755.469988] env[65726]: DEBUG oslo_vmware.api [None req-b479f246-d94b-4634-888f-dee4586c3272 tempest-ServerAddressesNegativeTestJSON-963263679 tempest-ServerAddressesNegativeTestJSON-963263679-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]520f214c-6959-bcf9-e2a6-2cc3b1430f72, 'name': SearchDatastore_Task, 'duration_secs': 0.011601} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 755.470319] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b479f246-d94b-4634-888f-dee4586c3272 tempest-ServerAddressesNegativeTestJSON-963263679 tempest-ServerAddressesNegativeTestJSON-963263679-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 755.470551] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-b479f246-d94b-4634-888f-dee4586c3272 tempest-ServerAddressesNegativeTestJSON-963263679 tempest-ServerAddressesNegativeTestJSON-963263679-project-member] [instance: 00746508-d0d6-4dfa-9026-772398d004ee] Processing image b52362a3-ee8a-4cbf-b06f-513b0cc8f95c {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 755.470807] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b479f246-d94b-4634-888f-dee4586c3272 tempest-ServerAddressesNegativeTestJSON-963263679 tempest-ServerAddressesNegativeTestJSON-963263679-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 755.470952] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b479f246-d94b-4634-888f-dee4586c3272 tempest-ServerAddressesNegativeTestJSON-963263679 tempest-ServerAddressesNegativeTestJSON-963263679-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 755.471145] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-b479f246-d94b-4634-888f-dee4586c3272 tempest-ServerAddressesNegativeTestJSON-963263679 tempest-ServerAddressesNegativeTestJSON-963263679-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 755.471409] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-56910e2c-d1f3-4d6a-8f21-36358d02821f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.481529] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-b479f246-d94b-4634-888f-dee4586c3272 tempest-ServerAddressesNegativeTestJSON-963263679 tempest-ServerAddressesNegativeTestJSON-963263679-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 755.481766] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-b479f246-d94b-4634-888f-dee4586c3272 tempest-ServerAddressesNegativeTestJSON-963263679 tempest-ServerAddressesNegativeTestJSON-963263679-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65726) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 755.482583] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9222f574-44bc-4136-8696-b07881ff5f5a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.488299] env[65726]: DEBUG oslo_vmware.api [None req-b479f246-d94b-4634-888f-dee4586c3272 tempest-ServerAddressesNegativeTestJSON-963263679 tempest-ServerAddressesNegativeTestJSON-963263679-project-member] Waiting for the task: (returnval){ [ 755.488299] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5261937e-7fe9-0775-b53f-eee7481b66cf" [ 755.488299] env[65726]: _type = "Task" [ 755.488299] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 755.497060] env[65726]: DEBUG oslo_vmware.api [None req-b479f246-d94b-4634-888f-dee4586c3272 tempest-ServerAddressesNegativeTestJSON-963263679 tempest-ServerAddressesNegativeTestJSON-963263679-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5261937e-7fe9-0775-b53f-eee7481b66cf, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 755.586290] env[65726]: WARNING openstack [None req-04afb47c-692a-46d9-a640-68cc2c90206f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 755.586668] env[65726]: WARNING openstack [None req-04afb47c-692a-46d9-a640-68cc2c90206f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 755.613019] env[65726]: ERROR nova.scheduler.client.report [None req-af2f0c1d-9bc5-40c3-a712-d408e73bc806 tempest-ServerDiagnosticsNegativeTest-1635132145 tempest-ServerDiagnosticsNegativeTest-1635132145-project-member] [req-37dd1a1d-8d2e-40a9-8296-1b194bd9ed7f] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 07c4692f-bdb4-4058-9173-ff9664830295. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-37dd1a1d-8d2e-40a9-8296-1b194bd9ed7f"}]} [ 755.630182] env[65726]: DEBUG nova.scheduler.client.report [None req-af2f0c1d-9bc5-40c3-a712-d408e73bc806 tempest-ServerDiagnosticsNegativeTest-1635132145 tempest-ServerDiagnosticsNegativeTest-1635132145-project-member] Refreshing inventories for resource provider 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 755.643853] env[65726]: DEBUG nova.scheduler.client.report [None req-af2f0c1d-9bc5-40c3-a712-d408e73bc806 tempest-ServerDiagnosticsNegativeTest-1635132145 tempest-ServerDiagnosticsNegativeTest-1635132145-project-member] Updating ProviderTree inventory for provider 07c4692f-bdb4-4058-9173-ff9664830295 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 755.644846] env[65726]: DEBUG nova.compute.provider_tree [None req-af2f0c1d-9bc5-40c3-a712-d408e73bc806 tempest-ServerDiagnosticsNegativeTest-1635132145 tempest-ServerDiagnosticsNegativeTest-1635132145-project-member] Updating inventory in ProviderTree for provider 07c4692f-bdb4-4058-9173-ff9664830295 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 755.657032] env[65726]: DEBUG nova.scheduler.client.report [None req-af2f0c1d-9bc5-40c3-a712-d408e73bc806 tempest-ServerDiagnosticsNegativeTest-1635132145 tempest-ServerDiagnosticsNegativeTest-1635132145-project-member] Refreshing aggregate associations for resource provider 07c4692f-bdb4-4058-9173-ff9664830295, aggregates: None {{(pid=65726) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 755.675765] env[65726]: DEBUG nova.scheduler.client.report [None req-af2f0c1d-9bc5-40c3-a712-d408e73bc806 tempest-ServerDiagnosticsNegativeTest-1635132145 tempest-ServerDiagnosticsNegativeTest-1635132145-project-member] Refreshing trait associations for resource provider 07c4692f-bdb4-4058-9173-ff9664830295, traits: COMPUTE_IMAGE_TYPE_ISO,HW_ARCH_X86_64,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=65726) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 755.820030] env[65726]: DEBUG oslo_concurrency.lockutils [None req-0b112b08-1593-41b5-b38e-86e5b4d22081 tempest-ServersV294TestFqdnHostnames-819811677 tempest-ServersV294TestFqdnHostnames-819811677-project-member] Acquiring lock "c8be64a3-1bfa-41a3-b9be-68a7e6e052e0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 755.820318] env[65726]: DEBUG oslo_concurrency.lockutils [None req-0b112b08-1593-41b5-b38e-86e5b4d22081 tempest-ServersV294TestFqdnHostnames-819811677 tempest-ServersV294TestFqdnHostnames-819811677-project-member] Lock "c8be64a3-1bfa-41a3-b9be-68a7e6e052e0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 755.851644] env[65726]: DEBUG nova.network.neutron [-] [instance: bd839fe1-8801-4ba1-9c23-288453258b66] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 755.866491] env[65726]: DEBUG nova.compute.manager [None req-44e83fc7-afcc-4134-bbbb-6bc576931bb8 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: 895797e4-2941-44cd-aab1-67afa4fac02d] Start spawning the instance on the hypervisor. {{(pid=65726) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 755.894704] env[65726]: DEBUG nova.virt.hardware [None req-44e83fc7-afcc-4134-bbbb-6bc576931bb8 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T19:28:34Z,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1539f06fa3534e90acbc3a60c4b8bd3f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T19:28:35Z,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 755.894704] env[65726]: DEBUG nova.virt.hardware [None req-44e83fc7-afcc-4134-bbbb-6bc576931bb8 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 755.894704] env[65726]: DEBUG nova.virt.hardware [None req-44e83fc7-afcc-4134-bbbb-6bc576931bb8 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 755.895237] env[65726]: DEBUG nova.virt.hardware [None req-44e83fc7-afcc-4134-bbbb-6bc576931bb8 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 755.895546] env[65726]: DEBUG nova.virt.hardware [None req-44e83fc7-afcc-4134-bbbb-6bc576931bb8 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 755.895826] env[65726]: DEBUG nova.virt.hardware [None req-44e83fc7-afcc-4134-bbbb-6bc576931bb8 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 755.896252] env[65726]: DEBUG nova.virt.hardware [None req-44e83fc7-afcc-4134-bbbb-6bc576931bb8 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 755.896593] env[65726]: DEBUG nova.virt.hardware [None req-44e83fc7-afcc-4134-bbbb-6bc576931bb8 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 755.896884] env[65726]: DEBUG nova.virt.hardware [None req-44e83fc7-afcc-4134-bbbb-6bc576931bb8 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 755.897164] env[65726]: DEBUG nova.virt.hardware [None req-44e83fc7-afcc-4134-bbbb-6bc576931bb8 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 755.897448] env[65726]: DEBUG nova.virt.hardware [None req-44e83fc7-afcc-4134-bbbb-6bc576931bb8 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 755.898498] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb2df9a6-aba4-4d31-80c2-25fbbc6c0fa1 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.911069] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07770d32-e014-4090-93e0-0626b2a0a8e6 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.999814] env[65726]: DEBUG oslo_vmware.api [None req-b479f246-d94b-4634-888f-dee4586c3272 tempest-ServerAddressesNegativeTestJSON-963263679 tempest-ServerAddressesNegativeTestJSON-963263679-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5261937e-7fe9-0775-b53f-eee7481b66cf, 'name': SearchDatastore_Task, 'duration_secs': 0.009912} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 756.003365] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-88ecc6c2-b3dd-4ce2-83f3-31dba259405d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.009863] env[65726]: DEBUG oslo_vmware.api [None req-b479f246-d94b-4634-888f-dee4586c3272 tempest-ServerAddressesNegativeTestJSON-963263679 tempest-ServerAddressesNegativeTestJSON-963263679-project-member] Waiting for the task: (returnval){ [ 756.009863] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52f8e735-65fb-f729-13fe-513610f70be7" [ 756.009863] env[65726]: _type = "Task" [ 756.009863] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 756.022366] env[65726]: DEBUG oslo_vmware.api [None req-b479f246-d94b-4634-888f-dee4586c3272 tempest-ServerAddressesNegativeTestJSON-963263679 tempest-ServerAddressesNegativeTestJSON-963263679-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52f8e735-65fb-f729-13fe-513610f70be7, 'name': SearchDatastore_Task, 'duration_secs': 0.010352} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 756.022642] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b479f246-d94b-4634-888f-dee4586c3272 tempest-ServerAddressesNegativeTestJSON-963263679 tempest-ServerAddressesNegativeTestJSON-963263679-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 756.022867] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-b479f246-d94b-4634-888f-dee4586c3272 tempest-ServerAddressesNegativeTestJSON-963263679 tempest-ServerAddressesNegativeTestJSON-963263679-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] 00746508-d0d6-4dfa-9026-772398d004ee/00746508-d0d6-4dfa-9026-772398d004ee.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 756.023193] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0175d9b1-2a85-428e-9779-eeda10196291 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.033441] env[65726]: DEBUG oslo_vmware.api [None req-b479f246-d94b-4634-888f-dee4586c3272 tempest-ServerAddressesNegativeTestJSON-963263679 tempest-ServerAddressesNegativeTestJSON-963263679-project-member] Waiting for the task: (returnval){ [ 756.033441] env[65726]: value = "task-5115777" [ 756.033441] env[65726]: _type = "Task" [ 756.033441] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 756.042915] env[65726]: DEBUG oslo_vmware.api [None req-b479f246-d94b-4634-888f-dee4586c3272 tempest-ServerAddressesNegativeTestJSON-963263679 tempest-ServerAddressesNegativeTestJSON-963263679-project-member] Task: {'id': task-5115777, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 756.194066] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-816e5be8-3271-42c1-97a4-c1d280ba4b3e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.203963] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1964c67d-ca51-4479-8df5-8a70d3887336 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.235430] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c81578c-cda5-46b6-ba40-4f6be26dff51 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.244389] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bfe3e7bc-707c-430f-921c-5768630f79df {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.260351] env[65726]: DEBUG nova.compute.provider_tree [None req-af2f0c1d-9bc5-40c3-a712-d408e73bc806 tempest-ServerDiagnosticsNegativeTest-1635132145 tempest-ServerDiagnosticsNegativeTest-1635132145-project-member] Updating inventory in ProviderTree for provider 07c4692f-bdb4-4058-9173-ff9664830295 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 756.358251] env[65726]: INFO nova.compute.manager [-] [instance: bd839fe1-8801-4ba1-9c23-288453258b66] Took 1.63 seconds to deallocate network for instance. [ 756.496211] env[65726]: DEBUG nova.network.neutron [None req-44e83fc7-afcc-4134-bbbb-6bc576931bb8 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: 895797e4-2941-44cd-aab1-67afa4fac02d] Successfully updated port: a1dec2db-0e1e-45dd-8587-a84e4a5e2ef4 {{(pid=65726) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 756.547861] env[65726]: DEBUG oslo_vmware.api [None req-b479f246-d94b-4634-888f-dee4586c3272 tempest-ServerAddressesNegativeTestJSON-963263679 tempest-ServerAddressesNegativeTestJSON-963263679-project-member] Task: {'id': task-5115777, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 756.553823] env[65726]: WARNING openstack [None req-04afb47c-692a-46d9-a640-68cc2c90206f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 756.554329] env[65726]: WARNING openstack [None req-04afb47c-692a-46d9-a640-68cc2c90206f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 756.700021] env[65726]: DEBUG nova.network.neutron [None req-04afb47c-692a-46d9-a640-68cc2c90206f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: c108f2a5-031b-47ee-9a5b-d62c2f42c26b] Updating instance_info_cache with network_info: [{"id": "981741c6-69f0-42d4-87be-e1c8f62d0e17", "address": "fa:16:3e:ca:6f:8a", "network": {"id": "3f7916f7-ee9c-4629-b3b3-1367de9f98ea", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-579713617-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b794c99309e94c1181fc9d8eeeb84702", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d0bdd11b-58af-4cc0-9d38-8322e1bb4e74", "external-id": "nsx-vlan-transportzone-398", "segmentation_id": 398, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap981741c6-69", "ovs_interfaceid": "981741c6-69f0-42d4-87be-e1c8f62d0e17", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 756.795172] env[65726]: DEBUG nova.scheduler.client.report [None req-af2f0c1d-9bc5-40c3-a712-d408e73bc806 tempest-ServerDiagnosticsNegativeTest-1635132145 tempest-ServerDiagnosticsNegativeTest-1635132145-project-member] Updated inventory for provider 07c4692f-bdb4-4058-9173-ff9664830295 with generation 70 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 756.795374] env[65726]: DEBUG nova.compute.provider_tree [None req-af2f0c1d-9bc5-40c3-a712-d408e73bc806 tempest-ServerDiagnosticsNegativeTest-1635132145 tempest-ServerDiagnosticsNegativeTest-1635132145-project-member] Updating resource provider 07c4692f-bdb4-4058-9173-ff9664830295 generation from 70 to 71 during operation: update_inventory {{(pid=65726) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 756.795544] env[65726]: DEBUG nova.compute.provider_tree [None req-af2f0c1d-9bc5-40c3-a712-d408e73bc806 tempest-ServerDiagnosticsNegativeTest-1635132145 tempest-ServerDiagnosticsNegativeTest-1635132145-project-member] Updating inventory in ProviderTree for provider 07c4692f-bdb4-4058-9173-ff9664830295 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 756.867722] env[65726]: DEBUG oslo_concurrency.lockutils [None req-1640464d-fd11-4b5f-b0e0-f0de74fa607d tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 757.000364] env[65726]: DEBUG oslo_concurrency.lockutils [None req-44e83fc7-afcc-4134-bbbb-6bc576931bb8 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Acquiring lock "refresh_cache-895797e4-2941-44cd-aab1-67afa4fac02d" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 757.000728] env[65726]: DEBUG oslo_concurrency.lockutils [None req-44e83fc7-afcc-4134-bbbb-6bc576931bb8 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Acquired lock "refresh_cache-895797e4-2941-44cd-aab1-67afa4fac02d" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 757.001038] env[65726]: DEBUG nova.network.neutron [None req-44e83fc7-afcc-4134-bbbb-6bc576931bb8 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: 895797e4-2941-44cd-aab1-67afa4fac02d] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 757.051817] env[65726]: DEBUG oslo_vmware.api [None req-b479f246-d94b-4634-888f-dee4586c3272 tempest-ServerAddressesNegativeTestJSON-963263679 tempest-ServerAddressesNegativeTestJSON-963263679-project-member] Task: {'id': task-5115777, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.611308} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 757.051817] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-b479f246-d94b-4634-888f-dee4586c3272 tempest-ServerAddressesNegativeTestJSON-963263679 tempest-ServerAddressesNegativeTestJSON-963263679-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] 00746508-d0d6-4dfa-9026-772398d004ee/00746508-d0d6-4dfa-9026-772398d004ee.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 757.051817] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-b479f246-d94b-4634-888f-dee4586c3272 tempest-ServerAddressesNegativeTestJSON-963263679 tempest-ServerAddressesNegativeTestJSON-963263679-project-member] [instance: 00746508-d0d6-4dfa-9026-772398d004ee] Extending root virtual disk to 1048576 {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 757.051817] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0eeb82e8-a0b0-4aae-9dd8-c6a69d1ba59a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.061810] env[65726]: DEBUG oslo_vmware.api [None req-b479f246-d94b-4634-888f-dee4586c3272 tempest-ServerAddressesNegativeTestJSON-963263679 tempest-ServerAddressesNegativeTestJSON-963263679-project-member] Waiting for the task: (returnval){ [ 757.061810] env[65726]: value = "task-5115778" [ 757.061810] env[65726]: _type = "Task" [ 757.061810] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 757.072349] env[65726]: DEBUG oslo_vmware.api [None req-b479f246-d94b-4634-888f-dee4586c3272 tempest-ServerAddressesNegativeTestJSON-963263679 tempest-ServerAddressesNegativeTestJSON-963263679-project-member] Task: {'id': task-5115778, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 757.206020] env[65726]: DEBUG oslo_concurrency.lockutils [None req-04afb47c-692a-46d9-a640-68cc2c90206f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Releasing lock "refresh_cache-c108f2a5-031b-47ee-9a5b-d62c2f42c26b" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 757.206020] env[65726]: DEBUG nova.compute.manager [None req-04afb47c-692a-46d9-a640-68cc2c90206f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: c108f2a5-031b-47ee-9a5b-d62c2f42c26b] Instance network_info: |[{"id": "981741c6-69f0-42d4-87be-e1c8f62d0e17", "address": "fa:16:3e:ca:6f:8a", "network": {"id": "3f7916f7-ee9c-4629-b3b3-1367de9f98ea", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-579713617-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b794c99309e94c1181fc9d8eeeb84702", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d0bdd11b-58af-4cc0-9d38-8322e1bb4e74", "external-id": "nsx-vlan-transportzone-398", "segmentation_id": 398, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap981741c6-69", "ovs_interfaceid": "981741c6-69f0-42d4-87be-e1c8f62d0e17", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 757.206292] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-04afb47c-692a-46d9-a640-68cc2c90206f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: c108f2a5-031b-47ee-9a5b-d62c2f42c26b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ca:6f:8a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd0bdd11b-58af-4cc0-9d38-8322e1bb4e74', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '981741c6-69f0-42d4-87be-e1c8f62d0e17', 'vif_model': 'vmxnet3'}] {{(pid=65726) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 757.213394] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-04afb47c-692a-46d9-a640-68cc2c90206f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 757.213795] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c108f2a5-031b-47ee-9a5b-d62c2f42c26b] Creating VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 757.214518] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b9c9892d-4121-4dd8-aec5-75c3815a5558 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.241323] env[65726]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 757.241323] env[65726]: value = "task-5115779" [ 757.241323] env[65726]: _type = "Task" [ 757.241323] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 757.251768] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115779, 'name': CreateVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 757.302528] env[65726]: DEBUG oslo_concurrency.lockutils [None req-af2f0c1d-9bc5-40c3-a712-d408e73bc806 tempest-ServerDiagnosticsNegativeTest-1635132145 tempest-ServerDiagnosticsNegativeTest-1635132145-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.479s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 757.303208] env[65726]: DEBUG nova.compute.manager [None req-af2f0c1d-9bc5-40c3-a712-d408e73bc806 tempest-ServerDiagnosticsNegativeTest-1635132145 tempest-ServerDiagnosticsNegativeTest-1635132145-project-member] [instance: 0d6f14c5-5770-476f-a186-4cf1b2d4fbe8] Start building networks asynchronously for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 757.306321] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b1ce1b66-ca58-46d0-b624-7c893de1adee tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 34.389s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 757.306481] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b1ce1b66-ca58-46d0-b624-7c893de1adee tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 757.308614] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5acbf047-f1f8-4297-909c-a3d7a9b0fdd6 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 33.738s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 757.308814] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5acbf047-f1f8-4297-909c-a3d7a9b0fdd6 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 757.311220] env[65726]: DEBUG oslo_concurrency.lockutils [None req-48153719-e51d-407a-98ef-93401c6270f2 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 30.243s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 757.314619] env[65726]: INFO nova.compute.claims [None req-48153719-e51d-407a-98ef-93401c6270f2 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] [instance: 65aaa2eb-3ca5-4c2b-92f9-b524366c10dd] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 757.349264] env[65726]: INFO nova.scheduler.client.report [None req-b1ce1b66-ca58-46d0-b624-7c893de1adee tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Deleted allocations for instance 423af936-a383-4652-8887-9c8b8f507909 [ 757.352487] env[65726]: INFO nova.scheduler.client.report [None req-5acbf047-f1f8-4297-909c-a3d7a9b0fdd6 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Deleted allocations for instance 815f056b-2d88-40dd-9ef7-a891ec8d92cf [ 757.516069] env[65726]: WARNING openstack [None req-44e83fc7-afcc-4134-bbbb-6bc576931bb8 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 757.516614] env[65726]: WARNING openstack [None req-44e83fc7-afcc-4134-bbbb-6bc576931bb8 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 757.522998] env[65726]: DEBUG nova.network.neutron [None req-44e83fc7-afcc-4134-bbbb-6bc576931bb8 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: 895797e4-2941-44cd-aab1-67afa4fac02d] Instance cache missing network info. {{(pid=65726) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 757.524711] env[65726]: DEBUG oslo_concurrency.lockutils [None req-d63650ad-6689-479f-81d4-98dcfe6418a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Acquiring lock "cd539d8b-921b-4947-954c-8573380dbdc8" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 757.525083] env[65726]: DEBUG oslo_concurrency.lockutils [None req-d63650ad-6689-479f-81d4-98dcfe6418a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Lock "cd539d8b-921b-4947-954c-8573380dbdc8" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 757.525720] env[65726]: DEBUG oslo_concurrency.lockutils [None req-d63650ad-6689-479f-81d4-98dcfe6418a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Acquiring lock "cd539d8b-921b-4947-954c-8573380dbdc8-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 757.526071] env[65726]: DEBUG oslo_concurrency.lockutils [None req-d63650ad-6689-479f-81d4-98dcfe6418a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Lock "cd539d8b-921b-4947-954c-8573380dbdc8-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 757.526341] env[65726]: DEBUG oslo_concurrency.lockutils [None req-d63650ad-6689-479f-81d4-98dcfe6418a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Lock "cd539d8b-921b-4947-954c-8573380dbdc8-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 757.528612] env[65726]: INFO nova.compute.manager [None req-d63650ad-6689-479f-81d4-98dcfe6418a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] [instance: cd539d8b-921b-4947-954c-8573380dbdc8] Terminating instance [ 757.556189] env[65726]: DEBUG nova.compute.manager [req-207d272f-0a18-4cd6-8840-84c1a6367cbd req-077126f5-a106-4772-9d7e-e3ed8f50e430 service nova] [instance: 00746508-d0d6-4dfa-9026-772398d004ee] Received event network-changed-1151551e-12f0-4e85-a385-f1ebf4e9bc72 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 757.556189] env[65726]: DEBUG nova.compute.manager [req-207d272f-0a18-4cd6-8840-84c1a6367cbd req-077126f5-a106-4772-9d7e-e3ed8f50e430 service nova] [instance: 00746508-d0d6-4dfa-9026-772398d004ee] Refreshing instance network info cache due to event network-changed-1151551e-12f0-4e85-a385-f1ebf4e9bc72. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 757.556189] env[65726]: DEBUG oslo_concurrency.lockutils [req-207d272f-0a18-4cd6-8840-84c1a6367cbd req-077126f5-a106-4772-9d7e-e3ed8f50e430 service nova] Acquiring lock "refresh_cache-00746508-d0d6-4dfa-9026-772398d004ee" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 757.556189] env[65726]: DEBUG oslo_concurrency.lockutils [req-207d272f-0a18-4cd6-8840-84c1a6367cbd req-077126f5-a106-4772-9d7e-e3ed8f50e430 service nova] Acquired lock "refresh_cache-00746508-d0d6-4dfa-9026-772398d004ee" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 757.556501] env[65726]: DEBUG nova.network.neutron [req-207d272f-0a18-4cd6-8840-84c1a6367cbd req-077126f5-a106-4772-9d7e-e3ed8f50e430 service nova] [instance: 00746508-d0d6-4dfa-9026-772398d004ee] Refreshing network info cache for port 1151551e-12f0-4e85-a385-f1ebf4e9bc72 {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 757.575060] env[65726]: DEBUG oslo_vmware.api [None req-b479f246-d94b-4634-888f-dee4586c3272 tempest-ServerAddressesNegativeTestJSON-963263679 tempest-ServerAddressesNegativeTestJSON-963263679-project-member] Task: {'id': task-5115778, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.112406} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 757.575513] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-b479f246-d94b-4634-888f-dee4586c3272 tempest-ServerAddressesNegativeTestJSON-963263679 tempest-ServerAddressesNegativeTestJSON-963263679-project-member] [instance: 00746508-d0d6-4dfa-9026-772398d004ee] Extended root virtual disk {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 757.576398] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c80d2125-73f8-4b69-8b57-26571c396e7c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.581413] env[65726]: WARNING openstack [None req-44e83fc7-afcc-4134-bbbb-6bc576931bb8 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 757.582254] env[65726]: WARNING openstack [None req-44e83fc7-afcc-4134-bbbb-6bc576931bb8 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 757.616314] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-b479f246-d94b-4634-888f-dee4586c3272 tempest-ServerAddressesNegativeTestJSON-963263679 tempest-ServerAddressesNegativeTestJSON-963263679-project-member] [instance: 00746508-d0d6-4dfa-9026-772398d004ee] Reconfiguring VM instance instance-00000026 to attach disk [datastore1] 00746508-d0d6-4dfa-9026-772398d004ee/00746508-d0d6-4dfa-9026-772398d004ee.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 757.617410] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c9cc880b-42e7-4875-9b8e-55f0109d1664 tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Acquiring lock "ca67d1a4-d304-4d06-a436-f257b60ca080" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 757.617813] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c9cc880b-42e7-4875-9b8e-55f0109d1664 tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Lock "ca67d1a4-d304-4d06-a436-f257b60ca080" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 757.618159] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c9cc880b-42e7-4875-9b8e-55f0109d1664 tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Acquiring lock "ca67d1a4-d304-4d06-a436-f257b60ca080-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 757.618442] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c9cc880b-42e7-4875-9b8e-55f0109d1664 tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Lock "ca67d1a4-d304-4d06-a436-f257b60ca080-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 757.618748] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c9cc880b-42e7-4875-9b8e-55f0109d1664 tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Lock "ca67d1a4-d304-4d06-a436-f257b60ca080-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 757.620739] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-14185496-a4f7-43f9-9319-214da23582f4 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.641609] env[65726]: INFO nova.compute.manager [None req-c9cc880b-42e7-4875-9b8e-55f0109d1664 tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] [instance: ca67d1a4-d304-4d06-a436-f257b60ca080] Terminating instance [ 757.650552] env[65726]: DEBUG oslo_vmware.api [None req-b479f246-d94b-4634-888f-dee4586c3272 tempest-ServerAddressesNegativeTestJSON-963263679 tempest-ServerAddressesNegativeTestJSON-963263679-project-member] Waiting for the task: (returnval){ [ 757.650552] env[65726]: value = "task-5115780" [ 757.650552] env[65726]: _type = "Task" [ 757.650552] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 757.661790] env[65726]: DEBUG oslo_vmware.api [None req-b479f246-d94b-4634-888f-dee4586c3272 tempest-ServerAddressesNegativeTestJSON-963263679 tempest-ServerAddressesNegativeTestJSON-963263679-project-member] Task: {'id': task-5115780, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 757.677934] env[65726]: WARNING openstack [None req-44e83fc7-afcc-4134-bbbb-6bc576931bb8 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 757.679403] env[65726]: WARNING openstack [None req-44e83fc7-afcc-4134-bbbb-6bc576931bb8 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 757.752447] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115779, 'name': CreateVM_Task} progress is 25%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 757.776241] env[65726]: DEBUG nova.network.neutron [None req-44e83fc7-afcc-4134-bbbb-6bc576931bb8 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: 895797e4-2941-44cd-aab1-67afa4fac02d] Updating instance_info_cache with network_info: [{"id": "a1dec2db-0e1e-45dd-8587-a84e4a5e2ef4", "address": "fa:16:3e:b9:47:02", "network": {"id": "ad1373b5-e0ed-4ee6-a97c-af931afbab35", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-909418417-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a090d553766847e58e3231b966c92565", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5f60c972-a72d-4c5f-a250-faadfd6eafbe", "external-id": "nsx-vlan-transportzone-932", "segmentation_id": 932, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa1dec2db-0e", "ovs_interfaceid": "a1dec2db-0e1e-45dd-8587-a84e4a5e2ef4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 757.821039] env[65726]: DEBUG nova.compute.utils [None req-af2f0c1d-9bc5-40c3-a712-d408e73bc806 tempest-ServerDiagnosticsNegativeTest-1635132145 tempest-ServerDiagnosticsNegativeTest-1635132145-project-member] Using /dev/sd instead of None {{(pid=65726) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 757.825851] env[65726]: DEBUG nova.compute.manager [None req-af2f0c1d-9bc5-40c3-a712-d408e73bc806 tempest-ServerDiagnosticsNegativeTest-1635132145 tempest-ServerDiagnosticsNegativeTest-1635132145-project-member] [instance: 0d6f14c5-5770-476f-a186-4cf1b2d4fbe8] Allocating IP information in the background. {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 757.826111] env[65726]: DEBUG nova.network.neutron [None req-af2f0c1d-9bc5-40c3-a712-d408e73bc806 tempest-ServerDiagnosticsNegativeTest-1635132145 tempest-ServerDiagnosticsNegativeTest-1635132145-project-member] [instance: 0d6f14c5-5770-476f-a186-4cf1b2d4fbe8] allocate_for_instance() {{(pid=65726) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 757.826586] env[65726]: WARNING neutronclient.v2_0.client [None req-af2f0c1d-9bc5-40c3-a712-d408e73bc806 tempest-ServerDiagnosticsNegativeTest-1635132145 tempest-ServerDiagnosticsNegativeTest-1635132145-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 757.827049] env[65726]: WARNING neutronclient.v2_0.client [None req-af2f0c1d-9bc5-40c3-a712-d408e73bc806 tempest-ServerDiagnosticsNegativeTest-1635132145 tempest-ServerDiagnosticsNegativeTest-1635132145-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 757.827831] env[65726]: WARNING openstack [None req-af2f0c1d-9bc5-40c3-a712-d408e73bc806 tempest-ServerDiagnosticsNegativeTest-1635132145 tempest-ServerDiagnosticsNegativeTest-1635132145-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 757.828355] env[65726]: WARNING openstack [None req-af2f0c1d-9bc5-40c3-a712-d408e73bc806 tempest-ServerDiagnosticsNegativeTest-1635132145 tempest-ServerDiagnosticsNegativeTest-1635132145-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 757.866277] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b1ce1b66-ca58-46d0-b624-7c893de1adee tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Lock "423af936-a383-4652-8887-9c8b8f507909" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 36.973s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 757.867902] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5acbf047-f1f8-4297-909c-a3d7a9b0fdd6 tempest-DeleteServersAdminTestJSON-1600913385 tempest-DeleteServersAdminTestJSON-1600913385-project-member] Lock "815f056b-2d88-40dd-9ef7-a891ec8d92cf" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 37.745s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 757.909745] env[65726]: DEBUG nova.policy [None req-af2f0c1d-9bc5-40c3-a712-d408e73bc806 tempest-ServerDiagnosticsNegativeTest-1635132145 tempest-ServerDiagnosticsNegativeTest-1635132145-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f035cafb2b2c4cf4a778edc716ee2349', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '877bbe4aeef145749b4156f8ab6f2202', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65726) authorize /opt/stack/nova/nova/policy.py:192}} [ 757.951668] env[65726]: DEBUG nova.compute.manager [req-6967800b-6b88-4dd5-bbee-b376f3cbd51a req-5ed96645-2638-4dc7-9e67-3b20c2608d40 service nova] [instance: bd839fe1-8801-4ba1-9c23-288453258b66] Received event network-vif-deleted-32d9913d-f98b-4e80-af5a-5f88083281b3 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 758.033832] env[65726]: DEBUG nova.compute.manager [None req-d63650ad-6689-479f-81d4-98dcfe6418a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] [instance: cd539d8b-921b-4947-954c-8573380dbdc8] Start destroying the instance on the hypervisor. {{(pid=65726) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 758.035469] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-d63650ad-6689-479f-81d4-98dcfe6418a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] [instance: cd539d8b-921b-4947-954c-8573380dbdc8] Destroying instance {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 758.035469] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df824cff-c10e-4612-9c0e-2d74b26d1542 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.044264] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-d63650ad-6689-479f-81d4-98dcfe6418a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] [instance: cd539d8b-921b-4947-954c-8573380dbdc8] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 758.044507] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-bed6983b-2e6b-4f38-9e9a-956db3cb69b2 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.054251] env[65726]: DEBUG oslo_vmware.api [None req-d63650ad-6689-479f-81d4-98dcfe6418a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Waiting for the task: (returnval){ [ 758.054251] env[65726]: value = "task-5115781" [ 758.054251] env[65726]: _type = "Task" [ 758.054251] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 758.063146] env[65726]: DEBUG oslo_vmware.api [None req-d63650ad-6689-479f-81d4-98dcfe6418a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Task: {'id': task-5115781, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 758.065398] env[65726]: WARNING openstack [req-207d272f-0a18-4cd6-8840-84c1a6367cbd req-077126f5-a106-4772-9d7e-e3ed8f50e430 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 758.065519] env[65726]: WARNING openstack [req-207d272f-0a18-4cd6-8840-84c1a6367cbd req-077126f5-a106-4772-9d7e-e3ed8f50e430 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 758.150035] env[65726]: DEBUG nova.compute.manager [None req-c9cc880b-42e7-4875-9b8e-55f0109d1664 tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] [instance: ca67d1a4-d304-4d06-a436-f257b60ca080] Start destroying the instance on the hypervisor. {{(pid=65726) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 758.150035] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-c9cc880b-42e7-4875-9b8e-55f0109d1664 tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] [instance: ca67d1a4-d304-4d06-a436-f257b60ca080] Destroying instance {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 758.150035] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e5c945e-356e-42f7-98c4-8bc017cd1264 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.163328] env[65726]: DEBUG oslo_vmware.api [None req-b479f246-d94b-4634-888f-dee4586c3272 tempest-ServerAddressesNegativeTestJSON-963263679 tempest-ServerAddressesNegativeTestJSON-963263679-project-member] Task: {'id': task-5115780, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 758.166764] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-c9cc880b-42e7-4875-9b8e-55f0109d1664 tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] [instance: ca67d1a4-d304-4d06-a436-f257b60ca080] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 758.167201] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0fd3d0eb-dd6f-43e3-b13d-a5bf9f344e5d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.177294] env[65726]: DEBUG oslo_vmware.api [None req-c9cc880b-42e7-4875-9b8e-55f0109d1664 tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Waiting for the task: (returnval){ [ 758.177294] env[65726]: value = "task-5115782" [ 758.177294] env[65726]: _type = "Task" [ 758.177294] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 758.188151] env[65726]: DEBUG oslo_vmware.api [None req-c9cc880b-42e7-4875-9b8e-55f0109d1664 tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Task: {'id': task-5115782, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 758.253855] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115779, 'name': CreateVM_Task, 'duration_secs': 0.785626} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 758.254147] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c108f2a5-031b-47ee-9a5b-d62c2f42c26b] Created VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 758.254952] env[65726]: WARNING openstack [None req-04afb47c-692a-46d9-a640-68cc2c90206f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 758.255350] env[65726]: WARNING openstack [None req-04afb47c-692a-46d9-a640-68cc2c90206f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 758.261262] env[65726]: DEBUG oslo_concurrency.lockutils [None req-04afb47c-692a-46d9-a640-68cc2c90206f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 758.261578] env[65726]: DEBUG oslo_concurrency.lockutils [None req-04afb47c-692a-46d9-a640-68cc2c90206f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 758.262089] env[65726]: DEBUG oslo_concurrency.lockutils [None req-04afb47c-692a-46d9-a640-68cc2c90206f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 758.262468] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-93aa9da0-0d21-49dc-a52b-fcf75625e7ed {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.269469] env[65726]: DEBUG oslo_vmware.api [None req-04afb47c-692a-46d9-a640-68cc2c90206f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Waiting for the task: (returnval){ [ 758.269469] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52bb8daa-7532-3cd7-de2d-0f0f28b888e5" [ 758.269469] env[65726]: _type = "Task" [ 758.269469] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 758.282520] env[65726]: DEBUG oslo_concurrency.lockutils [None req-44e83fc7-afcc-4134-bbbb-6bc576931bb8 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Releasing lock "refresh_cache-895797e4-2941-44cd-aab1-67afa4fac02d" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 758.282885] env[65726]: DEBUG nova.compute.manager [None req-44e83fc7-afcc-4134-bbbb-6bc576931bb8 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: 895797e4-2941-44cd-aab1-67afa4fac02d] Instance network_info: |[{"id": "a1dec2db-0e1e-45dd-8587-a84e4a5e2ef4", "address": "fa:16:3e:b9:47:02", "network": {"id": "ad1373b5-e0ed-4ee6-a97c-af931afbab35", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-909418417-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a090d553766847e58e3231b966c92565", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5f60c972-a72d-4c5f-a250-faadfd6eafbe", "external-id": "nsx-vlan-transportzone-932", "segmentation_id": 932, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa1dec2db-0e", "ovs_interfaceid": "a1dec2db-0e1e-45dd-8587-a84e4a5e2ef4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 758.283247] env[65726]: DEBUG oslo_vmware.api [None req-04afb47c-692a-46d9-a640-68cc2c90206f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52bb8daa-7532-3cd7-de2d-0f0f28b888e5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 758.283634] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-44e83fc7-afcc-4134-bbbb-6bc576931bb8 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: 895797e4-2941-44cd-aab1-67afa4fac02d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b9:47:02', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '5f60c972-a72d-4c5f-a250-faadfd6eafbe', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a1dec2db-0e1e-45dd-8587-a84e4a5e2ef4', 'vif_model': 'vmxnet3'}] {{(pid=65726) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 758.292672] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-44e83fc7-afcc-4134-bbbb-6bc576931bb8 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Creating folder: Project (a090d553766847e58e3231b966c92565). Parent ref: group-v995008. {{(pid=65726) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 758.293156] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-44c46825-fb54-477f-8e9c-275da9d955b8 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.308931] env[65726]: INFO nova.virt.vmwareapi.vm_util [None req-44e83fc7-afcc-4134-bbbb-6bc576931bb8 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Created folder: Project (a090d553766847e58e3231b966c92565) in parent group-v995008. [ 758.309322] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-44e83fc7-afcc-4134-bbbb-6bc576931bb8 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Creating folder: Instances. Parent ref: group-v995113. {{(pid=65726) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 758.309484] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-88b5e671-8004-47c9-8e00-af6c3d4342b8 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.323533] env[65726]: INFO nova.virt.vmwareapi.vm_util [None req-44e83fc7-afcc-4134-bbbb-6bc576931bb8 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Created folder: Instances in parent group-v995113. [ 758.323533] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-44e83fc7-afcc-4134-bbbb-6bc576931bb8 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 758.323533] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 895797e4-2941-44cd-aab1-67afa4fac02d] Creating VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 758.323533] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b42c5207-d7db-45a3-870d-66a92f2ae622 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.339297] env[65726]: DEBUG nova.compute.manager [None req-af2f0c1d-9bc5-40c3-a712-d408e73bc806 tempest-ServerDiagnosticsNegativeTest-1635132145 tempest-ServerDiagnosticsNegativeTest-1635132145-project-member] [instance: 0d6f14c5-5770-476f-a186-4cf1b2d4fbe8] Start building block device mappings for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 758.345037] env[65726]: DEBUG nova.network.neutron [None req-af2f0c1d-9bc5-40c3-a712-d408e73bc806 tempest-ServerDiagnosticsNegativeTest-1635132145 tempest-ServerDiagnosticsNegativeTest-1635132145-project-member] [instance: 0d6f14c5-5770-476f-a186-4cf1b2d4fbe8] Successfully created port: c8363d25-c607-4d04-ab53-412848d7cf4f {{(pid=65726) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 758.357183] env[65726]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 758.357183] env[65726]: value = "task-5115785" [ 758.357183] env[65726]: _type = "Task" [ 758.357183] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 758.366960] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115785, 'name': CreateVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 758.436480] env[65726]: WARNING openstack [req-207d272f-0a18-4cd6-8840-84c1a6367cbd req-077126f5-a106-4772-9d7e-e3ed8f50e430 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 758.436891] env[65726]: WARNING openstack [req-207d272f-0a18-4cd6-8840-84c1a6367cbd req-077126f5-a106-4772-9d7e-e3ed8f50e430 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 758.531842] env[65726]: DEBUG oslo_concurrency.lockutils [None req-645c8c71-cd4d-4583-8270-b88bb6120d10 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Acquiring lock "01d86089-6b9b-4588-864e-ef91375a1eea" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 758.532263] env[65726]: DEBUG oslo_concurrency.lockutils [None req-645c8c71-cd4d-4583-8270-b88bb6120d10 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Lock "01d86089-6b9b-4588-864e-ef91375a1eea" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 758.567079] env[65726]: DEBUG oslo_vmware.api [None req-d63650ad-6689-479f-81d4-98dcfe6418a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Task: {'id': task-5115781, 'name': PowerOffVM_Task, 'duration_secs': 0.318907} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 758.572218] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-d63650ad-6689-479f-81d4-98dcfe6418a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] [instance: cd539d8b-921b-4947-954c-8573380dbdc8] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 758.572397] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-d63650ad-6689-479f-81d4-98dcfe6418a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] [instance: cd539d8b-921b-4947-954c-8573380dbdc8] Unregistering the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 758.572903] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6fb89ae6-0579-4c20-a0f0-3c691fed8fb6 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.645495] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-d63650ad-6689-479f-81d4-98dcfe6418a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] [instance: cd539d8b-921b-4947-954c-8573380dbdc8] Unregistered the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 758.645803] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-d63650ad-6689-479f-81d4-98dcfe6418a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] [instance: cd539d8b-921b-4947-954c-8573380dbdc8] Deleting contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 758.646026] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-d63650ad-6689-479f-81d4-98dcfe6418a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Deleting the datastore file [datastore1] cd539d8b-921b-4947-954c-8573380dbdc8 {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 758.646425] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-892c1f62-d69d-4a89-b73f-ab0c9402585f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.660462] env[65726]: DEBUG oslo_vmware.api [None req-d63650ad-6689-479f-81d4-98dcfe6418a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Waiting for the task: (returnval){ [ 758.660462] env[65726]: value = "task-5115787" [ 758.660462] env[65726]: _type = "Task" [ 758.660462] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 758.670039] env[65726]: DEBUG oslo_vmware.api [None req-b479f246-d94b-4634-888f-dee4586c3272 tempest-ServerAddressesNegativeTestJSON-963263679 tempest-ServerAddressesNegativeTestJSON-963263679-project-member] Task: {'id': task-5115780, 'name': ReconfigVM_Task, 'duration_secs': 0.860923} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 758.670574] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-b479f246-d94b-4634-888f-dee4586c3272 tempest-ServerAddressesNegativeTestJSON-963263679 tempest-ServerAddressesNegativeTestJSON-963263679-project-member] [instance: 00746508-d0d6-4dfa-9026-772398d004ee] Reconfigured VM instance instance-00000026 to attach disk [datastore1] 00746508-d0d6-4dfa-9026-772398d004ee/00746508-d0d6-4dfa-9026-772398d004ee.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 758.671327] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0690275b-c778-42d3-b068-4fb0d391022b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.679521] env[65726]: DEBUG oslo_vmware.api [None req-d63650ad-6689-479f-81d4-98dcfe6418a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Task: {'id': task-5115787, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 758.686526] env[65726]: DEBUG oslo_vmware.api [None req-b479f246-d94b-4634-888f-dee4586c3272 tempest-ServerAddressesNegativeTestJSON-963263679 tempest-ServerAddressesNegativeTestJSON-963263679-project-member] Waiting for the task: (returnval){ [ 758.686526] env[65726]: value = "task-5115788" [ 758.686526] env[65726]: _type = "Task" [ 758.686526] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 758.693890] env[65726]: DEBUG oslo_vmware.api [None req-c9cc880b-42e7-4875-9b8e-55f0109d1664 tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Task: {'id': task-5115782, 'name': PowerOffVM_Task, 'duration_secs': 0.272928} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 758.697635] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-c9cc880b-42e7-4875-9b8e-55f0109d1664 tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] [instance: ca67d1a4-d304-4d06-a436-f257b60ca080] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 758.697926] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-c9cc880b-42e7-4875-9b8e-55f0109d1664 tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] [instance: ca67d1a4-d304-4d06-a436-f257b60ca080] Unregistering the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 758.698602] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e4b2e07d-d06a-40ac-858b-49ef6403923f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.704626] env[65726]: DEBUG oslo_vmware.api [None req-b479f246-d94b-4634-888f-dee4586c3272 tempest-ServerAddressesNegativeTestJSON-963263679 tempest-ServerAddressesNegativeTestJSON-963263679-project-member] Task: {'id': task-5115788, 'name': Rename_Task} progress is 10%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 758.781379] env[65726]: DEBUG oslo_vmware.api [None req-04afb47c-692a-46d9-a640-68cc2c90206f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52bb8daa-7532-3cd7-de2d-0f0f28b888e5, 'name': SearchDatastore_Task, 'duration_secs': 0.01305} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 758.784778] env[65726]: DEBUG oslo_concurrency.lockutils [None req-04afb47c-692a-46d9-a640-68cc2c90206f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 758.785089] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-04afb47c-692a-46d9-a640-68cc2c90206f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: c108f2a5-031b-47ee-9a5b-d62c2f42c26b] Processing image b52362a3-ee8a-4cbf-b06f-513b0cc8f95c {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 758.785354] env[65726]: DEBUG oslo_concurrency.lockutils [None req-04afb47c-692a-46d9-a640-68cc2c90206f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 758.785495] env[65726]: DEBUG oslo_concurrency.lockutils [None req-04afb47c-692a-46d9-a640-68cc2c90206f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 758.785689] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-04afb47c-692a-46d9-a640-68cc2c90206f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 758.786303] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0ea1c0cd-e75e-4979-b148-f51355477997 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.797369] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-c9cc880b-42e7-4875-9b8e-55f0109d1664 tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] [instance: ca67d1a4-d304-4d06-a436-f257b60ca080] Unregistered the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 758.797812] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-c9cc880b-42e7-4875-9b8e-55f0109d1664 tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] [instance: ca67d1a4-d304-4d06-a436-f257b60ca080] Deleting contents of the VM from datastore datastore2 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 758.798064] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-c9cc880b-42e7-4875-9b8e-55f0109d1664 tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Deleting the datastore file [datastore2] ca67d1a4-d304-4d06-a436-f257b60ca080 {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 758.802710] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e6fb7b9c-402c-476e-bef4-b5684aca520f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.807679] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-04afb47c-692a-46d9-a640-68cc2c90206f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 758.807797] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-04afb47c-692a-46d9-a640-68cc2c90206f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65726) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 758.809158] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ecc2cf20-f34a-465f-b4e9-75f432ea8505 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.821073] env[65726]: DEBUG oslo_vmware.api [None req-c9cc880b-42e7-4875-9b8e-55f0109d1664 tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Waiting for the task: (returnval){ [ 758.821073] env[65726]: value = "task-5115790" [ 758.821073] env[65726]: _type = "Task" [ 758.821073] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 758.821073] env[65726]: DEBUG oslo_vmware.api [None req-04afb47c-692a-46d9-a640-68cc2c90206f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Waiting for the task: (returnval){ [ 758.821073] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]520b6db1-b640-d411-e76a-cbb489ed2cc7" [ 758.821073] env[65726]: _type = "Task" [ 758.821073] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 758.843042] env[65726]: DEBUG oslo_vmware.api [None req-c9cc880b-42e7-4875-9b8e-55f0109d1664 tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Task: {'id': task-5115790, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 758.844244] env[65726]: DEBUG oslo_vmware.api [None req-04afb47c-692a-46d9-a640-68cc2c90206f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]520b6db1-b640-d411-e76a-cbb489ed2cc7, 'name': SearchDatastore_Task, 'duration_secs': 0.017378} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 758.848574] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c25df348-950a-42c3-9ed7-4693f24706cb {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.858689] env[65726]: DEBUG oslo_vmware.api [None req-04afb47c-692a-46d9-a640-68cc2c90206f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Waiting for the task: (returnval){ [ 758.858689] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52dc07df-1cf9-ecea-33ab-230a24a33d1e" [ 758.858689] env[65726]: _type = "Task" [ 758.858689] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 758.878464] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115785, 'name': CreateVM_Task, 'duration_secs': 0.508503} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 758.882263] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 895797e4-2941-44cd-aab1-67afa4fac02d] Created VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 758.882587] env[65726]: DEBUG oslo_vmware.api [None req-04afb47c-692a-46d9-a640-68cc2c90206f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52dc07df-1cf9-ecea-33ab-230a24a33d1e, 'name': SearchDatastore_Task, 'duration_secs': 0.012117} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 758.886423] env[65726]: WARNING openstack [None req-44e83fc7-afcc-4134-bbbb-6bc576931bb8 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 758.886952] env[65726]: WARNING openstack [None req-44e83fc7-afcc-4134-bbbb-6bc576931bb8 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 758.896025] env[65726]: DEBUG oslo_concurrency.lockutils [None req-44e83fc7-afcc-4134-bbbb-6bc576931bb8 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 758.896025] env[65726]: DEBUG oslo_concurrency.lockutils [None req-44e83fc7-afcc-4134-bbbb-6bc576931bb8 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 758.896025] env[65726]: DEBUG oslo_concurrency.lockutils [None req-44e83fc7-afcc-4134-bbbb-6bc576931bb8 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 758.896025] env[65726]: DEBUG oslo_concurrency.lockutils [None req-04afb47c-692a-46d9-a640-68cc2c90206f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 758.896569] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-04afb47c-692a-46d9-a640-68cc2c90206f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] c108f2a5-031b-47ee-9a5b-d62c2f42c26b/c108f2a5-031b-47ee-9a5b-d62c2f42c26b.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 758.896569] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3154f7b3-9e4b-432c-ab1a-4893ab996f0f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.897373] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0f44ab79-4677-4b4a-8134-37b61ffc0fb1 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.905759] env[65726]: DEBUG oslo_vmware.api [None req-44e83fc7-afcc-4134-bbbb-6bc576931bb8 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Waiting for the task: (returnval){ [ 758.905759] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52a79690-4550-acdd-9566-c6f759a0f185" [ 758.905759] env[65726]: _type = "Task" [ 758.905759] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 758.907980] env[65726]: DEBUG oslo_vmware.api [None req-04afb47c-692a-46d9-a640-68cc2c90206f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Waiting for the task: (returnval){ [ 758.907980] env[65726]: value = "task-5115791" [ 758.907980] env[65726]: _type = "Task" [ 758.907980] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 758.927690] env[65726]: DEBUG oslo_vmware.api [None req-44e83fc7-afcc-4134-bbbb-6bc576931bb8 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52a79690-4550-acdd-9566-c6f759a0f185, 'name': SearchDatastore_Task, 'duration_secs': 0.011302} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 758.934947] env[65726]: DEBUG oslo_concurrency.lockutils [None req-44e83fc7-afcc-4134-bbbb-6bc576931bb8 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 758.935411] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-44e83fc7-afcc-4134-bbbb-6bc576931bb8 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: 895797e4-2941-44cd-aab1-67afa4fac02d] Processing image b52362a3-ee8a-4cbf-b06f-513b0cc8f95c {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 758.935814] env[65726]: DEBUG oslo_concurrency.lockutils [None req-44e83fc7-afcc-4134-bbbb-6bc576931bb8 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 758.936089] env[65726]: DEBUG oslo_concurrency.lockutils [None req-44e83fc7-afcc-4134-bbbb-6bc576931bb8 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 758.936404] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-44e83fc7-afcc-4134-bbbb-6bc576931bb8 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 758.936852] env[65726]: DEBUG oslo_vmware.api [None req-04afb47c-692a-46d9-a640-68cc2c90206f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Task: {'id': task-5115791, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 758.937491] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c253eb11-af4d-45e2-846e-aa0e76986701 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.948445] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-44e83fc7-afcc-4134-bbbb-6bc576931bb8 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 758.948445] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-44e83fc7-afcc-4134-bbbb-6bc576931bb8 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65726) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 758.952318] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e15059b3-c7af-431d-8af5-fe6877bfae9f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.960358] env[65726]: DEBUG oslo_vmware.api [None req-44e83fc7-afcc-4134-bbbb-6bc576931bb8 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Waiting for the task: (returnval){ [ 758.960358] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52f98f3f-f5b6-5a71-f46f-7f1c7baa5d25" [ 758.960358] env[65726]: _type = "Task" [ 758.960358] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 758.970015] env[65726]: DEBUG oslo_vmware.api [None req-44e83fc7-afcc-4134-bbbb-6bc576931bb8 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52f98f3f-f5b6-5a71-f46f-7f1c7baa5d25, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 759.015391] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0822ef3b-a91a-4b7f-9191-a76a2fa42fbd {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.024216] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82571bb6-0a61-443b-a7cc-85571b3df294 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.056321] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f6ca132-de7a-4a85-abfd-f063341bdbe9 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.065636] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a43f924c-3035-4824-aaf2-08179c0ed0b0 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.081112] env[65726]: DEBUG nova.compute.provider_tree [None req-48153719-e51d-407a-98ef-93401c6270f2 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 759.180131] env[65726]: DEBUG oslo_vmware.api [None req-d63650ad-6689-479f-81d4-98dcfe6418a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Task: {'id': task-5115787, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.180428} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 759.180695] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-d63650ad-6689-479f-81d4-98dcfe6418a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Deleted the datastore file {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 759.181033] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-d63650ad-6689-479f-81d4-98dcfe6418a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] [instance: cd539d8b-921b-4947-954c-8573380dbdc8] Deleted contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 759.181311] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-d63650ad-6689-479f-81d4-98dcfe6418a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] [instance: cd539d8b-921b-4947-954c-8573380dbdc8] Instance destroyed {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 759.181606] env[65726]: INFO nova.compute.manager [None req-d63650ad-6689-479f-81d4-98dcfe6418a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] [instance: cd539d8b-921b-4947-954c-8573380dbdc8] Took 1.15 seconds to destroy the instance on the hypervisor. [ 759.182103] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-d63650ad-6689-479f-81d4-98dcfe6418a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 759.182453] env[65726]: DEBUG nova.compute.manager [-] [instance: cd539d8b-921b-4947-954c-8573380dbdc8] Deallocating network for instance {{(pid=65726) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 759.182611] env[65726]: DEBUG nova.network.neutron [-] [instance: cd539d8b-921b-4947-954c-8573380dbdc8] deallocate_for_instance() {{(pid=65726) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 759.183460] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 759.183875] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 759.209755] env[65726]: DEBUG oslo_vmware.api [None req-b479f246-d94b-4634-888f-dee4586c3272 tempest-ServerAddressesNegativeTestJSON-963263679 tempest-ServerAddressesNegativeTestJSON-963263679-project-member] Task: {'id': task-5115788, 'name': Rename_Task, 'duration_secs': 0.186521} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 759.210456] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-b479f246-d94b-4634-888f-dee4586c3272 tempest-ServerAddressesNegativeTestJSON-963263679 tempest-ServerAddressesNegativeTestJSON-963263679-project-member] [instance: 00746508-d0d6-4dfa-9026-772398d004ee] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 759.210559] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1e2fd6d0-d4ec-4803-bcdb-16f3753950f0 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.219063] env[65726]: DEBUG oslo_vmware.api [None req-b479f246-d94b-4634-888f-dee4586c3272 tempest-ServerAddressesNegativeTestJSON-963263679 tempest-ServerAddressesNegativeTestJSON-963263679-project-member] Waiting for the task: (returnval){ [ 759.219063] env[65726]: value = "task-5115792" [ 759.219063] env[65726]: _type = "Task" [ 759.219063] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 759.232087] env[65726]: DEBUG oslo_vmware.api [None req-b479f246-d94b-4634-888f-dee4586c3272 tempest-ServerAddressesNegativeTestJSON-963263679 tempest-ServerAddressesNegativeTestJSON-963263679-project-member] Task: {'id': task-5115792, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 759.331577] env[65726]: DEBUG oslo_vmware.api [None req-c9cc880b-42e7-4875-9b8e-55f0109d1664 tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Task: {'id': task-5115790, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.197602} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 759.331900] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-c9cc880b-42e7-4875-9b8e-55f0109d1664 tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Deleted the datastore file {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 759.332432] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-c9cc880b-42e7-4875-9b8e-55f0109d1664 tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] [instance: ca67d1a4-d304-4d06-a436-f257b60ca080] Deleted contents of the VM from datastore datastore2 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 759.332811] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-c9cc880b-42e7-4875-9b8e-55f0109d1664 tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] [instance: ca67d1a4-d304-4d06-a436-f257b60ca080] Instance destroyed {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 759.333639] env[65726]: INFO nova.compute.manager [None req-c9cc880b-42e7-4875-9b8e-55f0109d1664 tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] [instance: ca67d1a4-d304-4d06-a436-f257b60ca080] Took 1.19 seconds to destroy the instance on the hypervisor. [ 759.333983] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-c9cc880b-42e7-4875-9b8e-55f0109d1664 tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 759.334236] env[65726]: DEBUG nova.compute.manager [-] [instance: ca67d1a4-d304-4d06-a436-f257b60ca080] Deallocating network for instance {{(pid=65726) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 759.334342] env[65726]: DEBUG nova.network.neutron [-] [instance: ca67d1a4-d304-4d06-a436-f257b60ca080] deallocate_for_instance() {{(pid=65726) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 759.335021] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 759.335120] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 759.361220] env[65726]: DEBUG nova.compute.manager [None req-af2f0c1d-9bc5-40c3-a712-d408e73bc806 tempest-ServerDiagnosticsNegativeTest-1635132145 tempest-ServerDiagnosticsNegativeTest-1635132145-project-member] [instance: 0d6f14c5-5770-476f-a186-4cf1b2d4fbe8] Start spawning the instance on the hypervisor. {{(pid=65726) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 759.397387] env[65726]: DEBUG nova.virt.hardware [None req-af2f0c1d-9bc5-40c3-a712-d408e73bc806 tempest-ServerDiagnosticsNegativeTest-1635132145 tempest-ServerDiagnosticsNegativeTest-1635132145-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T19:28:34Z,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1539f06fa3534e90acbc3a60c4b8bd3f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T19:28:35Z,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 759.397680] env[65726]: DEBUG nova.virt.hardware [None req-af2f0c1d-9bc5-40c3-a712-d408e73bc806 tempest-ServerDiagnosticsNegativeTest-1635132145 tempest-ServerDiagnosticsNegativeTest-1635132145-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 759.397839] env[65726]: DEBUG nova.virt.hardware [None req-af2f0c1d-9bc5-40c3-a712-d408e73bc806 tempest-ServerDiagnosticsNegativeTest-1635132145 tempest-ServerDiagnosticsNegativeTest-1635132145-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 759.398026] env[65726]: DEBUG nova.virt.hardware [None req-af2f0c1d-9bc5-40c3-a712-d408e73bc806 tempest-ServerDiagnosticsNegativeTest-1635132145 tempest-ServerDiagnosticsNegativeTest-1635132145-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 759.398192] env[65726]: DEBUG nova.virt.hardware [None req-af2f0c1d-9bc5-40c3-a712-d408e73bc806 tempest-ServerDiagnosticsNegativeTest-1635132145 tempest-ServerDiagnosticsNegativeTest-1635132145-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 759.398348] env[65726]: DEBUG nova.virt.hardware [None req-af2f0c1d-9bc5-40c3-a712-d408e73bc806 tempest-ServerDiagnosticsNegativeTest-1635132145 tempest-ServerDiagnosticsNegativeTest-1635132145-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 759.398571] env[65726]: DEBUG nova.virt.hardware [None req-af2f0c1d-9bc5-40c3-a712-d408e73bc806 tempest-ServerDiagnosticsNegativeTest-1635132145 tempest-ServerDiagnosticsNegativeTest-1635132145-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 759.398722] env[65726]: DEBUG nova.virt.hardware [None req-af2f0c1d-9bc5-40c3-a712-d408e73bc806 tempest-ServerDiagnosticsNegativeTest-1635132145 tempest-ServerDiagnosticsNegativeTest-1635132145-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 759.398886] env[65726]: DEBUG nova.virt.hardware [None req-af2f0c1d-9bc5-40c3-a712-d408e73bc806 tempest-ServerDiagnosticsNegativeTest-1635132145 tempest-ServerDiagnosticsNegativeTest-1635132145-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 759.399056] env[65726]: DEBUG nova.virt.hardware [None req-af2f0c1d-9bc5-40c3-a712-d408e73bc806 tempest-ServerDiagnosticsNegativeTest-1635132145 tempest-ServerDiagnosticsNegativeTest-1635132145-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 759.399234] env[65726]: DEBUG nova.virt.hardware [None req-af2f0c1d-9bc5-40c3-a712-d408e73bc806 tempest-ServerDiagnosticsNegativeTest-1635132145 tempest-ServerDiagnosticsNegativeTest-1635132145-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 759.400257] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e043d6b-bc0d-4f95-8fe5-9dae7edeaa83 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.409999] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f6b7f24-076c-451b-bd74-cebe5dd56d70 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.431832] env[65726]: DEBUG oslo_vmware.api [None req-04afb47c-692a-46d9-a640-68cc2c90206f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Task: {'id': task-5115791, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 759.470855] env[65726]: DEBUG oslo_vmware.api [None req-44e83fc7-afcc-4134-bbbb-6bc576931bb8 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52f98f3f-f5b6-5a71-f46f-7f1c7baa5d25, 'name': SearchDatastore_Task, 'duration_secs': 0.010835} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 759.472161] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-35627a98-c80a-47b5-808f-b6a304e5eb1e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.478634] env[65726]: DEBUG oslo_vmware.api [None req-44e83fc7-afcc-4134-bbbb-6bc576931bb8 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Waiting for the task: (returnval){ [ 759.478634] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52781265-4438-3841-5bf3-ae3830247f24" [ 759.478634] env[65726]: _type = "Task" [ 759.478634] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 759.487063] env[65726]: DEBUG oslo_vmware.api [None req-44e83fc7-afcc-4134-bbbb-6bc576931bb8 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52781265-4438-3841-5bf3-ae3830247f24, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 759.585532] env[65726]: DEBUG nova.scheduler.client.report [None req-48153719-e51d-407a-98ef-93401c6270f2 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 759.635821] env[65726]: WARNING openstack [req-207d272f-0a18-4cd6-8840-84c1a6367cbd req-077126f5-a106-4772-9d7e-e3ed8f50e430 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 759.636331] env[65726]: WARNING openstack [req-207d272f-0a18-4cd6-8840-84c1a6367cbd req-077126f5-a106-4772-9d7e-e3ed8f50e430 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 759.711448] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a78d739e-5974-49c3-a6f0-74e291b9e87b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Acquiring lock "02351120-8385-4403-8464-a154f3da9380" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 759.711875] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a78d739e-5974-49c3-a6f0-74e291b9e87b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Lock "02351120-8385-4403-8464-a154f3da9380" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 759.712132] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a78d739e-5974-49c3-a6f0-74e291b9e87b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Acquiring lock "02351120-8385-4403-8464-a154f3da9380-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 759.712292] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a78d739e-5974-49c3-a6f0-74e291b9e87b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Lock "02351120-8385-4403-8464-a154f3da9380-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 759.712497] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a78d739e-5974-49c3-a6f0-74e291b9e87b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Lock "02351120-8385-4403-8464-a154f3da9380-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 759.715043] env[65726]: INFO nova.compute.manager [None req-a78d739e-5974-49c3-a6f0-74e291b9e87b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] [instance: 02351120-8385-4403-8464-a154f3da9380] Terminating instance [ 759.732191] env[65726]: DEBUG oslo_vmware.api [None req-b479f246-d94b-4634-888f-dee4586c3272 tempest-ServerAddressesNegativeTestJSON-963263679 tempest-ServerAddressesNegativeTestJSON-963263679-project-member] Task: {'id': task-5115792, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 759.757015] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 759.758020] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 759.846457] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 759.846757] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 759.926125] env[65726]: DEBUG oslo_vmware.api [None req-04afb47c-692a-46d9-a640-68cc2c90206f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Task: {'id': task-5115791, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.539826} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 759.926407] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-04afb47c-692a-46d9-a640-68cc2c90206f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] c108f2a5-031b-47ee-9a5b-d62c2f42c26b/c108f2a5-031b-47ee-9a5b-d62c2f42c26b.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 759.926654] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-04afb47c-692a-46d9-a640-68cc2c90206f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: c108f2a5-031b-47ee-9a5b-d62c2f42c26b] Extending root virtual disk to 1048576 {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 759.927045] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a557a7fc-57a2-4a80-b445-99df143fd826 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.940012] env[65726]: DEBUG oslo_vmware.api [None req-04afb47c-692a-46d9-a640-68cc2c90206f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Waiting for the task: (returnval){ [ 759.940012] env[65726]: value = "task-5115793" [ 759.940012] env[65726]: _type = "Task" [ 759.940012] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 759.951381] env[65726]: DEBUG oslo_vmware.api [None req-04afb47c-692a-46d9-a640-68cc2c90206f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Task: {'id': task-5115793, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 759.991830] env[65726]: DEBUG oslo_vmware.api [None req-44e83fc7-afcc-4134-bbbb-6bc576931bb8 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52781265-4438-3841-5bf3-ae3830247f24, 'name': SearchDatastore_Task, 'duration_secs': 0.012338} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 759.992116] env[65726]: DEBUG oslo_concurrency.lockutils [None req-44e83fc7-afcc-4134-bbbb-6bc576931bb8 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 759.992370] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-44e83fc7-afcc-4134-bbbb-6bc576931bb8 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore2] 895797e4-2941-44cd-aab1-67afa4fac02d/895797e4-2941-44cd-aab1-67afa4fac02d.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 759.996868] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b35fd4a9-31aa-4885-a3bd-b7204ad2a35e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.005154] env[65726]: DEBUG oslo_vmware.api [None req-44e83fc7-afcc-4134-bbbb-6bc576931bb8 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Waiting for the task: (returnval){ [ 760.005154] env[65726]: value = "task-5115794" [ 760.005154] env[65726]: _type = "Task" [ 760.005154] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 760.014205] env[65726]: DEBUG oslo_vmware.api [None req-44e83fc7-afcc-4134-bbbb-6bc576931bb8 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Task: {'id': task-5115794, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 760.101116] env[65726]: DEBUG oslo_concurrency.lockutils [None req-48153719-e51d-407a-98ef-93401c6270f2 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.790s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 760.102208] env[65726]: DEBUG nova.compute.manager [None req-48153719-e51d-407a-98ef-93401c6270f2 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] [instance: 65aaa2eb-3ca5-4c2b-92f9-b524366c10dd] Start building networks asynchronously for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 760.109551] env[65726]: DEBUG oslo_concurrency.lockutils [None req-6c979061-2d03-4525-8b87-afa801f4542e tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 32.689s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 760.109551] env[65726]: INFO nova.compute.claims [None req-6c979061-2d03-4525-8b87-afa801f4542e tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: 6d2b86c2-0a44-433f-8b3f-2e9c7f693baa] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 760.219800] env[65726]: DEBUG nova.compute.manager [None req-a78d739e-5974-49c3-a6f0-74e291b9e87b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] [instance: 02351120-8385-4403-8464-a154f3da9380] Start destroying the instance on the hypervisor. {{(pid=65726) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 760.220254] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-a78d739e-5974-49c3-a6f0-74e291b9e87b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] [instance: 02351120-8385-4403-8464-a154f3da9380] Destroying instance {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 760.221166] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-071a6f4a-cafc-45fe-8dce-e964add8cd26 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.237205] env[65726]: DEBUG oslo_vmware.api [None req-b479f246-d94b-4634-888f-dee4586c3272 tempest-ServerAddressesNegativeTestJSON-963263679 tempest-ServerAddressesNegativeTestJSON-963263679-project-member] Task: {'id': task-5115792, 'name': PowerOnVM_Task, 'duration_secs': 0.963497} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 760.239771] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-b479f246-d94b-4634-888f-dee4586c3272 tempest-ServerAddressesNegativeTestJSON-963263679 tempest-ServerAddressesNegativeTestJSON-963263679-project-member] [instance: 00746508-d0d6-4dfa-9026-772398d004ee] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 760.241058] env[65726]: INFO nova.compute.manager [None req-b479f246-d94b-4634-888f-dee4586c3272 tempest-ServerAddressesNegativeTestJSON-963263679 tempest-ServerAddressesNegativeTestJSON-963263679-project-member] [instance: 00746508-d0d6-4dfa-9026-772398d004ee] Took 10.00 seconds to spawn the instance on the hypervisor. [ 760.241058] env[65726]: DEBUG nova.compute.manager [None req-b479f246-d94b-4634-888f-dee4586c3272 tempest-ServerAddressesNegativeTestJSON-963263679 tempest-ServerAddressesNegativeTestJSON-963263679-project-member] [instance: 00746508-d0d6-4dfa-9026-772398d004ee] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 760.241058] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-a78d739e-5974-49c3-a6f0-74e291b9e87b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] [instance: 02351120-8385-4403-8464-a154f3da9380] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 760.242697] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ca4c253-210b-483f-b7bd-8dfa86367e6c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.245636] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4aa1063d-6ad8-4a1b-a724-cbef2353d5f9 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.252321] env[65726]: DEBUG nova.network.neutron [None req-af2f0c1d-9bc5-40c3-a712-d408e73bc806 tempest-ServerDiagnosticsNegativeTest-1635132145 tempest-ServerDiagnosticsNegativeTest-1635132145-project-member] [instance: 0d6f14c5-5770-476f-a186-4cf1b2d4fbe8] Successfully updated port: c8363d25-c607-4d04-ab53-412848d7cf4f {{(pid=65726) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 760.268953] env[65726]: DEBUG oslo_vmware.api [None req-a78d739e-5974-49c3-a6f0-74e291b9e87b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Waiting for the task: (returnval){ [ 760.268953] env[65726]: value = "task-5115795" [ 760.268953] env[65726]: _type = "Task" [ 760.268953] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 760.282944] env[65726]: DEBUG oslo_vmware.api [None req-a78d739e-5974-49c3-a6f0-74e291b9e87b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Task: {'id': task-5115795, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 760.452414] env[65726]: DEBUG oslo_vmware.api [None req-04afb47c-692a-46d9-a640-68cc2c90206f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Task: {'id': task-5115793, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.229225} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 760.452773] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-04afb47c-692a-46d9-a640-68cc2c90206f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: c108f2a5-031b-47ee-9a5b-d62c2f42c26b] Extended root virtual disk {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 760.453708] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56083f1a-a5d9-4aca-b90c-74568fbbb9ea {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.497023] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-04afb47c-692a-46d9-a640-68cc2c90206f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: c108f2a5-031b-47ee-9a5b-d62c2f42c26b] Reconfiguring VM instance instance-00000027 to attach disk [datastore1] c108f2a5-031b-47ee-9a5b-d62c2f42c26b/c108f2a5-031b-47ee-9a5b-d62c2f42c26b.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 760.497540] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bcfc651c-26b4-42d7-8f8f-fc99b56af510 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.534841] env[65726]: DEBUG oslo_vmware.api [None req-44e83fc7-afcc-4134-bbbb-6bc576931bb8 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Task: {'id': task-5115794, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 760.540382] env[65726]: DEBUG oslo_vmware.api [None req-04afb47c-692a-46d9-a640-68cc2c90206f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Waiting for the task: (returnval){ [ 760.540382] env[65726]: value = "task-5115796" [ 760.540382] env[65726]: _type = "Task" [ 760.540382] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 760.551539] env[65726]: DEBUG oslo_vmware.api [None req-04afb47c-692a-46d9-a640-68cc2c90206f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Task: {'id': task-5115796, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 760.576169] env[65726]: DEBUG nova.network.neutron [req-207d272f-0a18-4cd6-8840-84c1a6367cbd req-077126f5-a106-4772-9d7e-e3ed8f50e430 service nova] [instance: 00746508-d0d6-4dfa-9026-772398d004ee] Updated VIF entry in instance network info cache for port 1151551e-12f0-4e85-a385-f1ebf4e9bc72. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 760.576386] env[65726]: DEBUG nova.network.neutron [req-207d272f-0a18-4cd6-8840-84c1a6367cbd req-077126f5-a106-4772-9d7e-e3ed8f50e430 service nova] [instance: 00746508-d0d6-4dfa-9026-772398d004ee] Updating instance_info_cache with network_info: [{"id": "1151551e-12f0-4e85-a385-f1ebf4e9bc72", "address": "fa:16:3e:eb:da:97", "network": {"id": "833c6007-53da-4e80-95f8-cd5f400ba120", "bridge": "br-int", "label": "tempest-ServerAddressesNegativeTestJSON-1292797458-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0d2e9e9fe170484896239f5f9cc9dc09", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69f65356-c85e-4b7f-ad28-7c7b5e8cf50c", "external-id": "nsx-vlan-transportzone-281", "segmentation_id": 281, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1151551e-12", "ovs_interfaceid": "1151551e-12f0-4e85-a385-f1ebf4e9bc72", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 760.613921] env[65726]: DEBUG nova.compute.utils [None req-48153719-e51d-407a-98ef-93401c6270f2 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Using /dev/sd instead of None {{(pid=65726) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 760.619430] env[65726]: DEBUG nova.compute.manager [None req-48153719-e51d-407a-98ef-93401c6270f2 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] [instance: 65aaa2eb-3ca5-4c2b-92f9-b524366c10dd] Allocating IP information in the background. {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 760.619804] env[65726]: DEBUG nova.network.neutron [None req-48153719-e51d-407a-98ef-93401c6270f2 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] [instance: 65aaa2eb-3ca5-4c2b-92f9-b524366c10dd] allocate_for_instance() {{(pid=65726) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 760.620223] env[65726]: WARNING neutronclient.v2_0.client [None req-48153719-e51d-407a-98ef-93401c6270f2 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 760.620585] env[65726]: WARNING neutronclient.v2_0.client [None req-48153719-e51d-407a-98ef-93401c6270f2 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 760.621210] env[65726]: WARNING openstack [None req-48153719-e51d-407a-98ef-93401c6270f2 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 760.621673] env[65726]: WARNING openstack [None req-48153719-e51d-407a-98ef-93401c6270f2 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 760.721483] env[65726]: DEBUG nova.policy [None req-48153719-e51d-407a-98ef-93401c6270f2 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b23d26af16b54e6d9374fe28e5162d86', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'aac5fe6c99f740ada3747088b2b0ad89', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65726) authorize /opt/stack/nova/nova/policy.py:192}} [ 760.733564] env[65726]: DEBUG nova.network.neutron [-] [instance: cd539d8b-921b-4947-954c-8573380dbdc8] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 760.754744] env[65726]: DEBUG oslo_concurrency.lockutils [None req-af2f0c1d-9bc5-40c3-a712-d408e73bc806 tempest-ServerDiagnosticsNegativeTest-1635132145 tempest-ServerDiagnosticsNegativeTest-1635132145-project-member] Acquiring lock "refresh_cache-0d6f14c5-5770-476f-a186-4cf1b2d4fbe8" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 760.754955] env[65726]: DEBUG oslo_concurrency.lockutils [None req-af2f0c1d-9bc5-40c3-a712-d408e73bc806 tempest-ServerDiagnosticsNegativeTest-1635132145 tempest-ServerDiagnosticsNegativeTest-1635132145-project-member] Acquired lock "refresh_cache-0d6f14c5-5770-476f-a186-4cf1b2d4fbe8" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 760.755195] env[65726]: DEBUG nova.network.neutron [None req-af2f0c1d-9bc5-40c3-a712-d408e73bc806 tempest-ServerDiagnosticsNegativeTest-1635132145 tempest-ServerDiagnosticsNegativeTest-1635132145-project-member] [instance: 0d6f14c5-5770-476f-a186-4cf1b2d4fbe8] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 760.784757] env[65726]: INFO nova.compute.manager [None req-b479f246-d94b-4634-888f-dee4586c3272 tempest-ServerAddressesNegativeTestJSON-963263679 tempest-ServerAddressesNegativeTestJSON-963263679-project-member] [instance: 00746508-d0d6-4dfa-9026-772398d004ee] Took 56.42 seconds to build instance. [ 760.796405] env[65726]: DEBUG oslo_vmware.api [None req-a78d739e-5974-49c3-a6f0-74e291b9e87b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Task: {'id': task-5115795, 'name': PowerOffVM_Task, 'duration_secs': 0.405098} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 760.796635] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-a78d739e-5974-49c3-a6f0-74e291b9e87b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] [instance: 02351120-8385-4403-8464-a154f3da9380] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 760.796815] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-a78d739e-5974-49c3-a6f0-74e291b9e87b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] [instance: 02351120-8385-4403-8464-a154f3da9380] Unregistering the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 760.796955] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-36bcb800-bd0c-42e4-a003-9a662fce9a4c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.896131] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-a78d739e-5974-49c3-a6f0-74e291b9e87b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] [instance: 02351120-8385-4403-8464-a154f3da9380] Unregistered the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 760.896378] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-a78d739e-5974-49c3-a6f0-74e291b9e87b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] [instance: 02351120-8385-4403-8464-a154f3da9380] Deleting contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 760.896591] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-a78d739e-5974-49c3-a6f0-74e291b9e87b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Deleting the datastore file [datastore1] 02351120-8385-4403-8464-a154f3da9380 {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 760.896846] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7fe86b38-fd42-48f0-9a79-c6e1eafb2320 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.905954] env[65726]: DEBUG oslo_vmware.api [None req-a78d739e-5974-49c3-a6f0-74e291b9e87b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Waiting for the task: (returnval){ [ 760.905954] env[65726]: value = "task-5115798" [ 760.905954] env[65726]: _type = "Task" [ 760.905954] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 760.915731] env[65726]: DEBUG oslo_vmware.api [None req-a78d739e-5974-49c3-a6f0-74e291b9e87b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Task: {'id': task-5115798, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 761.015032] env[65726]: DEBUG nova.network.neutron [None req-48153719-e51d-407a-98ef-93401c6270f2 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] [instance: 65aaa2eb-3ca5-4c2b-92f9-b524366c10dd] Successfully created port: 862071bf-725c-4193-b60b-72b03b4c0553 {{(pid=65726) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 761.032543] env[65726]: DEBUG oslo_vmware.api [None req-44e83fc7-afcc-4134-bbbb-6bc576931bb8 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Task: {'id': task-5115794, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.672782} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 761.032808] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-44e83fc7-afcc-4134-bbbb-6bc576931bb8 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore2] 895797e4-2941-44cd-aab1-67afa4fac02d/895797e4-2941-44cd-aab1-67afa4fac02d.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 761.033028] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-44e83fc7-afcc-4134-bbbb-6bc576931bb8 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: 895797e4-2941-44cd-aab1-67afa4fac02d] Extending root virtual disk to 1048576 {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 761.033286] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5ada09c9-fe12-4828-b8bb-92f4a2187e70 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.041190] env[65726]: DEBUG nova.network.neutron [-] [instance: ca67d1a4-d304-4d06-a436-f257b60ca080] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 761.042536] env[65726]: DEBUG oslo_vmware.api [None req-44e83fc7-afcc-4134-bbbb-6bc576931bb8 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Waiting for the task: (returnval){ [ 761.042536] env[65726]: value = "task-5115799" [ 761.042536] env[65726]: _type = "Task" [ 761.042536] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 761.061426] env[65726]: DEBUG oslo_vmware.api [None req-04afb47c-692a-46d9-a640-68cc2c90206f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Task: {'id': task-5115796, 'name': ReconfigVM_Task, 'duration_secs': 0.347874} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 761.065131] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-04afb47c-692a-46d9-a640-68cc2c90206f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: c108f2a5-031b-47ee-9a5b-d62c2f42c26b] Reconfigured VM instance instance-00000027 to attach disk [datastore1] c108f2a5-031b-47ee-9a5b-d62c2f42c26b/c108f2a5-031b-47ee-9a5b-d62c2f42c26b.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 761.065827] env[65726]: DEBUG oslo_vmware.api [None req-44e83fc7-afcc-4134-bbbb-6bc576931bb8 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Task: {'id': task-5115799, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 761.066103] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6b14a20f-e28f-4c04-8f3c-57a9807a111f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.074426] env[65726]: DEBUG oslo_vmware.api [None req-04afb47c-692a-46d9-a640-68cc2c90206f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Waiting for the task: (returnval){ [ 761.074426] env[65726]: value = "task-5115800" [ 761.074426] env[65726]: _type = "Task" [ 761.074426] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 761.079505] env[65726]: DEBUG oslo_concurrency.lockutils [req-207d272f-0a18-4cd6-8840-84c1a6367cbd req-077126f5-a106-4772-9d7e-e3ed8f50e430 service nova] Releasing lock "refresh_cache-00746508-d0d6-4dfa-9026-772398d004ee" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 761.084259] env[65726]: DEBUG oslo_vmware.api [None req-04afb47c-692a-46d9-a640-68cc2c90206f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Task: {'id': task-5115800, 'name': Rename_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 761.121210] env[65726]: DEBUG nova.compute.manager [None req-48153719-e51d-407a-98ef-93401c6270f2 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] [instance: 65aaa2eb-3ca5-4c2b-92f9-b524366c10dd] Start building block device mappings for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 761.236651] env[65726]: INFO nova.compute.manager [-] [instance: cd539d8b-921b-4947-954c-8573380dbdc8] Took 2.05 seconds to deallocate network for instance. [ 761.257989] env[65726]: WARNING openstack [None req-af2f0c1d-9bc5-40c3-a712-d408e73bc806 tempest-ServerDiagnosticsNegativeTest-1635132145 tempest-ServerDiagnosticsNegativeTest-1635132145-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 761.259783] env[65726]: WARNING openstack [None req-af2f0c1d-9bc5-40c3-a712-d408e73bc806 tempest-ServerDiagnosticsNegativeTest-1635132145 tempest-ServerDiagnosticsNegativeTest-1635132145-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 761.268787] env[65726]: DEBUG nova.network.neutron [None req-af2f0c1d-9bc5-40c3-a712-d408e73bc806 tempest-ServerDiagnosticsNegativeTest-1635132145 tempest-ServerDiagnosticsNegativeTest-1635132145-project-member] [instance: 0d6f14c5-5770-476f-a186-4cf1b2d4fbe8] Instance cache missing network info. {{(pid=65726) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 761.290875] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b479f246-d94b-4634-888f-dee4586c3272 tempest-ServerAddressesNegativeTestJSON-963263679 tempest-ServerAddressesNegativeTestJSON-963263679-project-member] Lock "00746508-d0d6-4dfa-9026-772398d004ee" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 81.213s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 761.422981] env[65726]: DEBUG oslo_vmware.api [None req-a78d739e-5974-49c3-a6f0-74e291b9e87b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Task: {'id': task-5115798, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.319157} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 761.423340] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-a78d739e-5974-49c3-a6f0-74e291b9e87b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Deleted the datastore file {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 761.423558] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-a78d739e-5974-49c3-a6f0-74e291b9e87b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] [instance: 02351120-8385-4403-8464-a154f3da9380] Deleted contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 761.423784] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-a78d739e-5974-49c3-a6f0-74e291b9e87b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] [instance: 02351120-8385-4403-8464-a154f3da9380] Instance destroyed {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 761.423977] env[65726]: INFO nova.compute.manager [None req-a78d739e-5974-49c3-a6f0-74e291b9e87b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] [instance: 02351120-8385-4403-8464-a154f3da9380] Took 1.20 seconds to destroy the instance on the hypervisor. [ 761.426052] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-a78d739e-5974-49c3-a6f0-74e291b9e87b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 761.426052] env[65726]: DEBUG nova.compute.manager [-] [instance: 02351120-8385-4403-8464-a154f3da9380] Deallocating network for instance {{(pid=65726) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 761.426052] env[65726]: DEBUG nova.network.neutron [-] [instance: 02351120-8385-4403-8464-a154f3da9380] deallocate_for_instance() {{(pid=65726) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 761.426052] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 761.426052] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 761.544130] env[65726]: INFO nova.compute.manager [-] [instance: ca67d1a4-d304-4d06-a436-f257b60ca080] Took 2.21 seconds to deallocate network for instance. [ 761.565629] env[65726]: DEBUG oslo_vmware.api [None req-44e83fc7-afcc-4134-bbbb-6bc576931bb8 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Task: {'id': task-5115799, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.101368} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 761.566228] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-44e83fc7-afcc-4134-bbbb-6bc576931bb8 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: 895797e4-2941-44cd-aab1-67afa4fac02d] Extended root virtual disk {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 761.567610] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27d7f7aa-8621-4a80-9a01-3362649ffc46 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.599432] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-44e83fc7-afcc-4134-bbbb-6bc576931bb8 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: 895797e4-2941-44cd-aab1-67afa4fac02d] Reconfiguring VM instance instance-00000028 to attach disk [datastore2] 895797e4-2941-44cd-aab1-67afa4fac02d/895797e4-2941-44cd-aab1-67afa4fac02d.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 761.603710] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-541b1216-7c20-48ea-a5ce-e8db7e2238a7 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.623443] env[65726]: DEBUG oslo_vmware.api [None req-04afb47c-692a-46d9-a640-68cc2c90206f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Task: {'id': task-5115800, 'name': Rename_Task} progress is 99%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 761.630148] env[65726]: DEBUG oslo_vmware.api [None req-44e83fc7-afcc-4134-bbbb-6bc576931bb8 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Waiting for the task: (returnval){ [ 761.630148] env[65726]: value = "task-5115801" [ 761.630148] env[65726]: _type = "Task" [ 761.630148] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 761.643514] env[65726]: DEBUG oslo_vmware.api [None req-44e83fc7-afcc-4134-bbbb-6bc576931bb8 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Task: {'id': task-5115801, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 761.749945] env[65726]: DEBUG oslo_concurrency.lockutils [None req-d63650ad-6689-479f-81d4-98dcfe6418a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 761.792760] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e176fb55-f8f4-48bf-82be-e5f2dc58951f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.797633] env[65726]: DEBUG nova.compute.manager [None req-04d3888f-ac16-4533-b7db-d3c669b17255 tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] [instance: bc2c12e0-0d06-432f-b42f-be468e3b6ee4] Starting instance... {{(pid=65726) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 761.806611] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0cd7aedf-caa5-463a-b38f-02a35cb98dee {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.840244] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fce635f6-e707-4fba-8225-c5b829a625fb {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.851481] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d07bb8f5-e2f5-4942-9a1d-1b8a5edc6ecb {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.856985] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9ff22eed-6c42-49bb-b683-3f83a9ff00ca tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Acquiring lock "ee676e3b-3326-46a3-940d-ebbb8b108991" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 761.856985] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9ff22eed-6c42-49bb-b683-3f83a9ff00ca tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Lock "ee676e3b-3326-46a3-940d-ebbb8b108991" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 761.857205] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9ff22eed-6c42-49bb-b683-3f83a9ff00ca tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Acquiring lock "ee676e3b-3326-46a3-940d-ebbb8b108991-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 761.857246] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9ff22eed-6c42-49bb-b683-3f83a9ff00ca tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Lock "ee676e3b-3326-46a3-940d-ebbb8b108991-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 761.857387] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9ff22eed-6c42-49bb-b683-3f83a9ff00ca tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Lock "ee676e3b-3326-46a3-940d-ebbb8b108991-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 761.860819] env[65726]: INFO nova.compute.manager [None req-9ff22eed-6c42-49bb-b683-3f83a9ff00ca tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] [instance: ee676e3b-3326-46a3-940d-ebbb8b108991] Terminating instance [ 761.870587] env[65726]: DEBUG nova.compute.provider_tree [None req-6c979061-2d03-4525-8b87-afa801f4542e tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 762.040313] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 762.040577] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 762.052805] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c9cc880b-42e7-4875-9b8e-55f0109d1664 tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 762.084868] env[65726]: DEBUG oslo_vmware.api [None req-04afb47c-692a-46d9-a640-68cc2c90206f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Task: {'id': task-5115800, 'name': Rename_Task} progress is 99%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 762.139180] env[65726]: DEBUG nova.compute.manager [None req-48153719-e51d-407a-98ef-93401c6270f2 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] [instance: 65aaa2eb-3ca5-4c2b-92f9-b524366c10dd] Start spawning the instance on the hypervisor. {{(pid=65726) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 762.175041] env[65726]: DEBUG oslo_vmware.api [None req-44e83fc7-afcc-4134-bbbb-6bc576931bb8 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Task: {'id': task-5115801, 'name': ReconfigVM_Task, 'duration_secs': 0.356688} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 762.175222] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-44e83fc7-afcc-4134-bbbb-6bc576931bb8 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: 895797e4-2941-44cd-aab1-67afa4fac02d] Reconfigured VM instance instance-00000028 to attach disk [datastore2] 895797e4-2941-44cd-aab1-67afa4fac02d/895797e4-2941-44cd-aab1-67afa4fac02d.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 762.176639] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9d055e09-2dff-47a3-8acf-f1ec25413c0d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.186821] env[65726]: DEBUG oslo_vmware.api [None req-44e83fc7-afcc-4134-bbbb-6bc576931bb8 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Waiting for the task: (returnval){ [ 762.186821] env[65726]: value = "task-5115802" [ 762.186821] env[65726]: _type = "Task" [ 762.186821] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 762.209379] env[65726]: DEBUG oslo_vmware.api [None req-44e83fc7-afcc-4134-bbbb-6bc576931bb8 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Task: {'id': task-5115802, 'name': Rename_Task} progress is 10%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 762.215284] env[65726]: DEBUG nova.virt.hardware [None req-48153719-e51d-407a-98ef-93401c6270f2 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T19:28:34Z,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1539f06fa3534e90acbc3a60c4b8bd3f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T19:28:35Z,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 762.215744] env[65726]: DEBUG nova.virt.hardware [None req-48153719-e51d-407a-98ef-93401c6270f2 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 762.215898] env[65726]: DEBUG nova.virt.hardware [None req-48153719-e51d-407a-98ef-93401c6270f2 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 762.216281] env[65726]: DEBUG nova.virt.hardware [None req-48153719-e51d-407a-98ef-93401c6270f2 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 762.216418] env[65726]: DEBUG nova.virt.hardware [None req-48153719-e51d-407a-98ef-93401c6270f2 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 762.216552] env[65726]: DEBUG nova.virt.hardware [None req-48153719-e51d-407a-98ef-93401c6270f2 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 762.217318] env[65726]: DEBUG nova.virt.hardware [None req-48153719-e51d-407a-98ef-93401c6270f2 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 762.217478] env[65726]: DEBUG nova.virt.hardware [None req-48153719-e51d-407a-98ef-93401c6270f2 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 762.217772] env[65726]: DEBUG nova.virt.hardware [None req-48153719-e51d-407a-98ef-93401c6270f2 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 762.218600] env[65726]: DEBUG nova.virt.hardware [None req-48153719-e51d-407a-98ef-93401c6270f2 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 762.218835] env[65726]: DEBUG nova.virt.hardware [None req-48153719-e51d-407a-98ef-93401c6270f2 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 762.220913] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c56f1845-4b4d-44d9-b20c-1d3df206f34f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.241797] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b6d167c-4617-45b7-9763-8053a11d38f4 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.326898] env[65726]: DEBUG oslo_concurrency.lockutils [None req-04d3888f-ac16-4533-b7db-d3c669b17255 tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 762.374392] env[65726]: DEBUG nova.scheduler.client.report [None req-6c979061-2d03-4525-8b87-afa801f4542e tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 762.378153] env[65726]: DEBUG nova.compute.manager [None req-9ff22eed-6c42-49bb-b683-3f83a9ff00ca tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] [instance: ee676e3b-3326-46a3-940d-ebbb8b108991] Start destroying the instance on the hypervisor. {{(pid=65726) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 762.378362] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-9ff22eed-6c42-49bb-b683-3f83a9ff00ca tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] [instance: ee676e3b-3326-46a3-940d-ebbb8b108991] Destroying instance {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 762.379463] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81d0ab19-e1b5-424a-aaec-bf4ad9e741ca {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.387326] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-9ff22eed-6c42-49bb-b683-3f83a9ff00ca tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] [instance: ee676e3b-3326-46a3-940d-ebbb8b108991] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 762.387607] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7af0dbcd-dfe5-4ed2-928d-62180e956163 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.395990] env[65726]: DEBUG oslo_vmware.api [None req-9ff22eed-6c42-49bb-b683-3f83a9ff00ca tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Waiting for the task: (returnval){ [ 762.395990] env[65726]: value = "task-5115803" [ 762.395990] env[65726]: _type = "Task" [ 762.395990] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 762.406340] env[65726]: DEBUG oslo_vmware.api [None req-9ff22eed-6c42-49bb-b683-3f83a9ff00ca tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Task: {'id': task-5115803, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 762.587521] env[65726]: DEBUG oslo_vmware.api [None req-04afb47c-692a-46d9-a640-68cc2c90206f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Task: {'id': task-5115800, 'name': Rename_Task, 'duration_secs': 1.136509} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 762.587521] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-04afb47c-692a-46d9-a640-68cc2c90206f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: c108f2a5-031b-47ee-9a5b-d62c2f42c26b] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 762.587606] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e7a588fe-4517-42b5-ab28-f00211ff66ea {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.595475] env[65726]: DEBUG oslo_vmware.api [None req-04afb47c-692a-46d9-a640-68cc2c90206f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Waiting for the task: (returnval){ [ 762.595475] env[65726]: value = "task-5115804" [ 762.595475] env[65726]: _type = "Task" [ 762.595475] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 762.604743] env[65726]: DEBUG oslo_vmware.api [None req-04afb47c-692a-46d9-a640-68cc2c90206f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Task: {'id': task-5115804, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 762.702333] env[65726]: DEBUG oslo_vmware.api [None req-44e83fc7-afcc-4134-bbbb-6bc576931bb8 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Task: {'id': task-5115802, 'name': Rename_Task, 'duration_secs': 0.179537} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 762.702333] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-44e83fc7-afcc-4134-bbbb-6bc576931bb8 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: 895797e4-2941-44cd-aab1-67afa4fac02d] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 762.702333] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-dcda3496-f894-4ea0-ac7f-945d3b8a68fd {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.711235] env[65726]: DEBUG oslo_vmware.api [None req-44e83fc7-afcc-4134-bbbb-6bc576931bb8 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Waiting for the task: (returnval){ [ 762.711235] env[65726]: value = "task-5115805" [ 762.711235] env[65726]: _type = "Task" [ 762.711235] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 762.722653] env[65726]: DEBUG oslo_vmware.api [None req-44e83fc7-afcc-4134-bbbb-6bc576931bb8 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Task: {'id': task-5115805, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 762.802821] env[65726]: DEBUG nova.network.neutron [None req-48153719-e51d-407a-98ef-93401c6270f2 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] [instance: 65aaa2eb-3ca5-4c2b-92f9-b524366c10dd] Successfully updated port: 862071bf-725c-4193-b60b-72b03b4c0553 {{(pid=65726) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 762.820421] env[65726]: WARNING openstack [None req-af2f0c1d-9bc5-40c3-a712-d408e73bc806 tempest-ServerDiagnosticsNegativeTest-1635132145 tempest-ServerDiagnosticsNegativeTest-1635132145-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 762.820845] env[65726]: WARNING openstack [None req-af2f0c1d-9bc5-40c3-a712-d408e73bc806 tempest-ServerDiagnosticsNegativeTest-1635132145 tempest-ServerDiagnosticsNegativeTest-1635132145-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 762.879877] env[65726]: DEBUG oslo_concurrency.lockutils [None req-6c979061-2d03-4525-8b87-afa801f4542e tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.773s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 762.880452] env[65726]: DEBUG nova.compute.manager [None req-6c979061-2d03-4525-8b87-afa801f4542e tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: 6d2b86c2-0a44-433f-8b3f-2e9c7f693baa] Start building networks asynchronously for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 762.883258] env[65726]: DEBUG oslo_concurrency.lockutils [None req-d51d38e8-71be-4889-a206-d52403737076 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 34.601s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 762.885276] env[65726]: INFO nova.compute.claims [None req-d51d38e8-71be-4889-a206-d52403737076 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: 9fc0c9ee-9dc7-4059-b8d2-a5db965e00c2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 762.906511] env[65726]: DEBUG oslo_vmware.api [None req-9ff22eed-6c42-49bb-b683-3f83a9ff00ca tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Task: {'id': task-5115803, 'name': PowerOffVM_Task, 'duration_secs': 0.377601} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 762.906777] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-9ff22eed-6c42-49bb-b683-3f83a9ff00ca tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] [instance: ee676e3b-3326-46a3-940d-ebbb8b108991] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 762.906931] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-9ff22eed-6c42-49bb-b683-3f83a9ff00ca tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] [instance: ee676e3b-3326-46a3-940d-ebbb8b108991] Unregistering the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 762.907414] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-43117dfe-3996-48e0-bfd1-7eb0349cdb71 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.979163] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-9ff22eed-6c42-49bb-b683-3f83a9ff00ca tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] [instance: ee676e3b-3326-46a3-940d-ebbb8b108991] Unregistered the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 762.979449] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-9ff22eed-6c42-49bb-b683-3f83a9ff00ca tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] [instance: ee676e3b-3326-46a3-940d-ebbb8b108991] Deleting contents of the VM from datastore datastore2 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 762.979707] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-9ff22eed-6c42-49bb-b683-3f83a9ff00ca tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Deleting the datastore file [datastore2] ee676e3b-3326-46a3-940d-ebbb8b108991 {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 762.981099] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-79e50b27-88c0-4fa1-b6e1-33028961ba19 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.997318] env[65726]: DEBUG oslo_vmware.api [None req-9ff22eed-6c42-49bb-b683-3f83a9ff00ca tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Waiting for the task: (returnval){ [ 762.997318] env[65726]: value = "task-5115807" [ 762.997318] env[65726]: _type = "Task" [ 762.997318] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 763.010833] env[65726]: DEBUG oslo_vmware.api [None req-9ff22eed-6c42-49bb-b683-3f83a9ff00ca tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Task: {'id': task-5115807, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 763.066083] env[65726]: WARNING openstack [None req-af2f0c1d-9bc5-40c3-a712-d408e73bc806 tempest-ServerDiagnosticsNegativeTest-1635132145 tempest-ServerDiagnosticsNegativeTest-1635132145-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 763.067046] env[65726]: WARNING openstack [None req-af2f0c1d-9bc5-40c3-a712-d408e73bc806 tempest-ServerDiagnosticsNegativeTest-1635132145 tempest-ServerDiagnosticsNegativeTest-1635132145-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 763.108667] env[65726]: DEBUG oslo_vmware.api [None req-04afb47c-692a-46d9-a640-68cc2c90206f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Task: {'id': task-5115804, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 763.201108] env[65726]: DEBUG nova.network.neutron [None req-af2f0c1d-9bc5-40c3-a712-d408e73bc806 tempest-ServerDiagnosticsNegativeTest-1635132145 tempest-ServerDiagnosticsNegativeTest-1635132145-project-member] [instance: 0d6f14c5-5770-476f-a186-4cf1b2d4fbe8] Updating instance_info_cache with network_info: [{"id": "c8363d25-c607-4d04-ab53-412848d7cf4f", "address": "fa:16:3e:a8:64:02", "network": {"id": "55d629b8-a541-4720-9289-927bbc6e6f22", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.61", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "1539f06fa3534e90acbc3a60c4b8bd3f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "35342bcb-8b06-472e-b3c0-43fd3d6c4b30", "external-id": "nsx-vlan-transportzone-524", "segmentation_id": 524, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc8363d25-c6", "ovs_interfaceid": "c8363d25-c607-4d04-ab53-412848d7cf4f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 763.225041] env[65726]: DEBUG oslo_vmware.api [None req-44e83fc7-afcc-4134-bbbb-6bc576931bb8 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Task: {'id': task-5115805, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 763.267154] env[65726]: DEBUG nova.compute.manager [req-9dd6fa57-ace0-4209-bdb9-4c08cdca9942 req-38afafcd-4a90-4ef5-bf59-4ccbb9af9716 service nova] [instance: c108f2a5-031b-47ee-9a5b-d62c2f42c26b] Received event network-vif-plugged-981741c6-69f0-42d4-87be-e1c8f62d0e17 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 763.270565] env[65726]: DEBUG oslo_concurrency.lockutils [req-9dd6fa57-ace0-4209-bdb9-4c08cdca9942 req-38afafcd-4a90-4ef5-bf59-4ccbb9af9716 service nova] Acquiring lock "c108f2a5-031b-47ee-9a5b-d62c2f42c26b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 763.271496] env[65726]: DEBUG oslo_concurrency.lockutils [req-9dd6fa57-ace0-4209-bdb9-4c08cdca9942 req-38afafcd-4a90-4ef5-bf59-4ccbb9af9716 service nova] Lock "c108f2a5-031b-47ee-9a5b-d62c2f42c26b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 763.271692] env[65726]: DEBUG oslo_concurrency.lockutils [req-9dd6fa57-ace0-4209-bdb9-4c08cdca9942 req-38afafcd-4a90-4ef5-bf59-4ccbb9af9716 service nova] Lock "c108f2a5-031b-47ee-9a5b-d62c2f42c26b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.001s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 763.271871] env[65726]: DEBUG nova.compute.manager [req-9dd6fa57-ace0-4209-bdb9-4c08cdca9942 req-38afafcd-4a90-4ef5-bf59-4ccbb9af9716 service nova] [instance: c108f2a5-031b-47ee-9a5b-d62c2f42c26b] No waiting events found dispatching network-vif-plugged-981741c6-69f0-42d4-87be-e1c8f62d0e17 {{(pid=65726) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 763.272068] env[65726]: WARNING nova.compute.manager [req-9dd6fa57-ace0-4209-bdb9-4c08cdca9942 req-38afafcd-4a90-4ef5-bf59-4ccbb9af9716 service nova] [instance: c108f2a5-031b-47ee-9a5b-d62c2f42c26b] Received unexpected event network-vif-plugged-981741c6-69f0-42d4-87be-e1c8f62d0e17 for instance with vm_state building and task_state spawning. [ 763.272230] env[65726]: DEBUG nova.compute.manager [req-9dd6fa57-ace0-4209-bdb9-4c08cdca9942 req-38afafcd-4a90-4ef5-bf59-4ccbb9af9716 service nova] [instance: c108f2a5-031b-47ee-9a5b-d62c2f42c26b] Received event network-changed-981741c6-69f0-42d4-87be-e1c8f62d0e17 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 763.272372] env[65726]: DEBUG nova.compute.manager [req-9dd6fa57-ace0-4209-bdb9-4c08cdca9942 req-38afafcd-4a90-4ef5-bf59-4ccbb9af9716 service nova] [instance: c108f2a5-031b-47ee-9a5b-d62c2f42c26b] Refreshing instance network info cache due to event network-changed-981741c6-69f0-42d4-87be-e1c8f62d0e17. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 763.272551] env[65726]: DEBUG oslo_concurrency.lockutils [req-9dd6fa57-ace0-4209-bdb9-4c08cdca9942 req-38afafcd-4a90-4ef5-bf59-4ccbb9af9716 service nova] Acquiring lock "refresh_cache-c108f2a5-031b-47ee-9a5b-d62c2f42c26b" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 763.272743] env[65726]: DEBUG oslo_concurrency.lockutils [req-9dd6fa57-ace0-4209-bdb9-4c08cdca9942 req-38afafcd-4a90-4ef5-bf59-4ccbb9af9716 service nova] Acquired lock "refresh_cache-c108f2a5-031b-47ee-9a5b-d62c2f42c26b" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 763.273500] env[65726]: DEBUG nova.network.neutron [req-9dd6fa57-ace0-4209-bdb9-4c08cdca9942 req-38afafcd-4a90-4ef5-bf59-4ccbb9af9716 service nova] [instance: c108f2a5-031b-47ee-9a5b-d62c2f42c26b] Refreshing network info cache for port 981741c6-69f0-42d4-87be-e1c8f62d0e17 {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 763.306423] env[65726]: DEBUG oslo_concurrency.lockutils [None req-48153719-e51d-407a-98ef-93401c6270f2 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Acquiring lock "refresh_cache-65aaa2eb-3ca5-4c2b-92f9-b524366c10dd" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 763.306423] env[65726]: DEBUG oslo_concurrency.lockutils [None req-48153719-e51d-407a-98ef-93401c6270f2 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Acquired lock "refresh_cache-65aaa2eb-3ca5-4c2b-92f9-b524366c10dd" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 763.306575] env[65726]: DEBUG nova.network.neutron [None req-48153719-e51d-407a-98ef-93401c6270f2 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] [instance: 65aaa2eb-3ca5-4c2b-92f9-b524366c10dd] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 763.393435] env[65726]: DEBUG nova.compute.utils [None req-6c979061-2d03-4525-8b87-afa801f4542e tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Using /dev/sd instead of None {{(pid=65726) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 763.396987] env[65726]: DEBUG nova.compute.manager [None req-6c979061-2d03-4525-8b87-afa801f4542e tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: 6d2b86c2-0a44-433f-8b3f-2e9c7f693baa] Allocating IP information in the background. {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 763.399216] env[65726]: DEBUG nova.network.neutron [None req-6c979061-2d03-4525-8b87-afa801f4542e tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: 6d2b86c2-0a44-433f-8b3f-2e9c7f693baa] allocate_for_instance() {{(pid=65726) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 763.399216] env[65726]: WARNING neutronclient.v2_0.client [None req-6c979061-2d03-4525-8b87-afa801f4542e tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 763.399216] env[65726]: WARNING neutronclient.v2_0.client [None req-6c979061-2d03-4525-8b87-afa801f4542e tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 763.399216] env[65726]: WARNING openstack [None req-6c979061-2d03-4525-8b87-afa801f4542e tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 763.400228] env[65726]: WARNING openstack [None req-6c979061-2d03-4525-8b87-afa801f4542e tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 763.411508] env[65726]: DEBUG nova.compute.manager [req-beefbd62-56a8-4a78-a763-e39df8cb9486 req-c54bcb99-2361-48fd-8c06-27b5e5f431af service nova] [instance: cd539d8b-921b-4947-954c-8573380dbdc8] Received event network-vif-deleted-60d92cf3-683a-4291-ba51-fad4124f3c8c {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 763.460309] env[65726]: DEBUG nova.policy [None req-6c979061-2d03-4525-8b87-afa801f4542e tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'faf82c34c4884dc890edf0909b9f510d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a8362295a62f4b51bae719a7ef5a4656', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65726) authorize /opt/stack/nova/nova/policy.py:192}} [ 763.509615] env[65726]: DEBUG oslo_vmware.api [None req-9ff22eed-6c42-49bb-b683-3f83a9ff00ca tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Task: {'id': task-5115807, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.255233} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 763.510100] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-9ff22eed-6c42-49bb-b683-3f83a9ff00ca tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Deleted the datastore file {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 763.510384] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-9ff22eed-6c42-49bb-b683-3f83a9ff00ca tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] [instance: ee676e3b-3326-46a3-940d-ebbb8b108991] Deleted contents of the VM from datastore datastore2 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 763.510772] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-9ff22eed-6c42-49bb-b683-3f83a9ff00ca tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] [instance: ee676e3b-3326-46a3-940d-ebbb8b108991] Instance destroyed {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 763.511230] env[65726]: INFO nova.compute.manager [None req-9ff22eed-6c42-49bb-b683-3f83a9ff00ca tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] [instance: ee676e3b-3326-46a3-940d-ebbb8b108991] Took 1.13 seconds to destroy the instance on the hypervisor. [ 763.511384] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-9ff22eed-6c42-49bb-b683-3f83a9ff00ca tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 763.511566] env[65726]: DEBUG nova.compute.manager [-] [instance: ee676e3b-3326-46a3-940d-ebbb8b108991] Deallocating network for instance {{(pid=65726) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 763.511685] env[65726]: DEBUG nova.network.neutron [-] [instance: ee676e3b-3326-46a3-940d-ebbb8b108991] deallocate_for_instance() {{(pid=65726) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 763.512181] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 763.512439] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 763.570038] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 763.570038] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 763.583953] env[65726]: DEBUG nova.network.neutron [-] [instance: 02351120-8385-4403-8464-a154f3da9380] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 763.610873] env[65726]: DEBUG oslo_vmware.api [None req-04afb47c-692a-46d9-a640-68cc2c90206f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Task: {'id': task-5115804, 'name': PowerOnVM_Task, 'duration_secs': 0.585577} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 763.612025] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-04afb47c-692a-46d9-a640-68cc2c90206f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: c108f2a5-031b-47ee-9a5b-d62c2f42c26b] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 763.612025] env[65726]: INFO nova.compute.manager [None req-04afb47c-692a-46d9-a640-68cc2c90206f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: c108f2a5-031b-47ee-9a5b-d62c2f42c26b] Took 10.44 seconds to spawn the instance on the hypervisor. [ 763.612025] env[65726]: DEBUG nova.compute.manager [None req-04afb47c-692a-46d9-a640-68cc2c90206f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: c108f2a5-031b-47ee-9a5b-d62c2f42c26b] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 763.612283] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04f018b7-4b31-401a-aeb2-bae1215ab044 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.705289] env[65726]: DEBUG oslo_concurrency.lockutils [None req-af2f0c1d-9bc5-40c3-a712-d408e73bc806 tempest-ServerDiagnosticsNegativeTest-1635132145 tempest-ServerDiagnosticsNegativeTest-1635132145-project-member] Releasing lock "refresh_cache-0d6f14c5-5770-476f-a186-4cf1b2d4fbe8" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 763.705673] env[65726]: DEBUG nova.compute.manager [None req-af2f0c1d-9bc5-40c3-a712-d408e73bc806 tempest-ServerDiagnosticsNegativeTest-1635132145 tempest-ServerDiagnosticsNegativeTest-1635132145-project-member] [instance: 0d6f14c5-5770-476f-a186-4cf1b2d4fbe8] Instance network_info: |[{"id": "c8363d25-c607-4d04-ab53-412848d7cf4f", "address": "fa:16:3e:a8:64:02", "network": {"id": "55d629b8-a541-4720-9289-927bbc6e6f22", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.61", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "1539f06fa3534e90acbc3a60c4b8bd3f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "35342bcb-8b06-472e-b3c0-43fd3d6c4b30", "external-id": "nsx-vlan-transportzone-524", "segmentation_id": 524, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc8363d25-c6", "ovs_interfaceid": "c8363d25-c607-4d04-ab53-412848d7cf4f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 763.706174] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-af2f0c1d-9bc5-40c3-a712-d408e73bc806 tempest-ServerDiagnosticsNegativeTest-1635132145 tempest-ServerDiagnosticsNegativeTest-1635132145-project-member] [instance: 0d6f14c5-5770-476f-a186-4cf1b2d4fbe8] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a8:64:02', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '35342bcb-8b06-472e-b3c0-43fd3d6c4b30', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c8363d25-c607-4d04-ab53-412848d7cf4f', 'vif_model': 'vmxnet3'}] {{(pid=65726) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 763.714629] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-af2f0c1d-9bc5-40c3-a712-d408e73bc806 tempest-ServerDiagnosticsNegativeTest-1635132145 tempest-ServerDiagnosticsNegativeTest-1635132145-project-member] Creating folder: Project (877bbe4aeef145749b4156f8ab6f2202). Parent ref: group-v995008. {{(pid=65726) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 763.715085] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-8624e6f0-621a-4815-8ba1-3c5e0ae20632 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.728382] env[65726]: DEBUG oslo_vmware.api [None req-44e83fc7-afcc-4134-bbbb-6bc576931bb8 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Task: {'id': task-5115805, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 763.732334] env[65726]: DEBUG nova.network.neutron [None req-6c979061-2d03-4525-8b87-afa801f4542e tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: 6d2b86c2-0a44-433f-8b3f-2e9c7f693baa] Successfully created port: b487d53a-9bda-4e3c-be61-f26acbc8ae0f {{(pid=65726) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 763.737706] env[65726]: INFO nova.virt.vmwareapi.vm_util [None req-af2f0c1d-9bc5-40c3-a712-d408e73bc806 tempest-ServerDiagnosticsNegativeTest-1635132145 tempest-ServerDiagnosticsNegativeTest-1635132145-project-member] Created folder: Project (877bbe4aeef145749b4156f8ab6f2202) in parent group-v995008. [ 763.738092] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-af2f0c1d-9bc5-40c3-a712-d408e73bc806 tempest-ServerDiagnosticsNegativeTest-1635132145 tempest-ServerDiagnosticsNegativeTest-1635132145-project-member] Creating folder: Instances. Parent ref: group-v995116. {{(pid=65726) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 763.738374] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c8f2fa54-f0f4-482f-807e-2b0c0a2343cb {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.750394] env[65726]: INFO nova.virt.vmwareapi.vm_util [None req-af2f0c1d-9bc5-40c3-a712-d408e73bc806 tempest-ServerDiagnosticsNegativeTest-1635132145 tempest-ServerDiagnosticsNegativeTest-1635132145-project-member] Created folder: Instances in parent group-v995116. [ 763.750725] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-af2f0c1d-9bc5-40c3-a712-d408e73bc806 tempest-ServerDiagnosticsNegativeTest-1635132145 tempest-ServerDiagnosticsNegativeTest-1635132145-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 763.750890] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0d6f14c5-5770-476f-a186-4cf1b2d4fbe8] Creating VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 763.751194] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7bf5c1ae-6133-41c1-8b5a-7aeb6269d482 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.777420] env[65726]: WARNING openstack [req-9dd6fa57-ace0-4209-bdb9-4c08cdca9942 req-38afafcd-4a90-4ef5-bf59-4ccbb9af9716 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 763.777815] env[65726]: WARNING openstack [req-9dd6fa57-ace0-4209-bdb9-4c08cdca9942 req-38afafcd-4a90-4ef5-bf59-4ccbb9af9716 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 763.786789] env[65726]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 763.786789] env[65726]: value = "task-5115810" [ 763.786789] env[65726]: _type = "Task" [ 763.786789] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 763.799705] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115810, 'name': CreateVM_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 763.810251] env[65726]: WARNING openstack [None req-48153719-e51d-407a-98ef-93401c6270f2 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 763.811213] env[65726]: WARNING openstack [None req-48153719-e51d-407a-98ef-93401c6270f2 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 763.820996] env[65726]: DEBUG nova.network.neutron [None req-48153719-e51d-407a-98ef-93401c6270f2 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] [instance: 65aaa2eb-3ca5-4c2b-92f9-b524366c10dd] Instance cache missing network info. {{(pid=65726) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 763.911276] env[65726]: DEBUG nova.compute.manager [None req-6c979061-2d03-4525-8b87-afa801f4542e tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: 6d2b86c2-0a44-433f-8b3f-2e9c7f693baa] Start building block device mappings for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 763.985482] env[65726]: WARNING openstack [None req-48153719-e51d-407a-98ef-93401c6270f2 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 763.986031] env[65726]: WARNING openstack [None req-48153719-e51d-407a-98ef-93401c6270f2 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 764.018451] env[65726]: WARNING openstack [req-9dd6fa57-ace0-4209-bdb9-4c08cdca9942 req-38afafcd-4a90-4ef5-bf59-4ccbb9af9716 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 764.018451] env[65726]: WARNING openstack [req-9dd6fa57-ace0-4209-bdb9-4c08cdca9942 req-38afafcd-4a90-4ef5-bf59-4ccbb9af9716 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 764.087977] env[65726]: INFO nova.compute.manager [-] [instance: 02351120-8385-4403-8464-a154f3da9380] Took 2.66 seconds to deallocate network for instance. [ 764.141805] env[65726]: INFO nova.compute.manager [None req-04afb47c-692a-46d9-a640-68cc2c90206f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: c108f2a5-031b-47ee-9a5b-d62c2f42c26b] Took 58.50 seconds to build instance. [ 764.156101] env[65726]: WARNING openstack [None req-48153719-e51d-407a-98ef-93401c6270f2 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 764.156101] env[65726]: WARNING openstack [None req-48153719-e51d-407a-98ef-93401c6270f2 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 764.191574] env[65726]: WARNING openstack [req-9dd6fa57-ace0-4209-bdb9-4c08cdca9942 req-38afafcd-4a90-4ef5-bf59-4ccbb9af9716 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 764.192073] env[65726]: WARNING openstack [req-9dd6fa57-ace0-4209-bdb9-4c08cdca9942 req-38afafcd-4a90-4ef5-bf59-4ccbb9af9716 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 764.229680] env[65726]: DEBUG oslo_vmware.api [None req-44e83fc7-afcc-4134-bbbb-6bc576931bb8 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Task: {'id': task-5115805, 'name': PowerOnVM_Task, 'duration_secs': 1.114413} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 764.229986] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-44e83fc7-afcc-4134-bbbb-6bc576931bb8 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: 895797e4-2941-44cd-aab1-67afa4fac02d] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 764.230221] env[65726]: INFO nova.compute.manager [None req-44e83fc7-afcc-4134-bbbb-6bc576931bb8 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: 895797e4-2941-44cd-aab1-67afa4fac02d] Took 8.36 seconds to spawn the instance on the hypervisor. [ 764.230401] env[65726]: DEBUG nova.compute.manager [None req-44e83fc7-afcc-4134-bbbb-6bc576931bb8 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: 895797e4-2941-44cd-aab1-67afa4fac02d] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 764.231294] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4517fd91-8737-4b62-a081-24aec6ff36ec {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.308874] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115810, 'name': CreateVM_Task} progress is 99%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 764.322765] env[65726]: DEBUG nova.network.neutron [-] [instance: ee676e3b-3326-46a3-940d-ebbb8b108991] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 764.596487] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a78d739e-5974-49c3-a6f0-74e291b9e87b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 764.602168] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41d6e6cd-e772-4afe-ad0a-6bb41cf42923 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.611760] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25090313-651f-46fd-8b5c-dbef2b37f471 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.646532] env[65726]: DEBUG oslo_concurrency.lockutils [None req-04afb47c-692a-46d9-a640-68cc2c90206f tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Lock "c108f2a5-031b-47ee-9a5b-d62c2f42c26b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 83.528s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 764.653602] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91f114e0-1284-4d05-9d27-57c402d9b749 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.663640] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b15e8694-a6a1-4a44-b348-67bd1b74a245 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.679052] env[65726]: DEBUG nova.compute.provider_tree [None req-d51d38e8-71be-4889-a206-d52403737076 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 764.756921] env[65726]: INFO nova.compute.manager [None req-44e83fc7-afcc-4134-bbbb-6bc576931bb8 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: 895797e4-2941-44cd-aab1-67afa4fac02d] Took 47.48 seconds to build instance. [ 764.765606] env[65726]: DEBUG nova.network.neutron [None req-48153719-e51d-407a-98ef-93401c6270f2 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] [instance: 65aaa2eb-3ca5-4c2b-92f9-b524366c10dd] Updating instance_info_cache with network_info: [{"id": "862071bf-725c-4193-b60b-72b03b4c0553", "address": "fa:16:3e:9e:67:58", "network": {"id": "27801f9a-9912-41cf-be65-bc9903f4952f", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-781436951-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aac5fe6c99f740ada3747088b2b0ad89", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dd7d0d95-6848-4e69-ac21-75f8db82a3b5", "external-id": "nsx-vlan-transportzone-272", "segmentation_id": 272, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap862071bf-72", "ovs_interfaceid": "862071bf-725c-4193-b60b-72b03b4c0553", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 764.799096] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115810, 'name': CreateVM_Task, 'duration_secs': 0.565745} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 764.801628] env[65726]: DEBUG nova.network.neutron [req-9dd6fa57-ace0-4209-bdb9-4c08cdca9942 req-38afafcd-4a90-4ef5-bf59-4ccbb9af9716 service nova] [instance: c108f2a5-031b-47ee-9a5b-d62c2f42c26b] Updated VIF entry in instance network info cache for port 981741c6-69f0-42d4-87be-e1c8f62d0e17. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 764.802401] env[65726]: DEBUG nova.network.neutron [req-9dd6fa57-ace0-4209-bdb9-4c08cdca9942 req-38afafcd-4a90-4ef5-bf59-4ccbb9af9716 service nova] [instance: c108f2a5-031b-47ee-9a5b-d62c2f42c26b] Updating instance_info_cache with network_info: [{"id": "981741c6-69f0-42d4-87be-e1c8f62d0e17", "address": "fa:16:3e:ca:6f:8a", "network": {"id": "3f7916f7-ee9c-4629-b3b3-1367de9f98ea", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-579713617-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b794c99309e94c1181fc9d8eeeb84702", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d0bdd11b-58af-4cc0-9d38-8322e1bb4e74", "external-id": "nsx-vlan-transportzone-398", "segmentation_id": 398, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap981741c6-69", "ovs_interfaceid": "981741c6-69f0-42d4-87be-e1c8f62d0e17", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 764.803790] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0d6f14c5-5770-476f-a186-4cf1b2d4fbe8] Created VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 764.804859] env[65726]: WARNING openstack [None req-af2f0c1d-9bc5-40c3-a712-d408e73bc806 tempest-ServerDiagnosticsNegativeTest-1635132145 tempest-ServerDiagnosticsNegativeTest-1635132145-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 764.805770] env[65726]: WARNING openstack [None req-af2f0c1d-9bc5-40c3-a712-d408e73bc806 tempest-ServerDiagnosticsNegativeTest-1635132145 tempest-ServerDiagnosticsNegativeTest-1635132145-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 764.811426] env[65726]: DEBUG oslo_concurrency.lockutils [None req-af2f0c1d-9bc5-40c3-a712-d408e73bc806 tempest-ServerDiagnosticsNegativeTest-1635132145 tempest-ServerDiagnosticsNegativeTest-1635132145-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 764.811618] env[65726]: DEBUG oslo_concurrency.lockutils [None req-af2f0c1d-9bc5-40c3-a712-d408e73bc806 tempest-ServerDiagnosticsNegativeTest-1635132145 tempest-ServerDiagnosticsNegativeTest-1635132145-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 764.811957] env[65726]: DEBUG oslo_concurrency.lockutils [None req-af2f0c1d-9bc5-40c3-a712-d408e73bc806 tempest-ServerDiagnosticsNegativeTest-1635132145 tempest-ServerDiagnosticsNegativeTest-1635132145-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 764.812295] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e1e272fc-2dd7-4f2f-af34-ce75f2333c88 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.820054] env[65726]: DEBUG oslo_vmware.api [None req-af2f0c1d-9bc5-40c3-a712-d408e73bc806 tempest-ServerDiagnosticsNegativeTest-1635132145 tempest-ServerDiagnosticsNegativeTest-1635132145-project-member] Waiting for the task: (returnval){ [ 764.820054] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52c5cbfe-e011-9626-1955-41070618c36f" [ 764.820054] env[65726]: _type = "Task" [ 764.820054] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 764.832835] env[65726]: INFO nova.compute.manager [-] [instance: ee676e3b-3326-46a3-940d-ebbb8b108991] Took 1.32 seconds to deallocate network for instance. [ 764.832835] env[65726]: DEBUG oslo_vmware.api [None req-af2f0c1d-9bc5-40c3-a712-d408e73bc806 tempest-ServerDiagnosticsNegativeTest-1635132145 tempest-ServerDiagnosticsNegativeTest-1635132145-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52c5cbfe-e011-9626-1955-41070618c36f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 764.931399] env[65726]: DEBUG nova.compute.manager [None req-6c979061-2d03-4525-8b87-afa801f4542e tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: 6d2b86c2-0a44-433f-8b3f-2e9c7f693baa] Start spawning the instance on the hypervisor. {{(pid=65726) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 764.969635] env[65726]: DEBUG nova.virt.hardware [None req-6c979061-2d03-4525-8b87-afa801f4542e tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T19:28:34Z,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1539f06fa3534e90acbc3a60c4b8bd3f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T19:28:35Z,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 764.969971] env[65726]: DEBUG nova.virt.hardware [None req-6c979061-2d03-4525-8b87-afa801f4542e tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 764.970143] env[65726]: DEBUG nova.virt.hardware [None req-6c979061-2d03-4525-8b87-afa801f4542e tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 764.970366] env[65726]: DEBUG nova.virt.hardware [None req-6c979061-2d03-4525-8b87-afa801f4542e tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 764.970512] env[65726]: DEBUG nova.virt.hardware [None req-6c979061-2d03-4525-8b87-afa801f4542e tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 764.970660] env[65726]: DEBUG nova.virt.hardware [None req-6c979061-2d03-4525-8b87-afa801f4542e tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 764.970893] env[65726]: DEBUG nova.virt.hardware [None req-6c979061-2d03-4525-8b87-afa801f4542e tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 764.971058] env[65726]: DEBUG nova.virt.hardware [None req-6c979061-2d03-4525-8b87-afa801f4542e tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 764.971227] env[65726]: DEBUG nova.virt.hardware [None req-6c979061-2d03-4525-8b87-afa801f4542e tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 764.971379] env[65726]: DEBUG nova.virt.hardware [None req-6c979061-2d03-4525-8b87-afa801f4542e tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 764.971546] env[65726]: DEBUG nova.virt.hardware [None req-6c979061-2d03-4525-8b87-afa801f4542e tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 764.972552] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-420ed040-027c-442b-95aa-b40bfc98f186 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.981299] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2879c6b6-3fba-4311-b588-0f26ece4a389 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.001240] env[65726]: DEBUG oslo_concurrency.lockutils [None req-ebfc0e41-4845-41ca-9f74-d2a2946626c0 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Acquiring lock "1cfb1bba-1e21-47eb-b50f-a86575846a65" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 765.001478] env[65726]: DEBUG oslo_concurrency.lockutils [None req-ebfc0e41-4845-41ca-9f74-d2a2946626c0 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Lock "1cfb1bba-1e21-47eb-b50f-a86575846a65" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 765.158985] env[65726]: DEBUG nova.compute.manager [None req-a76cfabd-af7f-4569-b968-716d45bcf7f1 tempest-InstanceActionsNegativeTestJSON-844675881 tempest-InstanceActionsNegativeTestJSON-844675881-project-member] [instance: 1bd26aef-995e-43b6-af9f-077e878c8c44] Starting instance... {{(pid=65726) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 765.189026] env[65726]: DEBUG nova.scheduler.client.report [None req-d51d38e8-71be-4889-a206-d52403737076 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 765.259270] env[65726]: DEBUG oslo_concurrency.lockutils [None req-44e83fc7-afcc-4134-bbbb-6bc576931bb8 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Lock "895797e4-2941-44cd-aab1-67afa4fac02d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 84.117s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 765.268233] env[65726]: DEBUG oslo_concurrency.lockutils [None req-48153719-e51d-407a-98ef-93401c6270f2 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Releasing lock "refresh_cache-65aaa2eb-3ca5-4c2b-92f9-b524366c10dd" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 765.268663] env[65726]: DEBUG nova.compute.manager [None req-48153719-e51d-407a-98ef-93401c6270f2 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] [instance: 65aaa2eb-3ca5-4c2b-92f9-b524366c10dd] Instance network_info: |[{"id": "862071bf-725c-4193-b60b-72b03b4c0553", "address": "fa:16:3e:9e:67:58", "network": {"id": "27801f9a-9912-41cf-be65-bc9903f4952f", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-781436951-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aac5fe6c99f740ada3747088b2b0ad89", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dd7d0d95-6848-4e69-ac21-75f8db82a3b5", "external-id": "nsx-vlan-transportzone-272", "segmentation_id": 272, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap862071bf-72", "ovs_interfaceid": "862071bf-725c-4193-b60b-72b03b4c0553", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 765.269941] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-48153719-e51d-407a-98ef-93401c6270f2 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] [instance: 65aaa2eb-3ca5-4c2b-92f9-b524366c10dd] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:9e:67:58', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'dd7d0d95-6848-4e69-ac21-75f8db82a3b5', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '862071bf-725c-4193-b60b-72b03b4c0553', 'vif_model': 'vmxnet3'}] {{(pid=65726) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 765.278559] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-48153719-e51d-407a-98ef-93401c6270f2 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 765.279557] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 65aaa2eb-3ca5-4c2b-92f9-b524366c10dd] Creating VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 765.279899] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-24b4ad66-f5b2-4f9b-b776-6ec50cbd3081 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.304788] env[65726]: DEBUG oslo_concurrency.lockutils [req-9dd6fa57-ace0-4209-bdb9-4c08cdca9942 req-38afafcd-4a90-4ef5-bf59-4ccbb9af9716 service nova] Releasing lock "refresh_cache-c108f2a5-031b-47ee-9a5b-d62c2f42c26b" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 765.305108] env[65726]: DEBUG nova.compute.manager [req-9dd6fa57-ace0-4209-bdb9-4c08cdca9942 req-38afafcd-4a90-4ef5-bf59-4ccbb9af9716 service nova] [instance: 895797e4-2941-44cd-aab1-67afa4fac02d] Received event network-vif-plugged-a1dec2db-0e1e-45dd-8587-a84e4a5e2ef4 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 765.305270] env[65726]: DEBUG oslo_concurrency.lockutils [req-9dd6fa57-ace0-4209-bdb9-4c08cdca9942 req-38afafcd-4a90-4ef5-bf59-4ccbb9af9716 service nova] Acquiring lock "895797e4-2941-44cd-aab1-67afa4fac02d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 765.305473] env[65726]: DEBUG oslo_concurrency.lockutils [req-9dd6fa57-ace0-4209-bdb9-4c08cdca9942 req-38afafcd-4a90-4ef5-bf59-4ccbb9af9716 service nova] Lock "895797e4-2941-44cd-aab1-67afa4fac02d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 765.305626] env[65726]: DEBUG oslo_concurrency.lockutils [req-9dd6fa57-ace0-4209-bdb9-4c08cdca9942 req-38afafcd-4a90-4ef5-bf59-4ccbb9af9716 service nova] Lock "895797e4-2941-44cd-aab1-67afa4fac02d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 765.305785] env[65726]: DEBUG nova.compute.manager [req-9dd6fa57-ace0-4209-bdb9-4c08cdca9942 req-38afafcd-4a90-4ef5-bf59-4ccbb9af9716 service nova] [instance: 895797e4-2941-44cd-aab1-67afa4fac02d] No waiting events found dispatching network-vif-plugged-a1dec2db-0e1e-45dd-8587-a84e4a5e2ef4 {{(pid=65726) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 765.306040] env[65726]: WARNING nova.compute.manager [req-9dd6fa57-ace0-4209-bdb9-4c08cdca9942 req-38afafcd-4a90-4ef5-bf59-4ccbb9af9716 service nova] [instance: 895797e4-2941-44cd-aab1-67afa4fac02d] Received unexpected event network-vif-plugged-a1dec2db-0e1e-45dd-8587-a84e4a5e2ef4 for instance with vm_state building and task_state spawning. [ 765.306239] env[65726]: DEBUG nova.compute.manager [req-9dd6fa57-ace0-4209-bdb9-4c08cdca9942 req-38afafcd-4a90-4ef5-bf59-4ccbb9af9716 service nova] [instance: 895797e4-2941-44cd-aab1-67afa4fac02d] Received event network-changed-a1dec2db-0e1e-45dd-8587-a84e4a5e2ef4 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 765.306420] env[65726]: DEBUG nova.compute.manager [req-9dd6fa57-ace0-4209-bdb9-4c08cdca9942 req-38afafcd-4a90-4ef5-bf59-4ccbb9af9716 service nova] [instance: 895797e4-2941-44cd-aab1-67afa4fac02d] Refreshing instance network info cache due to event network-changed-a1dec2db-0e1e-45dd-8587-a84e4a5e2ef4. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 765.306621] env[65726]: DEBUG oslo_concurrency.lockutils [req-9dd6fa57-ace0-4209-bdb9-4c08cdca9942 req-38afafcd-4a90-4ef5-bf59-4ccbb9af9716 service nova] Acquiring lock "refresh_cache-895797e4-2941-44cd-aab1-67afa4fac02d" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 765.306780] env[65726]: DEBUG oslo_concurrency.lockutils [req-9dd6fa57-ace0-4209-bdb9-4c08cdca9942 req-38afafcd-4a90-4ef5-bf59-4ccbb9af9716 service nova] Acquired lock "refresh_cache-895797e4-2941-44cd-aab1-67afa4fac02d" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 765.307095] env[65726]: DEBUG nova.network.neutron [req-9dd6fa57-ace0-4209-bdb9-4c08cdca9942 req-38afafcd-4a90-4ef5-bf59-4ccbb9af9716 service nova] [instance: 895797e4-2941-44cd-aab1-67afa4fac02d] Refreshing network info cache for port a1dec2db-0e1e-45dd-8587-a84e4a5e2ef4 {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 765.310056] env[65726]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 765.310056] env[65726]: value = "task-5115811" [ 765.310056] env[65726]: _type = "Task" [ 765.310056] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 765.320589] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115811, 'name': CreateVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 765.330565] env[65726]: DEBUG oslo_vmware.api [None req-af2f0c1d-9bc5-40c3-a712-d408e73bc806 tempest-ServerDiagnosticsNegativeTest-1635132145 tempest-ServerDiagnosticsNegativeTest-1635132145-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52c5cbfe-e011-9626-1955-41070618c36f, 'name': SearchDatastore_Task, 'duration_secs': 0.03104} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 765.330942] env[65726]: DEBUG oslo_concurrency.lockutils [None req-af2f0c1d-9bc5-40c3-a712-d408e73bc806 tempest-ServerDiagnosticsNegativeTest-1635132145 tempest-ServerDiagnosticsNegativeTest-1635132145-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 765.331225] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-af2f0c1d-9bc5-40c3-a712-d408e73bc806 tempest-ServerDiagnosticsNegativeTest-1635132145 tempest-ServerDiagnosticsNegativeTest-1635132145-project-member] [instance: 0d6f14c5-5770-476f-a186-4cf1b2d4fbe8] Processing image b52362a3-ee8a-4cbf-b06f-513b0cc8f95c {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 765.331595] env[65726]: DEBUG oslo_concurrency.lockutils [None req-af2f0c1d-9bc5-40c3-a712-d408e73bc806 tempest-ServerDiagnosticsNegativeTest-1635132145 tempest-ServerDiagnosticsNegativeTest-1635132145-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 765.331595] env[65726]: DEBUG oslo_concurrency.lockutils [None req-af2f0c1d-9bc5-40c3-a712-d408e73bc806 tempest-ServerDiagnosticsNegativeTest-1635132145 tempest-ServerDiagnosticsNegativeTest-1635132145-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 765.332274] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-af2f0c1d-9bc5-40c3-a712-d408e73bc806 tempest-ServerDiagnosticsNegativeTest-1635132145 tempest-ServerDiagnosticsNegativeTest-1635132145-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 765.332274] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-aed0c9c2-d977-4772-a8b2-e3dd8e2d086a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.340183] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9ff22eed-6c42-49bb-b683-3f83a9ff00ca tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 765.344209] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-af2f0c1d-9bc5-40c3-a712-d408e73bc806 tempest-ServerDiagnosticsNegativeTest-1635132145 tempest-ServerDiagnosticsNegativeTest-1635132145-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 765.344408] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-af2f0c1d-9bc5-40c3-a712-d408e73bc806 tempest-ServerDiagnosticsNegativeTest-1635132145 tempest-ServerDiagnosticsNegativeTest-1635132145-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65726) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 765.345237] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e6139b4c-c1be-4d04-b04d-7d0563443e1c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.353179] env[65726]: DEBUG oslo_vmware.api [None req-af2f0c1d-9bc5-40c3-a712-d408e73bc806 tempest-ServerDiagnosticsNegativeTest-1635132145 tempest-ServerDiagnosticsNegativeTest-1635132145-project-member] Waiting for the task: (returnval){ [ 765.353179] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52ebcbbe-64d1-102d-9526-3fc9c31fde81" [ 765.353179] env[65726]: _type = "Task" [ 765.353179] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 765.362999] env[65726]: DEBUG oslo_vmware.api [None req-af2f0c1d-9bc5-40c3-a712-d408e73bc806 tempest-ServerDiagnosticsNegativeTest-1635132145 tempest-ServerDiagnosticsNegativeTest-1635132145-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52ebcbbe-64d1-102d-9526-3fc9c31fde81, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 765.409403] env[65726]: DEBUG nova.network.neutron [None req-6c979061-2d03-4525-8b87-afa801f4542e tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: 6d2b86c2-0a44-433f-8b3f-2e9c7f693baa] Successfully updated port: b487d53a-9bda-4e3c-be61-f26acbc8ae0f {{(pid=65726) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 765.685842] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a76cfabd-af7f-4569-b968-716d45bcf7f1 tempest-InstanceActionsNegativeTestJSON-844675881 tempest-InstanceActionsNegativeTestJSON-844675881-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 765.694520] env[65726]: DEBUG oslo_concurrency.lockutils [None req-d51d38e8-71be-4889-a206-d52403737076 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.811s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 765.695099] env[65726]: DEBUG nova.compute.manager [None req-d51d38e8-71be-4889-a206-d52403737076 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: 9fc0c9ee-9dc7-4059-b8d2-a5db965e00c2] Start building networks asynchronously for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 765.698315] env[65726]: DEBUG oslo_concurrency.lockutils [None req-cda5c6a3-fc21-4556-afc4-6ebd218aa8a5 tempest-ServerExternalEventsTest-1845125219 tempest-ServerExternalEventsTest-1845125219-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 32.714s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 765.698315] env[65726]: DEBUG nova.objects.instance [None req-cda5c6a3-fc21-4556-afc4-6ebd218aa8a5 tempest-ServerExternalEventsTest-1845125219 tempest-ServerExternalEventsTest-1845125219-project-member] Lazy-loading 'resources' on Instance uuid 96acc5f8-b411-408e-929d-d1035cfd50db {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 765.762594] env[65726]: DEBUG nova.compute.manager [None req-855996fc-ecea-49d8-a71b-8b4307731732 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] [instance: 6c8db442-daaa-4eca-ae24-b9d4afaf3a71] Starting instance... {{(pid=65726) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 765.812579] env[65726]: WARNING openstack [req-9dd6fa57-ace0-4209-bdb9-4c08cdca9942 req-38afafcd-4a90-4ef5-bf59-4ccbb9af9716 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 765.812694] env[65726]: WARNING openstack [req-9dd6fa57-ace0-4209-bdb9-4c08cdca9942 req-38afafcd-4a90-4ef5-bf59-4ccbb9af9716 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 765.830144] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115811, 'name': CreateVM_Task, 'duration_secs': 0.521914} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 765.830485] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 65aaa2eb-3ca5-4c2b-92f9-b524366c10dd] Created VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 765.831170] env[65726]: WARNING openstack [None req-48153719-e51d-407a-98ef-93401c6270f2 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 765.831537] env[65726]: WARNING openstack [None req-48153719-e51d-407a-98ef-93401c6270f2 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 765.837202] env[65726]: DEBUG oslo_concurrency.lockutils [None req-48153719-e51d-407a-98ef-93401c6270f2 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 765.838750] env[65726]: DEBUG oslo_concurrency.lockutils [None req-48153719-e51d-407a-98ef-93401c6270f2 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 765.838750] env[65726]: DEBUG oslo_concurrency.lockutils [None req-48153719-e51d-407a-98ef-93401c6270f2 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 765.838750] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a3ffcf6a-000e-40a6-8f55-bb6594223194 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.844730] env[65726]: DEBUG oslo_vmware.api [None req-48153719-e51d-407a-98ef-93401c6270f2 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Waiting for the task: (returnval){ [ 765.844730] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52b0f9d4-5a1a-52da-2b9d-c21ca35de2ba" [ 765.844730] env[65726]: _type = "Task" [ 765.844730] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 765.854468] env[65726]: DEBUG oslo_vmware.api [None req-48153719-e51d-407a-98ef-93401c6270f2 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52b0f9d4-5a1a-52da-2b9d-c21ca35de2ba, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 765.870725] env[65726]: DEBUG oslo_vmware.api [None req-af2f0c1d-9bc5-40c3-a712-d408e73bc806 tempest-ServerDiagnosticsNegativeTest-1635132145 tempest-ServerDiagnosticsNegativeTest-1635132145-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52ebcbbe-64d1-102d-9526-3fc9c31fde81, 'name': SearchDatastore_Task, 'duration_secs': 0.014939} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 765.871691] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-93e31b58-a53e-420b-bd3c-5c2efa78d966 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.878080] env[65726]: DEBUG oslo_vmware.api [None req-af2f0c1d-9bc5-40c3-a712-d408e73bc806 tempest-ServerDiagnosticsNegativeTest-1635132145 tempest-ServerDiagnosticsNegativeTest-1635132145-project-member] Waiting for the task: (returnval){ [ 765.878080] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52e9db48-3ae7-ad7e-9b3a-385a10646b53" [ 765.878080] env[65726]: _type = "Task" [ 765.878080] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 765.889220] env[65726]: DEBUG oslo_vmware.api [None req-af2f0c1d-9bc5-40c3-a712-d408e73bc806 tempest-ServerDiagnosticsNegativeTest-1635132145 tempest-ServerDiagnosticsNegativeTest-1635132145-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52e9db48-3ae7-ad7e-9b3a-385a10646b53, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 765.916021] env[65726]: DEBUG oslo_concurrency.lockutils [None req-6c979061-2d03-4525-8b87-afa801f4542e tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Acquiring lock "refresh_cache-6d2b86c2-0a44-433f-8b3f-2e9c7f693baa" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 765.916021] env[65726]: DEBUG oslo_concurrency.lockutils [None req-6c979061-2d03-4525-8b87-afa801f4542e tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Acquired lock "refresh_cache-6d2b86c2-0a44-433f-8b3f-2e9c7f693baa" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 765.916021] env[65726]: DEBUG nova.network.neutron [None req-6c979061-2d03-4525-8b87-afa801f4542e tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: 6d2b86c2-0a44-433f-8b3f-2e9c7f693baa] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 765.934042] env[65726]: WARNING openstack [req-9dd6fa57-ace0-4209-bdb9-4c08cdca9942 req-38afafcd-4a90-4ef5-bf59-4ccbb9af9716 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 765.934042] env[65726]: WARNING openstack [req-9dd6fa57-ace0-4209-bdb9-4c08cdca9942 req-38afafcd-4a90-4ef5-bf59-4ccbb9af9716 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 766.003969] env[65726]: WARNING openstack [req-9dd6fa57-ace0-4209-bdb9-4c08cdca9942 req-38afafcd-4a90-4ef5-bf59-4ccbb9af9716 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 766.004774] env[65726]: WARNING openstack [req-9dd6fa57-ace0-4209-bdb9-4c08cdca9942 req-38afafcd-4a90-4ef5-bf59-4ccbb9af9716 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 766.101091] env[65726]: DEBUG nova.network.neutron [req-9dd6fa57-ace0-4209-bdb9-4c08cdca9942 req-38afafcd-4a90-4ef5-bf59-4ccbb9af9716 service nova] [instance: 895797e4-2941-44cd-aab1-67afa4fac02d] Updated VIF entry in instance network info cache for port a1dec2db-0e1e-45dd-8587-a84e4a5e2ef4. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 766.101487] env[65726]: DEBUG nova.network.neutron [req-9dd6fa57-ace0-4209-bdb9-4c08cdca9942 req-38afafcd-4a90-4ef5-bf59-4ccbb9af9716 service nova] [instance: 895797e4-2941-44cd-aab1-67afa4fac02d] Updating instance_info_cache with network_info: [{"id": "a1dec2db-0e1e-45dd-8587-a84e4a5e2ef4", "address": "fa:16:3e:b9:47:02", "network": {"id": "ad1373b5-e0ed-4ee6-a97c-af931afbab35", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-909418417-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a090d553766847e58e3231b966c92565", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5f60c972-a72d-4c5f-a250-faadfd6eafbe", "external-id": "nsx-vlan-transportzone-932", "segmentation_id": 932, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa1dec2db-0e", "ovs_interfaceid": "a1dec2db-0e1e-45dd-8587-a84e4a5e2ef4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 766.204208] env[65726]: DEBUG nova.compute.utils [None req-d51d38e8-71be-4889-a206-d52403737076 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Using /dev/sd instead of None {{(pid=65726) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 766.207387] env[65726]: DEBUG nova.compute.manager [None req-d51d38e8-71be-4889-a206-d52403737076 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: 9fc0c9ee-9dc7-4059-b8d2-a5db965e00c2] Allocating IP information in the background. {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 766.207387] env[65726]: DEBUG nova.network.neutron [None req-d51d38e8-71be-4889-a206-d52403737076 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: 9fc0c9ee-9dc7-4059-b8d2-a5db965e00c2] allocate_for_instance() {{(pid=65726) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 766.207387] env[65726]: WARNING neutronclient.v2_0.client [None req-d51d38e8-71be-4889-a206-d52403737076 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 766.207508] env[65726]: WARNING neutronclient.v2_0.client [None req-d51d38e8-71be-4889-a206-d52403737076 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 766.208281] env[65726]: WARNING openstack [None req-d51d38e8-71be-4889-a206-d52403737076 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 766.208638] env[65726]: WARNING openstack [None req-d51d38e8-71be-4889-a206-d52403737076 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 766.216938] env[65726]: DEBUG nova.compute.manager [None req-d51d38e8-71be-4889-a206-d52403737076 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: 9fc0c9ee-9dc7-4059-b8d2-a5db965e00c2] Start building block device mappings for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 766.276223] env[65726]: DEBUG nova.policy [None req-d51d38e8-71be-4889-a206-d52403737076 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'faf82c34c4884dc890edf0909b9f510d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a8362295a62f4b51bae719a7ef5a4656', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65726) authorize /opt/stack/nova/nova/policy.py:192}} [ 766.296404] env[65726]: DEBUG oslo_concurrency.lockutils [None req-855996fc-ecea-49d8-a71b-8b4307731732 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 766.359169] env[65726]: DEBUG oslo_vmware.api [None req-48153719-e51d-407a-98ef-93401c6270f2 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52b0f9d4-5a1a-52da-2b9d-c21ca35de2ba, 'name': SearchDatastore_Task, 'duration_secs': 0.012944} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 766.359558] env[65726]: DEBUG oslo_concurrency.lockutils [None req-48153719-e51d-407a-98ef-93401c6270f2 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 766.359761] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-48153719-e51d-407a-98ef-93401c6270f2 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] [instance: 65aaa2eb-3ca5-4c2b-92f9-b524366c10dd] Processing image b52362a3-ee8a-4cbf-b06f-513b0cc8f95c {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 766.360016] env[65726]: DEBUG oslo_concurrency.lockutils [None req-48153719-e51d-407a-98ef-93401c6270f2 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 766.392460] env[65726]: DEBUG oslo_vmware.api [None req-af2f0c1d-9bc5-40c3-a712-d408e73bc806 tempest-ServerDiagnosticsNegativeTest-1635132145 tempest-ServerDiagnosticsNegativeTest-1635132145-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52e9db48-3ae7-ad7e-9b3a-385a10646b53, 'name': SearchDatastore_Task, 'duration_secs': 0.014442} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 766.392615] env[65726]: DEBUG oslo_concurrency.lockutils [None req-af2f0c1d-9bc5-40c3-a712-d408e73bc806 tempest-ServerDiagnosticsNegativeTest-1635132145 tempest-ServerDiagnosticsNegativeTest-1635132145-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 766.393195] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-af2f0c1d-9bc5-40c3-a712-d408e73bc806 tempest-ServerDiagnosticsNegativeTest-1635132145 tempest-ServerDiagnosticsNegativeTest-1635132145-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore2] 0d6f14c5-5770-476f-a186-4cf1b2d4fbe8/0d6f14c5-5770-476f-a186-4cf1b2d4fbe8.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 766.393195] env[65726]: DEBUG oslo_concurrency.lockutils [None req-48153719-e51d-407a-98ef-93401c6270f2 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 766.393384] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-48153719-e51d-407a-98ef-93401c6270f2 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 766.393525] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-00071b7b-5012-4c7d-9864-d830bcf5454a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.395794] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3c426ef7-9645-427c-9d77-1d0c289ee23b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.406327] env[65726]: DEBUG oslo_vmware.api [None req-af2f0c1d-9bc5-40c3-a712-d408e73bc806 tempest-ServerDiagnosticsNegativeTest-1635132145 tempest-ServerDiagnosticsNegativeTest-1635132145-project-member] Waiting for the task: (returnval){ [ 766.406327] env[65726]: value = "task-5115812" [ 766.406327] env[65726]: _type = "Task" [ 766.406327] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 766.411645] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-48153719-e51d-407a-98ef-93401c6270f2 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 766.411645] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-48153719-e51d-407a-98ef-93401c6270f2 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65726) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 766.415348] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-16878ec5-7299-4517-83a8-5c56a5a2335e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.418031] env[65726]: DEBUG oslo_vmware.api [None req-af2f0c1d-9bc5-40c3-a712-d408e73bc806 tempest-ServerDiagnosticsNegativeTest-1635132145 tempest-ServerDiagnosticsNegativeTest-1635132145-project-member] Task: {'id': task-5115812, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 766.422251] env[65726]: WARNING openstack [None req-6c979061-2d03-4525-8b87-afa801f4542e tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 766.422251] env[65726]: WARNING openstack [None req-6c979061-2d03-4525-8b87-afa801f4542e tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 766.428602] env[65726]: DEBUG nova.network.neutron [None req-6c979061-2d03-4525-8b87-afa801f4542e tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: 6d2b86c2-0a44-433f-8b3f-2e9c7f693baa] Instance cache missing network info. {{(pid=65726) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 766.437109] env[65726]: DEBUG oslo_vmware.api [None req-48153719-e51d-407a-98ef-93401c6270f2 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Waiting for the task: (returnval){ [ 766.437109] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]528148e5-216d-49ce-8d8e-c13a2fc16f00" [ 766.437109] env[65726]: _type = "Task" [ 766.437109] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 766.452513] env[65726]: DEBUG oslo_vmware.api [None req-48153719-e51d-407a-98ef-93401c6270f2 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]528148e5-216d-49ce-8d8e-c13a2fc16f00, 'name': SearchDatastore_Task, 'duration_secs': 0.013104} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 766.452714] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-74c129d8-c3b9-410f-bda2-a3a33f9dba66 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.463312] env[65726]: DEBUG oslo_vmware.api [None req-48153719-e51d-407a-98ef-93401c6270f2 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Waiting for the task: (returnval){ [ 766.463312] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5255e07b-28ea-9540-d5d3-5c1291b82b42" [ 766.463312] env[65726]: _type = "Task" [ 766.463312] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 766.479619] env[65726]: DEBUG oslo_vmware.api [None req-48153719-e51d-407a-98ef-93401c6270f2 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5255e07b-28ea-9540-d5d3-5c1291b82b42, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 766.517326] env[65726]: WARNING openstack [None req-6c979061-2d03-4525-8b87-afa801f4542e tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 766.518014] env[65726]: WARNING openstack [None req-6c979061-2d03-4525-8b87-afa801f4542e tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 766.604619] env[65726]: DEBUG oslo_concurrency.lockutils [req-9dd6fa57-ace0-4209-bdb9-4c08cdca9942 req-38afafcd-4a90-4ef5-bf59-4ccbb9af9716 service nova] Releasing lock "refresh_cache-895797e4-2941-44cd-aab1-67afa4fac02d" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 766.610739] env[65726]: WARNING openstack [None req-6c979061-2d03-4525-8b87-afa801f4542e tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 766.613171] env[65726]: WARNING openstack [None req-6c979061-2d03-4525-8b87-afa801f4542e tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 766.649522] env[65726]: DEBUG nova.network.neutron [None req-d51d38e8-71be-4889-a206-d52403737076 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: 9fc0c9ee-9dc7-4059-b8d2-a5db965e00c2] Successfully created port: e4315658-0732-463a-b063-b1c4cf7cdafd {{(pid=65726) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 766.662376] env[65726]: DEBUG nova.compute.manager [req-57a1ccc9-eb29-4ce1-aa59-ef1da805887f req-4983a4fb-f280-4f1f-bd93-a5414d986d2f service nova] [instance: 0d6f14c5-5770-476f-a186-4cf1b2d4fbe8] Received event network-vif-plugged-c8363d25-c607-4d04-ab53-412848d7cf4f {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 766.662596] env[65726]: DEBUG oslo_concurrency.lockutils [req-57a1ccc9-eb29-4ce1-aa59-ef1da805887f req-4983a4fb-f280-4f1f-bd93-a5414d986d2f service nova] Acquiring lock "0d6f14c5-5770-476f-a186-4cf1b2d4fbe8-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 766.663412] env[65726]: DEBUG oslo_concurrency.lockutils [req-57a1ccc9-eb29-4ce1-aa59-ef1da805887f req-4983a4fb-f280-4f1f-bd93-a5414d986d2f service nova] Lock "0d6f14c5-5770-476f-a186-4cf1b2d4fbe8-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 766.663412] env[65726]: DEBUG oslo_concurrency.lockutils [req-57a1ccc9-eb29-4ce1-aa59-ef1da805887f req-4983a4fb-f280-4f1f-bd93-a5414d986d2f service nova] Lock "0d6f14c5-5770-476f-a186-4cf1b2d4fbe8-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 766.663412] env[65726]: DEBUG nova.compute.manager [req-57a1ccc9-eb29-4ce1-aa59-ef1da805887f req-4983a4fb-f280-4f1f-bd93-a5414d986d2f service nova] [instance: 0d6f14c5-5770-476f-a186-4cf1b2d4fbe8] No waiting events found dispatching network-vif-plugged-c8363d25-c607-4d04-ab53-412848d7cf4f {{(pid=65726) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 766.663412] env[65726]: WARNING nova.compute.manager [req-57a1ccc9-eb29-4ce1-aa59-ef1da805887f req-4983a4fb-f280-4f1f-bd93-a5414d986d2f service nova] [instance: 0d6f14c5-5770-476f-a186-4cf1b2d4fbe8] Received unexpected event network-vif-plugged-c8363d25-c607-4d04-ab53-412848d7cf4f for instance with vm_state building and task_state spawning. [ 766.663600] env[65726]: DEBUG nova.compute.manager [req-57a1ccc9-eb29-4ce1-aa59-ef1da805887f req-4983a4fb-f280-4f1f-bd93-a5414d986d2f service nova] [instance: 0d6f14c5-5770-476f-a186-4cf1b2d4fbe8] Received event network-changed-c8363d25-c607-4d04-ab53-412848d7cf4f {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 766.663638] env[65726]: DEBUG nova.compute.manager [req-57a1ccc9-eb29-4ce1-aa59-ef1da805887f req-4983a4fb-f280-4f1f-bd93-a5414d986d2f service nova] [instance: 0d6f14c5-5770-476f-a186-4cf1b2d4fbe8] Refreshing instance network info cache due to event network-changed-c8363d25-c607-4d04-ab53-412848d7cf4f. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 766.664093] env[65726]: DEBUG oslo_concurrency.lockutils [req-57a1ccc9-eb29-4ce1-aa59-ef1da805887f req-4983a4fb-f280-4f1f-bd93-a5414d986d2f service nova] Acquiring lock "refresh_cache-0d6f14c5-5770-476f-a186-4cf1b2d4fbe8" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 766.664233] env[65726]: DEBUG oslo_concurrency.lockutils [req-57a1ccc9-eb29-4ce1-aa59-ef1da805887f req-4983a4fb-f280-4f1f-bd93-a5414d986d2f service nova] Acquired lock "refresh_cache-0d6f14c5-5770-476f-a186-4cf1b2d4fbe8" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 766.664395] env[65726]: DEBUG nova.network.neutron [req-57a1ccc9-eb29-4ce1-aa59-ef1da805887f req-4983a4fb-f280-4f1f-bd93-a5414d986d2f service nova] [instance: 0d6f14c5-5770-476f-a186-4cf1b2d4fbe8] Refreshing network info cache for port c8363d25-c607-4d04-ab53-412848d7cf4f {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 766.696374] env[65726]: DEBUG nova.compute.manager [req-263a0dbd-e9f8-43e6-bd76-20cca5a5b6bf req-c8cffdb2-7231-4aa7-84e0-df6e3779fa3e service nova] [instance: ca67d1a4-d304-4d06-a436-f257b60ca080] Received event network-vif-deleted-8789d2e9-5db3-4e5d-bdb2-14160bd43a1d {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 766.696635] env[65726]: DEBUG nova.compute.manager [req-263a0dbd-e9f8-43e6-bd76-20cca5a5b6bf req-c8cffdb2-7231-4aa7-84e0-df6e3779fa3e service nova] [instance: ee676e3b-3326-46a3-940d-ebbb8b108991] Received event network-vif-deleted-30651f8a-abdc-487b-9044-1c8a17beab23 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 766.761566] env[65726]: DEBUG nova.network.neutron [None req-6c979061-2d03-4525-8b87-afa801f4542e tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: 6d2b86c2-0a44-433f-8b3f-2e9c7f693baa] Updating instance_info_cache with network_info: [{"id": "b487d53a-9bda-4e3c-be61-f26acbc8ae0f", "address": "fa:16:3e:ad:db:36", "network": {"id": "9d5ad167-99fb-4208-ab30-7053f35774e4", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-2082516762-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a8362295a62f4b51bae719a7ef5a4656", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dabbac20-1723-40ad-9da0-e53b28073651", "external-id": "nsx-vlan-transportzone-790", "segmentation_id": 790, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb487d53a-9b", "ovs_interfaceid": "b487d53a-9bda-4e3c-be61-f26acbc8ae0f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 766.870238] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32bb5ee7-4555-4387-b473-62610b1d44ed {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.880821] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-091e52ba-e873-487c-8018-1588ac6ebbec {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.919308] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0870a428-4bb1-482e-a2a2-3ebbe307876e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.932300] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-120e950c-6c33-4edb-b1a9-983abf69a0b1 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.936917] env[65726]: DEBUG oslo_vmware.api [None req-af2f0c1d-9bc5-40c3-a712-d408e73bc806 tempest-ServerDiagnosticsNegativeTest-1635132145 tempest-ServerDiagnosticsNegativeTest-1635132145-project-member] Task: {'id': task-5115812, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 766.952425] env[65726]: DEBUG nova.compute.provider_tree [None req-cda5c6a3-fc21-4556-afc4-6ebd218aa8a5 tempest-ServerExternalEventsTest-1845125219 tempest-ServerExternalEventsTest-1845125219-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 766.977425] env[65726]: DEBUG oslo_vmware.api [None req-48153719-e51d-407a-98ef-93401c6270f2 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5255e07b-28ea-9540-d5d3-5c1291b82b42, 'name': SearchDatastore_Task, 'duration_secs': 0.011926} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 766.977425] env[65726]: DEBUG oslo_concurrency.lockutils [None req-48153719-e51d-407a-98ef-93401c6270f2 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 766.977425] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-48153719-e51d-407a-98ef-93401c6270f2 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore2] 65aaa2eb-3ca5-4c2b-92f9-b524366c10dd/65aaa2eb-3ca5-4c2b-92f9-b524366c10dd.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 766.977425] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0fe1e3a5-fb69-43f0-a239-888018d7280e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.985711] env[65726]: DEBUG oslo_vmware.api [None req-48153719-e51d-407a-98ef-93401c6270f2 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Waiting for the task: (returnval){ [ 766.985711] env[65726]: value = "task-5115813" [ 766.985711] env[65726]: _type = "Task" [ 766.985711] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 766.995019] env[65726]: DEBUG oslo_vmware.api [None req-48153719-e51d-407a-98ef-93401c6270f2 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Task: {'id': task-5115813, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 767.012847] env[65726]: DEBUG nova.compute.manager [None req-703975dc-3e07-4b77-b05e-15d0411bc7cd tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 608cc0ea-3f6b-4b4f-83c1-01aa50999d63] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 767.013500] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1494198-0cd1-4019-b24f-aafc7c61aee9 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.103303] env[65726]: INFO nova.compute.manager [None req-276e84bf-a104-4323-ae2b-430f1d490aad tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] [instance: 3827eed1-b7a1-4922-a4de-f0c038cb8566] Rebuilding instance [ 767.164835] env[65726]: DEBUG nova.compute.manager [None req-276e84bf-a104-4323-ae2b-430f1d490aad tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] [instance: 3827eed1-b7a1-4922-a4de-f0c038cb8566] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 767.165759] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cad727b-9040-4aac-ab7a-fd7e133c048e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.170194] env[65726]: WARNING openstack [req-57a1ccc9-eb29-4ce1-aa59-ef1da805887f req-4983a4fb-f280-4f1f-bd93-a5414d986d2f service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 767.170579] env[65726]: WARNING openstack [req-57a1ccc9-eb29-4ce1-aa59-ef1da805887f req-4983a4fb-f280-4f1f-bd93-a5414d986d2f service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 767.237735] env[65726]: DEBUG nova.compute.manager [None req-d51d38e8-71be-4889-a206-d52403737076 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: 9fc0c9ee-9dc7-4059-b8d2-a5db965e00c2] Start spawning the instance on the hypervisor. {{(pid=65726) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 767.267320] env[65726]: DEBUG oslo_concurrency.lockutils [None req-6c979061-2d03-4525-8b87-afa801f4542e tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Releasing lock "refresh_cache-6d2b86c2-0a44-433f-8b3f-2e9c7f693baa" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 767.267779] env[65726]: DEBUG nova.compute.manager [None req-6c979061-2d03-4525-8b87-afa801f4542e tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: 6d2b86c2-0a44-433f-8b3f-2e9c7f693baa] Instance network_info: |[{"id": "b487d53a-9bda-4e3c-be61-f26acbc8ae0f", "address": "fa:16:3e:ad:db:36", "network": {"id": "9d5ad167-99fb-4208-ab30-7053f35774e4", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-2082516762-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a8362295a62f4b51bae719a7ef5a4656", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dabbac20-1723-40ad-9da0-e53b28073651", "external-id": "nsx-vlan-transportzone-790", "segmentation_id": 790, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb487d53a-9b", "ovs_interfaceid": "b487d53a-9bda-4e3c-be61-f26acbc8ae0f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 767.269042] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-6c979061-2d03-4525-8b87-afa801f4542e tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: 6d2b86c2-0a44-433f-8b3f-2e9c7f693baa] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ad:db:36', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'dabbac20-1723-40ad-9da0-e53b28073651', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b487d53a-9bda-4e3c-be61-f26acbc8ae0f', 'vif_model': 'vmxnet3'}] {{(pid=65726) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 767.276681] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-6c979061-2d03-4525-8b87-afa801f4542e tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Creating folder: Project (a8362295a62f4b51bae719a7ef5a4656). Parent ref: group-v995008. {{(pid=65726) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 767.279440] env[65726]: DEBUG nova.virt.hardware [None req-d51d38e8-71be-4889-a206-d52403737076 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T19:28:34Z,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1539f06fa3534e90acbc3a60c4b8bd3f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T19:28:35Z,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 767.279722] env[65726]: DEBUG nova.virt.hardware [None req-d51d38e8-71be-4889-a206-d52403737076 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 767.279850] env[65726]: DEBUG nova.virt.hardware [None req-d51d38e8-71be-4889-a206-d52403737076 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 767.280048] env[65726]: DEBUG nova.virt.hardware [None req-d51d38e8-71be-4889-a206-d52403737076 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 767.280293] env[65726]: DEBUG nova.virt.hardware [None req-d51d38e8-71be-4889-a206-d52403737076 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 767.280468] env[65726]: DEBUG nova.virt.hardware [None req-d51d38e8-71be-4889-a206-d52403737076 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 767.281403] env[65726]: DEBUG nova.virt.hardware [None req-d51d38e8-71be-4889-a206-d52403737076 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 767.281665] env[65726]: DEBUG nova.virt.hardware [None req-d51d38e8-71be-4889-a206-d52403737076 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 767.281877] env[65726]: DEBUG nova.virt.hardware [None req-d51d38e8-71be-4889-a206-d52403737076 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 767.282143] env[65726]: DEBUG nova.virt.hardware [None req-d51d38e8-71be-4889-a206-d52403737076 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 767.282448] env[65726]: DEBUG nova.virt.hardware [None req-d51d38e8-71be-4889-a206-d52403737076 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 767.283462] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-96f885a4-02b0-4c8e-9fc7-643ebcc4ba1d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.286349] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb53ad4c-2829-4964-a4ec-b78309f96db3 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.298245] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9cf7f8c-e3f0-401f-9a39-cb21b79bacd0 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.304293] env[65726]: INFO nova.virt.vmwareapi.vm_util [None req-6c979061-2d03-4525-8b87-afa801f4542e tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Created folder: Project (a8362295a62f4b51bae719a7ef5a4656) in parent group-v995008. [ 767.304498] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-6c979061-2d03-4525-8b87-afa801f4542e tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Creating folder: Instances. Parent ref: group-v995120. {{(pid=65726) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 767.305295] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0417098a-4f97-4c40-b316-925661ab2cd1 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.328968] env[65726]: INFO nova.virt.vmwareapi.vm_util [None req-6c979061-2d03-4525-8b87-afa801f4542e tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Created folder: Instances in parent group-v995120. [ 767.329274] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-6c979061-2d03-4525-8b87-afa801f4542e tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 767.329459] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6d2b86c2-0a44-433f-8b3f-2e9c7f693baa] Creating VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 767.329708] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-4e072538-35dd-44f8-927b-e7fbdd9dadae {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.354874] env[65726]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 767.354874] env[65726]: value = "task-5115816" [ 767.354874] env[65726]: _type = "Task" [ 767.354874] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 767.365844] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115816, 'name': CreateVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 767.428867] env[65726]: DEBUG oslo_vmware.api [None req-af2f0c1d-9bc5-40c3-a712-d408e73bc806 tempest-ServerDiagnosticsNegativeTest-1635132145 tempest-ServerDiagnosticsNegativeTest-1635132145-project-member] Task: {'id': task-5115812, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.652104} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 767.429206] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-af2f0c1d-9bc5-40c3-a712-d408e73bc806 tempest-ServerDiagnosticsNegativeTest-1635132145 tempest-ServerDiagnosticsNegativeTest-1635132145-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore2] 0d6f14c5-5770-476f-a186-4cf1b2d4fbe8/0d6f14c5-5770-476f-a186-4cf1b2d4fbe8.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 767.429495] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-af2f0c1d-9bc5-40c3-a712-d408e73bc806 tempest-ServerDiagnosticsNegativeTest-1635132145 tempest-ServerDiagnosticsNegativeTest-1635132145-project-member] [instance: 0d6f14c5-5770-476f-a186-4cf1b2d4fbe8] Extending root virtual disk to 1048576 {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 767.429845] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-9e8fb594-85e3-4245-92f7-f5ec5ba8fa47 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.438686] env[65726]: DEBUG oslo_vmware.api [None req-af2f0c1d-9bc5-40c3-a712-d408e73bc806 tempest-ServerDiagnosticsNegativeTest-1635132145 tempest-ServerDiagnosticsNegativeTest-1635132145-project-member] Waiting for the task: (returnval){ [ 767.438686] env[65726]: value = "task-5115817" [ 767.438686] env[65726]: _type = "Task" [ 767.438686] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 767.451897] env[65726]: DEBUG oslo_vmware.api [None req-af2f0c1d-9bc5-40c3-a712-d408e73bc806 tempest-ServerDiagnosticsNegativeTest-1635132145 tempest-ServerDiagnosticsNegativeTest-1635132145-project-member] Task: {'id': task-5115817, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 767.458239] env[65726]: DEBUG nova.scheduler.client.report [None req-cda5c6a3-fc21-4556-afc4-6ebd218aa8a5 tempest-ServerExternalEventsTest-1845125219 tempest-ServerExternalEventsTest-1845125219-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 767.497819] env[65726]: DEBUG oslo_vmware.api [None req-48153719-e51d-407a-98ef-93401c6270f2 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Task: {'id': task-5115813, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 767.524693] env[65726]: INFO nova.compute.manager [None req-703975dc-3e07-4b77-b05e-15d0411bc7cd tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 608cc0ea-3f6b-4b4f-83c1-01aa50999d63] instance snapshotting [ 767.525477] env[65726]: DEBUG nova.objects.instance [None req-703975dc-3e07-4b77-b05e-15d0411bc7cd tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Lazy-loading 'flavor' on Instance uuid 608cc0ea-3f6b-4b4f-83c1-01aa50999d63 {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 767.712321] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a6bb6238-1018-4e0c-84f4-13a2ac492d77 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Acquiring lock "44d73b2d-2b6a-4501-9944-432da3c9330d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 767.712604] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a6bb6238-1018-4e0c-84f4-13a2ac492d77 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Lock "44d73b2d-2b6a-4501-9944-432da3c9330d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 767.868189] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115816, 'name': CreateVM_Task} progress is 99%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 767.952606] env[65726]: DEBUG oslo_vmware.api [None req-af2f0c1d-9bc5-40c3-a712-d408e73bc806 tempest-ServerDiagnosticsNegativeTest-1635132145 tempest-ServerDiagnosticsNegativeTest-1635132145-project-member] Task: {'id': task-5115817, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.273891} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 767.952923] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-af2f0c1d-9bc5-40c3-a712-d408e73bc806 tempest-ServerDiagnosticsNegativeTest-1635132145 tempest-ServerDiagnosticsNegativeTest-1635132145-project-member] [instance: 0d6f14c5-5770-476f-a186-4cf1b2d4fbe8] Extended root virtual disk {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 767.953765] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7922131-ed8f-4750-976a-baa80bfba53b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.971159] env[65726]: DEBUG oslo_concurrency.lockutils [None req-cda5c6a3-fc21-4556-afc4-6ebd218aa8a5 tempest-ServerExternalEventsTest-1845125219 tempest-ServerExternalEventsTest-1845125219-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.273s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 767.984486] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-af2f0c1d-9bc5-40c3-a712-d408e73bc806 tempest-ServerDiagnosticsNegativeTest-1635132145 tempest-ServerDiagnosticsNegativeTest-1635132145-project-member] [instance: 0d6f14c5-5770-476f-a186-4cf1b2d4fbe8] Reconfiguring VM instance instance-00000029 to attach disk [datastore2] 0d6f14c5-5770-476f-a186-4cf1b2d4fbe8/0d6f14c5-5770-476f-a186-4cf1b2d4fbe8.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 767.985538] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5ae556a9-54a3-4f16-bd92-05c390a03c24 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 32.365s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 767.985538] env[65726]: DEBUG nova.objects.instance [None req-5ae556a9-54a3-4f16-bd92-05c390a03c24 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Lazy-loading 'resources' on Instance uuid fd481728-3ef0-4a96-affd-ab2dd3f596bb {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 767.989162] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3d0bbfa7-8fbf-4714-a47f-367eadb112c9 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.010182] env[65726]: INFO nova.scheduler.client.report [None req-cda5c6a3-fc21-4556-afc4-6ebd218aa8a5 tempest-ServerExternalEventsTest-1845125219 tempest-ServerExternalEventsTest-1845125219-project-member] Deleted allocations for instance 96acc5f8-b411-408e-929d-d1035cfd50db [ 768.016516] env[65726]: DEBUG oslo_vmware.api [None req-af2f0c1d-9bc5-40c3-a712-d408e73bc806 tempest-ServerDiagnosticsNegativeTest-1635132145 tempest-ServerDiagnosticsNegativeTest-1635132145-project-member] Waiting for the task: (returnval){ [ 768.016516] env[65726]: value = "task-5115818" [ 768.016516] env[65726]: _type = "Task" [ 768.016516] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 768.023054] env[65726]: DEBUG oslo_vmware.api [None req-48153719-e51d-407a-98ef-93401c6270f2 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Task: {'id': task-5115813, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.666519} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 768.024366] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-48153719-e51d-407a-98ef-93401c6270f2 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore2] 65aaa2eb-3ca5-4c2b-92f9-b524366c10dd/65aaa2eb-3ca5-4c2b-92f9-b524366c10dd.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 768.024599] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-48153719-e51d-407a-98ef-93401c6270f2 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] [instance: 65aaa2eb-3ca5-4c2b-92f9-b524366c10dd] Extending root virtual disk to 1048576 {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 768.024974] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c6269a7e-f7bb-4fb1-89dd-1a53b62adeb0 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.043225] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08a503fe-e9b2-4333-8832-ebf3f351e9f9 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.047614] env[65726]: DEBUG oslo_vmware.api [None req-af2f0c1d-9bc5-40c3-a712-d408e73bc806 tempest-ServerDiagnosticsNegativeTest-1635132145 tempest-ServerDiagnosticsNegativeTest-1635132145-project-member] Task: {'id': task-5115818, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 768.050459] env[65726]: DEBUG oslo_vmware.api [None req-48153719-e51d-407a-98ef-93401c6270f2 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Waiting for the task: (returnval){ [ 768.050459] env[65726]: value = "task-5115819" [ 768.050459] env[65726]: _type = "Task" [ 768.050459] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 768.073914] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e051ea31-d5d4-4691-a764-30034c099741 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.080966] env[65726]: DEBUG oslo_vmware.api [None req-48153719-e51d-407a-98ef-93401c6270f2 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Task: {'id': task-5115819, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 768.187578] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-276e84bf-a104-4323-ae2b-430f1d490aad tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] [instance: 3827eed1-b7a1-4922-a4de-f0c038cb8566] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 768.187921] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-219fc97a-3cd0-4e8a-9753-ace6df23707a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.195058] env[65726]: DEBUG oslo_vmware.api [None req-276e84bf-a104-4323-ae2b-430f1d490aad tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] Waiting for the task: (returnval){ [ 768.195058] env[65726]: value = "task-5115820" [ 768.195058] env[65726]: _type = "Task" [ 768.195058] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 768.202135] env[65726]: WARNING openstack [req-57a1ccc9-eb29-4ce1-aa59-ef1da805887f req-4983a4fb-f280-4f1f-bd93-a5414d986d2f service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 768.202322] env[65726]: WARNING openstack [req-57a1ccc9-eb29-4ce1-aa59-ef1da805887f req-4983a4fb-f280-4f1f-bd93-a5414d986d2f service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 768.218930] env[65726]: DEBUG oslo_vmware.api [None req-276e84bf-a104-4323-ae2b-430f1d490aad tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] Task: {'id': task-5115820, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 768.370131] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115816, 'name': CreateVM_Task, 'duration_secs': 0.676088} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 768.370131] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6d2b86c2-0a44-433f-8b3f-2e9c7f693baa] Created VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 768.370824] env[65726]: WARNING openstack [None req-6c979061-2d03-4525-8b87-afa801f4542e tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 768.371018] env[65726]: WARNING openstack [None req-6c979061-2d03-4525-8b87-afa801f4542e tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 768.376710] env[65726]: DEBUG oslo_concurrency.lockutils [None req-6c979061-2d03-4525-8b87-afa801f4542e tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 768.376876] env[65726]: DEBUG oslo_concurrency.lockutils [None req-6c979061-2d03-4525-8b87-afa801f4542e tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 768.377308] env[65726]: DEBUG oslo_concurrency.lockutils [None req-6c979061-2d03-4525-8b87-afa801f4542e tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 768.377741] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-962fa79d-b529-4342-948f-5bf0b0413a36 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.385431] env[65726]: DEBUG oslo_vmware.api [None req-6c979061-2d03-4525-8b87-afa801f4542e tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Waiting for the task: (returnval){ [ 768.385431] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52d251a1-3d17-6ef2-1b57-e7a798b0b71a" [ 768.385431] env[65726]: _type = "Task" [ 768.385431] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 768.397606] env[65726]: DEBUG oslo_vmware.api [None req-6c979061-2d03-4525-8b87-afa801f4542e tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52d251a1-3d17-6ef2-1b57-e7a798b0b71a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 768.400913] env[65726]: DEBUG nova.network.neutron [None req-d51d38e8-71be-4889-a206-d52403737076 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: 9fc0c9ee-9dc7-4059-b8d2-a5db965e00c2] Successfully updated port: e4315658-0732-463a-b063-b1c4cf7cdafd {{(pid=65726) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 768.522555] env[65726]: WARNING openstack [req-57a1ccc9-eb29-4ce1-aa59-ef1da805887f req-4983a4fb-f280-4f1f-bd93-a5414d986d2f service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 768.523126] env[65726]: WARNING openstack [req-57a1ccc9-eb29-4ce1-aa59-ef1da805887f req-4983a4fb-f280-4f1f-bd93-a5414d986d2f service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 768.537871] env[65726]: DEBUG oslo_concurrency.lockutils [None req-cda5c6a3-fc21-4556-afc4-6ebd218aa8a5 tempest-ServerExternalEventsTest-1845125219 tempest-ServerExternalEventsTest-1845125219-project-member] Lock "96acc5f8-b411-408e-929d-d1035cfd50db" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 39.042s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 768.547159] env[65726]: DEBUG oslo_vmware.api [None req-af2f0c1d-9bc5-40c3-a712-d408e73bc806 tempest-ServerDiagnosticsNegativeTest-1635132145 tempest-ServerDiagnosticsNegativeTest-1635132145-project-member] Task: {'id': task-5115818, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 768.550203] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9ac23e9e-c303-4029-8e2c-3f5115af48ab tempest-ServerAddressesNegativeTestJSON-963263679 tempest-ServerAddressesNegativeTestJSON-963263679-project-member] Acquiring lock "00746508-d0d6-4dfa-9026-772398d004ee" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 768.550469] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9ac23e9e-c303-4029-8e2c-3f5115af48ab tempest-ServerAddressesNegativeTestJSON-963263679 tempest-ServerAddressesNegativeTestJSON-963263679-project-member] Lock "00746508-d0d6-4dfa-9026-772398d004ee" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 768.550713] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9ac23e9e-c303-4029-8e2c-3f5115af48ab tempest-ServerAddressesNegativeTestJSON-963263679 tempest-ServerAddressesNegativeTestJSON-963263679-project-member] Acquiring lock "00746508-d0d6-4dfa-9026-772398d004ee-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 768.550926] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9ac23e9e-c303-4029-8e2c-3f5115af48ab tempest-ServerAddressesNegativeTestJSON-963263679 tempest-ServerAddressesNegativeTestJSON-963263679-project-member] Lock "00746508-d0d6-4dfa-9026-772398d004ee-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 768.551110] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9ac23e9e-c303-4029-8e2c-3f5115af48ab tempest-ServerAddressesNegativeTestJSON-963263679 tempest-ServerAddressesNegativeTestJSON-963263679-project-member] Lock "00746508-d0d6-4dfa-9026-772398d004ee-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 768.553193] env[65726]: INFO nova.compute.manager [None req-9ac23e9e-c303-4029-8e2c-3f5115af48ab tempest-ServerAddressesNegativeTestJSON-963263679 tempest-ServerAddressesNegativeTestJSON-963263679-project-member] [instance: 00746508-d0d6-4dfa-9026-772398d004ee] Terminating instance [ 768.566641] env[65726]: DEBUG oslo_vmware.api [None req-48153719-e51d-407a-98ef-93401c6270f2 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Task: {'id': task-5115819, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.127777} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 768.566953] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-48153719-e51d-407a-98ef-93401c6270f2 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] [instance: 65aaa2eb-3ca5-4c2b-92f9-b524366c10dd] Extended root virtual disk {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 768.567794] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fff75ff5-39a7-455f-a9aa-081b7ee96922 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.592772] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-48153719-e51d-407a-98ef-93401c6270f2 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] [instance: 65aaa2eb-3ca5-4c2b-92f9-b524366c10dd] Reconfiguring VM instance instance-0000002a to attach disk [datastore2] 65aaa2eb-3ca5-4c2b-92f9-b524366c10dd/65aaa2eb-3ca5-4c2b-92f9-b524366c10dd.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 768.596730] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-703975dc-3e07-4b77-b05e-15d0411bc7cd tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 608cc0ea-3f6b-4b4f-83c1-01aa50999d63] Creating Snapshot of the VM instance {{(pid=65726) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 768.596995] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-117090e3-2b8a-4636-9b8c-099b1c6e5ea3 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.612865] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-de7ac76c-ce51-4408-912b-6e3b335286c1 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.620811] env[65726]: DEBUG oslo_vmware.api [None req-703975dc-3e07-4b77-b05e-15d0411bc7cd tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Waiting for the task: (returnval){ [ 768.620811] env[65726]: value = "task-5115821" [ 768.620811] env[65726]: _type = "Task" [ 768.620811] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 768.622321] env[65726]: DEBUG oslo_vmware.api [None req-48153719-e51d-407a-98ef-93401c6270f2 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Waiting for the task: (returnval){ [ 768.622321] env[65726]: value = "task-5115822" [ 768.622321] env[65726]: _type = "Task" [ 768.622321] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 768.639973] env[65726]: DEBUG oslo_vmware.api [None req-48153719-e51d-407a-98ef-93401c6270f2 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Task: {'id': task-5115822, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 768.643288] env[65726]: DEBUG oslo_vmware.api [None req-703975dc-3e07-4b77-b05e-15d0411bc7cd tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5115821, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 768.708062] env[65726]: DEBUG oslo_vmware.api [None req-276e84bf-a104-4323-ae2b-430f1d490aad tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] Task: {'id': task-5115820, 'name': PowerOffVM_Task, 'duration_secs': 0.276675} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 768.708367] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-276e84bf-a104-4323-ae2b-430f1d490aad tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] [instance: 3827eed1-b7a1-4922-a4de-f0c038cb8566] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 768.709460] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-276e84bf-a104-4323-ae2b-430f1d490aad tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] [instance: 3827eed1-b7a1-4922-a4de-f0c038cb8566] Destroying instance {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 768.710346] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b064fd30-f6bb-4da6-b531-732575f8dbcd {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.718387] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-276e84bf-a104-4323-ae2b-430f1d490aad tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] [instance: 3827eed1-b7a1-4922-a4de-f0c038cb8566] Unregistering the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 768.718689] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-70518581-53a4-4dd7-82f7-28b85e45f817 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.733955] env[65726]: DEBUG nova.network.neutron [req-57a1ccc9-eb29-4ce1-aa59-ef1da805887f req-4983a4fb-f280-4f1f-bd93-a5414d986d2f service nova] [instance: 0d6f14c5-5770-476f-a186-4cf1b2d4fbe8] Updated VIF entry in instance network info cache for port c8363d25-c607-4d04-ab53-412848d7cf4f. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 768.734482] env[65726]: DEBUG nova.network.neutron [req-57a1ccc9-eb29-4ce1-aa59-ef1da805887f req-4983a4fb-f280-4f1f-bd93-a5414d986d2f service nova] [instance: 0d6f14c5-5770-476f-a186-4cf1b2d4fbe8] Updating instance_info_cache with network_info: [{"id": "c8363d25-c607-4d04-ab53-412848d7cf4f", "address": "fa:16:3e:a8:64:02", "network": {"id": "55d629b8-a541-4720-9289-927bbc6e6f22", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.61", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "1539f06fa3534e90acbc3a60c4b8bd3f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "35342bcb-8b06-472e-b3c0-43fd3d6c4b30", "external-id": "nsx-vlan-transportzone-524", "segmentation_id": 524, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc8363d25-c6", "ovs_interfaceid": "c8363d25-c607-4d04-ab53-412848d7cf4f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 768.746554] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-276e84bf-a104-4323-ae2b-430f1d490aad tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] [instance: 3827eed1-b7a1-4922-a4de-f0c038cb8566] Unregistered the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 768.746878] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-276e84bf-a104-4323-ae2b-430f1d490aad tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] [instance: 3827eed1-b7a1-4922-a4de-f0c038cb8566] Deleting contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 768.747130] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-276e84bf-a104-4323-ae2b-430f1d490aad tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] Deleting the datastore file [datastore1] 3827eed1-b7a1-4922-a4de-f0c038cb8566 {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 768.747500] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-11c88ef4-cd22-4294-9b6c-a6206247bb9b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.758352] env[65726]: DEBUG oslo_vmware.api [None req-276e84bf-a104-4323-ae2b-430f1d490aad tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] Waiting for the task: (returnval){ [ 768.758352] env[65726]: value = "task-5115824" [ 768.758352] env[65726]: _type = "Task" [ 768.758352] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 768.772379] env[65726]: DEBUG oslo_vmware.api [None req-276e84bf-a104-4323-ae2b-430f1d490aad tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] Task: {'id': task-5115824, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 768.896989] env[65726]: DEBUG oslo_vmware.api [None req-6c979061-2d03-4525-8b87-afa801f4542e tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52d251a1-3d17-6ef2-1b57-e7a798b0b71a, 'name': SearchDatastore_Task, 'duration_secs': 0.021052} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 768.896989] env[65726]: DEBUG oslo_concurrency.lockutils [None req-6c979061-2d03-4525-8b87-afa801f4542e tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 768.896989] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-6c979061-2d03-4525-8b87-afa801f4542e tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: 6d2b86c2-0a44-433f-8b3f-2e9c7f693baa] Processing image b52362a3-ee8a-4cbf-b06f-513b0cc8f95c {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 768.897299] env[65726]: DEBUG oslo_concurrency.lockutils [None req-6c979061-2d03-4525-8b87-afa801f4542e tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 768.897340] env[65726]: DEBUG oslo_concurrency.lockutils [None req-6c979061-2d03-4525-8b87-afa801f4542e tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 768.897693] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-6c979061-2d03-4525-8b87-afa801f4542e tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 768.897877] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4a4c312a-1286-4f0b-9c4a-cda66791b058 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.903881] env[65726]: DEBUG oslo_concurrency.lockutils [None req-d51d38e8-71be-4889-a206-d52403737076 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Acquiring lock "refresh_cache-9fc0c9ee-9dc7-4059-b8d2-a5db965e00c2" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 768.903881] env[65726]: DEBUG oslo_concurrency.lockutils [None req-d51d38e8-71be-4889-a206-d52403737076 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Acquired lock "refresh_cache-9fc0c9ee-9dc7-4059-b8d2-a5db965e00c2" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 768.903881] env[65726]: DEBUG nova.network.neutron [None req-d51d38e8-71be-4889-a206-d52403737076 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: 9fc0c9ee-9dc7-4059-b8d2-a5db965e00c2] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 768.908149] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-6c979061-2d03-4525-8b87-afa801f4542e tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 768.908335] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-6c979061-2d03-4525-8b87-afa801f4542e tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65726) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 768.912428] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c87b64b5-7b9a-4628-8274-b60c5aa3e6b1 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.920544] env[65726]: DEBUG oslo_vmware.api [None req-6c979061-2d03-4525-8b87-afa801f4542e tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Waiting for the task: (returnval){ [ 768.920544] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]526429f9-5e88-76d9-b358-743f68d25d9b" [ 768.920544] env[65726]: _type = "Task" [ 768.920544] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 768.931052] env[65726]: DEBUG oslo_vmware.api [None req-6c979061-2d03-4525-8b87-afa801f4542e tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]526429f9-5e88-76d9-b358-743f68d25d9b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 769.044800] env[65726]: DEBUG oslo_vmware.api [None req-af2f0c1d-9bc5-40c3-a712-d408e73bc806 tempest-ServerDiagnosticsNegativeTest-1635132145 tempest-ServerDiagnosticsNegativeTest-1635132145-project-member] Task: {'id': task-5115818, 'name': ReconfigVM_Task, 'duration_secs': 0.812809} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 769.048865] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-af2f0c1d-9bc5-40c3-a712-d408e73bc806 tempest-ServerDiagnosticsNegativeTest-1635132145 tempest-ServerDiagnosticsNegativeTest-1635132145-project-member] [instance: 0d6f14c5-5770-476f-a186-4cf1b2d4fbe8] Reconfigured VM instance instance-00000029 to attach disk [datastore2] 0d6f14c5-5770-476f-a186-4cf1b2d4fbe8/0d6f14c5-5770-476f-a186-4cf1b2d4fbe8.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 769.050298] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9a7f1bcd-088a-4391-a642-99aaf75a82c1 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.059694] env[65726]: DEBUG oslo_vmware.api [None req-af2f0c1d-9bc5-40c3-a712-d408e73bc806 tempest-ServerDiagnosticsNegativeTest-1635132145 tempest-ServerDiagnosticsNegativeTest-1635132145-project-member] Waiting for the task: (returnval){ [ 769.059694] env[65726]: value = "task-5115825" [ 769.059694] env[65726]: _type = "Task" [ 769.059694] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 769.066718] env[65726]: DEBUG nova.compute.manager [None req-9ac23e9e-c303-4029-8e2c-3f5115af48ab tempest-ServerAddressesNegativeTestJSON-963263679 tempest-ServerAddressesNegativeTestJSON-963263679-project-member] [instance: 00746508-d0d6-4dfa-9026-772398d004ee] Start destroying the instance on the hypervisor. {{(pid=65726) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 769.066937] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-9ac23e9e-c303-4029-8e2c-3f5115af48ab tempest-ServerAddressesNegativeTestJSON-963263679 tempest-ServerAddressesNegativeTestJSON-963263679-project-member] [instance: 00746508-d0d6-4dfa-9026-772398d004ee] Destroying instance {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 769.068308] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1d29b2e-be22-415a-98ec-a0d584e15c23 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.079893] env[65726]: DEBUG oslo_vmware.api [None req-af2f0c1d-9bc5-40c3-a712-d408e73bc806 tempest-ServerDiagnosticsNegativeTest-1635132145 tempest-ServerDiagnosticsNegativeTest-1635132145-project-member] Task: {'id': task-5115825, 'name': Rename_Task} progress is 10%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 769.082905] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-9ac23e9e-c303-4029-8e2c-3f5115af48ab tempest-ServerAddressesNegativeTestJSON-963263679 tempest-ServerAddressesNegativeTestJSON-963263679-project-member] [instance: 00746508-d0d6-4dfa-9026-772398d004ee] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 769.083260] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e765c324-dd71-4870-9685-b02838782437 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.095254] env[65726]: DEBUG oslo_vmware.api [None req-9ac23e9e-c303-4029-8e2c-3f5115af48ab tempest-ServerAddressesNegativeTestJSON-963263679 tempest-ServerAddressesNegativeTestJSON-963263679-project-member] Waiting for the task: (returnval){ [ 769.095254] env[65726]: value = "task-5115826" [ 769.095254] env[65726]: _type = "Task" [ 769.095254] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 769.111307] env[65726]: DEBUG oslo_vmware.api [None req-9ac23e9e-c303-4029-8e2c-3f5115af48ab tempest-ServerAddressesNegativeTestJSON-963263679 tempest-ServerAddressesNegativeTestJSON-963263679-project-member] Task: {'id': task-5115826, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 769.141776] env[65726]: DEBUG oslo_vmware.api [None req-703975dc-3e07-4b77-b05e-15d0411bc7cd tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5115821, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 769.145219] env[65726]: DEBUG oslo_vmware.api [None req-48153719-e51d-407a-98ef-93401c6270f2 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Task: {'id': task-5115822, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 769.198850] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b222801-18dd-4318-9d63-06fa05dd808a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.211723] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13a4fd72-9857-4959-bafc-770a19af42cf {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.262126] env[65726]: DEBUG oslo_concurrency.lockutils [req-57a1ccc9-eb29-4ce1-aa59-ef1da805887f req-4983a4fb-f280-4f1f-bd93-a5414d986d2f service nova] Releasing lock "refresh_cache-0d6f14c5-5770-476f-a186-4cf1b2d4fbe8" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 769.262999] env[65726]: DEBUG nova.compute.manager [req-57a1ccc9-eb29-4ce1-aa59-ef1da805887f req-4983a4fb-f280-4f1f-bd93-a5414d986d2f service nova] [instance: 65aaa2eb-3ca5-4c2b-92f9-b524366c10dd] Received event network-vif-plugged-862071bf-725c-4193-b60b-72b03b4c0553 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 769.262999] env[65726]: DEBUG oslo_concurrency.lockutils [req-57a1ccc9-eb29-4ce1-aa59-ef1da805887f req-4983a4fb-f280-4f1f-bd93-a5414d986d2f service nova] Acquiring lock "65aaa2eb-3ca5-4c2b-92f9-b524366c10dd-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 769.263223] env[65726]: DEBUG oslo_concurrency.lockutils [req-57a1ccc9-eb29-4ce1-aa59-ef1da805887f req-4983a4fb-f280-4f1f-bd93-a5414d986d2f service nova] Lock "65aaa2eb-3ca5-4c2b-92f9-b524366c10dd-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 769.264113] env[65726]: DEBUG oslo_concurrency.lockutils [req-57a1ccc9-eb29-4ce1-aa59-ef1da805887f req-4983a4fb-f280-4f1f-bd93-a5414d986d2f service nova] Lock "65aaa2eb-3ca5-4c2b-92f9-b524366c10dd-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 769.264113] env[65726]: DEBUG nova.compute.manager [req-57a1ccc9-eb29-4ce1-aa59-ef1da805887f req-4983a4fb-f280-4f1f-bd93-a5414d986d2f service nova] [instance: 65aaa2eb-3ca5-4c2b-92f9-b524366c10dd] No waiting events found dispatching network-vif-plugged-862071bf-725c-4193-b60b-72b03b4c0553 {{(pid=65726) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 769.264113] env[65726]: WARNING nova.compute.manager [req-57a1ccc9-eb29-4ce1-aa59-ef1da805887f req-4983a4fb-f280-4f1f-bd93-a5414d986d2f service nova] [instance: 65aaa2eb-3ca5-4c2b-92f9-b524366c10dd] Received unexpected event network-vif-plugged-862071bf-725c-4193-b60b-72b03b4c0553 for instance with vm_state building and task_state spawning. [ 769.264675] env[65726]: DEBUG nova.compute.manager [req-57a1ccc9-eb29-4ce1-aa59-ef1da805887f req-4983a4fb-f280-4f1f-bd93-a5414d986d2f service nova] [instance: 65aaa2eb-3ca5-4c2b-92f9-b524366c10dd] Received event network-changed-862071bf-725c-4193-b60b-72b03b4c0553 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 769.264675] env[65726]: DEBUG nova.compute.manager [req-57a1ccc9-eb29-4ce1-aa59-ef1da805887f req-4983a4fb-f280-4f1f-bd93-a5414d986d2f service nova] [instance: 65aaa2eb-3ca5-4c2b-92f9-b524366c10dd] Refreshing instance network info cache due to event network-changed-862071bf-725c-4193-b60b-72b03b4c0553. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 769.264823] env[65726]: DEBUG oslo_concurrency.lockutils [req-57a1ccc9-eb29-4ce1-aa59-ef1da805887f req-4983a4fb-f280-4f1f-bd93-a5414d986d2f service nova] Acquiring lock "refresh_cache-65aaa2eb-3ca5-4c2b-92f9-b524366c10dd" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 769.265591] env[65726]: DEBUG oslo_concurrency.lockutils [req-57a1ccc9-eb29-4ce1-aa59-ef1da805887f req-4983a4fb-f280-4f1f-bd93-a5414d986d2f service nova] Acquired lock "refresh_cache-65aaa2eb-3ca5-4c2b-92f9-b524366c10dd" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 769.265591] env[65726]: DEBUG nova.network.neutron [req-57a1ccc9-eb29-4ce1-aa59-ef1da805887f req-4983a4fb-f280-4f1f-bd93-a5414d986d2f service nova] [instance: 65aaa2eb-3ca5-4c2b-92f9-b524366c10dd] Refreshing network info cache for port 862071bf-725c-4193-b60b-72b03b4c0553 {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 769.273811] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebdcf041-bdab-469b-846d-59fac4d84f91 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.290913] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cedef85a-07c6-43cc-a8df-fc37945738cf {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.295294] env[65726]: DEBUG oslo_vmware.api [None req-276e84bf-a104-4323-ae2b-430f1d490aad tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] Task: {'id': task-5115824, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.111128} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 769.295894] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-276e84bf-a104-4323-ae2b-430f1d490aad tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] Deleted the datastore file {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 769.296221] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-276e84bf-a104-4323-ae2b-430f1d490aad tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] [instance: 3827eed1-b7a1-4922-a4de-f0c038cb8566] Deleted contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 769.296557] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-276e84bf-a104-4323-ae2b-430f1d490aad tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] [instance: 3827eed1-b7a1-4922-a4de-f0c038cb8566] Instance destroyed {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 769.311262] env[65726]: DEBUG nova.compute.provider_tree [None req-5ae556a9-54a3-4f16-bd92-05c390a03c24 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Updating inventory in ProviderTree for provider 07c4692f-bdb4-4058-9173-ff9664830295 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 769.408330] env[65726]: WARNING openstack [None req-d51d38e8-71be-4889-a206-d52403737076 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 769.408763] env[65726]: WARNING openstack [None req-d51d38e8-71be-4889-a206-d52403737076 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 769.413922] env[65726]: DEBUG nova.network.neutron [None req-d51d38e8-71be-4889-a206-d52403737076 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: 9fc0c9ee-9dc7-4059-b8d2-a5db965e00c2] Instance cache missing network info. {{(pid=65726) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 769.431425] env[65726]: DEBUG oslo_vmware.api [None req-6c979061-2d03-4525-8b87-afa801f4542e tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]526429f9-5e88-76d9-b358-743f68d25d9b, 'name': SearchDatastore_Task, 'duration_secs': 0.011289} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 769.431795] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-850cb722-6551-4f05-a156-b195ecd5c007 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.438123] env[65726]: DEBUG oslo_vmware.api [None req-6c979061-2d03-4525-8b87-afa801f4542e tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Waiting for the task: (returnval){ [ 769.438123] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]520b4ff3-e5ff-4d5a-ef75-127f84154238" [ 769.438123] env[65726]: _type = "Task" [ 769.438123] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 769.449177] env[65726]: DEBUG oslo_vmware.api [None req-6c979061-2d03-4525-8b87-afa801f4542e tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]520b4ff3-e5ff-4d5a-ef75-127f84154238, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 769.476571] env[65726]: WARNING openstack [None req-d51d38e8-71be-4889-a206-d52403737076 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 769.476957] env[65726]: WARNING openstack [None req-d51d38e8-71be-4889-a206-d52403737076 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 769.544645] env[65726]: WARNING openstack [None req-d51d38e8-71be-4889-a206-d52403737076 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 769.545126] env[65726]: WARNING openstack [None req-d51d38e8-71be-4889-a206-d52403737076 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 769.571214] env[65726]: DEBUG oslo_vmware.api [None req-af2f0c1d-9bc5-40c3-a712-d408e73bc806 tempest-ServerDiagnosticsNegativeTest-1635132145 tempest-ServerDiagnosticsNegativeTest-1635132145-project-member] Task: {'id': task-5115825, 'name': Rename_Task, 'duration_secs': 0.20585} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 769.571578] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-af2f0c1d-9bc5-40c3-a712-d408e73bc806 tempest-ServerDiagnosticsNegativeTest-1635132145 tempest-ServerDiagnosticsNegativeTest-1635132145-project-member] [instance: 0d6f14c5-5770-476f-a186-4cf1b2d4fbe8] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 769.571960] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e69ec1ff-6a6b-44ff-a95d-4d4f0adc8fc0 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.585036] env[65726]: DEBUG oslo_vmware.api [None req-af2f0c1d-9bc5-40c3-a712-d408e73bc806 tempest-ServerDiagnosticsNegativeTest-1635132145 tempest-ServerDiagnosticsNegativeTest-1635132145-project-member] Waiting for the task: (returnval){ [ 769.585036] env[65726]: value = "task-5115827" [ 769.585036] env[65726]: _type = "Task" [ 769.585036] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 769.597487] env[65726]: DEBUG oslo_vmware.api [None req-af2f0c1d-9bc5-40c3-a712-d408e73bc806 tempest-ServerDiagnosticsNegativeTest-1635132145 tempest-ServerDiagnosticsNegativeTest-1635132145-project-member] Task: {'id': task-5115827, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 769.608967] env[65726]: DEBUG oslo_vmware.api [None req-9ac23e9e-c303-4029-8e2c-3f5115af48ab tempest-ServerAddressesNegativeTestJSON-963263679 tempest-ServerAddressesNegativeTestJSON-963263679-project-member] Task: {'id': task-5115826, 'name': PowerOffVM_Task, 'duration_secs': 0.267824} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 769.608967] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-9ac23e9e-c303-4029-8e2c-3f5115af48ab tempest-ServerAddressesNegativeTestJSON-963263679 tempest-ServerAddressesNegativeTestJSON-963263679-project-member] [instance: 00746508-d0d6-4dfa-9026-772398d004ee] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 769.608967] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-9ac23e9e-c303-4029-8e2c-3f5115af48ab tempest-ServerAddressesNegativeTestJSON-963263679 tempest-ServerAddressesNegativeTestJSON-963263679-project-member] [instance: 00746508-d0d6-4dfa-9026-772398d004ee] Unregistering the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 769.608967] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-81596a50-ab2b-416b-91f9-965141d05817 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.634769] env[65726]: DEBUG oslo_vmware.api [None req-703975dc-3e07-4b77-b05e-15d0411bc7cd tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5115821, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 769.637106] env[65726]: DEBUG nova.network.neutron [None req-d51d38e8-71be-4889-a206-d52403737076 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: 9fc0c9ee-9dc7-4059-b8d2-a5db965e00c2] Updating instance_info_cache with network_info: [{"id": "e4315658-0732-463a-b063-b1c4cf7cdafd", "address": "fa:16:3e:c1:e7:e1", "network": {"id": "9d5ad167-99fb-4208-ab30-7053f35774e4", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-2082516762-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a8362295a62f4b51bae719a7ef5a4656", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dabbac20-1723-40ad-9da0-e53b28073651", "external-id": "nsx-vlan-transportzone-790", "segmentation_id": 790, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape4315658-07", "ovs_interfaceid": "e4315658-0732-463a-b063-b1c4cf7cdafd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 769.648092] env[65726]: DEBUG oslo_vmware.api [None req-48153719-e51d-407a-98ef-93401c6270f2 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Task: {'id': task-5115822, 'name': ReconfigVM_Task, 'duration_secs': 0.697881} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 769.649447] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-48153719-e51d-407a-98ef-93401c6270f2 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] [instance: 65aaa2eb-3ca5-4c2b-92f9-b524366c10dd] Reconfigured VM instance instance-0000002a to attach disk [datastore2] 65aaa2eb-3ca5-4c2b-92f9-b524366c10dd/65aaa2eb-3ca5-4c2b-92f9-b524366c10dd.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 769.649825] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-5a06f412-a78d-4def-866a-f083d1b635de {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.658255] env[65726]: DEBUG oslo_vmware.api [None req-48153719-e51d-407a-98ef-93401c6270f2 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Waiting for the task: (returnval){ [ 769.658255] env[65726]: value = "task-5115829" [ 769.658255] env[65726]: _type = "Task" [ 769.658255] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 769.668850] env[65726]: DEBUG oslo_vmware.api [None req-48153719-e51d-407a-98ef-93401c6270f2 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Task: {'id': task-5115829, 'name': Rename_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 769.698229] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-9ac23e9e-c303-4029-8e2c-3f5115af48ab tempest-ServerAddressesNegativeTestJSON-963263679 tempest-ServerAddressesNegativeTestJSON-963263679-project-member] [instance: 00746508-d0d6-4dfa-9026-772398d004ee] Unregistered the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 769.698460] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-9ac23e9e-c303-4029-8e2c-3f5115af48ab tempest-ServerAddressesNegativeTestJSON-963263679 tempest-ServerAddressesNegativeTestJSON-963263679-project-member] [instance: 00746508-d0d6-4dfa-9026-772398d004ee] Deleting contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 769.699199] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-9ac23e9e-c303-4029-8e2c-3f5115af48ab tempest-ServerAddressesNegativeTestJSON-963263679 tempest-ServerAddressesNegativeTestJSON-963263679-project-member] Deleting the datastore file [datastore1] 00746508-d0d6-4dfa-9026-772398d004ee {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 769.699199] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ea04a1c7-7cb6-4636-b769-f17ccc20f86b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.707575] env[65726]: DEBUG oslo_vmware.api [None req-9ac23e9e-c303-4029-8e2c-3f5115af48ab tempest-ServerAddressesNegativeTestJSON-963263679 tempest-ServerAddressesNegativeTestJSON-963263679-project-member] Waiting for the task: (returnval){ [ 769.707575] env[65726]: value = "task-5115830" [ 769.707575] env[65726]: _type = "Task" [ 769.707575] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 769.719449] env[65726]: DEBUG oslo_vmware.api [None req-9ac23e9e-c303-4029-8e2c-3f5115af48ab tempest-ServerAddressesNegativeTestJSON-963263679 tempest-ServerAddressesNegativeTestJSON-963263679-project-member] Task: {'id': task-5115830, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 769.777088] env[65726]: WARNING openstack [req-57a1ccc9-eb29-4ce1-aa59-ef1da805887f req-4983a4fb-f280-4f1f-bd93-a5414d986d2f service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 769.777558] env[65726]: WARNING openstack [req-57a1ccc9-eb29-4ce1-aa59-ef1da805887f req-4983a4fb-f280-4f1f-bd93-a5414d986d2f service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 769.830575] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b0e194cf-16dc-40d0-b65c-88bd21cd63e3 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Acquiring lock "cb8855b6-589d-4863-b86f-7fb4e30fb29e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 769.830901] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b0e194cf-16dc-40d0-b65c-88bd21cd63e3 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Lock "cb8855b6-589d-4863-b86f-7fb4e30fb29e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 769.840047] env[65726]: ERROR nova.scheduler.client.report [None req-5ae556a9-54a3-4f16-bd92-05c390a03c24 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] [req-239e31ff-9744-4542-b8cc-9f32876c31ac] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 07c4692f-bdb4-4058-9173-ff9664830295. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-239e31ff-9744-4542-b8cc-9f32876c31ac"}]} [ 769.860234] env[65726]: DEBUG nova.scheduler.client.report [None req-5ae556a9-54a3-4f16-bd92-05c390a03c24 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Refreshing inventories for resource provider 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 769.881070] env[65726]: DEBUG nova.scheduler.client.report [None req-5ae556a9-54a3-4f16-bd92-05c390a03c24 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Updating ProviderTree inventory for provider 07c4692f-bdb4-4058-9173-ff9664830295 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 769.881070] env[65726]: DEBUG nova.compute.provider_tree [None req-5ae556a9-54a3-4f16-bd92-05c390a03c24 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Updating inventory in ProviderTree for provider 07c4692f-bdb4-4058-9173-ff9664830295 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 769.896558] env[65726]: DEBUG nova.scheduler.client.report [None req-5ae556a9-54a3-4f16-bd92-05c390a03c24 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Refreshing aggregate associations for resource provider 07c4692f-bdb4-4058-9173-ff9664830295, aggregates: None {{(pid=65726) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 769.913617] env[65726]: WARNING openstack [req-57a1ccc9-eb29-4ce1-aa59-ef1da805887f req-4983a4fb-f280-4f1f-bd93-a5414d986d2f service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 769.914032] env[65726]: WARNING openstack [req-57a1ccc9-eb29-4ce1-aa59-ef1da805887f req-4983a4fb-f280-4f1f-bd93-a5414d986d2f service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 769.923501] env[65726]: DEBUG nova.scheduler.client.report [None req-5ae556a9-54a3-4f16-bd92-05c390a03c24 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Refreshing trait associations for resource provider 07c4692f-bdb4-4058-9173-ff9664830295, traits: COMPUTE_IMAGE_TYPE_ISO,HW_ARCH_X86_64,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=65726) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 769.938658] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b86f24be-1076-4ee1-9368-176430971dfc tempest-VolumesAssistedSnapshotsTest-1477776432 tempest-VolumesAssistedSnapshotsTest-1477776432-project-admin] Acquiring lock "aa2716a9-2cf8-4e51-800d-1c8f98b2cf05" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 769.938909] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b86f24be-1076-4ee1-9368-176430971dfc tempest-VolumesAssistedSnapshotsTest-1477776432 tempest-VolumesAssistedSnapshotsTest-1477776432-project-admin] Lock "aa2716a9-2cf8-4e51-800d-1c8f98b2cf05" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 769.942943] env[65726]: DEBUG nova.compute.manager [req-da245a27-980c-420b-9648-990ba5e6a122 req-ba47e5fc-0777-4cc0-a801-66ce5cfe21e1 service nova] [instance: 6d2b86c2-0a44-433f-8b3f-2e9c7f693baa] Received event network-changed-b487d53a-9bda-4e3c-be61-f26acbc8ae0f {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 769.943167] env[65726]: DEBUG nova.compute.manager [req-da245a27-980c-420b-9648-990ba5e6a122 req-ba47e5fc-0777-4cc0-a801-66ce5cfe21e1 service nova] [instance: 6d2b86c2-0a44-433f-8b3f-2e9c7f693baa] Refreshing instance network info cache due to event network-changed-b487d53a-9bda-4e3c-be61-f26acbc8ae0f. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 769.943387] env[65726]: DEBUG oslo_concurrency.lockutils [req-da245a27-980c-420b-9648-990ba5e6a122 req-ba47e5fc-0777-4cc0-a801-66ce5cfe21e1 service nova] Acquiring lock "refresh_cache-6d2b86c2-0a44-433f-8b3f-2e9c7f693baa" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 769.943521] env[65726]: DEBUG oslo_concurrency.lockutils [req-da245a27-980c-420b-9648-990ba5e6a122 req-ba47e5fc-0777-4cc0-a801-66ce5cfe21e1 service nova] Acquired lock "refresh_cache-6d2b86c2-0a44-433f-8b3f-2e9c7f693baa" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 769.943960] env[65726]: DEBUG nova.network.neutron [req-da245a27-980c-420b-9648-990ba5e6a122 req-ba47e5fc-0777-4cc0-a801-66ce5cfe21e1 service nova] [instance: 6d2b86c2-0a44-433f-8b3f-2e9c7f693baa] Refreshing network info cache for port b487d53a-9bda-4e3c-be61-f26acbc8ae0f {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 769.959839] env[65726]: DEBUG oslo_vmware.api [None req-6c979061-2d03-4525-8b87-afa801f4542e tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]520b4ff3-e5ff-4d5a-ef75-127f84154238, 'name': SearchDatastore_Task, 'duration_secs': 0.010199} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 769.960146] env[65726]: DEBUG oslo_concurrency.lockutils [None req-6c979061-2d03-4525-8b87-afa801f4542e tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 769.960405] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-6c979061-2d03-4525-8b87-afa801f4542e tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore2] 6d2b86c2-0a44-433f-8b3f-2e9c7f693baa/6d2b86c2-0a44-433f-8b3f-2e9c7f693baa.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 769.960805] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-30be3ec4-8ebb-4ca4-ac88-17b7296b525a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.973609] env[65726]: DEBUG nova.compute.manager [req-767b54cd-afaa-479f-beab-3d140190858d req-67a7e1c2-d4f5-4635-9745-5320cfc755c6 service nova] [instance: 895797e4-2941-44cd-aab1-67afa4fac02d] Received event network-changed-a1dec2db-0e1e-45dd-8587-a84e4a5e2ef4 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 769.973804] env[65726]: DEBUG nova.compute.manager [req-767b54cd-afaa-479f-beab-3d140190858d req-67a7e1c2-d4f5-4635-9745-5320cfc755c6 service nova] [instance: 895797e4-2941-44cd-aab1-67afa4fac02d] Refreshing instance network info cache due to event network-changed-a1dec2db-0e1e-45dd-8587-a84e4a5e2ef4. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 769.974034] env[65726]: DEBUG oslo_concurrency.lockutils [req-767b54cd-afaa-479f-beab-3d140190858d req-67a7e1c2-d4f5-4635-9745-5320cfc755c6 service nova] Acquiring lock "refresh_cache-895797e4-2941-44cd-aab1-67afa4fac02d" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 769.974180] env[65726]: DEBUG oslo_concurrency.lockutils [req-767b54cd-afaa-479f-beab-3d140190858d req-67a7e1c2-d4f5-4635-9745-5320cfc755c6 service nova] Acquired lock "refresh_cache-895797e4-2941-44cd-aab1-67afa4fac02d" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 769.974342] env[65726]: DEBUG nova.network.neutron [req-767b54cd-afaa-479f-beab-3d140190858d req-67a7e1c2-d4f5-4635-9745-5320cfc755c6 service nova] [instance: 895797e4-2941-44cd-aab1-67afa4fac02d] Refreshing network info cache for port a1dec2db-0e1e-45dd-8587-a84e4a5e2ef4 {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 769.977940] env[65726]: DEBUG oslo_vmware.api [None req-6c979061-2d03-4525-8b87-afa801f4542e tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Waiting for the task: (returnval){ [ 769.977940] env[65726]: value = "task-5115831" [ 769.977940] env[65726]: _type = "Task" [ 769.977940] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 769.987949] env[65726]: DEBUG oslo_vmware.api [None req-6c979061-2d03-4525-8b87-afa801f4542e tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Task: {'id': task-5115831, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 770.006504] env[65726]: WARNING openstack [req-57a1ccc9-eb29-4ce1-aa59-ef1da805887f req-4983a4fb-f280-4f1f-bd93-a5414d986d2f service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 770.006878] env[65726]: WARNING openstack [req-57a1ccc9-eb29-4ce1-aa59-ef1da805887f req-4983a4fb-f280-4f1f-bd93-a5414d986d2f service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 770.094171] env[65726]: DEBUG oslo_vmware.api [None req-af2f0c1d-9bc5-40c3-a712-d408e73bc806 tempest-ServerDiagnosticsNegativeTest-1635132145 tempest-ServerDiagnosticsNegativeTest-1635132145-project-member] Task: {'id': task-5115827, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 770.120218] env[65726]: DEBUG nova.network.neutron [req-57a1ccc9-eb29-4ce1-aa59-ef1da805887f req-4983a4fb-f280-4f1f-bd93-a5414d986d2f service nova] [instance: 65aaa2eb-3ca5-4c2b-92f9-b524366c10dd] Updated VIF entry in instance network info cache for port 862071bf-725c-4193-b60b-72b03b4c0553. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 770.120630] env[65726]: DEBUG nova.network.neutron [req-57a1ccc9-eb29-4ce1-aa59-ef1da805887f req-4983a4fb-f280-4f1f-bd93-a5414d986d2f service nova] [instance: 65aaa2eb-3ca5-4c2b-92f9-b524366c10dd] Updating instance_info_cache with network_info: [{"id": "862071bf-725c-4193-b60b-72b03b4c0553", "address": "fa:16:3e:9e:67:58", "network": {"id": "27801f9a-9912-41cf-be65-bc9903f4952f", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-781436951-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aac5fe6c99f740ada3747088b2b0ad89", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dd7d0d95-6848-4e69-ac21-75f8db82a3b5", "external-id": "nsx-vlan-transportzone-272", "segmentation_id": 272, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap862071bf-72", "ovs_interfaceid": "862071bf-725c-4193-b60b-72b03b4c0553", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 770.141026] env[65726]: DEBUG oslo_vmware.api [None req-703975dc-3e07-4b77-b05e-15d0411bc7cd tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5115821, 'name': CreateSnapshot_Task, 'duration_secs': 1.16413} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 770.142042] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-703975dc-3e07-4b77-b05e-15d0411bc7cd tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 608cc0ea-3f6b-4b4f-83c1-01aa50999d63] Created Snapshot of the VM instance {{(pid=65726) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 770.142811] env[65726]: DEBUG oslo_concurrency.lockutils [None req-d51d38e8-71be-4889-a206-d52403737076 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Releasing lock "refresh_cache-9fc0c9ee-9dc7-4059-b8d2-a5db965e00c2" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 770.143022] env[65726]: DEBUG nova.compute.manager [None req-d51d38e8-71be-4889-a206-d52403737076 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: 9fc0c9ee-9dc7-4059-b8d2-a5db965e00c2] Instance network_info: |[{"id": "e4315658-0732-463a-b063-b1c4cf7cdafd", "address": "fa:16:3e:c1:e7:e1", "network": {"id": "9d5ad167-99fb-4208-ab30-7053f35774e4", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-2082516762-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a8362295a62f4b51bae719a7ef5a4656", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dabbac20-1723-40ad-9da0-e53b28073651", "external-id": "nsx-vlan-transportzone-790", "segmentation_id": 790, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape4315658-07", "ovs_interfaceid": "e4315658-0732-463a-b063-b1c4cf7cdafd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 770.144332] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80753e46-b478-4cb1-bc48-be005436cc67 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.147754] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-d51d38e8-71be-4889-a206-d52403737076 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: 9fc0c9ee-9dc7-4059-b8d2-a5db965e00c2] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c1:e7:e1', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'dabbac20-1723-40ad-9da0-e53b28073651', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e4315658-0732-463a-b063-b1c4cf7cdafd', 'vif_model': 'vmxnet3'}] {{(pid=65726) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 770.157410] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-d51d38e8-71be-4889-a206-d52403737076 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 770.161434] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9fc0c9ee-9dc7-4059-b8d2-a5db965e00c2] Creating VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 770.162690] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b24435b2-5649-4b23-a5f7-8ce254df6a0e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.020976] env[65726]: DEBUG nova.compute.utils [None req-b86f24be-1076-4ee1-9368-176430971dfc tempest-VolumesAssistedSnapshotsTest-1477776432 tempest-VolumesAssistedSnapshotsTest-1477776432-project-admin] Using /dev/sd instead of None {{(pid=65726) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 771.026121] env[65726]: WARNING openstack [req-da245a27-980c-420b-9648-990ba5e6a122 req-ba47e5fc-0777-4cc0-a801-66ce5cfe21e1 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 771.026121] env[65726]: WARNING openstack [req-da245a27-980c-420b-9648-990ba5e6a122 req-ba47e5fc-0777-4cc0-a801-66ce5cfe21e1 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 771.030723] env[65726]: WARNING openstack [req-767b54cd-afaa-479f-beab-3d140190858d req-67a7e1c2-d4f5-4635-9745-5320cfc755c6 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 771.031057] env[65726]: WARNING openstack [req-767b54cd-afaa-479f-beab-3d140190858d req-67a7e1c2-d4f5-4635-9745-5320cfc755c6 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 771.038374] env[65726]: DEBUG oslo_concurrency.lockutils [req-57a1ccc9-eb29-4ce1-aa59-ef1da805887f req-4983a4fb-f280-4f1f-bd93-a5414d986d2f service nova] Releasing lock "refresh_cache-65aaa2eb-3ca5-4c2b-92f9-b524366c10dd" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 771.038660] env[65726]: DEBUG nova.compute.manager [req-57a1ccc9-eb29-4ce1-aa59-ef1da805887f req-4983a4fb-f280-4f1f-bd93-a5414d986d2f service nova] [instance: 02351120-8385-4403-8464-a154f3da9380] Received event network-vif-deleted-f9ef8504-78db-4da2-abc7-94b9a07a014e {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 771.038869] env[65726]: DEBUG nova.compute.manager [req-57a1ccc9-eb29-4ce1-aa59-ef1da805887f req-4983a4fb-f280-4f1f-bd93-a5414d986d2f service nova] [instance: 6d2b86c2-0a44-433f-8b3f-2e9c7f693baa] Received event network-vif-plugged-b487d53a-9bda-4e3c-be61-f26acbc8ae0f {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 771.039542] env[65726]: DEBUG oslo_concurrency.lockutils [req-57a1ccc9-eb29-4ce1-aa59-ef1da805887f req-4983a4fb-f280-4f1f-bd93-a5414d986d2f service nova] Acquiring lock "6d2b86c2-0a44-433f-8b3f-2e9c7f693baa-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 771.039542] env[65726]: DEBUG oslo_concurrency.lockutils [req-57a1ccc9-eb29-4ce1-aa59-ef1da805887f req-4983a4fb-f280-4f1f-bd93-a5414d986d2f service nova] Lock "6d2b86c2-0a44-433f-8b3f-2e9c7f693baa-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 771.039542] env[65726]: DEBUG oslo_concurrency.lockutils [req-57a1ccc9-eb29-4ce1-aa59-ef1da805887f req-4983a4fb-f280-4f1f-bd93-a5414d986d2f service nova] Lock "6d2b86c2-0a44-433f-8b3f-2e9c7f693baa-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 771.039542] env[65726]: DEBUG nova.compute.manager [req-57a1ccc9-eb29-4ce1-aa59-ef1da805887f req-4983a4fb-f280-4f1f-bd93-a5414d986d2f service nova] [instance: 6d2b86c2-0a44-433f-8b3f-2e9c7f693baa] No waiting events found dispatching network-vif-plugged-b487d53a-9bda-4e3c-be61-f26acbc8ae0f {{(pid=65726) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 771.039806] env[65726]: WARNING nova.compute.manager [req-57a1ccc9-eb29-4ce1-aa59-ef1da805887f req-4983a4fb-f280-4f1f-bd93-a5414d986d2f service nova] [instance: 6d2b86c2-0a44-433f-8b3f-2e9c7f693baa] Received unexpected event network-vif-plugged-b487d53a-9bda-4e3c-be61-f26acbc8ae0f for instance with vm_state building and task_state spawning. [ 771.061174] env[65726]: DEBUG oslo_vmware.api [None req-48153719-e51d-407a-98ef-93401c6270f2 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Task: {'id': task-5115829, 'name': Rename_Task, 'duration_secs': 0.181237} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 771.069836] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-48153719-e51d-407a-98ef-93401c6270f2 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] [instance: 65aaa2eb-3ca5-4c2b-92f9-b524366c10dd] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 771.070294] env[65726]: DEBUG oslo_vmware.api [None req-9ac23e9e-c303-4029-8e2c-3f5115af48ab tempest-ServerAddressesNegativeTestJSON-963263679 tempest-ServerAddressesNegativeTestJSON-963263679-project-member] Task: {'id': task-5115830, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.259086} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 771.071545] env[65726]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 771.071545] env[65726]: value = "task-5115832" [ 771.071545] env[65726]: _type = "Task" [ 771.071545] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 771.071545] env[65726]: DEBUG oslo_vmware.api [None req-6c979061-2d03-4525-8b87-afa801f4542e tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Task: {'id': task-5115831, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.661687} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 771.075914] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-60eb46fa-a039-4f5c-8757-4f5840b514db {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.077752] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-9ac23e9e-c303-4029-8e2c-3f5115af48ab tempest-ServerAddressesNegativeTestJSON-963263679 tempest-ServerAddressesNegativeTestJSON-963263679-project-member] Deleted the datastore file {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 771.077971] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-9ac23e9e-c303-4029-8e2c-3f5115af48ab tempest-ServerAddressesNegativeTestJSON-963263679 tempest-ServerAddressesNegativeTestJSON-963263679-project-member] [instance: 00746508-d0d6-4dfa-9026-772398d004ee] Deleted contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 771.078166] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-9ac23e9e-c303-4029-8e2c-3f5115af48ab tempest-ServerAddressesNegativeTestJSON-963263679 tempest-ServerAddressesNegativeTestJSON-963263679-project-member] [instance: 00746508-d0d6-4dfa-9026-772398d004ee] Instance destroyed {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 771.078338] env[65726]: INFO nova.compute.manager [None req-9ac23e9e-c303-4029-8e2c-3f5115af48ab tempest-ServerAddressesNegativeTestJSON-963263679 tempest-ServerAddressesNegativeTestJSON-963263679-project-member] [instance: 00746508-d0d6-4dfa-9026-772398d004ee] Took 2.01 seconds to destroy the instance on the hypervisor. [ 771.078583] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-9ac23e9e-c303-4029-8e2c-3f5115af48ab tempest-ServerAddressesNegativeTestJSON-963263679 tempest-ServerAddressesNegativeTestJSON-963263679-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 771.078880] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-6c979061-2d03-4525-8b87-afa801f4542e tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore2] 6d2b86c2-0a44-433f-8b3f-2e9c7f693baa/6d2b86c2-0a44-433f-8b3f-2e9c7f693baa.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 771.079480] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-6c979061-2d03-4525-8b87-afa801f4542e tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: 6d2b86c2-0a44-433f-8b3f-2e9c7f693baa] Extending root virtual disk to 1048576 {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 771.083450] env[65726]: DEBUG nova.compute.manager [-] [instance: 00746508-d0d6-4dfa-9026-772398d004ee] Deallocating network for instance {{(pid=65726) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 771.083632] env[65726]: DEBUG nova.network.neutron [-] [instance: 00746508-d0d6-4dfa-9026-772398d004ee] deallocate_for_instance() {{(pid=65726) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 771.084098] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 771.084706] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 771.095055] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-59baa6c5-5f15-4a2c-9873-83815c4a3a2b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.097221] env[65726]: DEBUG oslo_vmware.api [None req-af2f0c1d-9bc5-40c3-a712-d408e73bc806 tempest-ServerDiagnosticsNegativeTest-1635132145 tempest-ServerDiagnosticsNegativeTest-1635132145-project-member] Task: {'id': task-5115827, 'name': PowerOnVM_Task, 'duration_secs': 0.664022} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 771.099608] env[65726]: DEBUG nova.virt.hardware [None req-276e84bf-a104-4323-ae2b-430f1d490aad tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T19:28:34Z,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1539f06fa3534e90acbc3a60c4b8bd3f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T19:28:35Z,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 771.099893] env[65726]: DEBUG nova.virt.hardware [None req-276e84bf-a104-4323-ae2b-430f1d490aad tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 771.100270] env[65726]: DEBUG nova.virt.hardware [None req-276e84bf-a104-4323-ae2b-430f1d490aad tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 771.100270] env[65726]: DEBUG nova.virt.hardware [None req-276e84bf-a104-4323-ae2b-430f1d490aad tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 771.101051] env[65726]: DEBUG nova.virt.hardware [None req-276e84bf-a104-4323-ae2b-430f1d490aad tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 771.101051] env[65726]: DEBUG nova.virt.hardware [None req-276e84bf-a104-4323-ae2b-430f1d490aad tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 771.101051] env[65726]: DEBUG nova.virt.hardware [None req-276e84bf-a104-4323-ae2b-430f1d490aad tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 771.101051] env[65726]: DEBUG nova.virt.hardware [None req-276e84bf-a104-4323-ae2b-430f1d490aad tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 771.101051] env[65726]: DEBUG nova.virt.hardware [None req-276e84bf-a104-4323-ae2b-430f1d490aad tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 771.101687] env[65726]: DEBUG nova.virt.hardware [None req-276e84bf-a104-4323-ae2b-430f1d490aad tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 771.101687] env[65726]: DEBUG nova.virt.hardware [None req-276e84bf-a104-4323-ae2b-430f1d490aad tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 771.110367] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-af2f0c1d-9bc5-40c3-a712-d408e73bc806 tempest-ServerDiagnosticsNegativeTest-1635132145 tempest-ServerDiagnosticsNegativeTest-1635132145-project-member] [instance: 0d6f14c5-5770-476f-a186-4cf1b2d4fbe8] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 771.110606] env[65726]: INFO nova.compute.manager [None req-af2f0c1d-9bc5-40c3-a712-d408e73bc806 tempest-ServerDiagnosticsNegativeTest-1635132145 tempest-ServerDiagnosticsNegativeTest-1635132145-project-member] [instance: 0d6f14c5-5770-476f-a186-4cf1b2d4fbe8] Took 11.75 seconds to spawn the instance on the hypervisor. [ 771.110782] env[65726]: DEBUG nova.compute.manager [None req-af2f0c1d-9bc5-40c3-a712-d408e73bc806 tempest-ServerDiagnosticsNegativeTest-1635132145 tempest-ServerDiagnosticsNegativeTest-1635132145-project-member] [instance: 0d6f14c5-5770-476f-a186-4cf1b2d4fbe8] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 771.111644] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1edbdeb-139d-4976-bacd-99ccd91ed5f6 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.120830] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c489149d-9585-4bbb-83c2-f88f28a64d5e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.126078] env[65726]: DEBUG oslo_vmware.api [None req-48153719-e51d-407a-98ef-93401c6270f2 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Waiting for the task: (returnval){ [ 771.126078] env[65726]: value = "task-5115833" [ 771.126078] env[65726]: _type = "Task" [ 771.126078] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 771.134666] env[65726]: DEBUG oslo_vmware.api [None req-6c979061-2d03-4525-8b87-afa801f4542e tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Waiting for the task: (returnval){ [ 771.134666] env[65726]: value = "task-5115834" [ 771.134666] env[65726]: _type = "Task" [ 771.134666] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 771.134666] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115832, 'name': CreateVM_Task} progress is 25%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 771.153453] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3afee7e3-1437-46c5-b6b4-23ebcdaa229d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.162930] env[65726]: DEBUG oslo_vmware.api [None req-48153719-e51d-407a-98ef-93401c6270f2 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Task: {'id': task-5115833, 'name': PowerOnVM_Task} progress is 33%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 771.180836] env[65726]: DEBUG oslo_vmware.api [None req-6c979061-2d03-4525-8b87-afa801f4542e tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Task: {'id': task-5115834, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 771.181773] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-276e84bf-a104-4323-ae2b-430f1d490aad tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] [instance: 3827eed1-b7a1-4922-a4de-f0c038cb8566] Instance VIF info [] {{(pid=65726) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 771.188408] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-276e84bf-a104-4323-ae2b-430f1d490aad tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 771.201645] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 771.201645] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 771.209115] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3827eed1-b7a1-4922-a4de-f0c038cb8566] Creating VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 771.211049] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ec8f298d-2696-41b6-b6e9-f9760868bb55 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.234219] env[65726]: WARNING openstack [req-da245a27-980c-420b-9648-990ba5e6a122 req-ba47e5fc-0777-4cc0-a801-66ce5cfe21e1 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 771.234622] env[65726]: WARNING openstack [req-da245a27-980c-420b-9648-990ba5e6a122 req-ba47e5fc-0777-4cc0-a801-66ce5cfe21e1 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 771.245972] env[65726]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 771.245972] env[65726]: value = "task-5115835" [ 771.245972] env[65726]: _type = "Task" [ 771.245972] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 771.259520] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115835, 'name': CreateVM_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 771.278778] env[65726]: WARNING openstack [req-767b54cd-afaa-479f-beab-3d140190858d req-67a7e1c2-d4f5-4635-9745-5320cfc755c6 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 771.279201] env[65726]: WARNING openstack [req-767b54cd-afaa-479f-beab-3d140190858d req-67a7e1c2-d4f5-4635-9745-5320cfc755c6 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 771.381170] env[65726]: WARNING openstack [req-da245a27-980c-420b-9648-990ba5e6a122 req-ba47e5fc-0777-4cc0-a801-66ce5cfe21e1 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 771.381596] env[65726]: WARNING openstack [req-da245a27-980c-420b-9648-990ba5e6a122 req-ba47e5fc-0777-4cc0-a801-66ce5cfe21e1 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 771.423892] env[65726]: WARNING openstack [req-767b54cd-afaa-479f-beab-3d140190858d req-67a7e1c2-d4f5-4635-9745-5320cfc755c6 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 771.424616] env[65726]: WARNING openstack [req-767b54cd-afaa-479f-beab-3d140190858d req-67a7e1c2-d4f5-4635-9745-5320cfc755c6 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 771.545147] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b86f24be-1076-4ee1-9368-176430971dfc tempest-VolumesAssistedSnapshotsTest-1477776432 tempest-VolumesAssistedSnapshotsTest-1477776432-project-admin] Lock "aa2716a9-2cf8-4e51-800d-1c8f98b2cf05" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.606s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 771.562396] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-703975dc-3e07-4b77-b05e-15d0411bc7cd tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 608cc0ea-3f6b-4b4f-83c1-01aa50999d63] Creating linked-clone VM from snapshot {{(pid=65726) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 771.562986] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-32c4099b-6f68-4f18-850a-2e5edc5009f3 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.572197] env[65726]: DEBUG oslo_vmware.api [None req-703975dc-3e07-4b77-b05e-15d0411bc7cd tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Waiting for the task: (returnval){ [ 771.572197] env[65726]: value = "task-5115836" [ 771.572197] env[65726]: _type = "Task" [ 771.572197] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 771.586423] env[65726]: DEBUG oslo_vmware.api [None req-703975dc-3e07-4b77-b05e-15d0411bc7cd tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5115836, 'name': CloneVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 771.592085] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115832, 'name': CreateVM_Task, 'duration_secs': 0.411364} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 771.593187] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9fc0c9ee-9dc7-4059-b8d2-a5db965e00c2] Created VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 771.593969] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72c17589-7f45-4e9d-a755-2b3bf2e93b6f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.597064] env[65726]: WARNING openstack [None req-d51d38e8-71be-4889-a206-d52403737076 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 771.597426] env[65726]: WARNING openstack [None req-d51d38e8-71be-4889-a206-d52403737076 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 771.602650] env[65726]: DEBUG oslo_concurrency.lockutils [None req-d51d38e8-71be-4889-a206-d52403737076 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 771.602844] env[65726]: DEBUG oslo_concurrency.lockutils [None req-d51d38e8-71be-4889-a206-d52403737076 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 771.603205] env[65726]: DEBUG oslo_concurrency.lockutils [None req-d51d38e8-71be-4889-a206-d52403737076 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 771.605540] env[65726]: DEBUG nova.network.neutron [req-da245a27-980c-420b-9648-990ba5e6a122 req-ba47e5fc-0777-4cc0-a801-66ce5cfe21e1 service nova] [instance: 6d2b86c2-0a44-433f-8b3f-2e9c7f693baa] Updated VIF entry in instance network info cache for port b487d53a-9bda-4e3c-be61-f26acbc8ae0f. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 771.605861] env[65726]: DEBUG nova.network.neutron [req-da245a27-980c-420b-9648-990ba5e6a122 req-ba47e5fc-0777-4cc0-a801-66ce5cfe21e1 service nova] [instance: 6d2b86c2-0a44-433f-8b3f-2e9c7f693baa] Updating instance_info_cache with network_info: [{"id": "b487d53a-9bda-4e3c-be61-f26acbc8ae0f", "address": "fa:16:3e:ad:db:36", "network": {"id": "9d5ad167-99fb-4208-ab30-7053f35774e4", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-2082516762-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a8362295a62f4b51bae719a7ef5a4656", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dabbac20-1723-40ad-9da0-e53b28073651", "external-id": "nsx-vlan-transportzone-790", "segmentation_id": 790, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb487d53a-9b", "ovs_interfaceid": "b487d53a-9bda-4e3c-be61-f26acbc8ae0f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 771.607188] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-090e2472-babf-4734-b04f-3416b5543d81 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.616620] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e141f1a2-841f-42c5-baef-0c13acacbade {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.620446] env[65726]: DEBUG oslo_vmware.api [None req-d51d38e8-71be-4889-a206-d52403737076 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Waiting for the task: (returnval){ [ 771.620446] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52620485-10d5-2507-1f5c-721092bee151" [ 771.620446] env[65726]: _type = "Task" [ 771.620446] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 771.653541] env[65726]: DEBUG nova.network.neutron [req-767b54cd-afaa-479f-beab-3d140190858d req-67a7e1c2-d4f5-4635-9745-5320cfc755c6 service nova] [instance: 895797e4-2941-44cd-aab1-67afa4fac02d] Updated VIF entry in instance network info cache for port a1dec2db-0e1e-45dd-8587-a84e4a5e2ef4. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 771.653939] env[65726]: DEBUG nova.network.neutron [req-767b54cd-afaa-479f-beab-3d140190858d req-67a7e1c2-d4f5-4635-9745-5320cfc755c6 service nova] [instance: 895797e4-2941-44cd-aab1-67afa4fac02d] Updating instance_info_cache with network_info: [{"id": "a1dec2db-0e1e-45dd-8587-a84e4a5e2ef4", "address": "fa:16:3e:b9:47:02", "network": {"id": "ad1373b5-e0ed-4ee6-a97c-af931afbab35", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-909418417-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.224", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a090d553766847e58e3231b966c92565", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5f60c972-a72d-4c5f-a250-faadfd6eafbe", "external-id": "nsx-vlan-transportzone-932", "segmentation_id": 932, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa1dec2db-0e", "ovs_interfaceid": "a1dec2db-0e1e-45dd-8587-a84e4a5e2ef4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 771.664058] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0344e2d6-b1d0-414e-9f8d-5554763d381a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.674902] env[65726]: DEBUG oslo_vmware.api [None req-d51d38e8-71be-4889-a206-d52403737076 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52620485-10d5-2507-1f5c-721092bee151, 'name': SearchDatastore_Task, 'duration_secs': 0.011662} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 771.678771] env[65726]: DEBUG oslo_concurrency.lockutils [None req-d51d38e8-71be-4889-a206-d52403737076 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 771.678771] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-d51d38e8-71be-4889-a206-d52403737076 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: 9fc0c9ee-9dc7-4059-b8d2-a5db965e00c2] Processing image b52362a3-ee8a-4cbf-b06f-513b0cc8f95c {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 771.678771] env[65726]: DEBUG oslo_concurrency.lockutils [None req-d51d38e8-71be-4889-a206-d52403737076 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 771.678771] env[65726]: DEBUG oslo_concurrency.lockutils [None req-d51d38e8-71be-4889-a206-d52403737076 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 771.678923] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-d51d38e8-71be-4889-a206-d52403737076 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 771.679784] env[65726]: INFO nova.compute.manager [None req-af2f0c1d-9bc5-40c3-a712-d408e73bc806 tempest-ServerDiagnosticsNegativeTest-1635132145 tempest-ServerDiagnosticsNegativeTest-1635132145-project-member] [instance: 0d6f14c5-5770-476f-a186-4cf1b2d4fbe8] Took 53.54 seconds to build instance. [ 771.681430] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9b14f3af-955a-4997-9e19-1e9278d23df3 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.694330] env[65726]: DEBUG oslo_vmware.api [None req-6c979061-2d03-4525-8b87-afa801f4542e tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Task: {'id': task-5115834, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.084939} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 771.694911] env[65726]: DEBUG oslo_vmware.api [None req-48153719-e51d-407a-98ef-93401c6270f2 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Task: {'id': task-5115833, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 771.695337] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-6c979061-2d03-4525-8b87-afa801f4542e tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: 6d2b86c2-0a44-433f-8b3f-2e9c7f693baa] Extended root virtual disk {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 771.696864] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13852fd7-c855-41fc-8e11-45c0f55e6346 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.702136] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-192d4feb-4f16-4d99-98aa-0db9c686fd8e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.708678] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-d51d38e8-71be-4889-a206-d52403737076 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 771.708678] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-d51d38e8-71be-4889-a206-d52403737076 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65726) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 771.708939] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a436948c-49cb-4451-8ddc-3c411de3b993 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.730282] env[65726]: DEBUG nova.compute.provider_tree [None req-5ae556a9-54a3-4f16-bd92-05c390a03c24 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 771.740912] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-6c979061-2d03-4525-8b87-afa801f4542e tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: 6d2b86c2-0a44-433f-8b3f-2e9c7f693baa] Reconfiguring VM instance instance-0000002b to attach disk [datastore2] 6d2b86c2-0a44-433f-8b3f-2e9c7f693baa/6d2b86c2-0a44-433f-8b3f-2e9c7f693baa.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 771.742838] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4c978704-cab0-421c-a9ac-98bb3c5a9553 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.760219] env[65726]: DEBUG oslo_vmware.api [None req-d51d38e8-71be-4889-a206-d52403737076 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Waiting for the task: (returnval){ [ 771.760219] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]529b7e8e-53ed-972b-d92a-f51066308bb7" [ 771.760219] env[65726]: _type = "Task" [ 771.760219] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 771.769819] env[65726]: DEBUG oslo_vmware.api [None req-6c979061-2d03-4525-8b87-afa801f4542e tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Waiting for the task: (returnval){ [ 771.769819] env[65726]: value = "task-5115837" [ 771.769819] env[65726]: _type = "Task" [ 771.769819] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 771.777286] env[65726]: DEBUG oslo_vmware.api [None req-d51d38e8-71be-4889-a206-d52403737076 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]529b7e8e-53ed-972b-d92a-f51066308bb7, 'name': SearchDatastore_Task, 'duration_secs': 0.011875} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 771.777606] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115835, 'name': CreateVM_Task, 'duration_secs': 0.349904} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 771.783253] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3827eed1-b7a1-4922-a4de-f0c038cb8566] Created VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 771.783527] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9bc8cf23-9941-4790-9d48-67f6c1bac332 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.787896] env[65726]: DEBUG oslo_concurrency.lockutils [None req-276e84bf-a104-4323-ae2b-430f1d490aad tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 771.787896] env[65726]: DEBUG oslo_concurrency.lockutils [None req-276e84bf-a104-4323-ae2b-430f1d490aad tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 771.787896] env[65726]: DEBUG oslo_concurrency.lockutils [None req-276e84bf-a104-4323-ae2b-430f1d490aad tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 771.788580] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a41454bd-f97a-4897-abb4-6f0a4d5bfdb6 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.795351] env[65726]: DEBUG oslo_vmware.api [None req-6c979061-2d03-4525-8b87-afa801f4542e tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Task: {'id': task-5115837, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 771.799582] env[65726]: DEBUG oslo_vmware.api [None req-d51d38e8-71be-4889-a206-d52403737076 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Waiting for the task: (returnval){ [ 771.799582] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]527506a8-cb13-6fad-5bb6-53af377d4ec7" [ 771.799582] env[65726]: _type = "Task" [ 771.799582] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 771.801283] env[65726]: DEBUG oslo_vmware.api [None req-276e84bf-a104-4323-ae2b-430f1d490aad tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] Waiting for the task: (returnval){ [ 771.801283] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]525a27c5-1a50-6bcc-5152-43a906926b22" [ 771.801283] env[65726]: _type = "Task" [ 771.801283] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 771.813091] env[65726]: DEBUG oslo_vmware.api [None req-d51d38e8-71be-4889-a206-d52403737076 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]527506a8-cb13-6fad-5bb6-53af377d4ec7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 771.816502] env[65726]: DEBUG oslo_vmware.api [None req-276e84bf-a104-4323-ae2b-430f1d490aad tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]525a27c5-1a50-6bcc-5152-43a906926b22, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 771.976788] env[65726]: DEBUG nova.network.neutron [-] [instance: 00746508-d0d6-4dfa-9026-772398d004ee] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 772.086895] env[65726]: DEBUG oslo_vmware.api [None req-703975dc-3e07-4b77-b05e-15d0411bc7cd tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5115836, 'name': CloneVM_Task} progress is 94%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 772.114386] env[65726]: DEBUG oslo_concurrency.lockutils [req-da245a27-980c-420b-9648-990ba5e6a122 req-ba47e5fc-0777-4cc0-a801-66ce5cfe21e1 service nova] Releasing lock "refresh_cache-6d2b86c2-0a44-433f-8b3f-2e9c7f693baa" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 772.114386] env[65726]: DEBUG nova.compute.manager [req-da245a27-980c-420b-9648-990ba5e6a122 req-ba47e5fc-0777-4cc0-a801-66ce5cfe21e1 service nova] [instance: 9fc0c9ee-9dc7-4059-b8d2-a5db965e00c2] Received event network-vif-plugged-e4315658-0732-463a-b063-b1c4cf7cdafd {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 772.114386] env[65726]: DEBUG oslo_concurrency.lockutils [req-da245a27-980c-420b-9648-990ba5e6a122 req-ba47e5fc-0777-4cc0-a801-66ce5cfe21e1 service nova] Acquiring lock "9fc0c9ee-9dc7-4059-b8d2-a5db965e00c2-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 772.114386] env[65726]: DEBUG oslo_concurrency.lockutils [req-da245a27-980c-420b-9648-990ba5e6a122 req-ba47e5fc-0777-4cc0-a801-66ce5cfe21e1 service nova] Lock "9fc0c9ee-9dc7-4059-b8d2-a5db965e00c2-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 772.114386] env[65726]: DEBUG oslo_concurrency.lockutils [req-da245a27-980c-420b-9648-990ba5e6a122 req-ba47e5fc-0777-4cc0-a801-66ce5cfe21e1 service nova] Lock "9fc0c9ee-9dc7-4059-b8d2-a5db965e00c2-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 772.114808] env[65726]: DEBUG nova.compute.manager [req-da245a27-980c-420b-9648-990ba5e6a122 req-ba47e5fc-0777-4cc0-a801-66ce5cfe21e1 service nova] [instance: 9fc0c9ee-9dc7-4059-b8d2-a5db965e00c2] No waiting events found dispatching network-vif-plugged-e4315658-0732-463a-b063-b1c4cf7cdafd {{(pid=65726) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 772.114808] env[65726]: WARNING nova.compute.manager [req-da245a27-980c-420b-9648-990ba5e6a122 req-ba47e5fc-0777-4cc0-a801-66ce5cfe21e1 service nova] [instance: 9fc0c9ee-9dc7-4059-b8d2-a5db965e00c2] Received unexpected event network-vif-plugged-e4315658-0732-463a-b063-b1c4cf7cdafd for instance with vm_state building and task_state spawning. [ 772.114808] env[65726]: DEBUG nova.compute.manager [req-da245a27-980c-420b-9648-990ba5e6a122 req-ba47e5fc-0777-4cc0-a801-66ce5cfe21e1 service nova] [instance: 9fc0c9ee-9dc7-4059-b8d2-a5db965e00c2] Received event network-changed-e4315658-0732-463a-b063-b1c4cf7cdafd {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 772.114808] env[65726]: DEBUG nova.compute.manager [req-da245a27-980c-420b-9648-990ba5e6a122 req-ba47e5fc-0777-4cc0-a801-66ce5cfe21e1 service nova] [instance: 9fc0c9ee-9dc7-4059-b8d2-a5db965e00c2] Refreshing instance network info cache due to event network-changed-e4315658-0732-463a-b063-b1c4cf7cdafd. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 772.114808] env[65726]: DEBUG oslo_concurrency.lockutils [req-da245a27-980c-420b-9648-990ba5e6a122 req-ba47e5fc-0777-4cc0-a801-66ce5cfe21e1 service nova] Acquiring lock "refresh_cache-9fc0c9ee-9dc7-4059-b8d2-a5db965e00c2" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 772.115126] env[65726]: DEBUG oslo_concurrency.lockutils [req-da245a27-980c-420b-9648-990ba5e6a122 req-ba47e5fc-0777-4cc0-a801-66ce5cfe21e1 service nova] Acquired lock "refresh_cache-9fc0c9ee-9dc7-4059-b8d2-a5db965e00c2" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 772.115126] env[65726]: DEBUG nova.network.neutron [req-da245a27-980c-420b-9648-990ba5e6a122 req-ba47e5fc-0777-4cc0-a801-66ce5cfe21e1 service nova] [instance: 9fc0c9ee-9dc7-4059-b8d2-a5db965e00c2] Refreshing network info cache for port e4315658-0732-463a-b063-b1c4cf7cdafd {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 772.164442] env[65726]: DEBUG oslo_vmware.api [None req-48153719-e51d-407a-98ef-93401c6270f2 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Task: {'id': task-5115833, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 772.167130] env[65726]: DEBUG oslo_concurrency.lockutils [req-767b54cd-afaa-479f-beab-3d140190858d req-67a7e1c2-d4f5-4635-9745-5320cfc755c6 service nova] Releasing lock "refresh_cache-895797e4-2941-44cd-aab1-67afa4fac02d" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 772.190257] env[65726]: DEBUG oslo_concurrency.lockutils [None req-af2f0c1d-9bc5-40c3-a712-d408e73bc806 tempest-ServerDiagnosticsNegativeTest-1635132145 tempest-ServerDiagnosticsNegativeTest-1635132145-project-member] Lock "0d6f14c5-5770-476f-a186-4cf1b2d4fbe8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 89.089s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 772.242910] env[65726]: DEBUG nova.scheduler.client.report [None req-5ae556a9-54a3-4f16-bd92-05c390a03c24 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 772.291106] env[65726]: DEBUG oslo_vmware.api [None req-6c979061-2d03-4525-8b87-afa801f4542e tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Task: {'id': task-5115837, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 772.315532] env[65726]: DEBUG oslo_vmware.api [None req-d51d38e8-71be-4889-a206-d52403737076 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]527506a8-cb13-6fad-5bb6-53af377d4ec7, 'name': SearchDatastore_Task, 'duration_secs': 0.015897} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 772.319676] env[65726]: DEBUG oslo_concurrency.lockutils [None req-d51d38e8-71be-4889-a206-d52403737076 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 772.319896] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-d51d38e8-71be-4889-a206-d52403737076 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore2] 9fc0c9ee-9dc7-4059-b8d2-a5db965e00c2/9fc0c9ee-9dc7-4059-b8d2-a5db965e00c2.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 772.320107] env[65726]: DEBUG oslo_vmware.api [None req-276e84bf-a104-4323-ae2b-430f1d490aad tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]525a27c5-1a50-6bcc-5152-43a906926b22, 'name': SearchDatastore_Task, 'duration_secs': 0.021137} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 772.320666] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7cf26e7f-51b6-4074-b0df-a703df07e880 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.322798] env[65726]: DEBUG oslo_concurrency.lockutils [None req-276e84bf-a104-4323-ae2b-430f1d490aad tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 772.323037] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-276e84bf-a104-4323-ae2b-430f1d490aad tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] [instance: 3827eed1-b7a1-4922-a4de-f0c038cb8566] Processing image b52362a3-ee8a-4cbf-b06f-513b0cc8f95c {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 772.323272] env[65726]: DEBUG oslo_concurrency.lockutils [None req-276e84bf-a104-4323-ae2b-430f1d490aad tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 772.323456] env[65726]: DEBUG oslo_concurrency.lockutils [None req-276e84bf-a104-4323-ae2b-430f1d490aad tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 772.323642] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-276e84bf-a104-4323-ae2b-430f1d490aad tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 772.323998] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f1cb0b77-f664-44ff-8fcf-b095994ffa1c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.331725] env[65726]: DEBUG oslo_vmware.api [None req-d51d38e8-71be-4889-a206-d52403737076 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Waiting for the task: (returnval){ [ 772.331725] env[65726]: value = "task-5115838" [ 772.331725] env[65726]: _type = "Task" [ 772.331725] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 772.336764] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-276e84bf-a104-4323-ae2b-430f1d490aad tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 772.337018] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-276e84bf-a104-4323-ae2b-430f1d490aad tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65726) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 772.338309] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c3a94f51-0ffa-44f5-ad12-3776fedbfd03 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.346352] env[65726]: DEBUG oslo_vmware.api [None req-d51d38e8-71be-4889-a206-d52403737076 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Task: {'id': task-5115838, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 772.352023] env[65726]: DEBUG oslo_vmware.api [None req-276e84bf-a104-4323-ae2b-430f1d490aad tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] Waiting for the task: (returnval){ [ 772.352023] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52d86aba-c2fb-dc68-d9c9-0cbd9a94113a" [ 772.352023] env[65726]: _type = "Task" [ 772.352023] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 772.361413] env[65726]: DEBUG oslo_vmware.api [None req-276e84bf-a104-4323-ae2b-430f1d490aad tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52d86aba-c2fb-dc68-d9c9-0cbd9a94113a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 772.481038] env[65726]: INFO nova.compute.manager [-] [instance: 00746508-d0d6-4dfa-9026-772398d004ee] Took 1.40 seconds to deallocate network for instance. [ 772.527730] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b86f24be-1076-4ee1-9368-176430971dfc tempest-VolumesAssistedSnapshotsTest-1477776432 tempest-VolumesAssistedSnapshotsTest-1477776432-project-admin] Acquiring lock "aa2716a9-2cf8-4e51-800d-1c8f98b2cf05" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 772.527730] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b86f24be-1076-4ee1-9368-176430971dfc tempest-VolumesAssistedSnapshotsTest-1477776432 tempest-VolumesAssistedSnapshotsTest-1477776432-project-admin] Lock "aa2716a9-2cf8-4e51-800d-1c8f98b2cf05" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 772.527730] env[65726]: INFO nova.compute.manager [None req-b86f24be-1076-4ee1-9368-176430971dfc tempest-VolumesAssistedSnapshotsTest-1477776432 tempest-VolumesAssistedSnapshotsTest-1477776432-project-admin] [instance: aa2716a9-2cf8-4e51-800d-1c8f98b2cf05] Attaching volume b7243e8e-0979-43c1-87dc-2570fca06458 to /dev/sdb [ 772.580435] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ce39159-2cc2-45e4-8be7-824538aeb7f6 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.590306] env[65726]: DEBUG oslo_vmware.api [None req-703975dc-3e07-4b77-b05e-15d0411bc7cd tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5115836, 'name': CloneVM_Task} progress is 94%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 772.592356] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4c3f997-6fef-42a3-bdd3-dd2f75a02903 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.608937] env[65726]: DEBUG nova.virt.block_device [None req-b86f24be-1076-4ee1-9368-176430971dfc tempest-VolumesAssistedSnapshotsTest-1477776432 tempest-VolumesAssistedSnapshotsTest-1477776432-project-admin] [instance: aa2716a9-2cf8-4e51-800d-1c8f98b2cf05] Updating existing volume attachment record: 6cff00df-d0bf-48dd-b7b9-d5fd7f7b72c2 {{(pid=65726) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 772.618102] env[65726]: WARNING openstack [req-da245a27-980c-420b-9648-990ba5e6a122 req-ba47e5fc-0777-4cc0-a801-66ce5cfe21e1 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 772.618560] env[65726]: WARNING openstack [req-da245a27-980c-420b-9648-990ba5e6a122 req-ba47e5fc-0777-4cc0-a801-66ce5cfe21e1 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 772.664515] env[65726]: DEBUG oslo_vmware.api [None req-48153719-e51d-407a-98ef-93401c6270f2 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Task: {'id': task-5115833, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 772.693969] env[65726]: DEBUG nova.compute.manager [None req-ae076e22-617f-47ad-a346-b93664495097 tempest-ServersNegativeTestMultiTenantJSON-944098855 tempest-ServersNegativeTestMultiTenantJSON-944098855-project-member] [instance: 018c8ac5-66c4-4a9a-ab45-85e84f92b4fe] Starting instance... {{(pid=65726) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 772.747879] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5ae556a9-54a3-4f16-bd92-05c390a03c24 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 4.763s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 772.751371] env[65726]: DEBUG oslo_concurrency.lockutils [None req-462b38ac-85e6-4728-81f8-4b729f9b1c26 tempest-ServersAaction247Test-712120018 tempest-ServersAaction247Test-712120018-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 36.217s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 772.752687] env[65726]: INFO nova.compute.claims [None req-462b38ac-85e6-4728-81f8-4b729f9b1c26 tempest-ServersAaction247Test-712120018 tempest-ServersAaction247Test-712120018-project-member] [instance: ad48cadd-9bb6-4191-a68a-5c8fdaeebd44] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 772.789117] env[65726]: INFO nova.scheduler.client.report [None req-5ae556a9-54a3-4f16-bd92-05c390a03c24 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Deleted allocations for instance fd481728-3ef0-4a96-affd-ab2dd3f596bb [ 772.806591] env[65726]: DEBUG oslo_vmware.api [None req-6c979061-2d03-4525-8b87-afa801f4542e tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Task: {'id': task-5115837, 'name': ReconfigVM_Task, 'duration_secs': 0.780426} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 772.807177] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-6c979061-2d03-4525-8b87-afa801f4542e tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: 6d2b86c2-0a44-433f-8b3f-2e9c7f693baa] Reconfigured VM instance instance-0000002b to attach disk [datastore2] 6d2b86c2-0a44-433f-8b3f-2e9c7f693baa/6d2b86c2-0a44-433f-8b3f-2e9c7f693baa.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 772.808520] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9eda6333-4cf6-4a57-bbe3-1d02d00c45b1 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.817929] env[65726]: DEBUG oslo_vmware.api [None req-6c979061-2d03-4525-8b87-afa801f4542e tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Waiting for the task: (returnval){ [ 772.817929] env[65726]: value = "task-5115840" [ 772.817929] env[65726]: _type = "Task" [ 772.817929] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 772.830672] env[65726]: DEBUG oslo_vmware.api [None req-6c979061-2d03-4525-8b87-afa801f4542e tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Task: {'id': task-5115840, 'name': Rename_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 772.843806] env[65726]: DEBUG oslo_vmware.api [None req-d51d38e8-71be-4889-a206-d52403737076 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Task: {'id': task-5115838, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 772.864097] env[65726]: DEBUG oslo_vmware.api [None req-276e84bf-a104-4323-ae2b-430f1d490aad tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52d86aba-c2fb-dc68-d9c9-0cbd9a94113a, 'name': SearchDatastore_Task, 'duration_secs': 0.011146} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 772.865028] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f33ea533-c197-4fc1-9935-221b6a2f7473 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.874369] env[65726]: DEBUG oslo_vmware.api [None req-276e84bf-a104-4323-ae2b-430f1d490aad tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] Waiting for the task: (returnval){ [ 772.874369] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]525458f4-da06-8b4b-7bb0-8703ea2449ae" [ 772.874369] env[65726]: _type = "Task" [ 772.874369] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 772.889907] env[65726]: DEBUG oslo_vmware.api [None req-276e84bf-a104-4323-ae2b-430f1d490aad tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]525458f4-da06-8b4b-7bb0-8703ea2449ae, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 772.898131] env[65726]: WARNING openstack [req-da245a27-980c-420b-9648-990ba5e6a122 req-ba47e5fc-0777-4cc0-a801-66ce5cfe21e1 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 772.898683] env[65726]: WARNING openstack [req-da245a27-980c-420b-9648-990ba5e6a122 req-ba47e5fc-0777-4cc0-a801-66ce5cfe21e1 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 772.987854] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9ac23e9e-c303-4029-8e2c-3f5115af48ab tempest-ServerAddressesNegativeTestJSON-963263679 tempest-ServerAddressesNegativeTestJSON-963263679-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 772.999195] env[65726]: WARNING openstack [req-da245a27-980c-420b-9648-990ba5e6a122 req-ba47e5fc-0777-4cc0-a801-66ce5cfe21e1 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 772.999602] env[65726]: WARNING openstack [req-da245a27-980c-420b-9648-990ba5e6a122 req-ba47e5fc-0777-4cc0-a801-66ce5cfe21e1 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 773.087881] env[65726]: DEBUG oslo_vmware.api [None req-703975dc-3e07-4b77-b05e-15d0411bc7cd tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5115836, 'name': CloneVM_Task} progress is 95%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 773.129165] env[65726]: DEBUG nova.network.neutron [req-da245a27-980c-420b-9648-990ba5e6a122 req-ba47e5fc-0777-4cc0-a801-66ce5cfe21e1 service nova] [instance: 9fc0c9ee-9dc7-4059-b8d2-a5db965e00c2] Updated VIF entry in instance network info cache for port e4315658-0732-463a-b063-b1c4cf7cdafd. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 773.129596] env[65726]: DEBUG nova.network.neutron [req-da245a27-980c-420b-9648-990ba5e6a122 req-ba47e5fc-0777-4cc0-a801-66ce5cfe21e1 service nova] [instance: 9fc0c9ee-9dc7-4059-b8d2-a5db965e00c2] Updating instance_info_cache with network_info: [{"id": "e4315658-0732-463a-b063-b1c4cf7cdafd", "address": "fa:16:3e:c1:e7:e1", "network": {"id": "9d5ad167-99fb-4208-ab30-7053f35774e4", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-2082516762-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a8362295a62f4b51bae719a7ef5a4656", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dabbac20-1723-40ad-9da0-e53b28073651", "external-id": "nsx-vlan-transportzone-790", "segmentation_id": 790, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape4315658-07", "ovs_interfaceid": "e4315658-0732-463a-b063-b1c4cf7cdafd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 773.163085] env[65726]: DEBUG oslo_vmware.api [None req-48153719-e51d-407a-98ef-93401c6270f2 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Task: {'id': task-5115833, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 773.226108] env[65726]: DEBUG oslo_concurrency.lockutils [None req-ae076e22-617f-47ad-a346-b93664495097 tempest-ServersNegativeTestMultiTenantJSON-944098855 tempest-ServersNegativeTestMultiTenantJSON-944098855-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 773.289788] env[65726]: DEBUG nova.compute.manager [req-d74dc8bc-58a3-43b7-9e52-7a0203478e32 req-850141b4-fe39-4ca6-adef-441408cfe2b1 service nova] [instance: 00746508-d0d6-4dfa-9026-772398d004ee] Received event network-vif-deleted-1151551e-12f0-4e85-a385-f1ebf4e9bc72 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 773.310649] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5ae556a9-54a3-4f16-bd92-05c390a03c24 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Lock "fd481728-3ef0-4a96-affd-ab2dd3f596bb" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 41.479s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 773.329512] env[65726]: DEBUG oslo_vmware.api [None req-6c979061-2d03-4525-8b87-afa801f4542e tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Task: {'id': task-5115840, 'name': Rename_Task, 'duration_secs': 0.209931} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 773.330630] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-6c979061-2d03-4525-8b87-afa801f4542e tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: 6d2b86c2-0a44-433f-8b3f-2e9c7f693baa] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 773.330833] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9982e833-30d1-4fe2-93d6-051296e2a680 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.340018] env[65726]: DEBUG oslo_vmware.api [None req-6c979061-2d03-4525-8b87-afa801f4542e tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Waiting for the task: (returnval){ [ 773.340018] env[65726]: value = "task-5115843" [ 773.340018] env[65726]: _type = "Task" [ 773.340018] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 773.345408] env[65726]: DEBUG oslo_vmware.api [None req-d51d38e8-71be-4889-a206-d52403737076 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Task: {'id': task-5115838, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.685143} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 773.346056] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-d51d38e8-71be-4889-a206-d52403737076 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore2] 9fc0c9ee-9dc7-4059-b8d2-a5db965e00c2/9fc0c9ee-9dc7-4059-b8d2-a5db965e00c2.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 773.346244] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-d51d38e8-71be-4889-a206-d52403737076 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: 9fc0c9ee-9dc7-4059-b8d2-a5db965e00c2] Extending root virtual disk to 1048576 {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 773.346495] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-10fe3c00-71ca-483d-944f-16324cc2e9c9 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.354288] env[65726]: DEBUG oslo_vmware.api [None req-6c979061-2d03-4525-8b87-afa801f4542e tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Task: {'id': task-5115843, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 773.356212] env[65726]: DEBUG oslo_vmware.api [None req-d51d38e8-71be-4889-a206-d52403737076 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Waiting for the task: (returnval){ [ 773.356212] env[65726]: value = "task-5115844" [ 773.356212] env[65726]: _type = "Task" [ 773.356212] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 773.372955] env[65726]: DEBUG oslo_vmware.api [None req-d51d38e8-71be-4889-a206-d52403737076 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Task: {'id': task-5115844, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 773.393525] env[65726]: DEBUG oslo_vmware.api [None req-276e84bf-a104-4323-ae2b-430f1d490aad tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]525458f4-da06-8b4b-7bb0-8703ea2449ae, 'name': SearchDatastore_Task, 'duration_secs': 0.039037} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 773.393973] env[65726]: DEBUG oslo_concurrency.lockutils [None req-276e84bf-a104-4323-ae2b-430f1d490aad tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 773.394379] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-276e84bf-a104-4323-ae2b-430f1d490aad tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] 3827eed1-b7a1-4922-a4de-f0c038cb8566/3827eed1-b7a1-4922-a4de-f0c038cb8566.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 773.394788] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8059fbbd-5162-40fa-bf24-6b1079fcae19 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.404741] env[65726]: DEBUG oslo_vmware.api [None req-276e84bf-a104-4323-ae2b-430f1d490aad tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] Waiting for the task: (returnval){ [ 773.404741] env[65726]: value = "task-5115845" [ 773.404741] env[65726]: _type = "Task" [ 773.404741] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 773.423201] env[65726]: DEBUG oslo_vmware.api [None req-276e84bf-a104-4323-ae2b-430f1d490aad tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] Task: {'id': task-5115845, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 773.587263] env[65726]: DEBUG oslo_vmware.api [None req-703975dc-3e07-4b77-b05e-15d0411bc7cd tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5115836, 'name': CloneVM_Task, 'duration_secs': 1.787685} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 773.587824] env[65726]: INFO nova.virt.vmwareapi.vmops [None req-703975dc-3e07-4b77-b05e-15d0411bc7cd tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 608cc0ea-3f6b-4b4f-83c1-01aa50999d63] Created linked-clone VM from snapshot [ 773.588741] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af69d761-e606-4402-ac9f-6ff718f10a1e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.597965] env[65726]: DEBUG nova.virt.vmwareapi.images [None req-703975dc-3e07-4b77-b05e-15d0411bc7cd tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 608cc0ea-3f6b-4b4f-83c1-01aa50999d63] Uploading image 5554ed4e-2d17-4d6f-946f-e0ce42b880fc {{(pid=65726) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 773.636381] env[65726]: DEBUG oslo_vmware.rw_handles [None req-703975dc-3e07-4b77-b05e-15d0411bc7cd tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 773.636381] env[65726]: value = "vm-995126" [ 773.636381] env[65726]: _type = "VirtualMachine" [ 773.636381] env[65726]: }. {{(pid=65726) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 773.637230] env[65726]: DEBUG oslo_concurrency.lockutils [req-da245a27-980c-420b-9648-990ba5e6a122 req-ba47e5fc-0777-4cc0-a801-66ce5cfe21e1 service nova] Releasing lock "refresh_cache-9fc0c9ee-9dc7-4059-b8d2-a5db965e00c2" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 773.640827] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-d3683f44-247b-45dd-a962-0d5b847bc833 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.648323] env[65726]: DEBUG oslo_vmware.rw_handles [None req-703975dc-3e07-4b77-b05e-15d0411bc7cd tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Lease: (returnval){ [ 773.648323] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]524c8784-b31a-8b27-42da-405625604eac" [ 773.648323] env[65726]: _type = "HttpNfcLease" [ 773.648323] env[65726]: } obtained for exporting VM: (result){ [ 773.648323] env[65726]: value = "vm-995126" [ 773.648323] env[65726]: _type = "VirtualMachine" [ 773.648323] env[65726]: }. {{(pid=65726) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 773.648323] env[65726]: DEBUG oslo_vmware.api [None req-703975dc-3e07-4b77-b05e-15d0411bc7cd tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Waiting for the lease: (returnval){ [ 773.648323] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]524c8784-b31a-8b27-42da-405625604eac" [ 773.648323] env[65726]: _type = "HttpNfcLease" [ 773.648323] env[65726]: } to be ready. {{(pid=65726) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 773.663865] env[65726]: DEBUG oslo_vmware.api [None req-48153719-e51d-407a-98ef-93401c6270f2 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Task: {'id': task-5115833, 'name': PowerOnVM_Task, 'duration_secs': 2.17664} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 773.667395] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-48153719-e51d-407a-98ef-93401c6270f2 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] [instance: 65aaa2eb-3ca5-4c2b-92f9-b524366c10dd] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 773.667876] env[65726]: INFO nova.compute.manager [None req-48153719-e51d-407a-98ef-93401c6270f2 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] [instance: 65aaa2eb-3ca5-4c2b-92f9-b524366c10dd] Took 11.53 seconds to spawn the instance on the hypervisor. [ 773.668204] env[65726]: DEBUG nova.compute.manager [None req-48153719-e51d-407a-98ef-93401c6270f2 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] [instance: 65aaa2eb-3ca5-4c2b-92f9-b524366c10dd] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 773.668609] env[65726]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 773.668609] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]524c8784-b31a-8b27-42da-405625604eac" [ 773.668609] env[65726]: _type = "HttpNfcLease" [ 773.668609] env[65726]: } is initializing. {{(pid=65726) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 773.669663] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fc193fa-03fc-4f23-ad7a-80a4a988b8f7 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.855520] env[65726]: DEBUG oslo_vmware.api [None req-6c979061-2d03-4525-8b87-afa801f4542e tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Task: {'id': task-5115843, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 773.873056] env[65726]: DEBUG oslo_vmware.api [None req-d51d38e8-71be-4889-a206-d52403737076 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Task: {'id': task-5115844, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.224981} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 773.873237] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-d51d38e8-71be-4889-a206-d52403737076 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: 9fc0c9ee-9dc7-4059-b8d2-a5db965e00c2] Extended root virtual disk {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 773.874561] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a031f3f6-d863-4506-85ce-73b6a6fb2d1f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.909270] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-d51d38e8-71be-4889-a206-d52403737076 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: 9fc0c9ee-9dc7-4059-b8d2-a5db965e00c2] Reconfiguring VM instance instance-0000002c to attach disk [datastore2] 9fc0c9ee-9dc7-4059-b8d2-a5db965e00c2/9fc0c9ee-9dc7-4059-b8d2-a5db965e00c2.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 773.913106] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bc5c02b9-d7cf-43c8-857c-1581f9514545 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.942127] env[65726]: DEBUG oslo_vmware.api [None req-276e84bf-a104-4323-ae2b-430f1d490aad tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] Task: {'id': task-5115845, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 773.942810] env[65726]: DEBUG oslo_vmware.api [None req-d51d38e8-71be-4889-a206-d52403737076 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Waiting for the task: (returnval){ [ 773.942810] env[65726]: value = "task-5115847" [ 773.942810] env[65726]: _type = "Task" [ 773.942810] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 773.955579] env[65726]: DEBUG oslo_vmware.api [None req-d51d38e8-71be-4889-a206-d52403737076 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Task: {'id': task-5115847, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 774.161026] env[65726]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 774.161026] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]524c8784-b31a-8b27-42da-405625604eac" [ 774.161026] env[65726]: _type = "HttpNfcLease" [ 774.161026] env[65726]: } is ready. {{(pid=65726) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 774.162223] env[65726]: DEBUG oslo_vmware.rw_handles [None req-703975dc-3e07-4b77-b05e-15d0411bc7cd tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 774.162223] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]524c8784-b31a-8b27-42da-405625604eac" [ 774.162223] env[65726]: _type = "HttpNfcLease" [ 774.162223] env[65726]: }. {{(pid=65726) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 774.164121] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-486f1fda-48e6-406a-bc5f-87a1d5bbc704 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.177796] env[65726]: DEBUG oslo_vmware.rw_handles [None req-703975dc-3e07-4b77-b05e-15d0411bc7cd tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52c3d80b-f9e4-46fc-efe3-f28a9a498520/disk-0.vmdk from lease info. {{(pid=65726) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 774.178185] env[65726]: DEBUG oslo_vmware.rw_handles [None req-703975dc-3e07-4b77-b05e-15d0411bc7cd tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52c3d80b-f9e4-46fc-efe3-f28a9a498520/disk-0.vmdk for reading. {{(pid=65726) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 774.266226] env[65726]: INFO nova.compute.manager [None req-48153719-e51d-407a-98ef-93401c6270f2 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] [instance: 65aaa2eb-3ca5-4c2b-92f9-b524366c10dd] Took 47.21 seconds to build instance. [ 774.329556] env[65726]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-0c282432-de20-45be-b741-d6fc94c1e137 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.364151] env[65726]: DEBUG oslo_vmware.api [None req-6c979061-2d03-4525-8b87-afa801f4542e tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Task: {'id': task-5115843, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 774.429614] env[65726]: DEBUG oslo_vmware.api [None req-276e84bf-a104-4323-ae2b-430f1d490aad tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] Task: {'id': task-5115845, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 774.456740] env[65726]: DEBUG oslo_vmware.api [None req-d51d38e8-71be-4889-a206-d52403737076 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Task: {'id': task-5115847, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 774.621514] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe194de2-a0e1-483a-9e98-ffb322f72670 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.631298] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9486c112-6995-46f1-bf01-cbba2f5ffe34 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.667236] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d11356c-4553-4771-bd63-3bdac6a0b482 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.674476] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c5cf455-26d2-4dd2-9316-c86a5521b8db {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.690349] env[65726]: DEBUG nova.compute.provider_tree [None req-462b38ac-85e6-4728-81f8-4b729f9b1c26 tempest-ServersAaction247Test-712120018 tempest-ServersAaction247Test-712120018-project-member] Updating inventory in ProviderTree for provider 07c4692f-bdb4-4058-9173-ff9664830295 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 774.772390] env[65726]: DEBUG oslo_concurrency.lockutils [None req-7b44120c-3963-45cf-8eb3-04f0d2e08d90 tempest-ServerDiagnosticsNegativeTest-1635132145 tempest-ServerDiagnosticsNegativeTest-1635132145-project-member] Acquiring lock "0d6f14c5-5770-476f-a186-4cf1b2d4fbe8" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 774.772670] env[65726]: DEBUG oslo_concurrency.lockutils [None req-7b44120c-3963-45cf-8eb3-04f0d2e08d90 tempest-ServerDiagnosticsNegativeTest-1635132145 tempest-ServerDiagnosticsNegativeTest-1635132145-project-member] Lock "0d6f14c5-5770-476f-a186-4cf1b2d4fbe8" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 774.772838] env[65726]: DEBUG oslo_concurrency.lockutils [None req-7b44120c-3963-45cf-8eb3-04f0d2e08d90 tempest-ServerDiagnosticsNegativeTest-1635132145 tempest-ServerDiagnosticsNegativeTest-1635132145-project-member] Acquiring lock "0d6f14c5-5770-476f-a186-4cf1b2d4fbe8-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 774.773020] env[65726]: DEBUG oslo_concurrency.lockutils [None req-7b44120c-3963-45cf-8eb3-04f0d2e08d90 tempest-ServerDiagnosticsNegativeTest-1635132145 tempest-ServerDiagnosticsNegativeTest-1635132145-project-member] Lock "0d6f14c5-5770-476f-a186-4cf1b2d4fbe8-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 774.773415] env[65726]: DEBUG oslo_concurrency.lockutils [None req-7b44120c-3963-45cf-8eb3-04f0d2e08d90 tempest-ServerDiagnosticsNegativeTest-1635132145 tempest-ServerDiagnosticsNegativeTest-1635132145-project-member] Lock "0d6f14c5-5770-476f-a186-4cf1b2d4fbe8-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 774.779047] env[65726]: DEBUG oslo_concurrency.lockutils [None req-48153719-e51d-407a-98ef-93401c6270f2 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Lock "65aaa2eb-3ca5-4c2b-92f9-b524366c10dd" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 87.579s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 774.779047] env[65726]: INFO nova.compute.manager [None req-7b44120c-3963-45cf-8eb3-04f0d2e08d90 tempest-ServerDiagnosticsNegativeTest-1635132145 tempest-ServerDiagnosticsNegativeTest-1635132145-project-member] [instance: 0d6f14c5-5770-476f-a186-4cf1b2d4fbe8] Terminating instance [ 774.856552] env[65726]: DEBUG oslo_vmware.api [None req-6c979061-2d03-4525-8b87-afa801f4542e tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Task: {'id': task-5115843, 'name': PowerOnVM_Task, 'duration_secs': 1.098807} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 774.856947] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-6c979061-2d03-4525-8b87-afa801f4542e tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: 6d2b86c2-0a44-433f-8b3f-2e9c7f693baa] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 774.857312] env[65726]: INFO nova.compute.manager [None req-6c979061-2d03-4525-8b87-afa801f4542e tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: 6d2b86c2-0a44-433f-8b3f-2e9c7f693baa] Took 9.93 seconds to spawn the instance on the hypervisor. [ 774.857519] env[65726]: DEBUG nova.compute.manager [None req-6c979061-2d03-4525-8b87-afa801f4542e tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: 6d2b86c2-0a44-433f-8b3f-2e9c7f693baa] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 774.858558] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4fcd30ae-23fb-4703-8c3c-3cb6adc82165 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.932102] env[65726]: DEBUG oslo_vmware.api [None req-276e84bf-a104-4323-ae2b-430f1d490aad tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] Task: {'id': task-5115845, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.182807} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 774.932102] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-276e84bf-a104-4323-ae2b-430f1d490aad tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] 3827eed1-b7a1-4922-a4de-f0c038cb8566/3827eed1-b7a1-4922-a4de-f0c038cb8566.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 774.932102] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-276e84bf-a104-4323-ae2b-430f1d490aad tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] [instance: 3827eed1-b7a1-4922-a4de-f0c038cb8566] Extending root virtual disk to 1048576 {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 774.932102] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-573f9f16-b9f6-45af-89a0-6ae6d189fbd0 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.940869] env[65726]: DEBUG oslo_vmware.api [None req-276e84bf-a104-4323-ae2b-430f1d490aad tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] Waiting for the task: (returnval){ [ 774.940869] env[65726]: value = "task-5115849" [ 774.940869] env[65726]: _type = "Task" [ 774.940869] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 774.959134] env[65726]: DEBUG oslo_vmware.api [None req-276e84bf-a104-4323-ae2b-430f1d490aad tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] Task: {'id': task-5115849, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 774.962707] env[65726]: DEBUG oslo_vmware.api [None req-d51d38e8-71be-4889-a206-d52403737076 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Task: {'id': task-5115847, 'name': ReconfigVM_Task, 'duration_secs': 0.656922} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 774.965766] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-d51d38e8-71be-4889-a206-d52403737076 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: 9fc0c9ee-9dc7-4059-b8d2-a5db965e00c2] Reconfigured VM instance instance-0000002c to attach disk [datastore2] 9fc0c9ee-9dc7-4059-b8d2-a5db965e00c2/9fc0c9ee-9dc7-4059-b8d2-a5db965e00c2.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 774.967120] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1189fae2-db97-4363-a4e8-9ae05a7892a6 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.976762] env[65726]: DEBUG oslo_vmware.api [None req-d51d38e8-71be-4889-a206-d52403737076 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Waiting for the task: (returnval){ [ 774.976762] env[65726]: value = "task-5115850" [ 774.976762] env[65726]: _type = "Task" [ 774.976762] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 774.988693] env[65726]: DEBUG oslo_vmware.api [None req-d51d38e8-71be-4889-a206-d52403737076 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Task: {'id': task-5115850, 'name': Rename_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 775.235137] env[65726]: DEBUG nova.scheduler.client.report [None req-462b38ac-85e6-4728-81f8-4b729f9b1c26 tempest-ServersAaction247Test-712120018 tempest-ServersAaction247Test-712120018-project-member] Updated inventory for provider 07c4692f-bdb4-4058-9173-ff9664830295 with generation 75 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 775.235447] env[65726]: DEBUG nova.compute.provider_tree [None req-462b38ac-85e6-4728-81f8-4b729f9b1c26 tempest-ServersAaction247Test-712120018 tempest-ServersAaction247Test-712120018-project-member] Updating resource provider 07c4692f-bdb4-4058-9173-ff9664830295 generation from 75 to 76 during operation: update_inventory {{(pid=65726) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 775.235805] env[65726]: DEBUG nova.compute.provider_tree [None req-462b38ac-85e6-4728-81f8-4b729f9b1c26 tempest-ServersAaction247Test-712120018 tempest-ServersAaction247Test-712120018-project-member] Updating inventory in ProviderTree for provider 07c4692f-bdb4-4058-9173-ff9664830295 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 775.281811] env[65726]: DEBUG nova.compute.manager [None req-81594943-9858-4765-a371-42f9ef0cbec9 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] [instance: 3cbbe146-be3a-43d5-867d-4d669884758c] Starting instance... {{(pid=65726) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 775.284406] env[65726]: DEBUG nova.compute.manager [None req-7b44120c-3963-45cf-8eb3-04f0d2e08d90 tempest-ServerDiagnosticsNegativeTest-1635132145 tempest-ServerDiagnosticsNegativeTest-1635132145-project-member] [instance: 0d6f14c5-5770-476f-a186-4cf1b2d4fbe8] Start destroying the instance on the hypervisor. {{(pid=65726) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 775.284843] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-7b44120c-3963-45cf-8eb3-04f0d2e08d90 tempest-ServerDiagnosticsNegativeTest-1635132145 tempest-ServerDiagnosticsNegativeTest-1635132145-project-member] [instance: 0d6f14c5-5770-476f-a186-4cf1b2d4fbe8] Destroying instance {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 775.286185] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d29fce28-171d-41b7-9318-66c5056a6a2c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.296912] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-7b44120c-3963-45cf-8eb3-04f0d2e08d90 tempest-ServerDiagnosticsNegativeTest-1635132145 tempest-ServerDiagnosticsNegativeTest-1635132145-project-member] [instance: 0d6f14c5-5770-476f-a186-4cf1b2d4fbe8] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 775.297396] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-baafaef5-3fb1-4905-80b5-752ebe3174e8 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.306121] env[65726]: DEBUG oslo_vmware.api [None req-7b44120c-3963-45cf-8eb3-04f0d2e08d90 tempest-ServerDiagnosticsNegativeTest-1635132145 tempest-ServerDiagnosticsNegativeTest-1635132145-project-member] Waiting for the task: (returnval){ [ 775.306121] env[65726]: value = "task-5115851" [ 775.306121] env[65726]: _type = "Task" [ 775.306121] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 775.316487] env[65726]: DEBUG oslo_vmware.api [None req-7b44120c-3963-45cf-8eb3-04f0d2e08d90 tempest-ServerDiagnosticsNegativeTest-1635132145 tempest-ServerDiagnosticsNegativeTest-1635132145-project-member] Task: {'id': task-5115851, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 775.381347] env[65726]: INFO nova.compute.manager [None req-6c979061-2d03-4525-8b87-afa801f4542e tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: 6d2b86c2-0a44-433f-8b3f-2e9c7f693baa] Took 47.99 seconds to build instance. [ 775.456722] env[65726]: DEBUG oslo_vmware.api [None req-276e84bf-a104-4323-ae2b-430f1d490aad tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] Task: {'id': task-5115849, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.187728} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 775.456986] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-276e84bf-a104-4323-ae2b-430f1d490aad tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] [instance: 3827eed1-b7a1-4922-a4de-f0c038cb8566] Extended root virtual disk {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 775.458162] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69ef0f46-1665-4922-b6ea-214ba25306c6 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.482123] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-276e84bf-a104-4323-ae2b-430f1d490aad tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] [instance: 3827eed1-b7a1-4922-a4de-f0c038cb8566] Reconfiguring VM instance instance-00000025 to attach disk [datastore1] 3827eed1-b7a1-4922-a4de-f0c038cb8566/3827eed1-b7a1-4922-a4de-f0c038cb8566.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 775.482539] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2a6e0f5e-1ce0-42e3-a3b4-2811406bc700 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.509318] env[65726]: DEBUG oslo_vmware.api [None req-d51d38e8-71be-4889-a206-d52403737076 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Task: {'id': task-5115850, 'name': Rename_Task, 'duration_secs': 0.22849} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 775.510062] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-d51d38e8-71be-4889-a206-d52403737076 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: 9fc0c9ee-9dc7-4059-b8d2-a5db965e00c2] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 775.513893] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-70f97ddf-3d5f-4963-8dae-6867e6e3f659 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.515465] env[65726]: DEBUG oslo_vmware.api [None req-276e84bf-a104-4323-ae2b-430f1d490aad tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] Waiting for the task: (returnval){ [ 775.515465] env[65726]: value = "task-5115852" [ 775.515465] env[65726]: _type = "Task" [ 775.515465] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 775.525183] env[65726]: DEBUG oslo_vmware.api [None req-d51d38e8-71be-4889-a206-d52403737076 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Waiting for the task: (returnval){ [ 775.525183] env[65726]: value = "task-5115853" [ 775.525183] env[65726]: _type = "Task" [ 775.525183] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 775.533389] env[65726]: DEBUG oslo_vmware.api [None req-276e84bf-a104-4323-ae2b-430f1d490aad tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] Task: {'id': task-5115852, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 775.537284] env[65726]: DEBUG oslo_vmware.api [None req-d51d38e8-71be-4889-a206-d52403737076 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Task: {'id': task-5115853, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 775.745060] env[65726]: DEBUG oslo_concurrency.lockutils [None req-462b38ac-85e6-4728-81f8-4b729f9b1c26 tempest-ServersAaction247Test-712120018 tempest-ServersAaction247Test-712120018-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.992s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 775.745060] env[65726]: DEBUG nova.compute.manager [None req-462b38ac-85e6-4728-81f8-4b729f9b1c26 tempest-ServersAaction247Test-712120018 tempest-ServersAaction247Test-712120018-project-member] [instance: ad48cadd-9bb6-4191-a68a-5c8fdaeebd44] Start building networks asynchronously for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 775.747100] env[65726]: DEBUG oslo_concurrency.lockutils [None req-11e4b8a8-60d9-4939-a6a9-6a357b669b13 tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 33.958s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 775.747490] env[65726]: DEBUG nova.objects.instance [None req-11e4b8a8-60d9-4939-a6a9-6a357b669b13 tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] [instance: 83573c2b-9448-456f-8fd2-b19661dd6cc4] Trying to apply a migration context that does not seem to be set for this instance {{(pid=65726) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 775.814659] env[65726]: DEBUG oslo_concurrency.lockutils [None req-81594943-9858-4765-a371-42f9ef0cbec9 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 775.820181] env[65726]: DEBUG oslo_vmware.api [None req-7b44120c-3963-45cf-8eb3-04f0d2e08d90 tempest-ServerDiagnosticsNegativeTest-1635132145 tempest-ServerDiagnosticsNegativeTest-1635132145-project-member] Task: {'id': task-5115851, 'name': PowerOffVM_Task, 'duration_secs': 0.363678} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 775.820181] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-7b44120c-3963-45cf-8eb3-04f0d2e08d90 tempest-ServerDiagnosticsNegativeTest-1635132145 tempest-ServerDiagnosticsNegativeTest-1635132145-project-member] [instance: 0d6f14c5-5770-476f-a186-4cf1b2d4fbe8] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 775.820308] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-7b44120c-3963-45cf-8eb3-04f0d2e08d90 tempest-ServerDiagnosticsNegativeTest-1635132145 tempest-ServerDiagnosticsNegativeTest-1635132145-project-member] [instance: 0d6f14c5-5770-476f-a186-4cf1b2d4fbe8] Unregistering the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 775.820928] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-bdf14ee3-8391-407d-9aaf-4ad39660f6f3 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.882554] env[65726]: DEBUG oslo_concurrency.lockutils [None req-6c979061-2d03-4525-8b87-afa801f4542e tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Lock "6d2b86c2-0a44-433f-8b3f-2e9c7f693baa" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 77.857s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 775.925630] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-7b44120c-3963-45cf-8eb3-04f0d2e08d90 tempest-ServerDiagnosticsNegativeTest-1635132145 tempest-ServerDiagnosticsNegativeTest-1635132145-project-member] [instance: 0d6f14c5-5770-476f-a186-4cf1b2d4fbe8] Unregistered the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 775.926187] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-7b44120c-3963-45cf-8eb3-04f0d2e08d90 tempest-ServerDiagnosticsNegativeTest-1635132145 tempest-ServerDiagnosticsNegativeTest-1635132145-project-member] [instance: 0d6f14c5-5770-476f-a186-4cf1b2d4fbe8] Deleting contents of the VM from datastore datastore2 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 775.926412] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-7b44120c-3963-45cf-8eb3-04f0d2e08d90 tempest-ServerDiagnosticsNegativeTest-1635132145 tempest-ServerDiagnosticsNegativeTest-1635132145-project-member] Deleting the datastore file [datastore2] 0d6f14c5-5770-476f-a186-4cf1b2d4fbe8 {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 775.926703] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a77b16b0-2be4-473a-a5e4-ae77bdd12750 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.939155] env[65726]: DEBUG oslo_vmware.api [None req-7b44120c-3963-45cf-8eb3-04f0d2e08d90 tempest-ServerDiagnosticsNegativeTest-1635132145 tempest-ServerDiagnosticsNegativeTest-1635132145-project-member] Waiting for the task: (returnval){ [ 775.939155] env[65726]: value = "task-5115855" [ 775.939155] env[65726]: _type = "Task" [ 775.939155] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 775.951018] env[65726]: DEBUG oslo_vmware.api [None req-7b44120c-3963-45cf-8eb3-04f0d2e08d90 tempest-ServerDiagnosticsNegativeTest-1635132145 tempest-ServerDiagnosticsNegativeTest-1635132145-project-member] Task: {'id': task-5115855, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 776.028913] env[65726]: DEBUG oslo_vmware.api [None req-276e84bf-a104-4323-ae2b-430f1d490aad tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] Task: {'id': task-5115852, 'name': ReconfigVM_Task, 'duration_secs': 0.384294} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 776.033444] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-276e84bf-a104-4323-ae2b-430f1d490aad tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] [instance: 3827eed1-b7a1-4922-a4de-f0c038cb8566] Reconfigured VM instance instance-00000025 to attach disk [datastore1] 3827eed1-b7a1-4922-a4de-f0c038cb8566/3827eed1-b7a1-4922-a4de-f0c038cb8566.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 776.035317] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8c4a9d26-0d5e-4f65-8d10-5a136a74aaec {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.048015] env[65726]: DEBUG oslo_vmware.api [None req-d51d38e8-71be-4889-a206-d52403737076 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Task: {'id': task-5115853, 'name': PowerOnVM_Task} progress is 88%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 776.048839] env[65726]: DEBUG oslo_vmware.api [None req-276e84bf-a104-4323-ae2b-430f1d490aad tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] Waiting for the task: (returnval){ [ 776.048839] env[65726]: value = "task-5115856" [ 776.048839] env[65726]: _type = "Task" [ 776.048839] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 776.066587] env[65726]: DEBUG oslo_vmware.api [None req-276e84bf-a104-4323-ae2b-430f1d490aad tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] Task: {'id': task-5115856, 'name': Rename_Task} progress is 6%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 776.110446] env[65726]: DEBUG oslo_concurrency.lockutils [None req-55d74d77-063b-418b-b818-59d05c490f34 tempest-ImagesNegativeTestJSON-1015404931 tempest-ImagesNegativeTestJSON-1015404931-project-member] Acquiring lock "c4177e20-b1bd-4b54-a275-c93582359a07" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 776.110446] env[65726]: DEBUG oslo_concurrency.lockutils [None req-55d74d77-063b-418b-b818-59d05c490f34 tempest-ImagesNegativeTestJSON-1015404931 tempest-ImagesNegativeTestJSON-1015404931-project-member] Lock "c4177e20-b1bd-4b54-a275-c93582359a07" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 776.255259] env[65726]: DEBUG nova.compute.utils [None req-462b38ac-85e6-4728-81f8-4b729f9b1c26 tempest-ServersAaction247Test-712120018 tempest-ServersAaction247Test-712120018-project-member] Using /dev/sd instead of None {{(pid=65726) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 776.268546] env[65726]: DEBUG nova.compute.manager [None req-462b38ac-85e6-4728-81f8-4b729f9b1c26 tempest-ServersAaction247Test-712120018 tempest-ServersAaction247Test-712120018-project-member] [instance: ad48cadd-9bb6-4191-a68a-5c8fdaeebd44] Not allocating networking since 'none' was specified. {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2015}} [ 776.386776] env[65726]: DEBUG nova.compute.manager [None req-0b112b08-1593-41b5-b38e-86e5b4d22081 tempest-ServersV294TestFqdnHostnames-819811677 tempest-ServersV294TestFqdnHostnames-819811677-project-member] [instance: c8be64a3-1bfa-41a3-b9be-68a7e6e052e0] Starting instance... {{(pid=65726) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 776.455537] env[65726]: DEBUG oslo_vmware.api [None req-7b44120c-3963-45cf-8eb3-04f0d2e08d90 tempest-ServerDiagnosticsNegativeTest-1635132145 tempest-ServerDiagnosticsNegativeTest-1635132145-project-member] Task: {'id': task-5115855, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.162482} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 776.455843] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-7b44120c-3963-45cf-8eb3-04f0d2e08d90 tempest-ServerDiagnosticsNegativeTest-1635132145 tempest-ServerDiagnosticsNegativeTest-1635132145-project-member] Deleted the datastore file {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 776.456073] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-7b44120c-3963-45cf-8eb3-04f0d2e08d90 tempest-ServerDiagnosticsNegativeTest-1635132145 tempest-ServerDiagnosticsNegativeTest-1635132145-project-member] [instance: 0d6f14c5-5770-476f-a186-4cf1b2d4fbe8] Deleted contents of the VM from datastore datastore2 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 776.456275] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-7b44120c-3963-45cf-8eb3-04f0d2e08d90 tempest-ServerDiagnosticsNegativeTest-1635132145 tempest-ServerDiagnosticsNegativeTest-1635132145-project-member] [instance: 0d6f14c5-5770-476f-a186-4cf1b2d4fbe8] Instance destroyed {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 776.456462] env[65726]: INFO nova.compute.manager [None req-7b44120c-3963-45cf-8eb3-04f0d2e08d90 tempest-ServerDiagnosticsNegativeTest-1635132145 tempest-ServerDiagnosticsNegativeTest-1635132145-project-member] [instance: 0d6f14c5-5770-476f-a186-4cf1b2d4fbe8] Took 1.17 seconds to destroy the instance on the hypervisor. [ 776.456763] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-7b44120c-3963-45cf-8eb3-04f0d2e08d90 tempest-ServerDiagnosticsNegativeTest-1635132145 tempest-ServerDiagnosticsNegativeTest-1635132145-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 776.456981] env[65726]: DEBUG nova.compute.manager [-] [instance: 0d6f14c5-5770-476f-a186-4cf1b2d4fbe8] Deallocating network for instance {{(pid=65726) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 776.457132] env[65726]: DEBUG nova.network.neutron [-] [instance: 0d6f14c5-5770-476f-a186-4cf1b2d4fbe8] deallocate_for_instance() {{(pid=65726) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 776.458139] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 776.458333] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 776.539592] env[65726]: DEBUG oslo_vmware.api [None req-d51d38e8-71be-4889-a206-d52403737076 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Task: {'id': task-5115853, 'name': PowerOnVM_Task, 'duration_secs': 0.934694} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 776.541368] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-d51d38e8-71be-4889-a206-d52403737076 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: 9fc0c9ee-9dc7-4059-b8d2-a5db965e00c2] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 776.541478] env[65726]: INFO nova.compute.manager [None req-d51d38e8-71be-4889-a206-d52403737076 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: 9fc0c9ee-9dc7-4059-b8d2-a5db965e00c2] Took 9.30 seconds to spawn the instance on the hypervisor. [ 776.541794] env[65726]: DEBUG nova.compute.manager [None req-d51d38e8-71be-4889-a206-d52403737076 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: 9fc0c9ee-9dc7-4059-b8d2-a5db965e00c2] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 776.542539] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f546fdc-d890-4cf3-b043-6a1c94e4d26e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.564889] env[65726]: DEBUG oslo_vmware.api [None req-276e84bf-a104-4323-ae2b-430f1d490aad tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] Task: {'id': task-5115856, 'name': Rename_Task, 'duration_secs': 0.166562} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 776.565466] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-276e84bf-a104-4323-ae2b-430f1d490aad tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] [instance: 3827eed1-b7a1-4922-a4de-f0c038cb8566] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 776.566337] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a22a5c00-57a1-4dc6-a8f4-468ba3944021 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.574223] env[65726]: DEBUG oslo_vmware.api [None req-276e84bf-a104-4323-ae2b-430f1d490aad tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] Waiting for the task: (returnval){ [ 776.574223] env[65726]: value = "task-5115857" [ 776.574223] env[65726]: _type = "Task" [ 776.574223] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 776.590490] env[65726]: DEBUG oslo_vmware.api [None req-276e84bf-a104-4323-ae2b-430f1d490aad tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] Task: {'id': task-5115857, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 776.696308] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 776.696590] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 776.762882] env[65726]: DEBUG nova.compute.manager [None req-462b38ac-85e6-4728-81f8-4b729f9b1c26 tempest-ServersAaction247Test-712120018 tempest-ServersAaction247Test-712120018-project-member] [instance: ad48cadd-9bb6-4191-a68a-5c8fdaeebd44] Start building block device mappings for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 776.771465] env[65726]: DEBUG oslo_concurrency.lockutils [None req-11e4b8a8-60d9-4939-a6a9-6a357b669b13 tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.024s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 776.773470] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5489bce2-093c-4c74-b145-6df2a52e95cf tempest-ServersTestBootFromVolume-762458528 tempest-ServersTestBootFromVolume-762458528-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 34.105s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 776.773470] env[65726]: DEBUG nova.objects.instance [None req-5489bce2-093c-4c74-b145-6df2a52e95cf tempest-ServersTestBootFromVolume-762458528 tempest-ServersTestBootFromVolume-762458528-project-member] Lazy-loading 'resources' on Instance uuid 5a8d4357-4b44-4a19-b1da-42d188c38adc {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 776.918595] env[65726]: DEBUG oslo_concurrency.lockutils [None req-0b112b08-1593-41b5-b38e-86e5b4d22081 tempest-ServersV294TestFqdnHostnames-819811677 tempest-ServersV294TestFqdnHostnames-819811677-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 777.075574] env[65726]: INFO nova.compute.manager [None req-d51d38e8-71be-4889-a206-d52403737076 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: 9fc0c9ee-9dc7-4059-b8d2-a5db965e00c2] Took 48.82 seconds to build instance. [ 777.089104] env[65726]: DEBUG oslo_vmware.api [None req-276e84bf-a104-4323-ae2b-430f1d490aad tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] Task: {'id': task-5115857, 'name': PowerOnVM_Task, 'duration_secs': 0.50665} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 777.090254] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-276e84bf-a104-4323-ae2b-430f1d490aad tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] [instance: 3827eed1-b7a1-4922-a4de-f0c038cb8566] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 777.090476] env[65726]: DEBUG nova.compute.manager [None req-276e84bf-a104-4323-ae2b-430f1d490aad tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] [instance: 3827eed1-b7a1-4922-a4de-f0c038cb8566] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 777.091381] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6124c9c-9024-4c99-a5d1-7fabd062a5a4 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.178290] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-b86f24be-1076-4ee1-9368-176430971dfc tempest-VolumesAssistedSnapshotsTest-1477776432 tempest-VolumesAssistedSnapshotsTest-1477776432-project-admin] [instance: aa2716a9-2cf8-4e51-800d-1c8f98b2cf05] Volume attach. Driver type: vmdk {{(pid=65726) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 777.178756] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-b86f24be-1076-4ee1-9368-176430971dfc tempest-VolumesAssistedSnapshotsTest-1477776432 tempest-VolumesAssistedSnapshotsTest-1477776432-project-admin] [instance: aa2716a9-2cf8-4e51-800d-1c8f98b2cf05] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-995129', 'volume_id': 'b7243e8e-0979-43c1-87dc-2570fca06458', 'name': 'volume-b7243e8e-0979-43c1-87dc-2570fca06458', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'aa2716a9-2cf8-4e51-800d-1c8f98b2cf05', 'attached_at': '', 'detached_at': '', 'volume_id': 'b7243e8e-0979-43c1-87dc-2570fca06458', 'serial': 'b7243e8e-0979-43c1-87dc-2570fca06458'} {{(pid=65726) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 777.180424] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a31a82c-90e5-43c7-a6ae-4ee27016199e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.206859] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a12906b5-56ac-41eb-9965-5ce97c494d22 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.247739] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-b86f24be-1076-4ee1-9368-176430971dfc tempest-VolumesAssistedSnapshotsTest-1477776432 tempest-VolumesAssistedSnapshotsTest-1477776432-project-admin] [instance: aa2716a9-2cf8-4e51-800d-1c8f98b2cf05] Reconfiguring VM instance instance-0000000b to attach disk [datastore1] volume-b7243e8e-0979-43c1-87dc-2570fca06458/volume-b7243e8e-0979-43c1-87dc-2570fca06458.vmdk or device None with type thin {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 777.248053] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c2cc2d92-c179-4a30-a6cb-2bf9ef02f7d0 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.281961] env[65726]: DEBUG oslo_vmware.api [None req-b86f24be-1076-4ee1-9368-176430971dfc tempest-VolumesAssistedSnapshotsTest-1477776432 tempest-VolumesAssistedSnapshotsTest-1477776432-project-admin] Waiting for the task: (returnval){ [ 777.281961] env[65726]: value = "task-5115858" [ 777.281961] env[65726]: _type = "Task" [ 777.281961] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 777.294183] env[65726]: DEBUG oslo_vmware.api [None req-b86f24be-1076-4ee1-9368-176430971dfc tempest-VolumesAssistedSnapshotsTest-1477776432 tempest-VolumesAssistedSnapshotsTest-1477776432-project-admin] Task: {'id': task-5115858, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 777.577671] env[65726]: DEBUG nova.network.neutron [-] [instance: 0d6f14c5-5770-476f-a186-4cf1b2d4fbe8] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 777.585073] env[65726]: DEBUG oslo_concurrency.lockutils [None req-d51d38e8-71be-4889-a206-d52403737076 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Lock "9fc0c9ee-9dc7-4059-b8d2-a5db965e00c2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 77.096s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 777.621062] env[65726]: DEBUG oslo_concurrency.lockutils [None req-276e84bf-a104-4323-ae2b-430f1d490aad tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 777.778842] env[65726]: DEBUG nova.compute.manager [None req-462b38ac-85e6-4728-81f8-4b729f9b1c26 tempest-ServersAaction247Test-712120018 tempest-ServersAaction247Test-712120018-project-member] [instance: ad48cadd-9bb6-4191-a68a-5c8fdaeebd44] Start spawning the instance on the hypervisor. {{(pid=65726) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 777.794972] env[65726]: DEBUG oslo_vmware.api [None req-b86f24be-1076-4ee1-9368-176430971dfc tempest-VolumesAssistedSnapshotsTest-1477776432 tempest-VolumesAssistedSnapshotsTest-1477776432-project-admin] Task: {'id': task-5115858, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 777.824095] env[65726]: DEBUG nova.virt.hardware [None req-462b38ac-85e6-4728-81f8-4b729f9b1c26 tempest-ServersAaction247Test-712120018 tempest-ServersAaction247Test-712120018-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T19:28:34Z,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1539f06fa3534e90acbc3a60c4b8bd3f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T19:28:35Z,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 777.824483] env[65726]: DEBUG nova.virt.hardware [None req-462b38ac-85e6-4728-81f8-4b729f9b1c26 tempest-ServersAaction247Test-712120018 tempest-ServersAaction247Test-712120018-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 777.824632] env[65726]: DEBUG nova.virt.hardware [None req-462b38ac-85e6-4728-81f8-4b729f9b1c26 tempest-ServersAaction247Test-712120018 tempest-ServersAaction247Test-712120018-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 777.824701] env[65726]: DEBUG nova.virt.hardware [None req-462b38ac-85e6-4728-81f8-4b729f9b1c26 tempest-ServersAaction247Test-712120018 tempest-ServersAaction247Test-712120018-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 777.824853] env[65726]: DEBUG nova.virt.hardware [None req-462b38ac-85e6-4728-81f8-4b729f9b1c26 tempest-ServersAaction247Test-712120018 tempest-ServersAaction247Test-712120018-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 777.824967] env[65726]: DEBUG nova.virt.hardware [None req-462b38ac-85e6-4728-81f8-4b729f9b1c26 tempest-ServersAaction247Test-712120018 tempest-ServersAaction247Test-712120018-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 777.825197] env[65726]: DEBUG nova.virt.hardware [None req-462b38ac-85e6-4728-81f8-4b729f9b1c26 tempest-ServersAaction247Test-712120018 tempest-ServersAaction247Test-712120018-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 777.825350] env[65726]: DEBUG nova.virt.hardware [None req-462b38ac-85e6-4728-81f8-4b729f9b1c26 tempest-ServersAaction247Test-712120018 tempest-ServersAaction247Test-712120018-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 777.825513] env[65726]: DEBUG nova.virt.hardware [None req-462b38ac-85e6-4728-81f8-4b729f9b1c26 tempest-ServersAaction247Test-712120018 tempest-ServersAaction247Test-712120018-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 777.826027] env[65726]: DEBUG nova.virt.hardware [None req-462b38ac-85e6-4728-81f8-4b729f9b1c26 tempest-ServersAaction247Test-712120018 tempest-ServersAaction247Test-712120018-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 777.826027] env[65726]: DEBUG nova.virt.hardware [None req-462b38ac-85e6-4728-81f8-4b729f9b1c26 tempest-ServersAaction247Test-712120018 tempest-ServersAaction247Test-712120018-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 777.826826] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0289d2c3-a3a5-4f7e-8c3c-e86996a13da6 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.841529] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afe4c4c6-e681-4b28-85d1-f6eb8aa99f75 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.859319] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-462b38ac-85e6-4728-81f8-4b729f9b1c26 tempest-ServersAaction247Test-712120018 tempest-ServersAaction247Test-712120018-project-member] [instance: ad48cadd-9bb6-4191-a68a-5c8fdaeebd44] Instance VIF info [] {{(pid=65726) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 777.870650] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-462b38ac-85e6-4728-81f8-4b729f9b1c26 tempest-ServersAaction247Test-712120018 tempest-ServersAaction247Test-712120018-project-member] Creating folder: Project (6367df51100e4445bab0dd00b260e03d). Parent ref: group-v995008. {{(pid=65726) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 777.873976] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a498161c-2002-41c5-ab56-c8e4cd8263c9 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.891171] env[65726]: INFO nova.virt.vmwareapi.vm_util [None req-462b38ac-85e6-4728-81f8-4b729f9b1c26 tempest-ServersAaction247Test-712120018 tempest-ServersAaction247Test-712120018-project-member] Created folder: Project (6367df51100e4445bab0dd00b260e03d) in parent group-v995008. [ 777.891613] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-462b38ac-85e6-4728-81f8-4b729f9b1c26 tempest-ServersAaction247Test-712120018 tempest-ServersAaction247Test-712120018-project-member] Creating folder: Instances. Parent ref: group-v995130. {{(pid=65726) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 777.891909] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-dd23250a-8ee7-4f04-abba-c0b090842098 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.910451] env[65726]: INFO nova.virt.vmwareapi.vm_util [None req-462b38ac-85e6-4728-81f8-4b729f9b1c26 tempest-ServersAaction247Test-712120018 tempest-ServersAaction247Test-712120018-project-member] Created folder: Instances in parent group-v995130. [ 777.910451] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-462b38ac-85e6-4728-81f8-4b729f9b1c26 tempest-ServersAaction247Test-712120018 tempest-ServersAaction247Test-712120018-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 777.910451] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ad48cadd-9bb6-4191-a68a-5c8fdaeebd44] Creating VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 777.910763] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-894de990-f8cc-454d-b2f6-c96cc1c44cbb {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.933319] env[65726]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 777.933319] env[65726]: value = "task-5115861" [ 777.933319] env[65726]: _type = "Task" [ 777.933319] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 777.944790] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115861, 'name': CreateVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 777.980045] env[65726]: DEBUG nova.compute.manager [req-1bb38af4-06de-4eb8-9912-e3f6610a7a03 req-7f40ebf5-a5da-45db-8bb8-fc3a1c7a55fa service nova] [instance: d7905c4f-1ba3-43b3-b4b3-aebacd2f9cc3] Received event network-changed-ee24bb04-7df3-4d06-9751-289b4bdeb753 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 777.980416] env[65726]: DEBUG nova.compute.manager [req-1bb38af4-06de-4eb8-9912-e3f6610a7a03 req-7f40ebf5-a5da-45db-8bb8-fc3a1c7a55fa service nova] [instance: d7905c4f-1ba3-43b3-b4b3-aebacd2f9cc3] Refreshing instance network info cache due to event network-changed-ee24bb04-7df3-4d06-9751-289b4bdeb753. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 777.980623] env[65726]: DEBUG oslo_concurrency.lockutils [req-1bb38af4-06de-4eb8-9912-e3f6610a7a03 req-7f40ebf5-a5da-45db-8bb8-fc3a1c7a55fa service nova] Acquiring lock "refresh_cache-d7905c4f-1ba3-43b3-b4b3-aebacd2f9cc3" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 777.980772] env[65726]: DEBUG oslo_concurrency.lockutils [req-1bb38af4-06de-4eb8-9912-e3f6610a7a03 req-7f40ebf5-a5da-45db-8bb8-fc3a1c7a55fa service nova] Acquired lock "refresh_cache-d7905c4f-1ba3-43b3-b4b3-aebacd2f9cc3" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 777.980876] env[65726]: DEBUG nova.network.neutron [req-1bb38af4-06de-4eb8-9912-e3f6610a7a03 req-7f40ebf5-a5da-45db-8bb8-fc3a1c7a55fa service nova] [instance: d7905c4f-1ba3-43b3-b4b3-aebacd2f9cc3] Refreshing network info cache for port ee24bb04-7df3-4d06-9751-289b4bdeb753 {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 778.051073] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a14df72-230f-4265-a396-0c6f902daebb {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.062020] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73487fc5-8b5a-4ed5-9e6d-ee438f1b3f49 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.104382] env[65726]: INFO nova.compute.manager [-] [instance: 0d6f14c5-5770-476f-a186-4cf1b2d4fbe8] Took 1.65 seconds to deallocate network for instance. [ 778.104382] env[65726]: DEBUG nova.compute.manager [None req-645c8c71-cd4d-4583-8270-b88bb6120d10 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 01d86089-6b9b-4588-864e-ef91375a1eea] Starting instance... {{(pid=65726) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 778.107845] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa214a00-7db9-4339-b5ac-21cbb4f0de92 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.121014] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bd48d76-4011-4218-adec-019d564792bf {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.142596] env[65726]: DEBUG nova.compute.provider_tree [None req-5489bce2-093c-4c74-b145-6df2a52e95cf tempest-ServersTestBootFromVolume-762458528 tempest-ServersTestBootFromVolume-762458528-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 778.296883] env[65726]: DEBUG oslo_vmware.api [None req-b86f24be-1076-4ee1-9368-176430971dfc tempest-VolumesAssistedSnapshotsTest-1477776432 tempest-VolumesAssistedSnapshotsTest-1477776432-project-admin] Task: {'id': task-5115858, 'name': ReconfigVM_Task, 'duration_secs': 0.901034} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 778.296883] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-b86f24be-1076-4ee1-9368-176430971dfc tempest-VolumesAssistedSnapshotsTest-1477776432 tempest-VolumesAssistedSnapshotsTest-1477776432-project-admin] [instance: aa2716a9-2cf8-4e51-800d-1c8f98b2cf05] Reconfigured VM instance instance-0000000b to attach disk [datastore1] volume-b7243e8e-0979-43c1-87dc-2570fca06458/volume-b7243e8e-0979-43c1-87dc-2570fca06458.vmdk or device None with type thin {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 778.302856] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-16fa274e-35f7-4d18-a508-99fe39931241 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.326120] env[65726]: DEBUG oslo_vmware.api [None req-b86f24be-1076-4ee1-9368-176430971dfc tempest-VolumesAssistedSnapshotsTest-1477776432 tempest-VolumesAssistedSnapshotsTest-1477776432-project-admin] Waiting for the task: (returnval){ [ 778.326120] env[65726]: value = "task-5115862" [ 778.326120] env[65726]: _type = "Task" [ 778.326120] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 778.340869] env[65726]: DEBUG oslo_vmware.api [None req-b86f24be-1076-4ee1-9368-176430971dfc tempest-VolumesAssistedSnapshotsTest-1477776432 tempest-VolumesAssistedSnapshotsTest-1477776432-project-admin] Task: {'id': task-5115862, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 778.444514] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115861, 'name': CreateVM_Task, 'duration_secs': 0.437938} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 778.444705] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ad48cadd-9bb6-4191-a68a-5c8fdaeebd44] Created VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 778.445164] env[65726]: DEBUG oslo_concurrency.lockutils [None req-462b38ac-85e6-4728-81f8-4b729f9b1c26 tempest-ServersAaction247Test-712120018 tempest-ServersAaction247Test-712120018-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 778.445325] env[65726]: DEBUG oslo_concurrency.lockutils [None req-462b38ac-85e6-4728-81f8-4b729f9b1c26 tempest-ServersAaction247Test-712120018 tempest-ServersAaction247Test-712120018-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 778.445861] env[65726]: DEBUG oslo_concurrency.lockutils [None req-462b38ac-85e6-4728-81f8-4b729f9b1c26 tempest-ServersAaction247Test-712120018 tempest-ServersAaction247Test-712120018-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 778.445940] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-12319976-c6ef-45f7-86b5-a36d76668365 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.452909] env[65726]: DEBUG oslo_vmware.api [None req-462b38ac-85e6-4728-81f8-4b729f9b1c26 tempest-ServersAaction247Test-712120018 tempest-ServersAaction247Test-712120018-project-member] Waiting for the task: (returnval){ [ 778.452909] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]523c6f1d-f996-1134-510e-33d3f4ce737f" [ 778.452909] env[65726]: _type = "Task" [ 778.452909] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 778.461449] env[65726]: DEBUG oslo_vmware.api [None req-462b38ac-85e6-4728-81f8-4b729f9b1c26 tempest-ServersAaction247Test-712120018 tempest-ServersAaction247Test-712120018-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]523c6f1d-f996-1134-510e-33d3f4ce737f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 778.484313] env[65726]: WARNING openstack [req-1bb38af4-06de-4eb8-9912-e3f6610a7a03 req-7f40ebf5-a5da-45db-8bb8-fc3a1c7a55fa service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 778.484868] env[65726]: WARNING openstack [req-1bb38af4-06de-4eb8-9912-e3f6610a7a03 req-7f40ebf5-a5da-45db-8bb8-fc3a1c7a55fa service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 778.622266] env[65726]: DEBUG oslo_concurrency.lockutils [None req-7b44120c-3963-45cf-8eb3-04f0d2e08d90 tempest-ServerDiagnosticsNegativeTest-1635132145 tempest-ServerDiagnosticsNegativeTest-1635132145-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 778.637593] env[65726]: DEBUG oslo_concurrency.lockutils [None req-645c8c71-cd4d-4583-8270-b88bb6120d10 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 778.644699] env[65726]: DEBUG nova.scheduler.client.report [None req-5489bce2-093c-4c74-b145-6df2a52e95cf tempest-ServersTestBootFromVolume-762458528 tempest-ServersTestBootFromVolume-762458528-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 778.818436] env[65726]: WARNING openstack [req-1bb38af4-06de-4eb8-9912-e3f6610a7a03 req-7f40ebf5-a5da-45db-8bb8-fc3a1c7a55fa service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 778.818909] env[65726]: WARNING openstack [req-1bb38af4-06de-4eb8-9912-e3f6610a7a03 req-7f40ebf5-a5da-45db-8bb8-fc3a1c7a55fa service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 778.842805] env[65726]: DEBUG oslo_vmware.api [None req-b86f24be-1076-4ee1-9368-176430971dfc tempest-VolumesAssistedSnapshotsTest-1477776432 tempest-VolumesAssistedSnapshotsTest-1477776432-project-admin] Task: {'id': task-5115862, 'name': ReconfigVM_Task, 'duration_secs': 0.253472} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 778.842805] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-b86f24be-1076-4ee1-9368-176430971dfc tempest-VolumesAssistedSnapshotsTest-1477776432 tempest-VolumesAssistedSnapshotsTest-1477776432-project-admin] [instance: aa2716a9-2cf8-4e51-800d-1c8f98b2cf05] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-995129', 'volume_id': 'b7243e8e-0979-43c1-87dc-2570fca06458', 'name': 'volume-b7243e8e-0979-43c1-87dc-2570fca06458', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'aa2716a9-2cf8-4e51-800d-1c8f98b2cf05', 'attached_at': '', 'detached_at': '', 'volume_id': 'b7243e8e-0979-43c1-87dc-2570fca06458', 'serial': 'b7243e8e-0979-43c1-87dc-2570fca06458'} {{(pid=65726) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 778.969028] env[65726]: DEBUG oslo_vmware.api [None req-462b38ac-85e6-4728-81f8-4b729f9b1c26 tempest-ServersAaction247Test-712120018 tempest-ServersAaction247Test-712120018-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]523c6f1d-f996-1134-510e-33d3f4ce737f, 'name': SearchDatastore_Task, 'duration_secs': 0.015043} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 778.969600] env[65726]: DEBUG oslo_concurrency.lockutils [None req-462b38ac-85e6-4728-81f8-4b729f9b1c26 tempest-ServersAaction247Test-712120018 tempest-ServersAaction247Test-712120018-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 778.969600] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-462b38ac-85e6-4728-81f8-4b729f9b1c26 tempest-ServersAaction247Test-712120018 tempest-ServersAaction247Test-712120018-project-member] [instance: ad48cadd-9bb6-4191-a68a-5c8fdaeebd44] Processing image b52362a3-ee8a-4cbf-b06f-513b0cc8f95c {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 778.969864] env[65726]: DEBUG oslo_concurrency.lockutils [None req-462b38ac-85e6-4728-81f8-4b729f9b1c26 tempest-ServersAaction247Test-712120018 tempest-ServersAaction247Test-712120018-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 778.970063] env[65726]: DEBUG oslo_concurrency.lockutils [None req-462b38ac-85e6-4728-81f8-4b729f9b1c26 tempest-ServersAaction247Test-712120018 tempest-ServersAaction247Test-712120018-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 778.970276] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-462b38ac-85e6-4728-81f8-4b729f9b1c26 tempest-ServersAaction247Test-712120018 tempest-ServersAaction247Test-712120018-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 778.970538] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1613fa72-62c2-4c1d-a448-80589fcdec40 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.982994] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-462b38ac-85e6-4728-81f8-4b729f9b1c26 tempest-ServersAaction247Test-712120018 tempest-ServersAaction247Test-712120018-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 778.983236] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-462b38ac-85e6-4728-81f8-4b729f9b1c26 tempest-ServersAaction247Test-712120018 tempest-ServersAaction247Test-712120018-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65726) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 778.984073] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d4409538-b6ee-4c36-97b8-2a2557931474 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.990460] env[65726]: DEBUG oslo_vmware.api [None req-462b38ac-85e6-4728-81f8-4b729f9b1c26 tempest-ServersAaction247Test-712120018 tempest-ServersAaction247Test-712120018-project-member] Waiting for the task: (returnval){ [ 778.990460] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5294b61b-bd2b-7b44-05bf-35f9404dddea" [ 778.990460] env[65726]: _type = "Task" [ 778.990460] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 779.000413] env[65726]: DEBUG oslo_vmware.api [None req-462b38ac-85e6-4728-81f8-4b729f9b1c26 tempest-ServersAaction247Test-712120018 tempest-ServersAaction247Test-712120018-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5294b61b-bd2b-7b44-05bf-35f9404dddea, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.017143] env[65726]: WARNING openstack [req-1bb38af4-06de-4eb8-9912-e3f6610a7a03 req-7f40ebf5-a5da-45db-8bb8-fc3a1c7a55fa service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 779.017655] env[65726]: WARNING openstack [req-1bb38af4-06de-4eb8-9912-e3f6610a7a03 req-7f40ebf5-a5da-45db-8bb8-fc3a1c7a55fa service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 779.151058] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5489bce2-093c-4c74-b145-6df2a52e95cf tempest-ServersTestBootFromVolume-762458528 tempest-ServersTestBootFromVolume-762458528-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.378s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 779.154740] env[65726]: DEBUG oslo_concurrency.lockutils [None req-fd5c9ab4-2b38-4729-88f4-0f437c7dacb5 tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 31.992s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 779.154740] env[65726]: DEBUG nova.objects.instance [None req-fd5c9ab4-2b38-4729-88f4-0f437c7dacb5 tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] Lazy-loading 'resources' on Instance uuid 83573c2b-9448-456f-8fd2-b19661dd6cc4 {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 779.184674] env[65726]: INFO nova.scheduler.client.report [None req-5489bce2-093c-4c74-b145-6df2a52e95cf tempest-ServersTestBootFromVolume-762458528 tempest-ServersTestBootFromVolume-762458528-project-member] Deleted allocations for instance 5a8d4357-4b44-4a19-b1da-42d188c38adc [ 779.248558] env[65726]: DEBUG nova.network.neutron [req-1bb38af4-06de-4eb8-9912-e3f6610a7a03 req-7f40ebf5-a5da-45db-8bb8-fc3a1c7a55fa service nova] [instance: d7905c4f-1ba3-43b3-b4b3-aebacd2f9cc3] Updated VIF entry in instance network info cache for port ee24bb04-7df3-4d06-9751-289b4bdeb753. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 779.248904] env[65726]: DEBUG nova.network.neutron [req-1bb38af4-06de-4eb8-9912-e3f6610a7a03 req-7f40ebf5-a5da-45db-8bb8-fc3a1c7a55fa service nova] [instance: d7905c4f-1ba3-43b3-b4b3-aebacd2f9cc3] Updating instance_info_cache with network_info: [{"id": "ee24bb04-7df3-4d06-9751-289b4bdeb753", "address": "fa:16:3e:82:91:65", "network": {"id": "27801f9a-9912-41cf-be65-bc9903f4952f", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-781436951-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.155", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aac5fe6c99f740ada3747088b2b0ad89", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dd7d0d95-6848-4e69-ac21-75f8db82a3b5", "external-id": "nsx-vlan-transportzone-272", "segmentation_id": 272, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapee24bb04-7d", "ovs_interfaceid": "ee24bb04-7df3-4d06-9751-289b4bdeb753", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 779.250532] env[65726]: DEBUG oslo_concurrency.lockutils [None req-84a40308-06fb-49aa-8717-9d81681505a7 tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] Acquiring lock "3827eed1-b7a1-4922-a4de-f0c038cb8566" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 779.250639] env[65726]: DEBUG oslo_concurrency.lockutils [None req-84a40308-06fb-49aa-8717-9d81681505a7 tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] Lock "3827eed1-b7a1-4922-a4de-f0c038cb8566" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 779.251382] env[65726]: DEBUG oslo_concurrency.lockutils [None req-84a40308-06fb-49aa-8717-9d81681505a7 tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] Acquiring lock "3827eed1-b7a1-4922-a4de-f0c038cb8566-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 779.251382] env[65726]: DEBUG oslo_concurrency.lockutils [None req-84a40308-06fb-49aa-8717-9d81681505a7 tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] Lock "3827eed1-b7a1-4922-a4de-f0c038cb8566-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 779.251382] env[65726]: DEBUG oslo_concurrency.lockutils [None req-84a40308-06fb-49aa-8717-9d81681505a7 tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] Lock "3827eed1-b7a1-4922-a4de-f0c038cb8566-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 779.253686] env[65726]: INFO nova.compute.manager [None req-84a40308-06fb-49aa-8717-9d81681505a7 tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] [instance: 3827eed1-b7a1-4922-a4de-f0c038cb8566] Terminating instance [ 779.448452] env[65726]: INFO nova.compute.manager [None req-00483101-e5ba-4006-8481-84d7c7b5b2b8 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: 9fc0c9ee-9dc7-4059-b8d2-a5db965e00c2] Rescuing [ 779.448834] env[65726]: DEBUG oslo_concurrency.lockutils [None req-00483101-e5ba-4006-8481-84d7c7b5b2b8 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Acquiring lock "refresh_cache-9fc0c9ee-9dc7-4059-b8d2-a5db965e00c2" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 779.448987] env[65726]: DEBUG oslo_concurrency.lockutils [None req-00483101-e5ba-4006-8481-84d7c7b5b2b8 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Acquired lock "refresh_cache-9fc0c9ee-9dc7-4059-b8d2-a5db965e00c2" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 779.449802] env[65726]: DEBUG nova.network.neutron [None req-00483101-e5ba-4006-8481-84d7c7b5b2b8 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: 9fc0c9ee-9dc7-4059-b8d2-a5db965e00c2] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 779.503176] env[65726]: DEBUG oslo_vmware.api [None req-462b38ac-85e6-4728-81f8-4b729f9b1c26 tempest-ServersAaction247Test-712120018 tempest-ServersAaction247Test-712120018-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5294b61b-bd2b-7b44-05bf-35f9404dddea, 'name': SearchDatastore_Task, 'duration_secs': 0.014514} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 779.515290] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bacc51e5-bcc9-4571-85e8-d9a12d46cfc6 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.536132] env[65726]: DEBUG oslo_vmware.api [None req-462b38ac-85e6-4728-81f8-4b729f9b1c26 tempest-ServersAaction247Test-712120018 tempest-ServersAaction247Test-712120018-project-member] Waiting for the task: (returnval){ [ 779.536132] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]529f2d61-5b39-754e-57f3-7d3acb02d463" [ 779.536132] env[65726]: _type = "Task" [ 779.536132] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 779.551651] env[65726]: DEBUG oslo_vmware.api [None req-462b38ac-85e6-4728-81f8-4b729f9b1c26 tempest-ServersAaction247Test-712120018 tempest-ServersAaction247Test-712120018-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]529f2d61-5b39-754e-57f3-7d3acb02d463, 'name': SearchDatastore_Task, 'duration_secs': 0.013544} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 779.552116] env[65726]: DEBUG oslo_concurrency.lockutils [None req-462b38ac-85e6-4728-81f8-4b729f9b1c26 tempest-ServersAaction247Test-712120018 tempest-ServersAaction247Test-712120018-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 779.552388] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-462b38ac-85e6-4728-81f8-4b729f9b1c26 tempest-ServersAaction247Test-712120018 tempest-ServersAaction247Test-712120018-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] ad48cadd-9bb6-4191-a68a-5c8fdaeebd44/ad48cadd-9bb6-4191-a68a-5c8fdaeebd44.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 779.552668] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9e884b28-89e7-46cc-a0be-d5fd0b918c2f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.560355] env[65726]: DEBUG oslo_vmware.api [None req-462b38ac-85e6-4728-81f8-4b729f9b1c26 tempest-ServersAaction247Test-712120018 tempest-ServersAaction247Test-712120018-project-member] Waiting for the task: (returnval){ [ 779.560355] env[65726]: value = "task-5115863" [ 779.560355] env[65726]: _type = "Task" [ 779.560355] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 779.569127] env[65726]: DEBUG oslo_vmware.api [None req-462b38ac-85e6-4728-81f8-4b729f9b1c26 tempest-ServersAaction247Test-712120018 tempest-ServersAaction247Test-712120018-project-member] Task: {'id': task-5115863, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.694890] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5489bce2-093c-4c74-b145-6df2a52e95cf tempest-ServersTestBootFromVolume-762458528 tempest-ServersTestBootFromVolume-762458528-project-member] Lock "5a8d4357-4b44-4a19-b1da-42d188c38adc" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 42.546s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 779.754338] env[65726]: DEBUG oslo_concurrency.lockutils [req-1bb38af4-06de-4eb8-9912-e3f6610a7a03 req-7f40ebf5-a5da-45db-8bb8-fc3a1c7a55fa service nova] Releasing lock "refresh_cache-d7905c4f-1ba3-43b3-b4b3-aebacd2f9cc3" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 779.758274] env[65726]: DEBUG oslo_concurrency.lockutils [None req-84a40308-06fb-49aa-8717-9d81681505a7 tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] Acquiring lock "refresh_cache-3827eed1-b7a1-4922-a4de-f0c038cb8566" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 779.758456] env[65726]: DEBUG oslo_concurrency.lockutils [None req-84a40308-06fb-49aa-8717-9d81681505a7 tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] Acquired lock "refresh_cache-3827eed1-b7a1-4922-a4de-f0c038cb8566" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 779.758653] env[65726]: DEBUG nova.network.neutron [None req-84a40308-06fb-49aa-8717-9d81681505a7 tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] [instance: 3827eed1-b7a1-4922-a4de-f0c038cb8566] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 779.893678] env[65726]: DEBUG nova.objects.instance [None req-b86f24be-1076-4ee1-9368-176430971dfc tempest-VolumesAssistedSnapshotsTest-1477776432 tempest-VolumesAssistedSnapshotsTest-1477776432-project-admin] Lazy-loading 'flavor' on Instance uuid aa2716a9-2cf8-4e51-800d-1c8f98b2cf05 {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 779.954685] env[65726]: WARNING openstack [None req-00483101-e5ba-4006-8481-84d7c7b5b2b8 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 779.955283] env[65726]: WARNING openstack [None req-00483101-e5ba-4006-8481-84d7c7b5b2b8 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 780.075790] env[65726]: DEBUG oslo_vmware.api [None req-462b38ac-85e6-4728-81f8-4b729f9b1c26 tempest-ServersAaction247Test-712120018 tempest-ServersAaction247Test-712120018-project-member] Task: {'id': task-5115863, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 780.151683] env[65726]: WARNING openstack [None req-00483101-e5ba-4006-8481-84d7c7b5b2b8 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 780.152139] env[65726]: WARNING openstack [None req-00483101-e5ba-4006-8481-84d7c7b5b2b8 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 780.261893] env[65726]: WARNING openstack [None req-84a40308-06fb-49aa-8717-9d81681505a7 tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 780.262284] env[65726]: WARNING openstack [None req-84a40308-06fb-49aa-8717-9d81681505a7 tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 780.267466] env[65726]: DEBUG nova.network.neutron [None req-84a40308-06fb-49aa-8717-9d81681505a7 tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] [instance: 3827eed1-b7a1-4922-a4de-f0c038cb8566] Instance cache missing network info. {{(pid=65726) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 780.306417] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce63db18-faf2-4083-a575-773980dce3c4 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.315175] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01990f9e-82be-45bf-a78e-75ff864bf0bc {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.353589] env[65726]: WARNING openstack [None req-00483101-e5ba-4006-8481-84d7c7b5b2b8 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 780.354308] env[65726]: WARNING openstack [None req-00483101-e5ba-4006-8481-84d7c7b5b2b8 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 780.365061] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dabb7904-e223-4922-ba03-eeb9182486a6 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.376019] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46e02379-f592-437f-a554-b755c99d6321 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.389474] env[65726]: DEBUG nova.compute.provider_tree [None req-fd5c9ab4-2b38-4729-88f4-0f437c7dacb5 tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 780.398185] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b86f24be-1076-4ee1-9368-176430971dfc tempest-VolumesAssistedSnapshotsTest-1477776432 tempest-VolumesAssistedSnapshotsTest-1477776432-project-admin] Lock "aa2716a9-2cf8-4e51-800d-1c8f98b2cf05" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.871s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 780.446580] env[65726]: DEBUG nova.network.neutron [None req-84a40308-06fb-49aa-8717-9d81681505a7 tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] [instance: 3827eed1-b7a1-4922-a4de-f0c038cb8566] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 780.530818] env[65726]: DEBUG nova.network.neutron [None req-00483101-e5ba-4006-8481-84d7c7b5b2b8 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: 9fc0c9ee-9dc7-4059-b8d2-a5db965e00c2] Updating instance_info_cache with network_info: [{"id": "e4315658-0732-463a-b063-b1c4cf7cdafd", "address": "fa:16:3e:c1:e7:e1", "network": {"id": "9d5ad167-99fb-4208-ab30-7053f35774e4", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-2082516762-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a8362295a62f4b51bae719a7ef5a4656", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dabbac20-1723-40ad-9da0-e53b28073651", "external-id": "nsx-vlan-transportzone-790", "segmentation_id": 790, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape4315658-07", "ovs_interfaceid": "e4315658-0732-463a-b063-b1c4cf7cdafd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 780.571496] env[65726]: DEBUG oslo_vmware.api [None req-462b38ac-85e6-4728-81f8-4b729f9b1c26 tempest-ServersAaction247Test-712120018 tempest-ServersAaction247Test-712120018-project-member] Task: {'id': task-5115863, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.592638} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 780.571773] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-462b38ac-85e6-4728-81f8-4b729f9b1c26 tempest-ServersAaction247Test-712120018 tempest-ServersAaction247Test-712120018-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] ad48cadd-9bb6-4191-a68a-5c8fdaeebd44/ad48cadd-9bb6-4191-a68a-5c8fdaeebd44.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 780.571988] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-462b38ac-85e6-4728-81f8-4b729f9b1c26 tempest-ServersAaction247Test-712120018 tempest-ServersAaction247Test-712120018-project-member] [instance: ad48cadd-9bb6-4191-a68a-5c8fdaeebd44] Extending root virtual disk to 1048576 {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 780.572422] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-142d9355-a51d-4f4a-8110-18098689729c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.581953] env[65726]: DEBUG oslo_vmware.api [None req-462b38ac-85e6-4728-81f8-4b729f9b1c26 tempest-ServersAaction247Test-712120018 tempest-ServersAaction247Test-712120018-project-member] Waiting for the task: (returnval){ [ 780.581953] env[65726]: value = "task-5115864" [ 780.581953] env[65726]: _type = "Task" [ 780.581953] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 780.592719] env[65726]: DEBUG oslo_vmware.api [None req-462b38ac-85e6-4728-81f8-4b729f9b1c26 tempest-ServersAaction247Test-712120018 tempest-ServersAaction247Test-712120018-project-member] Task: {'id': task-5115864, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 780.818047] env[65726]: DEBUG nova.compute.manager [req-ad486d87-e229-4fbb-976f-bb1d53038c5e req-b0bb59c0-c403-4e63-85df-ed37726b8079 service nova] [instance: d7905c4f-1ba3-43b3-b4b3-aebacd2f9cc3] Received event network-changed-ee24bb04-7df3-4d06-9751-289b4bdeb753 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 780.818677] env[65726]: DEBUG nova.compute.manager [req-ad486d87-e229-4fbb-976f-bb1d53038c5e req-b0bb59c0-c403-4e63-85df-ed37726b8079 service nova] [instance: d7905c4f-1ba3-43b3-b4b3-aebacd2f9cc3] Refreshing instance network info cache due to event network-changed-ee24bb04-7df3-4d06-9751-289b4bdeb753. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 780.819074] env[65726]: DEBUG oslo_concurrency.lockutils [req-ad486d87-e229-4fbb-976f-bb1d53038c5e req-b0bb59c0-c403-4e63-85df-ed37726b8079 service nova] Acquiring lock "refresh_cache-d7905c4f-1ba3-43b3-b4b3-aebacd2f9cc3" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 780.819224] env[65726]: DEBUG oslo_concurrency.lockutils [req-ad486d87-e229-4fbb-976f-bb1d53038c5e req-b0bb59c0-c403-4e63-85df-ed37726b8079 service nova] Acquired lock "refresh_cache-d7905c4f-1ba3-43b3-b4b3-aebacd2f9cc3" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 780.819407] env[65726]: DEBUG nova.network.neutron [req-ad486d87-e229-4fbb-976f-bb1d53038c5e req-b0bb59c0-c403-4e63-85df-ed37726b8079 service nova] [instance: d7905c4f-1ba3-43b3-b4b3-aebacd2f9cc3] Refreshing network info cache for port ee24bb04-7df3-4d06-9751-289b4bdeb753 {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 780.893795] env[65726]: DEBUG nova.scheduler.client.report [None req-fd5c9ab4-2b38-4729-88f4-0f437c7dacb5 tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 780.949842] env[65726]: DEBUG oslo_concurrency.lockutils [None req-84a40308-06fb-49aa-8717-9d81681505a7 tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] Releasing lock "refresh_cache-3827eed1-b7a1-4922-a4de-f0c038cb8566" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 780.950324] env[65726]: DEBUG nova.compute.manager [None req-84a40308-06fb-49aa-8717-9d81681505a7 tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] [instance: 3827eed1-b7a1-4922-a4de-f0c038cb8566] Start destroying the instance on the hypervisor. {{(pid=65726) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 780.954028] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-84a40308-06fb-49aa-8717-9d81681505a7 tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] [instance: 3827eed1-b7a1-4922-a4de-f0c038cb8566] Destroying instance {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 780.954028] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3aa7ced3-b090-48e4-975a-e7fdbd401d6b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.961091] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-84a40308-06fb-49aa-8717-9d81681505a7 tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] [instance: 3827eed1-b7a1-4922-a4de-f0c038cb8566] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 780.961392] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4f5cc96c-d108-4eff-b018-f9b8339072c4 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.968849] env[65726]: DEBUG oslo_vmware.api [None req-84a40308-06fb-49aa-8717-9d81681505a7 tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] Waiting for the task: (returnval){ [ 780.968849] env[65726]: value = "task-5115865" [ 780.968849] env[65726]: _type = "Task" [ 780.968849] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 780.979737] env[65726]: DEBUG oslo_vmware.api [None req-84a40308-06fb-49aa-8717-9d81681505a7 tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] Task: {'id': task-5115865, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.036109] env[65726]: DEBUG oslo_concurrency.lockutils [None req-00483101-e5ba-4006-8481-84d7c7b5b2b8 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Releasing lock "refresh_cache-9fc0c9ee-9dc7-4059-b8d2-a5db965e00c2" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 781.096508] env[65726]: DEBUG oslo_vmware.api [None req-462b38ac-85e6-4728-81f8-4b729f9b1c26 tempest-ServersAaction247Test-712120018 tempest-ServersAaction247Test-712120018-project-member] Task: {'id': task-5115864, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.382984} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 781.097642] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-462b38ac-85e6-4728-81f8-4b729f9b1c26 tempest-ServersAaction247Test-712120018 tempest-ServersAaction247Test-712120018-project-member] [instance: ad48cadd-9bb6-4191-a68a-5c8fdaeebd44] Extended root virtual disk {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 781.098648] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9aade4c0-2a54-4307-b195-0491889abb43 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.128531] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-462b38ac-85e6-4728-81f8-4b729f9b1c26 tempest-ServersAaction247Test-712120018 tempest-ServersAaction247Test-712120018-project-member] [instance: ad48cadd-9bb6-4191-a68a-5c8fdaeebd44] Reconfiguring VM instance instance-0000002d to attach disk [datastore1] ad48cadd-9bb6-4191-a68a-5c8fdaeebd44/ad48cadd-9bb6-4191-a68a-5c8fdaeebd44.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 781.129468] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d1538a87-8627-44d5-b173-b3036763bea9 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.161035] env[65726]: DEBUG oslo_vmware.api [None req-462b38ac-85e6-4728-81f8-4b729f9b1c26 tempest-ServersAaction247Test-712120018 tempest-ServersAaction247Test-712120018-project-member] Waiting for the task: (returnval){ [ 781.161035] env[65726]: value = "task-5115866" [ 781.161035] env[65726]: _type = "Task" [ 781.161035] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 781.171611] env[65726]: DEBUG oslo_vmware.api [None req-462b38ac-85e6-4728-81f8-4b729f9b1c26 tempest-ServersAaction247Test-712120018 tempest-ServersAaction247Test-712120018-project-member] Task: {'id': task-5115866, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.323207] env[65726]: WARNING openstack [req-ad486d87-e229-4fbb-976f-bb1d53038c5e req-b0bb59c0-c403-4e63-85df-ed37726b8079 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 781.323593] env[65726]: WARNING openstack [req-ad486d87-e229-4fbb-976f-bb1d53038c5e req-b0bb59c0-c403-4e63-85df-ed37726b8079 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 781.336028] env[65726]: DEBUG nova.compute.manager [req-9964352a-a95a-4496-babe-9cff6d620bee req-83856340-93e8-4bee-8855-f234cbb66b7c service nova] [instance: 0d6f14c5-5770-476f-a186-4cf1b2d4fbe8] Received event network-vif-deleted-c8363d25-c607-4d04-ab53-412848d7cf4f {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 781.398638] env[65726]: DEBUG oslo_concurrency.lockutils [None req-fd5c9ab4-2b38-4729-88f4-0f437c7dacb5 tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.244s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 781.401607] env[65726]: DEBUG oslo_concurrency.lockutils [None req-300cacab-e765-49c9-a8a1-8e47a20dc6a2 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 28.938s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 781.403598] env[65726]: INFO nova.compute.claims [None req-300cacab-e765-49c9-a8a1-8e47a20dc6a2 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] [instance: 76249623-6f83-46a3-b8c5-c001111aa698] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 781.431469] env[65726]: INFO nova.scheduler.client.report [None req-fd5c9ab4-2b38-4729-88f4-0f437c7dacb5 tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] Deleted allocations for instance 83573c2b-9448-456f-8fd2-b19661dd6cc4 [ 781.459599] env[65726]: WARNING openstack [req-ad486d87-e229-4fbb-976f-bb1d53038c5e req-b0bb59c0-c403-4e63-85df-ed37726b8079 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 781.459599] env[65726]: WARNING openstack [req-ad486d87-e229-4fbb-976f-bb1d53038c5e req-b0bb59c0-c403-4e63-85df-ed37726b8079 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 781.483831] env[65726]: DEBUG oslo_vmware.api [None req-84a40308-06fb-49aa-8717-9d81681505a7 tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] Task: {'id': task-5115865, 'name': PowerOffVM_Task, 'duration_secs': 0.142265} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 781.483993] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-84a40308-06fb-49aa-8717-9d81681505a7 tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] [instance: 3827eed1-b7a1-4922-a4de-f0c038cb8566] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 781.484271] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-84a40308-06fb-49aa-8717-9d81681505a7 tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] [instance: 3827eed1-b7a1-4922-a4de-f0c038cb8566] Unregistering the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 781.484540] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9a4a0305-a36d-468d-bd65-fde87e2c178a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.520771] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-84a40308-06fb-49aa-8717-9d81681505a7 tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] [instance: 3827eed1-b7a1-4922-a4de-f0c038cb8566] Unregistered the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 781.521193] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-84a40308-06fb-49aa-8717-9d81681505a7 tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] [instance: 3827eed1-b7a1-4922-a4de-f0c038cb8566] Deleting contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 781.521388] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-84a40308-06fb-49aa-8717-9d81681505a7 tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] Deleting the datastore file [datastore1] 3827eed1-b7a1-4922-a4de-f0c038cb8566 {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 781.521711] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-08322a59-5ad9-46f8-b135-8ea72773a0d5 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.531734] env[65726]: DEBUG oslo_vmware.api [None req-84a40308-06fb-49aa-8717-9d81681505a7 tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] Waiting for the task: (returnval){ [ 781.531734] env[65726]: value = "task-5115868" [ 781.531734] env[65726]: _type = "Task" [ 781.531734] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 781.543017] env[65726]: DEBUG oslo_vmware.api [None req-84a40308-06fb-49aa-8717-9d81681505a7 tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] Task: {'id': task-5115868, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.567588] env[65726]: WARNING openstack [req-ad486d87-e229-4fbb-976f-bb1d53038c5e req-b0bb59c0-c403-4e63-85df-ed37726b8079 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 781.568014] env[65726]: WARNING openstack [req-ad486d87-e229-4fbb-976f-bb1d53038c5e req-b0bb59c0-c403-4e63-85df-ed37726b8079 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 781.677409] env[65726]: DEBUG oslo_vmware.api [None req-462b38ac-85e6-4728-81f8-4b729f9b1c26 tempest-ServersAaction247Test-712120018 tempest-ServersAaction247Test-712120018-project-member] Task: {'id': task-5115866, 'name': ReconfigVM_Task, 'duration_secs': 0.467877} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 781.677754] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-462b38ac-85e6-4728-81f8-4b729f9b1c26 tempest-ServersAaction247Test-712120018 tempest-ServersAaction247Test-712120018-project-member] [instance: ad48cadd-9bb6-4191-a68a-5c8fdaeebd44] Reconfigured VM instance instance-0000002d to attach disk [datastore1] ad48cadd-9bb6-4191-a68a-5c8fdaeebd44/ad48cadd-9bb6-4191-a68a-5c8fdaeebd44.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 781.678478] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-34fea4bc-f0d5-4baa-93a2-af4b2dc6d5a0 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.690415] env[65726]: DEBUG oslo_vmware.api [None req-462b38ac-85e6-4728-81f8-4b729f9b1c26 tempest-ServersAaction247Test-712120018 tempest-ServersAaction247Test-712120018-project-member] Waiting for the task: (returnval){ [ 781.690415] env[65726]: value = "task-5115869" [ 781.690415] env[65726]: _type = "Task" [ 781.690415] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 781.703249] env[65726]: DEBUG oslo_vmware.api [None req-462b38ac-85e6-4728-81f8-4b729f9b1c26 tempest-ServersAaction247Test-712120018 tempest-ServersAaction247Test-712120018-project-member] Task: {'id': task-5115869, 'name': Rename_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.735492] env[65726]: DEBUG nova.network.neutron [req-ad486d87-e229-4fbb-976f-bb1d53038c5e req-b0bb59c0-c403-4e63-85df-ed37726b8079 service nova] [instance: d7905c4f-1ba3-43b3-b4b3-aebacd2f9cc3] Updated VIF entry in instance network info cache for port ee24bb04-7df3-4d06-9751-289b4bdeb753. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 781.736042] env[65726]: DEBUG nova.network.neutron [req-ad486d87-e229-4fbb-976f-bb1d53038c5e req-b0bb59c0-c403-4e63-85df-ed37726b8079 service nova] [instance: d7905c4f-1ba3-43b3-b4b3-aebacd2f9cc3] Updating instance_info_cache with network_info: [{"id": "ee24bb04-7df3-4d06-9751-289b4bdeb753", "address": "fa:16:3e:82:91:65", "network": {"id": "27801f9a-9912-41cf-be65-bc9903f4952f", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-781436951-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aac5fe6c99f740ada3747088b2b0ad89", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dd7d0d95-6848-4e69-ac21-75f8db82a3b5", "external-id": "nsx-vlan-transportzone-272", "segmentation_id": 272, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapee24bb04-7d", "ovs_interfaceid": "ee24bb04-7df3-4d06-9751-289b4bdeb753", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 781.944037] env[65726]: DEBUG oslo_concurrency.lockutils [None req-fd5c9ab4-2b38-4729-88f4-0f437c7dacb5 tempest-ServersListShow296Test-1066372181 tempest-ServersListShow296Test-1066372181-project-member] Lock "83573c2b-9448-456f-8fd2-b19661dd6cc4" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 39.092s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 782.047540] env[65726]: DEBUG oslo_vmware.api [None req-84a40308-06fb-49aa-8717-9d81681505a7 tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] Task: {'id': task-5115868, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.115925} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 782.047972] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-84a40308-06fb-49aa-8717-9d81681505a7 tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] Deleted the datastore file {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 782.048307] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-84a40308-06fb-49aa-8717-9d81681505a7 tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] [instance: 3827eed1-b7a1-4922-a4de-f0c038cb8566] Deleted contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 782.048378] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-84a40308-06fb-49aa-8717-9d81681505a7 tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] [instance: 3827eed1-b7a1-4922-a4de-f0c038cb8566] Instance destroyed {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 782.048573] env[65726]: INFO nova.compute.manager [None req-84a40308-06fb-49aa-8717-9d81681505a7 tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] [instance: 3827eed1-b7a1-4922-a4de-f0c038cb8566] Took 1.10 seconds to destroy the instance on the hypervisor. [ 782.048856] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-84a40308-06fb-49aa-8717-9d81681505a7 tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 782.049129] env[65726]: DEBUG nova.compute.manager [-] [instance: 3827eed1-b7a1-4922-a4de-f0c038cb8566] Deallocating network for instance {{(pid=65726) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 782.049240] env[65726]: DEBUG nova.network.neutron [-] [instance: 3827eed1-b7a1-4922-a4de-f0c038cb8566] deallocate_for_instance() {{(pid=65726) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 782.049869] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 782.050238] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 782.078396] env[65726]: DEBUG nova.network.neutron [-] [instance: 3827eed1-b7a1-4922-a4de-f0c038cb8566] Instance cache missing network info. {{(pid=65726) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 782.078948] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 782.079294] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 782.211433] env[65726]: DEBUG oslo_vmware.api [None req-462b38ac-85e6-4728-81f8-4b729f9b1c26 tempest-ServersAaction247Test-712120018 tempest-ServersAaction247Test-712120018-project-member] Task: {'id': task-5115869, 'name': Rename_Task, 'duration_secs': 0.237802} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 782.211780] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-462b38ac-85e6-4728-81f8-4b729f9b1c26 tempest-ServersAaction247Test-712120018 tempest-ServersAaction247Test-712120018-project-member] [instance: ad48cadd-9bb6-4191-a68a-5c8fdaeebd44] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 782.212060] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0207b1fa-5e72-48b7-bed3-895a0300e4fd {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.223390] env[65726]: DEBUG oslo_vmware.api [None req-462b38ac-85e6-4728-81f8-4b729f9b1c26 tempest-ServersAaction247Test-712120018 tempest-ServersAaction247Test-712120018-project-member] Waiting for the task: (returnval){ [ 782.223390] env[65726]: value = "task-5115870" [ 782.223390] env[65726]: _type = "Task" [ 782.223390] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 782.233049] env[65726]: DEBUG oslo_vmware.api [None req-462b38ac-85e6-4728-81f8-4b729f9b1c26 tempest-ServersAaction247Test-712120018 tempest-ServersAaction247Test-712120018-project-member] Task: {'id': task-5115870, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.241355] env[65726]: DEBUG oslo_concurrency.lockutils [req-ad486d87-e229-4fbb-976f-bb1d53038c5e req-b0bb59c0-c403-4e63-85df-ed37726b8079 service nova] Releasing lock "refresh_cache-d7905c4f-1ba3-43b3-b4b3-aebacd2f9cc3" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 782.587163] env[65726]: DEBUG nova.network.neutron [-] [instance: 3827eed1-b7a1-4922-a4de-f0c038cb8566] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 782.593666] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-00483101-e5ba-4006-8481-84d7c7b5b2b8 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: 9fc0c9ee-9dc7-4059-b8d2-a5db965e00c2] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 782.597178] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a53b8de7-39a2-4948-ae45-5cf9ab97bf4e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.606330] env[65726]: DEBUG oslo_vmware.api [None req-00483101-e5ba-4006-8481-84d7c7b5b2b8 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Waiting for the task: (returnval){ [ 782.606330] env[65726]: value = "task-5115871" [ 782.606330] env[65726]: _type = "Task" [ 782.606330] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 782.621651] env[65726]: DEBUG oslo_vmware.api [None req-00483101-e5ba-4006-8481-84d7c7b5b2b8 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Task: {'id': task-5115871, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.736858] env[65726]: DEBUG oslo_vmware.api [None req-462b38ac-85e6-4728-81f8-4b729f9b1c26 tempest-ServersAaction247Test-712120018 tempest-ServersAaction247Test-712120018-project-member] Task: {'id': task-5115870, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 783.029732] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a7c8af37-4b06-48cb-8e91-7cd44f1c2ae3 tempest-VolumesAssistedSnapshotsTest-1477776432 tempest-VolumesAssistedSnapshotsTest-1477776432-project-admin] Acquiring lock "aa2716a9-2cf8-4e51-800d-1c8f98b2cf05" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 783.031099] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a7c8af37-4b06-48cb-8e91-7cd44f1c2ae3 tempest-VolumesAssistedSnapshotsTest-1477776432 tempest-VolumesAssistedSnapshotsTest-1477776432-project-admin] Lock "aa2716a9-2cf8-4e51-800d-1c8f98b2cf05" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 783.090612] env[65726]: INFO nova.compute.manager [-] [instance: 3827eed1-b7a1-4922-a4de-f0c038cb8566] Took 1.04 seconds to deallocate network for instance. [ 783.120614] env[65726]: DEBUG oslo_vmware.api [None req-00483101-e5ba-4006-8481-84d7c7b5b2b8 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Task: {'id': task-5115871, 'name': PowerOffVM_Task, 'duration_secs': 0.257931} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 783.120959] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-00483101-e5ba-4006-8481-84d7c7b5b2b8 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: 9fc0c9ee-9dc7-4059-b8d2-a5db965e00c2] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 783.124794] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2dc89ece-2a8d-4c26-aa80-2517a4de56b8 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.132274] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8e450d2-0ad0-4027-92ee-fa422562a4d5 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.155642] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cec96866-2050-4d53-b9ed-83d5e0b8641d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.159415] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df4a7ee9-9923-4ab7-94cc-ad2e15e5bc91 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.202810] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8d8079b-5ed7-4ef4-a6f3-67c6401556f2 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.211911] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c393fd7a-bec5-438b-a7fb-c2b4ebd87230 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.228637] env[65726]: DEBUG nova.compute.provider_tree [None req-300cacab-e765-49c9-a8a1-8e47a20dc6a2 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Updating inventory in ProviderTree for provider 07c4692f-bdb4-4058-9173-ff9664830295 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 783.243591] env[65726]: DEBUG oslo_vmware.api [None req-462b38ac-85e6-4728-81f8-4b729f9b1c26 tempest-ServersAaction247Test-712120018 tempest-ServersAaction247Test-712120018-project-member] Task: {'id': task-5115870, 'name': PowerOnVM_Task, 'duration_secs': 0.795597} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 783.245985] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-00483101-e5ba-4006-8481-84d7c7b5b2b8 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: 9fc0c9ee-9dc7-4059-b8d2-a5db965e00c2] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 783.246983] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-462b38ac-85e6-4728-81f8-4b729f9b1c26 tempest-ServersAaction247Test-712120018 tempest-ServersAaction247Test-712120018-project-member] [instance: ad48cadd-9bb6-4191-a68a-5c8fdaeebd44] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 783.248174] env[65726]: INFO nova.compute.manager [None req-462b38ac-85e6-4728-81f8-4b729f9b1c26 tempest-ServersAaction247Test-712120018 tempest-ServersAaction247Test-712120018-project-member] [instance: ad48cadd-9bb6-4191-a68a-5c8fdaeebd44] Took 5.47 seconds to spawn the instance on the hypervisor. [ 783.248435] env[65726]: DEBUG nova.compute.manager [None req-462b38ac-85e6-4728-81f8-4b729f9b1c26 tempest-ServersAaction247Test-712120018 tempest-ServersAaction247Test-712120018-project-member] [instance: ad48cadd-9bb6-4191-a68a-5c8fdaeebd44] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 783.249419] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-37a2e08a-bfcd-41f9-9f79-1f6b6956fe98 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.251445] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0771a02d-60fd-412c-a20d-b956c8bdfd7a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.266960] env[65726]: DEBUG oslo_vmware.api [None req-00483101-e5ba-4006-8481-84d7c7b5b2b8 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Waiting for the task: (returnval){ [ 783.266960] env[65726]: value = "task-5115872" [ 783.266960] env[65726]: _type = "Task" [ 783.266960] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 783.276815] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-00483101-e5ba-4006-8481-84d7c7b5b2b8 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: 9fc0c9ee-9dc7-4059-b8d2-a5db965e00c2] VM already powered off {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 783.277080] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-00483101-e5ba-4006-8481-84d7c7b5b2b8 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: 9fc0c9ee-9dc7-4059-b8d2-a5db965e00c2] Processing image b52362a3-ee8a-4cbf-b06f-513b0cc8f95c {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 783.277330] env[65726]: DEBUG oslo_concurrency.lockutils [None req-00483101-e5ba-4006-8481-84d7c7b5b2b8 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 783.277489] env[65726]: DEBUG oslo_concurrency.lockutils [None req-00483101-e5ba-4006-8481-84d7c7b5b2b8 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 783.277639] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-00483101-e5ba-4006-8481-84d7c7b5b2b8 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 783.277888] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2508356d-a430-427a-bdc8-42b9e40e12ec {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.288886] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-00483101-e5ba-4006-8481-84d7c7b5b2b8 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 783.289104] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-00483101-e5ba-4006-8481-84d7c7b5b2b8 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65726) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 783.289889] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-65206ee7-6833-4b7e-af6f-e1b55ccbc1bf {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.297035] env[65726]: DEBUG oslo_vmware.api [None req-00483101-e5ba-4006-8481-84d7c7b5b2b8 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Waiting for the task: (returnval){ [ 783.297035] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]524121fd-3495-4471-7c8a-58485fb1c1f8" [ 783.297035] env[65726]: _type = "Task" [ 783.297035] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 783.305589] env[65726]: DEBUG oslo_vmware.api [None req-00483101-e5ba-4006-8481-84d7c7b5b2b8 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]524121fd-3495-4471-7c8a-58485fb1c1f8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 783.539036] env[65726]: INFO nova.compute.manager [None req-a7c8af37-4b06-48cb-8e91-7cd44f1c2ae3 tempest-VolumesAssistedSnapshotsTest-1477776432 tempest-VolumesAssistedSnapshotsTest-1477776432-project-admin] [instance: aa2716a9-2cf8-4e51-800d-1c8f98b2cf05] Detaching volume b7243e8e-0979-43c1-87dc-2570fca06458 [ 783.583063] env[65726]: INFO nova.virt.block_device [None req-a7c8af37-4b06-48cb-8e91-7cd44f1c2ae3 tempest-VolumesAssistedSnapshotsTest-1477776432 tempest-VolumesAssistedSnapshotsTest-1477776432-project-admin] [instance: aa2716a9-2cf8-4e51-800d-1c8f98b2cf05] Attempting to driver detach volume b7243e8e-0979-43c1-87dc-2570fca06458 from mountpoint /dev/sdb [ 783.583063] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-a7c8af37-4b06-48cb-8e91-7cd44f1c2ae3 tempest-VolumesAssistedSnapshotsTest-1477776432 tempest-VolumesAssistedSnapshotsTest-1477776432-project-admin] [instance: aa2716a9-2cf8-4e51-800d-1c8f98b2cf05] Volume detach. Driver type: vmdk {{(pid=65726) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 783.583063] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-a7c8af37-4b06-48cb-8e91-7cd44f1c2ae3 tempest-VolumesAssistedSnapshotsTest-1477776432 tempest-VolumesAssistedSnapshotsTest-1477776432-project-admin] [instance: aa2716a9-2cf8-4e51-800d-1c8f98b2cf05] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-995129', 'volume_id': 'b7243e8e-0979-43c1-87dc-2570fca06458', 'name': 'volume-b7243e8e-0979-43c1-87dc-2570fca06458', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'aa2716a9-2cf8-4e51-800d-1c8f98b2cf05', 'attached_at': '', 'detached_at': '', 'volume_id': 'b7243e8e-0979-43c1-87dc-2570fca06458', 'serial': 'b7243e8e-0979-43c1-87dc-2570fca06458'} {{(pid=65726) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 783.583063] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95bc9964-990e-4ea0-a20d-7146115b1411 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.606798] env[65726]: DEBUG oslo_concurrency.lockutils [None req-84a40308-06fb-49aa-8717-9d81681505a7 tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 783.607717] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f40de947-ee15-484b-8f7f-bf9638f3cb90 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.615634] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8cc8f8ec-7b55-4426-bae7-08b9a0ecb94e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.639122] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db0daf5f-1c42-49e7-ac36-231a8659d26b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.660452] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-a7c8af37-4b06-48cb-8e91-7cd44f1c2ae3 tempest-VolumesAssistedSnapshotsTest-1477776432 tempest-VolumesAssistedSnapshotsTest-1477776432-project-admin] The volume has not been displaced from its original location: [datastore1] volume-b7243e8e-0979-43c1-87dc-2570fca06458/volume-b7243e8e-0979-43c1-87dc-2570fca06458.vmdk. No consolidation needed. {{(pid=65726) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 783.665682] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-a7c8af37-4b06-48cb-8e91-7cd44f1c2ae3 tempest-VolumesAssistedSnapshotsTest-1477776432 tempest-VolumesAssistedSnapshotsTest-1477776432-project-admin] [instance: aa2716a9-2cf8-4e51-800d-1c8f98b2cf05] Reconfiguring VM instance instance-0000000b to detach disk 2001 {{(pid=65726) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 783.666133] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0f2ba14b-a087-4de1-83ff-8c23c89e9f59 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.685739] env[65726]: DEBUG oslo_vmware.api [None req-a7c8af37-4b06-48cb-8e91-7cd44f1c2ae3 tempest-VolumesAssistedSnapshotsTest-1477776432 tempest-VolumesAssistedSnapshotsTest-1477776432-project-admin] Waiting for the task: (returnval){ [ 783.685739] env[65726]: value = "task-5115873" [ 783.685739] env[65726]: _type = "Task" [ 783.685739] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 783.696404] env[65726]: DEBUG oslo_vmware.api [None req-a7c8af37-4b06-48cb-8e91-7cd44f1c2ae3 tempest-VolumesAssistedSnapshotsTest-1477776432 tempest-VolumesAssistedSnapshotsTest-1477776432-project-admin] Task: {'id': task-5115873, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 783.760702] env[65726]: ERROR nova.scheduler.client.report [None req-300cacab-e765-49c9-a8a1-8e47a20dc6a2 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] [req-55f926ff-e59c-41cc-9c1a-79c180c6ac5d] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 07c4692f-bdb4-4058-9173-ff9664830295. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-55f926ff-e59c-41cc-9c1a-79c180c6ac5d"}]} [ 783.777377] env[65726]: INFO nova.compute.manager [None req-462b38ac-85e6-4728-81f8-4b729f9b1c26 tempest-ServersAaction247Test-712120018 tempest-ServersAaction247Test-712120018-project-member] [instance: ad48cadd-9bb6-4191-a68a-5c8fdaeebd44] Took 47.28 seconds to build instance. [ 783.780551] env[65726]: DEBUG nova.scheduler.client.report [None req-300cacab-e765-49c9-a8a1-8e47a20dc6a2 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Refreshing inventories for resource provider 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 783.809267] env[65726]: DEBUG nova.scheduler.client.report [None req-300cacab-e765-49c9-a8a1-8e47a20dc6a2 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Updating ProviderTree inventory for provider 07c4692f-bdb4-4058-9173-ff9664830295 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 783.809589] env[65726]: DEBUG nova.compute.provider_tree [None req-300cacab-e765-49c9-a8a1-8e47a20dc6a2 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Updating inventory in ProviderTree for provider 07c4692f-bdb4-4058-9173-ff9664830295 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 783.816270] env[65726]: DEBUG oslo_vmware.api [None req-00483101-e5ba-4006-8481-84d7c7b5b2b8 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]524121fd-3495-4471-7c8a-58485fb1c1f8, 'name': SearchDatastore_Task, 'duration_secs': 0.010867} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 783.817712] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d6cc8a0a-9f8e-4340-aa90-676bf174996a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.824472] env[65726]: DEBUG oslo_vmware.api [None req-00483101-e5ba-4006-8481-84d7c7b5b2b8 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Waiting for the task: (returnval){ [ 783.824472] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52c551b3-80bd-03c2-fde5-b995eb85e7e1" [ 783.824472] env[65726]: _type = "Task" [ 783.824472] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 783.832281] env[65726]: DEBUG nova.scheduler.client.report [None req-300cacab-e765-49c9-a8a1-8e47a20dc6a2 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Refreshing aggregate associations for resource provider 07c4692f-bdb4-4058-9173-ff9664830295, aggregates: None {{(pid=65726) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 783.839669] env[65726]: DEBUG oslo_vmware.api [None req-00483101-e5ba-4006-8481-84d7c7b5b2b8 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52c551b3-80bd-03c2-fde5-b995eb85e7e1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 783.869547] env[65726]: DEBUG nova.scheduler.client.report [None req-300cacab-e765-49c9-a8a1-8e47a20dc6a2 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Refreshing trait associations for resource provider 07c4692f-bdb4-4058-9173-ff9664830295, traits: COMPUTE_IMAGE_TYPE_ISO,HW_ARCH_X86_64,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=65726) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 784.197785] env[65726]: DEBUG oslo_vmware.api [None req-a7c8af37-4b06-48cb-8e91-7cd44f1c2ae3 tempest-VolumesAssistedSnapshotsTest-1477776432 tempest-VolumesAssistedSnapshotsTest-1477776432-project-admin] Task: {'id': task-5115873, 'name': ReconfigVM_Task, 'duration_secs': 0.469206} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 784.198091] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-a7c8af37-4b06-48cb-8e91-7cd44f1c2ae3 tempest-VolumesAssistedSnapshotsTest-1477776432 tempest-VolumesAssistedSnapshotsTest-1477776432-project-admin] [instance: aa2716a9-2cf8-4e51-800d-1c8f98b2cf05] Reconfigured VM instance instance-0000000b to detach disk 2001 {{(pid=65726) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 784.205928] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f951ef65-c6f5-4a5c-b8a1-ae4802d30d91 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.224041] env[65726]: DEBUG oslo_vmware.api [None req-a7c8af37-4b06-48cb-8e91-7cd44f1c2ae3 tempest-VolumesAssistedSnapshotsTest-1477776432 tempest-VolumesAssistedSnapshotsTest-1477776432-project-admin] Waiting for the task: (returnval){ [ 784.224041] env[65726]: value = "task-5115874" [ 784.224041] env[65726]: _type = "Task" [ 784.224041] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 784.235594] env[65726]: DEBUG oslo_vmware.api [None req-a7c8af37-4b06-48cb-8e91-7cd44f1c2ae3 tempest-VolumesAssistedSnapshotsTest-1477776432 tempest-VolumesAssistedSnapshotsTest-1477776432-project-admin] Task: {'id': task-5115874, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.282346] env[65726]: DEBUG oslo_concurrency.lockutils [None req-462b38ac-85e6-4728-81f8-4b729f9b1c26 tempest-ServersAaction247Test-712120018 tempest-ServersAaction247Test-712120018-project-member] Lock "ad48cadd-9bb6-4191-a68a-5c8fdaeebd44" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 81.716s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 784.326305] env[65726]: DEBUG oslo_vmware.rw_handles [None req-703975dc-3e07-4b77-b05e-15d0411bc7cd tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52c3d80b-f9e4-46fc-efe3-f28a9a498520/disk-0.vmdk. {{(pid=65726) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 784.327770] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e683d1cd-5cbe-4def-b65b-b685f581abca {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.348863] env[65726]: DEBUG oslo_vmware.rw_handles [None req-703975dc-3e07-4b77-b05e-15d0411bc7cd tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52c3d80b-f9e4-46fc-efe3-f28a9a498520/disk-0.vmdk is in state: ready. {{(pid=65726) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 784.348974] env[65726]: ERROR oslo_vmware.rw_handles [None req-703975dc-3e07-4b77-b05e-15d0411bc7cd tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52c3d80b-f9e4-46fc-efe3-f28a9a498520/disk-0.vmdk due to incomplete transfer. [ 784.349681] env[65726]: DEBUG oslo_vmware.api [None req-00483101-e5ba-4006-8481-84d7c7b5b2b8 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52c551b3-80bd-03c2-fde5-b995eb85e7e1, 'name': SearchDatastore_Task, 'duration_secs': 0.01208} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 784.349681] env[65726]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-f521b482-78c1-4a03-98ff-1bec604556b7 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.351273] env[65726]: DEBUG oslo_concurrency.lockutils [None req-00483101-e5ba-4006-8481-84d7c7b5b2b8 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 784.351547] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-00483101-e5ba-4006-8481-84d7c7b5b2b8 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Copying virtual disk from [datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore2] 9fc0c9ee-9dc7-4059-b8d2-a5db965e00c2/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c-rescue.vmdk. {{(pid=65726) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 784.352577] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-bae54510-6ace-4226-bae6-9c183fec140b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.364627] env[65726]: DEBUG oslo_vmware.api [None req-00483101-e5ba-4006-8481-84d7c7b5b2b8 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Waiting for the task: (returnval){ [ 784.364627] env[65726]: value = "task-5115875" [ 784.364627] env[65726]: _type = "Task" [ 784.364627] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 784.366037] env[65726]: DEBUG oslo_vmware.rw_handles [None req-703975dc-3e07-4b77-b05e-15d0411bc7cd tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52c3d80b-f9e4-46fc-efe3-f28a9a498520/disk-0.vmdk. {{(pid=65726) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 784.366265] env[65726]: DEBUG nova.virt.vmwareapi.images [None req-703975dc-3e07-4b77-b05e-15d0411bc7cd tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 608cc0ea-3f6b-4b4f-83c1-01aa50999d63] Uploaded image 5554ed4e-2d17-4d6f-946f-e0ce42b880fc to the Glance image server {{(pid=65726) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 784.370737] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-703975dc-3e07-4b77-b05e-15d0411bc7cd tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 608cc0ea-3f6b-4b4f-83c1-01aa50999d63] Destroying the VM {{(pid=65726) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 784.372742] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-65981ff3-feb0-473a-b21b-3f5e0ac52a41 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.382000] env[65726]: DEBUG oslo_vmware.api [None req-00483101-e5ba-4006-8481-84d7c7b5b2b8 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Task: {'id': task-5115875, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.385847] env[65726]: DEBUG oslo_vmware.api [None req-703975dc-3e07-4b77-b05e-15d0411bc7cd tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Waiting for the task: (returnval){ [ 784.385847] env[65726]: value = "task-5115876" [ 784.385847] env[65726]: _type = "Task" [ 784.385847] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 784.397374] env[65726]: DEBUG oslo_vmware.api [None req-703975dc-3e07-4b77-b05e-15d0411bc7cd tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5115876, 'name': Destroy_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.668235] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cda66e16-31cb-4460-8ce8-2e430f57d37c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.678431] env[65726]: DEBUG nova.compute.manager [req-594ea47e-867f-4c55-a322-13356ca89398 req-2ed7db71-c87b-4007-a27b-cf0753b569e9 service nova] [instance: 65aaa2eb-3ca5-4c2b-92f9-b524366c10dd] Received event network-changed-862071bf-725c-4193-b60b-72b03b4c0553 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 784.678760] env[65726]: DEBUG nova.compute.manager [req-594ea47e-867f-4c55-a322-13356ca89398 req-2ed7db71-c87b-4007-a27b-cf0753b569e9 service nova] [instance: 65aaa2eb-3ca5-4c2b-92f9-b524366c10dd] Refreshing instance network info cache due to event network-changed-862071bf-725c-4193-b60b-72b03b4c0553. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 784.678854] env[65726]: DEBUG oslo_concurrency.lockutils [req-594ea47e-867f-4c55-a322-13356ca89398 req-2ed7db71-c87b-4007-a27b-cf0753b569e9 service nova] Acquiring lock "refresh_cache-65aaa2eb-3ca5-4c2b-92f9-b524366c10dd" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 784.684374] env[65726]: DEBUG oslo_concurrency.lockutils [req-594ea47e-867f-4c55-a322-13356ca89398 req-2ed7db71-c87b-4007-a27b-cf0753b569e9 service nova] Acquired lock "refresh_cache-65aaa2eb-3ca5-4c2b-92f9-b524366c10dd" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 784.684866] env[65726]: DEBUG nova.network.neutron [req-594ea47e-867f-4c55-a322-13356ca89398 req-2ed7db71-c87b-4007-a27b-cf0753b569e9 service nova] [instance: 65aaa2eb-3ca5-4c2b-92f9-b524366c10dd] Refreshing network info cache for port 862071bf-725c-4193-b60b-72b03b4c0553 {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 784.702524] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eec31855-53c1-4143-ac20-b31e86e9a979 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.748894] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad8f3f83-5544-4c88-8314-a23ceab45e56 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.762638] env[65726]: DEBUG oslo_vmware.api [None req-a7c8af37-4b06-48cb-8e91-7cd44f1c2ae3 tempest-VolumesAssistedSnapshotsTest-1477776432 tempest-VolumesAssistedSnapshotsTest-1477776432-project-admin] Task: {'id': task-5115874, 'name': ReconfigVM_Task, 'duration_secs': 0.242855} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 784.763875] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-a7c8af37-4b06-48cb-8e91-7cd44f1c2ae3 tempest-VolumesAssistedSnapshotsTest-1477776432 tempest-VolumesAssistedSnapshotsTest-1477776432-project-admin] [instance: aa2716a9-2cf8-4e51-800d-1c8f98b2cf05] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-995129', 'volume_id': 'b7243e8e-0979-43c1-87dc-2570fca06458', 'name': 'volume-b7243e8e-0979-43c1-87dc-2570fca06458', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'aa2716a9-2cf8-4e51-800d-1c8f98b2cf05', 'attached_at': '', 'detached_at': '', 'volume_id': 'b7243e8e-0979-43c1-87dc-2570fca06458', 'serial': 'b7243e8e-0979-43c1-87dc-2570fca06458'} {{(pid=65726) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 784.768126] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34aa4e4c-d3b0-495d-9801-815f0630f2c5 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.789396] env[65726]: DEBUG nova.compute.manager [None req-ebfc0e41-4845-41ca-9f74-d2a2946626c0 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] [instance: 1cfb1bba-1e21-47eb-b50f-a86575846a65] Starting instance... {{(pid=65726) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 784.793367] env[65726]: DEBUG nova.compute.provider_tree [None req-300cacab-e765-49c9-a8a1-8e47a20dc6a2 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Updating inventory in ProviderTree for provider 07c4692f-bdb4-4058-9173-ff9664830295 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 784.879141] env[65726]: DEBUG oslo_vmware.api [None req-00483101-e5ba-4006-8481-84d7c7b5b2b8 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Task: {'id': task-5115875, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.905054] env[65726]: DEBUG oslo_vmware.api [None req-703975dc-3e07-4b77-b05e-15d0411bc7cd tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5115876, 'name': Destroy_Task, 'duration_secs': 0.497055} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 784.905054] env[65726]: INFO nova.virt.vmwareapi.vm_util [None req-703975dc-3e07-4b77-b05e-15d0411bc7cd tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 608cc0ea-3f6b-4b4f-83c1-01aa50999d63] Destroyed the VM [ 784.905054] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-703975dc-3e07-4b77-b05e-15d0411bc7cd tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 608cc0ea-3f6b-4b4f-83c1-01aa50999d63] Deleting Snapshot of the VM instance {{(pid=65726) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 784.905054] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-abfbf84a-c52f-44c0-845c-799482f117e6 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.913140] env[65726]: DEBUG oslo_vmware.api [None req-703975dc-3e07-4b77-b05e-15d0411bc7cd tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Waiting for the task: (returnval){ [ 784.913140] env[65726]: value = "task-5115877" [ 784.913140] env[65726]: _type = "Task" [ 784.913140] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 784.928038] env[65726]: DEBUG oslo_vmware.api [None req-703975dc-3e07-4b77-b05e-15d0411bc7cd tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5115877, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.988448] env[65726]: DEBUG oslo_concurrency.lockutils [None req-cb2d094a-412f-492d-9ef1-aba8757b7146 tempest-ServerMetadataNegativeTestJSON-936475568 tempest-ServerMetadataNegativeTestJSON-936475568-project-member] Acquiring lock "d179030a-e5ba-45b1-ad11-4d2f71ed7bd4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 784.988779] env[65726]: DEBUG oslo_concurrency.lockutils [None req-cb2d094a-412f-492d-9ef1-aba8757b7146 tempest-ServerMetadataNegativeTestJSON-936475568 tempest-ServerMetadataNegativeTestJSON-936475568-project-member] Lock "d179030a-e5ba-45b1-ad11-4d2f71ed7bd4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 785.189108] env[65726]: WARNING openstack [req-594ea47e-867f-4c55-a322-13356ca89398 req-2ed7db71-c87b-4007-a27b-cf0753b569e9 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 785.190011] env[65726]: WARNING openstack [req-594ea47e-867f-4c55-a322-13356ca89398 req-2ed7db71-c87b-4007-a27b-cf0753b569e9 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 785.333786] env[65726]: DEBUG oslo_concurrency.lockutils [None req-ebfc0e41-4845-41ca-9f74-d2a2946626c0 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 785.341024] env[65726]: ERROR nova.scheduler.client.report [None req-300cacab-e765-49c9-a8a1-8e47a20dc6a2 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] [req-3c5723d8-5e2e-4ac7-8586-7aebdce162aa] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 07c4692f-bdb4-4058-9173-ff9664830295. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-3c5723d8-5e2e-4ac7-8586-7aebdce162aa"}]} [ 785.374780] env[65726]: DEBUG nova.scheduler.client.report [None req-300cacab-e765-49c9-a8a1-8e47a20dc6a2 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Refreshing inventories for resource provider 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 785.385774] env[65726]: DEBUG nova.objects.instance [None req-a7c8af37-4b06-48cb-8e91-7cd44f1c2ae3 tempest-VolumesAssistedSnapshotsTest-1477776432 tempest-VolumesAssistedSnapshotsTest-1477776432-project-admin] Lazy-loading 'flavor' on Instance uuid aa2716a9-2cf8-4e51-800d-1c8f98b2cf05 {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 785.395299] env[65726]: DEBUG oslo_vmware.api [None req-00483101-e5ba-4006-8481-84d7c7b5b2b8 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Task: {'id': task-5115875, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.53518} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 785.397076] env[65726]: WARNING openstack [req-594ea47e-867f-4c55-a322-13356ca89398 req-2ed7db71-c87b-4007-a27b-cf0753b569e9 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 785.397426] env[65726]: WARNING openstack [req-594ea47e-867f-4c55-a322-13356ca89398 req-2ed7db71-c87b-4007-a27b-cf0753b569e9 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 785.405759] env[65726]: INFO nova.virt.vmwareapi.ds_util [None req-00483101-e5ba-4006-8481-84d7c7b5b2b8 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Copied virtual disk from [datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore2] 9fc0c9ee-9dc7-4059-b8d2-a5db965e00c2/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c-rescue.vmdk. [ 785.407360] env[65726]: DEBUG nova.scheduler.client.report [None req-300cacab-e765-49c9-a8a1-8e47a20dc6a2 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Updating ProviderTree inventory for provider 07c4692f-bdb4-4058-9173-ff9664830295 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 785.407616] env[65726]: DEBUG nova.compute.provider_tree [None req-300cacab-e765-49c9-a8a1-8e47a20dc6a2 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Updating inventory in ProviderTree for provider 07c4692f-bdb4-4058-9173-ff9664830295 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 785.410797] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2c69f21-62b6-4827-8849-dffb4000f23e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.430288] env[65726]: DEBUG oslo_vmware.api [None req-703975dc-3e07-4b77-b05e-15d0411bc7cd tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5115877, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.431670] env[65726]: DEBUG nova.scheduler.client.report [None req-300cacab-e765-49c9-a8a1-8e47a20dc6a2 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Refreshing aggregate associations for resource provider 07c4692f-bdb4-4058-9173-ff9664830295, aggregates: None {{(pid=65726) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 785.460374] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-00483101-e5ba-4006-8481-84d7c7b5b2b8 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: 9fc0c9ee-9dc7-4059-b8d2-a5db965e00c2] Reconfiguring VM instance instance-0000002c to attach disk [datastore2] 9fc0c9ee-9dc7-4059-b8d2-a5db965e00c2/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c-rescue.vmdk or device None with type thin {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 785.465621] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8134a394-bd39-47d0-9055-1504ac723f3d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.491283] env[65726]: DEBUG nova.scheduler.client.report [None req-300cacab-e765-49c9-a8a1-8e47a20dc6a2 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Refreshing trait associations for resource provider 07c4692f-bdb4-4058-9173-ff9664830295, traits: COMPUTE_IMAGE_TYPE_ISO,HW_ARCH_X86_64,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=65726) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 785.503058] env[65726]: DEBUG oslo_vmware.api [None req-00483101-e5ba-4006-8481-84d7c7b5b2b8 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Waiting for the task: (returnval){ [ 785.503058] env[65726]: value = "task-5115878" [ 785.503058] env[65726]: _type = "Task" [ 785.503058] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 785.512966] env[65726]: DEBUG oslo_vmware.api [None req-00483101-e5ba-4006-8481-84d7c7b5b2b8 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Task: {'id': task-5115878, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.550140] env[65726]: WARNING openstack [req-594ea47e-867f-4c55-a322-13356ca89398 req-2ed7db71-c87b-4007-a27b-cf0753b569e9 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 785.550797] env[65726]: WARNING openstack [req-594ea47e-867f-4c55-a322-13356ca89398 req-2ed7db71-c87b-4007-a27b-cf0753b569e9 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 785.612299] env[65726]: DEBUG nova.compute.manager [req-beaaff5c-1a0e-45c7-8bb0-5ef3a129e2a9 req-2f49f460-09ee-4ae4-b892-e87939bd03bd service nova] [instance: 65aaa2eb-3ca5-4c2b-92f9-b524366c10dd] Received event network-changed-862071bf-725c-4193-b60b-72b03b4c0553 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 785.612563] env[65726]: DEBUG nova.compute.manager [req-beaaff5c-1a0e-45c7-8bb0-5ef3a129e2a9 req-2f49f460-09ee-4ae4-b892-e87939bd03bd service nova] [instance: 65aaa2eb-3ca5-4c2b-92f9-b524366c10dd] Refreshing instance network info cache due to event network-changed-862071bf-725c-4193-b60b-72b03b4c0553. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 785.612762] env[65726]: DEBUG oslo_concurrency.lockutils [req-beaaff5c-1a0e-45c7-8bb0-5ef3a129e2a9 req-2f49f460-09ee-4ae4-b892-e87939bd03bd service nova] Acquiring lock "refresh_cache-65aaa2eb-3ca5-4c2b-92f9-b524366c10dd" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 785.690273] env[65726]: DEBUG nova.network.neutron [req-594ea47e-867f-4c55-a322-13356ca89398 req-2ed7db71-c87b-4007-a27b-cf0753b569e9 service nova] [instance: 65aaa2eb-3ca5-4c2b-92f9-b524366c10dd] Updated VIF entry in instance network info cache for port 862071bf-725c-4193-b60b-72b03b4c0553. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 785.690750] env[65726]: DEBUG nova.network.neutron [req-594ea47e-867f-4c55-a322-13356ca89398 req-2ed7db71-c87b-4007-a27b-cf0753b569e9 service nova] [instance: 65aaa2eb-3ca5-4c2b-92f9-b524366c10dd] Updating instance_info_cache with network_info: [{"id": "862071bf-725c-4193-b60b-72b03b4c0553", "address": "fa:16:3e:9e:67:58", "network": {"id": "27801f9a-9912-41cf-be65-bc9903f4952f", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-781436951-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aac5fe6c99f740ada3747088b2b0ad89", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dd7d0d95-6848-4e69-ac21-75f8db82a3b5", "external-id": "nsx-vlan-transportzone-272", "segmentation_id": 272, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap862071bf-72", "ovs_interfaceid": "862071bf-725c-4193-b60b-72b03b4c0553", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 785.933577] env[65726]: DEBUG oslo_vmware.api [None req-703975dc-3e07-4b77-b05e-15d0411bc7cd tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5115877, 'name': RemoveSnapshot_Task, 'duration_secs': 0.967756} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 785.936723] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-703975dc-3e07-4b77-b05e-15d0411bc7cd tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 608cc0ea-3f6b-4b4f-83c1-01aa50999d63] Deleted Snapshot of the VM instance {{(pid=65726) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 785.936835] env[65726]: INFO nova.compute.manager [None req-703975dc-3e07-4b77-b05e-15d0411bc7cd tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 608cc0ea-3f6b-4b4f-83c1-01aa50999d63] Took 17.90 seconds to snapshot the instance on the hypervisor. [ 786.013957] env[65726]: DEBUG oslo_vmware.api [None req-00483101-e5ba-4006-8481-84d7c7b5b2b8 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Task: {'id': task-5115878, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.104869] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42e768b6-d4cc-44e3-8093-cc3cad0b27a6 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.114485] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e27eb6f-80ff-400f-96e5-c9307c99c8d5 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.155434] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-134aae54-9d1a-42e5-aab7-7e0f69e6deb3 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.168794] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00f54446-e788-4b9d-ab41-51754af70be1 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.186159] env[65726]: DEBUG nova.compute.provider_tree [None req-300cacab-e765-49c9-a8a1-8e47a20dc6a2 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Updating inventory in ProviderTree for provider 07c4692f-bdb4-4058-9173-ff9664830295 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 786.197554] env[65726]: DEBUG oslo_concurrency.lockutils [req-594ea47e-867f-4c55-a322-13356ca89398 req-2ed7db71-c87b-4007-a27b-cf0753b569e9 service nova] Releasing lock "refresh_cache-65aaa2eb-3ca5-4c2b-92f9-b524366c10dd" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 786.198870] env[65726]: DEBUG oslo_concurrency.lockutils [req-beaaff5c-1a0e-45c7-8bb0-5ef3a129e2a9 req-2f49f460-09ee-4ae4-b892-e87939bd03bd service nova] Acquired lock "refresh_cache-65aaa2eb-3ca5-4c2b-92f9-b524366c10dd" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 786.199084] env[65726]: DEBUG nova.network.neutron [req-beaaff5c-1a0e-45c7-8bb0-5ef3a129e2a9 req-2f49f460-09ee-4ae4-b892-e87939bd03bd service nova] [instance: 65aaa2eb-3ca5-4c2b-92f9-b524366c10dd] Refreshing network info cache for port 862071bf-725c-4193-b60b-72b03b4c0553 {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 786.213285] env[65726]: DEBUG oslo_concurrency.lockutils [None req-447a3897-f4ca-48fd-a799-fe155c3fcf63 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Acquiring lock "65aaa2eb-3ca5-4c2b-92f9-b524366c10dd" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 786.213548] env[65726]: DEBUG oslo_concurrency.lockutils [None req-447a3897-f4ca-48fd-a799-fe155c3fcf63 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Lock "65aaa2eb-3ca5-4c2b-92f9-b524366c10dd" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 786.213757] env[65726]: DEBUG oslo_concurrency.lockutils [None req-447a3897-f4ca-48fd-a799-fe155c3fcf63 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Acquiring lock "65aaa2eb-3ca5-4c2b-92f9-b524366c10dd-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 786.213936] env[65726]: DEBUG oslo_concurrency.lockutils [None req-447a3897-f4ca-48fd-a799-fe155c3fcf63 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Lock "65aaa2eb-3ca5-4c2b-92f9-b524366c10dd-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 786.214122] env[65726]: DEBUG oslo_concurrency.lockutils [None req-447a3897-f4ca-48fd-a799-fe155c3fcf63 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Lock "65aaa2eb-3ca5-4c2b-92f9-b524366c10dd-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 786.221737] env[65726]: INFO nova.compute.manager [None req-447a3897-f4ca-48fd-a799-fe155c3fcf63 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] [instance: 65aaa2eb-3ca5-4c2b-92f9-b524366c10dd] Terminating instance [ 786.400296] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a7c8af37-4b06-48cb-8e91-7cd44f1c2ae3 tempest-VolumesAssistedSnapshotsTest-1477776432 tempest-VolumesAssistedSnapshotsTest-1477776432-project-admin] Lock "aa2716a9-2cf8-4e51-800d-1c8f98b2cf05" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.366s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 786.516403] env[65726]: DEBUG oslo_vmware.api [None req-00483101-e5ba-4006-8481-84d7c7b5b2b8 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Task: {'id': task-5115878, 'name': ReconfigVM_Task, 'duration_secs': 0.673347} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 786.517900] env[65726]: DEBUG nova.compute.manager [None req-703975dc-3e07-4b77-b05e-15d0411bc7cd tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 608cc0ea-3f6b-4b4f-83c1-01aa50999d63] Found 1 images (rotation: 2) {{(pid=65726) _rotate_backups /opt/stack/nova/nova/compute/manager.py:5074}} [ 786.518782] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-00483101-e5ba-4006-8481-84d7c7b5b2b8 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: 9fc0c9ee-9dc7-4059-b8d2-a5db965e00c2] Reconfigured VM instance instance-0000002c to attach disk [datastore2] 9fc0c9ee-9dc7-4059-b8d2-a5db965e00c2/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c-rescue.vmdk or device None with type thin {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 786.519970] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-043c0b51-6a59-4d6e-8def-2215b9f831cb {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.555130] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-569d9bf2-d41b-47a8-a8ff-463ab7c449ef {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.572459] env[65726]: DEBUG oslo_vmware.api [None req-00483101-e5ba-4006-8481-84d7c7b5b2b8 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Waiting for the task: (returnval){ [ 786.572459] env[65726]: value = "task-5115880" [ 786.572459] env[65726]: _type = "Task" [ 786.572459] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 786.582993] env[65726]: DEBUG oslo_vmware.api [None req-00483101-e5ba-4006-8481-84d7c7b5b2b8 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Task: {'id': task-5115880, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.704083] env[65726]: WARNING openstack [req-beaaff5c-1a0e-45c7-8bb0-5ef3a129e2a9 req-2f49f460-09ee-4ae4-b892-e87939bd03bd service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 786.705234] env[65726]: WARNING openstack [req-beaaff5c-1a0e-45c7-8bb0-5ef3a129e2a9 req-2f49f460-09ee-4ae4-b892-e87939bd03bd service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 786.723081] env[65726]: DEBUG nova.compute.manager [None req-9401bfb8-fa64-40fc-97d6-85a4492f4163 tempest-ServersAaction247Test-712120018 tempest-ServersAaction247Test-712120018-project-member] [instance: ad48cadd-9bb6-4191-a68a-5c8fdaeebd44] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 786.724236] env[65726]: DEBUG nova.scheduler.client.report [None req-300cacab-e765-49c9-a8a1-8e47a20dc6a2 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Updated inventory for provider 07c4692f-bdb4-4058-9173-ff9664830295 with generation 78 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 786.724479] env[65726]: DEBUG nova.compute.provider_tree [None req-300cacab-e765-49c9-a8a1-8e47a20dc6a2 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Updating resource provider 07c4692f-bdb4-4058-9173-ff9664830295 generation from 78 to 79 during operation: update_inventory {{(pid=65726) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 786.724656] env[65726]: DEBUG nova.compute.provider_tree [None req-300cacab-e765-49c9-a8a1-8e47a20dc6a2 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Updating inventory in ProviderTree for provider 07c4692f-bdb4-4058-9173-ff9664830295 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 786.728784] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fdae8977-4fd9-4894-9aa3-de8bd932b615 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.733120] env[65726]: DEBUG nova.compute.manager [None req-447a3897-f4ca-48fd-a799-fe155c3fcf63 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] [instance: 65aaa2eb-3ca5-4c2b-92f9-b524366c10dd] Start destroying the instance on the hypervisor. {{(pid=65726) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 786.733324] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-447a3897-f4ca-48fd-a799-fe155c3fcf63 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] [instance: 65aaa2eb-3ca5-4c2b-92f9-b524366c10dd] Destroying instance {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 786.734463] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04a57a5d-9018-4c31-9f4c-4a9ff5620b98 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.747236] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-447a3897-f4ca-48fd-a799-fe155c3fcf63 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] [instance: 65aaa2eb-3ca5-4c2b-92f9-b524366c10dd] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 786.747872] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9807e122-b5b7-4b24-a28c-8708bbc6a351 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.756486] env[65726]: DEBUG oslo_vmware.api [None req-447a3897-f4ca-48fd-a799-fe155c3fcf63 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Waiting for the task: (returnval){ [ 786.756486] env[65726]: value = "task-5115881" [ 786.756486] env[65726]: _type = "Task" [ 786.756486] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 786.771229] env[65726]: DEBUG oslo_vmware.api [None req-447a3897-f4ca-48fd-a799-fe155c3fcf63 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Task: {'id': task-5115881, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.918157] env[65726]: WARNING openstack [req-beaaff5c-1a0e-45c7-8bb0-5ef3a129e2a9 req-2f49f460-09ee-4ae4-b892-e87939bd03bd service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 786.918157] env[65726]: WARNING openstack [req-beaaff5c-1a0e-45c7-8bb0-5ef3a129e2a9 req-2f49f460-09ee-4ae4-b892-e87939bd03bd service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 786.996343] env[65726]: WARNING openstack [req-beaaff5c-1a0e-45c7-8bb0-5ef3a129e2a9 req-2f49f460-09ee-4ae4-b892-e87939bd03bd service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 786.997143] env[65726]: WARNING openstack [req-beaaff5c-1a0e-45c7-8bb0-5ef3a129e2a9 req-2f49f460-09ee-4ae4-b892-e87939bd03bd service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 787.089231] env[65726]: DEBUG oslo_vmware.api [None req-00483101-e5ba-4006-8481-84d7c7b5b2b8 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Task: {'id': task-5115880, 'name': ReconfigVM_Task, 'duration_secs': 0.30844} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 787.089597] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-00483101-e5ba-4006-8481-84d7c7b5b2b8 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: 9fc0c9ee-9dc7-4059-b8d2-a5db965e00c2] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 787.089954] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6a20c4da-ff17-48dd-970f-696472db5308 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.099770] env[65726]: DEBUG nova.network.neutron [req-beaaff5c-1a0e-45c7-8bb0-5ef3a129e2a9 req-2f49f460-09ee-4ae4-b892-e87939bd03bd service nova] [instance: 65aaa2eb-3ca5-4c2b-92f9-b524366c10dd] Updated VIF entry in instance network info cache for port 862071bf-725c-4193-b60b-72b03b4c0553. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 787.101493] env[65726]: DEBUG nova.network.neutron [req-beaaff5c-1a0e-45c7-8bb0-5ef3a129e2a9 req-2f49f460-09ee-4ae4-b892-e87939bd03bd service nova] [instance: 65aaa2eb-3ca5-4c2b-92f9-b524366c10dd] Updating instance_info_cache with network_info: [{"id": "862071bf-725c-4193-b60b-72b03b4c0553", "address": "fa:16:3e:9e:67:58", "network": {"id": "27801f9a-9912-41cf-be65-bc9903f4952f", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-781436951-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aac5fe6c99f740ada3747088b2b0ad89", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dd7d0d95-6848-4e69-ac21-75f8db82a3b5", "external-id": "nsx-vlan-transportzone-272", "segmentation_id": 272, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap862071bf-72", "ovs_interfaceid": "862071bf-725c-4193-b60b-72b03b4c0553", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 787.103719] env[65726]: DEBUG oslo_vmware.api [None req-00483101-e5ba-4006-8481-84d7c7b5b2b8 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Waiting for the task: (returnval){ [ 787.103719] env[65726]: value = "task-5115882" [ 787.103719] env[65726]: _type = "Task" [ 787.103719] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 787.115599] env[65726]: DEBUG oslo_vmware.api [None req-00483101-e5ba-4006-8481-84d7c7b5b2b8 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Task: {'id': task-5115882, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.235537] env[65726]: DEBUG oslo_concurrency.lockutils [None req-300cacab-e765-49c9-a8a1-8e47a20dc6a2 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 5.834s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 787.236068] env[65726]: DEBUG nova.compute.manager [None req-300cacab-e765-49c9-a8a1-8e47a20dc6a2 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] [instance: 76249623-6f83-46a3-b8c5-c001111aa698] Start building networks asynchronously for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 787.239464] env[65726]: DEBUG oslo_concurrency.lockutils [None req-af1f7990-88d5-411a-be3e-a2d4ed6b5b5b tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 32.267s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 787.241554] env[65726]: INFO nova.compute.claims [None req-af1f7990-88d5-411a-be3e-a2d4ed6b5b5b tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: 621b899a-e9ec-4fc0-a574-4a08cecff6d9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 787.252019] env[65726]: INFO nova.compute.manager [None req-9401bfb8-fa64-40fc-97d6-85a4492f4163 tempest-ServersAaction247Test-712120018 tempest-ServersAaction247Test-712120018-project-member] [instance: ad48cadd-9bb6-4191-a68a-5c8fdaeebd44] instance snapshotting [ 787.252019] env[65726]: DEBUG nova.objects.instance [None req-9401bfb8-fa64-40fc-97d6-85a4492f4163 tempest-ServersAaction247Test-712120018 tempest-ServersAaction247Test-712120018-project-member] Lazy-loading 'flavor' on Instance uuid ad48cadd-9bb6-4191-a68a-5c8fdaeebd44 {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 787.271272] env[65726]: DEBUG oslo_vmware.api [None req-447a3897-f4ca-48fd-a799-fe155c3fcf63 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Task: {'id': task-5115881, 'name': PowerOffVM_Task, 'duration_secs': 0.242806} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 787.272745] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-447a3897-f4ca-48fd-a799-fe155c3fcf63 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] [instance: 65aaa2eb-3ca5-4c2b-92f9-b524366c10dd] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 787.272977] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-447a3897-f4ca-48fd-a799-fe155c3fcf63 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] [instance: 65aaa2eb-3ca5-4c2b-92f9-b524366c10dd] Unregistering the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 787.273639] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-cf5c0d16-0781-45e9-8397-18f35dee3f0f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.357110] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-447a3897-f4ca-48fd-a799-fe155c3fcf63 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] [instance: 65aaa2eb-3ca5-4c2b-92f9-b524366c10dd] Unregistered the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 787.357110] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-447a3897-f4ca-48fd-a799-fe155c3fcf63 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] [instance: 65aaa2eb-3ca5-4c2b-92f9-b524366c10dd] Deleting contents of the VM from datastore datastore2 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 787.357345] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-447a3897-f4ca-48fd-a799-fe155c3fcf63 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Deleting the datastore file [datastore2] 65aaa2eb-3ca5-4c2b-92f9-b524366c10dd {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 787.357449] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9c665811-e0a2-40dc-99d5-63cf2d9134a5 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.367712] env[65726]: DEBUG oslo_vmware.api [None req-447a3897-f4ca-48fd-a799-fe155c3fcf63 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Waiting for the task: (returnval){ [ 787.367712] env[65726]: value = "task-5115884" [ 787.367712] env[65726]: _type = "Task" [ 787.367712] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 787.377582] env[65726]: DEBUG oslo_vmware.api [None req-447a3897-f4ca-48fd-a799-fe155c3fcf63 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Task: {'id': task-5115884, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.606365] env[65726]: DEBUG oslo_concurrency.lockutils [req-beaaff5c-1a0e-45c7-8bb0-5ef3a129e2a9 req-2f49f460-09ee-4ae4-b892-e87939bd03bd service nova] Releasing lock "refresh_cache-65aaa2eb-3ca5-4c2b-92f9-b524366c10dd" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 787.616334] env[65726]: DEBUG oslo_vmware.api [None req-00483101-e5ba-4006-8481-84d7c7b5b2b8 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Task: {'id': task-5115882, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.747031] env[65726]: DEBUG nova.compute.utils [None req-300cacab-e765-49c9-a8a1-8e47a20dc6a2 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Using /dev/sd instead of None {{(pid=65726) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 787.752296] env[65726]: DEBUG nova.compute.manager [None req-300cacab-e765-49c9-a8a1-8e47a20dc6a2 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] [instance: 76249623-6f83-46a3-b8c5-c001111aa698] Allocating IP information in the background. {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 787.752296] env[65726]: DEBUG nova.network.neutron [None req-300cacab-e765-49c9-a8a1-8e47a20dc6a2 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] [instance: 76249623-6f83-46a3-b8c5-c001111aa698] allocate_for_instance() {{(pid=65726) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 787.752296] env[65726]: WARNING neutronclient.v2_0.client [None req-300cacab-e765-49c9-a8a1-8e47a20dc6a2 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 787.752296] env[65726]: WARNING neutronclient.v2_0.client [None req-300cacab-e765-49c9-a8a1-8e47a20dc6a2 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 787.752296] env[65726]: WARNING openstack [None req-300cacab-e765-49c9-a8a1-8e47a20dc6a2 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 787.752879] env[65726]: WARNING openstack [None req-300cacab-e765-49c9-a8a1-8e47a20dc6a2 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 787.765343] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-503587b3-ca3b-4124-8d8e-3215507f42fc {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.785808] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3260fa2a-3a51-48a8-9aea-39966d4e575e tempest-ServersAaction247Test-712120018 tempest-ServersAaction247Test-712120018-project-member] Acquiring lock "ad48cadd-9bb6-4191-a68a-5c8fdaeebd44" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 787.786079] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3260fa2a-3a51-48a8-9aea-39966d4e575e tempest-ServersAaction247Test-712120018 tempest-ServersAaction247Test-712120018-project-member] Lock "ad48cadd-9bb6-4191-a68a-5c8fdaeebd44" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 787.786288] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3260fa2a-3a51-48a8-9aea-39966d4e575e tempest-ServersAaction247Test-712120018 tempest-ServersAaction247Test-712120018-project-member] Acquiring lock "ad48cadd-9bb6-4191-a68a-5c8fdaeebd44-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 787.786464] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3260fa2a-3a51-48a8-9aea-39966d4e575e tempest-ServersAaction247Test-712120018 tempest-ServersAaction247Test-712120018-project-member] Lock "ad48cadd-9bb6-4191-a68a-5c8fdaeebd44-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 787.786626] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3260fa2a-3a51-48a8-9aea-39966d4e575e tempest-ServersAaction247Test-712120018 tempest-ServersAaction247Test-712120018-project-member] Lock "ad48cadd-9bb6-4191-a68a-5c8fdaeebd44-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 787.791601] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3aa67fb-a70f-4b38-992f-306c6de902fc {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.795388] env[65726]: INFO nova.compute.manager [None req-3260fa2a-3a51-48a8-9aea-39966d4e575e tempest-ServersAaction247Test-712120018 tempest-ServersAaction247Test-712120018-project-member] [instance: ad48cadd-9bb6-4191-a68a-5c8fdaeebd44] Terminating instance [ 787.836256] env[65726]: DEBUG nova.policy [None req-300cacab-e765-49c9-a8a1-8e47a20dc6a2 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '58e5e1469d2241d3ad3178f085a7bee6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'fae9ef0198a2468a9793374b76378abf', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65726) authorize /opt/stack/nova/nova/policy.py:192}} [ 787.879674] env[65726]: DEBUG oslo_vmware.api [None req-447a3897-f4ca-48fd-a799-fe155c3fcf63 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Task: {'id': task-5115884, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.149448} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 787.879947] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-447a3897-f4ca-48fd-a799-fe155c3fcf63 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Deleted the datastore file {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 787.880194] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-447a3897-f4ca-48fd-a799-fe155c3fcf63 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] [instance: 65aaa2eb-3ca5-4c2b-92f9-b524366c10dd] Deleted contents of the VM from datastore datastore2 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 787.880395] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-447a3897-f4ca-48fd-a799-fe155c3fcf63 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] [instance: 65aaa2eb-3ca5-4c2b-92f9-b524366c10dd] Instance destroyed {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 787.880568] env[65726]: INFO nova.compute.manager [None req-447a3897-f4ca-48fd-a799-fe155c3fcf63 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] [instance: 65aaa2eb-3ca5-4c2b-92f9-b524366c10dd] Took 1.15 seconds to destroy the instance on the hypervisor. [ 787.881835] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-447a3897-f4ca-48fd-a799-fe155c3fcf63 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 787.881835] env[65726]: DEBUG nova.compute.manager [-] [instance: 65aaa2eb-3ca5-4c2b-92f9-b524366c10dd] Deallocating network for instance {{(pid=65726) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 787.881835] env[65726]: DEBUG nova.network.neutron [-] [instance: 65aaa2eb-3ca5-4c2b-92f9-b524366c10dd] deallocate_for_instance() {{(pid=65726) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 787.881835] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 787.882242] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 787.919399] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 787.920146] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 788.066578] env[65726]: DEBUG nova.network.neutron [None req-300cacab-e765-49c9-a8a1-8e47a20dc6a2 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] [instance: 76249623-6f83-46a3-b8c5-c001111aa698] Successfully created port: e97a5da4-5540-4c0c-a33d-be1182e82aa7 {{(pid=65726) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 788.128837] env[65726]: DEBUG oslo_vmware.api [None req-00483101-e5ba-4006-8481-84d7c7b5b2b8 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Task: {'id': task-5115882, 'name': PowerOnVM_Task, 'duration_secs': 0.524356} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 788.132375] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-00483101-e5ba-4006-8481-84d7c7b5b2b8 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: 9fc0c9ee-9dc7-4059-b8d2-a5db965e00c2] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 788.135285] env[65726]: DEBUG nova.compute.manager [None req-00483101-e5ba-4006-8481-84d7c7b5b2b8 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: 9fc0c9ee-9dc7-4059-b8d2-a5db965e00c2] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 788.136233] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ad3eb0c-1175-466a-b433-93d5b2d5707d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.149249] env[65726]: DEBUG nova.compute.manager [req-27f338ba-e046-4e8d-a8dc-847096e17958 req-3c36a5e3-9c49-4cb6-8bdb-21bf26da997e service nova] [instance: d7905c4f-1ba3-43b3-b4b3-aebacd2f9cc3] Received event network-changed-ee24bb04-7df3-4d06-9751-289b4bdeb753 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 788.149249] env[65726]: DEBUG nova.compute.manager [req-27f338ba-e046-4e8d-a8dc-847096e17958 req-3c36a5e3-9c49-4cb6-8bdb-21bf26da997e service nova] [instance: d7905c4f-1ba3-43b3-b4b3-aebacd2f9cc3] Refreshing instance network info cache due to event network-changed-ee24bb04-7df3-4d06-9751-289b4bdeb753. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 788.149432] env[65726]: DEBUG oslo_concurrency.lockutils [req-27f338ba-e046-4e8d-a8dc-847096e17958 req-3c36a5e3-9c49-4cb6-8bdb-21bf26da997e service nova] Acquiring lock "refresh_cache-d7905c4f-1ba3-43b3-b4b3-aebacd2f9cc3" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 788.149789] env[65726]: DEBUG oslo_concurrency.lockutils [req-27f338ba-e046-4e8d-a8dc-847096e17958 req-3c36a5e3-9c49-4cb6-8bdb-21bf26da997e service nova] Acquired lock "refresh_cache-d7905c4f-1ba3-43b3-b4b3-aebacd2f9cc3" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 788.149789] env[65726]: DEBUG nova.network.neutron [req-27f338ba-e046-4e8d-a8dc-847096e17958 req-3c36a5e3-9c49-4cb6-8bdb-21bf26da997e service nova] [instance: d7905c4f-1ba3-43b3-b4b3-aebacd2f9cc3] Refreshing network info cache for port ee24bb04-7df3-4d06-9751-289b4bdeb753 {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 788.251085] env[65726]: DEBUG nova.compute.manager [None req-300cacab-e765-49c9-a8a1-8e47a20dc6a2 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] [instance: 76249623-6f83-46a3-b8c5-c001111aa698] Start building block device mappings for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 788.304927] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3260fa2a-3a51-48a8-9aea-39966d4e575e tempest-ServersAaction247Test-712120018 tempest-ServersAaction247Test-712120018-project-member] Acquiring lock "refresh_cache-ad48cadd-9bb6-4191-a68a-5c8fdaeebd44" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 788.305127] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3260fa2a-3a51-48a8-9aea-39966d4e575e tempest-ServersAaction247Test-712120018 tempest-ServersAaction247Test-712120018-project-member] Acquired lock "refresh_cache-ad48cadd-9bb6-4191-a68a-5c8fdaeebd44" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 788.305299] env[65726]: DEBUG nova.network.neutron [None req-3260fa2a-3a51-48a8-9aea-39966d4e575e tempest-ServersAaction247Test-712120018 tempest-ServersAaction247Test-712120018-project-member] [instance: ad48cadd-9bb6-4191-a68a-5c8fdaeebd44] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 788.306533] env[65726]: DEBUG nova.compute.manager [None req-9401bfb8-fa64-40fc-97d6-85a4492f4163 tempest-ServersAaction247Test-712120018 tempest-ServersAaction247Test-712120018-project-member] [instance: ad48cadd-9bb6-4191-a68a-5c8fdaeebd44] Instance disappeared during snapshot {{(pid=65726) _snapshot_instance /opt/stack/nova/nova/compute/manager.py:4652}} [ 788.466416] env[65726]: DEBUG nova.compute.manager [None req-9401bfb8-fa64-40fc-97d6-85a4492f4163 tempest-ServersAaction247Test-712120018 tempest-ServersAaction247Test-712120018-project-member] [instance: ad48cadd-9bb6-4191-a68a-5c8fdaeebd44] Found 0 images (rotation: 2) {{(pid=65726) _rotate_backups /opt/stack/nova/nova/compute/manager.py:5074}} [ 788.634625] env[65726]: DEBUG nova.compute.manager [req-b9847c63-ea51-4257-9755-e69b40d8a844 req-9f555806-8fff-4456-8d36-0a6531438309 service nova] [instance: 65aaa2eb-3ca5-4c2b-92f9-b524366c10dd] Received event network-vif-deleted-862071bf-725c-4193-b60b-72b03b4c0553 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 788.634925] env[65726]: INFO nova.compute.manager [req-b9847c63-ea51-4257-9755-e69b40d8a844 req-9f555806-8fff-4456-8d36-0a6531438309 service nova] [instance: 65aaa2eb-3ca5-4c2b-92f9-b524366c10dd] Neutron deleted interface 862071bf-725c-4193-b60b-72b03b4c0553; detaching it from the instance and deleting it from the info cache [ 788.635173] env[65726]: DEBUG nova.network.neutron [req-b9847c63-ea51-4257-9755-e69b40d8a844 req-9f555806-8fff-4456-8d36-0a6531438309 service nova] [instance: 65aaa2eb-3ca5-4c2b-92f9-b524366c10dd] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 788.637812] env[65726]: DEBUG nova.network.neutron [-] [instance: 65aaa2eb-3ca5-4c2b-92f9-b524366c10dd] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 788.655038] env[65726]: WARNING openstack [req-27f338ba-e046-4e8d-a8dc-847096e17958 req-3c36a5e3-9c49-4cb6-8bdb-21bf26da997e service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 788.655038] env[65726]: WARNING openstack [req-27f338ba-e046-4e8d-a8dc-847096e17958 req-3c36a5e3-9c49-4cb6-8bdb-21bf26da997e service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 788.811582] env[65726]: WARNING openstack [None req-3260fa2a-3a51-48a8-9aea-39966d4e575e tempest-ServersAaction247Test-712120018 tempest-ServersAaction247Test-712120018-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 788.812130] env[65726]: WARNING openstack [None req-3260fa2a-3a51-48a8-9aea-39966d4e575e tempest-ServersAaction247Test-712120018 tempest-ServersAaction247Test-712120018-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 788.817398] env[65726]: DEBUG nova.network.neutron [None req-3260fa2a-3a51-48a8-9aea-39966d4e575e tempest-ServersAaction247Test-712120018 tempest-ServersAaction247Test-712120018-project-member] [instance: ad48cadd-9bb6-4191-a68a-5c8fdaeebd44] Instance cache missing network info. {{(pid=65726) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 788.824615] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de80eb45-0525-4acf-8d5e-eeaae2835017 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.832827] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c92e00e4-f2cd-46c4-8e03-1e3c759c58b1 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.865306] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad5bea55-e033-47c6-ac4d-91ac3cc4055c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.875884] env[65726]: WARNING openstack [req-27f338ba-e046-4e8d-a8dc-847096e17958 req-3c36a5e3-9c49-4cb6-8bdb-21bf26da997e service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 788.876307] env[65726]: WARNING openstack [req-27f338ba-e046-4e8d-a8dc-847096e17958 req-3c36a5e3-9c49-4cb6-8bdb-21bf26da997e service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 788.883739] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5178cdb0-e46d-49ea-ac8a-e8a549c7c485 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.900935] env[65726]: DEBUG nova.compute.provider_tree [None req-af1f7990-88d5-411a-be3e-a2d4ed6b5b5b tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 789.041808] env[65726]: WARNING openstack [req-27f338ba-e046-4e8d-a8dc-847096e17958 req-3c36a5e3-9c49-4cb6-8bdb-21bf26da997e service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 789.042879] env[65726]: WARNING openstack [req-27f338ba-e046-4e8d-a8dc-847096e17958 req-3c36a5e3-9c49-4cb6-8bdb-21bf26da997e service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 789.077257] env[65726]: DEBUG nova.network.neutron [None req-3260fa2a-3a51-48a8-9aea-39966d4e575e tempest-ServersAaction247Test-712120018 tempest-ServersAaction247Test-712120018-project-member] [instance: ad48cadd-9bb6-4191-a68a-5c8fdaeebd44] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 789.140200] env[65726]: INFO nova.compute.manager [-] [instance: 65aaa2eb-3ca5-4c2b-92f9-b524366c10dd] Took 1.26 seconds to deallocate network for instance. [ 789.140483] env[65726]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a2784f89-4567-4964-8bc6-f0653c4ebec6 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.153773] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69cfd1c6-0ce9-4695-b411-534553f2049b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.178931] env[65726]: DEBUG nova.compute.manager [None req-2135c7ab-2ff3-4310-b64c-188bebed565e tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 608cc0ea-3f6b-4b4f-83c1-01aa50999d63] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 789.179960] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b12852b-2f50-4d1f-8357-84f519eebab1 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.197655] env[65726]: DEBUG nova.compute.manager [req-b9847c63-ea51-4257-9755-e69b40d8a844 req-9f555806-8fff-4456-8d36-0a6531438309 service nova] [instance: 65aaa2eb-3ca5-4c2b-92f9-b524366c10dd] Detach interface failed, port_id=862071bf-725c-4193-b60b-72b03b4c0553, reason: Instance 65aaa2eb-3ca5-4c2b-92f9-b524366c10dd could not be found. {{(pid=65726) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11669}} [ 789.259190] env[65726]: DEBUG nova.compute.manager [None req-300cacab-e765-49c9-a8a1-8e47a20dc6a2 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] [instance: 76249623-6f83-46a3-b8c5-c001111aa698] Start spawning the instance on the hypervisor. {{(pid=65726) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 789.285914] env[65726]: DEBUG nova.virt.hardware [None req-300cacab-e765-49c9-a8a1-8e47a20dc6a2 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:32:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='d5a398f3-3467-4467-9a69-ab6d1bafa43b',id=37,is_public=True,memory_mb=192,name='tempest-test_resize_flavor_-1568765588',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T19:28:34Z,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1539f06fa3534e90acbc3a60c4b8bd3f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T19:28:35Z,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 789.286206] env[65726]: DEBUG nova.virt.hardware [None req-300cacab-e765-49c9-a8a1-8e47a20dc6a2 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 789.286349] env[65726]: DEBUG nova.virt.hardware [None req-300cacab-e765-49c9-a8a1-8e47a20dc6a2 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 789.286802] env[65726]: DEBUG nova.virt.hardware [None req-300cacab-e765-49c9-a8a1-8e47a20dc6a2 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 789.286802] env[65726]: DEBUG nova.virt.hardware [None req-300cacab-e765-49c9-a8a1-8e47a20dc6a2 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 789.286802] env[65726]: DEBUG nova.virt.hardware [None req-300cacab-e765-49c9-a8a1-8e47a20dc6a2 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 789.286985] env[65726]: DEBUG nova.virt.hardware [None req-300cacab-e765-49c9-a8a1-8e47a20dc6a2 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 789.287422] env[65726]: DEBUG nova.virt.hardware [None req-300cacab-e765-49c9-a8a1-8e47a20dc6a2 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 789.287422] env[65726]: DEBUG nova.virt.hardware [None req-300cacab-e765-49c9-a8a1-8e47a20dc6a2 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 789.287534] env[65726]: DEBUG nova.virt.hardware [None req-300cacab-e765-49c9-a8a1-8e47a20dc6a2 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 789.287665] env[65726]: DEBUG nova.virt.hardware [None req-300cacab-e765-49c9-a8a1-8e47a20dc6a2 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 789.288573] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74ada57c-1979-433c-9590-b2ccda9884bb {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.298806] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e367e988-eda1-49d1-8a62-94279ad2149f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.404469] env[65726]: DEBUG nova.scheduler.client.report [None req-af1f7990-88d5-411a-be3e-a2d4ed6b5b5b tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 789.505555] env[65726]: DEBUG nova.network.neutron [req-27f338ba-e046-4e8d-a8dc-847096e17958 req-3c36a5e3-9c49-4cb6-8bdb-21bf26da997e service nova] [instance: d7905c4f-1ba3-43b3-b4b3-aebacd2f9cc3] Updated VIF entry in instance network info cache for port ee24bb04-7df3-4d06-9751-289b4bdeb753. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 789.506077] env[65726]: DEBUG nova.network.neutron [req-27f338ba-e046-4e8d-a8dc-847096e17958 req-3c36a5e3-9c49-4cb6-8bdb-21bf26da997e service nova] [instance: d7905c4f-1ba3-43b3-b4b3-aebacd2f9cc3] Updating instance_info_cache with network_info: [{"id": "ee24bb04-7df3-4d06-9751-289b4bdeb753", "address": "fa:16:3e:82:91:65", "network": {"id": "27801f9a-9912-41cf-be65-bc9903f4952f", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-781436951-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.155", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aac5fe6c99f740ada3747088b2b0ad89", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dd7d0d95-6848-4e69-ac21-75f8db82a3b5", "external-id": "nsx-vlan-transportzone-272", "segmentation_id": 272, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapee24bb04-7d", "ovs_interfaceid": "ee24bb04-7df3-4d06-9751-289b4bdeb753", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 789.580523] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3260fa2a-3a51-48a8-9aea-39966d4e575e tempest-ServersAaction247Test-712120018 tempest-ServersAaction247Test-712120018-project-member] Releasing lock "refresh_cache-ad48cadd-9bb6-4191-a68a-5c8fdaeebd44" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 789.580969] env[65726]: DEBUG nova.compute.manager [None req-3260fa2a-3a51-48a8-9aea-39966d4e575e tempest-ServersAaction247Test-712120018 tempest-ServersAaction247Test-712120018-project-member] [instance: ad48cadd-9bb6-4191-a68a-5c8fdaeebd44] Start destroying the instance on the hypervisor. {{(pid=65726) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 789.581230] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-3260fa2a-3a51-48a8-9aea-39966d4e575e tempest-ServersAaction247Test-712120018 tempest-ServersAaction247Test-712120018-project-member] [instance: ad48cadd-9bb6-4191-a68a-5c8fdaeebd44] Destroying instance {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 789.582132] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61884b42-b5bb-4e01-9f61-ef1bc3efaf1f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.591871] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-3260fa2a-3a51-48a8-9aea-39966d4e575e tempest-ServersAaction247Test-712120018 tempest-ServersAaction247Test-712120018-project-member] [instance: ad48cadd-9bb6-4191-a68a-5c8fdaeebd44] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 789.591871] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4c57a905-a233-4162-9b7b-cc4c544d4490 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.604522] env[65726]: DEBUG oslo_vmware.api [None req-3260fa2a-3a51-48a8-9aea-39966d4e575e tempest-ServersAaction247Test-712120018 tempest-ServersAaction247Test-712120018-project-member] Waiting for the task: (returnval){ [ 789.604522] env[65726]: value = "task-5115885" [ 789.604522] env[65726]: _type = "Task" [ 789.604522] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 789.615788] env[65726]: DEBUG oslo_vmware.api [None req-3260fa2a-3a51-48a8-9aea-39966d4e575e tempest-ServersAaction247Test-712120018 tempest-ServersAaction247Test-712120018-project-member] Task: {'id': task-5115885, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.653882] env[65726]: DEBUG oslo_concurrency.lockutils [None req-447a3897-f4ca-48fd-a799-fe155c3fcf63 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 789.706993] env[65726]: INFO nova.compute.manager [None req-2135c7ab-2ff3-4310-b64c-188bebed565e tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 608cc0ea-3f6b-4b4f-83c1-01aa50999d63] instance snapshotting [ 789.707440] env[65726]: DEBUG nova.objects.instance [None req-2135c7ab-2ff3-4310-b64c-188bebed565e tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Lazy-loading 'flavor' on Instance uuid 608cc0ea-3f6b-4b4f-83c1-01aa50999d63 {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 789.723597] env[65726]: DEBUG nova.network.neutron [None req-300cacab-e765-49c9-a8a1-8e47a20dc6a2 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] [instance: 76249623-6f83-46a3-b8c5-c001111aa698] Successfully updated port: e97a5da4-5540-4c0c-a33d-be1182e82aa7 {{(pid=65726) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 789.910224] env[65726]: DEBUG oslo_concurrency.lockutils [None req-af1f7990-88d5-411a-be3e-a2d4ed6b5b5b tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.671s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 789.910818] env[65726]: DEBUG nova.compute.manager [None req-af1f7990-88d5-411a-be3e-a2d4ed6b5b5b tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: 621b899a-e9ec-4fc0-a574-4a08cecff6d9] Start building networks asynchronously for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 789.913700] env[65726]: DEBUG oslo_concurrency.lockutils [None req-1640464d-fd11-4b5f-b0e0-f0de74fa607d tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 33.046s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 789.913957] env[65726]: DEBUG nova.objects.instance [None req-1640464d-fd11-4b5f-b0e0-f0de74fa607d tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Lazy-loading 'resources' on Instance uuid bd839fe1-8801-4ba1-9c23-288453258b66 {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 790.009298] env[65726]: DEBUG oslo_concurrency.lockutils [req-27f338ba-e046-4e8d-a8dc-847096e17958 req-3c36a5e3-9c49-4cb6-8bdb-21bf26da997e service nova] Releasing lock "refresh_cache-d7905c4f-1ba3-43b3-b4b3-aebacd2f9cc3" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 790.116509] env[65726]: DEBUG oslo_vmware.api [None req-3260fa2a-3a51-48a8-9aea-39966d4e575e tempest-ServersAaction247Test-712120018 tempest-ServersAaction247Test-712120018-project-member] Task: {'id': task-5115885, 'name': PowerOffVM_Task, 'duration_secs': 0.233303} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 790.116822] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-3260fa2a-3a51-48a8-9aea-39966d4e575e tempest-ServersAaction247Test-712120018 tempest-ServersAaction247Test-712120018-project-member] [instance: ad48cadd-9bb6-4191-a68a-5c8fdaeebd44] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 790.117050] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-3260fa2a-3a51-48a8-9aea-39966d4e575e tempest-ServersAaction247Test-712120018 tempest-ServersAaction247Test-712120018-project-member] [instance: ad48cadd-9bb6-4191-a68a-5c8fdaeebd44] Unregistering the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 790.117353] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8147cf83-721f-45b1-9291-17054c384c80 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.150816] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-3260fa2a-3a51-48a8-9aea-39966d4e575e tempest-ServersAaction247Test-712120018 tempest-ServersAaction247Test-712120018-project-member] [instance: ad48cadd-9bb6-4191-a68a-5c8fdaeebd44] Unregistered the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 790.150816] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-3260fa2a-3a51-48a8-9aea-39966d4e575e tempest-ServersAaction247Test-712120018 tempest-ServersAaction247Test-712120018-project-member] [instance: ad48cadd-9bb6-4191-a68a-5c8fdaeebd44] Deleting contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 790.151218] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-3260fa2a-3a51-48a8-9aea-39966d4e575e tempest-ServersAaction247Test-712120018 tempest-ServersAaction247Test-712120018-project-member] Deleting the datastore file [datastore1] ad48cadd-9bb6-4191-a68a-5c8fdaeebd44 {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 790.151633] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d82f174d-0c30-407a-a6e1-7d299d151baf {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.163106] env[65726]: DEBUG oslo_vmware.api [None req-3260fa2a-3a51-48a8-9aea-39966d4e575e tempest-ServersAaction247Test-712120018 tempest-ServersAaction247Test-712120018-project-member] Waiting for the task: (returnval){ [ 790.163106] env[65726]: value = "task-5115887" [ 790.163106] env[65726]: _type = "Task" [ 790.163106] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 790.172706] env[65726]: DEBUG oslo_vmware.api [None req-3260fa2a-3a51-48a8-9aea-39966d4e575e tempest-ServersAaction247Test-712120018 tempest-ServersAaction247Test-712120018-project-member] Task: {'id': task-5115887, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.214650] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d34926b-44c8-48c5-ab4d-a6d3ef04f27a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.237944] env[65726]: DEBUG oslo_concurrency.lockutils [None req-300cacab-e765-49c9-a8a1-8e47a20dc6a2 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Acquiring lock "refresh_cache-76249623-6f83-46a3-b8c5-c001111aa698" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 790.238139] env[65726]: DEBUG oslo_concurrency.lockutils [None req-300cacab-e765-49c9-a8a1-8e47a20dc6a2 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Acquired lock "refresh_cache-76249623-6f83-46a3-b8c5-c001111aa698" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 790.238317] env[65726]: DEBUG nova.network.neutron [None req-300cacab-e765-49c9-a8a1-8e47a20dc6a2 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] [instance: 76249623-6f83-46a3-b8c5-c001111aa698] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 790.241248] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b37edd3-1821-4951-bda3-33ecac0ced29 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.417829] env[65726]: DEBUG nova.compute.utils [None req-af1f7990-88d5-411a-be3e-a2d4ed6b5b5b tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Using /dev/sd instead of None {{(pid=65726) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 790.418769] env[65726]: DEBUG nova.compute.manager [None req-af1f7990-88d5-411a-be3e-a2d4ed6b5b5b tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: 621b899a-e9ec-4fc0-a574-4a08cecff6d9] Allocating IP information in the background. {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 790.418969] env[65726]: DEBUG nova.network.neutron [None req-af1f7990-88d5-411a-be3e-a2d4ed6b5b5b tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: 621b899a-e9ec-4fc0-a574-4a08cecff6d9] allocate_for_instance() {{(pid=65726) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 790.419318] env[65726]: WARNING neutronclient.v2_0.client [None req-af1f7990-88d5-411a-be3e-a2d4ed6b5b5b tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 790.419616] env[65726]: WARNING neutronclient.v2_0.client [None req-af1f7990-88d5-411a-be3e-a2d4ed6b5b5b tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 790.420297] env[65726]: WARNING openstack [None req-af1f7990-88d5-411a-be3e-a2d4ed6b5b5b tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 790.420769] env[65726]: WARNING openstack [None req-af1f7990-88d5-411a-be3e-a2d4ed6b5b5b tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 790.487218] env[65726]: DEBUG nova.policy [None req-af1f7990-88d5-411a-be3e-a2d4ed6b5b5b tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4cd1ed63db174f21be78f86554fada0a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'cbb17a09b35c4c22ade5c4082b10bda9', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65726) authorize /opt/stack/nova/nova/policy.py:192}} [ 790.494602] env[65726]: DEBUG nova.compute.manager [req-7e0f3c08-e637-4a7a-8686-651b076bf4cb req-c32e8064-4a17-48a0-a836-1c6ed27d6b03 service nova] [instance: 76249623-6f83-46a3-b8c5-c001111aa698] Received event network-vif-plugged-e97a5da4-5540-4c0c-a33d-be1182e82aa7 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 790.494602] env[65726]: DEBUG oslo_concurrency.lockutils [req-7e0f3c08-e637-4a7a-8686-651b076bf4cb req-c32e8064-4a17-48a0-a836-1c6ed27d6b03 service nova] Acquiring lock "76249623-6f83-46a3-b8c5-c001111aa698-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 790.495322] env[65726]: DEBUG oslo_concurrency.lockutils [req-7e0f3c08-e637-4a7a-8686-651b076bf4cb req-c32e8064-4a17-48a0-a836-1c6ed27d6b03 service nova] Lock "76249623-6f83-46a3-b8c5-c001111aa698-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 790.495322] env[65726]: DEBUG oslo_concurrency.lockutils [req-7e0f3c08-e637-4a7a-8686-651b076bf4cb req-c32e8064-4a17-48a0-a836-1c6ed27d6b03 service nova] Lock "76249623-6f83-46a3-b8c5-c001111aa698-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 790.495637] env[65726]: DEBUG nova.compute.manager [req-7e0f3c08-e637-4a7a-8686-651b076bf4cb req-c32e8064-4a17-48a0-a836-1c6ed27d6b03 service nova] [instance: 76249623-6f83-46a3-b8c5-c001111aa698] No waiting events found dispatching network-vif-plugged-e97a5da4-5540-4c0c-a33d-be1182e82aa7 {{(pid=65726) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 790.495637] env[65726]: WARNING nova.compute.manager [req-7e0f3c08-e637-4a7a-8686-651b076bf4cb req-c32e8064-4a17-48a0-a836-1c6ed27d6b03 service nova] [instance: 76249623-6f83-46a3-b8c5-c001111aa698] Received unexpected event network-vif-plugged-e97a5da4-5540-4c0c-a33d-be1182e82aa7 for instance with vm_state building and task_state spawning. [ 790.495791] env[65726]: DEBUG nova.compute.manager [req-7e0f3c08-e637-4a7a-8686-651b076bf4cb req-c32e8064-4a17-48a0-a836-1c6ed27d6b03 service nova] [instance: 76249623-6f83-46a3-b8c5-c001111aa698] Received event network-changed-e97a5da4-5540-4c0c-a33d-be1182e82aa7 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 790.496022] env[65726]: DEBUG nova.compute.manager [req-7e0f3c08-e637-4a7a-8686-651b076bf4cb req-c32e8064-4a17-48a0-a836-1c6ed27d6b03 service nova] [instance: 76249623-6f83-46a3-b8c5-c001111aa698] Refreshing instance network info cache due to event network-changed-e97a5da4-5540-4c0c-a33d-be1182e82aa7. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 790.496273] env[65726]: DEBUG oslo_concurrency.lockutils [req-7e0f3c08-e637-4a7a-8686-651b076bf4cb req-c32e8064-4a17-48a0-a836-1c6ed27d6b03 service nova] Acquiring lock "refresh_cache-76249623-6f83-46a3-b8c5-c001111aa698" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 790.682026] env[65726]: DEBUG oslo_vmware.api [None req-3260fa2a-3a51-48a8-9aea-39966d4e575e tempest-ServersAaction247Test-712120018 tempest-ServersAaction247Test-712120018-project-member] Task: {'id': task-5115887, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.124268} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 790.683799] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-3260fa2a-3a51-48a8-9aea-39966d4e575e tempest-ServersAaction247Test-712120018 tempest-ServersAaction247Test-712120018-project-member] Deleted the datastore file {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 790.683799] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-3260fa2a-3a51-48a8-9aea-39966d4e575e tempest-ServersAaction247Test-712120018 tempest-ServersAaction247Test-712120018-project-member] [instance: ad48cadd-9bb6-4191-a68a-5c8fdaeebd44] Deleted contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 790.683799] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-3260fa2a-3a51-48a8-9aea-39966d4e575e tempest-ServersAaction247Test-712120018 tempest-ServersAaction247Test-712120018-project-member] [instance: ad48cadd-9bb6-4191-a68a-5c8fdaeebd44] Instance destroyed {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 790.683799] env[65726]: INFO nova.compute.manager [None req-3260fa2a-3a51-48a8-9aea-39966d4e575e tempest-ServersAaction247Test-712120018 tempest-ServersAaction247Test-712120018-project-member] [instance: ad48cadd-9bb6-4191-a68a-5c8fdaeebd44] Took 1.10 seconds to destroy the instance on the hypervisor. [ 790.683799] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-3260fa2a-3a51-48a8-9aea-39966d4e575e tempest-ServersAaction247Test-712120018 tempest-ServersAaction247Test-712120018-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 790.684042] env[65726]: DEBUG nova.compute.manager [-] [instance: ad48cadd-9bb6-4191-a68a-5c8fdaeebd44] Deallocating network for instance {{(pid=65726) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 790.684042] env[65726]: DEBUG nova.network.neutron [-] [instance: ad48cadd-9bb6-4191-a68a-5c8fdaeebd44] deallocate_for_instance() {{(pid=65726) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 790.684042] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 790.684243] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 790.733221] env[65726]: DEBUG nova.network.neutron [-] [instance: ad48cadd-9bb6-4191-a68a-5c8fdaeebd44] Instance cache missing network info. {{(pid=65726) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 790.734146] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 790.734146] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 790.745217] env[65726]: DEBUG oslo_concurrency.lockutils [None req-d4c6362a-f5ab-42dd-9c7f-bfc3bee5f94c tempest-VolumesAssistedSnapshotsTest-1875841935 tempest-VolumesAssistedSnapshotsTest-1875841935-project-member] Acquiring lock "aa2716a9-2cf8-4e51-800d-1c8f98b2cf05" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 790.745921] env[65726]: DEBUG oslo_concurrency.lockutils [None req-d4c6362a-f5ab-42dd-9c7f-bfc3bee5f94c tempest-VolumesAssistedSnapshotsTest-1875841935 tempest-VolumesAssistedSnapshotsTest-1875841935-project-member] Lock "aa2716a9-2cf8-4e51-800d-1c8f98b2cf05" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 790.745921] env[65726]: DEBUG oslo_concurrency.lockutils [None req-d4c6362a-f5ab-42dd-9c7f-bfc3bee5f94c tempest-VolumesAssistedSnapshotsTest-1875841935 tempest-VolumesAssistedSnapshotsTest-1875841935-project-member] Acquiring lock "aa2716a9-2cf8-4e51-800d-1c8f98b2cf05-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 790.745921] env[65726]: DEBUG oslo_concurrency.lockutils [None req-d4c6362a-f5ab-42dd-9c7f-bfc3bee5f94c tempest-VolumesAssistedSnapshotsTest-1875841935 tempest-VolumesAssistedSnapshotsTest-1875841935-project-member] Lock "aa2716a9-2cf8-4e51-800d-1c8f98b2cf05-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 790.746203] env[65726]: DEBUG oslo_concurrency.lockutils [None req-d4c6362a-f5ab-42dd-9c7f-bfc3bee5f94c tempest-VolumesAssistedSnapshotsTest-1875841935 tempest-VolumesAssistedSnapshotsTest-1875841935-project-member] Lock "aa2716a9-2cf8-4e51-800d-1c8f98b2cf05-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 790.748083] env[65726]: WARNING openstack [None req-300cacab-e765-49c9-a8a1-8e47a20dc6a2 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 790.748652] env[65726]: WARNING openstack [None req-300cacab-e765-49c9-a8a1-8e47a20dc6a2 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 790.753325] env[65726]: DEBUG nova.network.neutron [None req-300cacab-e765-49c9-a8a1-8e47a20dc6a2 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] [instance: 76249623-6f83-46a3-b8c5-c001111aa698] Instance cache missing network info. {{(pid=65726) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 790.758653] env[65726]: INFO nova.compute.manager [None req-d4c6362a-f5ab-42dd-9c7f-bfc3bee5f94c tempest-VolumesAssistedSnapshotsTest-1875841935 tempest-VolumesAssistedSnapshotsTest-1875841935-project-member] [instance: aa2716a9-2cf8-4e51-800d-1c8f98b2cf05] Terminating instance [ 790.761313] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-2135c7ab-2ff3-4310-b64c-188bebed565e tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 608cc0ea-3f6b-4b4f-83c1-01aa50999d63] Creating Snapshot of the VM instance {{(pid=65726) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 790.762996] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-cb16b4e7-a461-48a4-9127-efa3d27f1b3a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.767706] env[65726]: DEBUG nova.compute.manager [req-3db31e76-7964-4ab8-8b0f-01c60dec0dc7 req-018b92a4-404c-4c21-9f8d-308ec492a017 service nova] [instance: d7905c4f-1ba3-43b3-b4b3-aebacd2f9cc3] Received event network-changed-ee24bb04-7df3-4d06-9751-289b4bdeb753 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 790.767901] env[65726]: DEBUG nova.compute.manager [req-3db31e76-7964-4ab8-8b0f-01c60dec0dc7 req-018b92a4-404c-4c21-9f8d-308ec492a017 service nova] [instance: d7905c4f-1ba3-43b3-b4b3-aebacd2f9cc3] Refreshing instance network info cache due to event network-changed-ee24bb04-7df3-4d06-9751-289b4bdeb753. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 790.768143] env[65726]: DEBUG oslo_concurrency.lockutils [req-3db31e76-7964-4ab8-8b0f-01c60dec0dc7 req-018b92a4-404c-4c21-9f8d-308ec492a017 service nova] Acquiring lock "refresh_cache-d7905c4f-1ba3-43b3-b4b3-aebacd2f9cc3" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 790.768284] env[65726]: DEBUG oslo_concurrency.lockutils [req-3db31e76-7964-4ab8-8b0f-01c60dec0dc7 req-018b92a4-404c-4c21-9f8d-308ec492a017 service nova] Acquired lock "refresh_cache-d7905c4f-1ba3-43b3-b4b3-aebacd2f9cc3" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 790.768443] env[65726]: DEBUG nova.network.neutron [req-3db31e76-7964-4ab8-8b0f-01c60dec0dc7 req-018b92a4-404c-4c21-9f8d-308ec492a017 service nova] [instance: d7905c4f-1ba3-43b3-b4b3-aebacd2f9cc3] Refreshing network info cache for port ee24bb04-7df3-4d06-9751-289b4bdeb753 {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 790.776483] env[65726]: DEBUG oslo_vmware.api [None req-2135c7ab-2ff3-4310-b64c-188bebed565e tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Waiting for the task: (returnval){ [ 790.776483] env[65726]: value = "task-5115888" [ 790.776483] env[65726]: _type = "Task" [ 790.776483] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 790.795479] env[65726]: DEBUG oslo_vmware.api [None req-2135c7ab-2ff3-4310-b64c-188bebed565e tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5115888, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.823856] env[65726]: WARNING openstack [None req-300cacab-e765-49c9-a8a1-8e47a20dc6a2 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 790.824264] env[65726]: WARNING openstack [None req-300cacab-e765-49c9-a8a1-8e47a20dc6a2 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 790.867136] env[65726]: DEBUG nova.network.neutron [None req-af1f7990-88d5-411a-be3e-a2d4ed6b5b5b tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: 621b899a-e9ec-4fc0-a574-4a08cecff6d9] Successfully created port: dcdfc13e-45af-4bc2-970d-b60727f42176 {{(pid=65726) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 790.930314] env[65726]: DEBUG nova.compute.manager [None req-af1f7990-88d5-411a-be3e-a2d4ed6b5b5b tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: 621b899a-e9ec-4fc0-a574-4a08cecff6d9] Start building block device mappings for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 790.956676] env[65726]: DEBUG oslo_concurrency.lockutils [None req-6397aebb-7e67-4113-8134-9f602fab9ef7 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Acquiring lock "65aaa2eb-3ca5-4c2b-92f9-b524366c10dd" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 791.067595] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a1b8b8f-74d6-492b-b41e-4d200e7ba6c0 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.078105] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73b257cd-da88-46a7-bd1c-1c21c00228d6 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.131840] env[65726]: WARNING openstack [None req-300cacab-e765-49c9-a8a1-8e47a20dc6a2 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 791.132507] env[65726]: WARNING openstack [None req-300cacab-e765-49c9-a8a1-8e47a20dc6a2 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 791.147416] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd8587a3-f4c5-40e9-8124-89225ea4e4b0 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.161021] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad69161f-7e5d-476e-ae2a-53a2ec374119 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.179681] env[65726]: DEBUG nova.compute.provider_tree [None req-1640464d-fd11-4b5f-b0e0-f0de74fa607d tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 791.256060] env[65726]: DEBUG nova.network.neutron [-] [instance: ad48cadd-9bb6-4191-a68a-5c8fdaeebd44] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 791.266772] env[65726]: DEBUG nova.compute.manager [None req-d4c6362a-f5ab-42dd-9c7f-bfc3bee5f94c tempest-VolumesAssistedSnapshotsTest-1875841935 tempest-VolumesAssistedSnapshotsTest-1875841935-project-member] [instance: aa2716a9-2cf8-4e51-800d-1c8f98b2cf05] Start destroying the instance on the hypervisor. {{(pid=65726) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 791.267084] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-d4c6362a-f5ab-42dd-9c7f-bfc3bee5f94c tempest-VolumesAssistedSnapshotsTest-1875841935 tempest-VolumesAssistedSnapshotsTest-1875841935-project-member] [instance: aa2716a9-2cf8-4e51-800d-1c8f98b2cf05] Destroying instance {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 791.268155] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e067aefd-7c5d-4e35-85b3-9d4e0006c0ec {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.272708] env[65726]: WARNING openstack [req-3db31e76-7964-4ab8-8b0f-01c60dec0dc7 req-018b92a4-404c-4c21-9f8d-308ec492a017 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 791.273116] env[65726]: WARNING openstack [req-3db31e76-7964-4ab8-8b0f-01c60dec0dc7 req-018b92a4-404c-4c21-9f8d-308ec492a017 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 791.286879] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-d4c6362a-f5ab-42dd-9c7f-bfc3bee5f94c tempest-VolumesAssistedSnapshotsTest-1875841935 tempest-VolumesAssistedSnapshotsTest-1875841935-project-member] [instance: aa2716a9-2cf8-4e51-800d-1c8f98b2cf05] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 791.287570] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f4a25aa5-13b7-47d7-82e2-b85e931dbc78 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.293558] env[65726]: DEBUG oslo_vmware.api [None req-2135c7ab-2ff3-4310-b64c-188bebed565e tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5115888, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 791.295101] env[65726]: DEBUG oslo_vmware.api [None req-d4c6362a-f5ab-42dd-9c7f-bfc3bee5f94c tempest-VolumesAssistedSnapshotsTest-1875841935 tempest-VolumesAssistedSnapshotsTest-1875841935-project-member] Waiting for the task: (returnval){ [ 791.295101] env[65726]: value = "task-5115889" [ 791.295101] env[65726]: _type = "Task" [ 791.295101] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 791.314404] env[65726]: DEBUG oslo_vmware.api [None req-d4c6362a-f5ab-42dd-9c7f-bfc3bee5f94c tempest-VolumesAssistedSnapshotsTest-1875841935 tempest-VolumesAssistedSnapshotsTest-1875841935-project-member] Task: {'id': task-5115889, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 791.440717] env[65726]: DEBUG nova.network.neutron [None req-300cacab-e765-49c9-a8a1-8e47a20dc6a2 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] [instance: 76249623-6f83-46a3-b8c5-c001111aa698] Updating instance_info_cache with network_info: [{"id": "e97a5da4-5540-4c0c-a33d-be1182e82aa7", "address": "fa:16:3e:af:b6:0f", "network": {"id": "55d629b8-a541-4720-9289-927bbc6e6f22", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.253", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "1539f06fa3534e90acbc3a60c4b8bd3f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "35342bcb-8b06-472e-b3c0-43fd3d6c4b30", "external-id": "nsx-vlan-transportzone-524", "segmentation_id": 524, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape97a5da4-55", "ovs_interfaceid": "e97a5da4-5540-4c0c-a33d-be1182e82aa7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 791.543847] env[65726]: WARNING openstack [req-3db31e76-7964-4ab8-8b0f-01c60dec0dc7 req-018b92a4-404c-4c21-9f8d-308ec492a017 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 791.544248] env[65726]: WARNING openstack [req-3db31e76-7964-4ab8-8b0f-01c60dec0dc7 req-018b92a4-404c-4c21-9f8d-308ec492a017 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 791.623936] env[65726]: WARNING openstack [req-3db31e76-7964-4ab8-8b0f-01c60dec0dc7 req-018b92a4-404c-4c21-9f8d-308ec492a017 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 791.624354] env[65726]: WARNING openstack [req-3db31e76-7964-4ab8-8b0f-01c60dec0dc7 req-018b92a4-404c-4c21-9f8d-308ec492a017 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 791.686094] env[65726]: DEBUG nova.scheduler.client.report [None req-1640464d-fd11-4b5f-b0e0-f0de74fa607d tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 791.728211] env[65726]: DEBUG nova.network.neutron [req-3db31e76-7964-4ab8-8b0f-01c60dec0dc7 req-018b92a4-404c-4c21-9f8d-308ec492a017 service nova] [instance: d7905c4f-1ba3-43b3-b4b3-aebacd2f9cc3] Updated VIF entry in instance network info cache for port ee24bb04-7df3-4d06-9751-289b4bdeb753. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 791.728938] env[65726]: DEBUG nova.network.neutron [req-3db31e76-7964-4ab8-8b0f-01c60dec0dc7 req-018b92a4-404c-4c21-9f8d-308ec492a017 service nova] [instance: d7905c4f-1ba3-43b3-b4b3-aebacd2f9cc3] Updating instance_info_cache with network_info: [{"id": "ee24bb04-7df3-4d06-9751-289b4bdeb753", "address": "fa:16:3e:82:91:65", "network": {"id": "27801f9a-9912-41cf-be65-bc9903f4952f", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-781436951-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aac5fe6c99f740ada3747088b2b0ad89", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dd7d0d95-6848-4e69-ac21-75f8db82a3b5", "external-id": "nsx-vlan-transportzone-272", "segmentation_id": 272, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapee24bb04-7d", "ovs_interfaceid": "ee24bb04-7df3-4d06-9751-289b4bdeb753", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 791.758737] env[65726]: INFO nova.compute.manager [-] [instance: ad48cadd-9bb6-4191-a68a-5c8fdaeebd44] Took 1.08 seconds to deallocate network for instance. [ 791.793152] env[65726]: DEBUG oslo_vmware.api [None req-2135c7ab-2ff3-4310-b64c-188bebed565e tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5115888, 'name': CreateSnapshot_Task, 'duration_secs': 0.854839} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 791.793152] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-2135c7ab-2ff3-4310-b64c-188bebed565e tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 608cc0ea-3f6b-4b4f-83c1-01aa50999d63] Created Snapshot of the VM instance {{(pid=65726) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 791.793744] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-162386cf-0684-4303-8108-e3a98241595e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.812400] env[65726]: DEBUG oslo_vmware.api [None req-d4c6362a-f5ab-42dd-9c7f-bfc3bee5f94c tempest-VolumesAssistedSnapshotsTest-1875841935 tempest-VolumesAssistedSnapshotsTest-1875841935-project-member] Task: {'id': task-5115889, 'name': PowerOffVM_Task, 'duration_secs': 0.201141} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 791.812855] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-d4c6362a-f5ab-42dd-9c7f-bfc3bee5f94c tempest-VolumesAssistedSnapshotsTest-1875841935 tempest-VolumesAssistedSnapshotsTest-1875841935-project-member] [instance: aa2716a9-2cf8-4e51-800d-1c8f98b2cf05] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 791.813092] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-d4c6362a-f5ab-42dd-9c7f-bfc3bee5f94c tempest-VolumesAssistedSnapshotsTest-1875841935 tempest-VolumesAssistedSnapshotsTest-1875841935-project-member] [instance: aa2716a9-2cf8-4e51-800d-1c8f98b2cf05] Unregistering the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 791.813570] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-859a403e-1b1c-4662-8d80-870e6957fa6c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.919409] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-d4c6362a-f5ab-42dd-9c7f-bfc3bee5f94c tempest-VolumesAssistedSnapshotsTest-1875841935 tempest-VolumesAssistedSnapshotsTest-1875841935-project-member] [instance: aa2716a9-2cf8-4e51-800d-1c8f98b2cf05] Unregistered the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 791.919409] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-d4c6362a-f5ab-42dd-9c7f-bfc3bee5f94c tempest-VolumesAssistedSnapshotsTest-1875841935 tempest-VolumesAssistedSnapshotsTest-1875841935-project-member] [instance: aa2716a9-2cf8-4e51-800d-1c8f98b2cf05] Deleting contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 791.919628] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-d4c6362a-f5ab-42dd-9c7f-bfc3bee5f94c tempest-VolumesAssistedSnapshotsTest-1875841935 tempest-VolumesAssistedSnapshotsTest-1875841935-project-member] Deleting the datastore file [datastore1] aa2716a9-2cf8-4e51-800d-1c8f98b2cf05 {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 791.921645] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2d7b06fd-25a1-4f38-a017-7dd110b27a41 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.927584] env[65726]: DEBUG oslo_vmware.api [None req-d4c6362a-f5ab-42dd-9c7f-bfc3bee5f94c tempest-VolumesAssistedSnapshotsTest-1875841935 tempest-VolumesAssistedSnapshotsTest-1875841935-project-member] Waiting for the task: (returnval){ [ 791.927584] env[65726]: value = "task-5115891" [ 791.927584] env[65726]: _type = "Task" [ 791.927584] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 791.937239] env[65726]: DEBUG oslo_vmware.api [None req-d4c6362a-f5ab-42dd-9c7f-bfc3bee5f94c tempest-VolumesAssistedSnapshotsTest-1875841935 tempest-VolumesAssistedSnapshotsTest-1875841935-project-member] Task: {'id': task-5115891, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 791.947773] env[65726]: DEBUG nova.compute.manager [None req-af1f7990-88d5-411a-be3e-a2d4ed6b5b5b tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: 621b899a-e9ec-4fc0-a574-4a08cecff6d9] Start spawning the instance on the hypervisor. {{(pid=65726) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 791.950408] env[65726]: DEBUG oslo_concurrency.lockutils [None req-300cacab-e765-49c9-a8a1-8e47a20dc6a2 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Releasing lock "refresh_cache-76249623-6f83-46a3-b8c5-c001111aa698" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 791.950979] env[65726]: DEBUG nova.compute.manager [None req-300cacab-e765-49c9-a8a1-8e47a20dc6a2 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] [instance: 76249623-6f83-46a3-b8c5-c001111aa698] Instance network_info: |[{"id": "e97a5da4-5540-4c0c-a33d-be1182e82aa7", "address": "fa:16:3e:af:b6:0f", "network": {"id": "55d629b8-a541-4720-9289-927bbc6e6f22", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.253", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "1539f06fa3534e90acbc3a60c4b8bd3f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "35342bcb-8b06-472e-b3c0-43fd3d6c4b30", "external-id": "nsx-vlan-transportzone-524", "segmentation_id": 524, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape97a5da4-55", "ovs_interfaceid": "e97a5da4-5540-4c0c-a33d-be1182e82aa7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 791.951131] env[65726]: DEBUG oslo_concurrency.lockutils [req-7e0f3c08-e637-4a7a-8686-651b076bf4cb req-c32e8064-4a17-48a0-a836-1c6ed27d6b03 service nova] Acquired lock "refresh_cache-76249623-6f83-46a3-b8c5-c001111aa698" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 791.951302] env[65726]: DEBUG nova.network.neutron [req-7e0f3c08-e637-4a7a-8686-651b076bf4cb req-c32e8064-4a17-48a0-a836-1c6ed27d6b03 service nova] [instance: 76249623-6f83-46a3-b8c5-c001111aa698] Refreshing network info cache for port e97a5da4-5540-4c0c-a33d-be1182e82aa7 {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 791.952520] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-300cacab-e765-49c9-a8a1-8e47a20dc6a2 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] [instance: 76249623-6f83-46a3-b8c5-c001111aa698] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:af:b6:0f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '35342bcb-8b06-472e-b3c0-43fd3d6c4b30', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e97a5da4-5540-4c0c-a33d-be1182e82aa7', 'vif_model': 'vmxnet3'}] {{(pid=65726) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 791.968520] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-300cacab-e765-49c9-a8a1-8e47a20dc6a2 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 791.969714] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 76249623-6f83-46a3-b8c5-c001111aa698] Creating VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 791.972345] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-362b236c-67d2-4d8b-9fca-0c34ad846f50 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.993506] env[65726]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 791.993506] env[65726]: value = "task-5115892" [ 791.993506] env[65726]: _type = "Task" [ 791.993506] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 791.999188] env[65726]: DEBUG nova.virt.hardware [None req-af1f7990-88d5-411a-be3e-a2d4ed6b5b5b tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T19:28:34Z,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1539f06fa3534e90acbc3a60c4b8bd3f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T19:28:35Z,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 791.999429] env[65726]: DEBUG nova.virt.hardware [None req-af1f7990-88d5-411a-be3e-a2d4ed6b5b5b tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 791.999581] env[65726]: DEBUG nova.virt.hardware [None req-af1f7990-88d5-411a-be3e-a2d4ed6b5b5b tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 791.999754] env[65726]: DEBUG nova.virt.hardware [None req-af1f7990-88d5-411a-be3e-a2d4ed6b5b5b tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 791.999913] env[65726]: DEBUG nova.virt.hardware [None req-af1f7990-88d5-411a-be3e-a2d4ed6b5b5b tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 792.000078] env[65726]: DEBUG nova.virt.hardware [None req-af1f7990-88d5-411a-be3e-a2d4ed6b5b5b tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 792.000291] env[65726]: DEBUG nova.virt.hardware [None req-af1f7990-88d5-411a-be3e-a2d4ed6b5b5b tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 792.000446] env[65726]: DEBUG nova.virt.hardware [None req-af1f7990-88d5-411a-be3e-a2d4ed6b5b5b tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 792.000609] env[65726]: DEBUG nova.virt.hardware [None req-af1f7990-88d5-411a-be3e-a2d4ed6b5b5b tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 792.000766] env[65726]: DEBUG nova.virt.hardware [None req-af1f7990-88d5-411a-be3e-a2d4ed6b5b5b tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 792.000942] env[65726]: DEBUG nova.virt.hardware [None req-af1f7990-88d5-411a-be3e-a2d4ed6b5b5b tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 792.002096] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48aac2c8-d482-4de0-b52c-60fd467d9abd {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.010750] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115892, 'name': CreateVM_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.014108] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-726f2ad7-1ac0-47f7-a091-26e331d399e3 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.192773] env[65726]: DEBUG oslo_concurrency.lockutils [None req-1640464d-fd11-4b5f-b0e0-f0de74fa607d tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.278s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 792.194293] env[65726]: DEBUG oslo_concurrency.lockutils [None req-d63650ad-6689-479f-81d4-98dcfe6418a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 30.445s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 792.194529] env[65726]: DEBUG nova.objects.instance [None req-d63650ad-6689-479f-81d4-98dcfe6418a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Lazy-loading 'resources' on Instance uuid cd539d8b-921b-4947-954c-8573380dbdc8 {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 792.224109] env[65726]: INFO nova.scheduler.client.report [None req-1640464d-fd11-4b5f-b0e0-f0de74fa607d tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Deleted allocations for instance bd839fe1-8801-4ba1-9c23-288453258b66 [ 792.233788] env[65726]: DEBUG oslo_concurrency.lockutils [req-3db31e76-7964-4ab8-8b0f-01c60dec0dc7 req-018b92a4-404c-4c21-9f8d-308ec492a017 service nova] Releasing lock "refresh_cache-d7905c4f-1ba3-43b3-b4b3-aebacd2f9cc3" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 792.267551] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3260fa2a-3a51-48a8-9aea-39966d4e575e tempest-ServersAaction247Test-712120018 tempest-ServersAaction247Test-712120018-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 792.323747] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-2135c7ab-2ff3-4310-b64c-188bebed565e tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 608cc0ea-3f6b-4b4f-83c1-01aa50999d63] Creating linked-clone VM from snapshot {{(pid=65726) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 792.324697] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-00acd200-8e9f-449c-8a14-d517ce6da10e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.335913] env[65726]: DEBUG oslo_vmware.api [None req-2135c7ab-2ff3-4310-b64c-188bebed565e tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Waiting for the task: (returnval){ [ 792.335913] env[65726]: value = "task-5115893" [ 792.335913] env[65726]: _type = "Task" [ 792.335913] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 792.345971] env[65726]: DEBUG oslo_vmware.api [None req-2135c7ab-2ff3-4310-b64c-188bebed565e tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5115893, 'name': CloneVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.440937] env[65726]: DEBUG oslo_vmware.api [None req-d4c6362a-f5ab-42dd-9c7f-bfc3bee5f94c tempest-VolumesAssistedSnapshotsTest-1875841935 tempest-VolumesAssistedSnapshotsTest-1875841935-project-member] Task: {'id': task-5115891, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.143828} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 792.440937] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-d4c6362a-f5ab-42dd-9c7f-bfc3bee5f94c tempest-VolumesAssistedSnapshotsTest-1875841935 tempest-VolumesAssistedSnapshotsTest-1875841935-project-member] Deleted the datastore file {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 792.440937] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-d4c6362a-f5ab-42dd-9c7f-bfc3bee5f94c tempest-VolumesAssistedSnapshotsTest-1875841935 tempest-VolumesAssistedSnapshotsTest-1875841935-project-member] [instance: aa2716a9-2cf8-4e51-800d-1c8f98b2cf05] Deleted contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 792.440937] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-d4c6362a-f5ab-42dd-9c7f-bfc3bee5f94c tempest-VolumesAssistedSnapshotsTest-1875841935 tempest-VolumesAssistedSnapshotsTest-1875841935-project-member] [instance: aa2716a9-2cf8-4e51-800d-1c8f98b2cf05] Instance destroyed {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 792.440937] env[65726]: INFO nova.compute.manager [None req-d4c6362a-f5ab-42dd-9c7f-bfc3bee5f94c tempest-VolumesAssistedSnapshotsTest-1875841935 tempest-VolumesAssistedSnapshotsTest-1875841935-project-member] [instance: aa2716a9-2cf8-4e51-800d-1c8f98b2cf05] Took 1.17 seconds to destroy the instance on the hypervisor. [ 792.441636] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-d4c6362a-f5ab-42dd-9c7f-bfc3bee5f94c tempest-VolumesAssistedSnapshotsTest-1875841935 tempest-VolumesAssistedSnapshotsTest-1875841935-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 792.441636] env[65726]: DEBUG nova.compute.manager [-] [instance: aa2716a9-2cf8-4e51-800d-1c8f98b2cf05] Deallocating network for instance {{(pid=65726) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 792.441719] env[65726]: DEBUG nova.network.neutron [-] [instance: aa2716a9-2cf8-4e51-800d-1c8f98b2cf05] deallocate_for_instance() {{(pid=65726) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 792.442289] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 792.442613] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 792.474029] env[65726]: WARNING openstack [req-7e0f3c08-e637-4a7a-8686-651b076bf4cb req-c32e8064-4a17-48a0-a836-1c6ed27d6b03 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 792.474029] env[65726]: WARNING openstack [req-7e0f3c08-e637-4a7a-8686-651b076bf4cb req-c32e8064-4a17-48a0-a836-1c6ed27d6b03 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 792.487181] env[65726]: DEBUG nova.network.neutron [None req-af1f7990-88d5-411a-be3e-a2d4ed6b5b5b tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: 621b899a-e9ec-4fc0-a574-4a08cecff6d9] Successfully updated port: dcdfc13e-45af-4bc2-970d-b60727f42176 {{(pid=65726) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 792.492914] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 792.493662] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 792.504055] env[65726]: DEBUG oslo_concurrency.lockutils [None req-af1f7990-88d5-411a-be3e-a2d4ed6b5b5b tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Acquiring lock "refresh_cache-621b899a-e9ec-4fc0-a574-4a08cecff6d9" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 792.504055] env[65726]: DEBUG oslo_concurrency.lockutils [None req-af1f7990-88d5-411a-be3e-a2d4ed6b5b5b tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Acquired lock "refresh_cache-621b899a-e9ec-4fc0-a574-4a08cecff6d9" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 792.504055] env[65726]: DEBUG nova.network.neutron [None req-af1f7990-88d5-411a-be3e-a2d4ed6b5b5b tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: 621b899a-e9ec-4fc0-a574-4a08cecff6d9] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 792.515867] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115892, 'name': CreateVM_Task, 'duration_secs': 0.425079} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 792.516069] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 76249623-6f83-46a3-b8c5-c001111aa698] Created VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 792.517056] env[65726]: WARNING openstack [None req-300cacab-e765-49c9-a8a1-8e47a20dc6a2 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 792.517610] env[65726]: WARNING openstack [None req-300cacab-e765-49c9-a8a1-8e47a20dc6a2 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 792.523154] env[65726]: DEBUG oslo_concurrency.lockutils [None req-300cacab-e765-49c9-a8a1-8e47a20dc6a2 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 792.523324] env[65726]: DEBUG oslo_concurrency.lockutils [None req-300cacab-e765-49c9-a8a1-8e47a20dc6a2 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 792.523685] env[65726]: DEBUG oslo_concurrency.lockutils [None req-300cacab-e765-49c9-a8a1-8e47a20dc6a2 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 792.528724] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-83cd6136-a6ca-42b4-92d5-61c07a9fd610 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.534898] env[65726]: DEBUG oslo_vmware.api [None req-300cacab-e765-49c9-a8a1-8e47a20dc6a2 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Waiting for the task: (returnval){ [ 792.534898] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52ed3dfe-2eac-db4d-aab3-d4ebe8cdbc69" [ 792.534898] env[65726]: _type = "Task" [ 792.534898] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 792.545234] env[65726]: DEBUG oslo_vmware.api [None req-300cacab-e765-49c9-a8a1-8e47a20dc6a2 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52ed3dfe-2eac-db4d-aab3-d4ebe8cdbc69, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.613445] env[65726]: INFO nova.compute.manager [None req-1427c99f-55ff-4567-8d49-f5859451e059 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: 6d2b86c2-0a44-433f-8b3f-2e9c7f693baa] Rescuing [ 792.613821] env[65726]: DEBUG oslo_concurrency.lockutils [None req-1427c99f-55ff-4567-8d49-f5859451e059 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Acquiring lock "refresh_cache-6d2b86c2-0a44-433f-8b3f-2e9c7f693baa" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 792.613996] env[65726]: DEBUG oslo_concurrency.lockutils [None req-1427c99f-55ff-4567-8d49-f5859451e059 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Acquired lock "refresh_cache-6d2b86c2-0a44-433f-8b3f-2e9c7f693baa" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 792.614233] env[65726]: DEBUG nova.network.neutron [None req-1427c99f-55ff-4567-8d49-f5859451e059 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: 6d2b86c2-0a44-433f-8b3f-2e9c7f693baa] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 792.621430] env[65726]: WARNING openstack [req-7e0f3c08-e637-4a7a-8686-651b076bf4cb req-c32e8064-4a17-48a0-a836-1c6ed27d6b03 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 792.621706] env[65726]: WARNING openstack [req-7e0f3c08-e637-4a7a-8686-651b076bf4cb req-c32e8064-4a17-48a0-a836-1c6ed27d6b03 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 792.708065] env[65726]: WARNING openstack [req-7e0f3c08-e637-4a7a-8686-651b076bf4cb req-c32e8064-4a17-48a0-a836-1c6ed27d6b03 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 792.708472] env[65726]: WARNING openstack [req-7e0f3c08-e637-4a7a-8686-651b076bf4cb req-c32e8064-4a17-48a0-a836-1c6ed27d6b03 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 792.736239] env[65726]: DEBUG oslo_concurrency.lockutils [None req-1640464d-fd11-4b5f-b0e0-f0de74fa607d tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Lock "bd839fe1-8801-4ba1-9c23-288453258b66" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 39.128s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 792.827698] env[65726]: DEBUG nova.network.neutron [req-7e0f3c08-e637-4a7a-8686-651b076bf4cb req-c32e8064-4a17-48a0-a836-1c6ed27d6b03 service nova] [instance: 76249623-6f83-46a3-b8c5-c001111aa698] Updated VIF entry in instance network info cache for port e97a5da4-5540-4c0c-a33d-be1182e82aa7. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 792.828193] env[65726]: DEBUG nova.network.neutron [req-7e0f3c08-e637-4a7a-8686-651b076bf4cb req-c32e8064-4a17-48a0-a836-1c6ed27d6b03 service nova] [instance: 76249623-6f83-46a3-b8c5-c001111aa698] Updating instance_info_cache with network_info: [{"id": "e97a5da4-5540-4c0c-a33d-be1182e82aa7", "address": "fa:16:3e:af:b6:0f", "network": {"id": "55d629b8-a541-4720-9289-927bbc6e6f22", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.253", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "1539f06fa3534e90acbc3a60c4b8bd3f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "35342bcb-8b06-472e-b3c0-43fd3d6c4b30", "external-id": "nsx-vlan-transportzone-524", "segmentation_id": 524, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape97a5da4-55", "ovs_interfaceid": "e97a5da4-5540-4c0c-a33d-be1182e82aa7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 792.851050] env[65726]: DEBUG oslo_vmware.api [None req-2135c7ab-2ff3-4310-b64c-188bebed565e tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5115893, 'name': CloneVM_Task} progress is 94%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.868940] env[65726]: DEBUG nova.compute.manager [req-94092359-eb8a-422b-be74-4dbdcd79c3cc req-68bc4dca-1a0d-46c9-94f0-ba930ad342a3 service nova] [instance: 621b899a-e9ec-4fc0-a574-4a08cecff6d9] Received event network-vif-plugged-dcdfc13e-45af-4bc2-970d-b60727f42176 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 792.869196] env[65726]: DEBUG oslo_concurrency.lockutils [req-94092359-eb8a-422b-be74-4dbdcd79c3cc req-68bc4dca-1a0d-46c9-94f0-ba930ad342a3 service nova] Acquiring lock "621b899a-e9ec-4fc0-a574-4a08cecff6d9-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 792.869467] env[65726]: DEBUG oslo_concurrency.lockutils [req-94092359-eb8a-422b-be74-4dbdcd79c3cc req-68bc4dca-1a0d-46c9-94f0-ba930ad342a3 service nova] Lock "621b899a-e9ec-4fc0-a574-4a08cecff6d9-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 792.869563] env[65726]: DEBUG oslo_concurrency.lockutils [req-94092359-eb8a-422b-be74-4dbdcd79c3cc req-68bc4dca-1a0d-46c9-94f0-ba930ad342a3 service nova] Lock "621b899a-e9ec-4fc0-a574-4a08cecff6d9-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 792.869684] env[65726]: DEBUG nova.compute.manager [req-94092359-eb8a-422b-be74-4dbdcd79c3cc req-68bc4dca-1a0d-46c9-94f0-ba930ad342a3 service nova] [instance: 621b899a-e9ec-4fc0-a574-4a08cecff6d9] No waiting events found dispatching network-vif-plugged-dcdfc13e-45af-4bc2-970d-b60727f42176 {{(pid=65726) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 792.869843] env[65726]: WARNING nova.compute.manager [req-94092359-eb8a-422b-be74-4dbdcd79c3cc req-68bc4dca-1a0d-46c9-94f0-ba930ad342a3 service nova] [instance: 621b899a-e9ec-4fc0-a574-4a08cecff6d9] Received unexpected event network-vif-plugged-dcdfc13e-45af-4bc2-970d-b60727f42176 for instance with vm_state building and task_state spawning. [ 792.870072] env[65726]: DEBUG nova.compute.manager [req-94092359-eb8a-422b-be74-4dbdcd79c3cc req-68bc4dca-1a0d-46c9-94f0-ba930ad342a3 service nova] [instance: 621b899a-e9ec-4fc0-a574-4a08cecff6d9] Received event network-changed-dcdfc13e-45af-4bc2-970d-b60727f42176 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 792.870358] env[65726]: DEBUG nova.compute.manager [req-94092359-eb8a-422b-be74-4dbdcd79c3cc req-68bc4dca-1a0d-46c9-94f0-ba930ad342a3 service nova] [instance: 621b899a-e9ec-4fc0-a574-4a08cecff6d9] Refreshing instance network info cache due to event network-changed-dcdfc13e-45af-4bc2-970d-b60727f42176. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 792.870554] env[65726]: DEBUG oslo_concurrency.lockutils [req-94092359-eb8a-422b-be74-4dbdcd79c3cc req-68bc4dca-1a0d-46c9-94f0-ba930ad342a3 service nova] Acquiring lock "refresh_cache-621b899a-e9ec-4fc0-a574-4a08cecff6d9" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 793.011950] env[65726]: WARNING openstack [None req-af1f7990-88d5-411a-be3e-a2d4ed6b5b5b tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 793.011950] env[65726]: WARNING openstack [None req-af1f7990-88d5-411a-be3e-a2d4ed6b5b5b tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 793.017198] env[65726]: DEBUG nova.network.neutron [None req-af1f7990-88d5-411a-be3e-a2d4ed6b5b5b tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: 621b899a-e9ec-4fc0-a574-4a08cecff6d9] Instance cache missing network info. {{(pid=65726) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 793.024507] env[65726]: DEBUG nova.compute.manager [req-aa17bc71-7e24-4f3d-8d3c-515895445965 req-4b0e90d4-b822-43b3-b375-a3ae1ed1c2f2 service nova] [instance: aa2716a9-2cf8-4e51-800d-1c8f98b2cf05] Received event network-vif-deleted-c4f80985-65f1-478f-9944-afb3538efb48 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 793.024951] env[65726]: INFO nova.compute.manager [req-aa17bc71-7e24-4f3d-8d3c-515895445965 req-4b0e90d4-b822-43b3-b375-a3ae1ed1c2f2 service nova] [instance: aa2716a9-2cf8-4e51-800d-1c8f98b2cf05] Neutron deleted interface c4f80985-65f1-478f-9944-afb3538efb48; detaching it from the instance and deleting it from the info cache [ 793.024951] env[65726]: DEBUG nova.network.neutron [req-aa17bc71-7e24-4f3d-8d3c-515895445965 req-4b0e90d4-b822-43b3-b375-a3ae1ed1c2f2 service nova] [instance: aa2716a9-2cf8-4e51-800d-1c8f98b2cf05] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 793.059980] env[65726]: DEBUG oslo_vmware.api [None req-300cacab-e765-49c9-a8a1-8e47a20dc6a2 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52ed3dfe-2eac-db4d-aab3-d4ebe8cdbc69, 'name': SearchDatastore_Task, 'duration_secs': 0.012169} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 793.064165] env[65726]: DEBUG oslo_concurrency.lockutils [None req-300cacab-e765-49c9-a8a1-8e47a20dc6a2 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 793.064525] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-300cacab-e765-49c9-a8a1-8e47a20dc6a2 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] [instance: 76249623-6f83-46a3-b8c5-c001111aa698] Processing image b52362a3-ee8a-4cbf-b06f-513b0cc8f95c {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 793.064775] env[65726]: DEBUG oslo_concurrency.lockutils [None req-300cacab-e765-49c9-a8a1-8e47a20dc6a2 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 793.064917] env[65726]: DEBUG oslo_concurrency.lockutils [None req-300cacab-e765-49c9-a8a1-8e47a20dc6a2 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 793.065112] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-300cacab-e765-49c9-a8a1-8e47a20dc6a2 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 793.066090] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e9b07a92-974d-4641-94fe-dfd250af8c5f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.071430] env[65726]: WARNING openstack [None req-af1f7990-88d5-411a-be3e-a2d4ed6b5b5b tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 793.071900] env[65726]: WARNING openstack [None req-af1f7990-88d5-411a-be3e-a2d4ed6b5b5b tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 793.082841] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-300cacab-e765-49c9-a8a1-8e47a20dc6a2 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 793.082841] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-300cacab-e765-49c9-a8a1-8e47a20dc6a2 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65726) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 793.084054] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2ca52fca-d2e1-41fd-ba8b-5738ca5b6ee9 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.092684] env[65726]: DEBUG oslo_vmware.api [None req-300cacab-e765-49c9-a8a1-8e47a20dc6a2 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Waiting for the task: (returnval){ [ 793.092684] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5296c426-0715-7bc9-b24c-2f7dfdc88258" [ 793.092684] env[65726]: _type = "Task" [ 793.092684] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 793.110350] env[65726]: DEBUG oslo_vmware.api [None req-300cacab-e765-49c9-a8a1-8e47a20dc6a2 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5296c426-0715-7bc9-b24c-2f7dfdc88258, 'name': SearchDatastore_Task, 'duration_secs': 0.011587} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 793.117031] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-aa978295-c01f-4138-82a7-103759a24004 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.122914] env[65726]: WARNING openstack [None req-1427c99f-55ff-4567-8d49-f5859451e059 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 793.126645] env[65726]: WARNING openstack [None req-1427c99f-55ff-4567-8d49-f5859451e059 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 793.136503] env[65726]: DEBUG oslo_vmware.api [None req-300cacab-e765-49c9-a8a1-8e47a20dc6a2 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Waiting for the task: (returnval){ [ 793.136503] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]529b522b-a0a0-f891-fb93-cd84b5ccb2ae" [ 793.136503] env[65726]: _type = "Task" [ 793.136503] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 793.147502] env[65726]: DEBUG oslo_vmware.api [None req-300cacab-e765-49c9-a8a1-8e47a20dc6a2 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]529b522b-a0a0-f891-fb93-cd84b5ccb2ae, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.174691] env[65726]: WARNING openstack [None req-af1f7990-88d5-411a-be3e-a2d4ed6b5b5b tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 793.175256] env[65726]: WARNING openstack [None req-af1f7990-88d5-411a-be3e-a2d4ed6b5b5b tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 793.281755] env[65726]: DEBUG nova.network.neutron [None req-af1f7990-88d5-411a-be3e-a2d4ed6b5b5b tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: 621b899a-e9ec-4fc0-a574-4a08cecff6d9] Updating instance_info_cache with network_info: [{"id": "dcdfc13e-45af-4bc2-970d-b60727f42176", "address": "fa:16:3e:7a:7e:fb", "network": {"id": "4c972d9d-db3b-4779-ac39-39d9a2d141dd", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1418913861-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cbb17a09b35c4c22ade5c4082b10bda9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "50cf0a70-948d-4611-af05-94c1483064ed", "external-id": "nsx-vlan-transportzone-536", "segmentation_id": 536, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdcdfc13e-45", "ovs_interfaceid": "dcdfc13e-45af-4bc2-970d-b60727f42176", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 793.292977] env[65726]: WARNING openstack [None req-1427c99f-55ff-4567-8d49-f5859451e059 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 793.293398] env[65726]: WARNING openstack [None req-1427c99f-55ff-4567-8d49-f5859451e059 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 793.331697] env[65726]: DEBUG oslo_concurrency.lockutils [req-7e0f3c08-e637-4a7a-8686-651b076bf4cb req-c32e8064-4a17-48a0-a836-1c6ed27d6b03 service nova] Releasing lock "refresh_cache-76249623-6f83-46a3-b8c5-c001111aa698" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 793.334749] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9aa8744-e345-4d23-b9a4-2ea425248c02 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.358176] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9d9d7a6-8ff9-4de5-a4c9-0ed2194436e8 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.361927] env[65726]: DEBUG oslo_vmware.api [None req-2135c7ab-2ff3-4310-b64c-188bebed565e tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5115893, 'name': CloneVM_Task} progress is 94%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.416987] env[65726]: WARNING openstack [None req-1427c99f-55ff-4567-8d49-f5859451e059 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 793.416987] env[65726]: WARNING openstack [None req-1427c99f-55ff-4567-8d49-f5859451e059 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 793.431470] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6bd99528-b9b4-47ad-ae09-70c17a3900de {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.435896] env[65726]: DEBUG nova.network.neutron [-] [instance: aa2716a9-2cf8-4e51-800d-1c8f98b2cf05] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 793.447259] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e22847d3-6636-417d-8884-bc2a82bdb295 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.477802] env[65726]: DEBUG nova.compute.provider_tree [None req-d63650ad-6689-479f-81d4-98dcfe6418a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 793.528671] env[65726]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d6fee938-a21e-4754-9b83-dd9daf81d02d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.549173] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88f3717c-26e8-483d-a7a4-d912cc076dd5 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.573326] env[65726]: DEBUG nova.network.neutron [None req-1427c99f-55ff-4567-8d49-f5859451e059 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: 6d2b86c2-0a44-433f-8b3f-2e9c7f693baa] Updating instance_info_cache with network_info: [{"id": "b487d53a-9bda-4e3c-be61-f26acbc8ae0f", "address": "fa:16:3e:ad:db:36", "network": {"id": "9d5ad167-99fb-4208-ab30-7053f35774e4", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-2082516762-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a8362295a62f4b51bae719a7ef5a4656", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dabbac20-1723-40ad-9da0-e53b28073651", "external-id": "nsx-vlan-transportzone-790", "segmentation_id": 790, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb487d53a-9b", "ovs_interfaceid": "b487d53a-9bda-4e3c-be61-f26acbc8ae0f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 793.608047] env[65726]: DEBUG nova.compute.manager [req-aa17bc71-7e24-4f3d-8d3c-515895445965 req-4b0e90d4-b822-43b3-b375-a3ae1ed1c2f2 service nova] [instance: aa2716a9-2cf8-4e51-800d-1c8f98b2cf05] Detach interface failed, port_id=c4f80985-65f1-478f-9944-afb3538efb48, reason: Instance aa2716a9-2cf8-4e51-800d-1c8f98b2cf05 could not be found. {{(pid=65726) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11669}} [ 793.647988] env[65726]: DEBUG oslo_vmware.api [None req-300cacab-e765-49c9-a8a1-8e47a20dc6a2 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]529b522b-a0a0-f891-fb93-cd84b5ccb2ae, 'name': SearchDatastore_Task, 'duration_secs': 0.01883} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 793.648364] env[65726]: DEBUG oslo_concurrency.lockutils [None req-300cacab-e765-49c9-a8a1-8e47a20dc6a2 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 793.648685] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-300cacab-e765-49c9-a8a1-8e47a20dc6a2 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] 76249623-6f83-46a3-b8c5-c001111aa698/76249623-6f83-46a3-b8c5-c001111aa698.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 793.649011] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d3ec70ed-277b-465f-a2a4-33954acc0c5e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.657485] env[65726]: DEBUG oslo_vmware.api [None req-300cacab-e765-49c9-a8a1-8e47a20dc6a2 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Waiting for the task: (returnval){ [ 793.657485] env[65726]: value = "task-5115894" [ 793.657485] env[65726]: _type = "Task" [ 793.657485] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 793.666815] env[65726]: DEBUG oslo_vmware.api [None req-300cacab-e765-49c9-a8a1-8e47a20dc6a2 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Task: {'id': task-5115894, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.786834] env[65726]: DEBUG oslo_concurrency.lockutils [None req-af1f7990-88d5-411a-be3e-a2d4ed6b5b5b tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Releasing lock "refresh_cache-621b899a-e9ec-4fc0-a574-4a08cecff6d9" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 793.786834] env[65726]: DEBUG nova.compute.manager [None req-af1f7990-88d5-411a-be3e-a2d4ed6b5b5b tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: 621b899a-e9ec-4fc0-a574-4a08cecff6d9] Instance network_info: |[{"id": "dcdfc13e-45af-4bc2-970d-b60727f42176", "address": "fa:16:3e:7a:7e:fb", "network": {"id": "4c972d9d-db3b-4779-ac39-39d9a2d141dd", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1418913861-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cbb17a09b35c4c22ade5c4082b10bda9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "50cf0a70-948d-4611-af05-94c1483064ed", "external-id": "nsx-vlan-transportzone-536", "segmentation_id": 536, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdcdfc13e-45", "ovs_interfaceid": "dcdfc13e-45af-4bc2-970d-b60727f42176", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 793.787087] env[65726]: DEBUG oslo_concurrency.lockutils [req-94092359-eb8a-422b-be74-4dbdcd79c3cc req-68bc4dca-1a0d-46c9-94f0-ba930ad342a3 service nova] Acquired lock "refresh_cache-621b899a-e9ec-4fc0-a574-4a08cecff6d9" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 793.787087] env[65726]: DEBUG nova.network.neutron [req-94092359-eb8a-422b-be74-4dbdcd79c3cc req-68bc4dca-1a0d-46c9-94f0-ba930ad342a3 service nova] [instance: 621b899a-e9ec-4fc0-a574-4a08cecff6d9] Refreshing network info cache for port dcdfc13e-45af-4bc2-970d-b60727f42176 {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 793.788077] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-af1f7990-88d5-411a-be3e-a2d4ed6b5b5b tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: 621b899a-e9ec-4fc0-a574-4a08cecff6d9] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:7a:7e:fb', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '50cf0a70-948d-4611-af05-94c1483064ed', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'dcdfc13e-45af-4bc2-970d-b60727f42176', 'vif_model': 'vmxnet3'}] {{(pid=65726) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 793.795957] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-af1f7990-88d5-411a-be3e-a2d4ed6b5b5b tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Creating folder: Project (cbb17a09b35c4c22ade5c4082b10bda9). Parent ref: group-v995008. {{(pid=65726) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 793.797850] env[65726]: WARNING openstack [req-94092359-eb8a-422b-be74-4dbdcd79c3cc req-68bc4dca-1a0d-46c9-94f0-ba930ad342a3 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 793.798361] env[65726]: WARNING openstack [req-94092359-eb8a-422b-be74-4dbdcd79c3cc req-68bc4dca-1a0d-46c9-94f0-ba930ad342a3 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 793.805123] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-bba0ff3d-7d76-44e0-a31c-70552c9a3e0f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.819845] env[65726]: INFO nova.virt.vmwareapi.vm_util [None req-af1f7990-88d5-411a-be3e-a2d4ed6b5b5b tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Created folder: Project (cbb17a09b35c4c22ade5c4082b10bda9) in parent group-v995008. [ 793.820155] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-af1f7990-88d5-411a-be3e-a2d4ed6b5b5b tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Creating folder: Instances. Parent ref: group-v995136. {{(pid=65726) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 793.820618] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a469c61e-cbd2-46bd-976f-b55dbd14dea9 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.832985] env[65726]: INFO nova.virt.vmwareapi.vm_util [None req-af1f7990-88d5-411a-be3e-a2d4ed6b5b5b tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Created folder: Instances in parent group-v995136. [ 793.833402] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-af1f7990-88d5-411a-be3e-a2d4ed6b5b5b tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 793.833574] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 621b899a-e9ec-4fc0-a574-4a08cecff6d9] Creating VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 793.833851] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-0c5e2958-2336-478d-bf11-1a16c2bc2999 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.871415] env[65726]: DEBUG oslo_vmware.api [None req-2135c7ab-2ff3-4310-b64c-188bebed565e tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5115893, 'name': CloneVM_Task} progress is 95%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.873602] env[65726]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 793.873602] env[65726]: value = "task-5115897" [ 793.873602] env[65726]: _type = "Task" [ 793.873602] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 793.884093] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115897, 'name': CreateVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.929754] env[65726]: WARNING openstack [req-94092359-eb8a-422b-be74-4dbdcd79c3cc req-68bc4dca-1a0d-46c9-94f0-ba930ad342a3 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 793.930345] env[65726]: WARNING openstack [req-94092359-eb8a-422b-be74-4dbdcd79c3cc req-68bc4dca-1a0d-46c9-94f0-ba930ad342a3 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 793.938893] env[65726]: INFO nova.compute.manager [-] [instance: aa2716a9-2cf8-4e51-800d-1c8f98b2cf05] Took 1.50 seconds to deallocate network for instance. [ 793.992400] env[65726]: DEBUG nova.scheduler.client.report [None req-d63650ad-6689-479f-81d4-98dcfe6418a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 794.010542] env[65726]: WARNING openstack [req-94092359-eb8a-422b-be74-4dbdcd79c3cc req-68bc4dca-1a0d-46c9-94f0-ba930ad342a3 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 794.011068] env[65726]: WARNING openstack [req-94092359-eb8a-422b-be74-4dbdcd79c3cc req-68bc4dca-1a0d-46c9-94f0-ba930ad342a3 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 794.077826] env[65726]: DEBUG oslo_concurrency.lockutils [None req-1427c99f-55ff-4567-8d49-f5859451e059 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Releasing lock "refresh_cache-6d2b86c2-0a44-433f-8b3f-2e9c7f693baa" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 794.107182] env[65726]: DEBUG nova.network.neutron [req-94092359-eb8a-422b-be74-4dbdcd79c3cc req-68bc4dca-1a0d-46c9-94f0-ba930ad342a3 service nova] [instance: 621b899a-e9ec-4fc0-a574-4a08cecff6d9] Updated VIF entry in instance network info cache for port dcdfc13e-45af-4bc2-970d-b60727f42176. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 794.107648] env[65726]: DEBUG nova.network.neutron [req-94092359-eb8a-422b-be74-4dbdcd79c3cc req-68bc4dca-1a0d-46c9-94f0-ba930ad342a3 service nova] [instance: 621b899a-e9ec-4fc0-a574-4a08cecff6d9] Updating instance_info_cache with network_info: [{"id": "dcdfc13e-45af-4bc2-970d-b60727f42176", "address": "fa:16:3e:7a:7e:fb", "network": {"id": "4c972d9d-db3b-4779-ac39-39d9a2d141dd", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1418913861-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cbb17a09b35c4c22ade5c4082b10bda9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "50cf0a70-948d-4611-af05-94c1483064ed", "external-id": "nsx-vlan-transportzone-536", "segmentation_id": 536, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdcdfc13e-45", "ovs_interfaceid": "dcdfc13e-45af-4bc2-970d-b60727f42176", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 794.171618] env[65726]: DEBUG oslo_vmware.api [None req-300cacab-e765-49c9-a8a1-8e47a20dc6a2 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Task: {'id': task-5115894, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.369039] env[65726]: DEBUG oslo_vmware.api [None req-2135c7ab-2ff3-4310-b64c-188bebed565e tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5115893, 'name': CloneVM_Task, 'duration_secs': 1.735645} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 794.369299] env[65726]: INFO nova.virt.vmwareapi.vmops [None req-2135c7ab-2ff3-4310-b64c-188bebed565e tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 608cc0ea-3f6b-4b4f-83c1-01aa50999d63] Created linked-clone VM from snapshot [ 794.370251] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b284186-b167-4f1e-ac34-06ec9f528857 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.381714] env[65726]: DEBUG nova.virt.vmwareapi.images [None req-2135c7ab-2ff3-4310-b64c-188bebed565e tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 608cc0ea-3f6b-4b4f-83c1-01aa50999d63] Uploading image 9fc0cda5-a631-4766-b141-daaef80ccc69 {{(pid=65726) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 794.389663] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115897, 'name': CreateVM_Task} progress is 99%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.410053] env[65726]: DEBUG oslo_vmware.rw_handles [None req-2135c7ab-2ff3-4310-b64c-188bebed565e tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 794.410053] env[65726]: value = "vm-995135" [ 794.410053] env[65726]: _type = "VirtualMachine" [ 794.410053] env[65726]: }. {{(pid=65726) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 794.410383] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-f41d9c61-41b4-4d52-820c-b39d78b415b8 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.420038] env[65726]: DEBUG oslo_vmware.rw_handles [None req-2135c7ab-2ff3-4310-b64c-188bebed565e tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Lease: (returnval){ [ 794.420038] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52a894b2-92f5-2be5-df6b-62efd9d73678" [ 794.420038] env[65726]: _type = "HttpNfcLease" [ 794.420038] env[65726]: } obtained for exporting VM: (result){ [ 794.420038] env[65726]: value = "vm-995135" [ 794.420038] env[65726]: _type = "VirtualMachine" [ 794.420038] env[65726]: }. {{(pid=65726) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 794.420334] env[65726]: DEBUG oslo_vmware.api [None req-2135c7ab-2ff3-4310-b64c-188bebed565e tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Waiting for the lease: (returnval){ [ 794.420334] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52a894b2-92f5-2be5-df6b-62efd9d73678" [ 794.420334] env[65726]: _type = "HttpNfcLease" [ 794.420334] env[65726]: } to be ready. {{(pid=65726) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 794.426848] env[65726]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 794.426848] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52a894b2-92f5-2be5-df6b-62efd9d73678" [ 794.426848] env[65726]: _type = "HttpNfcLease" [ 794.426848] env[65726]: } is initializing. {{(pid=65726) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 794.447583] env[65726]: DEBUG oslo_concurrency.lockutils [None req-d4c6362a-f5ab-42dd-9c7f-bfc3bee5f94c tempest-VolumesAssistedSnapshotsTest-1875841935 tempest-VolumesAssistedSnapshotsTest-1875841935-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 794.498947] env[65726]: DEBUG oslo_concurrency.lockutils [None req-d63650ad-6689-479f-81d4-98dcfe6418a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.305s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 794.501829] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c9cc880b-42e7-4875-9b8e-55f0109d1664 tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 32.449s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 794.502059] env[65726]: DEBUG nova.objects.instance [None req-c9cc880b-42e7-4875-9b8e-55f0109d1664 tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Lazy-loading 'resources' on Instance uuid ca67d1a4-d304-4d06-a436-f257b60ca080 {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 794.524414] env[65726]: INFO nova.scheduler.client.report [None req-d63650ad-6689-479f-81d4-98dcfe6418a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Deleted allocations for instance cd539d8b-921b-4947-954c-8573380dbdc8 [ 794.614643] env[65726]: DEBUG oslo_concurrency.lockutils [req-94092359-eb8a-422b-be74-4dbdcd79c3cc req-68bc4dca-1a0d-46c9-94f0-ba930ad342a3 service nova] Releasing lock "refresh_cache-621b899a-e9ec-4fc0-a574-4a08cecff6d9" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 794.669643] env[65726]: DEBUG oslo_vmware.api [None req-300cacab-e765-49c9-a8a1-8e47a20dc6a2 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Task: {'id': task-5115894, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.516497} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 794.669737] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-300cacab-e765-49c9-a8a1-8e47a20dc6a2 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] 76249623-6f83-46a3-b8c5-c001111aa698/76249623-6f83-46a3-b8c5-c001111aa698.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 794.669885] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-300cacab-e765-49c9-a8a1-8e47a20dc6a2 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] [instance: 76249623-6f83-46a3-b8c5-c001111aa698] Extending root virtual disk to 1048576 {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 794.670199] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0470605c-2c6c-4ec4-829a-e84005682c20 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.677429] env[65726]: DEBUG oslo_vmware.api [None req-300cacab-e765-49c9-a8a1-8e47a20dc6a2 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Waiting for the task: (returnval){ [ 794.677429] env[65726]: value = "task-5115899" [ 794.677429] env[65726]: _type = "Task" [ 794.677429] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 794.688017] env[65726]: DEBUG oslo_vmware.api [None req-300cacab-e765-49c9-a8a1-8e47a20dc6a2 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Task: {'id': task-5115899, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.887200] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115897, 'name': CreateVM_Task, 'duration_secs': 0.533324} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 794.887384] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 621b899a-e9ec-4fc0-a574-4a08cecff6d9] Created VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 794.888216] env[65726]: WARNING openstack [None req-af1f7990-88d5-411a-be3e-a2d4ed6b5b5b tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 794.888582] env[65726]: WARNING openstack [None req-af1f7990-88d5-411a-be3e-a2d4ed6b5b5b tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 794.893638] env[65726]: DEBUG oslo_concurrency.lockutils [None req-af1f7990-88d5-411a-be3e-a2d4ed6b5b5b tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 794.893797] env[65726]: DEBUG oslo_concurrency.lockutils [None req-af1f7990-88d5-411a-be3e-a2d4ed6b5b5b tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 794.894252] env[65726]: DEBUG oslo_concurrency.lockutils [None req-af1f7990-88d5-411a-be3e-a2d4ed6b5b5b tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 794.894591] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-49fb2122-ef27-427e-8e1c-6dc885db9e77 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.901031] env[65726]: DEBUG oslo_vmware.api [None req-af1f7990-88d5-411a-be3e-a2d4ed6b5b5b tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Waiting for the task: (returnval){ [ 794.901031] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52f74575-9db6-bfed-789a-6c025a3415f4" [ 794.901031] env[65726]: _type = "Task" [ 794.901031] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 794.909839] env[65726]: DEBUG oslo_vmware.api [None req-af1f7990-88d5-411a-be3e-a2d4ed6b5b5b tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52f74575-9db6-bfed-789a-6c025a3415f4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.928988] env[65726]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 794.928988] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52a894b2-92f5-2be5-df6b-62efd9d73678" [ 794.928988] env[65726]: _type = "HttpNfcLease" [ 794.928988] env[65726]: } is ready. {{(pid=65726) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 794.928988] env[65726]: DEBUG oslo_vmware.rw_handles [None req-2135c7ab-2ff3-4310-b64c-188bebed565e tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 794.928988] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52a894b2-92f5-2be5-df6b-62efd9d73678" [ 794.928988] env[65726]: _type = "HttpNfcLease" [ 794.928988] env[65726]: }. {{(pid=65726) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 794.929784] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22569fe5-0b15-403f-bcf7-dd9196085b0e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.937879] env[65726]: DEBUG oslo_vmware.rw_handles [None req-2135c7ab-2ff3-4310-b64c-188bebed565e tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/523612c5-2df1-5f7a-27a4-fa9c623a4605/disk-0.vmdk from lease info. {{(pid=65726) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 794.938199] env[65726]: DEBUG oslo_vmware.rw_handles [None req-2135c7ab-2ff3-4310-b64c-188bebed565e tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/523612c5-2df1-5f7a-27a4-fa9c623a4605/disk-0.vmdk for reading. {{(pid=65726) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 795.035761] env[65726]: DEBUG oslo_concurrency.lockutils [None req-d63650ad-6689-479f-81d4-98dcfe6418a3 tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Lock "cd539d8b-921b-4947-954c-8573380dbdc8" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 37.511s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 795.107445] env[65726]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-9eaaf29a-f0e3-4169-a4ff-6e9a8d2c53c6 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.189175] env[65726]: DEBUG oslo_vmware.api [None req-300cacab-e765-49c9-a8a1-8e47a20dc6a2 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Task: {'id': task-5115899, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.073331} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 795.192444] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-300cacab-e765-49c9-a8a1-8e47a20dc6a2 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] [instance: 76249623-6f83-46a3-b8c5-c001111aa698] Extended root virtual disk {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 795.193718] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f7628de-da15-49e9-aa09-a72f07191777 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.224428] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-300cacab-e765-49c9-a8a1-8e47a20dc6a2 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] [instance: 76249623-6f83-46a3-b8c5-c001111aa698] Reconfiguring VM instance instance-0000002e to attach disk [datastore1] 76249623-6f83-46a3-b8c5-c001111aa698/76249623-6f83-46a3-b8c5-c001111aa698.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 795.227245] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b11abd56-1cf7-4598-9ed8-e8f0013a6519 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.251609] env[65726]: DEBUG oslo_vmware.api [None req-300cacab-e765-49c9-a8a1-8e47a20dc6a2 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Waiting for the task: (returnval){ [ 795.251609] env[65726]: value = "task-5115900" [ 795.251609] env[65726]: _type = "Task" [ 795.251609] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 795.267531] env[65726]: DEBUG oslo_vmware.api [None req-300cacab-e765-49c9-a8a1-8e47a20dc6a2 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Task: {'id': task-5115900, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.424072] env[65726]: DEBUG oslo_vmware.api [None req-af1f7990-88d5-411a-be3e-a2d4ed6b5b5b tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52f74575-9db6-bfed-789a-6c025a3415f4, 'name': SearchDatastore_Task, 'duration_secs': 0.011093} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 795.425316] env[65726]: DEBUG oslo_concurrency.lockutils [None req-af1f7990-88d5-411a-be3e-a2d4ed6b5b5b tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 795.425817] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-af1f7990-88d5-411a-be3e-a2d4ed6b5b5b tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: 621b899a-e9ec-4fc0-a574-4a08cecff6d9] Processing image b52362a3-ee8a-4cbf-b06f-513b0cc8f95c {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 795.426283] env[65726]: DEBUG oslo_concurrency.lockutils [None req-af1f7990-88d5-411a-be3e-a2d4ed6b5b5b tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 795.426425] env[65726]: DEBUG oslo_concurrency.lockutils [None req-af1f7990-88d5-411a-be3e-a2d4ed6b5b5b tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 795.426711] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-af1f7990-88d5-411a-be3e-a2d4ed6b5b5b tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 795.430146] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4c2107cd-9336-46ef-83fa-0582e6e7282b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.445141] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-af1f7990-88d5-411a-be3e-a2d4ed6b5b5b tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 795.445141] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-af1f7990-88d5-411a-be3e-a2d4ed6b5b5b tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65726) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 795.445141] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ba162a77-d0bc-4ae7-82f8-fe0046536c52 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.461645] env[65726]: DEBUG oslo_vmware.api [None req-af1f7990-88d5-411a-be3e-a2d4ed6b5b5b tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Waiting for the task: (returnval){ [ 795.461645] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52cd14ff-7653-464c-8bf8-10c62c0b15ea" [ 795.461645] env[65726]: _type = "Task" [ 795.461645] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 795.473566] env[65726]: DEBUG oslo_vmware.api [None req-af1f7990-88d5-411a-be3e-a2d4ed6b5b5b tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52cd14ff-7653-464c-8bf8-10c62c0b15ea, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.593601] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9626b235-42a0-481c-9034-0b075a8d2a0f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.603047] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8ed3617-26af-4b5b-8db9-3457a6a0c5b5 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.635474] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-1427c99f-55ff-4567-8d49-f5859451e059 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: 6d2b86c2-0a44-433f-8b3f-2e9c7f693baa] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 795.635981] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7e89bc63-fcf2-4223-8341-c5ea24ea0eec {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.638355] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5796c9a-febb-4a45-83c2-ab710d91693f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.649255] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a0d9306-0fa8-4f5b-b497-0b60122c464e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.656397] env[65726]: DEBUG oslo_vmware.api [None req-1427c99f-55ff-4567-8d49-f5859451e059 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Waiting for the task: (returnval){ [ 795.656397] env[65726]: value = "task-5115901" [ 795.656397] env[65726]: _type = "Task" [ 795.656397] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 795.674945] env[65726]: DEBUG nova.compute.provider_tree [None req-c9cc880b-42e7-4875-9b8e-55f0109d1664 tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 795.683439] env[65726]: DEBUG oslo_vmware.api [None req-1427c99f-55ff-4567-8d49-f5859451e059 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Task: {'id': task-5115901, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.767143] env[65726]: DEBUG oslo_vmware.api [None req-300cacab-e765-49c9-a8a1-8e47a20dc6a2 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Task: {'id': task-5115900, 'name': ReconfigVM_Task} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 795.767971] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-300cacab-e765-49c9-a8a1-8e47a20dc6a2 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] [instance: 76249623-6f83-46a3-b8c5-c001111aa698] Reconfigured VM instance instance-0000002e to attach disk [datastore1] 76249623-6f83-46a3-b8c5-c001111aa698/76249623-6f83-46a3-b8c5-c001111aa698.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 795.768730] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-aa6ab892-b666-4c07-aee0-e0c509127a7f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.777614] env[65726]: DEBUG oslo_vmware.api [None req-300cacab-e765-49c9-a8a1-8e47a20dc6a2 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Waiting for the task: (returnval){ [ 795.777614] env[65726]: value = "task-5115902" [ 795.777614] env[65726]: _type = "Task" [ 795.777614] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 795.787305] env[65726]: DEBUG oslo_vmware.api [None req-300cacab-e765-49c9-a8a1-8e47a20dc6a2 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Task: {'id': task-5115902, 'name': Rename_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.946174] env[65726]: DEBUG oslo_concurrency.lockutils [None req-debf7b46-904e-41a8-8c0e-839a5a6575cb tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Acquiring lock "703fff2d-5aec-4ed1-b987-9e1cc3a25f67" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 795.946670] env[65726]: DEBUG oslo_concurrency.lockutils [None req-debf7b46-904e-41a8-8c0e-839a5a6575cb tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Lock "703fff2d-5aec-4ed1-b987-9e1cc3a25f67" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 795.947040] env[65726]: DEBUG oslo_concurrency.lockutils [None req-debf7b46-904e-41a8-8c0e-839a5a6575cb tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Acquiring lock "703fff2d-5aec-4ed1-b987-9e1cc3a25f67-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 795.947321] env[65726]: DEBUG oslo_concurrency.lockutils [None req-debf7b46-904e-41a8-8c0e-839a5a6575cb tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Lock "703fff2d-5aec-4ed1-b987-9e1cc3a25f67-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 795.947862] env[65726]: DEBUG oslo_concurrency.lockutils [None req-debf7b46-904e-41a8-8c0e-839a5a6575cb tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Lock "703fff2d-5aec-4ed1-b987-9e1cc3a25f67-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 795.951945] env[65726]: INFO nova.compute.manager [None req-debf7b46-904e-41a8-8c0e-839a5a6575cb tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] [instance: 703fff2d-5aec-4ed1-b987-9e1cc3a25f67] Terminating instance [ 795.975577] env[65726]: DEBUG oslo_vmware.api [None req-af1f7990-88d5-411a-be3e-a2d4ed6b5b5b tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52cd14ff-7653-464c-8bf8-10c62c0b15ea, 'name': SearchDatastore_Task, 'duration_secs': 0.012282} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 795.976453] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-71243377-c020-484c-9a67-bcad771c2805 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.983091] env[65726]: DEBUG oslo_vmware.api [None req-af1f7990-88d5-411a-be3e-a2d4ed6b5b5b tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Waiting for the task: (returnval){ [ 795.983091] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]520d9031-ef7f-3d2e-8ed9-e6213e23f095" [ 795.983091] env[65726]: _type = "Task" [ 795.983091] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 795.992777] env[65726]: DEBUG oslo_vmware.api [None req-af1f7990-88d5-411a-be3e-a2d4ed6b5b5b tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]520d9031-ef7f-3d2e-8ed9-e6213e23f095, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.167909] env[65726]: DEBUG oslo_vmware.api [None req-1427c99f-55ff-4567-8d49-f5859451e059 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Task: {'id': task-5115901, 'name': PowerOffVM_Task, 'duration_secs': 0.237209} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 796.169110] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-1427c99f-55ff-4567-8d49-f5859451e059 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: 6d2b86c2-0a44-433f-8b3f-2e9c7f693baa] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 796.169651] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad1fd404-805d-4050-9ac4-d020efcaecbb {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.190939] env[65726]: DEBUG nova.scheduler.client.report [None req-c9cc880b-42e7-4875-9b8e-55f0109d1664 tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 796.195195] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7734d513-0e22-4249-a22d-06de0e3d122e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.234403] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-1427c99f-55ff-4567-8d49-f5859451e059 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: 6d2b86c2-0a44-433f-8b3f-2e9c7f693baa] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 796.234868] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c999efd1-7dd8-4212-b531-f3b1c44a66b2 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.245070] env[65726]: DEBUG oslo_vmware.api [None req-1427c99f-55ff-4567-8d49-f5859451e059 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Waiting for the task: (returnval){ [ 796.245070] env[65726]: value = "task-5115903" [ 796.245070] env[65726]: _type = "Task" [ 796.245070] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 796.257585] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-1427c99f-55ff-4567-8d49-f5859451e059 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: 6d2b86c2-0a44-433f-8b3f-2e9c7f693baa] VM already powered off {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 796.257872] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-1427c99f-55ff-4567-8d49-f5859451e059 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: 6d2b86c2-0a44-433f-8b3f-2e9c7f693baa] Processing image b52362a3-ee8a-4cbf-b06f-513b0cc8f95c {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 796.258149] env[65726]: DEBUG oslo_concurrency.lockutils [None req-1427c99f-55ff-4567-8d49-f5859451e059 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 796.258396] env[65726]: DEBUG oslo_concurrency.lockutils [None req-1427c99f-55ff-4567-8d49-f5859451e059 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 796.258585] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-1427c99f-55ff-4567-8d49-f5859451e059 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 796.258995] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c1bf6258-09bd-470d-88ca-4e4f36a56987 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.269172] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-1427c99f-55ff-4567-8d49-f5859451e059 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 796.269432] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-1427c99f-55ff-4567-8d49-f5859451e059 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65726) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 796.270259] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-aebf2b85-b2d1-4d81-9ff7-33edab315176 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.277576] env[65726]: DEBUG oslo_vmware.api [None req-1427c99f-55ff-4567-8d49-f5859451e059 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Waiting for the task: (returnval){ [ 796.277576] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52c9ba62-c2a7-a34b-7b90-6c0a7e37504f" [ 796.277576] env[65726]: _type = "Task" [ 796.277576] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 796.299698] env[65726]: DEBUG oslo_vmware.api [None req-1427c99f-55ff-4567-8d49-f5859451e059 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52c9ba62-c2a7-a34b-7b90-6c0a7e37504f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.300196] env[65726]: DEBUG oslo_vmware.api [None req-300cacab-e765-49c9-a8a1-8e47a20dc6a2 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Task: {'id': task-5115902, 'name': Rename_Task, 'duration_secs': 0.234207} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 796.300668] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-300cacab-e765-49c9-a8a1-8e47a20dc6a2 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] [instance: 76249623-6f83-46a3-b8c5-c001111aa698] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 796.301155] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-92475fed-a18f-4f0a-b89c-ce29a9dd672f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.309419] env[65726]: DEBUG oslo_vmware.api [None req-300cacab-e765-49c9-a8a1-8e47a20dc6a2 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Waiting for the task: (returnval){ [ 796.309419] env[65726]: value = "task-5115904" [ 796.309419] env[65726]: _type = "Task" [ 796.309419] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 796.317781] env[65726]: DEBUG oslo_vmware.api [None req-300cacab-e765-49c9-a8a1-8e47a20dc6a2 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Task: {'id': task-5115904, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.455801] env[65726]: DEBUG nova.compute.manager [None req-debf7b46-904e-41a8-8c0e-839a5a6575cb tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] [instance: 703fff2d-5aec-4ed1-b987-9e1cc3a25f67] Start destroying the instance on the hypervisor. {{(pid=65726) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 796.456082] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-debf7b46-904e-41a8-8c0e-839a5a6575cb tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] [instance: 703fff2d-5aec-4ed1-b987-9e1cc3a25f67] Destroying instance {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 796.457075] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-986f4d55-3bc7-410e-bf0c-1817cc219859 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.465927] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-debf7b46-904e-41a8-8c0e-839a5a6575cb tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] [instance: 703fff2d-5aec-4ed1-b987-9e1cc3a25f67] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 796.466485] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c6e888c4-ca63-428e-9d4d-8127bfc17050 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.473938] env[65726]: DEBUG oslo_vmware.api [None req-debf7b46-904e-41a8-8c0e-839a5a6575cb tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Waiting for the task: (returnval){ [ 796.473938] env[65726]: value = "task-5115905" [ 796.473938] env[65726]: _type = "Task" [ 796.473938] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 796.484200] env[65726]: DEBUG oslo_vmware.api [None req-debf7b46-904e-41a8-8c0e-839a5a6575cb tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Task: {'id': task-5115905, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.494284] env[65726]: DEBUG oslo_vmware.api [None req-af1f7990-88d5-411a-be3e-a2d4ed6b5b5b tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]520d9031-ef7f-3d2e-8ed9-e6213e23f095, 'name': SearchDatastore_Task, 'duration_secs': 0.010393} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 796.495063] env[65726]: DEBUG oslo_concurrency.lockutils [None req-af1f7990-88d5-411a-be3e-a2d4ed6b5b5b tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 796.495063] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-af1f7990-88d5-411a-be3e-a2d4ed6b5b5b tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] 621b899a-e9ec-4fc0-a574-4a08cecff6d9/621b899a-e9ec-4fc0-a574-4a08cecff6d9.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 796.495492] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-824d43ac-bbc3-43c6-8f7f-5f740fa3e813 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.503550] env[65726]: DEBUG oslo_vmware.api [None req-af1f7990-88d5-411a-be3e-a2d4ed6b5b5b tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Waiting for the task: (returnval){ [ 796.503550] env[65726]: value = "task-5115906" [ 796.503550] env[65726]: _type = "Task" [ 796.503550] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 796.514407] env[65726]: DEBUG oslo_vmware.api [None req-af1f7990-88d5-411a-be3e-a2d4ed6b5b5b tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': task-5115906, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.699649] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c9cc880b-42e7-4875-9b8e-55f0109d1664 tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.198s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 796.703046] env[65726]: DEBUG oslo_concurrency.lockutils [None req-04d3888f-ac16-4533-b7db-d3c669b17255 tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 34.376s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 796.704791] env[65726]: INFO nova.compute.claims [None req-04d3888f-ac16-4533-b7db-d3c669b17255 tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] [instance: bc2c12e0-0d06-432f-b42f-be468e3b6ee4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 796.730295] env[65726]: INFO nova.scheduler.client.report [None req-c9cc880b-42e7-4875-9b8e-55f0109d1664 tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Deleted allocations for instance ca67d1a4-d304-4d06-a436-f257b60ca080 [ 796.793117] env[65726]: DEBUG oslo_vmware.api [None req-1427c99f-55ff-4567-8d49-f5859451e059 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52c9ba62-c2a7-a34b-7b90-6c0a7e37504f, 'name': SearchDatastore_Task, 'duration_secs': 0.016843} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 796.794628] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c9937055-e20c-46b2-bcc4-179973ef5949 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.805927] env[65726]: DEBUG oslo_vmware.api [None req-1427c99f-55ff-4567-8d49-f5859451e059 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Waiting for the task: (returnval){ [ 796.805927] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]520e3b2f-6567-7f0f-cb03-fd98a8d0b9f0" [ 796.805927] env[65726]: _type = "Task" [ 796.805927] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 796.819093] env[65726]: DEBUG oslo_vmware.api [None req-1427c99f-55ff-4567-8d49-f5859451e059 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]520e3b2f-6567-7f0f-cb03-fd98a8d0b9f0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.826212] env[65726]: DEBUG oslo_vmware.api [None req-300cacab-e765-49c9-a8a1-8e47a20dc6a2 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Task: {'id': task-5115904, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.852381] env[65726]: DEBUG oslo_service.periodic_task [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=65726) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 796.852697] env[65726]: DEBUG oslo_service.periodic_task [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=65726) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 796.987391] env[65726]: DEBUG oslo_vmware.api [None req-debf7b46-904e-41a8-8c0e-839a5a6575cb tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Task: {'id': task-5115905, 'name': PowerOffVM_Task, 'duration_secs': 0.207289} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 796.987886] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-debf7b46-904e-41a8-8c0e-839a5a6575cb tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] [instance: 703fff2d-5aec-4ed1-b987-9e1cc3a25f67] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 796.988187] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-debf7b46-904e-41a8-8c0e-839a5a6575cb tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] [instance: 703fff2d-5aec-4ed1-b987-9e1cc3a25f67] Unregistering the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 796.988599] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c51b0589-fb19-4195-93d6-660eff8cdd0e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.019702] env[65726]: DEBUG oslo_vmware.api [None req-af1f7990-88d5-411a-be3e-a2d4ed6b5b5b tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': task-5115906, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.077529] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-debf7b46-904e-41a8-8c0e-839a5a6575cb tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] [instance: 703fff2d-5aec-4ed1-b987-9e1cc3a25f67] Unregistered the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 797.077709] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-debf7b46-904e-41a8-8c0e-839a5a6575cb tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] [instance: 703fff2d-5aec-4ed1-b987-9e1cc3a25f67] Deleting contents of the VM from datastore datastore2 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 797.077921] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-debf7b46-904e-41a8-8c0e-839a5a6575cb tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Deleting the datastore file [datastore2] 703fff2d-5aec-4ed1-b987-9e1cc3a25f67 {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 797.078231] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-db41d995-124f-4701-a92e-561255f6daa7 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.087582] env[65726]: DEBUG oslo_vmware.api [None req-debf7b46-904e-41a8-8c0e-839a5a6575cb tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Waiting for the task: (returnval){ [ 797.087582] env[65726]: value = "task-5115908" [ 797.087582] env[65726]: _type = "Task" [ 797.087582] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 797.098173] env[65726]: DEBUG oslo_vmware.api [None req-debf7b46-904e-41a8-8c0e-839a5a6575cb tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Task: {'id': task-5115908, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.243169] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c9cc880b-42e7-4875-9b8e-55f0109d1664 tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Lock "ca67d1a4-d304-4d06-a436-f257b60ca080" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 39.625s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 797.317871] env[65726]: DEBUG oslo_vmware.api [None req-1427c99f-55ff-4567-8d49-f5859451e059 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]520e3b2f-6567-7f0f-cb03-fd98a8d0b9f0, 'name': SearchDatastore_Task, 'duration_secs': 0.03447} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 797.318671] env[65726]: DEBUG oslo_concurrency.lockutils [None req-1427c99f-55ff-4567-8d49-f5859451e059 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 797.318943] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-1427c99f-55ff-4567-8d49-f5859451e059 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Copying virtual disk from [datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore2] 6d2b86c2-0a44-433f-8b3f-2e9c7f693baa/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c-rescue.vmdk. {{(pid=65726) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 797.319289] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d045d336-1281-4eec-9683-9c6627823426 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.324817] env[65726]: DEBUG oslo_vmware.api [None req-300cacab-e765-49c9-a8a1-8e47a20dc6a2 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Task: {'id': task-5115904, 'name': PowerOnVM_Task, 'duration_secs': 0.716913} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 797.325560] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-300cacab-e765-49c9-a8a1-8e47a20dc6a2 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] [instance: 76249623-6f83-46a3-b8c5-c001111aa698] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 797.325767] env[65726]: INFO nova.compute.manager [None req-300cacab-e765-49c9-a8a1-8e47a20dc6a2 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] [instance: 76249623-6f83-46a3-b8c5-c001111aa698] Took 8.07 seconds to spawn the instance on the hypervisor. [ 797.325949] env[65726]: DEBUG nova.compute.manager [None req-300cacab-e765-49c9-a8a1-8e47a20dc6a2 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] [instance: 76249623-6f83-46a3-b8c5-c001111aa698] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 797.326920] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e39b9144-a054-46c7-ac65-6829c1bae1dc {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.331612] env[65726]: DEBUG oslo_vmware.api [None req-1427c99f-55ff-4567-8d49-f5859451e059 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Waiting for the task: (returnval){ [ 797.331612] env[65726]: value = "task-5115909" [ 797.331612] env[65726]: _type = "Task" [ 797.331612] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 797.345912] env[65726]: DEBUG oslo_vmware.api [None req-1427c99f-55ff-4567-8d49-f5859451e059 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Task: {'id': task-5115909, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.363977] env[65726]: DEBUG oslo_service.periodic_task [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=65726) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 797.364101] env[65726]: DEBUG oslo_service.periodic_task [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=65726) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 797.364252] env[65726]: DEBUG oslo_service.periodic_task [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=65726) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 797.365566] env[65726]: DEBUG oslo_service.periodic_task [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=65726) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 797.365826] env[65726]: DEBUG oslo_service.periodic_task [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=65726) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 797.365989] env[65726]: DEBUG oslo_service.periodic_task [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=65726) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 797.366148] env[65726]: DEBUG nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=65726) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11251}} [ 797.366295] env[65726]: DEBUG oslo_service.periodic_task [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Running periodic task ComputeManager.update_available_resource {{(pid=65726) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 797.520963] env[65726]: DEBUG oslo_vmware.api [None req-af1f7990-88d5-411a-be3e-a2d4ed6b5b5b tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': task-5115906, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.598431] env[65726]: DEBUG oslo_vmware.api [None req-debf7b46-904e-41a8-8c0e-839a5a6575cb tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Task: {'id': task-5115908, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.177924} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 797.598761] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-debf7b46-904e-41a8-8c0e-839a5a6575cb tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Deleted the datastore file {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 797.598958] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-debf7b46-904e-41a8-8c0e-839a5a6575cb tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] [instance: 703fff2d-5aec-4ed1-b987-9e1cc3a25f67] Deleted contents of the VM from datastore datastore2 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 797.599223] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-debf7b46-904e-41a8-8c0e-839a5a6575cb tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] [instance: 703fff2d-5aec-4ed1-b987-9e1cc3a25f67] Instance destroyed {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 797.599316] env[65726]: INFO nova.compute.manager [None req-debf7b46-904e-41a8-8c0e-839a5a6575cb tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] [instance: 703fff2d-5aec-4ed1-b987-9e1cc3a25f67] Took 1.14 seconds to destroy the instance on the hypervisor. [ 797.599562] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-debf7b46-904e-41a8-8c0e-839a5a6575cb tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 797.599776] env[65726]: DEBUG nova.compute.manager [-] [instance: 703fff2d-5aec-4ed1-b987-9e1cc3a25f67] Deallocating network for instance {{(pid=65726) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 797.599939] env[65726]: DEBUG nova.network.neutron [-] [instance: 703fff2d-5aec-4ed1-b987-9e1cc3a25f67] deallocate_for_instance() {{(pid=65726) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 797.600479] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 797.600741] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 797.646931] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 797.647532] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 797.846249] env[65726]: DEBUG oslo_vmware.api [None req-1427c99f-55ff-4567-8d49-f5859451e059 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Task: {'id': task-5115909, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.853441] env[65726]: INFO nova.compute.manager [None req-300cacab-e765-49c9-a8a1-8e47a20dc6a2 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] [instance: 76249623-6f83-46a3-b8c5-c001111aa698] Took 45.42 seconds to build instance. [ 797.869335] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 797.905716] env[65726]: DEBUG nova.compute.manager [req-d106d27d-8ee6-4f4c-8ffa-37a81929edf5 req-8637e8bc-f6d6-4799-8c8f-bfe50a07504a service nova] [instance: 703fff2d-5aec-4ed1-b987-9e1cc3a25f67] Received event network-vif-deleted-29330b5b-6171-4de8-bf77-6c57f6a8bd58 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 797.905921] env[65726]: INFO nova.compute.manager [req-d106d27d-8ee6-4f4c-8ffa-37a81929edf5 req-8637e8bc-f6d6-4799-8c8f-bfe50a07504a service nova] [instance: 703fff2d-5aec-4ed1-b987-9e1cc3a25f67] Neutron deleted interface 29330b5b-6171-4de8-bf77-6c57f6a8bd58; detaching it from the instance and deleting it from the info cache [ 797.906104] env[65726]: DEBUG nova.network.neutron [req-d106d27d-8ee6-4f4c-8ffa-37a81929edf5 req-8637e8bc-f6d6-4799-8c8f-bfe50a07504a service nova] [instance: 703fff2d-5aec-4ed1-b987-9e1cc3a25f67] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 798.020592] env[65726]: DEBUG oslo_vmware.api [None req-af1f7990-88d5-411a-be3e-a2d4ed6b5b5b tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': task-5115906, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.224676] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01b6701a-2a09-4dad-bb0d-2c1d428f131a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.234221] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18ac227f-f1d3-47d0-a4bd-7fe0bf1b9c05 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.268701] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8d11002-6639-4575-a3a9-090077c0ddb5 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.278197] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41a174a1-027f-40bd-baf4-6cc3e91439c9 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.294747] env[65726]: DEBUG nova.compute.provider_tree [None req-04d3888f-ac16-4533-b7db-d3c669b17255 tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 798.343976] env[65726]: DEBUG oslo_vmware.api [None req-1427c99f-55ff-4567-8d49-f5859451e059 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Task: {'id': task-5115909, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.353741] env[65726]: DEBUG nova.network.neutron [-] [instance: 703fff2d-5aec-4ed1-b987-9e1cc3a25f67] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 798.357309] env[65726]: DEBUG oslo_concurrency.lockutils [None req-300cacab-e765-49c9-a8a1-8e47a20dc6a2 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Lock "76249623-6f83-46a3-b8c5-c001111aa698" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 92.514s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 798.410197] env[65726]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-55e37d6f-b1b1-4352-b207-c3c238345795 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.421727] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c5951a2-9594-4c07-b05c-d8ab83ddd6db {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.464727] env[65726]: DEBUG nova.compute.manager [req-d106d27d-8ee6-4f4c-8ffa-37a81929edf5 req-8637e8bc-f6d6-4799-8c8f-bfe50a07504a service nova] [instance: 703fff2d-5aec-4ed1-b987-9e1cc3a25f67] Detach interface failed, port_id=29330b5b-6171-4de8-bf77-6c57f6a8bd58, reason: Instance 703fff2d-5aec-4ed1-b987-9e1cc3a25f67 could not be found. {{(pid=65726) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11669}} [ 798.517751] env[65726]: DEBUG oslo_vmware.api [None req-af1f7990-88d5-411a-be3e-a2d4ed6b5b5b tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': task-5115906, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.552772} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 798.518227] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-af1f7990-88d5-411a-be3e-a2d4ed6b5b5b tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] 621b899a-e9ec-4fc0-a574-4a08cecff6d9/621b899a-e9ec-4fc0-a574-4a08cecff6d9.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 798.518384] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-af1f7990-88d5-411a-be3e-a2d4ed6b5b5b tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: 621b899a-e9ec-4fc0-a574-4a08cecff6d9] Extending root virtual disk to 1048576 {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 798.518890] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-dca87c4a-2cc7-4c72-9387-d245ea33f75f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.527156] env[65726]: DEBUG oslo_vmware.api [None req-af1f7990-88d5-411a-be3e-a2d4ed6b5b5b tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Waiting for the task: (returnval){ [ 798.527156] env[65726]: value = "task-5115910" [ 798.527156] env[65726]: _type = "Task" [ 798.527156] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 798.538169] env[65726]: DEBUG oslo_vmware.api [None req-af1f7990-88d5-411a-be3e-a2d4ed6b5b5b tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': task-5115910, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.798484] env[65726]: DEBUG nova.scheduler.client.report [None req-04d3888f-ac16-4533-b7db-d3c669b17255 tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 798.843632] env[65726]: DEBUG oslo_vmware.api [None req-1427c99f-55ff-4567-8d49-f5859451e059 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Task: {'id': task-5115909, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.190736} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 798.843921] env[65726]: INFO nova.virt.vmwareapi.ds_util [None req-1427c99f-55ff-4567-8d49-f5859451e059 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Copied virtual disk from [datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore2] 6d2b86c2-0a44-433f-8b3f-2e9c7f693baa/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c-rescue.vmdk. [ 798.844783] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db43102b-8c9e-4336-8299-534f74f67283 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.865785] env[65726]: INFO nova.compute.manager [-] [instance: 703fff2d-5aec-4ed1-b987-9e1cc3a25f67] Took 1.27 seconds to deallocate network for instance. [ 798.866562] env[65726]: DEBUG nova.compute.manager [None req-a6bb6238-1018-4e0c-84f4-13a2ac492d77 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] [instance: 44d73b2d-2b6a-4501-9944-432da3c9330d] Starting instance... {{(pid=65726) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 798.876410] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-1427c99f-55ff-4567-8d49-f5859451e059 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: 6d2b86c2-0a44-433f-8b3f-2e9c7f693baa] Reconfiguring VM instance instance-0000002b to attach disk [datastore2] 6d2b86c2-0a44-433f-8b3f-2e9c7f693baa/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c-rescue.vmdk or device None with type thin {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 798.878358] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fa860ddd-4572-4b1a-b87d-e5c786ec544a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.902625] env[65726]: DEBUG oslo_vmware.api [None req-1427c99f-55ff-4567-8d49-f5859451e059 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Waiting for the task: (returnval){ [ 798.902625] env[65726]: value = "task-5115911" [ 798.902625] env[65726]: _type = "Task" [ 798.902625] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 798.912323] env[65726]: DEBUG oslo_vmware.api [None req-1427c99f-55ff-4567-8d49-f5859451e059 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Task: {'id': task-5115911, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.038512] env[65726]: DEBUG oslo_vmware.api [None req-af1f7990-88d5-411a-be3e-a2d4ed6b5b5b tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': task-5115910, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.075146} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 799.038888] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-af1f7990-88d5-411a-be3e-a2d4ed6b5b5b tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: 621b899a-e9ec-4fc0-a574-4a08cecff6d9] Extended root virtual disk {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 799.039869] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc7c29e8-b045-46eb-9318-e5cbfb4d981d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.064124] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-af1f7990-88d5-411a-be3e-a2d4ed6b5b5b tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: 621b899a-e9ec-4fc0-a574-4a08cecff6d9] Reconfiguring VM instance instance-0000002f to attach disk [datastore1] 621b899a-e9ec-4fc0-a574-4a08cecff6d9/621b899a-e9ec-4fc0-a574-4a08cecff6d9.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 799.064813] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4efe11e4-547e-4b60-adbc-bca322bab04b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.086992] env[65726]: DEBUG oslo_vmware.api [None req-af1f7990-88d5-411a-be3e-a2d4ed6b5b5b tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Waiting for the task: (returnval){ [ 799.086992] env[65726]: value = "task-5115912" [ 799.086992] env[65726]: _type = "Task" [ 799.086992] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 799.097832] env[65726]: DEBUG oslo_vmware.api [None req-af1f7990-88d5-411a-be3e-a2d4ed6b5b5b tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': task-5115912, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.235569] env[65726]: DEBUG nova.compute.manager [None req-4562f28d-bad2-4bca-833c-c069e59f6cc6 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] [instance: 76249623-6f83-46a3-b8c5-c001111aa698] Stashing vm_state: active {{(pid=65726) _prep_resize /opt/stack/nova/nova/compute/manager.py:6193}} [ 799.304902] env[65726]: DEBUG oslo_concurrency.lockutils [None req-04d3888f-ac16-4533-b7db-d3c669b17255 tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.602s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 799.305561] env[65726]: DEBUG nova.compute.manager [None req-04d3888f-ac16-4533-b7db-d3c669b17255 tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] [instance: bc2c12e0-0d06-432f-b42f-be468e3b6ee4] Start building networks asynchronously for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 799.311252] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a78d739e-5974-49c3-a6f0-74e291b9e87b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 34.714s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 799.311252] env[65726]: DEBUG nova.objects.instance [None req-a78d739e-5974-49c3-a6f0-74e291b9e87b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Lazy-loading 'resources' on Instance uuid 02351120-8385-4403-8464-a154f3da9380 {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 799.398855] env[65726]: DEBUG oslo_concurrency.lockutils [None req-debf7b46-904e-41a8-8c0e-839a5a6575cb tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 799.400180] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a6bb6238-1018-4e0c-84f4-13a2ac492d77 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 799.413664] env[65726]: DEBUG oslo_vmware.api [None req-1427c99f-55ff-4567-8d49-f5859451e059 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Task: {'id': task-5115911, 'name': ReconfigVM_Task, 'duration_secs': 0.441886} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 799.413933] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-1427c99f-55ff-4567-8d49-f5859451e059 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: 6d2b86c2-0a44-433f-8b3f-2e9c7f693baa] Reconfigured VM instance instance-0000002b to attach disk [datastore2] 6d2b86c2-0a44-433f-8b3f-2e9c7f693baa/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c-rescue.vmdk or device None with type thin {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 799.414791] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca721bd6-e884-4afc-ac0e-afcc8debe6c4 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.440661] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c15f7a4f-f2ae-4d1e-8977-8da640de097f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.457784] env[65726]: DEBUG oslo_vmware.api [None req-1427c99f-55ff-4567-8d49-f5859451e059 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Waiting for the task: (returnval){ [ 799.457784] env[65726]: value = "task-5115913" [ 799.457784] env[65726]: _type = "Task" [ 799.457784] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 799.467048] env[65726]: DEBUG oslo_vmware.api [None req-1427c99f-55ff-4567-8d49-f5859451e059 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Task: {'id': task-5115913, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.598641] env[65726]: DEBUG oslo_vmware.api [None req-af1f7990-88d5-411a-be3e-a2d4ed6b5b5b tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': task-5115912, 'name': ReconfigVM_Task, 'duration_secs': 0.298099} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 799.598958] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-af1f7990-88d5-411a-be3e-a2d4ed6b5b5b tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: 621b899a-e9ec-4fc0-a574-4a08cecff6d9] Reconfigured VM instance instance-0000002f to attach disk [datastore1] 621b899a-e9ec-4fc0-a574-4a08cecff6d9/621b899a-e9ec-4fc0-a574-4a08cecff6d9.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 799.599697] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c146b69f-451c-420a-8cf7-c84ffb81a0df {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.607722] env[65726]: DEBUG oslo_vmware.api [None req-af1f7990-88d5-411a-be3e-a2d4ed6b5b5b tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Waiting for the task: (returnval){ [ 799.607722] env[65726]: value = "task-5115914" [ 799.607722] env[65726]: _type = "Task" [ 799.607722] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 799.618075] env[65726]: DEBUG oslo_vmware.api [None req-af1f7990-88d5-411a-be3e-a2d4ed6b5b5b tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': task-5115914, 'name': Rename_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.757759] env[65726]: DEBUG oslo_concurrency.lockutils [None req-4562f28d-bad2-4bca-833c-c069e59f6cc6 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 799.812850] env[65726]: DEBUG nova.compute.utils [None req-04d3888f-ac16-4533-b7db-d3c669b17255 tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Using /dev/sd instead of None {{(pid=65726) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 799.813855] env[65726]: DEBUG nova.compute.manager [None req-04d3888f-ac16-4533-b7db-d3c669b17255 tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] [instance: bc2c12e0-0d06-432f-b42f-be468e3b6ee4] Allocating IP information in the background. {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 799.814071] env[65726]: DEBUG nova.network.neutron [None req-04d3888f-ac16-4533-b7db-d3c669b17255 tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] [instance: bc2c12e0-0d06-432f-b42f-be468e3b6ee4] allocate_for_instance() {{(pid=65726) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 799.814396] env[65726]: WARNING neutronclient.v2_0.client [None req-04d3888f-ac16-4533-b7db-d3c669b17255 tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 799.814702] env[65726]: WARNING neutronclient.v2_0.client [None req-04d3888f-ac16-4533-b7db-d3c669b17255 tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 799.815291] env[65726]: WARNING openstack [None req-04d3888f-ac16-4533-b7db-d3c669b17255 tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 799.815641] env[65726]: WARNING openstack [None req-04d3888f-ac16-4533-b7db-d3c669b17255 tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 799.872923] env[65726]: DEBUG nova.policy [None req-04d3888f-ac16-4533-b7db-d3c669b17255 tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e5c535534cca4b6f9e55c2ada589dc05', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '55f93cdf51e84cf683879668c2acecf4', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65726) authorize /opt/stack/nova/nova/policy.py:192}} [ 799.968617] env[65726]: DEBUG oslo_vmware.api [None req-1427c99f-55ff-4567-8d49-f5859451e059 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Task: {'id': task-5115913, 'name': ReconfigVM_Task, 'duration_secs': 0.227489} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 799.971357] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-1427c99f-55ff-4567-8d49-f5859451e059 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: 6d2b86c2-0a44-433f-8b3f-2e9c7f693baa] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 799.972063] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e4adb6e3-a9f8-45ca-bdde-8c6f097731ec {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.978946] env[65726]: DEBUG oslo_vmware.api [None req-1427c99f-55ff-4567-8d49-f5859451e059 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Waiting for the task: (returnval){ [ 799.978946] env[65726]: value = "task-5115915" [ 799.978946] env[65726]: _type = "Task" [ 799.978946] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 799.991912] env[65726]: DEBUG oslo_vmware.api [None req-1427c99f-55ff-4567-8d49-f5859451e059 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Task: {'id': task-5115915, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.111395] env[65726]: DEBUG nova.network.neutron [None req-04d3888f-ac16-4533-b7db-d3c669b17255 tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] [instance: bc2c12e0-0d06-432f-b42f-be468e3b6ee4] Successfully created port: a1c6ed0b-f1f6-4da3-aeaf-9f815ef90b64 {{(pid=65726) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 800.127273] env[65726]: DEBUG oslo_vmware.api [None req-af1f7990-88d5-411a-be3e-a2d4ed6b5b5b tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': task-5115914, 'name': Rename_Task, 'duration_secs': 0.150345} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 800.127893] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-af1f7990-88d5-411a-be3e-a2d4ed6b5b5b tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: 621b899a-e9ec-4fc0-a574-4a08cecff6d9] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 800.127893] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-99b42cfb-8a2a-4888-87cf-dc7dcd15d168 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.138347] env[65726]: DEBUG oslo_vmware.api [None req-af1f7990-88d5-411a-be3e-a2d4ed6b5b5b tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Waiting for the task: (returnval){ [ 800.138347] env[65726]: value = "task-5115916" [ 800.138347] env[65726]: _type = "Task" [ 800.138347] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 800.151168] env[65726]: DEBUG oslo_vmware.api [None req-af1f7990-88d5-411a-be3e-a2d4ed6b5b5b tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': task-5115916, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.326556] env[65726]: DEBUG nova.compute.manager [None req-04d3888f-ac16-4533-b7db-d3c669b17255 tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] [instance: bc2c12e0-0d06-432f-b42f-be468e3b6ee4] Start building block device mappings for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 800.333027] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ceecc283-03d2-4cf8-991d-6ab4da8f3fb8 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.340444] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a87d95f1-3f80-4c67-813e-5f88fd0e802b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.375445] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0b8bb9d-0c90-40de-93ce-1e455596f9d4 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.384822] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9a494ca-23d1-4f5c-b5ec-88be7d6f4cbe {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.400265] env[65726]: DEBUG nova.compute.provider_tree [None req-a78d739e-5974-49c3-a6f0-74e291b9e87b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 800.491080] env[65726]: DEBUG oslo_vmware.api [None req-1427c99f-55ff-4567-8d49-f5859451e059 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Task: {'id': task-5115915, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.658122] env[65726]: DEBUG oslo_vmware.api [None req-af1f7990-88d5-411a-be3e-a2d4ed6b5b5b tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': task-5115916, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.904863] env[65726]: DEBUG nova.scheduler.client.report [None req-a78d739e-5974-49c3-a6f0-74e291b9e87b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 800.990559] env[65726]: DEBUG oslo_vmware.api [None req-1427c99f-55ff-4567-8d49-f5859451e059 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Task: {'id': task-5115915, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.153972] env[65726]: DEBUG oslo_vmware.api [None req-af1f7990-88d5-411a-be3e-a2d4ed6b5b5b tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': task-5115916, 'name': PowerOnVM_Task, 'duration_secs': 0.528694} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 801.154137] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-af1f7990-88d5-411a-be3e-a2d4ed6b5b5b tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: 621b899a-e9ec-4fc0-a574-4a08cecff6d9] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 801.154432] env[65726]: INFO nova.compute.manager [None req-af1f7990-88d5-411a-be3e-a2d4ed6b5b5b tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: 621b899a-e9ec-4fc0-a574-4a08cecff6d9] Took 9.21 seconds to spawn the instance on the hypervisor. [ 801.154725] env[65726]: DEBUG nova.compute.manager [None req-af1f7990-88d5-411a-be3e-a2d4ed6b5b5b tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: 621b899a-e9ec-4fc0-a574-4a08cecff6d9] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 801.155817] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8ab56c6-ff81-4b25-bbb5-e5487d3531f6 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.343570] env[65726]: DEBUG nova.compute.manager [None req-04d3888f-ac16-4533-b7db-d3c669b17255 tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] [instance: bc2c12e0-0d06-432f-b42f-be468e3b6ee4] Start spawning the instance on the hypervisor. {{(pid=65726) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 801.371364] env[65726]: DEBUG nova.virt.hardware [None req-04d3888f-ac16-4533-b7db-d3c669b17255 tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T19:28:34Z,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1539f06fa3534e90acbc3a60c4b8bd3f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T19:28:35Z,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 801.371647] env[65726]: DEBUG nova.virt.hardware [None req-04d3888f-ac16-4533-b7db-d3c669b17255 tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 801.371776] env[65726]: DEBUG nova.virt.hardware [None req-04d3888f-ac16-4533-b7db-d3c669b17255 tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 801.371947] env[65726]: DEBUG nova.virt.hardware [None req-04d3888f-ac16-4533-b7db-d3c669b17255 tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 801.372177] env[65726]: DEBUG nova.virt.hardware [None req-04d3888f-ac16-4533-b7db-d3c669b17255 tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 801.372350] env[65726]: DEBUG nova.virt.hardware [None req-04d3888f-ac16-4533-b7db-d3c669b17255 tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 801.372567] env[65726]: DEBUG nova.virt.hardware [None req-04d3888f-ac16-4533-b7db-d3c669b17255 tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 801.372720] env[65726]: DEBUG nova.virt.hardware [None req-04d3888f-ac16-4533-b7db-d3c669b17255 tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 801.372885] env[65726]: DEBUG nova.virt.hardware [None req-04d3888f-ac16-4533-b7db-d3c669b17255 tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 801.373058] env[65726]: DEBUG nova.virt.hardware [None req-04d3888f-ac16-4533-b7db-d3c669b17255 tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 801.373239] env[65726]: DEBUG nova.virt.hardware [None req-04d3888f-ac16-4533-b7db-d3c669b17255 tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 801.374560] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3134d7e9-de2d-4b7b-bd7d-d56ef7b371d0 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.383518] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35952c22-955e-465e-989e-5fc41f5afb12 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.409499] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a78d739e-5974-49c3-a6f0-74e291b9e87b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.099s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 801.411813] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9ff22eed-6c42-49bb-b683-3f83a9ff00ca tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 36.072s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 801.412059] env[65726]: DEBUG nova.objects.instance [None req-9ff22eed-6c42-49bb-b683-3f83a9ff00ca tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Lazy-loading 'resources' on Instance uuid ee676e3b-3326-46a3-940d-ebbb8b108991 {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 801.434673] env[65726]: INFO nova.scheduler.client.report [None req-a78d739e-5974-49c3-a6f0-74e291b9e87b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Deleted allocations for instance 02351120-8385-4403-8464-a154f3da9380 [ 801.491309] env[65726]: DEBUG oslo_vmware.api [None req-1427c99f-55ff-4567-8d49-f5859451e059 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Task: {'id': task-5115915, 'name': PowerOnVM_Task, 'duration_secs': 1.237225} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 801.491597] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-1427c99f-55ff-4567-8d49-f5859451e059 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: 6d2b86c2-0a44-433f-8b3f-2e9c7f693baa] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 801.494694] env[65726]: DEBUG nova.compute.manager [None req-1427c99f-55ff-4567-8d49-f5859451e059 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: 6d2b86c2-0a44-433f-8b3f-2e9c7f693baa] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 801.495727] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27387d9e-587f-4445-9c69-f8dd6f745ca6 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.682825] env[65726]: INFO nova.compute.manager [None req-af1f7990-88d5-411a-be3e-a2d4ed6b5b5b tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: 621b899a-e9ec-4fc0-a574-4a08cecff6d9] Took 46.73 seconds to build instance. [ 801.717376] env[65726]: DEBUG nova.network.neutron [None req-04d3888f-ac16-4533-b7db-d3c669b17255 tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] [instance: bc2c12e0-0d06-432f-b42f-be468e3b6ee4] Successfully updated port: a1c6ed0b-f1f6-4da3-aeaf-9f815ef90b64 {{(pid=65726) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 801.851466] env[65726]: DEBUG oslo_concurrency.lockutils [None req-2c0d0141-cf5e-454f-809c-529966ff3002 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Acquiring lock "interface-49bcab1f-7b64-4999-abff-37771c58a271-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 801.851466] env[65726]: DEBUG oslo_concurrency.lockutils [None req-2c0d0141-cf5e-454f-809c-529966ff3002 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Lock "interface-49bcab1f-7b64-4999-abff-37771c58a271-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 801.851658] env[65726]: DEBUG nova.objects.instance [None req-2c0d0141-cf5e-454f-809c-529966ff3002 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Lazy-loading 'flavor' on Instance uuid 49bcab1f-7b64-4999-abff-37771c58a271 {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 801.857209] env[65726]: DEBUG nova.compute.manager [req-3231e419-ee7e-4784-b78d-2d5afaea93b7 req-89933359-19c6-436e-ae82-c9a91cd6a4f7 service nova] [instance: bc2c12e0-0d06-432f-b42f-be468e3b6ee4] Received event network-vif-plugged-a1c6ed0b-f1f6-4da3-aeaf-9f815ef90b64 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 801.857457] env[65726]: DEBUG oslo_concurrency.lockutils [req-3231e419-ee7e-4784-b78d-2d5afaea93b7 req-89933359-19c6-436e-ae82-c9a91cd6a4f7 service nova] Acquiring lock "bc2c12e0-0d06-432f-b42f-be468e3b6ee4-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 801.857680] env[65726]: DEBUG oslo_concurrency.lockutils [req-3231e419-ee7e-4784-b78d-2d5afaea93b7 req-89933359-19c6-436e-ae82-c9a91cd6a4f7 service nova] Lock "bc2c12e0-0d06-432f-b42f-be468e3b6ee4-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 801.857761] env[65726]: DEBUG oslo_concurrency.lockutils [req-3231e419-ee7e-4784-b78d-2d5afaea93b7 req-89933359-19c6-436e-ae82-c9a91cd6a4f7 service nova] Lock "bc2c12e0-0d06-432f-b42f-be468e3b6ee4-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 801.857993] env[65726]: DEBUG nova.compute.manager [req-3231e419-ee7e-4784-b78d-2d5afaea93b7 req-89933359-19c6-436e-ae82-c9a91cd6a4f7 service nova] [instance: bc2c12e0-0d06-432f-b42f-be468e3b6ee4] No waiting events found dispatching network-vif-plugged-a1c6ed0b-f1f6-4da3-aeaf-9f815ef90b64 {{(pid=65726) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 801.858156] env[65726]: WARNING nova.compute.manager [req-3231e419-ee7e-4784-b78d-2d5afaea93b7 req-89933359-19c6-436e-ae82-c9a91cd6a4f7 service nova] [instance: bc2c12e0-0d06-432f-b42f-be468e3b6ee4] Received unexpected event network-vif-plugged-a1c6ed0b-f1f6-4da3-aeaf-9f815ef90b64 for instance with vm_state building and task_state spawning. [ 801.945622] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a78d739e-5974-49c3-a6f0-74e291b9e87b tempest-MultipleCreateTestJSON-365649581 tempest-MultipleCreateTestJSON-365649581-project-member] Lock "02351120-8385-4403-8464-a154f3da9380" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 42.234s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 802.185328] env[65726]: DEBUG oslo_concurrency.lockutils [None req-af1f7990-88d5-411a-be3e-a2d4ed6b5b5b tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Lock "621b899a-e9ec-4fc0-a574-4a08cecff6d9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 83.769s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 802.220736] env[65726]: DEBUG oslo_concurrency.lockutils [None req-04d3888f-ac16-4533-b7db-d3c669b17255 tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Acquiring lock "refresh_cache-bc2c12e0-0d06-432f-b42f-be468e3b6ee4" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 802.220736] env[65726]: DEBUG oslo_concurrency.lockutils [None req-04d3888f-ac16-4533-b7db-d3c669b17255 tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Acquired lock "refresh_cache-bc2c12e0-0d06-432f-b42f-be468e3b6ee4" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 802.220736] env[65726]: DEBUG nova.network.neutron [None req-04d3888f-ac16-4533-b7db-d3c669b17255 tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] [instance: bc2c12e0-0d06-432f-b42f-be468e3b6ee4] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 802.355265] env[65726]: WARNING openstack [None req-2c0d0141-cf5e-454f-809c-529966ff3002 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 802.355540] env[65726]: WARNING openstack [None req-2c0d0141-cf5e-454f-809c-529966ff3002 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 802.397054] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad1e2a5a-a089-4f18-acbc-f6808b373dd3 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.407811] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bc77166-aaf3-4e50-b6ff-2fc26f24352a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.442280] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d093d89-e80f-4ba3-87cd-8530b47e3bec {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.451909] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0776222-5500-4e24-9e60-2409ba505db0 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.467063] env[65726]: DEBUG nova.compute.provider_tree [None req-9ff22eed-6c42-49bb-b683-3f83a9ff00ca tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 802.484017] env[65726]: DEBUG nova.objects.instance [None req-2c0d0141-cf5e-454f-809c-529966ff3002 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Lazy-loading 'pci_requests' on Instance uuid 49bcab1f-7b64-4999-abff-37771c58a271 {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 802.670308] env[65726]: DEBUG oslo_concurrency.lockutils [None req-861b1b19-8cf2-4cb5-a3b7-93ab69618cb7 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Acquiring lock "621b899a-e9ec-4fc0-a574-4a08cecff6d9" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 802.670644] env[65726]: DEBUG oslo_concurrency.lockutils [None req-861b1b19-8cf2-4cb5-a3b7-93ab69618cb7 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Lock "621b899a-e9ec-4fc0-a574-4a08cecff6d9" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 802.670790] env[65726]: DEBUG oslo_concurrency.lockutils [None req-861b1b19-8cf2-4cb5-a3b7-93ab69618cb7 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Acquiring lock "621b899a-e9ec-4fc0-a574-4a08cecff6d9-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 802.670983] env[65726]: DEBUG oslo_concurrency.lockutils [None req-861b1b19-8cf2-4cb5-a3b7-93ab69618cb7 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Lock "621b899a-e9ec-4fc0-a574-4a08cecff6d9-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 802.671193] env[65726]: DEBUG oslo_concurrency.lockutils [None req-861b1b19-8cf2-4cb5-a3b7-93ab69618cb7 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Lock "621b899a-e9ec-4fc0-a574-4a08cecff6d9-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 802.675652] env[65726]: INFO nova.compute.manager [None req-861b1b19-8cf2-4cb5-a3b7-93ab69618cb7 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: 621b899a-e9ec-4fc0-a574-4a08cecff6d9] Terminating instance [ 802.687589] env[65726]: DEBUG nova.compute.manager [None req-b0e194cf-16dc-40d0-b65c-88bd21cd63e3 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: cb8855b6-589d-4863-b86f-7fb4e30fb29e] Starting instance... {{(pid=65726) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 802.723016] env[65726]: WARNING openstack [None req-04d3888f-ac16-4533-b7db-d3c669b17255 tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 802.723490] env[65726]: WARNING openstack [None req-04d3888f-ac16-4533-b7db-d3c669b17255 tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 802.728680] env[65726]: DEBUG nova.network.neutron [None req-04d3888f-ac16-4533-b7db-d3c669b17255 tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] [instance: bc2c12e0-0d06-432f-b42f-be468e3b6ee4] Instance cache missing network info. {{(pid=65726) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 802.785884] env[65726]: WARNING openstack [None req-04d3888f-ac16-4533-b7db-d3c669b17255 tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 802.786304] env[65726]: WARNING openstack [None req-04d3888f-ac16-4533-b7db-d3c669b17255 tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 802.883030] env[65726]: WARNING openstack [None req-04d3888f-ac16-4533-b7db-d3c669b17255 tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 802.884860] env[65726]: WARNING openstack [None req-04d3888f-ac16-4533-b7db-d3c669b17255 tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 802.970294] env[65726]: DEBUG nova.scheduler.client.report [None req-9ff22eed-6c42-49bb-b683-3f83a9ff00ca tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 802.982730] env[65726]: DEBUG nova.network.neutron [None req-04d3888f-ac16-4533-b7db-d3c669b17255 tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] [instance: bc2c12e0-0d06-432f-b42f-be468e3b6ee4] Updating instance_info_cache with network_info: [{"id": "a1c6ed0b-f1f6-4da3-aeaf-9f815ef90b64", "address": "fa:16:3e:09:eb:ee", "network": {"id": "dd4bbaef-0a1e-4e00-b65a-f19e7a837cc1", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-255385230-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "55f93cdf51e84cf683879668c2acecf4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "04ccbc7a-cf8d-4ea2-8411-291a1e27df7b", "external-id": "nsx-vlan-transportzone-998", "segmentation_id": 998, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa1c6ed0b-f1", "ovs_interfaceid": "a1c6ed0b-f1f6-4da3-aeaf-9f815ef90b64", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 802.986993] env[65726]: DEBUG nova.objects.base [None req-2c0d0141-cf5e-454f-809c-529966ff3002 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Object Instance<49bcab1f-7b64-4999-abff-37771c58a271> lazy-loaded attributes: flavor,pci_requests {{(pid=65726) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 802.987262] env[65726]: DEBUG nova.network.neutron [None req-2c0d0141-cf5e-454f-809c-529966ff3002 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: 49bcab1f-7b64-4999-abff-37771c58a271] allocate_for_instance() {{(pid=65726) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 802.987701] env[65726]: WARNING neutronclient.v2_0.client [None req-2c0d0141-cf5e-454f-809c-529966ff3002 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 802.988137] env[65726]: WARNING neutronclient.v2_0.client [None req-2c0d0141-cf5e-454f-809c-529966ff3002 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 802.988865] env[65726]: WARNING openstack [None req-2c0d0141-cf5e-454f-809c-529966ff3002 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 802.989474] env[65726]: WARNING openstack [None req-2c0d0141-cf5e-454f-809c-529966ff3002 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 803.046053] env[65726]: DEBUG nova.policy [None req-2c0d0141-cf5e-454f-809c-529966ff3002 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9b65b9194933469c908c7cc04478444b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd46ccb00794f458b85da4a93879139ee', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65726) authorize /opt/stack/nova/nova/policy.py:192}} [ 803.182163] env[65726]: DEBUG nova.compute.manager [None req-861b1b19-8cf2-4cb5-a3b7-93ab69618cb7 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: 621b899a-e9ec-4fc0-a574-4a08cecff6d9] Start destroying the instance on the hypervisor. {{(pid=65726) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 803.182163] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-861b1b19-8cf2-4cb5-a3b7-93ab69618cb7 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: 621b899a-e9ec-4fc0-a574-4a08cecff6d9] Destroying instance {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 803.182163] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb175267-75c3-44d5-bbec-a83daa8adce1 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.189808] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-861b1b19-8cf2-4cb5-a3b7-93ab69618cb7 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: 621b899a-e9ec-4fc0-a574-4a08cecff6d9] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 803.190375] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e12098c6-5956-40d4-8633-a8e4b16239e7 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.202214] env[65726]: DEBUG oslo_vmware.api [None req-861b1b19-8cf2-4cb5-a3b7-93ab69618cb7 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Waiting for the task: (returnval){ [ 803.202214] env[65726]: value = "task-5115917" [ 803.202214] env[65726]: _type = "Task" [ 803.202214] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 803.211470] env[65726]: DEBUG oslo_vmware.api [None req-861b1b19-8cf2-4cb5-a3b7-93ab69618cb7 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': task-5115917, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.216575] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b0e194cf-16dc-40d0-b65c-88bd21cd63e3 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 803.320734] env[65726]: DEBUG nova.network.neutron [None req-2c0d0141-cf5e-454f-809c-529966ff3002 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: 49bcab1f-7b64-4999-abff-37771c58a271] Successfully created port: f2a8ac3c-ad7e-4eb4-aec3-db29631fd7ad {{(pid=65726) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 803.488025] env[65726]: DEBUG oslo_concurrency.lockutils [None req-04d3888f-ac16-4533-b7db-d3c669b17255 tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Releasing lock "refresh_cache-bc2c12e0-0d06-432f-b42f-be468e3b6ee4" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 803.488025] env[65726]: DEBUG nova.compute.manager [None req-04d3888f-ac16-4533-b7db-d3c669b17255 tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] [instance: bc2c12e0-0d06-432f-b42f-be468e3b6ee4] Instance network_info: |[{"id": "a1c6ed0b-f1f6-4da3-aeaf-9f815ef90b64", "address": "fa:16:3e:09:eb:ee", "network": {"id": "dd4bbaef-0a1e-4e00-b65a-f19e7a837cc1", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-255385230-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "55f93cdf51e84cf683879668c2acecf4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "04ccbc7a-cf8d-4ea2-8411-291a1e27df7b", "external-id": "nsx-vlan-transportzone-998", "segmentation_id": 998, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa1c6ed0b-f1", "ovs_interfaceid": "a1c6ed0b-f1f6-4da3-aeaf-9f815ef90b64", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 803.488522] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9ff22eed-6c42-49bb-b683-3f83a9ff00ca tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.075s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 803.489868] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-04d3888f-ac16-4533-b7db-d3c669b17255 tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] [instance: bc2c12e0-0d06-432f-b42f-be468e3b6ee4] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:09:eb:ee', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '04ccbc7a-cf8d-4ea2-8411-291a1e27df7b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a1c6ed0b-f1f6-4da3-aeaf-9f815ef90b64', 'vif_model': 'vmxnet3'}] {{(pid=65726) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 803.499767] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-04d3888f-ac16-4533-b7db-d3c669b17255 tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Creating folder: Project (55f93cdf51e84cf683879668c2acecf4). Parent ref: group-v995008. {{(pid=65726) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 803.500374] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a76cfabd-af7f-4569-b968-716d45bcf7f1 tempest-InstanceActionsNegativeTestJSON-844675881 tempest-InstanceActionsNegativeTestJSON-844675881-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 37.815s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 803.507206] env[65726]: INFO nova.compute.claims [None req-a76cfabd-af7f-4569-b968-716d45bcf7f1 tempest-InstanceActionsNegativeTestJSON-844675881 tempest-InstanceActionsNegativeTestJSON-844675881-project-member] [instance: 1bd26aef-995e-43b6-af9f-077e878c8c44] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 803.511077] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-00e6a6af-c9a7-4b30-8bf9-13305c0bff0a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.524293] env[65726]: INFO nova.scheduler.client.report [None req-9ff22eed-6c42-49bb-b683-3f83a9ff00ca tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Deleted allocations for instance ee676e3b-3326-46a3-940d-ebbb8b108991 [ 803.531802] env[65726]: INFO nova.virt.vmwareapi.vm_util [None req-04d3888f-ac16-4533-b7db-d3c669b17255 tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Created folder: Project (55f93cdf51e84cf683879668c2acecf4) in parent group-v995008. [ 803.532611] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-04d3888f-ac16-4533-b7db-d3c669b17255 tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Creating folder: Instances. Parent ref: group-v995139. {{(pid=65726) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 803.533385] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-731a1a69-9237-4198-acc5-deba22bfaea5 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.548407] env[65726]: INFO nova.virt.vmwareapi.vm_util [None req-04d3888f-ac16-4533-b7db-d3c669b17255 tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Created folder: Instances in parent group-v995139. [ 803.548407] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-04d3888f-ac16-4533-b7db-d3c669b17255 tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 803.548407] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bc2c12e0-0d06-432f-b42f-be468e3b6ee4] Creating VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 803.548718] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ad075d06-8c12-4d01-9967-19f684f3bc0b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.574008] env[65726]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 803.574008] env[65726]: value = "task-5115920" [ 803.574008] env[65726]: _type = "Task" [ 803.574008] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 803.584621] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115920, 'name': CreateVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.714620] env[65726]: DEBUG oslo_vmware.api [None req-861b1b19-8cf2-4cb5-a3b7-93ab69618cb7 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': task-5115917, 'name': PowerOffVM_Task, 'duration_secs': 0.194019} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 803.714620] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-861b1b19-8cf2-4cb5-a3b7-93ab69618cb7 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: 621b899a-e9ec-4fc0-a574-4a08cecff6d9] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 803.714620] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-861b1b19-8cf2-4cb5-a3b7-93ab69618cb7 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: 621b899a-e9ec-4fc0-a574-4a08cecff6d9] Unregistering the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 803.714620] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-313d1a53-d181-4f46-8826-9a061818c82a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.782465] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-861b1b19-8cf2-4cb5-a3b7-93ab69618cb7 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: 621b899a-e9ec-4fc0-a574-4a08cecff6d9] Unregistered the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 803.782465] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-861b1b19-8cf2-4cb5-a3b7-93ab69618cb7 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: 621b899a-e9ec-4fc0-a574-4a08cecff6d9] Deleting contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 803.782465] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-861b1b19-8cf2-4cb5-a3b7-93ab69618cb7 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Deleting the datastore file [datastore1] 621b899a-e9ec-4fc0-a574-4a08cecff6d9 {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 803.782848] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ec433762-3092-4f1e-819d-f7b5e4a13599 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.800738] env[65726]: DEBUG oslo_vmware.api [None req-861b1b19-8cf2-4cb5-a3b7-93ab69618cb7 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Waiting for the task: (returnval){ [ 803.800738] env[65726]: value = "task-5115922" [ 803.800738] env[65726]: _type = "Task" [ 803.800738] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 803.813074] env[65726]: DEBUG oslo_vmware.api [None req-861b1b19-8cf2-4cb5-a3b7-93ab69618cb7 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': task-5115922, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.037042] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9ff22eed-6c42-49bb-b683-3f83a9ff00ca tempest-ServersAdminNegativeTestJSON-1259632363 tempest-ServersAdminNegativeTestJSON-1259632363-project-member] Lock "ee676e3b-3326-46a3-940d-ebbb8b108991" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 42.180s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 804.084489] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115920, 'name': CreateVM_Task, 'duration_secs': 0.359043} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 804.084624] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bc2c12e0-0d06-432f-b42f-be468e3b6ee4] Created VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 804.085616] env[65726]: WARNING openstack [None req-04d3888f-ac16-4533-b7db-d3c669b17255 tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 804.085988] env[65726]: WARNING openstack [None req-04d3888f-ac16-4533-b7db-d3c669b17255 tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 804.092515] env[65726]: DEBUG oslo_concurrency.lockutils [None req-04d3888f-ac16-4533-b7db-d3c669b17255 tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 804.092635] env[65726]: DEBUG oslo_concurrency.lockutils [None req-04d3888f-ac16-4533-b7db-d3c669b17255 tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 804.092888] env[65726]: DEBUG oslo_concurrency.lockutils [None req-04d3888f-ac16-4533-b7db-d3c669b17255 tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 804.093564] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-34afd465-a0fc-4710-a758-ce1ea07c8fdc {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.099484] env[65726]: DEBUG oslo_vmware.api [None req-04d3888f-ac16-4533-b7db-d3c669b17255 tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Waiting for the task: (returnval){ [ 804.099484] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]521354a2-92b5-4835-bbd9-70d4b3cef34c" [ 804.099484] env[65726]: _type = "Task" [ 804.099484] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 804.108932] env[65726]: DEBUG oslo_vmware.api [None req-04d3888f-ac16-4533-b7db-d3c669b17255 tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]521354a2-92b5-4835-bbd9-70d4b3cef34c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.313542] env[65726]: DEBUG oslo_vmware.api [None req-861b1b19-8cf2-4cb5-a3b7-93ab69618cb7 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': task-5115922, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.186868} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 804.313813] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-861b1b19-8cf2-4cb5-a3b7-93ab69618cb7 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Deleted the datastore file {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 804.314076] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-861b1b19-8cf2-4cb5-a3b7-93ab69618cb7 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: 621b899a-e9ec-4fc0-a574-4a08cecff6d9] Deleted contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 804.314269] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-861b1b19-8cf2-4cb5-a3b7-93ab69618cb7 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: 621b899a-e9ec-4fc0-a574-4a08cecff6d9] Instance destroyed {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 804.314566] env[65726]: INFO nova.compute.manager [None req-861b1b19-8cf2-4cb5-a3b7-93ab69618cb7 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: 621b899a-e9ec-4fc0-a574-4a08cecff6d9] Took 1.13 seconds to destroy the instance on the hypervisor. [ 804.314860] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-861b1b19-8cf2-4cb5-a3b7-93ab69618cb7 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 804.315105] env[65726]: DEBUG nova.compute.manager [-] [instance: 621b899a-e9ec-4fc0-a574-4a08cecff6d9] Deallocating network for instance {{(pid=65726) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 804.315204] env[65726]: DEBUG nova.network.neutron [-] [instance: 621b899a-e9ec-4fc0-a574-4a08cecff6d9] deallocate_for_instance() {{(pid=65726) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 804.315728] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 804.315954] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 804.368597] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 804.368872] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 804.502722] env[65726]: DEBUG oslo_vmware.rw_handles [None req-2135c7ab-2ff3-4310-b64c-188bebed565e tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/523612c5-2df1-5f7a-27a4-fa9c623a4605/disk-0.vmdk. {{(pid=65726) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 804.503792] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa753267-0017-4a93-a1d4-a011d2f09b57 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.510853] env[65726]: DEBUG oslo_vmware.rw_handles [None req-2135c7ab-2ff3-4310-b64c-188bebed565e tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/523612c5-2df1-5f7a-27a4-fa9c623a4605/disk-0.vmdk is in state: ready. {{(pid=65726) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 804.511117] env[65726]: ERROR oslo_vmware.rw_handles [None req-2135c7ab-2ff3-4310-b64c-188bebed565e tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/523612c5-2df1-5f7a-27a4-fa9c623a4605/disk-0.vmdk due to incomplete transfer. [ 804.511314] env[65726]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-8b655c3a-3ae5-4076-a742-c4057c8b7b30 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.525616] env[65726]: DEBUG oslo_vmware.rw_handles [None req-2135c7ab-2ff3-4310-b64c-188bebed565e tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/523612c5-2df1-5f7a-27a4-fa9c623a4605/disk-0.vmdk. {{(pid=65726) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 804.525976] env[65726]: DEBUG nova.virt.vmwareapi.images [None req-2135c7ab-2ff3-4310-b64c-188bebed565e tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 608cc0ea-3f6b-4b4f-83c1-01aa50999d63] Uploaded image 9fc0cda5-a631-4766-b141-daaef80ccc69 to the Glance image server {{(pid=65726) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 804.531461] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-2135c7ab-2ff3-4310-b64c-188bebed565e tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 608cc0ea-3f6b-4b4f-83c1-01aa50999d63] Destroying the VM {{(pid=65726) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 804.531796] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-9981c27a-38ab-4a08-9cdd-aa898c51db1b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.539545] env[65726]: DEBUG oslo_vmware.api [None req-2135c7ab-2ff3-4310-b64c-188bebed565e tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Waiting for the task: (returnval){ [ 804.539545] env[65726]: value = "task-5115923" [ 804.539545] env[65726]: _type = "Task" [ 804.539545] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 804.551240] env[65726]: DEBUG oslo_vmware.api [None req-2135c7ab-2ff3-4310-b64c-188bebed565e tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5115923, 'name': Destroy_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.619901] env[65726]: DEBUG oslo_vmware.api [None req-04d3888f-ac16-4533-b7db-d3c669b17255 tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]521354a2-92b5-4835-bbd9-70d4b3cef34c, 'name': SearchDatastore_Task, 'duration_secs': 0.01062} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 804.627627] env[65726]: DEBUG oslo_concurrency.lockutils [None req-04d3888f-ac16-4533-b7db-d3c669b17255 tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 804.627627] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-04d3888f-ac16-4533-b7db-d3c669b17255 tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] [instance: bc2c12e0-0d06-432f-b42f-be468e3b6ee4] Processing image b52362a3-ee8a-4cbf-b06f-513b0cc8f95c {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 804.627841] env[65726]: DEBUG oslo_concurrency.lockutils [None req-04d3888f-ac16-4533-b7db-d3c669b17255 tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 804.627968] env[65726]: DEBUG oslo_concurrency.lockutils [None req-04d3888f-ac16-4533-b7db-d3c669b17255 tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 804.628231] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-04d3888f-ac16-4533-b7db-d3c669b17255 tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 804.629196] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-41022636-517f-4fff-9e16-deae387d4e08 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.645849] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-04d3888f-ac16-4533-b7db-d3c669b17255 tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 804.645849] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-04d3888f-ac16-4533-b7db-d3c669b17255 tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65726) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 804.645849] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0db3ec0e-eca4-4cf0-87de-5539affbd2da {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.647054] env[65726]: DEBUG oslo_vmware.api [None req-04d3888f-ac16-4533-b7db-d3c669b17255 tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Waiting for the task: (returnval){ [ 804.647054] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52674a85-47fe-c975-5b87-d4daf7828410" [ 804.647054] env[65726]: _type = "Task" [ 804.647054] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 804.664608] env[65726]: DEBUG oslo_vmware.api [None req-04d3888f-ac16-4533-b7db-d3c669b17255 tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52674a85-47fe-c975-5b87-d4daf7828410, 'name': SearchDatastore_Task} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 804.664958] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-543e8fba-9657-47c2-b69e-0071a447861f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.676753] env[65726]: DEBUG oslo_vmware.api [None req-04d3888f-ac16-4533-b7db-d3c669b17255 tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Waiting for the task: (returnval){ [ 804.676753] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52688dee-e4ad-f528-1c97-90c2066c9123" [ 804.676753] env[65726]: _type = "Task" [ 804.676753] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 804.693464] env[65726]: DEBUG oslo_vmware.api [None req-04d3888f-ac16-4533-b7db-d3c669b17255 tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52688dee-e4ad-f528-1c97-90c2066c9123, 'name': SearchDatastore_Task, 'duration_secs': 0.014989} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 804.693743] env[65726]: DEBUG oslo_concurrency.lockutils [None req-04d3888f-ac16-4533-b7db-d3c669b17255 tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 804.693996] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-04d3888f-ac16-4533-b7db-d3c669b17255 tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] bc2c12e0-0d06-432f-b42f-be468e3b6ee4/bc2c12e0-0d06-432f-b42f-be468e3b6ee4.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 804.694293] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ac76ee38-c951-4a91-9039-b41540369586 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.704230] env[65726]: DEBUG oslo_vmware.api [None req-04d3888f-ac16-4533-b7db-d3c669b17255 tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Waiting for the task: (returnval){ [ 804.704230] env[65726]: value = "task-5115924" [ 804.704230] env[65726]: _type = "Task" [ 804.704230] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 804.718747] env[65726]: DEBUG oslo_vmware.api [None req-04d3888f-ac16-4533-b7db-d3c669b17255 tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Task: {'id': task-5115924, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.969290] env[65726]: DEBUG nova.network.neutron [None req-2c0d0141-cf5e-454f-809c-529966ff3002 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: 49bcab1f-7b64-4999-abff-37771c58a271] Successfully updated port: f2a8ac3c-ad7e-4eb4-aec3-db29631fd7ad {{(pid=65726) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 805.052426] env[65726]: DEBUG oslo_vmware.api [None req-2135c7ab-2ff3-4310-b64c-188bebed565e tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5115923, 'name': Destroy_Task} progress is 33%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.081055] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9175f430-7707-4f23-ba82-d1f840dd5db9 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.088936] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ccdfd0d-1dfe-4857-ad73-48ff56d7d646 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.125783] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8278cb2-cb3a-475a-ad2b-1d35c422e352 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.129269] env[65726]: DEBUG nova.network.neutron [-] [instance: 621b899a-e9ec-4fc0-a574-4a08cecff6d9] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 805.135541] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9aa7b32b-c3bd-4e1b-9702-2a15e7f8f6d7 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.155186] env[65726]: DEBUG nova.compute.provider_tree [None req-a76cfabd-af7f-4569-b968-716d45bcf7f1 tempest-InstanceActionsNegativeTestJSON-844675881 tempest-InstanceActionsNegativeTestJSON-844675881-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 805.219690] env[65726]: DEBUG oslo_vmware.api [None req-04d3888f-ac16-4533-b7db-d3c669b17255 tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Task: {'id': task-5115924, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.471853] env[65726]: DEBUG oslo_concurrency.lockutils [None req-2c0d0141-cf5e-454f-809c-529966ff3002 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Acquiring lock "refresh_cache-49bcab1f-7b64-4999-abff-37771c58a271" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 805.472317] env[65726]: DEBUG oslo_concurrency.lockutils [None req-2c0d0141-cf5e-454f-809c-529966ff3002 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Acquired lock "refresh_cache-49bcab1f-7b64-4999-abff-37771c58a271" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 805.472317] env[65726]: DEBUG nova.network.neutron [None req-2c0d0141-cf5e-454f-809c-529966ff3002 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: 49bcab1f-7b64-4999-abff-37771c58a271] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 805.552784] env[65726]: DEBUG oslo_vmware.api [None req-2135c7ab-2ff3-4310-b64c-188bebed565e tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5115923, 'name': Destroy_Task} progress is 100%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.631792] env[65726]: INFO nova.compute.manager [-] [instance: 621b899a-e9ec-4fc0-a574-4a08cecff6d9] Took 1.32 seconds to deallocate network for instance. [ 805.656101] env[65726]: DEBUG nova.compute.manager [req-948320b2-f874-4d43-931c-aa6b6a7ae00d req-596be536-d18d-4ba4-8f76-b5548892e4d6 service nova] [instance: bc2c12e0-0d06-432f-b42f-be468e3b6ee4] Received event network-changed-a1c6ed0b-f1f6-4da3-aeaf-9f815ef90b64 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 805.657032] env[65726]: DEBUG nova.compute.manager [req-948320b2-f874-4d43-931c-aa6b6a7ae00d req-596be536-d18d-4ba4-8f76-b5548892e4d6 service nova] [instance: bc2c12e0-0d06-432f-b42f-be468e3b6ee4] Refreshing instance network info cache due to event network-changed-a1c6ed0b-f1f6-4da3-aeaf-9f815ef90b64. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 805.657032] env[65726]: DEBUG oslo_concurrency.lockutils [req-948320b2-f874-4d43-931c-aa6b6a7ae00d req-596be536-d18d-4ba4-8f76-b5548892e4d6 service nova] Acquiring lock "refresh_cache-bc2c12e0-0d06-432f-b42f-be468e3b6ee4" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 805.657032] env[65726]: DEBUG oslo_concurrency.lockutils [req-948320b2-f874-4d43-931c-aa6b6a7ae00d req-596be536-d18d-4ba4-8f76-b5548892e4d6 service nova] Acquired lock "refresh_cache-bc2c12e0-0d06-432f-b42f-be468e3b6ee4" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 805.657032] env[65726]: DEBUG nova.network.neutron [req-948320b2-f874-4d43-931c-aa6b6a7ae00d req-596be536-d18d-4ba4-8f76-b5548892e4d6 service nova] [instance: bc2c12e0-0d06-432f-b42f-be468e3b6ee4] Refreshing network info cache for port a1c6ed0b-f1f6-4da3-aeaf-9f815ef90b64 {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 805.659170] env[65726]: DEBUG nova.scheduler.client.report [None req-a76cfabd-af7f-4569-b968-716d45bcf7f1 tempest-InstanceActionsNegativeTestJSON-844675881 tempest-InstanceActionsNegativeTestJSON-844675881-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 805.718625] env[65726]: DEBUG oslo_vmware.api [None req-04d3888f-ac16-4533-b7db-d3c669b17255 tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Task: {'id': task-5115924, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.793791} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 805.718625] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-04d3888f-ac16-4533-b7db-d3c669b17255 tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] bc2c12e0-0d06-432f-b42f-be468e3b6ee4/bc2c12e0-0d06-432f-b42f-be468e3b6ee4.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 805.718805] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-04d3888f-ac16-4533-b7db-d3c669b17255 tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] [instance: bc2c12e0-0d06-432f-b42f-be468e3b6ee4] Extending root virtual disk to 1048576 {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 805.719202] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d32cfaa1-0d2a-416a-aa84-79fafad3ff38 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.726778] env[65726]: DEBUG oslo_vmware.api [None req-04d3888f-ac16-4533-b7db-d3c669b17255 tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Waiting for the task: (returnval){ [ 805.726778] env[65726]: value = "task-5115925" [ 805.726778] env[65726]: _type = "Task" [ 805.726778] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.736404] env[65726]: DEBUG oslo_vmware.api [None req-04d3888f-ac16-4533-b7db-d3c669b17255 tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Task: {'id': task-5115925, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.972109] env[65726]: DEBUG nova.compute.manager [req-76b20960-676d-46ee-b4a4-3cec01289e6e req-a16abbaf-a446-4adc-9811-29e7f3f67506 service nova] [instance: 49bcab1f-7b64-4999-abff-37771c58a271] Received event network-vif-plugged-f2a8ac3c-ad7e-4eb4-aec3-db29631fd7ad {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 805.972361] env[65726]: DEBUG oslo_concurrency.lockutils [req-76b20960-676d-46ee-b4a4-3cec01289e6e req-a16abbaf-a446-4adc-9811-29e7f3f67506 service nova] Acquiring lock "49bcab1f-7b64-4999-abff-37771c58a271-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 805.973026] env[65726]: DEBUG oslo_concurrency.lockutils [req-76b20960-676d-46ee-b4a4-3cec01289e6e req-a16abbaf-a446-4adc-9811-29e7f3f67506 service nova] Lock "49bcab1f-7b64-4999-abff-37771c58a271-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 805.973026] env[65726]: DEBUG oslo_concurrency.lockutils [req-76b20960-676d-46ee-b4a4-3cec01289e6e req-a16abbaf-a446-4adc-9811-29e7f3f67506 service nova] Lock "49bcab1f-7b64-4999-abff-37771c58a271-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 805.973026] env[65726]: DEBUG nova.compute.manager [req-76b20960-676d-46ee-b4a4-3cec01289e6e req-a16abbaf-a446-4adc-9811-29e7f3f67506 service nova] [instance: 49bcab1f-7b64-4999-abff-37771c58a271] No waiting events found dispatching network-vif-plugged-f2a8ac3c-ad7e-4eb4-aec3-db29631fd7ad {{(pid=65726) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 805.974145] env[65726]: WARNING nova.compute.manager [req-76b20960-676d-46ee-b4a4-3cec01289e6e req-a16abbaf-a446-4adc-9811-29e7f3f67506 service nova] [instance: 49bcab1f-7b64-4999-abff-37771c58a271] Received unexpected event network-vif-plugged-f2a8ac3c-ad7e-4eb4-aec3-db29631fd7ad for instance with vm_state active and task_state None. [ 805.976711] env[65726]: WARNING openstack [None req-2c0d0141-cf5e-454f-809c-529966ff3002 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 805.977158] env[65726]: WARNING openstack [None req-2c0d0141-cf5e-454f-809c-529966ff3002 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 806.019757] env[65726]: WARNING nova.network.neutron [None req-2c0d0141-cf5e-454f-809c-529966ff3002 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: 49bcab1f-7b64-4999-abff-37771c58a271] 51a66448-0f72-4266-8868-856b48be839f already exists in list: networks containing: ['51a66448-0f72-4266-8868-856b48be839f']. ignoring it [ 806.052134] env[65726]: DEBUG oslo_vmware.api [None req-2135c7ab-2ff3-4310-b64c-188bebed565e tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5115923, 'name': Destroy_Task} progress is 100%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.058723] env[65726]: WARNING openstack [None req-2c0d0141-cf5e-454f-809c-529966ff3002 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 806.059227] env[65726]: WARNING openstack [None req-2c0d0141-cf5e-454f-809c-529966ff3002 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 806.142508] env[65726]: DEBUG oslo_concurrency.lockutils [None req-861b1b19-8cf2-4cb5-a3b7-93ab69618cb7 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 806.149689] env[65726]: WARNING openstack [None req-2c0d0141-cf5e-454f-809c-529966ff3002 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 806.150237] env[65726]: WARNING openstack [None req-2c0d0141-cf5e-454f-809c-529966ff3002 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 806.163643] env[65726]: WARNING openstack [req-948320b2-f874-4d43-931c-aa6b6a7ae00d req-596be536-d18d-4ba4-8f76-b5548892e4d6 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 806.164018] env[65726]: WARNING openstack [req-948320b2-f874-4d43-931c-aa6b6a7ae00d req-596be536-d18d-4ba4-8f76-b5548892e4d6 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 806.171432] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a76cfabd-af7f-4569-b968-716d45bcf7f1 tempest-InstanceActionsNegativeTestJSON-844675881 tempest-InstanceActionsNegativeTestJSON-844675881-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.671s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 806.171584] env[65726]: DEBUG nova.compute.manager [None req-a76cfabd-af7f-4569-b968-716d45bcf7f1 tempest-InstanceActionsNegativeTestJSON-844675881 tempest-InstanceActionsNegativeTestJSON-844675881-project-member] [instance: 1bd26aef-995e-43b6-af9f-077e878c8c44] Start building networks asynchronously for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 806.174336] env[65726]: DEBUG oslo_concurrency.lockutils [None req-855996fc-ecea-49d8-a71b-8b4307731732 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 39.878s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 806.176357] env[65726]: INFO nova.compute.claims [None req-855996fc-ecea-49d8-a71b-8b4307731732 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] [instance: 6c8db442-daaa-4eca-ae24-b9d4afaf3a71] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 806.243676] env[65726]: DEBUG oslo_vmware.api [None req-04d3888f-ac16-4533-b7db-d3c669b17255 tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Task: {'id': task-5115925, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.071544} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 806.244157] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-04d3888f-ac16-4533-b7db-d3c669b17255 tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] [instance: bc2c12e0-0d06-432f-b42f-be468e3b6ee4] Extended root virtual disk {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 806.245264] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15ab61bf-ec7f-4aef-b4c0-2376582fb2ed {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.271628] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-04d3888f-ac16-4533-b7db-d3c669b17255 tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] [instance: bc2c12e0-0d06-432f-b42f-be468e3b6ee4] Reconfiguring VM instance instance-00000030 to attach disk [datastore1] bc2c12e0-0d06-432f-b42f-be468e3b6ee4/bc2c12e0-0d06-432f-b42f-be468e3b6ee4.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 806.273984] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2e7290b9-c93e-4260-b4bb-8ed714391f8f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.306309] env[65726]: DEBUG oslo_vmware.api [None req-04d3888f-ac16-4533-b7db-d3c669b17255 tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Waiting for the task: (returnval){ [ 806.306309] env[65726]: value = "task-5115926" [ 806.306309] env[65726]: _type = "Task" [ 806.306309] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 806.316651] env[65726]: DEBUG oslo_vmware.api [None req-04d3888f-ac16-4533-b7db-d3c669b17255 tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Task: {'id': task-5115926, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.321815] env[65726]: WARNING openstack [None req-2c0d0141-cf5e-454f-809c-529966ff3002 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 806.322361] env[65726]: WARNING openstack [None req-2c0d0141-cf5e-454f-809c-529966ff3002 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 806.334053] env[65726]: WARNING openstack [req-948320b2-f874-4d43-931c-aa6b6a7ae00d req-596be536-d18d-4ba4-8f76-b5548892e4d6 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 806.335059] env[65726]: WARNING openstack [req-948320b2-f874-4d43-931c-aa6b6a7ae00d req-596be536-d18d-4ba4-8f76-b5548892e4d6 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 806.446766] env[65726]: WARNING openstack [None req-2c0d0141-cf5e-454f-809c-529966ff3002 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 806.447191] env[65726]: WARNING openstack [None req-2c0d0141-cf5e-454f-809c-529966ff3002 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 806.478107] env[65726]: WARNING openstack [req-948320b2-f874-4d43-931c-aa6b6a7ae00d req-596be536-d18d-4ba4-8f76-b5548892e4d6 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 806.478548] env[65726]: WARNING openstack [req-948320b2-f874-4d43-931c-aa6b6a7ae00d req-596be536-d18d-4ba4-8f76-b5548892e4d6 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 806.555040] env[65726]: DEBUG oslo_vmware.api [None req-2135c7ab-2ff3-4310-b64c-188bebed565e tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5115923, 'name': Destroy_Task, 'duration_secs': 1.664865} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 806.555040] env[65726]: INFO nova.virt.vmwareapi.vm_util [None req-2135c7ab-2ff3-4310-b64c-188bebed565e tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 608cc0ea-3f6b-4b4f-83c1-01aa50999d63] Destroyed the VM [ 806.555040] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-2135c7ab-2ff3-4310-b64c-188bebed565e tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 608cc0ea-3f6b-4b4f-83c1-01aa50999d63] Deleting Snapshot of the VM instance {{(pid=65726) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 806.555040] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-b0f192f4-055a-4e11-9eb7-8236c7922f68 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.564342] env[65726]: DEBUG oslo_vmware.api [None req-2135c7ab-2ff3-4310-b64c-188bebed565e tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Waiting for the task: (returnval){ [ 806.564342] env[65726]: value = "task-5115927" [ 806.564342] env[65726]: _type = "Task" [ 806.564342] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 806.577403] env[65726]: DEBUG oslo_vmware.api [None req-2135c7ab-2ff3-4310-b64c-188bebed565e tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5115927, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.609414] env[65726]: DEBUG nova.network.neutron [None req-2c0d0141-cf5e-454f-809c-529966ff3002 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: 49bcab1f-7b64-4999-abff-37771c58a271] Updating instance_info_cache with network_info: [{"id": "45612b7e-c052-4346-a779-5f94f9adc084", "address": "fa:16:3e:9d:95:60", "network": {"id": "51a66448-0f72-4266-8868-856b48be839f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-155658961-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.161", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d46ccb00794f458b85da4a93879139ee", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b1f3e6c3-5584-4852-9017-476ab8ac4946", "external-id": "nsx-vlan-transportzone-304", "segmentation_id": 304, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap45612b7e-c0", "ovs_interfaceid": "45612b7e-c052-4346-a779-5f94f9adc084", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "f2a8ac3c-ad7e-4eb4-aec3-db29631fd7ad", "address": "fa:16:3e:22:ca:22", "network": {"id": "51a66448-0f72-4266-8868-856b48be839f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-155658961-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d46ccb00794f458b85da4a93879139ee", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b1f3e6c3-5584-4852-9017-476ab8ac4946", "external-id": "nsx-vlan-transportzone-304", "segmentation_id": 304, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf2a8ac3c-ad", "ovs_interfaceid": "f2a8ac3c-ad7e-4eb4-aec3-db29631fd7ad", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 806.642322] env[65726]: DEBUG nova.network.neutron [req-948320b2-f874-4d43-931c-aa6b6a7ae00d req-596be536-d18d-4ba4-8f76-b5548892e4d6 service nova] [instance: bc2c12e0-0d06-432f-b42f-be468e3b6ee4] Updated VIF entry in instance network info cache for port a1c6ed0b-f1f6-4da3-aeaf-9f815ef90b64. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 806.642684] env[65726]: DEBUG nova.network.neutron [req-948320b2-f874-4d43-931c-aa6b6a7ae00d req-596be536-d18d-4ba4-8f76-b5548892e4d6 service nova] [instance: bc2c12e0-0d06-432f-b42f-be468e3b6ee4] Updating instance_info_cache with network_info: [{"id": "a1c6ed0b-f1f6-4da3-aeaf-9f815ef90b64", "address": "fa:16:3e:09:eb:ee", "network": {"id": "dd4bbaef-0a1e-4e00-b65a-f19e7a837cc1", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-255385230-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "55f93cdf51e84cf683879668c2acecf4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "04ccbc7a-cf8d-4ea2-8411-291a1e27df7b", "external-id": "nsx-vlan-transportzone-998", "segmentation_id": 998, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa1c6ed0b-f1", "ovs_interfaceid": "a1c6ed0b-f1f6-4da3-aeaf-9f815ef90b64", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 806.686950] env[65726]: DEBUG nova.compute.utils [None req-a76cfabd-af7f-4569-b968-716d45bcf7f1 tempest-InstanceActionsNegativeTestJSON-844675881 tempest-InstanceActionsNegativeTestJSON-844675881-project-member] Using /dev/sd instead of None {{(pid=65726) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 806.687880] env[65726]: DEBUG nova.compute.manager [None req-a76cfabd-af7f-4569-b968-716d45bcf7f1 tempest-InstanceActionsNegativeTestJSON-844675881 tempest-InstanceActionsNegativeTestJSON-844675881-project-member] [instance: 1bd26aef-995e-43b6-af9f-077e878c8c44] Allocating IP information in the background. {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 806.688140] env[65726]: DEBUG nova.network.neutron [None req-a76cfabd-af7f-4569-b968-716d45bcf7f1 tempest-InstanceActionsNegativeTestJSON-844675881 tempest-InstanceActionsNegativeTestJSON-844675881-project-member] [instance: 1bd26aef-995e-43b6-af9f-077e878c8c44] allocate_for_instance() {{(pid=65726) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 806.688544] env[65726]: WARNING neutronclient.v2_0.client [None req-a76cfabd-af7f-4569-b968-716d45bcf7f1 tempest-InstanceActionsNegativeTestJSON-844675881 tempest-InstanceActionsNegativeTestJSON-844675881-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 806.688980] env[65726]: WARNING neutronclient.v2_0.client [None req-a76cfabd-af7f-4569-b968-716d45bcf7f1 tempest-InstanceActionsNegativeTestJSON-844675881 tempest-InstanceActionsNegativeTestJSON-844675881-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 806.689781] env[65726]: WARNING openstack [None req-a76cfabd-af7f-4569-b968-716d45bcf7f1 tempest-InstanceActionsNegativeTestJSON-844675881 tempest-InstanceActionsNegativeTestJSON-844675881-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 806.690167] env[65726]: WARNING openstack [None req-a76cfabd-af7f-4569-b968-716d45bcf7f1 tempest-InstanceActionsNegativeTestJSON-844675881 tempest-InstanceActionsNegativeTestJSON-844675881-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 806.751899] env[65726]: DEBUG nova.policy [None req-a76cfabd-af7f-4569-b968-716d45bcf7f1 tempest-InstanceActionsNegativeTestJSON-844675881 tempest-InstanceActionsNegativeTestJSON-844675881-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9751cd7ebb36402ba57447ae51392920', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'faa59ae0ed7f4c86939a84de1e3672e9', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65726) authorize /opt/stack/nova/nova/policy.py:192}} [ 806.818666] env[65726]: DEBUG oslo_vmware.api [None req-04d3888f-ac16-4533-b7db-d3c669b17255 tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Task: {'id': task-5115926, 'name': ReconfigVM_Task, 'duration_secs': 0.298877} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 806.818955] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-04d3888f-ac16-4533-b7db-d3c669b17255 tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] [instance: bc2c12e0-0d06-432f-b42f-be468e3b6ee4] Reconfigured VM instance instance-00000030 to attach disk [datastore1] bc2c12e0-0d06-432f-b42f-be468e3b6ee4/bc2c12e0-0d06-432f-b42f-be468e3b6ee4.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 806.819658] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7d64aae5-3625-46f5-bf58-d3fc45d53288 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.827384] env[65726]: DEBUG oslo_vmware.api [None req-04d3888f-ac16-4533-b7db-d3c669b17255 tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Waiting for the task: (returnval){ [ 806.827384] env[65726]: value = "task-5115928" [ 806.827384] env[65726]: _type = "Task" [ 806.827384] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 806.837553] env[65726]: DEBUG oslo_vmware.api [None req-04d3888f-ac16-4533-b7db-d3c669b17255 tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Task: {'id': task-5115928, 'name': Rename_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.844647] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b3ea747b-af8b-4592-b662-6643bc7f39d3 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] Acquiring lock "2cbee359-a1fc-45c0-bcc5-a2c24ded9c69" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 806.844902] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b3ea747b-af8b-4592-b662-6643bc7f39d3 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] Lock "2cbee359-a1fc-45c0-bcc5-a2c24ded9c69" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 807.049744] env[65726]: DEBUG nova.network.neutron [None req-a76cfabd-af7f-4569-b968-716d45bcf7f1 tempest-InstanceActionsNegativeTestJSON-844675881 tempest-InstanceActionsNegativeTestJSON-844675881-project-member] [instance: 1bd26aef-995e-43b6-af9f-077e878c8c44] Successfully created port: 61bb4097-108f-4424-9e34-38ed6f561f0c {{(pid=65726) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 807.076857] env[65726]: DEBUG oslo_vmware.api [None req-2135c7ab-2ff3-4310-b64c-188bebed565e tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5115927, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.112766] env[65726]: DEBUG oslo_concurrency.lockutils [None req-2c0d0141-cf5e-454f-809c-529966ff3002 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Releasing lock "refresh_cache-49bcab1f-7b64-4999-abff-37771c58a271" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 807.113720] env[65726]: DEBUG oslo_concurrency.lockutils [None req-2c0d0141-cf5e-454f-809c-529966ff3002 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Acquiring lock "49bcab1f-7b64-4999-abff-37771c58a271" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 807.113878] env[65726]: DEBUG oslo_concurrency.lockutils [None req-2c0d0141-cf5e-454f-809c-529966ff3002 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Acquired lock "49bcab1f-7b64-4999-abff-37771c58a271" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 807.114744] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-078eeb61-e15e-462c-baa1-b5760e5e6a07 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.135529] env[65726]: DEBUG nova.virt.hardware [None req-2c0d0141-cf5e-454f-809c-529966ff3002 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 807.135780] env[65726]: DEBUG nova.virt.hardware [None req-2c0d0141-cf5e-454f-809c-529966ff3002 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 807.135924] env[65726]: DEBUG nova.virt.hardware [None req-2c0d0141-cf5e-454f-809c-529966ff3002 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 807.136127] env[65726]: DEBUG nova.virt.hardware [None req-2c0d0141-cf5e-454f-809c-529966ff3002 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 807.136272] env[65726]: DEBUG nova.virt.hardware [None req-2c0d0141-cf5e-454f-809c-529966ff3002 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 807.136778] env[65726]: DEBUG nova.virt.hardware [None req-2c0d0141-cf5e-454f-809c-529966ff3002 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 807.136778] env[65726]: DEBUG nova.virt.hardware [None req-2c0d0141-cf5e-454f-809c-529966ff3002 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 807.136942] env[65726]: DEBUG nova.virt.hardware [None req-2c0d0141-cf5e-454f-809c-529966ff3002 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 807.137083] env[65726]: DEBUG nova.virt.hardware [None req-2c0d0141-cf5e-454f-809c-529966ff3002 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 807.137251] env[65726]: DEBUG nova.virt.hardware [None req-2c0d0141-cf5e-454f-809c-529966ff3002 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 807.137431] env[65726]: DEBUG nova.virt.hardware [None req-2c0d0141-cf5e-454f-809c-529966ff3002 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 807.144890] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-2c0d0141-cf5e-454f-809c-529966ff3002 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: 49bcab1f-7b64-4999-abff-37771c58a271] Reconfiguring VM to attach interface {{(pid=65726) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 807.148884] env[65726]: DEBUG oslo_concurrency.lockutils [req-948320b2-f874-4d43-931c-aa6b6a7ae00d req-596be536-d18d-4ba4-8f76-b5548892e4d6 service nova] Releasing lock "refresh_cache-bc2c12e0-0d06-432f-b42f-be468e3b6ee4" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 807.154485] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f27d5d39-0aae-462a-9c26-883d2bd4e881 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.176343] env[65726]: DEBUG oslo_vmware.api [None req-2c0d0141-cf5e-454f-809c-529966ff3002 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Waiting for the task: (returnval){ [ 807.176343] env[65726]: value = "task-5115929" [ 807.176343] env[65726]: _type = "Task" [ 807.176343] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.185810] env[65726]: DEBUG oslo_vmware.api [None req-2c0d0141-cf5e-454f-809c-529966ff3002 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': task-5115929, 'name': ReconfigVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.199977] env[65726]: DEBUG nova.compute.manager [None req-a76cfabd-af7f-4569-b968-716d45bcf7f1 tempest-InstanceActionsNegativeTestJSON-844675881 tempest-InstanceActionsNegativeTestJSON-844675881-project-member] [instance: 1bd26aef-995e-43b6-af9f-077e878c8c44] Start building block device mappings for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 807.339585] env[65726]: DEBUG oslo_vmware.api [None req-04d3888f-ac16-4533-b7db-d3c669b17255 tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Task: {'id': task-5115928, 'name': Rename_Task, 'duration_secs': 0.169331} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 807.339892] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-04d3888f-ac16-4533-b7db-d3c669b17255 tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] [instance: bc2c12e0-0d06-432f-b42f-be468e3b6ee4] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 807.340215] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ff203feb-55f4-4957-b86b-afae631ec0ed {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.348800] env[65726]: DEBUG oslo_vmware.api [None req-04d3888f-ac16-4533-b7db-d3c669b17255 tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Waiting for the task: (returnval){ [ 807.348800] env[65726]: value = "task-5115930" [ 807.348800] env[65726]: _type = "Task" [ 807.348800] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.367858] env[65726]: DEBUG oslo_vmware.api [None req-04d3888f-ac16-4533-b7db-d3c669b17255 tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Task: {'id': task-5115930, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.576808] env[65726]: DEBUG oslo_vmware.api [None req-2135c7ab-2ff3-4310-b64c-188bebed565e tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5115927, 'name': RemoveSnapshot_Task, 'duration_secs': 0.792601} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 807.580550] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-2135c7ab-2ff3-4310-b64c-188bebed565e tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 608cc0ea-3f6b-4b4f-83c1-01aa50999d63] Deleted Snapshot of the VM instance {{(pid=65726) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 807.580890] env[65726]: INFO nova.compute.manager [None req-2135c7ab-2ff3-4310-b64c-188bebed565e tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 608cc0ea-3f6b-4b4f-83c1-01aa50999d63] Took 17.37 seconds to snapshot the instance on the hypervisor. [ 807.689906] env[65726]: DEBUG oslo_vmware.api [None req-2c0d0141-cf5e-454f-809c-529966ff3002 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': task-5115929, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.711158] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85a76a40-368b-4600-93fa-464d8b46c5dc {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.718092] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b820a033-307a-44f5-a8ff-2e106e44048d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.757966] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac6c5fea-9660-4be7-8659-061bdcca5fdf {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.769184] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-041bbd24-eb5a-48f0-bb0c-ac037154e704 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.784873] env[65726]: DEBUG nova.compute.provider_tree [None req-855996fc-ecea-49d8-a71b-8b4307731732 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 807.863031] env[65726]: DEBUG oslo_vmware.api [None req-04d3888f-ac16-4533-b7db-d3c669b17255 tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Task: {'id': task-5115930, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.101846] env[65726]: INFO nova.compute.manager [None req-c9bfb110-bdb5-4ce1-88bb-ec98452720c3 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: 6d2b86c2-0a44-433f-8b3f-2e9c7f693baa] Unrescuing [ 808.102412] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c9bfb110-bdb5-4ce1-88bb-ec98452720c3 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Acquiring lock "refresh_cache-6d2b86c2-0a44-433f-8b3f-2e9c7f693baa" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 808.102544] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c9bfb110-bdb5-4ce1-88bb-ec98452720c3 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Acquired lock "refresh_cache-6d2b86c2-0a44-433f-8b3f-2e9c7f693baa" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 808.102779] env[65726]: DEBUG nova.network.neutron [None req-c9bfb110-bdb5-4ce1-88bb-ec98452720c3 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: 6d2b86c2-0a44-433f-8b3f-2e9c7f693baa] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 808.146347] env[65726]: DEBUG nova.compute.manager [None req-2135c7ab-2ff3-4310-b64c-188bebed565e tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 608cc0ea-3f6b-4b4f-83c1-01aa50999d63] Found 2 images (rotation: 2) {{(pid=65726) _rotate_backups /opt/stack/nova/nova/compute/manager.py:5074}} [ 808.192063] env[65726]: DEBUG oslo_vmware.api [None req-2c0d0141-cf5e-454f-809c-529966ff3002 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': task-5115929, 'name': ReconfigVM_Task, 'duration_secs': 0.86176} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 808.196286] env[65726]: WARNING openstack [None req-2c0d0141-cf5e-454f-809c-529966ff3002 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 808.196663] env[65726]: WARNING openstack [None req-2c0d0141-cf5e-454f-809c-529966ff3002 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 808.204068] env[65726]: DEBUG oslo_concurrency.lockutils [None req-2c0d0141-cf5e-454f-809c-529966ff3002 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Releasing lock "49bcab1f-7b64-4999-abff-37771c58a271" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 808.204354] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-2c0d0141-cf5e-454f-809c-529966ff3002 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: 49bcab1f-7b64-4999-abff-37771c58a271] Reconfigured VM to attach interface {{(pid=65726) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 808.209462] env[65726]: DEBUG nova.compute.manager [None req-a76cfabd-af7f-4569-b968-716d45bcf7f1 tempest-InstanceActionsNegativeTestJSON-844675881 tempest-InstanceActionsNegativeTestJSON-844675881-project-member] [instance: 1bd26aef-995e-43b6-af9f-077e878c8c44] Start spawning the instance on the hypervisor. {{(pid=65726) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 808.240436] env[65726]: DEBUG nova.virt.hardware [None req-a76cfabd-af7f-4569-b968-716d45bcf7f1 tempest-InstanceActionsNegativeTestJSON-844675881 tempest-InstanceActionsNegativeTestJSON-844675881-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T19:28:34Z,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1539f06fa3534e90acbc3a60c4b8bd3f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T19:28:35Z,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 808.240687] env[65726]: DEBUG nova.virt.hardware [None req-a76cfabd-af7f-4569-b968-716d45bcf7f1 tempest-InstanceActionsNegativeTestJSON-844675881 tempest-InstanceActionsNegativeTestJSON-844675881-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 808.240839] env[65726]: DEBUG nova.virt.hardware [None req-a76cfabd-af7f-4569-b968-716d45bcf7f1 tempest-InstanceActionsNegativeTestJSON-844675881 tempest-InstanceActionsNegativeTestJSON-844675881-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 808.241167] env[65726]: DEBUG nova.virt.hardware [None req-a76cfabd-af7f-4569-b968-716d45bcf7f1 tempest-InstanceActionsNegativeTestJSON-844675881 tempest-InstanceActionsNegativeTestJSON-844675881-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 808.241167] env[65726]: DEBUG nova.virt.hardware [None req-a76cfabd-af7f-4569-b968-716d45bcf7f1 tempest-InstanceActionsNegativeTestJSON-844675881 tempest-InstanceActionsNegativeTestJSON-844675881-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 808.241302] env[65726]: DEBUG nova.virt.hardware [None req-a76cfabd-af7f-4569-b968-716d45bcf7f1 tempest-InstanceActionsNegativeTestJSON-844675881 tempest-InstanceActionsNegativeTestJSON-844675881-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 808.241503] env[65726]: DEBUG nova.virt.hardware [None req-a76cfabd-af7f-4569-b968-716d45bcf7f1 tempest-InstanceActionsNegativeTestJSON-844675881 tempest-InstanceActionsNegativeTestJSON-844675881-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 808.241658] env[65726]: DEBUG nova.virt.hardware [None req-a76cfabd-af7f-4569-b968-716d45bcf7f1 tempest-InstanceActionsNegativeTestJSON-844675881 tempest-InstanceActionsNegativeTestJSON-844675881-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 808.241822] env[65726]: DEBUG nova.virt.hardware [None req-a76cfabd-af7f-4569-b968-716d45bcf7f1 tempest-InstanceActionsNegativeTestJSON-844675881 tempest-InstanceActionsNegativeTestJSON-844675881-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 808.241982] env[65726]: DEBUG nova.virt.hardware [None req-a76cfabd-af7f-4569-b968-716d45bcf7f1 tempest-InstanceActionsNegativeTestJSON-844675881 tempest-InstanceActionsNegativeTestJSON-844675881-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 808.242165] env[65726]: DEBUG nova.virt.hardware [None req-a76cfabd-af7f-4569-b968-716d45bcf7f1 tempest-InstanceActionsNegativeTestJSON-844675881 tempest-InstanceActionsNegativeTestJSON-844675881-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 808.243361] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-117d7c25-e866-4226-b234-c612c6954f11 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.251860] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53f86c2a-3a0a-4dd9-bc8d-2cfb9fb2e0e0 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.288090] env[65726]: DEBUG nova.scheduler.client.report [None req-855996fc-ecea-49d8-a71b-8b4307731732 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 808.362621] env[65726]: DEBUG oslo_vmware.api [None req-04d3888f-ac16-4533-b7db-d3c669b17255 tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Task: {'id': task-5115930, 'name': PowerOnVM_Task, 'duration_secs': 0.524787} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 808.363299] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-04d3888f-ac16-4533-b7db-d3c669b17255 tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] [instance: bc2c12e0-0d06-432f-b42f-be468e3b6ee4] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 808.366282] env[65726]: INFO nova.compute.manager [None req-04d3888f-ac16-4533-b7db-d3c669b17255 tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] [instance: bc2c12e0-0d06-432f-b42f-be468e3b6ee4] Took 7.02 seconds to spawn the instance on the hypervisor. [ 808.366282] env[65726]: DEBUG nova.compute.manager [None req-04d3888f-ac16-4533-b7db-d3c669b17255 tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] [instance: bc2c12e0-0d06-432f-b42f-be468e3b6ee4] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 808.366282] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f996c303-433b-4460-a3ad-0fb8dda85c91 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.608456] env[65726]: WARNING openstack [None req-c9bfb110-bdb5-4ce1-88bb-ec98452720c3 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 808.609115] env[65726]: WARNING openstack [None req-c9bfb110-bdb5-4ce1-88bb-ec98452720c3 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 808.684883] env[65726]: DEBUG nova.network.neutron [None req-a76cfabd-af7f-4569-b968-716d45bcf7f1 tempest-InstanceActionsNegativeTestJSON-844675881 tempest-InstanceActionsNegativeTestJSON-844675881-project-member] [instance: 1bd26aef-995e-43b6-af9f-077e878c8c44] Successfully updated port: 61bb4097-108f-4424-9e34-38ed6f561f0c {{(pid=65726) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 808.714092] env[65726]: DEBUG oslo_concurrency.lockutils [None req-2c0d0141-cf5e-454f-809c-529966ff3002 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Lock "interface-49bcab1f-7b64-4999-abff-37771c58a271-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 6.863s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 808.737798] env[65726]: WARNING openstack [None req-c9bfb110-bdb5-4ce1-88bb-ec98452720c3 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 808.738244] env[65726]: WARNING openstack [None req-c9bfb110-bdb5-4ce1-88bb-ec98452720c3 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 808.796022] env[65726]: DEBUG oslo_concurrency.lockutils [None req-855996fc-ecea-49d8-a71b-8b4307731732 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.620s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 808.796022] env[65726]: DEBUG nova.compute.manager [None req-855996fc-ecea-49d8-a71b-8b4307731732 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] [instance: 6c8db442-daaa-4eca-ae24-b9d4afaf3a71] Start building networks asynchronously for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 808.799308] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9ac23e9e-c303-4029-8e2c-3f5115af48ab tempest-ServerAddressesNegativeTestJSON-963263679 tempest-ServerAddressesNegativeTestJSON-963263679-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 35.812s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 808.799590] env[65726]: DEBUG nova.objects.instance [None req-9ac23e9e-c303-4029-8e2c-3f5115af48ab tempest-ServerAddressesNegativeTestJSON-963263679 tempest-ServerAddressesNegativeTestJSON-963263679-project-member] Lazy-loading 'resources' on Instance uuid 00746508-d0d6-4dfa-9026-772398d004ee {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 808.820873] env[65726]: WARNING openstack [None req-c9bfb110-bdb5-4ce1-88bb-ec98452720c3 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 808.822940] env[65726]: WARNING openstack [None req-c9bfb110-bdb5-4ce1-88bb-ec98452720c3 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 808.882817] env[65726]: INFO nova.compute.manager [None req-04d3888f-ac16-4533-b7db-d3c669b17255 tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] [instance: bc2c12e0-0d06-432f-b42f-be468e3b6ee4] Took 46.58 seconds to build instance. [ 808.920150] env[65726]: DEBUG nova.network.neutron [None req-c9bfb110-bdb5-4ce1-88bb-ec98452720c3 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: 6d2b86c2-0a44-433f-8b3f-2e9c7f693baa] Updating instance_info_cache with network_info: [{"id": "b487d53a-9bda-4e3c-be61-f26acbc8ae0f", "address": "fa:16:3e:ad:db:36", "network": {"id": "9d5ad167-99fb-4208-ab30-7053f35774e4", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-2082516762-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a8362295a62f4b51bae719a7ef5a4656", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dabbac20-1723-40ad-9da0-e53b28073651", "external-id": "nsx-vlan-transportzone-790", "segmentation_id": 790, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb487d53a-9b", "ovs_interfaceid": "b487d53a-9bda-4e3c-be61-f26acbc8ae0f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 809.187843] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a76cfabd-af7f-4569-b968-716d45bcf7f1 tempest-InstanceActionsNegativeTestJSON-844675881 tempest-InstanceActionsNegativeTestJSON-844675881-project-member] Acquiring lock "refresh_cache-1bd26aef-995e-43b6-af9f-077e878c8c44" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 809.188192] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a76cfabd-af7f-4569-b968-716d45bcf7f1 tempest-InstanceActionsNegativeTestJSON-844675881 tempest-InstanceActionsNegativeTestJSON-844675881-project-member] Acquired lock "refresh_cache-1bd26aef-995e-43b6-af9f-077e878c8c44" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 809.188306] env[65726]: DEBUG nova.network.neutron [None req-a76cfabd-af7f-4569-b968-716d45bcf7f1 tempest-InstanceActionsNegativeTestJSON-844675881 tempest-InstanceActionsNegativeTestJSON-844675881-project-member] [instance: 1bd26aef-995e-43b6-af9f-077e878c8c44] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 809.302076] env[65726]: DEBUG nova.compute.utils [None req-855996fc-ecea-49d8-a71b-8b4307731732 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Using /dev/sd instead of None {{(pid=65726) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 809.303764] env[65726]: DEBUG nova.compute.manager [None req-855996fc-ecea-49d8-a71b-8b4307731732 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] [instance: 6c8db442-daaa-4eca-ae24-b9d4afaf3a71] Allocating IP information in the background. {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 809.303941] env[65726]: DEBUG nova.network.neutron [None req-855996fc-ecea-49d8-a71b-8b4307731732 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] [instance: 6c8db442-daaa-4eca-ae24-b9d4afaf3a71] allocate_for_instance() {{(pid=65726) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 809.304330] env[65726]: WARNING neutronclient.v2_0.client [None req-855996fc-ecea-49d8-a71b-8b4307731732 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 809.305057] env[65726]: WARNING neutronclient.v2_0.client [None req-855996fc-ecea-49d8-a71b-8b4307731732 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 809.305672] env[65726]: WARNING openstack [None req-855996fc-ecea-49d8-a71b-8b4307731732 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 809.306025] env[65726]: WARNING openstack [None req-855996fc-ecea-49d8-a71b-8b4307731732 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 809.361092] env[65726]: DEBUG nova.policy [None req-855996fc-ecea-49d8-a71b-8b4307731732 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'cea37b69d6cc4784a4c6a8889a718bfd', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '79d3e960f190439fae7c61d693e9356e', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65726) authorize /opt/stack/nova/nova/policy.py:192}} [ 809.384447] env[65726]: DEBUG oslo_concurrency.lockutils [None req-04d3888f-ac16-4533-b7db-d3c669b17255 tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Lock "bc2c12e0-0d06-432f-b42f-be468e3b6ee4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 79.802s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 809.423861] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c9bfb110-bdb5-4ce1-88bb-ec98452720c3 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Releasing lock "refresh_cache-6d2b86c2-0a44-433f-8b3f-2e9c7f693baa" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 809.424596] env[65726]: DEBUG nova.objects.instance [None req-c9bfb110-bdb5-4ce1-88bb-ec98452720c3 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Lazy-loading 'flavor' on Instance uuid 6d2b86c2-0a44-433f-8b3f-2e9c7f693baa {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 809.683779] env[65726]: DEBUG nova.network.neutron [None req-855996fc-ecea-49d8-a71b-8b4307731732 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] [instance: 6c8db442-daaa-4eca-ae24-b9d4afaf3a71] Successfully created port: 6f422d37-8262-454d-aa2d-1569bacee1b7 {{(pid=65726) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 809.691936] env[65726]: WARNING openstack [None req-a76cfabd-af7f-4569-b968-716d45bcf7f1 tempest-InstanceActionsNegativeTestJSON-844675881 tempest-InstanceActionsNegativeTestJSON-844675881-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 809.692069] env[65726]: WARNING openstack [None req-a76cfabd-af7f-4569-b968-716d45bcf7f1 tempest-InstanceActionsNegativeTestJSON-844675881 tempest-InstanceActionsNegativeTestJSON-844675881-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 809.699216] env[65726]: DEBUG nova.network.neutron [None req-a76cfabd-af7f-4569-b968-716d45bcf7f1 tempest-InstanceActionsNegativeTestJSON-844675881 tempest-InstanceActionsNegativeTestJSON-844675881-project-member] [instance: 1bd26aef-995e-43b6-af9f-077e878c8c44] Instance cache missing network info. {{(pid=65726) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 809.704314] env[65726]: DEBUG nova.compute.manager [req-1e310770-08ad-4a87-878f-0d88015cf431 req-55fccb6f-e263-4b9d-af38-6c9c566cc966 service nova] [instance: 621b899a-e9ec-4fc0-a574-4a08cecff6d9] Received event network-vif-deleted-dcdfc13e-45af-4bc2-970d-b60727f42176 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 809.783135] env[65726]: WARNING openstack [None req-a76cfabd-af7f-4569-b968-716d45bcf7f1 tempest-InstanceActionsNegativeTestJSON-844675881 tempest-InstanceActionsNegativeTestJSON-844675881-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 809.783982] env[65726]: WARNING openstack [None req-a76cfabd-af7f-4569-b968-716d45bcf7f1 tempest-InstanceActionsNegativeTestJSON-844675881 tempest-InstanceActionsNegativeTestJSON-844675881-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 809.816487] env[65726]: DEBUG nova.compute.utils [None req-855996fc-ecea-49d8-a71b-8b4307731732 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Using /dev/sd instead of None {{(pid=65726) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 809.857364] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55b93561-c279-4ddd-a473-7cdb86daff6a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.866210] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d67e8c5c-c1fb-49c0-aab1-90cb0fc08af9 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.875705] env[65726]: DEBUG nova.compute.manager [req-e6888c71-8951-4071-9e26-cd2d50c819e0 req-d6474ee1-c83a-421c-8586-6bf1a0942440 service nova] [instance: 49bcab1f-7b64-4999-abff-37771c58a271] Received event network-changed-f2a8ac3c-ad7e-4eb4-aec3-db29631fd7ad {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 809.875931] env[65726]: DEBUG nova.compute.manager [req-e6888c71-8951-4071-9e26-cd2d50c819e0 req-d6474ee1-c83a-421c-8586-6bf1a0942440 service nova] [instance: 49bcab1f-7b64-4999-abff-37771c58a271] Refreshing instance network info cache due to event network-changed-f2a8ac3c-ad7e-4eb4-aec3-db29631fd7ad. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 809.876169] env[65726]: DEBUG oslo_concurrency.lockutils [req-e6888c71-8951-4071-9e26-cd2d50c819e0 req-d6474ee1-c83a-421c-8586-6bf1a0942440 service nova] Acquiring lock "refresh_cache-49bcab1f-7b64-4999-abff-37771c58a271" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 809.876363] env[65726]: DEBUG oslo_concurrency.lockutils [req-e6888c71-8951-4071-9e26-cd2d50c819e0 req-d6474ee1-c83a-421c-8586-6bf1a0942440 service nova] Acquired lock "refresh_cache-49bcab1f-7b64-4999-abff-37771c58a271" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 809.876489] env[65726]: DEBUG nova.network.neutron [req-e6888c71-8951-4071-9e26-cd2d50c819e0 req-d6474ee1-c83a-421c-8586-6bf1a0942440 service nova] [instance: 49bcab1f-7b64-4999-abff-37771c58a271] Refreshing network info cache for port f2a8ac3c-ad7e-4eb4-aec3-db29631fd7ad {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 809.902597] env[65726]: DEBUG nova.compute.manager [None req-55d74d77-063b-418b-b818-59d05c490f34 tempest-ImagesNegativeTestJSON-1015404931 tempest-ImagesNegativeTestJSON-1015404931-project-member] [instance: c4177e20-b1bd-4b54-a275-c93582359a07] Starting instance... {{(pid=65726) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 809.906593] env[65726]: WARNING openstack [req-e6888c71-8951-4071-9e26-cd2d50c819e0 req-d6474ee1-c83a-421c-8586-6bf1a0942440 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 809.906940] env[65726]: WARNING openstack [req-e6888c71-8951-4071-9e26-cd2d50c819e0 req-d6474ee1-c83a-421c-8586-6bf1a0942440 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 809.916858] env[65726]: WARNING openstack [None req-a76cfabd-af7f-4569-b968-716d45bcf7f1 tempest-InstanceActionsNegativeTestJSON-844675881 tempest-InstanceActionsNegativeTestJSON-844675881-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 809.917233] env[65726]: WARNING openstack [None req-a76cfabd-af7f-4569-b968-716d45bcf7f1 tempest-InstanceActionsNegativeTestJSON-844675881 tempest-InstanceActionsNegativeTestJSON-844675881-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 809.925869] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41df466e-b224-4401-81c2-12a465c7a7f9 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.932928] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06e9c97c-5f44-4d1a-a694-ecb900d46210 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.942091] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86e77693-b1eb-4876-8a09-bcd816ea6d6a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.965197] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-c9bfb110-bdb5-4ce1-88bb-ec98452720c3 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: 6d2b86c2-0a44-433f-8b3f-2e9c7f693baa] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 809.966246] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-69cbf69b-b1b6-445b-8306-5d8b81efbb67 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.977117] env[65726]: DEBUG nova.compute.provider_tree [None req-9ac23e9e-c303-4029-8e2c-3f5115af48ab tempest-ServerAddressesNegativeTestJSON-963263679 tempest-ServerAddressesNegativeTestJSON-963263679-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 809.980421] env[65726]: DEBUG oslo_vmware.api [None req-c9bfb110-bdb5-4ce1-88bb-ec98452720c3 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Waiting for the task: (returnval){ [ 809.980421] env[65726]: value = "task-5115931" [ 809.980421] env[65726]: _type = "Task" [ 809.980421] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.996721] env[65726]: DEBUG oslo_vmware.api [None req-c9bfb110-bdb5-4ce1-88bb-ec98452720c3 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Task: {'id': task-5115931, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.064286] env[65726]: DEBUG nova.network.neutron [None req-a76cfabd-af7f-4569-b968-716d45bcf7f1 tempest-InstanceActionsNegativeTestJSON-844675881 tempest-InstanceActionsNegativeTestJSON-844675881-project-member] [instance: 1bd26aef-995e-43b6-af9f-077e878c8c44] Updating instance_info_cache with network_info: [{"id": "61bb4097-108f-4424-9e34-38ed6f561f0c", "address": "fa:16:3e:a6:40:8f", "network": {"id": "1efad9b9-de6f-4953-905d-49886cbc9e17", "bridge": "br-int", "label": "tempest-InstanceActionsNegativeTestJSON-637686941-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "faa59ae0ed7f4c86939a84de1e3672e9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e5d88cd9-35a3-4ac3-9d6d-756464cd6cc5", "external-id": "nsx-vlan-transportzone-685", "segmentation_id": 685, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap61bb4097-10", "ovs_interfaceid": "61bb4097-108f-4424-9e34-38ed6f561f0c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 810.213749] env[65726]: WARNING openstack [req-e6888c71-8951-4071-9e26-cd2d50c819e0 req-d6474ee1-c83a-421c-8586-6bf1a0942440 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 810.214229] env[65726]: WARNING openstack [req-e6888c71-8951-4071-9e26-cd2d50c819e0 req-d6474ee1-c83a-421c-8586-6bf1a0942440 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 810.322527] env[65726]: DEBUG nova.compute.manager [None req-855996fc-ecea-49d8-a71b-8b4307731732 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] [instance: 6c8db442-daaa-4eca-ae24-b9d4afaf3a71] Start building block device mappings for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 810.452219] env[65726]: DEBUG oslo_concurrency.lockutils [None req-55d74d77-063b-418b-b818-59d05c490f34 tempest-ImagesNegativeTestJSON-1015404931 tempest-ImagesNegativeTestJSON-1015404931-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 810.482824] env[65726]: DEBUG nova.scheduler.client.report [None req-9ac23e9e-c303-4029-8e2c-3f5115af48ab tempest-ServerAddressesNegativeTestJSON-963263679 tempest-ServerAddressesNegativeTestJSON-963263679-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 810.509886] env[65726]: DEBUG oslo_vmware.api [None req-c9bfb110-bdb5-4ce1-88bb-ec98452720c3 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Task: {'id': task-5115931, 'name': PowerOffVM_Task, 'duration_secs': 0.240982} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.510384] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-c9bfb110-bdb5-4ce1-88bb-ec98452720c3 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: 6d2b86c2-0a44-433f-8b3f-2e9c7f693baa] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 810.520583] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-c9bfb110-bdb5-4ce1-88bb-ec98452720c3 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: 6d2b86c2-0a44-433f-8b3f-2e9c7f693baa] Reconfiguring VM instance instance-0000002b to detach disk 2001 {{(pid=65726) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 810.521575] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-142b32d4-759a-429d-a3af-854cc9fb2910 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.556286] env[65726]: DEBUG oslo_vmware.api [None req-c9bfb110-bdb5-4ce1-88bb-ec98452720c3 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Waiting for the task: (returnval){ [ 810.556286] env[65726]: value = "task-5115932" [ 810.556286] env[65726]: _type = "Task" [ 810.556286] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.573905] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a76cfabd-af7f-4569-b968-716d45bcf7f1 tempest-InstanceActionsNegativeTestJSON-844675881 tempest-InstanceActionsNegativeTestJSON-844675881-project-member] Releasing lock "refresh_cache-1bd26aef-995e-43b6-af9f-077e878c8c44" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 810.574762] env[65726]: DEBUG nova.compute.manager [None req-a76cfabd-af7f-4569-b968-716d45bcf7f1 tempest-InstanceActionsNegativeTestJSON-844675881 tempest-InstanceActionsNegativeTestJSON-844675881-project-member] [instance: 1bd26aef-995e-43b6-af9f-077e878c8c44] Instance network_info: |[{"id": "61bb4097-108f-4424-9e34-38ed6f561f0c", "address": "fa:16:3e:a6:40:8f", "network": {"id": "1efad9b9-de6f-4953-905d-49886cbc9e17", "bridge": "br-int", "label": "tempest-InstanceActionsNegativeTestJSON-637686941-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "faa59ae0ed7f4c86939a84de1e3672e9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e5d88cd9-35a3-4ac3-9d6d-756464cd6cc5", "external-id": "nsx-vlan-transportzone-685", "segmentation_id": 685, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap61bb4097-10", "ovs_interfaceid": "61bb4097-108f-4424-9e34-38ed6f561f0c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 810.575314] env[65726]: DEBUG oslo_vmware.api [None req-c9bfb110-bdb5-4ce1-88bb-ec98452720c3 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Task: {'id': task-5115932, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.575931] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-a76cfabd-af7f-4569-b968-716d45bcf7f1 tempest-InstanceActionsNegativeTestJSON-844675881 tempest-InstanceActionsNegativeTestJSON-844675881-project-member] [instance: 1bd26aef-995e-43b6-af9f-077e878c8c44] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a6:40:8f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'e5d88cd9-35a3-4ac3-9d6d-756464cd6cc5', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '61bb4097-108f-4424-9e34-38ed6f561f0c', 'vif_model': 'vmxnet3'}] {{(pid=65726) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 810.592756] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-a76cfabd-af7f-4569-b968-716d45bcf7f1 tempest-InstanceActionsNegativeTestJSON-844675881 tempest-InstanceActionsNegativeTestJSON-844675881-project-member] Creating folder: Project (faa59ae0ed7f4c86939a84de1e3672e9). Parent ref: group-v995008. {{(pid=65726) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 810.593190] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9acc4cf9-93fb-481c-bc11-9aaa29b4fa68 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.607128] env[65726]: INFO nova.virt.vmwareapi.vm_util [None req-a76cfabd-af7f-4569-b968-716d45bcf7f1 tempest-InstanceActionsNegativeTestJSON-844675881 tempest-InstanceActionsNegativeTestJSON-844675881-project-member] Created folder: Project (faa59ae0ed7f4c86939a84de1e3672e9) in parent group-v995008. [ 810.607360] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-a76cfabd-af7f-4569-b968-716d45bcf7f1 tempest-InstanceActionsNegativeTestJSON-844675881 tempest-InstanceActionsNegativeTestJSON-844675881-project-member] Creating folder: Instances. Parent ref: group-v995142. {{(pid=65726) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 810.607661] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9e0fbe4a-8349-458b-ab08-397b685a802a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.620123] env[65726]: INFO nova.virt.vmwareapi.vm_util [None req-a76cfabd-af7f-4569-b968-716d45bcf7f1 tempest-InstanceActionsNegativeTestJSON-844675881 tempest-InstanceActionsNegativeTestJSON-844675881-project-member] Created folder: Instances in parent group-v995142. [ 810.620476] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-a76cfabd-af7f-4569-b968-716d45bcf7f1 tempest-InstanceActionsNegativeTestJSON-844675881 tempest-InstanceActionsNegativeTestJSON-844675881-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 810.620711] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1bd26aef-995e-43b6-af9f-077e878c8c44] Creating VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 810.620994] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-90179b9f-0dca-450c-9d11-96eba6106d05 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.642894] env[65726]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 810.642894] env[65726]: value = "task-5115935" [ 810.642894] env[65726]: _type = "Task" [ 810.642894] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.654601] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115935, 'name': CreateVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.921863] env[65726]: WARNING openstack [req-e6888c71-8951-4071-9e26-cd2d50c819e0 req-d6474ee1-c83a-421c-8586-6bf1a0942440 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 810.922685] env[65726]: WARNING openstack [req-e6888c71-8951-4071-9e26-cd2d50c819e0 req-d6474ee1-c83a-421c-8586-6bf1a0942440 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 810.989512] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9ac23e9e-c303-4029-8e2c-3f5115af48ab tempest-ServerAddressesNegativeTestJSON-963263679 tempest-ServerAddressesNegativeTestJSON-963263679-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.190s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 810.992238] env[65726]: DEBUG oslo_concurrency.lockutils [None req-ae076e22-617f-47ad-a346-b93664495097 tempest-ServersNegativeTestMultiTenantJSON-944098855 tempest-ServersNegativeTestMultiTenantJSON-944098855-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 37.766s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 810.993830] env[65726]: INFO nova.compute.claims [None req-ae076e22-617f-47ad-a346-b93664495097 tempest-ServersNegativeTestMultiTenantJSON-944098855 tempest-ServersNegativeTestMultiTenantJSON-944098855-project-member] [instance: 018c8ac5-66c4-4a9a-ab45-85e84f92b4fe] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 811.028928] env[65726]: DEBUG nova.network.neutron [req-e6888c71-8951-4071-9e26-cd2d50c819e0 req-d6474ee1-c83a-421c-8586-6bf1a0942440 service nova] [instance: 49bcab1f-7b64-4999-abff-37771c58a271] Updated VIF entry in instance network info cache for port f2a8ac3c-ad7e-4eb4-aec3-db29631fd7ad. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 811.029499] env[65726]: DEBUG nova.network.neutron [req-e6888c71-8951-4071-9e26-cd2d50c819e0 req-d6474ee1-c83a-421c-8586-6bf1a0942440 service nova] [instance: 49bcab1f-7b64-4999-abff-37771c58a271] Updating instance_info_cache with network_info: [{"id": "45612b7e-c052-4346-a779-5f94f9adc084", "address": "fa:16:3e:9d:95:60", "network": {"id": "51a66448-0f72-4266-8868-856b48be839f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-155658961-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.161", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d46ccb00794f458b85da4a93879139ee", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b1f3e6c3-5584-4852-9017-476ab8ac4946", "external-id": "nsx-vlan-transportzone-304", "segmentation_id": 304, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap45612b7e-c0", "ovs_interfaceid": "45612b7e-c052-4346-a779-5f94f9adc084", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "f2a8ac3c-ad7e-4eb4-aec3-db29631fd7ad", "address": "fa:16:3e:22:ca:22", "network": {"id": "51a66448-0f72-4266-8868-856b48be839f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-155658961-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d46ccb00794f458b85da4a93879139ee", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b1f3e6c3-5584-4852-9017-476ab8ac4946", "external-id": "nsx-vlan-transportzone-304", "segmentation_id": 304, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf2a8ac3c-ad", "ovs_interfaceid": "f2a8ac3c-ad7e-4eb4-aec3-db29631fd7ad", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 811.035201] env[65726]: INFO nova.scheduler.client.report [None req-9ac23e9e-c303-4029-8e2c-3f5115af48ab tempest-ServerAddressesNegativeTestJSON-963263679 tempest-ServerAddressesNegativeTestJSON-963263679-project-member] Deleted allocations for instance 00746508-d0d6-4dfa-9026-772398d004ee [ 811.068673] env[65726]: DEBUG oslo_vmware.api [None req-c9bfb110-bdb5-4ce1-88bb-ec98452720c3 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Task: {'id': task-5115932, 'name': ReconfigVM_Task, 'duration_secs': 0.384503} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 811.068950] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-c9bfb110-bdb5-4ce1-88bb-ec98452720c3 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: 6d2b86c2-0a44-433f-8b3f-2e9c7f693baa] Reconfigured VM instance instance-0000002b to detach disk 2001 {{(pid=65726) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 811.069910] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-c9bfb110-bdb5-4ce1-88bb-ec98452720c3 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: 6d2b86c2-0a44-433f-8b3f-2e9c7f693baa] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 811.070566] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-edd346b2-45cd-49f7-a68e-bfbdcc6d5059 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.079552] env[65726]: DEBUG oslo_vmware.api [None req-c9bfb110-bdb5-4ce1-88bb-ec98452720c3 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Waiting for the task: (returnval){ [ 811.079552] env[65726]: value = "task-5115936" [ 811.079552] env[65726]: _type = "Task" [ 811.079552] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 811.089472] env[65726]: DEBUG oslo_vmware.api [None req-c9bfb110-bdb5-4ce1-88bb-ec98452720c3 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Task: {'id': task-5115936, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.157020] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115935, 'name': CreateVM_Task, 'duration_secs': 0.495475} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 811.157020] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1bd26aef-995e-43b6-af9f-077e878c8c44] Created VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 811.157359] env[65726]: WARNING openstack [None req-a76cfabd-af7f-4569-b968-716d45bcf7f1 tempest-InstanceActionsNegativeTestJSON-844675881 tempest-InstanceActionsNegativeTestJSON-844675881-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 811.158048] env[65726]: WARNING openstack [None req-a76cfabd-af7f-4569-b968-716d45bcf7f1 tempest-InstanceActionsNegativeTestJSON-844675881 tempest-InstanceActionsNegativeTestJSON-844675881-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 811.165503] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a76cfabd-af7f-4569-b968-716d45bcf7f1 tempest-InstanceActionsNegativeTestJSON-844675881 tempest-InstanceActionsNegativeTestJSON-844675881-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 811.165503] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a76cfabd-af7f-4569-b968-716d45bcf7f1 tempest-InstanceActionsNegativeTestJSON-844675881 tempest-InstanceActionsNegativeTestJSON-844675881-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 811.165503] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a76cfabd-af7f-4569-b968-716d45bcf7f1 tempest-InstanceActionsNegativeTestJSON-844675881 tempest-InstanceActionsNegativeTestJSON-844675881-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 811.165503] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-100f853c-058c-4ed8-8e30-b3cc94a87dd1 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.172798] env[65726]: DEBUG oslo_vmware.api [None req-a76cfabd-af7f-4569-b968-716d45bcf7f1 tempest-InstanceActionsNegativeTestJSON-844675881 tempest-InstanceActionsNegativeTestJSON-844675881-project-member] Waiting for the task: (returnval){ [ 811.172798] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]521cc49b-693e-520d-30ca-3fa6fa9933dd" [ 811.172798] env[65726]: _type = "Task" [ 811.172798] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 811.185244] env[65726]: DEBUG oslo_vmware.api [None req-a76cfabd-af7f-4569-b968-716d45bcf7f1 tempest-InstanceActionsNegativeTestJSON-844675881 tempest-InstanceActionsNegativeTestJSON-844675881-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]521cc49b-693e-520d-30ca-3fa6fa9933dd, 'name': SearchDatastore_Task} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 811.185694] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a76cfabd-af7f-4569-b968-716d45bcf7f1 tempest-InstanceActionsNegativeTestJSON-844675881 tempest-InstanceActionsNegativeTestJSON-844675881-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 811.185975] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-a76cfabd-af7f-4569-b968-716d45bcf7f1 tempest-InstanceActionsNegativeTestJSON-844675881 tempest-InstanceActionsNegativeTestJSON-844675881-project-member] [instance: 1bd26aef-995e-43b6-af9f-077e878c8c44] Processing image b52362a3-ee8a-4cbf-b06f-513b0cc8f95c {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 811.186448] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a76cfabd-af7f-4569-b968-716d45bcf7f1 tempest-InstanceActionsNegativeTestJSON-844675881 tempest-InstanceActionsNegativeTestJSON-844675881-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 811.186530] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a76cfabd-af7f-4569-b968-716d45bcf7f1 tempest-InstanceActionsNegativeTestJSON-844675881 tempest-InstanceActionsNegativeTestJSON-844675881-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 811.186782] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-a76cfabd-af7f-4569-b968-716d45bcf7f1 tempest-InstanceActionsNegativeTestJSON-844675881 tempest-InstanceActionsNegativeTestJSON-844675881-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 811.187124] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-97a60f2a-2752-41b5-a4c1-b4642a415eba {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.198585] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-a76cfabd-af7f-4569-b968-716d45bcf7f1 tempest-InstanceActionsNegativeTestJSON-844675881 tempest-InstanceActionsNegativeTestJSON-844675881-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 811.198923] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-a76cfabd-af7f-4569-b968-716d45bcf7f1 tempest-InstanceActionsNegativeTestJSON-844675881 tempest-InstanceActionsNegativeTestJSON-844675881-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65726) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 811.201205] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1f860fe6-6de5-437f-a0d5-a0adf4a23265 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.209934] env[65726]: DEBUG oslo_vmware.api [None req-a76cfabd-af7f-4569-b968-716d45bcf7f1 tempest-InstanceActionsNegativeTestJSON-844675881 tempest-InstanceActionsNegativeTestJSON-844675881-project-member] Waiting for the task: (returnval){ [ 811.209934] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52c0cb35-3daf-7394-c3eb-171c2f880066" [ 811.209934] env[65726]: _type = "Task" [ 811.209934] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 811.220452] env[65726]: DEBUG oslo_vmware.api [None req-a76cfabd-af7f-4569-b968-716d45bcf7f1 tempest-InstanceActionsNegativeTestJSON-844675881 tempest-InstanceActionsNegativeTestJSON-844675881-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52c0cb35-3daf-7394-c3eb-171c2f880066, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.305808] env[65726]: DEBUG nova.network.neutron [None req-855996fc-ecea-49d8-a71b-8b4307731732 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] [instance: 6c8db442-daaa-4eca-ae24-b9d4afaf3a71] Successfully updated port: 6f422d37-8262-454d-aa2d-1569bacee1b7 {{(pid=65726) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 811.332322] env[65726]: DEBUG nova.compute.manager [None req-855996fc-ecea-49d8-a71b-8b4307731732 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] [instance: 6c8db442-daaa-4eca-ae24-b9d4afaf3a71] Start spawning the instance on the hypervisor. {{(pid=65726) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 811.367707] env[65726]: DEBUG nova.virt.hardware [None req-855996fc-ecea-49d8-a71b-8b4307731732 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:30:59Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=1,extra_specs={hw_rng:allowed='True'},flavorid='2054834516',id=20,is_public=True,memory_mb=192,name='tempest-flavor_with_ephemeral_1-1964852925',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T19:28:34Z,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1539f06fa3534e90acbc3a60c4b8bd3f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T19:28:35Z,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 811.367917] env[65726]: DEBUG nova.virt.hardware [None req-855996fc-ecea-49d8-a71b-8b4307731732 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 811.367967] env[65726]: DEBUG nova.virt.hardware [None req-855996fc-ecea-49d8-a71b-8b4307731732 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 811.368668] env[65726]: DEBUG nova.virt.hardware [None req-855996fc-ecea-49d8-a71b-8b4307731732 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 811.368668] env[65726]: DEBUG nova.virt.hardware [None req-855996fc-ecea-49d8-a71b-8b4307731732 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 811.368668] env[65726]: DEBUG nova.virt.hardware [None req-855996fc-ecea-49d8-a71b-8b4307731732 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 811.368668] env[65726]: DEBUG nova.virt.hardware [None req-855996fc-ecea-49d8-a71b-8b4307731732 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 811.368813] env[65726]: DEBUG nova.virt.hardware [None req-855996fc-ecea-49d8-a71b-8b4307731732 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 811.368957] env[65726]: DEBUG nova.virt.hardware [None req-855996fc-ecea-49d8-a71b-8b4307731732 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 811.369163] env[65726]: DEBUG nova.virt.hardware [None req-855996fc-ecea-49d8-a71b-8b4307731732 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 811.369350] env[65726]: DEBUG nova.virt.hardware [None req-855996fc-ecea-49d8-a71b-8b4307731732 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 811.370498] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-847d1bdb-9591-4dd1-9d42-c2039e3aab42 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.379164] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe5aae32-2f23-4db8-a89e-5e48998412a9 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.535924] env[65726]: DEBUG oslo_concurrency.lockutils [req-e6888c71-8951-4071-9e26-cd2d50c819e0 req-d6474ee1-c83a-421c-8586-6bf1a0942440 service nova] Releasing lock "refresh_cache-49bcab1f-7b64-4999-abff-37771c58a271" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 811.547666] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9ac23e9e-c303-4029-8e2c-3f5115af48ab tempest-ServerAddressesNegativeTestJSON-963263679 tempest-ServerAddressesNegativeTestJSON-963263679-project-member] Lock "00746508-d0d6-4dfa-9026-772398d004ee" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 42.997s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 811.591701] env[65726]: DEBUG oslo_vmware.api [None req-c9bfb110-bdb5-4ce1-88bb-ec98452720c3 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Task: {'id': task-5115936, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.722307] env[65726]: DEBUG oslo_vmware.api [None req-a76cfabd-af7f-4569-b968-716d45bcf7f1 tempest-InstanceActionsNegativeTestJSON-844675881 tempest-InstanceActionsNegativeTestJSON-844675881-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52c0cb35-3daf-7394-c3eb-171c2f880066, 'name': SearchDatastore_Task, 'duration_secs': 0.012388} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 811.723236] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-254ba33f-4930-4019-8697-9f4ab706810b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.729954] env[65726]: DEBUG oslo_vmware.api [None req-a76cfabd-af7f-4569-b968-716d45bcf7f1 tempest-InstanceActionsNegativeTestJSON-844675881 tempest-InstanceActionsNegativeTestJSON-844675881-project-member] Waiting for the task: (returnval){ [ 811.729954] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]520da572-dccd-1fd1-f86b-e0186dc69437" [ 811.729954] env[65726]: _type = "Task" [ 811.729954] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 811.738785] env[65726]: DEBUG oslo_vmware.api [None req-a76cfabd-af7f-4569-b968-716d45bcf7f1 tempest-InstanceActionsNegativeTestJSON-844675881 tempest-InstanceActionsNegativeTestJSON-844675881-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]520da572-dccd-1fd1-f86b-e0186dc69437, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.810879] env[65726]: DEBUG oslo_concurrency.lockutils [None req-855996fc-ecea-49d8-a71b-8b4307731732 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Acquiring lock "refresh_cache-6c8db442-daaa-4eca-ae24-b9d4afaf3a71" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 811.810879] env[65726]: DEBUG oslo_concurrency.lockutils [None req-855996fc-ecea-49d8-a71b-8b4307731732 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Acquired lock "refresh_cache-6c8db442-daaa-4eca-ae24-b9d4afaf3a71" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 811.810879] env[65726]: DEBUG nova.network.neutron [None req-855996fc-ecea-49d8-a71b-8b4307731732 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] [instance: 6c8db442-daaa-4eca-ae24-b9d4afaf3a71] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 811.895165] env[65726]: DEBUG oslo_concurrency.lockutils [None req-49b35bf1-9410-49fb-9a5b-632c41ecd3a6 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Acquiring lock "c4cc4f85-cb35-4edc-a58b-adfee0ce1265" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 811.895390] env[65726]: DEBUG oslo_concurrency.lockutils [None req-49b35bf1-9410-49fb-9a5b-632c41ecd3a6 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Lock "c4cc4f85-cb35-4edc-a58b-adfee0ce1265" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 812.090721] env[65726]: DEBUG oslo_vmware.api [None req-c9bfb110-bdb5-4ce1-88bb-ec98452720c3 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Task: {'id': task-5115936, 'name': PowerOnVM_Task, 'duration_secs': 0.606955} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 812.094389] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-c9bfb110-bdb5-4ce1-88bb-ec98452720c3 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: 6d2b86c2-0a44-433f-8b3f-2e9c7f693baa] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 812.094389] env[65726]: DEBUG nova.compute.manager [None req-c9bfb110-bdb5-4ce1-88bb-ec98452720c3 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: 6d2b86c2-0a44-433f-8b3f-2e9c7f693baa] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 812.095640] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b817cf2c-e191-4c74-949f-76e01088df05 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.205140] env[65726]: DEBUG nova.compute.manager [None req-c3e5127a-ba4f-4e76-986d-1724943cb777 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 608cc0ea-3f6b-4b4f-83c1-01aa50999d63] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 812.206270] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c901bb1-43ff-4afa-8c16-009c39b19c9a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.246773] env[65726]: DEBUG oslo_vmware.api [None req-a76cfabd-af7f-4569-b968-716d45bcf7f1 tempest-InstanceActionsNegativeTestJSON-844675881 tempest-InstanceActionsNegativeTestJSON-844675881-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]520da572-dccd-1fd1-f86b-e0186dc69437, 'name': SearchDatastore_Task, 'duration_secs': 0.038373} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 812.248276] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a76cfabd-af7f-4569-b968-716d45bcf7f1 tempest-InstanceActionsNegativeTestJSON-844675881 tempest-InstanceActionsNegativeTestJSON-844675881-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 812.248276] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-a76cfabd-af7f-4569-b968-716d45bcf7f1 tempest-InstanceActionsNegativeTestJSON-844675881 tempest-InstanceActionsNegativeTestJSON-844675881-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore2] 1bd26aef-995e-43b6-af9f-077e878c8c44/1bd26aef-995e-43b6-af9f-077e878c8c44.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 812.248276] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e5ed50b4-d50d-4a31-9a6d-ae4d1d65ad04 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.257403] env[65726]: DEBUG oslo_vmware.api [None req-a76cfabd-af7f-4569-b968-716d45bcf7f1 tempest-InstanceActionsNegativeTestJSON-844675881 tempest-InstanceActionsNegativeTestJSON-844675881-project-member] Waiting for the task: (returnval){ [ 812.257403] env[65726]: value = "task-5115937" [ 812.257403] env[65726]: _type = "Task" [ 812.257403] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 812.274330] env[65726]: DEBUG oslo_vmware.api [None req-a76cfabd-af7f-4569-b968-716d45bcf7f1 tempest-InstanceActionsNegativeTestJSON-844675881 tempest-InstanceActionsNegativeTestJSON-844675881-project-member] Task: {'id': task-5115937, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.313794] env[65726]: WARNING openstack [None req-855996fc-ecea-49d8-a71b-8b4307731732 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 812.314267] env[65726]: WARNING openstack [None req-855996fc-ecea-49d8-a71b-8b4307731732 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 812.321798] env[65726]: DEBUG nova.network.neutron [None req-855996fc-ecea-49d8-a71b-8b4307731732 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] [instance: 6c8db442-daaa-4eca-ae24-b9d4afaf3a71] Instance cache missing network info. {{(pid=65726) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 812.432377] env[65726]: WARNING openstack [None req-855996fc-ecea-49d8-a71b-8b4307731732 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 812.432798] env[65726]: WARNING openstack [None req-855996fc-ecea-49d8-a71b-8b4307731732 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 812.559772] env[65726]: WARNING openstack [None req-855996fc-ecea-49d8-a71b-8b4307731732 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 812.559772] env[65726]: WARNING openstack [None req-855996fc-ecea-49d8-a71b-8b4307731732 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 812.586490] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-653cbc57-b4bd-4773-8bd4-343686bfeeb8 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.593327] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d44a4b21-9e5a-4f0d-92ea-b644f6bc5c26 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.635284] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3ef0154-530e-4cfd-b730-34d22abfc2cd {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.638327] env[65726]: DEBUG nova.compute.manager [req-b57f9b61-2b31-4154-9b89-8c7097d3e874 req-44c53ae4-827b-4b8e-935a-7913af33d4c0 service nova] [instance: 1bd26aef-995e-43b6-af9f-077e878c8c44] Received event network-vif-plugged-61bb4097-108f-4424-9e34-38ed6f561f0c {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 812.638537] env[65726]: DEBUG oslo_concurrency.lockutils [req-b57f9b61-2b31-4154-9b89-8c7097d3e874 req-44c53ae4-827b-4b8e-935a-7913af33d4c0 service nova] Acquiring lock "1bd26aef-995e-43b6-af9f-077e878c8c44-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 812.638767] env[65726]: DEBUG oslo_concurrency.lockutils [req-b57f9b61-2b31-4154-9b89-8c7097d3e874 req-44c53ae4-827b-4b8e-935a-7913af33d4c0 service nova] Lock "1bd26aef-995e-43b6-af9f-077e878c8c44-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 812.639074] env[65726]: DEBUG oslo_concurrency.lockutils [req-b57f9b61-2b31-4154-9b89-8c7097d3e874 req-44c53ae4-827b-4b8e-935a-7913af33d4c0 service nova] Lock "1bd26aef-995e-43b6-af9f-077e878c8c44-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 812.639124] env[65726]: DEBUG nova.compute.manager [req-b57f9b61-2b31-4154-9b89-8c7097d3e874 req-44c53ae4-827b-4b8e-935a-7913af33d4c0 service nova] [instance: 1bd26aef-995e-43b6-af9f-077e878c8c44] No waiting events found dispatching network-vif-plugged-61bb4097-108f-4424-9e34-38ed6f561f0c {{(pid=65726) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 812.639246] env[65726]: WARNING nova.compute.manager [req-b57f9b61-2b31-4154-9b89-8c7097d3e874 req-44c53ae4-827b-4b8e-935a-7913af33d4c0 service nova] [instance: 1bd26aef-995e-43b6-af9f-077e878c8c44] Received unexpected event network-vif-plugged-61bb4097-108f-4424-9e34-38ed6f561f0c for instance with vm_state building and task_state spawning. [ 812.639396] env[65726]: DEBUG nova.compute.manager [req-b57f9b61-2b31-4154-9b89-8c7097d3e874 req-44c53ae4-827b-4b8e-935a-7913af33d4c0 service nova] [instance: 1bd26aef-995e-43b6-af9f-077e878c8c44] Received event network-changed-61bb4097-108f-4424-9e34-38ed6f561f0c {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 812.639539] env[65726]: DEBUG nova.compute.manager [req-b57f9b61-2b31-4154-9b89-8c7097d3e874 req-44c53ae4-827b-4b8e-935a-7913af33d4c0 service nova] [instance: 1bd26aef-995e-43b6-af9f-077e878c8c44] Refreshing instance network info cache due to event network-changed-61bb4097-108f-4424-9e34-38ed6f561f0c. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 812.639712] env[65726]: DEBUG oslo_concurrency.lockutils [req-b57f9b61-2b31-4154-9b89-8c7097d3e874 req-44c53ae4-827b-4b8e-935a-7913af33d4c0 service nova] Acquiring lock "refresh_cache-1bd26aef-995e-43b6-af9f-077e878c8c44" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 812.639872] env[65726]: DEBUG oslo_concurrency.lockutils [req-b57f9b61-2b31-4154-9b89-8c7097d3e874 req-44c53ae4-827b-4b8e-935a-7913af33d4c0 service nova] Acquired lock "refresh_cache-1bd26aef-995e-43b6-af9f-077e878c8c44" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 812.639975] env[65726]: DEBUG nova.network.neutron [req-b57f9b61-2b31-4154-9b89-8c7097d3e874 req-44c53ae4-827b-4b8e-935a-7913af33d4c0 service nova] [instance: 1bd26aef-995e-43b6-af9f-077e878c8c44] Refreshing network info cache for port 61bb4097-108f-4424-9e34-38ed6f561f0c {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 812.651592] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26aa5c5e-4143-4e13-aef4-2af4ef6f051b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.669187] env[65726]: DEBUG nova.compute.provider_tree [None req-ae076e22-617f-47ad-a346-b93664495097 tempest-ServersNegativeTestMultiTenantJSON-944098855 tempest-ServersNegativeTestMultiTenantJSON-944098855-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 812.679155] env[65726]: DEBUG nova.compute.manager [req-c1900bad-2528-41d0-b6d8-4b6ec6788ecd req-b536691f-3cf0-48e2-b37a-3385e624207d service nova] [instance: bc2c12e0-0d06-432f-b42f-be468e3b6ee4] Received event network-changed-a1c6ed0b-f1f6-4da3-aeaf-9f815ef90b64 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 812.679360] env[65726]: DEBUG nova.compute.manager [req-c1900bad-2528-41d0-b6d8-4b6ec6788ecd req-b536691f-3cf0-48e2-b37a-3385e624207d service nova] [instance: bc2c12e0-0d06-432f-b42f-be468e3b6ee4] Refreshing instance network info cache due to event network-changed-a1c6ed0b-f1f6-4da3-aeaf-9f815ef90b64. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 812.679568] env[65726]: DEBUG oslo_concurrency.lockutils [req-c1900bad-2528-41d0-b6d8-4b6ec6788ecd req-b536691f-3cf0-48e2-b37a-3385e624207d service nova] Acquiring lock "refresh_cache-bc2c12e0-0d06-432f-b42f-be468e3b6ee4" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 812.679699] env[65726]: DEBUG oslo_concurrency.lockutils [req-c1900bad-2528-41d0-b6d8-4b6ec6788ecd req-b536691f-3cf0-48e2-b37a-3385e624207d service nova] Acquired lock "refresh_cache-bc2c12e0-0d06-432f-b42f-be468e3b6ee4" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 812.679845] env[65726]: DEBUG nova.network.neutron [req-c1900bad-2528-41d0-b6d8-4b6ec6788ecd req-b536691f-3cf0-48e2-b37a-3385e624207d service nova] [instance: bc2c12e0-0d06-432f-b42f-be468e3b6ee4] Refreshing network info cache for port a1c6ed0b-f1f6-4da3-aeaf-9f815ef90b64 {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 812.683080] env[65726]: DEBUG nova.network.neutron [None req-855996fc-ecea-49d8-a71b-8b4307731732 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] [instance: 6c8db442-daaa-4eca-ae24-b9d4afaf3a71] Updating instance_info_cache with network_info: [{"id": "6f422d37-8262-454d-aa2d-1569bacee1b7", "address": "fa:16:3e:19:b2:c8", "network": {"id": "d37113a1-409c-4c2c-81f9-8f80822b85b9", "bridge": "br-int", "label": "tempest-ServersWithSpecificFlavorTestJSON-471231219-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "79d3e960f190439fae7c61d693e9356e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "13af9422-d668-4413-b63a-766558d83a3b", "external-id": "nsx-vlan-transportzone-842", "segmentation_id": 842, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6f422d37-82", "ovs_interfaceid": "6f422d37-8262-454d-aa2d-1569bacee1b7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 812.723073] env[65726]: INFO nova.compute.manager [None req-c3e5127a-ba4f-4e76-986d-1724943cb777 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 608cc0ea-3f6b-4b4f-83c1-01aa50999d63] instance snapshotting [ 812.723686] env[65726]: DEBUG nova.objects.instance [None req-c3e5127a-ba4f-4e76-986d-1724943cb777 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Lazy-loading 'flavor' on Instance uuid 608cc0ea-3f6b-4b4f-83c1-01aa50999d63 {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 812.767923] env[65726]: DEBUG oslo_vmware.api [None req-a76cfabd-af7f-4569-b968-716d45bcf7f1 tempest-InstanceActionsNegativeTestJSON-844675881 tempest-InstanceActionsNegativeTestJSON-844675881-project-member] Task: {'id': task-5115937, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.145261] env[65726]: WARNING openstack [req-b57f9b61-2b31-4154-9b89-8c7097d3e874 req-44c53ae4-827b-4b8e-935a-7913af33d4c0 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 813.146095] env[65726]: WARNING openstack [req-b57f9b61-2b31-4154-9b89-8c7097d3e874 req-44c53ae4-827b-4b8e-935a-7913af33d4c0 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 813.173085] env[65726]: DEBUG nova.scheduler.client.report [None req-ae076e22-617f-47ad-a346-b93664495097 tempest-ServersNegativeTestMultiTenantJSON-944098855 tempest-ServersNegativeTestMultiTenantJSON-944098855-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 813.187843] env[65726]: WARNING openstack [req-c1900bad-2528-41d0-b6d8-4b6ec6788ecd req-b536691f-3cf0-48e2-b37a-3385e624207d service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 813.187843] env[65726]: WARNING openstack [req-c1900bad-2528-41d0-b6d8-4b6ec6788ecd req-b536691f-3cf0-48e2-b37a-3385e624207d service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 813.194635] env[65726]: DEBUG oslo_concurrency.lockutils [None req-855996fc-ecea-49d8-a71b-8b4307731732 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Releasing lock "refresh_cache-6c8db442-daaa-4eca-ae24-b9d4afaf3a71" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 813.194906] env[65726]: DEBUG nova.compute.manager [None req-855996fc-ecea-49d8-a71b-8b4307731732 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] [instance: 6c8db442-daaa-4eca-ae24-b9d4afaf3a71] Instance network_info: |[{"id": "6f422d37-8262-454d-aa2d-1569bacee1b7", "address": "fa:16:3e:19:b2:c8", "network": {"id": "d37113a1-409c-4c2c-81f9-8f80822b85b9", "bridge": "br-int", "label": "tempest-ServersWithSpecificFlavorTestJSON-471231219-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "79d3e960f190439fae7c61d693e9356e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "13af9422-d668-4413-b63a-766558d83a3b", "external-id": "nsx-vlan-transportzone-842", "segmentation_id": 842, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6f422d37-82", "ovs_interfaceid": "6f422d37-8262-454d-aa2d-1569bacee1b7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 813.202794] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-855996fc-ecea-49d8-a71b-8b4307731732 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] [instance: 6c8db442-daaa-4eca-ae24-b9d4afaf3a71] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:19:b2:c8', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '13af9422-d668-4413-b63a-766558d83a3b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '6f422d37-8262-454d-aa2d-1569bacee1b7', 'vif_model': 'vmxnet3'}] {{(pid=65726) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 813.214248] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-855996fc-ecea-49d8-a71b-8b4307731732 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 813.216777] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6c8db442-daaa-4eca-ae24-b9d4afaf3a71] Creating VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 813.217505] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-add76456-bf1b-486f-b87c-0e2282ddfc7c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.239917] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a4f1270-dea7-44ee-8df9-f822943941bf {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.245720] env[65726]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 813.245720] env[65726]: value = "task-5115938" [ 813.245720] env[65726]: _type = "Task" [ 813.245720] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.280176] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf79aa8a-f5dd-4e56-baf4-6d1492a26cbb {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.291317] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115938, 'name': CreateVM_Task} progress is 10%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.304434] env[65726]: DEBUG oslo_vmware.api [None req-a76cfabd-af7f-4569-b968-716d45bcf7f1 tempest-InstanceActionsNegativeTestJSON-844675881 tempest-InstanceActionsNegativeTestJSON-844675881-project-member] Task: {'id': task-5115937, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.393108] env[65726]: WARNING openstack [req-b57f9b61-2b31-4154-9b89-8c7097d3e874 req-44c53ae4-827b-4b8e-935a-7913af33d4c0 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 813.393531] env[65726]: WARNING openstack [req-b57f9b61-2b31-4154-9b89-8c7097d3e874 req-44c53ae4-827b-4b8e-935a-7913af33d4c0 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 813.540450] env[65726]: WARNING openstack [req-c1900bad-2528-41d0-b6d8-4b6ec6788ecd req-b536691f-3cf0-48e2-b37a-3385e624207d service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 813.540896] env[65726]: WARNING openstack [req-c1900bad-2528-41d0-b6d8-4b6ec6788ecd req-b536691f-3cf0-48e2-b37a-3385e624207d service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 813.633092] env[65726]: WARNING openstack [req-b57f9b61-2b31-4154-9b89-8c7097d3e874 req-44c53ae4-827b-4b8e-935a-7913af33d4c0 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 813.633619] env[65726]: WARNING openstack [req-b57f9b61-2b31-4154-9b89-8c7097d3e874 req-44c53ae4-827b-4b8e-935a-7913af33d4c0 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 813.677197] env[65726]: WARNING openstack [req-c1900bad-2528-41d0-b6d8-4b6ec6788ecd req-b536691f-3cf0-48e2-b37a-3385e624207d service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 813.678494] env[65726]: WARNING openstack [req-c1900bad-2528-41d0-b6d8-4b6ec6788ecd req-b536691f-3cf0-48e2-b37a-3385e624207d service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 813.689679] env[65726]: DEBUG oslo_concurrency.lockutils [None req-ae076e22-617f-47ad-a346-b93664495097 tempest-ServersNegativeTestMultiTenantJSON-944098855 tempest-ServersNegativeTestMultiTenantJSON-944098855-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.697s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 813.690359] env[65726]: DEBUG nova.compute.manager [None req-ae076e22-617f-47ad-a346-b93664495097 tempest-ServersNegativeTestMultiTenantJSON-944098855 tempest-ServersNegativeTestMultiTenantJSON-944098855-project-member] [instance: 018c8ac5-66c4-4a9a-ab45-85e84f92b4fe] Start building networks asynchronously for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 813.694309] env[65726]: DEBUG oslo_concurrency.lockutils [None req-81594943-9858-4765-a371-42f9ef0cbec9 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 37.880s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 813.696124] env[65726]: INFO nova.compute.claims [None req-81594943-9858-4765-a371-42f9ef0cbec9 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] [instance: 3cbbe146-be3a-43d5-867d-4d669884758c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 813.761657] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115938, 'name': CreateVM_Task, 'duration_secs': 0.492805} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 813.761844] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6c8db442-daaa-4eca-ae24-b9d4afaf3a71] Created VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 813.764358] env[65726]: WARNING openstack [None req-855996fc-ecea-49d8-a71b-8b4307731732 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 813.764358] env[65726]: WARNING openstack [None req-855996fc-ecea-49d8-a71b-8b4307731732 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 813.768215] env[65726]: DEBUG oslo_concurrency.lockutils [None req-855996fc-ecea-49d8-a71b-8b4307731732 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 813.768382] env[65726]: DEBUG oslo_concurrency.lockutils [None req-855996fc-ecea-49d8-a71b-8b4307731732 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 813.768698] env[65726]: DEBUG oslo_concurrency.lockutils [None req-855996fc-ecea-49d8-a71b-8b4307731732 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 813.773858] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d97d38ba-ca60-4736-9100-e08e2a2a4937 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.783514] env[65726]: DEBUG oslo_vmware.api [None req-855996fc-ecea-49d8-a71b-8b4307731732 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Waiting for the task: (returnval){ [ 813.783514] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]525e2f8e-5ecc-88a7-0c0e-53d912fb6bb8" [ 813.783514] env[65726]: _type = "Task" [ 813.783514] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.789011] env[65726]: DEBUG oslo_vmware.api [None req-a76cfabd-af7f-4569-b968-716d45bcf7f1 tempest-InstanceActionsNegativeTestJSON-844675881 tempest-InstanceActionsNegativeTestJSON-844675881-project-member] Task: {'id': task-5115937, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.04559} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 813.791432] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-a76cfabd-af7f-4569-b968-716d45bcf7f1 tempest-InstanceActionsNegativeTestJSON-844675881 tempest-InstanceActionsNegativeTestJSON-844675881-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore2] 1bd26aef-995e-43b6-af9f-077e878c8c44/1bd26aef-995e-43b6-af9f-077e878c8c44.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 813.791684] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-a76cfabd-af7f-4569-b968-716d45bcf7f1 tempest-InstanceActionsNegativeTestJSON-844675881 tempest-InstanceActionsNegativeTestJSON-844675881-project-member] [instance: 1bd26aef-995e-43b6-af9f-077e878c8c44] Extending root virtual disk to 1048576 {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 813.792399] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2865fcfe-262e-4688-8689-f2a63f505076 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.803742] env[65726]: DEBUG oslo_vmware.api [None req-855996fc-ecea-49d8-a71b-8b4307731732 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]525e2f8e-5ecc-88a7-0c0e-53d912fb6bb8, 'name': SearchDatastore_Task, 'duration_secs': 0.017718} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 813.806318] env[65726]: DEBUG oslo_concurrency.lockutils [None req-855996fc-ecea-49d8-a71b-8b4307731732 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 813.806668] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-855996fc-ecea-49d8-a71b-8b4307731732 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] [instance: 6c8db442-daaa-4eca-ae24-b9d4afaf3a71] Processing image b52362a3-ee8a-4cbf-b06f-513b0cc8f95c {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 813.807069] env[65726]: DEBUG oslo_concurrency.lockutils [None req-855996fc-ecea-49d8-a71b-8b4307731732 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 813.807296] env[65726]: DEBUG oslo_concurrency.lockutils [None req-855996fc-ecea-49d8-a71b-8b4307731732 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 813.807570] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-855996fc-ecea-49d8-a71b-8b4307731732 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 813.808073] env[65726]: DEBUG oslo_vmware.api [None req-a76cfabd-af7f-4569-b968-716d45bcf7f1 tempest-InstanceActionsNegativeTestJSON-844675881 tempest-InstanceActionsNegativeTestJSON-844675881-project-member] Waiting for the task: (returnval){ [ 813.808073] env[65726]: value = "task-5115939" [ 813.808073] env[65726]: _type = "Task" [ 813.808073] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.809555] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-c3e5127a-ba4f-4e76-986d-1724943cb777 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 608cc0ea-3f6b-4b4f-83c1-01aa50999d63] Creating Snapshot of the VM instance {{(pid=65726) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 813.809930] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1e76857d-0a63-4774-9a87-74d277195f41 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.813993] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-6dc1bef3-a196-46b3-9725-22d70e76fa95 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.829157] env[65726]: DEBUG oslo_vmware.api [None req-a76cfabd-af7f-4569-b968-716d45bcf7f1 tempest-InstanceActionsNegativeTestJSON-844675881 tempest-InstanceActionsNegativeTestJSON-844675881-project-member] Task: {'id': task-5115939, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.830906] env[65726]: DEBUG oslo_vmware.api [None req-c3e5127a-ba4f-4e76-986d-1724943cb777 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Waiting for the task: (returnval){ [ 813.830906] env[65726]: value = "task-5115940" [ 813.830906] env[65726]: _type = "Task" [ 813.830906] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.832437] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-855996fc-ecea-49d8-a71b-8b4307731732 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 813.832603] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-855996fc-ecea-49d8-a71b-8b4307731732 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65726) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 813.836465] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d07538e4-c9ec-4b69-884c-2049e0956d9d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.844116] env[65726]: DEBUG oslo_vmware.api [None req-855996fc-ecea-49d8-a71b-8b4307731732 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Waiting for the task: (returnval){ [ 813.844116] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5298df08-e4e7-c552-b2a4-c4458b55be2e" [ 813.844116] env[65726]: _type = "Task" [ 813.844116] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.848617] env[65726]: DEBUG oslo_vmware.api [None req-c3e5127a-ba4f-4e76-986d-1724943cb777 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5115940, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.860626] env[65726]: DEBUG oslo_vmware.api [None req-855996fc-ecea-49d8-a71b-8b4307731732 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5298df08-e4e7-c552-b2a4-c4458b55be2e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.882144] env[65726]: DEBUG nova.network.neutron [req-b57f9b61-2b31-4154-9b89-8c7097d3e874 req-44c53ae4-827b-4b8e-935a-7913af33d4c0 service nova] [instance: 1bd26aef-995e-43b6-af9f-077e878c8c44] Updated VIF entry in instance network info cache for port 61bb4097-108f-4424-9e34-38ed6f561f0c. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 813.882631] env[65726]: DEBUG nova.network.neutron [req-b57f9b61-2b31-4154-9b89-8c7097d3e874 req-44c53ae4-827b-4b8e-935a-7913af33d4c0 service nova] [instance: 1bd26aef-995e-43b6-af9f-077e878c8c44] Updating instance_info_cache with network_info: [{"id": "61bb4097-108f-4424-9e34-38ed6f561f0c", "address": "fa:16:3e:a6:40:8f", "network": {"id": "1efad9b9-de6f-4953-905d-49886cbc9e17", "bridge": "br-int", "label": "tempest-InstanceActionsNegativeTestJSON-637686941-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "faa59ae0ed7f4c86939a84de1e3672e9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e5d88cd9-35a3-4ac3-9d6d-756464cd6cc5", "external-id": "nsx-vlan-transportzone-685", "segmentation_id": 685, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap61bb4097-10", "ovs_interfaceid": "61bb4097-108f-4424-9e34-38ed6f561f0c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 813.891621] env[65726]: DEBUG nova.network.neutron [req-c1900bad-2528-41d0-b6d8-4b6ec6788ecd req-b536691f-3cf0-48e2-b37a-3385e624207d service nova] [instance: bc2c12e0-0d06-432f-b42f-be468e3b6ee4] Updated VIF entry in instance network info cache for port a1c6ed0b-f1f6-4da3-aeaf-9f815ef90b64. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 813.891621] env[65726]: DEBUG nova.network.neutron [req-c1900bad-2528-41d0-b6d8-4b6ec6788ecd req-b536691f-3cf0-48e2-b37a-3385e624207d service nova] [instance: bc2c12e0-0d06-432f-b42f-be468e3b6ee4] Updating instance_info_cache with network_info: [{"id": "a1c6ed0b-f1f6-4da3-aeaf-9f815ef90b64", "address": "fa:16:3e:09:eb:ee", "network": {"id": "dd4bbaef-0a1e-4e00-b65a-f19e7a837cc1", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-255385230-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "55f93cdf51e84cf683879668c2acecf4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "04ccbc7a-cf8d-4ea2-8411-291a1e27df7b", "external-id": "nsx-vlan-transportzone-998", "segmentation_id": 998, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa1c6ed0b-f1", "ovs_interfaceid": "a1c6ed0b-f1f6-4da3-aeaf-9f815ef90b64", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 814.168150] env[65726]: DEBUG oslo_concurrency.lockutils [None req-85226e56-668f-4c9e-b188-d90c53166469 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Acquiring lock "interface-49bcab1f-7b64-4999-abff-37771c58a271-f2a8ac3c-ad7e-4eb4-aec3-db29631fd7ad" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 814.168481] env[65726]: DEBUG oslo_concurrency.lockutils [None req-85226e56-668f-4c9e-b188-d90c53166469 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Lock "interface-49bcab1f-7b64-4999-abff-37771c58a271-f2a8ac3c-ad7e-4eb4-aec3-db29631fd7ad" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.001s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 814.205350] env[65726]: DEBUG nova.compute.utils [None req-ae076e22-617f-47ad-a346-b93664495097 tempest-ServersNegativeTestMultiTenantJSON-944098855 tempest-ServersNegativeTestMultiTenantJSON-944098855-project-member] Using /dev/sd instead of None {{(pid=65726) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 814.208440] env[65726]: DEBUG nova.compute.manager [None req-ae076e22-617f-47ad-a346-b93664495097 tempest-ServersNegativeTestMultiTenantJSON-944098855 tempest-ServersNegativeTestMultiTenantJSON-944098855-project-member] [instance: 018c8ac5-66c4-4a9a-ab45-85e84f92b4fe] Allocating IP information in the background. {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 814.208440] env[65726]: DEBUG nova.network.neutron [None req-ae076e22-617f-47ad-a346-b93664495097 tempest-ServersNegativeTestMultiTenantJSON-944098855 tempest-ServersNegativeTestMultiTenantJSON-944098855-project-member] [instance: 018c8ac5-66c4-4a9a-ab45-85e84f92b4fe] allocate_for_instance() {{(pid=65726) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 814.208563] env[65726]: WARNING neutronclient.v2_0.client [None req-ae076e22-617f-47ad-a346-b93664495097 tempest-ServersNegativeTestMultiTenantJSON-944098855 tempest-ServersNegativeTestMultiTenantJSON-944098855-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 814.212019] env[65726]: WARNING neutronclient.v2_0.client [None req-ae076e22-617f-47ad-a346-b93664495097 tempest-ServersNegativeTestMultiTenantJSON-944098855 tempest-ServersNegativeTestMultiTenantJSON-944098855-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 814.212019] env[65726]: WARNING openstack [None req-ae076e22-617f-47ad-a346-b93664495097 tempest-ServersNegativeTestMultiTenantJSON-944098855 tempest-ServersNegativeTestMultiTenantJSON-944098855-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 814.212019] env[65726]: WARNING openstack [None req-ae076e22-617f-47ad-a346-b93664495097 tempest-ServersNegativeTestMultiTenantJSON-944098855 tempest-ServersNegativeTestMultiTenantJSON-944098855-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 814.226632] env[65726]: DEBUG nova.compute.manager [None req-ae076e22-617f-47ad-a346-b93664495097 tempest-ServersNegativeTestMultiTenantJSON-944098855 tempest-ServersNegativeTestMultiTenantJSON-944098855-project-member] [instance: 018c8ac5-66c4-4a9a-ab45-85e84f92b4fe] Start building block device mappings for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 814.289728] env[65726]: DEBUG nova.policy [None req-ae076e22-617f-47ad-a346-b93664495097 tempest-ServersNegativeTestMultiTenantJSON-944098855 tempest-ServersNegativeTestMultiTenantJSON-944098855-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f33379b2dab94fbba9816d7b3f2315dd', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6fc182ce253b4da98ffe8e8d9d224156', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65726) authorize /opt/stack/nova/nova/policy.py:192}} [ 814.323862] env[65726]: DEBUG oslo_vmware.api [None req-a76cfabd-af7f-4569-b968-716d45bcf7f1 tempest-InstanceActionsNegativeTestJSON-844675881 tempest-InstanceActionsNegativeTestJSON-844675881-project-member] Task: {'id': task-5115939, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.080763} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 814.324287] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-a76cfabd-af7f-4569-b968-716d45bcf7f1 tempest-InstanceActionsNegativeTestJSON-844675881 tempest-InstanceActionsNegativeTestJSON-844675881-project-member] [instance: 1bd26aef-995e-43b6-af9f-077e878c8c44] Extended root virtual disk {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 814.326490] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1aa0b04-7ffd-4f39-972e-d3ab3ff812b5 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.352714] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-a76cfabd-af7f-4569-b968-716d45bcf7f1 tempest-InstanceActionsNegativeTestJSON-844675881 tempest-InstanceActionsNegativeTestJSON-844675881-project-member] [instance: 1bd26aef-995e-43b6-af9f-077e878c8c44] Reconfiguring VM instance instance-00000031 to attach disk [datastore2] 1bd26aef-995e-43b6-af9f-077e878c8c44/1bd26aef-995e-43b6-af9f-077e878c8c44.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 814.357086] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-77e3315f-f89c-4b58-ad96-4de32a44fb44 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.387073] env[65726]: DEBUG oslo_concurrency.lockutils [req-b57f9b61-2b31-4154-9b89-8c7097d3e874 req-44c53ae4-827b-4b8e-935a-7913af33d4c0 service nova] Releasing lock "refresh_cache-1bd26aef-995e-43b6-af9f-077e878c8c44" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 814.387299] env[65726]: DEBUG nova.compute.manager [req-b57f9b61-2b31-4154-9b89-8c7097d3e874 req-44c53ae4-827b-4b8e-935a-7913af33d4c0 service nova] [instance: 6c8db442-daaa-4eca-ae24-b9d4afaf3a71] Received event network-vif-plugged-6f422d37-8262-454d-aa2d-1569bacee1b7 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 814.387508] env[65726]: DEBUG oslo_concurrency.lockutils [req-b57f9b61-2b31-4154-9b89-8c7097d3e874 req-44c53ae4-827b-4b8e-935a-7913af33d4c0 service nova] Acquiring lock "6c8db442-daaa-4eca-ae24-b9d4afaf3a71-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 814.388663] env[65726]: DEBUG oslo_concurrency.lockutils [req-b57f9b61-2b31-4154-9b89-8c7097d3e874 req-44c53ae4-827b-4b8e-935a-7913af33d4c0 service nova] Lock "6c8db442-daaa-4eca-ae24-b9d4afaf3a71-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 814.388663] env[65726]: DEBUG oslo_concurrency.lockutils [req-b57f9b61-2b31-4154-9b89-8c7097d3e874 req-44c53ae4-827b-4b8e-935a-7913af33d4c0 service nova] Lock "6c8db442-daaa-4eca-ae24-b9d4afaf3a71-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 814.388663] env[65726]: DEBUG nova.compute.manager [req-b57f9b61-2b31-4154-9b89-8c7097d3e874 req-44c53ae4-827b-4b8e-935a-7913af33d4c0 service nova] [instance: 6c8db442-daaa-4eca-ae24-b9d4afaf3a71] No waiting events found dispatching network-vif-plugged-6f422d37-8262-454d-aa2d-1569bacee1b7 {{(pid=65726) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 814.388663] env[65726]: WARNING nova.compute.manager [req-b57f9b61-2b31-4154-9b89-8c7097d3e874 req-44c53ae4-827b-4b8e-935a-7913af33d4c0 service nova] [instance: 6c8db442-daaa-4eca-ae24-b9d4afaf3a71] Received unexpected event network-vif-plugged-6f422d37-8262-454d-aa2d-1569bacee1b7 for instance with vm_state building and task_state spawning. [ 814.388663] env[65726]: DEBUG nova.compute.manager [req-b57f9b61-2b31-4154-9b89-8c7097d3e874 req-44c53ae4-827b-4b8e-935a-7913af33d4c0 service nova] [instance: 6c8db442-daaa-4eca-ae24-b9d4afaf3a71] Received event network-changed-6f422d37-8262-454d-aa2d-1569bacee1b7 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 814.388663] env[65726]: DEBUG nova.compute.manager [req-b57f9b61-2b31-4154-9b89-8c7097d3e874 req-44c53ae4-827b-4b8e-935a-7913af33d4c0 service nova] [instance: 6c8db442-daaa-4eca-ae24-b9d4afaf3a71] Refreshing instance network info cache due to event network-changed-6f422d37-8262-454d-aa2d-1569bacee1b7. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 814.389023] env[65726]: DEBUG oslo_concurrency.lockutils [req-b57f9b61-2b31-4154-9b89-8c7097d3e874 req-44c53ae4-827b-4b8e-935a-7913af33d4c0 service nova] Acquiring lock "refresh_cache-6c8db442-daaa-4eca-ae24-b9d4afaf3a71" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 814.389023] env[65726]: DEBUG oslo_concurrency.lockutils [req-b57f9b61-2b31-4154-9b89-8c7097d3e874 req-44c53ae4-827b-4b8e-935a-7913af33d4c0 service nova] Acquired lock "refresh_cache-6c8db442-daaa-4eca-ae24-b9d4afaf3a71" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 814.389023] env[65726]: DEBUG nova.network.neutron [req-b57f9b61-2b31-4154-9b89-8c7097d3e874 req-44c53ae4-827b-4b8e-935a-7913af33d4c0 service nova] [instance: 6c8db442-daaa-4eca-ae24-b9d4afaf3a71] Refreshing network info cache for port 6f422d37-8262-454d-aa2d-1569bacee1b7 {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 814.390468] env[65726]: DEBUG oslo_vmware.api [None req-855996fc-ecea-49d8-a71b-8b4307731732 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5298df08-e4e7-c552-b2a4-c4458b55be2e, 'name': SearchDatastore_Task, 'duration_secs': 0.015513} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 814.395272] env[65726]: DEBUG oslo_vmware.api [None req-c3e5127a-ba4f-4e76-986d-1724943cb777 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5115940, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.395596] env[65726]: DEBUG oslo_vmware.api [None req-a76cfabd-af7f-4569-b968-716d45bcf7f1 tempest-InstanceActionsNegativeTestJSON-844675881 tempest-InstanceActionsNegativeTestJSON-844675881-project-member] Waiting for the task: (returnval){ [ 814.395596] env[65726]: value = "task-5115941" [ 814.395596] env[65726]: _type = "Task" [ 814.395596] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 814.396232] env[65726]: DEBUG oslo_concurrency.lockutils [req-c1900bad-2528-41d0-b6d8-4b6ec6788ecd req-b536691f-3cf0-48e2-b37a-3385e624207d service nova] Releasing lock "refresh_cache-bc2c12e0-0d06-432f-b42f-be468e3b6ee4" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 814.396598] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c3e5b784-511c-4cc1-9f80-994612ba3aab {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.407876] env[65726]: DEBUG oslo_vmware.api [None req-855996fc-ecea-49d8-a71b-8b4307731732 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Waiting for the task: (returnval){ [ 814.407876] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5226f614-2d69-1744-856a-61562d76e206" [ 814.407876] env[65726]: _type = "Task" [ 814.407876] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 814.415631] env[65726]: DEBUG oslo_vmware.api [None req-a76cfabd-af7f-4569-b968-716d45bcf7f1 tempest-InstanceActionsNegativeTestJSON-844675881 tempest-InstanceActionsNegativeTestJSON-844675881-project-member] Task: {'id': task-5115941, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.422437] env[65726]: DEBUG oslo_vmware.api [None req-855996fc-ecea-49d8-a71b-8b4307731732 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5226f614-2d69-1744-856a-61562d76e206, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.589318] env[65726]: DEBUG nova.network.neutron [None req-ae076e22-617f-47ad-a346-b93664495097 tempest-ServersNegativeTestMultiTenantJSON-944098855 tempest-ServersNegativeTestMultiTenantJSON-944098855-project-member] [instance: 018c8ac5-66c4-4a9a-ab45-85e84f92b4fe] Successfully created port: 5a241c1a-c51d-49fc-8387-7dd3cf7f38af {{(pid=65726) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 814.671660] env[65726]: DEBUG oslo_concurrency.lockutils [None req-85226e56-668f-4c9e-b188-d90c53166469 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Acquiring lock "49bcab1f-7b64-4999-abff-37771c58a271" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 814.671866] env[65726]: DEBUG oslo_concurrency.lockutils [None req-85226e56-668f-4c9e-b188-d90c53166469 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Acquired lock "49bcab1f-7b64-4999-abff-37771c58a271" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 814.672824] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db2f60d1-f0c6-4d8f-841d-a56874f887ed {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.692264] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c5609f8-96cd-4bae-84e1-862249effbea {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.724594] env[65726]: WARNING openstack [None req-85226e56-668f-4c9e-b188-d90c53166469 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 814.725053] env[65726]: WARNING openstack [None req-85226e56-668f-4c9e-b188-d90c53166469 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 814.735482] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-85226e56-668f-4c9e-b188-d90c53166469 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: 49bcab1f-7b64-4999-abff-37771c58a271] Reconfiguring VM to detach interface {{(pid=65726) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1968}} [ 814.740489] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e70aa7cf-dafc-4cc2-b1d3-cf78760ac1a8 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.761070] env[65726]: DEBUG oslo_vmware.api [None req-85226e56-668f-4c9e-b188-d90c53166469 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Waiting for the task: (returnval){ [ 814.761070] env[65726]: value = "task-5115942" [ 814.761070] env[65726]: _type = "Task" [ 814.761070] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 814.774047] env[65726]: DEBUG oslo_vmware.api [None req-85226e56-668f-4c9e-b188-d90c53166469 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': task-5115942, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.856466] env[65726]: DEBUG oslo_vmware.api [None req-c3e5127a-ba4f-4e76-986d-1724943cb777 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5115940, 'name': CreateSnapshot_Task, 'duration_secs': 0.595835} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 814.856809] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-c3e5127a-ba4f-4e76-986d-1724943cb777 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 608cc0ea-3f6b-4b4f-83c1-01aa50999d63] Created Snapshot of the VM instance {{(pid=65726) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 814.857775] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da261db2-6474-49df-b8f5-d8d8a629e428 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.897999] env[65726]: WARNING openstack [req-b57f9b61-2b31-4154-9b89-8c7097d3e874 req-44c53ae4-827b-4b8e-935a-7913af33d4c0 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 814.898573] env[65726]: WARNING openstack [req-b57f9b61-2b31-4154-9b89-8c7097d3e874 req-44c53ae4-827b-4b8e-935a-7913af33d4c0 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 814.926997] env[65726]: DEBUG oslo_vmware.api [None req-855996fc-ecea-49d8-a71b-8b4307731732 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5226f614-2d69-1744-856a-61562d76e206, 'name': SearchDatastore_Task, 'duration_secs': 0.017506} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 814.930493] env[65726]: DEBUG oslo_concurrency.lockutils [None req-855996fc-ecea-49d8-a71b-8b4307731732 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 814.930711] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-855996fc-ecea-49d8-a71b-8b4307731732 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore2] 6c8db442-daaa-4eca-ae24-b9d4afaf3a71/6c8db442-daaa-4eca-ae24-b9d4afaf3a71.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 814.931019] env[65726]: DEBUG oslo_vmware.api [None req-a76cfabd-af7f-4569-b968-716d45bcf7f1 tempest-InstanceActionsNegativeTestJSON-844675881 tempest-InstanceActionsNegativeTestJSON-844675881-project-member] Task: {'id': task-5115941, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.934178] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-db078349-1627-4224-a89b-64b652dd6e07 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.943681] env[65726]: DEBUG oslo_vmware.api [None req-855996fc-ecea-49d8-a71b-8b4307731732 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Waiting for the task: (returnval){ [ 814.943681] env[65726]: value = "task-5115943" [ 814.943681] env[65726]: _type = "Task" [ 814.943681] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 814.955312] env[65726]: DEBUG oslo_vmware.api [None req-855996fc-ecea-49d8-a71b-8b4307731732 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Task: {'id': task-5115943, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.980807] env[65726]: DEBUG nova.compute.manager [req-8660dbe1-3675-4ab9-826a-0f55c51d6e09 req-7a640ba2-57af-4f31-b674-e0efbdfac702 service nova] [instance: bc2c12e0-0d06-432f-b42f-be468e3b6ee4] Received event network-changed-a1c6ed0b-f1f6-4da3-aeaf-9f815ef90b64 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 814.981097] env[65726]: DEBUG nova.compute.manager [req-8660dbe1-3675-4ab9-826a-0f55c51d6e09 req-7a640ba2-57af-4f31-b674-e0efbdfac702 service nova] [instance: bc2c12e0-0d06-432f-b42f-be468e3b6ee4] Refreshing instance network info cache due to event network-changed-a1c6ed0b-f1f6-4da3-aeaf-9f815ef90b64. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 814.981293] env[65726]: DEBUG oslo_concurrency.lockutils [req-8660dbe1-3675-4ab9-826a-0f55c51d6e09 req-7a640ba2-57af-4f31-b674-e0efbdfac702 service nova] Acquiring lock "refresh_cache-bc2c12e0-0d06-432f-b42f-be468e3b6ee4" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 814.981473] env[65726]: DEBUG oslo_concurrency.lockutils [req-8660dbe1-3675-4ab9-826a-0f55c51d6e09 req-7a640ba2-57af-4f31-b674-e0efbdfac702 service nova] Acquired lock "refresh_cache-bc2c12e0-0d06-432f-b42f-be468e3b6ee4" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 814.981594] env[65726]: DEBUG nova.network.neutron [req-8660dbe1-3675-4ab9-826a-0f55c51d6e09 req-7a640ba2-57af-4f31-b674-e0efbdfac702 service nova] [instance: bc2c12e0-0d06-432f-b42f-be468e3b6ee4] Refreshing network info cache for port a1c6ed0b-f1f6-4da3-aeaf-9f815ef90b64 {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 815.256193] env[65726]: DEBUG nova.compute.manager [None req-ae076e22-617f-47ad-a346-b93664495097 tempest-ServersNegativeTestMultiTenantJSON-944098855 tempest-ServersNegativeTestMultiTenantJSON-944098855-project-member] [instance: 018c8ac5-66c4-4a9a-ab45-85e84f92b4fe] Start spawning the instance on the hypervisor. {{(pid=65726) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 815.273737] env[65726]: DEBUG oslo_vmware.api [None req-85226e56-668f-4c9e-b188-d90c53166469 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': task-5115942, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.284643] env[65726]: DEBUG nova.virt.hardware [None req-ae076e22-617f-47ad-a346-b93664495097 tempest-ServersNegativeTestMultiTenantJSON-944098855 tempest-ServersNegativeTestMultiTenantJSON-944098855-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T19:28:34Z,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1539f06fa3534e90acbc3a60c4b8bd3f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T19:28:35Z,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 815.284887] env[65726]: DEBUG nova.virt.hardware [None req-ae076e22-617f-47ad-a346-b93664495097 tempest-ServersNegativeTestMultiTenantJSON-944098855 tempest-ServersNegativeTestMultiTenantJSON-944098855-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 815.285048] env[65726]: DEBUG nova.virt.hardware [None req-ae076e22-617f-47ad-a346-b93664495097 tempest-ServersNegativeTestMultiTenantJSON-944098855 tempest-ServersNegativeTestMultiTenantJSON-944098855-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 815.285229] env[65726]: DEBUG nova.virt.hardware [None req-ae076e22-617f-47ad-a346-b93664495097 tempest-ServersNegativeTestMultiTenantJSON-944098855 tempest-ServersNegativeTestMultiTenantJSON-944098855-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 815.285372] env[65726]: DEBUG nova.virt.hardware [None req-ae076e22-617f-47ad-a346-b93664495097 tempest-ServersNegativeTestMultiTenantJSON-944098855 tempest-ServersNegativeTestMultiTenantJSON-944098855-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 815.285515] env[65726]: DEBUG nova.virt.hardware [None req-ae076e22-617f-47ad-a346-b93664495097 tempest-ServersNegativeTestMultiTenantJSON-944098855 tempest-ServersNegativeTestMultiTenantJSON-944098855-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 815.285718] env[65726]: DEBUG nova.virt.hardware [None req-ae076e22-617f-47ad-a346-b93664495097 tempest-ServersNegativeTestMultiTenantJSON-944098855 tempest-ServersNegativeTestMultiTenantJSON-944098855-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 815.285861] env[65726]: DEBUG nova.virt.hardware [None req-ae076e22-617f-47ad-a346-b93664495097 tempest-ServersNegativeTestMultiTenantJSON-944098855 tempest-ServersNegativeTestMultiTenantJSON-944098855-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 815.286035] env[65726]: DEBUG nova.virt.hardware [None req-ae076e22-617f-47ad-a346-b93664495097 tempest-ServersNegativeTestMultiTenantJSON-944098855 tempest-ServersNegativeTestMultiTenantJSON-944098855-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 815.286205] env[65726]: DEBUG nova.virt.hardware [None req-ae076e22-617f-47ad-a346-b93664495097 tempest-ServersNegativeTestMultiTenantJSON-944098855 tempest-ServersNegativeTestMultiTenantJSON-944098855-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 815.286373] env[65726]: DEBUG nova.virt.hardware [None req-ae076e22-617f-47ad-a346-b93664495097 tempest-ServersNegativeTestMultiTenantJSON-944098855 tempest-ServersNegativeTestMultiTenantJSON-944098855-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 815.287246] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62e28b5f-57ab-4671-a6e9-0df45debf22c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.298601] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a880793f-4fb0-43aa-9861-5fe8350d4dba {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.306761] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f1456b1-9c93-4cab-821d-31e73fb19819 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.330555] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a6fe25d-66eb-4079-96b4-87da5012fdbf {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.365648] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78c04571-6506-45d8-9d12-e6968fd7756a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.377417] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f402f27-69ad-48ef-9ff4-16e9d39402e5 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.395090] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-c3e5127a-ba4f-4e76-986d-1724943cb777 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 608cc0ea-3f6b-4b4f-83c1-01aa50999d63] Creating linked-clone VM from snapshot {{(pid=65726) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 815.395523] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-4ad6563b-fb50-40c9-b02b-6860f669b3a6 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.416345] env[65726]: DEBUG nova.compute.provider_tree [None req-81594943-9858-4765-a371-42f9ef0cbec9 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 815.424798] env[65726]: DEBUG oslo_vmware.api [None req-c3e5127a-ba4f-4e76-986d-1724943cb777 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Waiting for the task: (returnval){ [ 815.424798] env[65726]: value = "task-5115944" [ 815.424798] env[65726]: _type = "Task" [ 815.424798] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.437150] env[65726]: DEBUG oslo_vmware.api [None req-a76cfabd-af7f-4569-b968-716d45bcf7f1 tempest-InstanceActionsNegativeTestJSON-844675881 tempest-InstanceActionsNegativeTestJSON-844675881-project-member] Task: {'id': task-5115941, 'name': ReconfigVM_Task, 'duration_secs': 0.83905} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 815.437600] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-a76cfabd-af7f-4569-b968-716d45bcf7f1 tempest-InstanceActionsNegativeTestJSON-844675881 tempest-InstanceActionsNegativeTestJSON-844675881-project-member] [instance: 1bd26aef-995e-43b6-af9f-077e878c8c44] Reconfigured VM instance instance-00000031 to attach disk [datastore2] 1bd26aef-995e-43b6-af9f-077e878c8c44/1bd26aef-995e-43b6-af9f-077e878c8c44.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 815.438519] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-50314a67-658c-4521-b8f8-25af5525c897 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.446506] env[65726]: DEBUG oslo_vmware.api [None req-c3e5127a-ba4f-4e76-986d-1724943cb777 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5115944, 'name': CloneVM_Task} progress is 10%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.457497] env[65726]: DEBUG oslo_vmware.api [None req-855996fc-ecea-49d8-a71b-8b4307731732 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Task: {'id': task-5115943, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.459268] env[65726]: DEBUG oslo_vmware.api [None req-a76cfabd-af7f-4569-b968-716d45bcf7f1 tempest-InstanceActionsNegativeTestJSON-844675881 tempest-InstanceActionsNegativeTestJSON-844675881-project-member] Waiting for the task: (returnval){ [ 815.459268] env[65726]: value = "task-5115945" [ 815.459268] env[65726]: _type = "Task" [ 815.459268] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.475646] env[65726]: DEBUG oslo_vmware.api [None req-a76cfabd-af7f-4569-b968-716d45bcf7f1 tempest-InstanceActionsNegativeTestJSON-844675881 tempest-InstanceActionsNegativeTestJSON-844675881-project-member] Task: {'id': task-5115945, 'name': Rename_Task} progress is 10%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.485304] env[65726]: WARNING openstack [req-8660dbe1-3675-4ab9-826a-0f55c51d6e09 req-7a640ba2-57af-4f31-b674-e0efbdfac702 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 815.485976] env[65726]: WARNING openstack [req-8660dbe1-3675-4ab9-826a-0f55c51d6e09 req-7a640ba2-57af-4f31-b674-e0efbdfac702 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 815.528774] env[65726]: WARNING openstack [req-b57f9b61-2b31-4154-9b89-8c7097d3e874 req-44c53ae4-827b-4b8e-935a-7913af33d4c0 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 815.529219] env[65726]: WARNING openstack [req-b57f9b61-2b31-4154-9b89-8c7097d3e874 req-44c53ae4-827b-4b8e-935a-7913af33d4c0 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 815.773679] env[65726]: DEBUG oslo_vmware.api [None req-85226e56-668f-4c9e-b188-d90c53166469 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': task-5115942, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.927126] env[65726]: DEBUG nova.scheduler.client.report [None req-81594943-9858-4765-a371-42f9ef0cbec9 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 815.941061] env[65726]: DEBUG oslo_vmware.api [None req-c3e5127a-ba4f-4e76-986d-1724943cb777 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5115944, 'name': CloneVM_Task} progress is 94%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.957933] env[65726]: DEBUG oslo_vmware.api [None req-855996fc-ecea-49d8-a71b-8b4307731732 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Task: {'id': task-5115943, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.754309} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 815.958231] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-855996fc-ecea-49d8-a71b-8b4307731732 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore2] 6c8db442-daaa-4eca-ae24-b9d4afaf3a71/6c8db442-daaa-4eca-ae24-b9d4afaf3a71.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 815.958652] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-855996fc-ecea-49d8-a71b-8b4307731732 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] [instance: 6c8db442-daaa-4eca-ae24-b9d4afaf3a71] Extending root virtual disk to 1048576 {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 815.959201] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-dd6cfafe-90ff-4fc0-a703-de44dfd9467a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.971662] env[65726]: DEBUG oslo_vmware.api [None req-a76cfabd-af7f-4569-b968-716d45bcf7f1 tempest-InstanceActionsNegativeTestJSON-844675881 tempest-InstanceActionsNegativeTestJSON-844675881-project-member] Task: {'id': task-5115945, 'name': Rename_Task, 'duration_secs': 0.396118} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 815.974114] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-a76cfabd-af7f-4569-b968-716d45bcf7f1 tempest-InstanceActionsNegativeTestJSON-844675881 tempest-InstanceActionsNegativeTestJSON-844675881-project-member] [instance: 1bd26aef-995e-43b6-af9f-077e878c8c44] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 815.974114] env[65726]: DEBUG oslo_vmware.api [None req-855996fc-ecea-49d8-a71b-8b4307731732 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Waiting for the task: (returnval){ [ 815.974114] env[65726]: value = "task-5115946" [ 815.974114] env[65726]: _type = "Task" [ 815.974114] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.974114] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-509faef8-6ee3-4f76-8414-a5b3e77620e8 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.985538] env[65726]: DEBUG oslo_vmware.api [None req-855996fc-ecea-49d8-a71b-8b4307731732 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Task: {'id': task-5115946, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.987397] env[65726]: DEBUG oslo_vmware.api [None req-a76cfabd-af7f-4569-b968-716d45bcf7f1 tempest-InstanceActionsNegativeTestJSON-844675881 tempest-InstanceActionsNegativeTestJSON-844675881-project-member] Waiting for the task: (returnval){ [ 815.987397] env[65726]: value = "task-5115947" [ 815.987397] env[65726]: _type = "Task" [ 815.987397] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.997078] env[65726]: DEBUG oslo_vmware.api [None req-a76cfabd-af7f-4569-b968-716d45bcf7f1 tempest-InstanceActionsNegativeTestJSON-844675881 tempest-InstanceActionsNegativeTestJSON-844675881-project-member] Task: {'id': task-5115947, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.239238] env[65726]: DEBUG nova.network.neutron [None req-ae076e22-617f-47ad-a346-b93664495097 tempest-ServersNegativeTestMultiTenantJSON-944098855 tempest-ServersNegativeTestMultiTenantJSON-944098855-project-member] [instance: 018c8ac5-66c4-4a9a-ab45-85e84f92b4fe] Successfully updated port: 5a241c1a-c51d-49fc-8387-7dd3cf7f38af {{(pid=65726) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 816.278399] env[65726]: DEBUG oslo_vmware.api [None req-85226e56-668f-4c9e-b188-d90c53166469 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': task-5115942, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.372126] env[65726]: WARNING openstack [req-b57f9b61-2b31-4154-9b89-8c7097d3e874 req-44c53ae4-827b-4b8e-935a-7913af33d4c0 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 816.372565] env[65726]: WARNING openstack [req-b57f9b61-2b31-4154-9b89-8c7097d3e874 req-44c53ae4-827b-4b8e-935a-7913af33d4c0 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 816.385745] env[65726]: WARNING openstack [req-8660dbe1-3675-4ab9-826a-0f55c51d6e09 req-7a640ba2-57af-4f31-b674-e0efbdfac702 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 816.386181] env[65726]: WARNING openstack [req-8660dbe1-3675-4ab9-826a-0f55c51d6e09 req-7a640ba2-57af-4f31-b674-e0efbdfac702 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 816.436025] env[65726]: DEBUG oslo_concurrency.lockutils [None req-81594943-9858-4765-a371-42f9ef0cbec9 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.742s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 816.436521] env[65726]: DEBUG nova.compute.manager [None req-81594943-9858-4765-a371-42f9ef0cbec9 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] [instance: 3cbbe146-be3a-43d5-867d-4d669884758c] Start building networks asynchronously for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 816.440209] env[65726]: DEBUG oslo_concurrency.lockutils [None req-0b112b08-1593-41b5-b38e-86e5b4d22081 tempest-ServersV294TestFqdnHostnames-819811677 tempest-ServersV294TestFqdnHostnames-819811677-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 39.521s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 816.444062] env[65726]: INFO nova.compute.claims [None req-0b112b08-1593-41b5-b38e-86e5b4d22081 tempest-ServersV294TestFqdnHostnames-819811677 tempest-ServersV294TestFqdnHostnames-819811677-project-member] [instance: c8be64a3-1bfa-41a3-b9be-68a7e6e052e0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 816.452180] env[65726]: DEBUG oslo_vmware.api [None req-c3e5127a-ba4f-4e76-986d-1724943cb777 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5115944, 'name': CloneVM_Task} progress is 94%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.487038] env[65726]: DEBUG oslo_vmware.api [None req-855996fc-ecea-49d8-a71b-8b4307731732 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Task: {'id': task-5115946, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.180727} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.487425] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-855996fc-ecea-49d8-a71b-8b4307731732 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] [instance: 6c8db442-daaa-4eca-ae24-b9d4afaf3a71] Extended root virtual disk {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 816.488766] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c23b04f9-fb8a-473a-9bd7-7b8f84855cf2 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.523413] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-855996fc-ecea-49d8-a71b-8b4307731732 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] [instance: 6c8db442-daaa-4eca-ae24-b9d4afaf3a71] Reconfiguring VM instance instance-00000032 to attach disk [datastore2] 6c8db442-daaa-4eca-ae24-b9d4afaf3a71/6c8db442-daaa-4eca-ae24-b9d4afaf3a71.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 816.523593] env[65726]: DEBUG oslo_vmware.api [None req-a76cfabd-af7f-4569-b968-716d45bcf7f1 tempest-InstanceActionsNegativeTestJSON-844675881 tempest-InstanceActionsNegativeTestJSON-844675881-project-member] Task: {'id': task-5115947, 'name': PowerOnVM_Task, 'duration_secs': 0.516756} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.525290] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d8086235-4f2a-440e-a851-c028b682b6a7 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.543660] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-a76cfabd-af7f-4569-b968-716d45bcf7f1 tempest-InstanceActionsNegativeTestJSON-844675881 tempest-InstanceActionsNegativeTestJSON-844675881-project-member] [instance: 1bd26aef-995e-43b6-af9f-077e878c8c44] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 816.543897] env[65726]: INFO nova.compute.manager [None req-a76cfabd-af7f-4569-b968-716d45bcf7f1 tempest-InstanceActionsNegativeTestJSON-844675881 tempest-InstanceActionsNegativeTestJSON-844675881-project-member] [instance: 1bd26aef-995e-43b6-af9f-077e878c8c44] Took 8.33 seconds to spawn the instance on the hypervisor. [ 816.544320] env[65726]: DEBUG nova.compute.manager [None req-a76cfabd-af7f-4569-b968-716d45bcf7f1 tempest-InstanceActionsNegativeTestJSON-844675881 tempest-InstanceActionsNegativeTestJSON-844675881-project-member] [instance: 1bd26aef-995e-43b6-af9f-077e878c8c44] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 816.547148] env[65726]: DEBUG nova.network.neutron [req-b57f9b61-2b31-4154-9b89-8c7097d3e874 req-44c53ae4-827b-4b8e-935a-7913af33d4c0 service nova] [instance: 6c8db442-daaa-4eca-ae24-b9d4afaf3a71] Updated VIF entry in instance network info cache for port 6f422d37-8262-454d-aa2d-1569bacee1b7. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 816.548888] env[65726]: DEBUG nova.network.neutron [req-b57f9b61-2b31-4154-9b89-8c7097d3e874 req-44c53ae4-827b-4b8e-935a-7913af33d4c0 service nova] [instance: 6c8db442-daaa-4eca-ae24-b9d4afaf3a71] Updating instance_info_cache with network_info: [{"id": "6f422d37-8262-454d-aa2d-1569bacee1b7", "address": "fa:16:3e:19:b2:c8", "network": {"id": "d37113a1-409c-4c2c-81f9-8f80822b85b9", "bridge": "br-int", "label": "tempest-ServersWithSpecificFlavorTestJSON-471231219-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "79d3e960f190439fae7c61d693e9356e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "13af9422-d668-4413-b63a-766558d83a3b", "external-id": "nsx-vlan-transportzone-842", "segmentation_id": 842, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6f422d37-82", "ovs_interfaceid": "6f422d37-8262-454d-aa2d-1569bacee1b7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 816.551460] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed380663-2448-4a81-a1bf-b38d1a159d1c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.567461] env[65726]: DEBUG oslo_vmware.api [None req-855996fc-ecea-49d8-a71b-8b4307731732 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Waiting for the task: (returnval){ [ 816.567461] env[65726]: value = "task-5115948" [ 816.567461] env[65726]: _type = "Task" [ 816.567461] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.577584] env[65726]: DEBUG oslo_vmware.api [None req-855996fc-ecea-49d8-a71b-8b4307731732 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Task: {'id': task-5115948, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.624897] env[65726]: WARNING openstack [req-8660dbe1-3675-4ab9-826a-0f55c51d6e09 req-7a640ba2-57af-4f31-b674-e0efbdfac702 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 816.625359] env[65726]: WARNING openstack [req-8660dbe1-3675-4ab9-826a-0f55c51d6e09 req-7a640ba2-57af-4f31-b674-e0efbdfac702 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 816.745705] env[65726]: DEBUG oslo_concurrency.lockutils [None req-ae076e22-617f-47ad-a346-b93664495097 tempest-ServersNegativeTestMultiTenantJSON-944098855 tempest-ServersNegativeTestMultiTenantJSON-944098855-project-member] Acquiring lock "refresh_cache-018c8ac5-66c4-4a9a-ab45-85e84f92b4fe" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 816.745907] env[65726]: DEBUG oslo_concurrency.lockutils [None req-ae076e22-617f-47ad-a346-b93664495097 tempest-ServersNegativeTestMultiTenantJSON-944098855 tempest-ServersNegativeTestMultiTenantJSON-944098855-project-member] Acquired lock "refresh_cache-018c8ac5-66c4-4a9a-ab45-85e84f92b4fe" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 816.746098] env[65726]: DEBUG nova.network.neutron [None req-ae076e22-617f-47ad-a346-b93664495097 tempest-ServersNegativeTestMultiTenantJSON-944098855 tempest-ServersNegativeTestMultiTenantJSON-944098855-project-member] [instance: 018c8ac5-66c4-4a9a-ab45-85e84f92b4fe] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 816.778065] env[65726]: DEBUG oslo_vmware.api [None req-85226e56-668f-4c9e-b188-d90c53166469 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': task-5115942, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.943151] env[65726]: DEBUG oslo_vmware.api [None req-c3e5127a-ba4f-4e76-986d-1724943cb777 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5115944, 'name': CloneVM_Task} progress is 100%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.947526] env[65726]: DEBUG nova.compute.utils [None req-81594943-9858-4765-a371-42f9ef0cbec9 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] Using /dev/sd instead of None {{(pid=65726) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 816.952373] env[65726]: DEBUG nova.compute.manager [None req-81594943-9858-4765-a371-42f9ef0cbec9 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] [instance: 3cbbe146-be3a-43d5-867d-4d669884758c] Allocating IP information in the background. {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 816.952373] env[65726]: DEBUG nova.network.neutron [None req-81594943-9858-4765-a371-42f9ef0cbec9 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] [instance: 3cbbe146-be3a-43d5-867d-4d669884758c] allocate_for_instance() {{(pid=65726) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 816.952373] env[65726]: WARNING neutronclient.v2_0.client [None req-81594943-9858-4765-a371-42f9ef0cbec9 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 816.952561] env[65726]: WARNING neutronclient.v2_0.client [None req-81594943-9858-4765-a371-42f9ef0cbec9 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 816.953258] env[65726]: WARNING openstack [None req-81594943-9858-4765-a371-42f9ef0cbec9 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 816.956025] env[65726]: WARNING openstack [None req-81594943-9858-4765-a371-42f9ef0cbec9 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 817.011025] env[65726]: DEBUG nova.network.neutron [req-8660dbe1-3675-4ab9-826a-0f55c51d6e09 req-7a640ba2-57af-4f31-b674-e0efbdfac702 service nova] [instance: bc2c12e0-0d06-432f-b42f-be468e3b6ee4] Updated VIF entry in instance network info cache for port a1c6ed0b-f1f6-4da3-aeaf-9f815ef90b64. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 817.011025] env[65726]: DEBUG nova.network.neutron [req-8660dbe1-3675-4ab9-826a-0f55c51d6e09 req-7a640ba2-57af-4f31-b674-e0efbdfac702 service nova] [instance: bc2c12e0-0d06-432f-b42f-be468e3b6ee4] Updating instance_info_cache with network_info: [{"id": "a1c6ed0b-f1f6-4da3-aeaf-9f815ef90b64", "address": "fa:16:3e:09:eb:ee", "network": {"id": "dd4bbaef-0a1e-4e00-b65a-f19e7a837cc1", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-255385230-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "55f93cdf51e84cf683879668c2acecf4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "04ccbc7a-cf8d-4ea2-8411-291a1e27df7b", "external-id": "nsx-vlan-transportzone-998", "segmentation_id": 998, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa1c6ed0b-f1", "ovs_interfaceid": "a1c6ed0b-f1f6-4da3-aeaf-9f815ef90b64", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 817.056423] env[65726]: DEBUG oslo_concurrency.lockutils [req-b57f9b61-2b31-4154-9b89-8c7097d3e874 req-44c53ae4-827b-4b8e-935a-7913af33d4c0 service nova] Releasing lock "refresh_cache-6c8db442-daaa-4eca-ae24-b9d4afaf3a71" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 817.082323] env[65726]: INFO nova.compute.manager [None req-a76cfabd-af7f-4569-b968-716d45bcf7f1 tempest-InstanceActionsNegativeTestJSON-844675881 tempest-InstanceActionsNegativeTestJSON-844675881-project-member] [instance: 1bd26aef-995e-43b6-af9f-077e878c8c44] Took 51.42 seconds to build instance. [ 817.089213] env[65726]: DEBUG oslo_vmware.api [None req-855996fc-ecea-49d8-a71b-8b4307731732 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Task: {'id': task-5115948, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.250604] env[65726]: WARNING openstack [None req-ae076e22-617f-47ad-a346-b93664495097 tempest-ServersNegativeTestMultiTenantJSON-944098855 tempest-ServersNegativeTestMultiTenantJSON-944098855-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 817.251186] env[65726]: WARNING openstack [None req-ae076e22-617f-47ad-a346-b93664495097 tempest-ServersNegativeTestMultiTenantJSON-944098855 tempest-ServersNegativeTestMultiTenantJSON-944098855-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 817.260445] env[65726]: DEBUG nova.network.neutron [None req-ae076e22-617f-47ad-a346-b93664495097 tempest-ServersNegativeTestMultiTenantJSON-944098855 tempest-ServersNegativeTestMultiTenantJSON-944098855-project-member] [instance: 018c8ac5-66c4-4a9a-ab45-85e84f92b4fe] Instance cache missing network info. {{(pid=65726) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 817.276190] env[65726]: DEBUG oslo_vmware.api [None req-85226e56-668f-4c9e-b188-d90c53166469 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': task-5115942, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.298212] env[65726]: DEBUG nova.policy [None req-81594943-9858-4765-a371-42f9ef0cbec9 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '63038bd048c744a987997a8315870eeb', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '92c1cedbc08545d58fc7172ac95e2c23', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65726) authorize /opt/stack/nova/nova/policy.py:192}} [ 817.444699] env[65726]: DEBUG oslo_vmware.api [None req-c3e5127a-ba4f-4e76-986d-1724943cb777 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5115944, 'name': CloneVM_Task, 'duration_secs': 1.538417} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.445566] env[65726]: INFO nova.virt.vmwareapi.vmops [None req-c3e5127a-ba4f-4e76-986d-1724943cb777 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 608cc0ea-3f6b-4b4f-83c1-01aa50999d63] Created linked-clone VM from snapshot [ 817.446539] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2a7283e-584f-4510-b588-4f2e5581b518 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.453567] env[65726]: DEBUG nova.compute.manager [None req-81594943-9858-4765-a371-42f9ef0cbec9 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] [instance: 3cbbe146-be3a-43d5-867d-4d669884758c] Start building block device mappings for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 817.460104] env[65726]: DEBUG nova.virt.vmwareapi.images [None req-c3e5127a-ba4f-4e76-986d-1724943cb777 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 608cc0ea-3f6b-4b4f-83c1-01aa50999d63] Uploading image 122c2b99-e3fb-499d-a847-a4c088b862f5 {{(pid=65726) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 817.497199] env[65726]: DEBUG oslo_vmware.rw_handles [None req-c3e5127a-ba4f-4e76-986d-1724943cb777 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 817.497199] env[65726]: value = "vm-995147" [ 817.497199] env[65726]: _type = "VirtualMachine" [ 817.497199] env[65726]: }. {{(pid=65726) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 817.497537] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-b6aae829-bbdc-4415-91cc-e02f18b8565c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.506373] env[65726]: DEBUG oslo_vmware.rw_handles [None req-c3e5127a-ba4f-4e76-986d-1724943cb777 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Lease: (returnval){ [ 817.506373] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]523334ec-eb3a-e2cf-b07f-1d5ad99658a9" [ 817.506373] env[65726]: _type = "HttpNfcLease" [ 817.506373] env[65726]: } obtained for exporting VM: (result){ [ 817.506373] env[65726]: value = "vm-995147" [ 817.506373] env[65726]: _type = "VirtualMachine" [ 817.506373] env[65726]: }. {{(pid=65726) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 817.506650] env[65726]: DEBUG oslo_vmware.api [None req-c3e5127a-ba4f-4e76-986d-1724943cb777 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Waiting for the lease: (returnval){ [ 817.506650] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]523334ec-eb3a-e2cf-b07f-1d5ad99658a9" [ 817.506650] env[65726]: _type = "HttpNfcLease" [ 817.506650] env[65726]: } to be ready. {{(pid=65726) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 817.512952] env[65726]: DEBUG oslo_concurrency.lockutils [req-8660dbe1-3675-4ab9-826a-0f55c51d6e09 req-7a640ba2-57af-4f31-b674-e0efbdfac702 service nova] Releasing lock "refresh_cache-bc2c12e0-0d06-432f-b42f-be468e3b6ee4" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 817.515065] env[65726]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 817.515065] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]523334ec-eb3a-e2cf-b07f-1d5ad99658a9" [ 817.515065] env[65726]: _type = "HttpNfcLease" [ 817.515065] env[65726]: } is initializing. {{(pid=65726) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 817.571374] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de02e5c9-0e54-4326-adcd-f124a1d97661 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.585086] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ba6ddd1-252a-4c97-8db3-7ec76147159d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.588993] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a76cfabd-af7f-4569-b968-716d45bcf7f1 tempest-InstanceActionsNegativeTestJSON-844675881 tempest-InstanceActionsNegativeTestJSON-844675881-project-member] Lock "1bd26aef-995e-43b6-af9f-077e878c8c44" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 87.101s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 817.589361] env[65726]: DEBUG oslo_vmware.api [None req-855996fc-ecea-49d8-a71b-8b4307731732 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Task: {'id': task-5115948, 'name': ReconfigVM_Task, 'duration_secs': 0.755365} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.590616] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-855996fc-ecea-49d8-a71b-8b4307731732 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] [instance: 6c8db442-daaa-4eca-ae24-b9d4afaf3a71] Reconfigured VM instance instance-00000032 to attach disk [datastore2] 6c8db442-daaa-4eca-ae24-b9d4afaf3a71/6c8db442-daaa-4eca-ae24-b9d4afaf3a71.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 817.590930] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-855996fc-ecea-49d8-a71b-8b4307731732 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Creating Virtual Disk of size 1048576 KB and adapter type paraVirtual on the data store {{(pid=65726) create_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1383}} [ 817.593251] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CreateVirtualDisk_Task with opID=oslo.vmware-ea14d68a-a064-4e9b-9504-9f06b9713d69 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.623914] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-566ae425-addf-423f-b0ee-67603c59ce55 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.628498] env[65726]: DEBUG oslo_vmware.api [None req-855996fc-ecea-49d8-a71b-8b4307731732 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Waiting for the task: (returnval){ [ 817.628498] env[65726]: value = "task-5115950" [ 817.628498] env[65726]: _type = "Task" [ 817.628498] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 817.636463] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1678add-dcac-4c94-acd6-57a7e9ce0c8b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.642426] env[65726]: DEBUG nova.network.neutron [None req-81594943-9858-4765-a371-42f9ef0cbec9 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] [instance: 3cbbe146-be3a-43d5-867d-4d669884758c] Successfully created port: 40240c4e-9f3e-46aa-bd75-c62c2776d9b6 {{(pid=65726) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 817.652789] env[65726]: DEBUG oslo_vmware.api [None req-855996fc-ecea-49d8-a71b-8b4307731732 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Task: {'id': task-5115950, 'name': CreateVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.661824] env[65726]: DEBUG nova.compute.provider_tree [None req-0b112b08-1593-41b5-b38e-86e5b4d22081 tempest-ServersV294TestFqdnHostnames-819811677 tempest-ServersV294TestFqdnHostnames-819811677-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 817.682949] env[65726]: WARNING openstack [None req-ae076e22-617f-47ad-a346-b93664495097 tempest-ServersNegativeTestMultiTenantJSON-944098855 tempest-ServersNegativeTestMultiTenantJSON-944098855-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 817.683314] env[65726]: WARNING openstack [None req-ae076e22-617f-47ad-a346-b93664495097 tempest-ServersNegativeTestMultiTenantJSON-944098855 tempest-ServersNegativeTestMultiTenantJSON-944098855-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 817.780638] env[65726]: DEBUG oslo_vmware.api [None req-85226e56-668f-4c9e-b188-d90c53166469 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': task-5115942, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.827885] env[65726]: WARNING openstack [None req-ae076e22-617f-47ad-a346-b93664495097 tempest-ServersNegativeTestMultiTenantJSON-944098855 tempest-ServersNegativeTestMultiTenantJSON-944098855-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 817.828302] env[65726]: WARNING openstack [None req-ae076e22-617f-47ad-a346-b93664495097 tempest-ServersNegativeTestMultiTenantJSON-944098855 tempest-ServersNegativeTestMultiTenantJSON-944098855-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 817.942818] env[65726]: DEBUG nova.network.neutron [None req-ae076e22-617f-47ad-a346-b93664495097 tempest-ServersNegativeTestMultiTenantJSON-944098855 tempest-ServersNegativeTestMultiTenantJSON-944098855-project-member] [instance: 018c8ac5-66c4-4a9a-ab45-85e84f92b4fe] Updating instance_info_cache with network_info: [{"id": "5a241c1a-c51d-49fc-8387-7dd3cf7f38af", "address": "fa:16:3e:c7:57:f7", "network": {"id": "f9e2bbb2-f31e-4160-a1c2-d6d4079859ed", "bridge": "br-int", "label": "tempest-ServersNegativeTestMultiTenantJSON-1422595507-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6fc182ce253b4da98ffe8e8d9d224156", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ea00b53a-9c9b-4592-ab95-7e10473f338d", "external-id": "nsx-vlan-transportzone-235", "segmentation_id": 235, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5a241c1a-c5", "ovs_interfaceid": "5a241c1a-c51d-49fc-8387-7dd3cf7f38af", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 818.014697] env[65726]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 818.014697] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]523334ec-eb3a-e2cf-b07f-1d5ad99658a9" [ 818.014697] env[65726]: _type = "HttpNfcLease" [ 818.014697] env[65726]: } is ready. {{(pid=65726) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 818.015011] env[65726]: DEBUG oslo_vmware.rw_handles [None req-c3e5127a-ba4f-4e76-986d-1724943cb777 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 818.015011] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]523334ec-eb3a-e2cf-b07f-1d5ad99658a9" [ 818.015011] env[65726]: _type = "HttpNfcLease" [ 818.015011] env[65726]: }. {{(pid=65726) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 818.016584] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82343fa3-fb42-4942-adce-f1c410500612 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.025049] env[65726]: DEBUG oslo_vmware.rw_handles [None req-c3e5127a-ba4f-4e76-986d-1724943cb777 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5270d0a4-127e-6e1d-921f-59f48964187b/disk-0.vmdk from lease info. {{(pid=65726) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 818.025233] env[65726]: DEBUG oslo_vmware.rw_handles [None req-c3e5127a-ba4f-4e76-986d-1724943cb777 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5270d0a4-127e-6e1d-921f-59f48964187b/disk-0.vmdk for reading. {{(pid=65726) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 818.092372] env[65726]: DEBUG nova.compute.manager [None req-cb2d094a-412f-492d-9ef1-aba8757b7146 tempest-ServerMetadataNegativeTestJSON-936475568 tempest-ServerMetadataNegativeTestJSON-936475568-project-member] [instance: d179030a-e5ba-45b1-ad11-4d2f71ed7bd4] Starting instance... {{(pid=65726) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 818.121516] env[65726]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-3cda8179-7bac-4b6c-a7ad-c5f96ad3e285 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.146202] env[65726]: DEBUG oslo_vmware.api [None req-855996fc-ecea-49d8-a71b-8b4307731732 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Task: {'id': task-5115950, 'name': CreateVirtualDisk_Task, 'duration_secs': 0.057102} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 818.151018] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-855996fc-ecea-49d8-a71b-8b4307731732 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Created Virtual Disk of size 1048576 KB and type thin {{(pid=65726) create_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1404}} [ 818.155122] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92072399-e371-4cfd-88d7-389132ebf853 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.179355] env[65726]: DEBUG nova.scheduler.client.report [None req-0b112b08-1593-41b5-b38e-86e5b4d22081 tempest-ServersV294TestFqdnHostnames-819811677 tempest-ServersV294TestFqdnHostnames-819811677-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 818.190926] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-855996fc-ecea-49d8-a71b-8b4307731732 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] [instance: 6c8db442-daaa-4eca-ae24-b9d4afaf3a71] Reconfiguring VM instance instance-00000032 to attach disk [datastore2] 6c8db442-daaa-4eca-ae24-b9d4afaf3a71/ephemeral_0.vmdk or device None with type thin {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 818.196694] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-18105e82-30c9-442e-939a-d6e8cef7143d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.213767] env[65726]: DEBUG nova.compute.manager [req-2e1beea3-0cb4-4885-b6dd-be0109e02193 req-4572596d-ede6-4372-8371-945f0c36cbac service nova] [instance: 018c8ac5-66c4-4a9a-ab45-85e84f92b4fe] Received event network-vif-plugged-5a241c1a-c51d-49fc-8387-7dd3cf7f38af {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 818.213767] env[65726]: DEBUG oslo_concurrency.lockutils [req-2e1beea3-0cb4-4885-b6dd-be0109e02193 req-4572596d-ede6-4372-8371-945f0c36cbac service nova] Acquiring lock "018c8ac5-66c4-4a9a-ab45-85e84f92b4fe-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 818.213945] env[65726]: DEBUG oslo_concurrency.lockutils [req-2e1beea3-0cb4-4885-b6dd-be0109e02193 req-4572596d-ede6-4372-8371-945f0c36cbac service nova] Lock "018c8ac5-66c4-4a9a-ab45-85e84f92b4fe-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 818.214534] env[65726]: DEBUG oslo_concurrency.lockutils [req-2e1beea3-0cb4-4885-b6dd-be0109e02193 req-4572596d-ede6-4372-8371-945f0c36cbac service nova] Lock "018c8ac5-66c4-4a9a-ab45-85e84f92b4fe-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 818.214534] env[65726]: DEBUG nova.compute.manager [req-2e1beea3-0cb4-4885-b6dd-be0109e02193 req-4572596d-ede6-4372-8371-945f0c36cbac service nova] [instance: 018c8ac5-66c4-4a9a-ab45-85e84f92b4fe] No waiting events found dispatching network-vif-plugged-5a241c1a-c51d-49fc-8387-7dd3cf7f38af {{(pid=65726) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 818.214534] env[65726]: WARNING nova.compute.manager [req-2e1beea3-0cb4-4885-b6dd-be0109e02193 req-4572596d-ede6-4372-8371-945f0c36cbac service nova] [instance: 018c8ac5-66c4-4a9a-ab45-85e84f92b4fe] Received unexpected event network-vif-plugged-5a241c1a-c51d-49fc-8387-7dd3cf7f38af for instance with vm_state building and task_state spawning. [ 818.214534] env[65726]: DEBUG nova.compute.manager [req-2e1beea3-0cb4-4885-b6dd-be0109e02193 req-4572596d-ede6-4372-8371-945f0c36cbac service nova] [instance: 018c8ac5-66c4-4a9a-ab45-85e84f92b4fe] Received event network-changed-5a241c1a-c51d-49fc-8387-7dd3cf7f38af {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 818.216775] env[65726]: DEBUG nova.compute.manager [req-2e1beea3-0cb4-4885-b6dd-be0109e02193 req-4572596d-ede6-4372-8371-945f0c36cbac service nova] [instance: 018c8ac5-66c4-4a9a-ab45-85e84f92b4fe] Refreshing instance network info cache due to event network-changed-5a241c1a-c51d-49fc-8387-7dd3cf7f38af. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 818.216775] env[65726]: DEBUG oslo_concurrency.lockutils [req-2e1beea3-0cb4-4885-b6dd-be0109e02193 req-4572596d-ede6-4372-8371-945f0c36cbac service nova] Acquiring lock "refresh_cache-018c8ac5-66c4-4a9a-ab45-85e84f92b4fe" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 818.222926] env[65726]: DEBUG oslo_vmware.api [None req-855996fc-ecea-49d8-a71b-8b4307731732 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Waiting for the task: (returnval){ [ 818.222926] env[65726]: value = "task-5115951" [ 818.222926] env[65726]: _type = "Task" [ 818.222926] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 818.235469] env[65726]: DEBUG oslo_vmware.api [None req-855996fc-ecea-49d8-a71b-8b4307731732 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Task: {'id': task-5115951, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.277724] env[65726]: DEBUG oslo_vmware.api [None req-85226e56-668f-4c9e-b188-d90c53166469 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': task-5115942, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.446159] env[65726]: DEBUG oslo_concurrency.lockutils [None req-ae076e22-617f-47ad-a346-b93664495097 tempest-ServersNegativeTestMultiTenantJSON-944098855 tempest-ServersNegativeTestMultiTenantJSON-944098855-project-member] Releasing lock "refresh_cache-018c8ac5-66c4-4a9a-ab45-85e84f92b4fe" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 818.446537] env[65726]: DEBUG nova.compute.manager [None req-ae076e22-617f-47ad-a346-b93664495097 tempest-ServersNegativeTestMultiTenantJSON-944098855 tempest-ServersNegativeTestMultiTenantJSON-944098855-project-member] [instance: 018c8ac5-66c4-4a9a-ab45-85e84f92b4fe] Instance network_info: |[{"id": "5a241c1a-c51d-49fc-8387-7dd3cf7f38af", "address": "fa:16:3e:c7:57:f7", "network": {"id": "f9e2bbb2-f31e-4160-a1c2-d6d4079859ed", "bridge": "br-int", "label": "tempest-ServersNegativeTestMultiTenantJSON-1422595507-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6fc182ce253b4da98ffe8e8d9d224156", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ea00b53a-9c9b-4592-ab95-7e10473f338d", "external-id": "nsx-vlan-transportzone-235", "segmentation_id": 235, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5a241c1a-c5", "ovs_interfaceid": "5a241c1a-c51d-49fc-8387-7dd3cf7f38af", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 818.447159] env[65726]: DEBUG oslo_concurrency.lockutils [req-2e1beea3-0cb4-4885-b6dd-be0109e02193 req-4572596d-ede6-4372-8371-945f0c36cbac service nova] Acquired lock "refresh_cache-018c8ac5-66c4-4a9a-ab45-85e84f92b4fe" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 818.447159] env[65726]: DEBUG nova.network.neutron [req-2e1beea3-0cb4-4885-b6dd-be0109e02193 req-4572596d-ede6-4372-8371-945f0c36cbac service nova] [instance: 018c8ac5-66c4-4a9a-ab45-85e84f92b4fe] Refreshing network info cache for port 5a241c1a-c51d-49fc-8387-7dd3cf7f38af {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 818.448173] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-ae076e22-617f-47ad-a346-b93664495097 tempest-ServersNegativeTestMultiTenantJSON-944098855 tempest-ServersNegativeTestMultiTenantJSON-944098855-project-member] [instance: 018c8ac5-66c4-4a9a-ab45-85e84f92b4fe] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c7:57:f7', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ea00b53a-9c9b-4592-ab95-7e10473f338d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '5a241c1a-c51d-49fc-8387-7dd3cf7f38af', 'vif_model': 'vmxnet3'}] {{(pid=65726) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 818.455923] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae076e22-617f-47ad-a346-b93664495097 tempest-ServersNegativeTestMultiTenantJSON-944098855 tempest-ServersNegativeTestMultiTenantJSON-944098855-project-member] Creating folder: Project (6fc182ce253b4da98ffe8e8d9d224156). Parent ref: group-v995008. {{(pid=65726) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 818.456625] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-87a5f7bc-dd94-4fdc-9bb2-07f4ce164d24 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.475173] env[65726]: DEBUG nova.compute.manager [None req-81594943-9858-4765-a371-42f9ef0cbec9 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] [instance: 3cbbe146-be3a-43d5-867d-4d669884758c] Start spawning the instance on the hypervisor. {{(pid=65726) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 818.478207] env[65726]: INFO nova.virt.vmwareapi.vm_util [None req-ae076e22-617f-47ad-a346-b93664495097 tempest-ServersNegativeTestMultiTenantJSON-944098855 tempest-ServersNegativeTestMultiTenantJSON-944098855-project-member] Created folder: Project (6fc182ce253b4da98ffe8e8d9d224156) in parent group-v995008. [ 818.478362] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae076e22-617f-47ad-a346-b93664495097 tempest-ServersNegativeTestMultiTenantJSON-944098855 tempest-ServersNegativeTestMultiTenantJSON-944098855-project-member] Creating folder: Instances. Parent ref: group-v995148. {{(pid=65726) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 818.479291] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-01252c38-abd6-4cb7-92f4-938a92332741 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.493415] env[65726]: INFO nova.virt.vmwareapi.vm_util [None req-ae076e22-617f-47ad-a346-b93664495097 tempest-ServersNegativeTestMultiTenantJSON-944098855 tempest-ServersNegativeTestMultiTenantJSON-944098855-project-member] Created folder: Instances in parent group-v995148. [ 818.493909] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-ae076e22-617f-47ad-a346-b93664495097 tempest-ServersNegativeTestMultiTenantJSON-944098855 tempest-ServersNegativeTestMultiTenantJSON-944098855-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 818.496501] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 018c8ac5-66c4-4a9a-ab45-85e84f92b4fe] Creating VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 818.497660] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-02841b87-9c61-4609-9325-74a5c65a5cc9 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.516518] env[65726]: DEBUG nova.virt.hardware [None req-81594943-9858-4765-a371-42f9ef0cbec9 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T19:28:34Z,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1539f06fa3534e90acbc3a60c4b8bd3f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T19:28:35Z,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 818.516846] env[65726]: DEBUG nova.virt.hardware [None req-81594943-9858-4765-a371-42f9ef0cbec9 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 818.517093] env[65726]: DEBUG nova.virt.hardware [None req-81594943-9858-4765-a371-42f9ef0cbec9 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 818.517343] env[65726]: DEBUG nova.virt.hardware [None req-81594943-9858-4765-a371-42f9ef0cbec9 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 818.517539] env[65726]: DEBUG nova.virt.hardware [None req-81594943-9858-4765-a371-42f9ef0cbec9 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 818.517734] env[65726]: DEBUG nova.virt.hardware [None req-81594943-9858-4765-a371-42f9ef0cbec9 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 818.517984] env[65726]: DEBUG nova.virt.hardware [None req-81594943-9858-4765-a371-42f9ef0cbec9 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 818.518250] env[65726]: DEBUG nova.virt.hardware [None req-81594943-9858-4765-a371-42f9ef0cbec9 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 818.518446] env[65726]: DEBUG nova.virt.hardware [None req-81594943-9858-4765-a371-42f9ef0cbec9 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 818.518668] env[65726]: DEBUG nova.virt.hardware [None req-81594943-9858-4765-a371-42f9ef0cbec9 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 818.518868] env[65726]: DEBUG nova.virt.hardware [None req-81594943-9858-4765-a371-42f9ef0cbec9 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 818.519991] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43caa640-42ab-48a8-ad00-6eb3f06a3edc {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.531544] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7fb7ca5-837f-46d2-ba73-917d5093679d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.538496] env[65726]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 818.538496] env[65726]: value = "task-5115954" [ 818.538496] env[65726]: _type = "Task" [ 818.538496] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 818.558668] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115954, 'name': CreateVM_Task} progress is 6%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.624844] env[65726]: DEBUG oslo_concurrency.lockutils [None req-cb2d094a-412f-492d-9ef1-aba8757b7146 tempest-ServerMetadataNegativeTestJSON-936475568 tempest-ServerMetadataNegativeTestJSON-936475568-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 818.698810] env[65726]: DEBUG oslo_concurrency.lockutils [None req-0b112b08-1593-41b5-b38e-86e5b4d22081 tempest-ServersV294TestFqdnHostnames-819811677 tempest-ServersV294TestFqdnHostnames-819811677-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.259s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 818.699620] env[65726]: DEBUG nova.compute.manager [None req-0b112b08-1593-41b5-b38e-86e5b4d22081 tempest-ServersV294TestFqdnHostnames-819811677 tempest-ServersV294TestFqdnHostnames-819811677-project-member] [instance: c8be64a3-1bfa-41a3-b9be-68a7e6e052e0] Start building networks asynchronously for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 818.704051] env[65726]: DEBUG oslo_concurrency.lockutils [None req-276e84bf-a104-4323-ae2b-430f1d490aad tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 41.083s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 818.706334] env[65726]: DEBUG nova.objects.instance [None req-276e84bf-a104-4323-ae2b-430f1d490aad tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] [instance: 3827eed1-b7a1-4922-a4de-f0c038cb8566] Trying to apply a migration context that does not seem to be set for this instance {{(pid=65726) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 818.735115] env[65726]: DEBUG oslo_vmware.api [None req-855996fc-ecea-49d8-a71b-8b4307731732 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Task: {'id': task-5115951, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.783913] env[65726]: DEBUG oslo_vmware.api [None req-85226e56-668f-4c9e-b188-d90c53166469 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': task-5115942, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.958199] env[65726]: WARNING openstack [req-2e1beea3-0cb4-4885-b6dd-be0109e02193 req-4572596d-ede6-4372-8371-945f0c36cbac service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 818.959313] env[65726]: WARNING openstack [req-2e1beea3-0cb4-4885-b6dd-be0109e02193 req-4572596d-ede6-4372-8371-945f0c36cbac service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 819.054206] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115954, 'name': CreateVM_Task, 'duration_secs': 0.431306} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 819.054641] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 018c8ac5-66c4-4a9a-ab45-85e84f92b4fe] Created VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 819.055580] env[65726]: WARNING openstack [None req-ae076e22-617f-47ad-a346-b93664495097 tempest-ServersNegativeTestMultiTenantJSON-944098855 tempest-ServersNegativeTestMultiTenantJSON-944098855-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 819.056085] env[65726]: WARNING openstack [None req-ae076e22-617f-47ad-a346-b93664495097 tempest-ServersNegativeTestMultiTenantJSON-944098855 tempest-ServersNegativeTestMultiTenantJSON-944098855-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 819.062066] env[65726]: DEBUG oslo_concurrency.lockutils [None req-ae076e22-617f-47ad-a346-b93664495097 tempest-ServersNegativeTestMultiTenantJSON-944098855 tempest-ServersNegativeTestMultiTenantJSON-944098855-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 819.062337] env[65726]: DEBUG oslo_concurrency.lockutils [None req-ae076e22-617f-47ad-a346-b93664495097 tempest-ServersNegativeTestMultiTenantJSON-944098855 tempest-ServersNegativeTestMultiTenantJSON-944098855-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 819.062726] env[65726]: DEBUG oslo_concurrency.lockutils [None req-ae076e22-617f-47ad-a346-b93664495097 tempest-ServersNegativeTestMultiTenantJSON-944098855 tempest-ServersNegativeTestMultiTenantJSON-944098855-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 819.063090] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ffb8ea1d-9633-473d-96cc-01abce45112b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.069620] env[65726]: DEBUG oslo_vmware.api [None req-ae076e22-617f-47ad-a346-b93664495097 tempest-ServersNegativeTestMultiTenantJSON-944098855 tempest-ServersNegativeTestMultiTenantJSON-944098855-project-member] Waiting for the task: (returnval){ [ 819.069620] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5266c23c-f827-8085-339c-7ae42906b6ea" [ 819.069620] env[65726]: _type = "Task" [ 819.069620] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 819.080624] env[65726]: DEBUG oslo_vmware.api [None req-ae076e22-617f-47ad-a346-b93664495097 tempest-ServersNegativeTestMultiTenantJSON-944098855 tempest-ServersNegativeTestMultiTenantJSON-944098855-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5266c23c-f827-8085-339c-7ae42906b6ea, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.212738] env[65726]: DEBUG nova.compute.utils [None req-0b112b08-1593-41b5-b38e-86e5b4d22081 tempest-ServersV294TestFqdnHostnames-819811677 tempest-ServersV294TestFqdnHostnames-819811677-project-member] Using /dev/sd instead of None {{(pid=65726) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 819.217851] env[65726]: DEBUG nova.compute.manager [None req-0b112b08-1593-41b5-b38e-86e5b4d22081 tempest-ServersV294TestFqdnHostnames-819811677 tempest-ServersV294TestFqdnHostnames-819811677-project-member] [instance: c8be64a3-1bfa-41a3-b9be-68a7e6e052e0] Allocating IP information in the background. {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 819.219200] env[65726]: DEBUG nova.network.neutron [None req-0b112b08-1593-41b5-b38e-86e5b4d22081 tempest-ServersV294TestFqdnHostnames-819811677 tempest-ServersV294TestFqdnHostnames-819811677-project-member] [instance: c8be64a3-1bfa-41a3-b9be-68a7e6e052e0] allocate_for_instance() {{(pid=65726) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 819.219200] env[65726]: WARNING neutronclient.v2_0.client [None req-0b112b08-1593-41b5-b38e-86e5b4d22081 tempest-ServersV294TestFqdnHostnames-819811677 tempest-ServersV294TestFqdnHostnames-819811677-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 819.219200] env[65726]: WARNING neutronclient.v2_0.client [None req-0b112b08-1593-41b5-b38e-86e5b4d22081 tempest-ServersV294TestFqdnHostnames-819811677 tempest-ServersV294TestFqdnHostnames-819811677-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 819.219792] env[65726]: WARNING openstack [None req-0b112b08-1593-41b5-b38e-86e5b4d22081 tempest-ServersV294TestFqdnHostnames-819811677 tempest-ServersV294TestFqdnHostnames-819811677-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 819.220259] env[65726]: WARNING openstack [None req-0b112b08-1593-41b5-b38e-86e5b4d22081 tempest-ServersV294TestFqdnHostnames-819811677 tempest-ServersV294TestFqdnHostnames-819811677-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 819.240164] env[65726]: DEBUG oslo_vmware.api [None req-855996fc-ecea-49d8-a71b-8b4307731732 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Task: {'id': task-5115951, 'name': ReconfigVM_Task, 'duration_secs': 0.557287} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 819.240501] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-855996fc-ecea-49d8-a71b-8b4307731732 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] [instance: 6c8db442-daaa-4eca-ae24-b9d4afaf3a71] Reconfigured VM instance instance-00000032 to attach disk [datastore2] 6c8db442-daaa-4eca-ae24-b9d4afaf3a71/ephemeral_0.vmdk or device None with type thin {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 819.242923] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e6de6a68-a605-4a31-9d02-18c60f545098 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.249269] env[65726]: DEBUG oslo_vmware.api [None req-855996fc-ecea-49d8-a71b-8b4307731732 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Waiting for the task: (returnval){ [ 819.249269] env[65726]: value = "task-5115955" [ 819.249269] env[65726]: _type = "Task" [ 819.249269] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 819.262501] env[65726]: DEBUG oslo_vmware.api [None req-855996fc-ecea-49d8-a71b-8b4307731732 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Task: {'id': task-5115955, 'name': Rename_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.279207] env[65726]: DEBUG oslo_vmware.api [None req-85226e56-668f-4c9e-b188-d90c53166469 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': task-5115942, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.305793] env[65726]: DEBUG nova.network.neutron [None req-81594943-9858-4765-a371-42f9ef0cbec9 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] [instance: 3cbbe146-be3a-43d5-867d-4d669884758c] Successfully updated port: 40240c4e-9f3e-46aa-bd75-c62c2776d9b6 {{(pid=65726) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 819.396067] env[65726]: DEBUG nova.policy [None req-0b112b08-1593-41b5-b38e-86e5b4d22081 tempest-ServersV294TestFqdnHostnames-819811677 tempest-ServersV294TestFqdnHostnames-819811677-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'da2b0e957e104e8ebf5f2f8589768082', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b8855e0d30be40b49e32a3eb0b6eef30', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65726) authorize /opt/stack/nova/nova/policy.py:192}} [ 819.582173] env[65726]: DEBUG oslo_vmware.api [None req-ae076e22-617f-47ad-a346-b93664495097 tempest-ServersNegativeTestMultiTenantJSON-944098855 tempest-ServersNegativeTestMultiTenantJSON-944098855-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5266c23c-f827-8085-339c-7ae42906b6ea, 'name': SearchDatastore_Task, 'duration_secs': 0.014104} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 819.582173] env[65726]: DEBUG oslo_concurrency.lockutils [None req-ae076e22-617f-47ad-a346-b93664495097 tempest-ServersNegativeTestMultiTenantJSON-944098855 tempest-ServersNegativeTestMultiTenantJSON-944098855-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 819.582173] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-ae076e22-617f-47ad-a346-b93664495097 tempest-ServersNegativeTestMultiTenantJSON-944098855 tempest-ServersNegativeTestMultiTenantJSON-944098855-project-member] [instance: 018c8ac5-66c4-4a9a-ab45-85e84f92b4fe] Processing image b52362a3-ee8a-4cbf-b06f-513b0cc8f95c {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 819.582680] env[65726]: DEBUG oslo_concurrency.lockutils [None req-ae076e22-617f-47ad-a346-b93664495097 tempest-ServersNegativeTestMultiTenantJSON-944098855 tempest-ServersNegativeTestMultiTenantJSON-944098855-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 819.582680] env[65726]: DEBUG oslo_concurrency.lockutils [None req-ae076e22-617f-47ad-a346-b93664495097 tempest-ServersNegativeTestMultiTenantJSON-944098855 tempest-ServersNegativeTestMultiTenantJSON-944098855-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 819.582680] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-ae076e22-617f-47ad-a346-b93664495097 tempest-ServersNegativeTestMultiTenantJSON-944098855 tempest-ServersNegativeTestMultiTenantJSON-944098855-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 819.583035] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-046d535d-f065-4d81-9656-4823df6e2ed0 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.598020] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-ae076e22-617f-47ad-a346-b93664495097 tempest-ServersNegativeTestMultiTenantJSON-944098855 tempest-ServersNegativeTestMultiTenantJSON-944098855-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 819.598430] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-ae076e22-617f-47ad-a346-b93664495097 tempest-ServersNegativeTestMultiTenantJSON-944098855 tempest-ServersNegativeTestMultiTenantJSON-944098855-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65726) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 819.599666] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-632adef4-1105-4d29-a44e-4401a8c4da7e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.608894] env[65726]: DEBUG oslo_vmware.api [None req-ae076e22-617f-47ad-a346-b93664495097 tempest-ServersNegativeTestMultiTenantJSON-944098855 tempest-ServersNegativeTestMultiTenantJSON-944098855-project-member] Waiting for the task: (returnval){ [ 819.608894] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52a52774-fe7e-7190-083b-6b4a23fcbdc9" [ 819.608894] env[65726]: _type = "Task" [ 819.608894] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 819.621271] env[65726]: DEBUG oslo_vmware.api [None req-ae076e22-617f-47ad-a346-b93664495097 tempest-ServersNegativeTestMultiTenantJSON-944098855 tempest-ServersNegativeTestMultiTenantJSON-944098855-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52a52774-fe7e-7190-083b-6b4a23fcbdc9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.687656] env[65726]: WARNING openstack [req-2e1beea3-0cb4-4885-b6dd-be0109e02193 req-4572596d-ede6-4372-8371-945f0c36cbac service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 819.688311] env[65726]: WARNING openstack [req-2e1beea3-0cb4-4885-b6dd-be0109e02193 req-4572596d-ede6-4372-8371-945f0c36cbac service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 819.718095] env[65726]: DEBUG nova.compute.manager [None req-0b112b08-1593-41b5-b38e-86e5b4d22081 tempest-ServersV294TestFqdnHostnames-819811677 tempest-ServersV294TestFqdnHostnames-819811677-project-member] [instance: c8be64a3-1bfa-41a3-b9be-68a7e6e052e0] Start building block device mappings for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 819.731023] env[65726]: DEBUG oslo_concurrency.lockutils [None req-276e84bf-a104-4323-ae2b-430f1d490aad tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.026s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 819.734145] env[65726]: DEBUG oslo_concurrency.lockutils [None req-7b44120c-3963-45cf-8eb3-04f0d2e08d90 tempest-ServerDiagnosticsNegativeTest-1635132145 tempest-ServerDiagnosticsNegativeTest-1635132145-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 41.115s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 819.734145] env[65726]: DEBUG nova.objects.instance [None req-7b44120c-3963-45cf-8eb3-04f0d2e08d90 tempest-ServerDiagnosticsNegativeTest-1635132145 tempest-ServerDiagnosticsNegativeTest-1635132145-project-member] Lazy-loading 'resources' on Instance uuid 0d6f14c5-5770-476f-a186-4cf1b2d4fbe8 {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 819.766929] env[65726]: DEBUG oslo_vmware.api [None req-855996fc-ecea-49d8-a71b-8b4307731732 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Task: {'id': task-5115955, 'name': Rename_Task, 'duration_secs': 0.206261} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 819.767919] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-855996fc-ecea-49d8-a71b-8b4307731732 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] [instance: 6c8db442-daaa-4eca-ae24-b9d4afaf3a71] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 819.768291] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-de81ee4f-3e2b-4edc-8262-4405d396810b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.781635] env[65726]: DEBUG oslo_vmware.api [None req-85226e56-668f-4c9e-b188-d90c53166469 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': task-5115942, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.783877] env[65726]: DEBUG oslo_vmware.api [None req-855996fc-ecea-49d8-a71b-8b4307731732 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Waiting for the task: (returnval){ [ 819.783877] env[65726]: value = "task-5115956" [ 819.783877] env[65726]: _type = "Task" [ 819.783877] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 819.801289] env[65726]: DEBUG oslo_vmware.api [None req-855996fc-ecea-49d8-a71b-8b4307731732 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Task: {'id': task-5115956, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.807176] env[65726]: DEBUG oslo_concurrency.lockutils [None req-81594943-9858-4765-a371-42f9ef0cbec9 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] Acquiring lock "refresh_cache-3cbbe146-be3a-43d5-867d-4d669884758c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 819.807362] env[65726]: DEBUG oslo_concurrency.lockutils [None req-81594943-9858-4765-a371-42f9ef0cbec9 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] Acquired lock "refresh_cache-3cbbe146-be3a-43d5-867d-4d669884758c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 819.807790] env[65726]: DEBUG nova.network.neutron [None req-81594943-9858-4765-a371-42f9ef0cbec9 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] [instance: 3cbbe146-be3a-43d5-867d-4d669884758c] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 819.949801] env[65726]: WARNING openstack [req-2e1beea3-0cb4-4885-b6dd-be0109e02193 req-4572596d-ede6-4372-8371-945f0c36cbac service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 819.950688] env[65726]: WARNING openstack [req-2e1beea3-0cb4-4885-b6dd-be0109e02193 req-4572596d-ede6-4372-8371-945f0c36cbac service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 820.076726] env[65726]: DEBUG nova.network.neutron [None req-0b112b08-1593-41b5-b38e-86e5b4d22081 tempest-ServersV294TestFqdnHostnames-819811677 tempest-ServersV294TestFqdnHostnames-819811677-project-member] [instance: c8be64a3-1bfa-41a3-b9be-68a7e6e052e0] Successfully created port: e9d6f425-5466-4d5d-ae7f-80a433defe57 {{(pid=65726) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 820.123383] env[65726]: DEBUG oslo_vmware.api [None req-ae076e22-617f-47ad-a346-b93664495097 tempest-ServersNegativeTestMultiTenantJSON-944098855 tempest-ServersNegativeTestMultiTenantJSON-944098855-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52a52774-fe7e-7190-083b-6b4a23fcbdc9, 'name': SearchDatastore_Task, 'duration_secs': 0.025709} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.124738] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bd32311e-0a64-4726-8a55-967224a4c2c7 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.133544] env[65726]: DEBUG oslo_vmware.api [None req-ae076e22-617f-47ad-a346-b93664495097 tempest-ServersNegativeTestMultiTenantJSON-944098855 tempest-ServersNegativeTestMultiTenantJSON-944098855-project-member] Waiting for the task: (returnval){ [ 820.133544] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52b151b1-5236-c544-6218-d12b51844385" [ 820.133544] env[65726]: _type = "Task" [ 820.133544] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 820.150965] env[65726]: DEBUG oslo_vmware.api [None req-ae076e22-617f-47ad-a346-b93664495097 tempest-ServersNegativeTestMultiTenantJSON-944098855 tempest-ServersNegativeTestMultiTenantJSON-944098855-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52b151b1-5236-c544-6218-d12b51844385, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.177565] env[65726]: DEBUG nova.network.neutron [req-2e1beea3-0cb4-4885-b6dd-be0109e02193 req-4572596d-ede6-4372-8371-945f0c36cbac service nova] [instance: 018c8ac5-66c4-4a9a-ab45-85e84f92b4fe] Updated VIF entry in instance network info cache for port 5a241c1a-c51d-49fc-8387-7dd3cf7f38af. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 820.177930] env[65726]: DEBUG nova.network.neutron [req-2e1beea3-0cb4-4885-b6dd-be0109e02193 req-4572596d-ede6-4372-8371-945f0c36cbac service nova] [instance: 018c8ac5-66c4-4a9a-ab45-85e84f92b4fe] Updating instance_info_cache with network_info: [{"id": "5a241c1a-c51d-49fc-8387-7dd3cf7f38af", "address": "fa:16:3e:c7:57:f7", "network": {"id": "f9e2bbb2-f31e-4160-a1c2-d6d4079859ed", "bridge": "br-int", "label": "tempest-ServersNegativeTestMultiTenantJSON-1422595507-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6fc182ce253b4da98ffe8e8d9d224156", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ea00b53a-9c9b-4592-ab95-7e10473f338d", "external-id": "nsx-vlan-transportzone-235", "segmentation_id": 235, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5a241c1a-c5", "ovs_interfaceid": "5a241c1a-c51d-49fc-8387-7dd3cf7f38af", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 820.288542] env[65726]: DEBUG oslo_vmware.api [None req-85226e56-668f-4c9e-b188-d90c53166469 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': task-5115942, 'name': ReconfigVM_Task} progress is 18%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.308892] env[65726]: DEBUG oslo_vmware.api [None req-855996fc-ecea-49d8-a71b-8b4307731732 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Task: {'id': task-5115956, 'name': PowerOnVM_Task} progress is 88%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.312837] env[65726]: WARNING openstack [None req-81594943-9858-4765-a371-42f9ef0cbec9 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 820.313211] env[65726]: WARNING openstack [None req-81594943-9858-4765-a371-42f9ef0cbec9 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 820.319523] env[65726]: DEBUG nova.network.neutron [None req-81594943-9858-4765-a371-42f9ef0cbec9 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] [instance: 3cbbe146-be3a-43d5-867d-4d669884758c] Instance cache missing network info. {{(pid=65726) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 820.390331] env[65726]: WARNING openstack [None req-81594943-9858-4765-a371-42f9ef0cbec9 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 820.393784] env[65726]: WARNING openstack [None req-81594943-9858-4765-a371-42f9ef0cbec9 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 820.526167] env[65726]: WARNING openstack [None req-81594943-9858-4765-a371-42f9ef0cbec9 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 820.526746] env[65726]: WARNING openstack [None req-81594943-9858-4765-a371-42f9ef0cbec9 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 820.650152] env[65726]: DEBUG oslo_vmware.api [None req-ae076e22-617f-47ad-a346-b93664495097 tempest-ServersNegativeTestMultiTenantJSON-944098855 tempest-ServersNegativeTestMultiTenantJSON-944098855-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52b151b1-5236-c544-6218-d12b51844385, 'name': SearchDatastore_Task, 'duration_secs': 0.014451} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.653296] env[65726]: DEBUG nova.network.neutron [None req-81594943-9858-4765-a371-42f9ef0cbec9 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] [instance: 3cbbe146-be3a-43d5-867d-4d669884758c] Updating instance_info_cache with network_info: [{"id": "40240c4e-9f3e-46aa-bd75-c62c2776d9b6", "address": "fa:16:3e:8f:bd:09", "network": {"id": "2266eb6e-4c8f-4a3f-85d3-0c06022c0d5e", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-716412627-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "92c1cedbc08545d58fc7172ac95e2c23", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "24376631-ee89-4ff1-b8ac-f09911fc8329", "external-id": "nsx-vlan-transportzone-960", "segmentation_id": 960, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap40240c4e-9f", "ovs_interfaceid": "40240c4e-9f3e-46aa-bd75-c62c2776d9b6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 820.656263] env[65726]: DEBUG oslo_concurrency.lockutils [None req-ae076e22-617f-47ad-a346-b93664495097 tempest-ServersNegativeTestMultiTenantJSON-944098855 tempest-ServersNegativeTestMultiTenantJSON-944098855-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 820.656263] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae076e22-617f-47ad-a346-b93664495097 tempest-ServersNegativeTestMultiTenantJSON-944098855 tempest-ServersNegativeTestMultiTenantJSON-944098855-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore2] 018c8ac5-66c4-4a9a-ab45-85e84f92b4fe/018c8ac5-66c4-4a9a-ab45-85e84f92b4fe.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 820.656263] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f7bb53dd-c8f8-4438-8202-b5134cf9c8bb {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.666280] env[65726]: DEBUG oslo_vmware.api [None req-ae076e22-617f-47ad-a346-b93664495097 tempest-ServersNegativeTestMultiTenantJSON-944098855 tempest-ServersNegativeTestMultiTenantJSON-944098855-project-member] Waiting for the task: (returnval){ [ 820.666280] env[65726]: value = "task-5115957" [ 820.666280] env[65726]: _type = "Task" [ 820.666280] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 820.683031] env[65726]: DEBUG oslo_concurrency.lockutils [req-2e1beea3-0cb4-4885-b6dd-be0109e02193 req-4572596d-ede6-4372-8371-945f0c36cbac service nova] Releasing lock "refresh_cache-018c8ac5-66c4-4a9a-ab45-85e84f92b4fe" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 820.684132] env[65726]: DEBUG oslo_vmware.api [None req-ae076e22-617f-47ad-a346-b93664495097 tempest-ServersNegativeTestMultiTenantJSON-944098855 tempest-ServersNegativeTestMultiTenantJSON-944098855-project-member] Task: {'id': task-5115957, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.734202] env[65726]: DEBUG nova.compute.manager [None req-0b112b08-1593-41b5-b38e-86e5b4d22081 tempest-ServersV294TestFqdnHostnames-819811677 tempest-ServersV294TestFqdnHostnames-819811677-project-member] [instance: c8be64a3-1bfa-41a3-b9be-68a7e6e052e0] Start spawning the instance on the hypervisor. {{(pid=65726) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 820.763181] env[65726]: DEBUG nova.virt.hardware [None req-0b112b08-1593-41b5-b38e-86e5b4d22081 tempest-ServersV294TestFqdnHostnames-819811677 tempest-ServersV294TestFqdnHostnames-819811677-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T19:28:34Z,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1539f06fa3534e90acbc3a60c4b8bd3f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T19:28:35Z,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 820.763441] env[65726]: DEBUG nova.virt.hardware [None req-0b112b08-1593-41b5-b38e-86e5b4d22081 tempest-ServersV294TestFqdnHostnames-819811677 tempest-ServersV294TestFqdnHostnames-819811677-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 820.763604] env[65726]: DEBUG nova.virt.hardware [None req-0b112b08-1593-41b5-b38e-86e5b4d22081 tempest-ServersV294TestFqdnHostnames-819811677 tempest-ServersV294TestFqdnHostnames-819811677-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 820.763782] env[65726]: DEBUG nova.virt.hardware [None req-0b112b08-1593-41b5-b38e-86e5b4d22081 tempest-ServersV294TestFqdnHostnames-819811677 tempest-ServersV294TestFqdnHostnames-819811677-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 820.763934] env[65726]: DEBUG nova.virt.hardware [None req-0b112b08-1593-41b5-b38e-86e5b4d22081 tempest-ServersV294TestFqdnHostnames-819811677 tempest-ServersV294TestFqdnHostnames-819811677-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 820.764088] env[65726]: DEBUG nova.virt.hardware [None req-0b112b08-1593-41b5-b38e-86e5b4d22081 tempest-ServersV294TestFqdnHostnames-819811677 tempest-ServersV294TestFqdnHostnames-819811677-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 820.764302] env[65726]: DEBUG nova.virt.hardware [None req-0b112b08-1593-41b5-b38e-86e5b4d22081 tempest-ServersV294TestFqdnHostnames-819811677 tempest-ServersV294TestFqdnHostnames-819811677-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 820.764468] env[65726]: DEBUG nova.virt.hardware [None req-0b112b08-1593-41b5-b38e-86e5b4d22081 tempest-ServersV294TestFqdnHostnames-819811677 tempest-ServersV294TestFqdnHostnames-819811677-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 820.764640] env[65726]: DEBUG nova.virt.hardware [None req-0b112b08-1593-41b5-b38e-86e5b4d22081 tempest-ServersV294TestFqdnHostnames-819811677 tempest-ServersV294TestFqdnHostnames-819811677-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 820.764798] env[65726]: DEBUG nova.virt.hardware [None req-0b112b08-1593-41b5-b38e-86e5b4d22081 tempest-ServersV294TestFqdnHostnames-819811677 tempest-ServersV294TestFqdnHostnames-819811677-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 820.765115] env[65726]: DEBUG nova.virt.hardware [None req-0b112b08-1593-41b5-b38e-86e5b4d22081 tempest-ServersV294TestFqdnHostnames-819811677 tempest-ServersV294TestFqdnHostnames-819811677-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 820.766242] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45dd743c-6cbf-48e1-be5a-f7d372fbd5d4 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.783505] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fededf67-cbe2-4f13-9f92-aafd938a758e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.788927] env[65726]: DEBUG nova.compute.manager [req-b2148284-fd08-4267-ade8-0b14b32ffa5f req-e43f44f9-9e24-4b6c-9de3-1eacc01d8a7a service nova] [instance: 3cbbe146-be3a-43d5-867d-4d669884758c] Received event network-vif-plugged-40240c4e-9f3e-46aa-bd75-c62c2776d9b6 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 820.789143] env[65726]: DEBUG oslo_concurrency.lockutils [req-b2148284-fd08-4267-ade8-0b14b32ffa5f req-e43f44f9-9e24-4b6c-9de3-1eacc01d8a7a service nova] Acquiring lock "3cbbe146-be3a-43d5-867d-4d669884758c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 820.789350] env[65726]: DEBUG oslo_concurrency.lockutils [req-b2148284-fd08-4267-ade8-0b14b32ffa5f req-e43f44f9-9e24-4b6c-9de3-1eacc01d8a7a service nova] Lock "3cbbe146-be3a-43d5-867d-4d669884758c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 820.789523] env[65726]: DEBUG oslo_concurrency.lockutils [req-b2148284-fd08-4267-ade8-0b14b32ffa5f req-e43f44f9-9e24-4b6c-9de3-1eacc01d8a7a service nova] Lock "3cbbe146-be3a-43d5-867d-4d669884758c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 820.789684] env[65726]: DEBUG nova.compute.manager [req-b2148284-fd08-4267-ade8-0b14b32ffa5f req-e43f44f9-9e24-4b6c-9de3-1eacc01d8a7a service nova] [instance: 3cbbe146-be3a-43d5-867d-4d669884758c] No waiting events found dispatching network-vif-plugged-40240c4e-9f3e-46aa-bd75-c62c2776d9b6 {{(pid=65726) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 820.789855] env[65726]: WARNING nova.compute.manager [req-b2148284-fd08-4267-ade8-0b14b32ffa5f req-e43f44f9-9e24-4b6c-9de3-1eacc01d8a7a service nova] [instance: 3cbbe146-be3a-43d5-867d-4d669884758c] Received unexpected event network-vif-plugged-40240c4e-9f3e-46aa-bd75-c62c2776d9b6 for instance with vm_state building and task_state spawning. [ 820.808074] env[65726]: DEBUG oslo_vmware.api [None req-85226e56-668f-4c9e-b188-d90c53166469 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': task-5115942, 'name': ReconfigVM_Task, 'duration_secs': 5.910509} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.811213] env[65726]: DEBUG oslo_concurrency.lockutils [None req-85226e56-668f-4c9e-b188-d90c53166469 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Releasing lock "49bcab1f-7b64-4999-abff-37771c58a271" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 820.811461] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-85226e56-668f-4c9e-b188-d90c53166469 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: 49bcab1f-7b64-4999-abff-37771c58a271] Reconfigured VM to detach interface {{(pid=65726) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1978}} [ 820.812154] env[65726]: WARNING openstack [None req-85226e56-668f-4c9e-b188-d90c53166469 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 820.812518] env[65726]: WARNING openstack [None req-85226e56-668f-4c9e-b188-d90c53166469 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 820.824441] env[65726]: DEBUG oslo_vmware.api [None req-855996fc-ecea-49d8-a71b-8b4307731732 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Task: {'id': task-5115956, 'name': PowerOnVM_Task, 'duration_secs': 0.753477} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.824731] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-855996fc-ecea-49d8-a71b-8b4307731732 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] [instance: 6c8db442-daaa-4eca-ae24-b9d4afaf3a71] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 820.824942] env[65726]: INFO nova.compute.manager [None req-855996fc-ecea-49d8-a71b-8b4307731732 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] [instance: 6c8db442-daaa-4eca-ae24-b9d4afaf3a71] Took 9.49 seconds to spawn the instance on the hypervisor. [ 820.825149] env[65726]: DEBUG nova.compute.manager [None req-855996fc-ecea-49d8-a71b-8b4307731732 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] [instance: 6c8db442-daaa-4eca-ae24-b9d4afaf3a71] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 820.826165] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e282c4c1-0c60-4a6b-a0a5-34908b01fcbd {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.883185] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f81b2590-a32c-48a2-9a0b-fdba73cac5f8 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.892903] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3c1e60d-23f1-4248-b650-f605a66e69fd {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.933530] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ba2269e-a22c-4d23-800d-c0d58d72cd24 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.944679] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04e4d4a4-a5bf-401c-8a4f-4ec5719e5b22 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.968758] env[65726]: DEBUG nova.compute.provider_tree [None req-7b44120c-3963-45cf-8eb3-04f0d2e08d90 tempest-ServerDiagnosticsNegativeTest-1635132145 tempest-ServerDiagnosticsNegativeTest-1635132145-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 821.034341] env[65726]: DEBUG oslo_concurrency.lockutils [None req-73b0a627-5d81-4bf2-833a-caff3534c1e3 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Acquiring lock "cf950ec3-9914-4eb9-99db-048b3969bb21" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 821.034639] env[65726]: DEBUG oslo_concurrency.lockutils [None req-73b0a627-5d81-4bf2-833a-caff3534c1e3 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Lock "cf950ec3-9914-4eb9-99db-048b3969bb21" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 821.164357] env[65726]: DEBUG oslo_concurrency.lockutils [None req-81594943-9858-4765-a371-42f9ef0cbec9 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] Releasing lock "refresh_cache-3cbbe146-be3a-43d5-867d-4d669884758c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 821.164810] env[65726]: DEBUG nova.compute.manager [None req-81594943-9858-4765-a371-42f9ef0cbec9 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] [instance: 3cbbe146-be3a-43d5-867d-4d669884758c] Instance network_info: |[{"id": "40240c4e-9f3e-46aa-bd75-c62c2776d9b6", "address": "fa:16:3e:8f:bd:09", "network": {"id": "2266eb6e-4c8f-4a3f-85d3-0c06022c0d5e", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-716412627-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "92c1cedbc08545d58fc7172ac95e2c23", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "24376631-ee89-4ff1-b8ac-f09911fc8329", "external-id": "nsx-vlan-transportzone-960", "segmentation_id": 960, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap40240c4e-9f", "ovs_interfaceid": "40240c4e-9f3e-46aa-bd75-c62c2776d9b6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 821.166033] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-81594943-9858-4765-a371-42f9ef0cbec9 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] [instance: 3cbbe146-be3a-43d5-867d-4d669884758c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:8f:bd:09', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '24376631-ee89-4ff1-b8ac-f09911fc8329', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '40240c4e-9f3e-46aa-bd75-c62c2776d9b6', 'vif_model': 'vmxnet3'}] {{(pid=65726) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 821.176288] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-81594943-9858-4765-a371-42f9ef0cbec9 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] Creating folder: Project (92c1cedbc08545d58fc7172ac95e2c23). Parent ref: group-v995008. {{(pid=65726) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 821.177634] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-111ced43-c08b-4354-9913-9c2fc2b71667 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.192911] env[65726]: DEBUG oslo_vmware.api [None req-ae076e22-617f-47ad-a346-b93664495097 tempest-ServersNegativeTestMultiTenantJSON-944098855 tempest-ServersNegativeTestMultiTenantJSON-944098855-project-member] Task: {'id': task-5115957, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.199545] env[65726]: INFO nova.virt.vmwareapi.vm_util [None req-81594943-9858-4765-a371-42f9ef0cbec9 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] Created folder: Project (92c1cedbc08545d58fc7172ac95e2c23) in parent group-v995008. [ 821.199788] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-81594943-9858-4765-a371-42f9ef0cbec9 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] Creating folder: Instances. Parent ref: group-v995151. {{(pid=65726) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 821.200149] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9e24f798-604c-423e-a427-c4bf567f8668 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.215563] env[65726]: INFO nova.virt.vmwareapi.vm_util [None req-81594943-9858-4765-a371-42f9ef0cbec9 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] Created folder: Instances in parent group-v995151. [ 821.215889] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-81594943-9858-4765-a371-42f9ef0cbec9 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 821.216137] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3cbbe146-be3a-43d5-867d-4d669884758c] Creating VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 821.216397] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-00f5f51e-e63a-4589-93aa-a860af37ed1f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.242031] env[65726]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 821.242031] env[65726]: value = "task-5115960" [ 821.242031] env[65726]: _type = "Task" [ 821.242031] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 821.252813] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115960, 'name': CreateVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.357414] env[65726]: INFO nova.compute.manager [None req-855996fc-ecea-49d8-a71b-8b4307731732 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] [instance: 6c8db442-daaa-4eca-ae24-b9d4afaf3a71] Took 55.09 seconds to build instance. [ 821.472977] env[65726]: DEBUG nova.scheduler.client.report [None req-7b44120c-3963-45cf-8eb3-04f0d2e08d90 tempest-ServerDiagnosticsNegativeTest-1635132145 tempest-ServerDiagnosticsNegativeTest-1635132145-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 821.690750] env[65726]: DEBUG oslo_vmware.api [None req-ae076e22-617f-47ad-a346-b93664495097 tempest-ServersNegativeTestMultiTenantJSON-944098855 tempest-ServersNegativeTestMultiTenantJSON-944098855-project-member] Task: {'id': task-5115957, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.66508} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 821.691152] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae076e22-617f-47ad-a346-b93664495097 tempest-ServersNegativeTestMultiTenantJSON-944098855 tempest-ServersNegativeTestMultiTenantJSON-944098855-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore2] 018c8ac5-66c4-4a9a-ab45-85e84f92b4fe/018c8ac5-66c4-4a9a-ab45-85e84f92b4fe.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 821.691343] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-ae076e22-617f-47ad-a346-b93664495097 tempest-ServersNegativeTestMultiTenantJSON-944098855 tempest-ServersNegativeTestMultiTenantJSON-944098855-project-member] [instance: 018c8ac5-66c4-4a9a-ab45-85e84f92b4fe] Extending root virtual disk to 1048576 {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 821.691610] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-994c51b4-e59a-4d56-90d0-ecc0f125f41e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.701406] env[65726]: DEBUG oslo_vmware.api [None req-ae076e22-617f-47ad-a346-b93664495097 tempest-ServersNegativeTestMultiTenantJSON-944098855 tempest-ServersNegativeTestMultiTenantJSON-944098855-project-member] Waiting for the task: (returnval){ [ 821.701406] env[65726]: value = "task-5115961" [ 821.701406] env[65726]: _type = "Task" [ 821.701406] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 821.713538] env[65726]: DEBUG oslo_vmware.api [None req-ae076e22-617f-47ad-a346-b93664495097 tempest-ServersNegativeTestMultiTenantJSON-944098855 tempest-ServersNegativeTestMultiTenantJSON-944098855-project-member] Task: {'id': task-5115961, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.760562] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115960, 'name': CreateVM_Task, 'duration_secs': 0.456799} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 821.762473] env[65726]: DEBUG nova.network.neutron [None req-0b112b08-1593-41b5-b38e-86e5b4d22081 tempest-ServersV294TestFqdnHostnames-819811677 tempest-ServersV294TestFqdnHostnames-819811677-project-member] [instance: c8be64a3-1bfa-41a3-b9be-68a7e6e052e0] Successfully updated port: e9d6f425-5466-4d5d-ae7f-80a433defe57 {{(pid=65726) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 821.764739] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3cbbe146-be3a-43d5-867d-4d669884758c] Created VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 821.766384] env[65726]: WARNING openstack [None req-81594943-9858-4765-a371-42f9ef0cbec9 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 821.766968] env[65726]: WARNING openstack [None req-81594943-9858-4765-a371-42f9ef0cbec9 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 821.773111] env[65726]: DEBUG oslo_concurrency.lockutils [None req-81594943-9858-4765-a371-42f9ef0cbec9 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 821.773267] env[65726]: DEBUG oslo_concurrency.lockutils [None req-81594943-9858-4765-a371-42f9ef0cbec9 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 821.773844] env[65726]: DEBUG oslo_concurrency.lockutils [None req-81594943-9858-4765-a371-42f9ef0cbec9 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 821.774187] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4b36f5a3-e450-4a8b-b49b-ffc9281055d5 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.784175] env[65726]: DEBUG oslo_vmware.api [None req-81594943-9858-4765-a371-42f9ef0cbec9 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] Waiting for the task: (returnval){ [ 821.784175] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5282c7e4-47f6-f082-d930-9d90259f7431" [ 821.784175] env[65726]: _type = "Task" [ 821.784175] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 821.801021] env[65726]: DEBUG oslo_vmware.api [None req-81594943-9858-4765-a371-42f9ef0cbec9 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5282c7e4-47f6-f082-d930-9d90259f7431, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.860309] env[65726]: DEBUG oslo_concurrency.lockutils [None req-855996fc-ecea-49d8-a71b-8b4307731732 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Lock "6c8db442-daaa-4eca-ae24-b9d4afaf3a71" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 88.536s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 821.979175] env[65726]: DEBUG oslo_concurrency.lockutils [None req-7b44120c-3963-45cf-8eb3-04f0d2e08d90 tempest-ServerDiagnosticsNegativeTest-1635132145 tempest-ServerDiagnosticsNegativeTest-1635132145-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.246s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 821.981875] env[65726]: DEBUG oslo_concurrency.lockutils [None req-645c8c71-cd4d-4583-8270-b88bb6120d10 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 43.345s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 821.983440] env[65726]: INFO nova.compute.claims [None req-645c8c71-cd4d-4583-8270-b88bb6120d10 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 01d86089-6b9b-4588-864e-ef91375a1eea] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 822.007880] env[65726]: INFO nova.scheduler.client.report [None req-7b44120c-3963-45cf-8eb3-04f0d2e08d90 tempest-ServerDiagnosticsNegativeTest-1635132145 tempest-ServerDiagnosticsNegativeTest-1635132145-project-member] Deleted allocations for instance 0d6f14c5-5770-476f-a186-4cf1b2d4fbe8 [ 822.169987] env[65726]: DEBUG oslo_concurrency.lockutils [None req-85226e56-668f-4c9e-b188-d90c53166469 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Acquiring lock "refresh_cache-49bcab1f-7b64-4999-abff-37771c58a271" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 822.170230] env[65726]: DEBUG oslo_concurrency.lockutils [None req-85226e56-668f-4c9e-b188-d90c53166469 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Acquired lock "refresh_cache-49bcab1f-7b64-4999-abff-37771c58a271" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 822.170541] env[65726]: DEBUG nova.network.neutron [None req-85226e56-668f-4c9e-b188-d90c53166469 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: 49bcab1f-7b64-4999-abff-37771c58a271] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 822.214814] env[65726]: DEBUG oslo_vmware.api [None req-ae076e22-617f-47ad-a346-b93664495097 tempest-ServersNegativeTestMultiTenantJSON-944098855 tempest-ServersNegativeTestMultiTenantJSON-944098855-project-member] Task: {'id': task-5115961, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.123431} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.215960] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-ae076e22-617f-47ad-a346-b93664495097 tempest-ServersNegativeTestMultiTenantJSON-944098855 tempest-ServersNegativeTestMultiTenantJSON-944098855-project-member] [instance: 018c8ac5-66c4-4a9a-ab45-85e84f92b4fe] Extended root virtual disk {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 822.218229] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18b8a80c-85da-472a-90df-62d483607ab8 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.247371] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-ae076e22-617f-47ad-a346-b93664495097 tempest-ServersNegativeTestMultiTenantJSON-944098855 tempest-ServersNegativeTestMultiTenantJSON-944098855-project-member] [instance: 018c8ac5-66c4-4a9a-ab45-85e84f92b4fe] Reconfiguring VM instance instance-00000033 to attach disk [datastore2] 018c8ac5-66c4-4a9a-ab45-85e84f92b4fe/018c8ac5-66c4-4a9a-ab45-85e84f92b4fe.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 822.248433] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5ec1dd0e-8fc4-4a08-9de2-779e4a46d0f8 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.268770] env[65726]: DEBUG oslo_concurrency.lockutils [None req-0b112b08-1593-41b5-b38e-86e5b4d22081 tempest-ServersV294TestFqdnHostnames-819811677 tempest-ServersV294TestFqdnHostnames-819811677-project-member] Acquiring lock "refresh_cache-c8be64a3-1bfa-41a3-b9be-68a7e6e052e0" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 822.268770] env[65726]: DEBUG oslo_concurrency.lockutils [None req-0b112b08-1593-41b5-b38e-86e5b4d22081 tempest-ServersV294TestFqdnHostnames-819811677 tempest-ServersV294TestFqdnHostnames-819811677-project-member] Acquired lock "refresh_cache-c8be64a3-1bfa-41a3-b9be-68a7e6e052e0" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 822.268770] env[65726]: DEBUG nova.network.neutron [None req-0b112b08-1593-41b5-b38e-86e5b4d22081 tempest-ServersV294TestFqdnHostnames-819811677 tempest-ServersV294TestFqdnHostnames-819811677-project-member] [instance: c8be64a3-1bfa-41a3-b9be-68a7e6e052e0] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 822.270496] env[65726]: DEBUG nova.compute.manager [req-8ebb2d88-34af-49ee-8873-236e3f06e184 req-26c9eed3-e1b9-4710-a686-90735091f157 service nova] [instance: 49bcab1f-7b64-4999-abff-37771c58a271] Received event network-vif-deleted-f2a8ac3c-ad7e-4eb4-aec3-db29631fd7ad {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 822.270726] env[65726]: INFO nova.compute.manager [req-8ebb2d88-34af-49ee-8873-236e3f06e184 req-26c9eed3-e1b9-4710-a686-90735091f157 service nova] [instance: 49bcab1f-7b64-4999-abff-37771c58a271] Neutron deleted interface f2a8ac3c-ad7e-4eb4-aec3-db29631fd7ad; detaching it from the instance and deleting it from the info cache [ 822.271379] env[65726]: DEBUG nova.network.neutron [req-8ebb2d88-34af-49ee-8873-236e3f06e184 req-26c9eed3-e1b9-4710-a686-90735091f157 service nova] [instance: 49bcab1f-7b64-4999-abff-37771c58a271] Updating instance_info_cache with network_info: [{"id": "45612b7e-c052-4346-a779-5f94f9adc084", "address": "fa:16:3e:9d:95:60", "network": {"id": "51a66448-0f72-4266-8868-856b48be839f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-155658961-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.161", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d46ccb00794f458b85da4a93879139ee", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b1f3e6c3-5584-4852-9017-476ab8ac4946", "external-id": "nsx-vlan-transportzone-304", "segmentation_id": 304, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap45612b7e-c0", "ovs_interfaceid": "45612b7e-c052-4346-a779-5f94f9adc084", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 822.284947] env[65726]: DEBUG oslo_vmware.api [None req-ae076e22-617f-47ad-a346-b93664495097 tempest-ServersNegativeTestMultiTenantJSON-944098855 tempest-ServersNegativeTestMultiTenantJSON-944098855-project-member] Waiting for the task: (returnval){ [ 822.284947] env[65726]: value = "task-5115962" [ 822.284947] env[65726]: _type = "Task" [ 822.284947] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.299366] env[65726]: DEBUG oslo_vmware.api [None req-ae076e22-617f-47ad-a346-b93664495097 tempest-ServersNegativeTestMultiTenantJSON-944098855 tempest-ServersNegativeTestMultiTenantJSON-944098855-project-member] Task: {'id': task-5115962, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.303985] env[65726]: DEBUG oslo_vmware.api [None req-81594943-9858-4765-a371-42f9ef0cbec9 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5282c7e4-47f6-f082-d930-9d90259f7431, 'name': SearchDatastore_Task, 'duration_secs': 0.018499} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.305142] env[65726]: DEBUG oslo_concurrency.lockutils [None req-81594943-9858-4765-a371-42f9ef0cbec9 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 822.305457] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-81594943-9858-4765-a371-42f9ef0cbec9 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] [instance: 3cbbe146-be3a-43d5-867d-4d669884758c] Processing image b52362a3-ee8a-4cbf-b06f-513b0cc8f95c {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 822.305700] env[65726]: DEBUG oslo_concurrency.lockutils [None req-81594943-9858-4765-a371-42f9ef0cbec9 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 822.305848] env[65726]: DEBUG oslo_concurrency.lockutils [None req-81594943-9858-4765-a371-42f9ef0cbec9 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 822.306272] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-81594943-9858-4765-a371-42f9ef0cbec9 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 822.306957] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9c13c8a4-263b-4543-bfb0-b5a9fb2274f3 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.319273] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-81594943-9858-4765-a371-42f9ef0cbec9 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 822.319504] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-81594943-9858-4765-a371-42f9ef0cbec9 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65726) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 822.320437] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a9fae3db-0d47-4464-b0d8-288c24e7e143 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.327607] env[65726]: DEBUG oslo_vmware.api [None req-81594943-9858-4765-a371-42f9ef0cbec9 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] Waiting for the task: (returnval){ [ 822.327607] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]528884ef-5013-04c5-11a1-6e08eeba61da" [ 822.327607] env[65726]: _type = "Task" [ 822.327607] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.339992] env[65726]: DEBUG oslo_vmware.api [None req-81594943-9858-4765-a371-42f9ef0cbec9 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]528884ef-5013-04c5-11a1-6e08eeba61da, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.363718] env[65726]: DEBUG nova.compute.manager [None req-b3ea747b-af8b-4592-b662-6643bc7f39d3 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] [instance: 2cbee359-a1fc-45c0-bcc5-a2c24ded9c69] Starting instance... {{(pid=65726) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 822.517902] env[65726]: DEBUG oslo_concurrency.lockutils [None req-7b44120c-3963-45cf-8eb3-04f0d2e08d90 tempest-ServerDiagnosticsNegativeTest-1635132145 tempest-ServerDiagnosticsNegativeTest-1635132145-project-member] Lock "0d6f14c5-5770-476f-a186-4cf1b2d4fbe8" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 47.745s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 822.674542] env[65726]: WARNING openstack [None req-85226e56-668f-4c9e-b188-d90c53166469 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 822.675732] env[65726]: WARNING openstack [None req-85226e56-668f-4c9e-b188-d90c53166469 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 822.705490] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c4393951-32f6-44d6-b9c7-1ad480073dbe tempest-InstanceActionsNegativeTestJSON-844675881 tempest-InstanceActionsNegativeTestJSON-844675881-project-member] Acquiring lock "1bd26aef-995e-43b6-af9f-077e878c8c44" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 822.705490] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c4393951-32f6-44d6-b9c7-1ad480073dbe tempest-InstanceActionsNegativeTestJSON-844675881 tempest-InstanceActionsNegativeTestJSON-844675881-project-member] Lock "1bd26aef-995e-43b6-af9f-077e878c8c44" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 822.705490] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c4393951-32f6-44d6-b9c7-1ad480073dbe tempest-InstanceActionsNegativeTestJSON-844675881 tempest-InstanceActionsNegativeTestJSON-844675881-project-member] Acquiring lock "1bd26aef-995e-43b6-af9f-077e878c8c44-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 822.705950] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c4393951-32f6-44d6-b9c7-1ad480073dbe tempest-InstanceActionsNegativeTestJSON-844675881 tempest-InstanceActionsNegativeTestJSON-844675881-project-member] Lock "1bd26aef-995e-43b6-af9f-077e878c8c44-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 822.705950] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c4393951-32f6-44d6-b9c7-1ad480073dbe tempest-InstanceActionsNegativeTestJSON-844675881 tempest-InstanceActionsNegativeTestJSON-844675881-project-member] Lock "1bd26aef-995e-43b6-af9f-077e878c8c44-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 822.708815] env[65726]: INFO nova.compute.manager [None req-c4393951-32f6-44d6-b9c7-1ad480073dbe tempest-InstanceActionsNegativeTestJSON-844675881 tempest-InstanceActionsNegativeTestJSON-844675881-project-member] [instance: 1bd26aef-995e-43b6-af9f-077e878c8c44] Terminating instance [ 822.777360] env[65726]: WARNING openstack [None req-0b112b08-1593-41b5-b38e-86e5b4d22081 tempest-ServersV294TestFqdnHostnames-819811677 tempest-ServersV294TestFqdnHostnames-819811677-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 822.778050] env[65726]: WARNING openstack [None req-0b112b08-1593-41b5-b38e-86e5b4d22081 tempest-ServersV294TestFqdnHostnames-819811677 tempest-ServersV294TestFqdnHostnames-819811677-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 822.785708] env[65726]: DEBUG nova.network.neutron [None req-0b112b08-1593-41b5-b38e-86e5b4d22081 tempest-ServersV294TestFqdnHostnames-819811677 tempest-ServersV294TestFqdnHostnames-819811677-project-member] [instance: c8be64a3-1bfa-41a3-b9be-68a7e6e052e0] Instance cache missing network info. {{(pid=65726) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 822.786700] env[65726]: DEBUG oslo_concurrency.lockutils [req-8ebb2d88-34af-49ee-8873-236e3f06e184 req-26c9eed3-e1b9-4710-a686-90735091f157 service nova] Acquiring lock "49bcab1f-7b64-4999-abff-37771c58a271" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 822.787295] env[65726]: DEBUG oslo_concurrency.lockutils [req-8ebb2d88-34af-49ee-8873-236e3f06e184 req-26c9eed3-e1b9-4710-a686-90735091f157 service nova] Acquired lock "49bcab1f-7b64-4999-abff-37771c58a271" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 822.787866] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2bf2dbeb-9d4a-4df6-b2c1-95fcbfb35f6c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.803417] env[65726]: DEBUG oslo_vmware.api [None req-ae076e22-617f-47ad-a346-b93664495097 tempest-ServersNegativeTestMultiTenantJSON-944098855 tempest-ServersNegativeTestMultiTenantJSON-944098855-project-member] Task: {'id': task-5115962, 'name': ReconfigVM_Task, 'duration_secs': 0.374708} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.815438] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-ae076e22-617f-47ad-a346-b93664495097 tempest-ServersNegativeTestMultiTenantJSON-944098855 tempest-ServersNegativeTestMultiTenantJSON-944098855-project-member] [instance: 018c8ac5-66c4-4a9a-ab45-85e84f92b4fe] Reconfigured VM instance instance-00000033 to attach disk [datastore2] 018c8ac5-66c4-4a9a-ab45-85e84f92b4fe/018c8ac5-66c4-4a9a-ab45-85e84f92b4fe.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 822.816463] env[65726]: DEBUG oslo_concurrency.lockutils [req-8ebb2d88-34af-49ee-8873-236e3f06e184 req-26c9eed3-e1b9-4710-a686-90735091f157 service nova] Releasing lock "49bcab1f-7b64-4999-abff-37771c58a271" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 822.816583] env[65726]: WARNING nova.compute.manager [req-8ebb2d88-34af-49ee-8873-236e3f06e184 req-26c9eed3-e1b9-4710-a686-90735091f157 service nova] [instance: 49bcab1f-7b64-4999-abff-37771c58a271] Detach interface failed, port_id=f2a8ac3c-ad7e-4eb4-aec3-db29631fd7ad, reason: No device with interface-id f2a8ac3c-ad7e-4eb4-aec3-db29631fd7ad exists on VM: nova.exception.NotFound: No device with interface-id f2a8ac3c-ad7e-4eb4-aec3-db29631fd7ad exists on VM [ 822.821119] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2c7492e0-a667-4f1f-ad0e-cf8b9231a071 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.835118] env[65726]: DEBUG oslo_vmware.api [None req-ae076e22-617f-47ad-a346-b93664495097 tempest-ServersNegativeTestMultiTenantJSON-944098855 tempest-ServersNegativeTestMultiTenantJSON-944098855-project-member] Waiting for the task: (returnval){ [ 822.835118] env[65726]: value = "task-5115963" [ 822.835118] env[65726]: _type = "Task" [ 822.835118] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.843111] env[65726]: DEBUG oslo_vmware.api [None req-81594943-9858-4765-a371-42f9ef0cbec9 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]528884ef-5013-04c5-11a1-6e08eeba61da, 'name': SearchDatastore_Task, 'duration_secs': 0.018477} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.844448] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dae449e0-4414-40f4-81d7-95b7dbe1c58b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.852208] env[65726]: DEBUG oslo_vmware.api [None req-ae076e22-617f-47ad-a346-b93664495097 tempest-ServersNegativeTestMultiTenantJSON-944098855 tempest-ServersNegativeTestMultiTenantJSON-944098855-project-member] Task: {'id': task-5115963, 'name': Rename_Task} progress is 10%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.856722] env[65726]: DEBUG oslo_vmware.api [None req-81594943-9858-4765-a371-42f9ef0cbec9 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] Waiting for the task: (returnval){ [ 822.856722] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]522b4f76-55ce-9af2-b166-1679159c2478" [ 822.856722] env[65726]: _type = "Task" [ 822.856722] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.871194] env[65726]: DEBUG oslo_vmware.api [None req-81594943-9858-4765-a371-42f9ef0cbec9 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]522b4f76-55ce-9af2-b166-1679159c2478, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.878359] env[65726]: WARNING openstack [None req-85226e56-668f-4c9e-b188-d90c53166469 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 822.879056] env[65726]: WARNING openstack [None req-85226e56-668f-4c9e-b188-d90c53166469 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 822.886855] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b3ea747b-af8b-4592-b662-6643bc7f39d3 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 822.897040] env[65726]: WARNING openstack [None req-0b112b08-1593-41b5-b38e-86e5b4d22081 tempest-ServersV294TestFqdnHostnames-819811677 tempest-ServersV294TestFqdnHostnames-819811677-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 822.899324] env[65726]: WARNING openstack [None req-0b112b08-1593-41b5-b38e-86e5b4d22081 tempest-ServersV294TestFqdnHostnames-819811677 tempest-ServersV294TestFqdnHostnames-819811677-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 823.006436] env[65726]: WARNING openstack [None req-85226e56-668f-4c9e-b188-d90c53166469 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 823.006931] env[65726]: WARNING openstack [None req-85226e56-668f-4c9e-b188-d90c53166469 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 823.103384] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b20cc073-fafc-44fa-bcf9-9893dc98ebec tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Acquiring lock "2fe6fc5a-f5c7-4f8e-96df-4e621a252f04" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 823.103700] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b20cc073-fafc-44fa-bcf9-9893dc98ebec tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Lock "2fe6fc5a-f5c7-4f8e-96df-4e621a252f04" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 823.114821] env[65726]: WARNING openstack [None req-0b112b08-1593-41b5-b38e-86e5b4d22081 tempest-ServersV294TestFqdnHostnames-819811677 tempest-ServersV294TestFqdnHostnames-819811677-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 823.114954] env[65726]: WARNING openstack [None req-0b112b08-1593-41b5-b38e-86e5b4d22081 tempest-ServersV294TestFqdnHostnames-819811677 tempest-ServersV294TestFqdnHostnames-819811677-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 823.214068] env[65726]: DEBUG nova.compute.manager [None req-c4393951-32f6-44d6-b9c7-1ad480073dbe tempest-InstanceActionsNegativeTestJSON-844675881 tempest-InstanceActionsNegativeTestJSON-844675881-project-member] [instance: 1bd26aef-995e-43b6-af9f-077e878c8c44] Start destroying the instance on the hypervisor. {{(pid=65726) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 823.214383] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-c4393951-32f6-44d6-b9c7-1ad480073dbe tempest-InstanceActionsNegativeTestJSON-844675881 tempest-InstanceActionsNegativeTestJSON-844675881-project-member] [instance: 1bd26aef-995e-43b6-af9f-077e878c8c44] Destroying instance {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 823.215484] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e23e1a06-198f-489b-893b-6029f9b04314 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.225290] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-c4393951-32f6-44d6-b9c7-1ad480073dbe tempest-InstanceActionsNegativeTestJSON-844675881 tempest-InstanceActionsNegativeTestJSON-844675881-project-member] [instance: 1bd26aef-995e-43b6-af9f-077e878c8c44] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 823.225597] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c36f2161-4ed5-4d3c-941e-54f1a8146e4d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.234284] env[65726]: DEBUG oslo_vmware.api [None req-c4393951-32f6-44d6-b9c7-1ad480073dbe tempest-InstanceActionsNegativeTestJSON-844675881 tempest-InstanceActionsNegativeTestJSON-844675881-project-member] Waiting for the task: (returnval){ [ 823.234284] env[65726]: value = "task-5115964" [ 823.234284] env[65726]: _type = "Task" [ 823.234284] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 823.245727] env[65726]: DEBUG oslo_vmware.api [None req-c4393951-32f6-44d6-b9c7-1ad480073dbe tempest-InstanceActionsNegativeTestJSON-844675881 tempest-InstanceActionsNegativeTestJSON-844675881-project-member] Task: {'id': task-5115964, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.295704] env[65726]: INFO nova.network.neutron [None req-85226e56-668f-4c9e-b188-d90c53166469 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: 49bcab1f-7b64-4999-abff-37771c58a271] Port f2a8ac3c-ad7e-4eb4-aec3-db29631fd7ad from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 823.296113] env[65726]: DEBUG nova.network.neutron [None req-85226e56-668f-4c9e-b188-d90c53166469 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: 49bcab1f-7b64-4999-abff-37771c58a271] Updating instance_info_cache with network_info: [{"id": "45612b7e-c052-4346-a779-5f94f9adc084", "address": "fa:16:3e:9d:95:60", "network": {"id": "51a66448-0f72-4266-8868-856b48be839f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-155658961-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.161", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d46ccb00794f458b85da4a93879139ee", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b1f3e6c3-5584-4852-9017-476ab8ac4946", "external-id": "nsx-vlan-transportzone-304", "segmentation_id": 304, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap45612b7e-c0", "ovs_interfaceid": "45612b7e-c052-4346-a779-5f94f9adc084", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 823.300083] env[65726]: DEBUG nova.network.neutron [None req-0b112b08-1593-41b5-b38e-86e5b4d22081 tempest-ServersV294TestFqdnHostnames-819811677 tempest-ServersV294TestFqdnHostnames-819811677-project-member] [instance: c8be64a3-1bfa-41a3-b9be-68a7e6e052e0] Updating instance_info_cache with network_info: [{"id": "e9d6f425-5466-4d5d-ae7f-80a433defe57", "address": "fa:16:3e:7d:00:be", "network": {"id": "00391368-0b6f-45a0-9510-7b6ac2ab47ca", "bridge": "br-int", "label": "tempest-ServersV294TestFqdnHostnames-611600830-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b8855e0d30be40b49e32a3eb0b6eef30", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d6e940e5-e083-4238-973e-f1b4e2a3a5c7", "external-id": "nsx-vlan-transportzone-64", "segmentation_id": 64, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape9d6f425-54", "ovs_interfaceid": "e9d6f425-5466-4d5d-ae7f-80a433defe57", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 823.347628] env[65726]: DEBUG oslo_vmware.api [None req-ae076e22-617f-47ad-a346-b93664495097 tempest-ServersNegativeTestMultiTenantJSON-944098855 tempest-ServersNegativeTestMultiTenantJSON-944098855-project-member] Task: {'id': task-5115963, 'name': Rename_Task} progress is 99%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.377934] env[65726]: DEBUG oslo_vmware.api [None req-81594943-9858-4765-a371-42f9ef0cbec9 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]522b4f76-55ce-9af2-b166-1679159c2478, 'name': SearchDatastore_Task, 'duration_secs': 0.016049} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 823.380898] env[65726]: DEBUG oslo_concurrency.lockutils [None req-81594943-9858-4765-a371-42f9ef0cbec9 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 823.381051] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-81594943-9858-4765-a371-42f9ef0cbec9 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] 3cbbe146-be3a-43d5-867d-4d669884758c/3cbbe146-be3a-43d5-867d-4d669884758c.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 823.382411] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c6905df5-8676-4b12-be55-7d878f25a5f5 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.392555] env[65726]: DEBUG oslo_vmware.api [None req-81594943-9858-4765-a371-42f9ef0cbec9 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] Waiting for the task: (returnval){ [ 823.392555] env[65726]: value = "task-5115965" [ 823.392555] env[65726]: _type = "Task" [ 823.392555] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 823.407034] env[65726]: DEBUG oslo_vmware.api [None req-81594943-9858-4765-a371-42f9ef0cbec9 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] Task: {'id': task-5115965, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.455625] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e1137661-56aa-48f2-a91c-68a6f9993ed2 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Acquiring lock "49bcab1f-7b64-4999-abff-37771c58a271" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 823.456918] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e1137661-56aa-48f2-a91c-68a6f9993ed2 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Lock "49bcab1f-7b64-4999-abff-37771c58a271" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 823.456918] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e1137661-56aa-48f2-a91c-68a6f9993ed2 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Acquiring lock "49bcab1f-7b64-4999-abff-37771c58a271-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 823.456918] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e1137661-56aa-48f2-a91c-68a6f9993ed2 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Lock "49bcab1f-7b64-4999-abff-37771c58a271-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 823.456918] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e1137661-56aa-48f2-a91c-68a6f9993ed2 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Lock "49bcab1f-7b64-4999-abff-37771c58a271-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 823.461080] env[65726]: INFO nova.compute.manager [None req-e1137661-56aa-48f2-a91c-68a6f9993ed2 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: 49bcab1f-7b64-4999-abff-37771c58a271] Terminating instance [ 823.742835] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eddcf3dd-3352-48e8-b9ed-6bcd1ce6d10d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.750923] env[65726]: DEBUG oslo_vmware.api [None req-c4393951-32f6-44d6-b9c7-1ad480073dbe tempest-InstanceActionsNegativeTestJSON-844675881 tempest-InstanceActionsNegativeTestJSON-844675881-project-member] Task: {'id': task-5115964, 'name': PowerOffVM_Task, 'duration_secs': 0.355723} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 823.752400] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-c4393951-32f6-44d6-b9c7-1ad480073dbe tempest-InstanceActionsNegativeTestJSON-844675881 tempest-InstanceActionsNegativeTestJSON-844675881-project-member] [instance: 1bd26aef-995e-43b6-af9f-077e878c8c44] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 823.752400] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-c4393951-32f6-44d6-b9c7-1ad480073dbe tempest-InstanceActionsNegativeTestJSON-844675881 tempest-InstanceActionsNegativeTestJSON-844675881-project-member] [instance: 1bd26aef-995e-43b6-af9f-077e878c8c44] Unregistering the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 823.752796] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a47be5fa-7949-42ba-8207-c144e69a761b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.759691] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a0db06d-5fb6-491b-8c87-dba84be5d3da {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.800297] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ba46a92-4002-434f-9223-d3052d4bd50f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.806462] env[65726]: DEBUG oslo_concurrency.lockutils [None req-0b112b08-1593-41b5-b38e-86e5b4d22081 tempest-ServersV294TestFqdnHostnames-819811677 tempest-ServersV294TestFqdnHostnames-819811677-project-member] Releasing lock "refresh_cache-c8be64a3-1bfa-41a3-b9be-68a7e6e052e0" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 823.806856] env[65726]: DEBUG nova.compute.manager [None req-0b112b08-1593-41b5-b38e-86e5b4d22081 tempest-ServersV294TestFqdnHostnames-819811677 tempest-ServersV294TestFqdnHostnames-819811677-project-member] [instance: c8be64a3-1bfa-41a3-b9be-68a7e6e052e0] Instance network_info: |[{"id": "e9d6f425-5466-4d5d-ae7f-80a433defe57", "address": "fa:16:3e:7d:00:be", "network": {"id": "00391368-0b6f-45a0-9510-7b6ac2ab47ca", "bridge": "br-int", "label": "tempest-ServersV294TestFqdnHostnames-611600830-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b8855e0d30be40b49e32a3eb0b6eef30", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d6e940e5-e083-4238-973e-f1b4e2a3a5c7", "external-id": "nsx-vlan-transportzone-64", "segmentation_id": 64, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape9d6f425-54", "ovs_interfaceid": "e9d6f425-5466-4d5d-ae7f-80a433defe57", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 823.807483] env[65726]: DEBUG oslo_concurrency.lockutils [None req-85226e56-668f-4c9e-b188-d90c53166469 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Releasing lock "refresh_cache-49bcab1f-7b64-4999-abff-37771c58a271" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 823.812803] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-0b112b08-1593-41b5-b38e-86e5b4d22081 tempest-ServersV294TestFqdnHostnames-819811677 tempest-ServersV294TestFqdnHostnames-819811677-project-member] [instance: c8be64a3-1bfa-41a3-b9be-68a7e6e052e0] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:7d:00:be', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd6e940e5-e083-4238-973e-f1b4e2a3a5c7', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e9d6f425-5466-4d5d-ae7f-80a433defe57', 'vif_model': 'vmxnet3'}] {{(pid=65726) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 823.822915] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-0b112b08-1593-41b5-b38e-86e5b4d22081 tempest-ServersV294TestFqdnHostnames-819811677 tempest-ServersV294TestFqdnHostnames-819811677-project-member] Creating folder: Project (b8855e0d30be40b49e32a3eb0b6eef30). Parent ref: group-v995008. {{(pid=65726) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 823.827252] env[65726]: DEBUG oslo_concurrency.lockutils [None req-85226e56-668f-4c9e-b188-d90c53166469 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Lock "interface-49bcab1f-7b64-4999-abff-37771c58a271-f2a8ac3c-ad7e-4eb4-aec3-db29631fd7ad" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 9.657s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 823.827252] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e3d4a858-07b4-4117-8033-d1959671e42d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.830023] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32af2a78-10e6-4f3b-b0b0-f85f47ac9f32 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.839356] env[65726]: DEBUG nova.compute.manager [req-e8483eb8-e308-4a11-ab4b-e9a6d2595303 req-acf05380-44eb-416c-8e00-1e52695eb917 service nova] [instance: 3cbbe146-be3a-43d5-867d-4d669884758c] Received event network-changed-40240c4e-9f3e-46aa-bd75-c62c2776d9b6 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 823.842155] env[65726]: DEBUG nova.compute.manager [req-e8483eb8-e308-4a11-ab4b-e9a6d2595303 req-acf05380-44eb-416c-8e00-1e52695eb917 service nova] [instance: 3cbbe146-be3a-43d5-867d-4d669884758c] Refreshing instance network info cache due to event network-changed-40240c4e-9f3e-46aa-bd75-c62c2776d9b6. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 823.842155] env[65726]: DEBUG oslo_concurrency.lockutils [req-e8483eb8-e308-4a11-ab4b-e9a6d2595303 req-acf05380-44eb-416c-8e00-1e52695eb917 service nova] Acquiring lock "refresh_cache-3cbbe146-be3a-43d5-867d-4d669884758c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 823.842155] env[65726]: DEBUG oslo_concurrency.lockutils [req-e8483eb8-e308-4a11-ab4b-e9a6d2595303 req-acf05380-44eb-416c-8e00-1e52695eb917 service nova] Acquired lock "refresh_cache-3cbbe146-be3a-43d5-867d-4d669884758c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 823.842155] env[65726]: DEBUG nova.network.neutron [req-e8483eb8-e308-4a11-ab4b-e9a6d2595303 req-acf05380-44eb-416c-8e00-1e52695eb917 service nova] [instance: 3cbbe146-be3a-43d5-867d-4d669884758c] Refreshing network info cache for port 40240c4e-9f3e-46aa-bd75-c62c2776d9b6 {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 823.854283] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-c4393951-32f6-44d6-b9c7-1ad480073dbe tempest-InstanceActionsNegativeTestJSON-844675881 tempest-InstanceActionsNegativeTestJSON-844675881-project-member] [instance: 1bd26aef-995e-43b6-af9f-077e878c8c44] Unregistered the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 823.854642] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-c4393951-32f6-44d6-b9c7-1ad480073dbe tempest-InstanceActionsNegativeTestJSON-844675881 tempest-InstanceActionsNegativeTestJSON-844675881-project-member] [instance: 1bd26aef-995e-43b6-af9f-077e878c8c44] Deleting contents of the VM from datastore datastore2 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 823.854908] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-c4393951-32f6-44d6-b9c7-1ad480073dbe tempest-InstanceActionsNegativeTestJSON-844675881 tempest-InstanceActionsNegativeTestJSON-844675881-project-member] Deleting the datastore file [datastore2] 1bd26aef-995e-43b6-af9f-077e878c8c44 {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 823.855894] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f57728bd-801e-4b23-99d5-6a81096af596 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.870193] env[65726]: INFO nova.virt.vmwareapi.vm_util [None req-0b112b08-1593-41b5-b38e-86e5b4d22081 tempest-ServersV294TestFqdnHostnames-819811677 tempest-ServersV294TestFqdnHostnames-819811677-project-member] Created folder: Project (b8855e0d30be40b49e32a3eb0b6eef30) in parent group-v995008. [ 823.870193] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-0b112b08-1593-41b5-b38e-86e5b4d22081 tempest-ServersV294TestFqdnHostnames-819811677 tempest-ServersV294TestFqdnHostnames-819811677-project-member] Creating folder: Instances. Parent ref: group-v995154. {{(pid=65726) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 823.870585] env[65726]: DEBUG nova.compute.provider_tree [None req-645c8c71-cd4d-4583-8270-b88bb6120d10 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 823.876198] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9c006f82-93b0-404d-9bac-7fe9bba2ffa5 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.878851] env[65726]: DEBUG oslo_vmware.api [None req-ae076e22-617f-47ad-a346-b93664495097 tempest-ServersNegativeTestMultiTenantJSON-944098855 tempest-ServersNegativeTestMultiTenantJSON-944098855-project-member] Task: {'id': task-5115963, 'name': Rename_Task} progress is 99%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.882050] env[65726]: DEBUG oslo_vmware.api [None req-c4393951-32f6-44d6-b9c7-1ad480073dbe tempest-InstanceActionsNegativeTestJSON-844675881 tempest-InstanceActionsNegativeTestJSON-844675881-project-member] Waiting for the task: (returnval){ [ 823.882050] env[65726]: value = "task-5115968" [ 823.882050] env[65726]: _type = "Task" [ 823.882050] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 823.890354] env[65726]: INFO nova.virt.vmwareapi.vm_util [None req-0b112b08-1593-41b5-b38e-86e5b4d22081 tempest-ServersV294TestFqdnHostnames-819811677 tempest-ServersV294TestFqdnHostnames-819811677-project-member] Created folder: Instances in parent group-v995154. [ 823.890828] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-0b112b08-1593-41b5-b38e-86e5b4d22081 tempest-ServersV294TestFqdnHostnames-819811677 tempest-ServersV294TestFqdnHostnames-819811677-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 823.890828] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c8be64a3-1bfa-41a3-b9be-68a7e6e052e0] Creating VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 823.891951] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-89a3af90-7639-4dd6-a4f8-3c7711b4316d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.914405] env[65726]: DEBUG oslo_vmware.api [None req-c4393951-32f6-44d6-b9c7-1ad480073dbe tempest-InstanceActionsNegativeTestJSON-844675881 tempest-InstanceActionsNegativeTestJSON-844675881-project-member] Task: {'id': task-5115968, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.925221] env[65726]: DEBUG oslo_vmware.api [None req-81594943-9858-4765-a371-42f9ef0cbec9 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] Task: {'id': task-5115965, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.927319] env[65726]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 823.927319] env[65726]: value = "task-5115970" [ 823.927319] env[65726]: _type = "Task" [ 823.927319] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 823.936900] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115970, 'name': CreateVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.970375] env[65726]: DEBUG nova.compute.manager [None req-e1137661-56aa-48f2-a91c-68a6f9993ed2 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: 49bcab1f-7b64-4999-abff-37771c58a271] Start destroying the instance on the hypervisor. {{(pid=65726) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 823.970721] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-e1137661-56aa-48f2-a91c-68a6f9993ed2 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: 49bcab1f-7b64-4999-abff-37771c58a271] Destroying instance {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 823.971721] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-edb43269-4ef5-465b-9397-d8229dd96258 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.982127] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-e1137661-56aa-48f2-a91c-68a6f9993ed2 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: 49bcab1f-7b64-4999-abff-37771c58a271] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 823.982327] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3526c3da-8fd7-43cf-8a93-27810c63c605 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.989918] env[65726]: DEBUG oslo_vmware.api [None req-e1137661-56aa-48f2-a91c-68a6f9993ed2 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Waiting for the task: (returnval){ [ 823.989918] env[65726]: value = "task-5115971" [ 823.989918] env[65726]: _type = "Task" [ 823.989918] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 824.001576] env[65726]: DEBUG oslo_vmware.api [None req-e1137661-56aa-48f2-a91c-68a6f9993ed2 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': task-5115971, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.345180] env[65726]: WARNING openstack [req-e8483eb8-e308-4a11-ab4b-e9a6d2595303 req-acf05380-44eb-416c-8e00-1e52695eb917 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 824.345667] env[65726]: WARNING openstack [req-e8483eb8-e308-4a11-ab4b-e9a6d2595303 req-acf05380-44eb-416c-8e00-1e52695eb917 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 824.365628] env[65726]: DEBUG oslo_vmware.api [None req-ae076e22-617f-47ad-a346-b93664495097 tempest-ServersNegativeTestMultiTenantJSON-944098855 tempest-ServersNegativeTestMultiTenantJSON-944098855-project-member] Task: {'id': task-5115963, 'name': Rename_Task, 'duration_secs': 1.232931} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.365999] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae076e22-617f-47ad-a346-b93664495097 tempest-ServersNegativeTestMultiTenantJSON-944098855 tempest-ServersNegativeTestMultiTenantJSON-944098855-project-member] [instance: 018c8ac5-66c4-4a9a-ab45-85e84f92b4fe] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 824.366293] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4853c607-d198-4bdf-b7cb-6cf0592d0de0 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.375908] env[65726]: DEBUG oslo_vmware.api [None req-ae076e22-617f-47ad-a346-b93664495097 tempest-ServersNegativeTestMultiTenantJSON-944098855 tempest-ServersNegativeTestMultiTenantJSON-944098855-project-member] Waiting for the task: (returnval){ [ 824.375908] env[65726]: value = "task-5115972" [ 824.375908] env[65726]: _type = "Task" [ 824.375908] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 824.380595] env[65726]: DEBUG nova.scheduler.client.report [None req-645c8c71-cd4d-4583-8270-b88bb6120d10 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 824.402841] env[65726]: DEBUG oslo_vmware.api [None req-ae076e22-617f-47ad-a346-b93664495097 tempest-ServersNegativeTestMultiTenantJSON-944098855 tempest-ServersNegativeTestMultiTenantJSON-944098855-project-member] Task: {'id': task-5115972, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.409738] env[65726]: DEBUG oslo_vmware.api [None req-c4393951-32f6-44d6-b9c7-1ad480073dbe tempest-InstanceActionsNegativeTestJSON-844675881 tempest-InstanceActionsNegativeTestJSON-844675881-project-member] Task: {'id': task-5115968, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.311898} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.410846] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-c4393951-32f6-44d6-b9c7-1ad480073dbe tempest-InstanceActionsNegativeTestJSON-844675881 tempest-InstanceActionsNegativeTestJSON-844675881-project-member] Deleted the datastore file {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 824.411061] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-c4393951-32f6-44d6-b9c7-1ad480073dbe tempest-InstanceActionsNegativeTestJSON-844675881 tempest-InstanceActionsNegativeTestJSON-844675881-project-member] [instance: 1bd26aef-995e-43b6-af9f-077e878c8c44] Deleted contents of the VM from datastore datastore2 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 824.411371] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-c4393951-32f6-44d6-b9c7-1ad480073dbe tempest-InstanceActionsNegativeTestJSON-844675881 tempest-InstanceActionsNegativeTestJSON-844675881-project-member] [instance: 1bd26aef-995e-43b6-af9f-077e878c8c44] Instance destroyed {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 824.411427] env[65726]: INFO nova.compute.manager [None req-c4393951-32f6-44d6-b9c7-1ad480073dbe tempest-InstanceActionsNegativeTestJSON-844675881 tempest-InstanceActionsNegativeTestJSON-844675881-project-member] [instance: 1bd26aef-995e-43b6-af9f-077e878c8c44] Took 1.20 seconds to destroy the instance on the hypervisor. [ 824.411903] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-c4393951-32f6-44d6-b9c7-1ad480073dbe tempest-InstanceActionsNegativeTestJSON-844675881 tempest-InstanceActionsNegativeTestJSON-844675881-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 824.411903] env[65726]: DEBUG nova.compute.manager [-] [instance: 1bd26aef-995e-43b6-af9f-077e878c8c44] Deallocating network for instance {{(pid=65726) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 824.412278] env[65726]: DEBUG nova.network.neutron [-] [instance: 1bd26aef-995e-43b6-af9f-077e878c8c44] deallocate_for_instance() {{(pid=65726) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 824.412466] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 824.412736] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 824.434837] env[65726]: DEBUG oslo_vmware.api [None req-81594943-9858-4765-a371-42f9ef0cbec9 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] Task: {'id': task-5115965, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.619719} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.435588] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-81594943-9858-4765-a371-42f9ef0cbec9 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] 3cbbe146-be3a-43d5-867d-4d669884758c/3cbbe146-be3a-43d5-867d-4d669884758c.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 824.436980] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-81594943-9858-4765-a371-42f9ef0cbec9 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] [instance: 3cbbe146-be3a-43d5-867d-4d669884758c] Extending root virtual disk to 1048576 {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 824.436980] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-9ec7a846-27d6-47e4-802b-16ff52a26c6a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.447484] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115970, 'name': CreateVM_Task, 'duration_secs': 0.487533} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.449036] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c8be64a3-1bfa-41a3-b9be-68a7e6e052e0] Created VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 824.449036] env[65726]: WARNING openstack [None req-0b112b08-1593-41b5-b38e-86e5b4d22081 tempest-ServersV294TestFqdnHostnames-819811677 tempest-ServersV294TestFqdnHostnames-819811677-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 824.449508] env[65726]: WARNING openstack [None req-0b112b08-1593-41b5-b38e-86e5b4d22081 tempest-ServersV294TestFqdnHostnames-819811677 tempest-ServersV294TestFqdnHostnames-819811677-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 824.456112] env[65726]: DEBUG oslo_concurrency.lockutils [None req-0b112b08-1593-41b5-b38e-86e5b4d22081 tempest-ServersV294TestFqdnHostnames-819811677 tempest-ServersV294TestFqdnHostnames-819811677-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 824.456351] env[65726]: DEBUG oslo_concurrency.lockutils [None req-0b112b08-1593-41b5-b38e-86e5b4d22081 tempest-ServersV294TestFqdnHostnames-819811677 tempest-ServersV294TestFqdnHostnames-819811677-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 824.456573] env[65726]: DEBUG oslo_concurrency.lockutils [None req-0b112b08-1593-41b5-b38e-86e5b4d22081 tempest-ServersV294TestFqdnHostnames-819811677 tempest-ServersV294TestFqdnHostnames-819811677-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 824.458486] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7b7134d2-e261-4bbc-9e3c-98f85fb4136d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.461145] env[65726]: DEBUG oslo_vmware.api [None req-81594943-9858-4765-a371-42f9ef0cbec9 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] Waiting for the task: (returnval){ [ 824.461145] env[65726]: value = "task-5115973" [ 824.461145] env[65726]: _type = "Task" [ 824.461145] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 824.467839] env[65726]: DEBUG oslo_vmware.api [None req-0b112b08-1593-41b5-b38e-86e5b4d22081 tempest-ServersV294TestFqdnHostnames-819811677 tempest-ServersV294TestFqdnHostnames-819811677-project-member] Waiting for the task: (returnval){ [ 824.467839] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5207624c-fd0e-4069-e3fe-7703d04b2c6b" [ 824.467839] env[65726]: _type = "Task" [ 824.467839] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 824.479121] env[65726]: DEBUG oslo_vmware.api [None req-81594943-9858-4765-a371-42f9ef0cbec9 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] Task: {'id': task-5115973, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.485089] env[65726]: DEBUG oslo_vmware.api [None req-0b112b08-1593-41b5-b38e-86e5b4d22081 tempest-ServersV294TestFqdnHostnames-819811677 tempest-ServersV294TestFqdnHostnames-819811677-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5207624c-fd0e-4069-e3fe-7703d04b2c6b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.501368] env[65726]: DEBUG oslo_vmware.api [None req-e1137661-56aa-48f2-a91c-68a6f9993ed2 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': task-5115971, 'name': PowerOffVM_Task, 'duration_secs': 0.262588} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.501368] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-e1137661-56aa-48f2-a91c-68a6f9993ed2 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: 49bcab1f-7b64-4999-abff-37771c58a271] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 824.501636] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-e1137661-56aa-48f2-a91c-68a6f9993ed2 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: 49bcab1f-7b64-4999-abff-37771c58a271] Unregistering the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 824.501775] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-03ef47bd-47c3-4735-82c4-396fa8faf6f9 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.508830] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 824.510260] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 824.534413] env[65726]: WARNING openstack [req-e8483eb8-e308-4a11-ab4b-e9a6d2595303 req-acf05380-44eb-416c-8e00-1e52695eb917 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 824.534851] env[65726]: WARNING openstack [req-e8483eb8-e308-4a11-ab4b-e9a6d2595303 req-acf05380-44eb-416c-8e00-1e52695eb917 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 824.590511] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-e1137661-56aa-48f2-a91c-68a6f9993ed2 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: 49bcab1f-7b64-4999-abff-37771c58a271] Unregistered the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 824.590769] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-e1137661-56aa-48f2-a91c-68a6f9993ed2 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: 49bcab1f-7b64-4999-abff-37771c58a271] Deleting contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 824.591067] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-e1137661-56aa-48f2-a91c-68a6f9993ed2 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Deleting the datastore file [datastore1] 49bcab1f-7b64-4999-abff-37771c58a271 {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 824.591456] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b70cda60-4ec6-41e6-bf1e-dd391af36dbb {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.600465] env[65726]: DEBUG oslo_vmware.api [None req-e1137661-56aa-48f2-a91c-68a6f9993ed2 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Waiting for the task: (returnval){ [ 824.600465] env[65726]: value = "task-5115975" [ 824.600465] env[65726]: _type = "Task" [ 824.600465] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 824.612356] env[65726]: DEBUG oslo_vmware.api [None req-e1137661-56aa-48f2-a91c-68a6f9993ed2 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': task-5115975, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.761504] env[65726]: WARNING openstack [req-e8483eb8-e308-4a11-ab4b-e9a6d2595303 req-acf05380-44eb-416c-8e00-1e52695eb917 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 824.761930] env[65726]: WARNING openstack [req-e8483eb8-e308-4a11-ab4b-e9a6d2595303 req-acf05380-44eb-416c-8e00-1e52695eb917 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 824.873958] env[65726]: DEBUG nova.network.neutron [req-e8483eb8-e308-4a11-ab4b-e9a6d2595303 req-acf05380-44eb-416c-8e00-1e52695eb917 service nova] [instance: 3cbbe146-be3a-43d5-867d-4d669884758c] Updated VIF entry in instance network info cache for port 40240c4e-9f3e-46aa-bd75-c62c2776d9b6. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 824.874382] env[65726]: DEBUG nova.network.neutron [req-e8483eb8-e308-4a11-ab4b-e9a6d2595303 req-acf05380-44eb-416c-8e00-1e52695eb917 service nova] [instance: 3cbbe146-be3a-43d5-867d-4d669884758c] Updating instance_info_cache with network_info: [{"id": "40240c4e-9f3e-46aa-bd75-c62c2776d9b6", "address": "fa:16:3e:8f:bd:09", "network": {"id": "2266eb6e-4c8f-4a3f-85d3-0c06022c0d5e", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-716412627-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "92c1cedbc08545d58fc7172ac95e2c23", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "24376631-ee89-4ff1-b8ac-f09911fc8329", "external-id": "nsx-vlan-transportzone-960", "segmentation_id": 960, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap40240c4e-9f", "ovs_interfaceid": "40240c4e-9f3e-46aa-bd75-c62c2776d9b6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 824.886425] env[65726]: DEBUG oslo_vmware.api [None req-ae076e22-617f-47ad-a346-b93664495097 tempest-ServersNegativeTestMultiTenantJSON-944098855 tempest-ServersNegativeTestMultiTenantJSON-944098855-project-member] Task: {'id': task-5115972, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.889156] env[65726]: DEBUG oslo_concurrency.lockutils [None req-645c8c71-cd4d-4583-8270-b88bb6120d10 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.907s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 824.889765] env[65726]: DEBUG nova.compute.manager [None req-645c8c71-cd4d-4583-8270-b88bb6120d10 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 01d86089-6b9b-4588-864e-ef91375a1eea] Start building networks asynchronously for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 824.893076] env[65726]: DEBUG oslo_concurrency.lockutils [None req-84a40308-06fb-49aa-8717-9d81681505a7 tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 41.286s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 824.893335] env[65726]: DEBUG nova.objects.instance [None req-84a40308-06fb-49aa-8717-9d81681505a7 tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] Lazy-loading 'resources' on Instance uuid 3827eed1-b7a1-4922-a4de-f0c038cb8566 {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 824.976929] env[65726]: DEBUG oslo_vmware.api [None req-81594943-9858-4765-a371-42f9ef0cbec9 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] Task: {'id': task-5115973, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.124582} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.977428] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-81594943-9858-4765-a371-42f9ef0cbec9 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] [instance: 3cbbe146-be3a-43d5-867d-4d669884758c] Extended root virtual disk {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 824.978834] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9be1a10d-a765-4bf6-acb4-eb23691c1763 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.985804] env[65726]: DEBUG oslo_vmware.api [None req-0b112b08-1593-41b5-b38e-86e5b4d22081 tempest-ServersV294TestFqdnHostnames-819811677 tempest-ServersV294TestFqdnHostnames-819811677-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5207624c-fd0e-4069-e3fe-7703d04b2c6b, 'name': SearchDatastore_Task, 'duration_secs': 0.030976} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.986785] env[65726]: DEBUG oslo_concurrency.lockutils [None req-0b112b08-1593-41b5-b38e-86e5b4d22081 tempest-ServersV294TestFqdnHostnames-819811677 tempest-ServersV294TestFqdnHostnames-819811677-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 824.987202] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-0b112b08-1593-41b5-b38e-86e5b4d22081 tempest-ServersV294TestFqdnHostnames-819811677 tempest-ServersV294TestFqdnHostnames-819811677-project-member] [instance: c8be64a3-1bfa-41a3-b9be-68a7e6e052e0] Processing image b52362a3-ee8a-4cbf-b06f-513b0cc8f95c {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 824.987688] env[65726]: DEBUG oslo_concurrency.lockutils [None req-0b112b08-1593-41b5-b38e-86e5b4d22081 tempest-ServersV294TestFqdnHostnames-819811677 tempest-ServersV294TestFqdnHostnames-819811677-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 824.988139] env[65726]: DEBUG oslo_concurrency.lockutils [None req-0b112b08-1593-41b5-b38e-86e5b4d22081 tempest-ServersV294TestFqdnHostnames-819811677 tempest-ServersV294TestFqdnHostnames-819811677-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 824.988515] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-0b112b08-1593-41b5-b38e-86e5b4d22081 tempest-ServersV294TestFqdnHostnames-819811677 tempest-ServersV294TestFqdnHostnames-819811677-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 824.988949] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-40ae4e98-6611-492f-ad9d-6f784e50e2d9 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.017847] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-81594943-9858-4765-a371-42f9ef0cbec9 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] [instance: 3cbbe146-be3a-43d5-867d-4d669884758c] Reconfiguring VM instance instance-00000034 to attach disk [datastore1] 3cbbe146-be3a-43d5-867d-4d669884758c/3cbbe146-be3a-43d5-867d-4d669884758c.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 825.019351] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6400a871-7f78-47f8-952c-aeb81d88860f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.037693] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-0b112b08-1593-41b5-b38e-86e5b4d22081 tempest-ServersV294TestFqdnHostnames-819811677 tempest-ServersV294TestFqdnHostnames-819811677-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 825.037913] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-0b112b08-1593-41b5-b38e-86e5b4d22081 tempest-ServersV294TestFqdnHostnames-819811677 tempest-ServersV294TestFqdnHostnames-819811677-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65726) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 825.039813] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c390d529-6ce1-42a3-ad46-a065b6ad1118 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.049491] env[65726]: DEBUG oslo_vmware.api [None req-81594943-9858-4765-a371-42f9ef0cbec9 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] Waiting for the task: (returnval){ [ 825.049491] env[65726]: value = "task-5115976" [ 825.049491] env[65726]: _type = "Task" [ 825.049491] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.051843] env[65726]: DEBUG oslo_vmware.api [None req-0b112b08-1593-41b5-b38e-86e5b4d22081 tempest-ServersV294TestFqdnHostnames-819811677 tempest-ServersV294TestFqdnHostnames-819811677-project-member] Waiting for the task: (returnval){ [ 825.051843] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52c75d3e-6361-2504-b7f7-214ea2d63598" [ 825.051843] env[65726]: _type = "Task" [ 825.051843] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.067671] env[65726]: DEBUG oslo_vmware.api [None req-81594943-9858-4765-a371-42f9ef0cbec9 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] Task: {'id': task-5115976, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.071842] env[65726]: DEBUG oslo_vmware.api [None req-0b112b08-1593-41b5-b38e-86e5b4d22081 tempest-ServersV294TestFqdnHostnames-819811677 tempest-ServersV294TestFqdnHostnames-819811677-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52c75d3e-6361-2504-b7f7-214ea2d63598, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.113946] env[65726]: DEBUG oslo_vmware.api [None req-e1137661-56aa-48f2-a91c-68a6f9993ed2 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': task-5115975, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.351238} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 825.113946] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-e1137661-56aa-48f2-a91c-68a6f9993ed2 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Deleted the datastore file {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 825.114345] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-e1137661-56aa-48f2-a91c-68a6f9993ed2 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: 49bcab1f-7b64-4999-abff-37771c58a271] Deleted contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 825.114345] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-e1137661-56aa-48f2-a91c-68a6f9993ed2 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: 49bcab1f-7b64-4999-abff-37771c58a271] Instance destroyed {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 825.114540] env[65726]: INFO nova.compute.manager [None req-e1137661-56aa-48f2-a91c-68a6f9993ed2 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: 49bcab1f-7b64-4999-abff-37771c58a271] Took 1.14 seconds to destroy the instance on the hypervisor. [ 825.114796] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-e1137661-56aa-48f2-a91c-68a6f9993ed2 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 825.114996] env[65726]: DEBUG nova.compute.manager [-] [instance: 49bcab1f-7b64-4999-abff-37771c58a271] Deallocating network for instance {{(pid=65726) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 825.115115] env[65726]: DEBUG nova.network.neutron [-] [instance: 49bcab1f-7b64-4999-abff-37771c58a271] deallocate_for_instance() {{(pid=65726) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 825.115641] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 825.115897] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 825.157592] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 825.157592] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 825.173614] env[65726]: DEBUG nova.compute.manager [req-dc7a0b44-a57c-49e4-b360-702b8cdb9e8a req-25c08c64-b367-4771-8f94-58f026904773 service nova] [instance: 6c8db442-daaa-4eca-ae24-b9d4afaf3a71] Received event network-changed-6f422d37-8262-454d-aa2d-1569bacee1b7 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 825.174405] env[65726]: DEBUG nova.compute.manager [req-dc7a0b44-a57c-49e4-b360-702b8cdb9e8a req-25c08c64-b367-4771-8f94-58f026904773 service nova] [instance: 6c8db442-daaa-4eca-ae24-b9d4afaf3a71] Refreshing instance network info cache due to event network-changed-6f422d37-8262-454d-aa2d-1569bacee1b7. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 825.174787] env[65726]: DEBUG oslo_concurrency.lockutils [req-dc7a0b44-a57c-49e4-b360-702b8cdb9e8a req-25c08c64-b367-4771-8f94-58f026904773 service nova] Acquiring lock "refresh_cache-6c8db442-daaa-4eca-ae24-b9d4afaf3a71" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 825.175392] env[65726]: DEBUG oslo_concurrency.lockutils [req-dc7a0b44-a57c-49e4-b360-702b8cdb9e8a req-25c08c64-b367-4771-8f94-58f026904773 service nova] Acquired lock "refresh_cache-6c8db442-daaa-4eca-ae24-b9d4afaf3a71" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 825.175920] env[65726]: DEBUG nova.network.neutron [req-dc7a0b44-a57c-49e4-b360-702b8cdb9e8a req-25c08c64-b367-4771-8f94-58f026904773 service nova] [instance: 6c8db442-daaa-4eca-ae24-b9d4afaf3a71] Refreshing network info cache for port 6f422d37-8262-454d-aa2d-1569bacee1b7 {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 825.253806] env[65726]: DEBUG nova.network.neutron [-] [instance: 1bd26aef-995e-43b6-af9f-077e878c8c44] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 825.377948] env[65726]: DEBUG oslo_concurrency.lockutils [req-e8483eb8-e308-4a11-ab4b-e9a6d2595303 req-acf05380-44eb-416c-8e00-1e52695eb917 service nova] Releasing lock "refresh_cache-3cbbe146-be3a-43d5-867d-4d669884758c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 825.377948] env[65726]: DEBUG nova.compute.manager [req-e8483eb8-e308-4a11-ab4b-e9a6d2595303 req-acf05380-44eb-416c-8e00-1e52695eb917 service nova] [instance: c8be64a3-1bfa-41a3-b9be-68a7e6e052e0] Received event network-vif-plugged-e9d6f425-5466-4d5d-ae7f-80a433defe57 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 825.377948] env[65726]: DEBUG oslo_concurrency.lockutils [req-e8483eb8-e308-4a11-ab4b-e9a6d2595303 req-acf05380-44eb-416c-8e00-1e52695eb917 service nova] Acquiring lock "c8be64a3-1bfa-41a3-b9be-68a7e6e052e0-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 825.378218] env[65726]: DEBUG oslo_concurrency.lockutils [req-e8483eb8-e308-4a11-ab4b-e9a6d2595303 req-acf05380-44eb-416c-8e00-1e52695eb917 service nova] Lock "c8be64a3-1bfa-41a3-b9be-68a7e6e052e0-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 825.378277] env[65726]: DEBUG oslo_concurrency.lockutils [req-e8483eb8-e308-4a11-ab4b-e9a6d2595303 req-acf05380-44eb-416c-8e00-1e52695eb917 service nova] Lock "c8be64a3-1bfa-41a3-b9be-68a7e6e052e0-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 825.381043] env[65726]: DEBUG nova.compute.manager [req-e8483eb8-e308-4a11-ab4b-e9a6d2595303 req-acf05380-44eb-416c-8e00-1e52695eb917 service nova] [instance: c8be64a3-1bfa-41a3-b9be-68a7e6e052e0] No waiting events found dispatching network-vif-plugged-e9d6f425-5466-4d5d-ae7f-80a433defe57 {{(pid=65726) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 825.381043] env[65726]: WARNING nova.compute.manager [req-e8483eb8-e308-4a11-ab4b-e9a6d2595303 req-acf05380-44eb-416c-8e00-1e52695eb917 service nova] [instance: c8be64a3-1bfa-41a3-b9be-68a7e6e052e0] Received unexpected event network-vif-plugged-e9d6f425-5466-4d5d-ae7f-80a433defe57 for instance with vm_state building and task_state spawning. [ 825.381043] env[65726]: DEBUG nova.compute.manager [req-e8483eb8-e308-4a11-ab4b-e9a6d2595303 req-acf05380-44eb-416c-8e00-1e52695eb917 service nova] [instance: c8be64a3-1bfa-41a3-b9be-68a7e6e052e0] Received event network-changed-e9d6f425-5466-4d5d-ae7f-80a433defe57 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 825.381043] env[65726]: DEBUG nova.compute.manager [req-e8483eb8-e308-4a11-ab4b-e9a6d2595303 req-acf05380-44eb-416c-8e00-1e52695eb917 service nova] [instance: c8be64a3-1bfa-41a3-b9be-68a7e6e052e0] Refreshing instance network info cache due to event network-changed-e9d6f425-5466-4d5d-ae7f-80a433defe57. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 825.381043] env[65726]: DEBUG oslo_concurrency.lockutils [req-e8483eb8-e308-4a11-ab4b-e9a6d2595303 req-acf05380-44eb-416c-8e00-1e52695eb917 service nova] Acquiring lock "refresh_cache-c8be64a3-1bfa-41a3-b9be-68a7e6e052e0" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 825.381043] env[65726]: DEBUG oslo_concurrency.lockutils [req-e8483eb8-e308-4a11-ab4b-e9a6d2595303 req-acf05380-44eb-416c-8e00-1e52695eb917 service nova] Acquired lock "refresh_cache-c8be64a3-1bfa-41a3-b9be-68a7e6e052e0" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 825.381043] env[65726]: DEBUG nova.network.neutron [req-e8483eb8-e308-4a11-ab4b-e9a6d2595303 req-acf05380-44eb-416c-8e00-1e52695eb917 service nova] [instance: c8be64a3-1bfa-41a3-b9be-68a7e6e052e0] Refreshing network info cache for port e9d6f425-5466-4d5d-ae7f-80a433defe57 {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 825.397586] env[65726]: DEBUG nova.compute.utils [None req-645c8c71-cd4d-4583-8270-b88bb6120d10 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Using /dev/sd instead of None {{(pid=65726) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 825.402667] env[65726]: DEBUG oslo_vmware.api [None req-ae076e22-617f-47ad-a346-b93664495097 tempest-ServersNegativeTestMultiTenantJSON-944098855 tempest-ServersNegativeTestMultiTenantJSON-944098855-project-member] Task: {'id': task-5115972, 'name': PowerOnVM_Task, 'duration_secs': 0.539801} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 825.403884] env[65726]: DEBUG nova.compute.manager [None req-645c8c71-cd4d-4583-8270-b88bb6120d10 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 01d86089-6b9b-4588-864e-ef91375a1eea] Allocating IP information in the background. {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 825.403884] env[65726]: DEBUG nova.network.neutron [None req-645c8c71-cd4d-4583-8270-b88bb6120d10 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 01d86089-6b9b-4588-864e-ef91375a1eea] allocate_for_instance() {{(pid=65726) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 825.404091] env[65726]: WARNING neutronclient.v2_0.client [None req-645c8c71-cd4d-4583-8270-b88bb6120d10 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 825.404442] env[65726]: WARNING neutronclient.v2_0.client [None req-645c8c71-cd4d-4583-8270-b88bb6120d10 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 825.405211] env[65726]: WARNING openstack [None req-645c8c71-cd4d-4583-8270-b88bb6120d10 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 825.405464] env[65726]: WARNING openstack [None req-645c8c71-cd4d-4583-8270-b88bb6120d10 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 825.413996] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae076e22-617f-47ad-a346-b93664495097 tempest-ServersNegativeTestMultiTenantJSON-944098855 tempest-ServersNegativeTestMultiTenantJSON-944098855-project-member] [instance: 018c8ac5-66c4-4a9a-ab45-85e84f92b4fe] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 825.413996] env[65726]: INFO nova.compute.manager [None req-ae076e22-617f-47ad-a346-b93664495097 tempest-ServersNegativeTestMultiTenantJSON-944098855 tempest-ServersNegativeTestMultiTenantJSON-944098855-project-member] [instance: 018c8ac5-66c4-4a9a-ab45-85e84f92b4fe] Took 10.16 seconds to spawn the instance on the hypervisor. [ 825.414222] env[65726]: DEBUG nova.compute.manager [None req-ae076e22-617f-47ad-a346-b93664495097 tempest-ServersNegativeTestMultiTenantJSON-944098855 tempest-ServersNegativeTestMultiTenantJSON-944098855-project-member] [instance: 018c8ac5-66c4-4a9a-ab45-85e84f92b4fe] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 825.416122] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d39abdeb-f943-4221-a114-08f924cde911 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.501184] env[65726]: DEBUG nova.policy [None req-645c8c71-cd4d-4583-8270-b88bb6120d10 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9b6087cfa55d41a4bad9f6cee89865d5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '39067912d091464192516c08eb576696', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65726) authorize /opt/stack/nova/nova/policy.py:192}} [ 825.571944] env[65726]: DEBUG oslo_vmware.api [None req-81594943-9858-4765-a371-42f9ef0cbec9 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] Task: {'id': task-5115976, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.579586] env[65726]: DEBUG oslo_vmware.api [None req-0b112b08-1593-41b5-b38e-86e5b4d22081 tempest-ServersV294TestFqdnHostnames-819811677 tempest-ServersV294TestFqdnHostnames-819811677-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52c75d3e-6361-2504-b7f7-214ea2d63598, 'name': SearchDatastore_Task, 'duration_secs': 0.02106} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 825.581295] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-02ef1e50-9f8a-499d-8aed-7ce53e2026da {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.590024] env[65726]: DEBUG oslo_vmware.api [None req-0b112b08-1593-41b5-b38e-86e5b4d22081 tempest-ServersV294TestFqdnHostnames-819811677 tempest-ServersV294TestFqdnHostnames-819811677-project-member] Waiting for the task: (returnval){ [ 825.590024] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5256c697-5e7d-d50f-4a5e-04ac81c915d2" [ 825.590024] env[65726]: _type = "Task" [ 825.590024] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.603123] env[65726]: DEBUG oslo_vmware.api [None req-0b112b08-1593-41b5-b38e-86e5b4d22081 tempest-ServersV294TestFqdnHostnames-819811677 tempest-ServersV294TestFqdnHostnames-819811677-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5256c697-5e7d-d50f-4a5e-04ac81c915d2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.680388] env[65726]: WARNING openstack [req-dc7a0b44-a57c-49e4-b360-702b8cdb9e8a req-25c08c64-b367-4771-8f94-58f026904773 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 825.680784] env[65726]: WARNING openstack [req-dc7a0b44-a57c-49e4-b360-702b8cdb9e8a req-25c08c64-b367-4771-8f94-58f026904773 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 825.758499] env[65726]: INFO nova.compute.manager [-] [instance: 1bd26aef-995e-43b6-af9f-077e878c8c44] Took 1.35 seconds to deallocate network for instance. [ 825.843799] env[65726]: DEBUG nova.network.neutron [None req-645c8c71-cd4d-4583-8270-b88bb6120d10 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 01d86089-6b9b-4588-864e-ef91375a1eea] Successfully created port: beb674aa-cade-455d-a97a-4da9699c9a70 {{(pid=65726) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 825.886069] env[65726]: WARNING openstack [req-e8483eb8-e308-4a11-ab4b-e9a6d2595303 req-acf05380-44eb-416c-8e00-1e52695eb917 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 825.889143] env[65726]: WARNING openstack [req-e8483eb8-e308-4a11-ab4b-e9a6d2595303 req-acf05380-44eb-416c-8e00-1e52695eb917 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 825.905786] env[65726]: DEBUG nova.compute.manager [None req-645c8c71-cd4d-4583-8270-b88bb6120d10 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 01d86089-6b9b-4588-864e-ef91375a1eea] Start building block device mappings for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 825.947785] env[65726]: INFO nova.compute.manager [None req-ae076e22-617f-47ad-a346-b93664495097 tempest-ServersNegativeTestMultiTenantJSON-944098855 tempest-ServersNegativeTestMultiTenantJSON-944098855-project-member] [instance: 018c8ac5-66c4-4a9a-ab45-85e84f92b4fe] Took 52.74 seconds to build instance. [ 825.980770] env[65726]: WARNING openstack [req-dc7a0b44-a57c-49e4-b360-702b8cdb9e8a req-25c08c64-b367-4771-8f94-58f026904773 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 825.981282] env[65726]: WARNING openstack [req-dc7a0b44-a57c-49e4-b360-702b8cdb9e8a req-25c08c64-b367-4771-8f94-58f026904773 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 826.043695] env[65726]: DEBUG oslo_vmware.rw_handles [None req-c3e5127a-ba4f-4e76-986d-1724943cb777 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5270d0a4-127e-6e1d-921f-59f48964187b/disk-0.vmdk. {{(pid=65726) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 826.044697] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52a8b576-31fa-4de0-a0ff-bb6094a0f3d2 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.059323] env[65726]: DEBUG oslo_vmware.rw_handles [None req-c3e5127a-ba4f-4e76-986d-1724943cb777 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5270d0a4-127e-6e1d-921f-59f48964187b/disk-0.vmdk is in state: ready. {{(pid=65726) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 826.059577] env[65726]: ERROR oslo_vmware.rw_handles [None req-c3e5127a-ba4f-4e76-986d-1724943cb777 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5270d0a4-127e-6e1d-921f-59f48964187b/disk-0.vmdk due to incomplete transfer. [ 826.060776] env[65726]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-6f816097-c876-403d-9fed-615632733739 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.066486] env[65726]: DEBUG oslo_vmware.api [None req-81594943-9858-4765-a371-42f9ef0cbec9 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] Task: {'id': task-5115976, 'name': ReconfigVM_Task, 'duration_secs': 0.542687} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.066862] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-81594943-9858-4765-a371-42f9ef0cbec9 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] [instance: 3cbbe146-be3a-43d5-867d-4d669884758c] Reconfigured VM instance instance-00000034 to attach disk [datastore1] 3cbbe146-be3a-43d5-867d-4d669884758c/3cbbe146-be3a-43d5-867d-4d669884758c.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 826.067591] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d4b2d1ec-5945-45ac-8b2e-af18d5d42dd1 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.072755] env[65726]: DEBUG oslo_vmware.rw_handles [None req-c3e5127a-ba4f-4e76-986d-1724943cb777 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5270d0a4-127e-6e1d-921f-59f48964187b/disk-0.vmdk. {{(pid=65726) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 826.072755] env[65726]: DEBUG nova.virt.vmwareapi.images [None req-c3e5127a-ba4f-4e76-986d-1724943cb777 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 608cc0ea-3f6b-4b4f-83c1-01aa50999d63] Uploaded image 122c2b99-e3fb-499d-a847-a4c088b862f5 to the Glance image server {{(pid=65726) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 826.074212] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-c3e5127a-ba4f-4e76-986d-1724943cb777 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 608cc0ea-3f6b-4b4f-83c1-01aa50999d63] Destroying the VM {{(pid=65726) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 826.078352] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-6a3a139a-7a5a-456f-9901-b35897e36a61 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.083418] env[65726]: DEBUG oslo_vmware.api [None req-81594943-9858-4765-a371-42f9ef0cbec9 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] Waiting for the task: (returnval){ [ 826.083418] env[65726]: value = "task-5115977" [ 826.083418] env[65726]: _type = "Task" [ 826.083418] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.088289] env[65726]: DEBUG oslo_vmware.api [None req-c3e5127a-ba4f-4e76-986d-1724943cb777 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Waiting for the task: (returnval){ [ 826.088289] env[65726]: value = "task-5115978" [ 826.088289] env[65726]: _type = "Task" [ 826.088289] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.102824] env[65726]: DEBUG oslo_vmware.api [None req-81594943-9858-4765-a371-42f9ef0cbec9 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] Task: {'id': task-5115977, 'name': Rename_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.113080] env[65726]: DEBUG oslo_vmware.api [None req-c3e5127a-ba4f-4e76-986d-1724943cb777 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5115978, 'name': Destroy_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.118461] env[65726]: DEBUG oslo_vmware.api [None req-0b112b08-1593-41b5-b38e-86e5b4d22081 tempest-ServersV294TestFqdnHostnames-819811677 tempest-ServersV294TestFqdnHostnames-819811677-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5256c697-5e7d-d50f-4a5e-04ac81c915d2, 'name': SearchDatastore_Task, 'duration_secs': 0.017268} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.118461] env[65726]: DEBUG oslo_concurrency.lockutils [None req-0b112b08-1593-41b5-b38e-86e5b4d22081 tempest-ServersV294TestFqdnHostnames-819811677 tempest-ServersV294TestFqdnHostnames-819811677-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 826.118629] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-0b112b08-1593-41b5-b38e-86e5b4d22081 tempest-ServersV294TestFqdnHostnames-819811677 tempest-ServersV294TestFqdnHostnames-819811677-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] c8be64a3-1bfa-41a3-b9be-68a7e6e052e0/c8be64a3-1bfa-41a3-b9be-68a7e6e052e0.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 826.118856] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-cc2ba817-2393-4314-84bf-04fa161c1656 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.133375] env[65726]: DEBUG oslo_vmware.api [None req-0b112b08-1593-41b5-b38e-86e5b4d22081 tempest-ServersV294TestFqdnHostnames-819811677 tempest-ServersV294TestFqdnHostnames-819811677-project-member] Waiting for the task: (returnval){ [ 826.133375] env[65726]: value = "task-5115979" [ 826.133375] env[65726]: _type = "Task" [ 826.133375] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.143894] env[65726]: DEBUG oslo_vmware.api [None req-0b112b08-1593-41b5-b38e-86e5b4d22081 tempest-ServersV294TestFqdnHostnames-819811677 tempest-ServersV294TestFqdnHostnames-819811677-project-member] Task: {'id': task-5115979, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.146252] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f696e448-0482-487d-8767-0d64b27bea34 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.161056] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b99a3ab-6d06-48ab-8340-3bdba00e3e1a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.198622] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9586e81-c2ae-4ef5-8e62-198c9ea23ee6 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.201788] env[65726]: DEBUG nova.network.neutron [-] [instance: 49bcab1f-7b64-4999-abff-37771c58a271] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 826.211035] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e24deb7e-0e7f-4b3b-8393-4286fc4b21b6 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.229508] env[65726]: DEBUG nova.compute.provider_tree [None req-84a40308-06fb-49aa-8717-9d81681505a7 tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 826.236262] env[65726]: WARNING openstack [req-dc7a0b44-a57c-49e4-b360-702b8cdb9e8a req-25c08c64-b367-4771-8f94-58f026904773 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 826.238421] env[65726]: WARNING openstack [req-dc7a0b44-a57c-49e4-b360-702b8cdb9e8a req-25c08c64-b367-4771-8f94-58f026904773 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 826.249732] env[65726]: WARNING openstack [req-e8483eb8-e308-4a11-ab4b-e9a6d2595303 req-acf05380-44eb-416c-8e00-1e52695eb917 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 826.250383] env[65726]: WARNING openstack [req-e8483eb8-e308-4a11-ab4b-e9a6d2595303 req-acf05380-44eb-416c-8e00-1e52695eb917 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 826.277020] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c4393951-32f6-44d6-b9c7-1ad480073dbe tempest-InstanceActionsNegativeTestJSON-844675881 tempest-InstanceActionsNegativeTestJSON-844675881-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 826.450547] env[65726]: DEBUG oslo_concurrency.lockutils [None req-ae076e22-617f-47ad-a346-b93664495097 tempest-ServersNegativeTestMultiTenantJSON-944098855 tempest-ServersNegativeTestMultiTenantJSON-944098855-project-member] Lock "018c8ac5-66c4-4a9a-ab45-85e84f92b4fe" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 85.402s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 826.518754] env[65726]: DEBUG nova.network.neutron [req-dc7a0b44-a57c-49e4-b360-702b8cdb9e8a req-25c08c64-b367-4771-8f94-58f026904773 service nova] [instance: 6c8db442-daaa-4eca-ae24-b9d4afaf3a71] Updated VIF entry in instance network info cache for port 6f422d37-8262-454d-aa2d-1569bacee1b7. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 826.519114] env[65726]: DEBUG nova.network.neutron [req-dc7a0b44-a57c-49e4-b360-702b8cdb9e8a req-25c08c64-b367-4771-8f94-58f026904773 service nova] [instance: 6c8db442-daaa-4eca-ae24-b9d4afaf3a71] Updating instance_info_cache with network_info: [{"id": "6f422d37-8262-454d-aa2d-1569bacee1b7", "address": "fa:16:3e:19:b2:c8", "network": {"id": "d37113a1-409c-4c2c-81f9-8f80822b85b9", "bridge": "br-int", "label": "tempest-ServersWithSpecificFlavorTestJSON-471231219-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.139", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "79d3e960f190439fae7c61d693e9356e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "13af9422-d668-4413-b63a-766558d83a3b", "external-id": "nsx-vlan-transportzone-842", "segmentation_id": 842, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6f422d37-82", "ovs_interfaceid": "6f422d37-8262-454d-aa2d-1569bacee1b7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 826.551995] env[65726]: WARNING openstack [req-e8483eb8-e308-4a11-ab4b-e9a6d2595303 req-acf05380-44eb-416c-8e00-1e52695eb917 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 826.552527] env[65726]: WARNING openstack [req-e8483eb8-e308-4a11-ab4b-e9a6d2595303 req-acf05380-44eb-416c-8e00-1e52695eb917 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 826.596682] env[65726]: DEBUG oslo_vmware.api [None req-81594943-9858-4765-a371-42f9ef0cbec9 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] Task: {'id': task-5115977, 'name': Rename_Task, 'duration_secs': 0.218752} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.596749] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-81594943-9858-4765-a371-42f9ef0cbec9 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] [instance: 3cbbe146-be3a-43d5-867d-4d669884758c] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 826.597122] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e4b78798-563c-4dd4-b5d7-8345a8a4fcc1 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.602365] env[65726]: DEBUG oslo_vmware.api [None req-c3e5127a-ba4f-4e76-986d-1724943cb777 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5115978, 'name': Destroy_Task, 'duration_secs': 0.370656} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.603056] env[65726]: INFO nova.virt.vmwareapi.vm_util [None req-c3e5127a-ba4f-4e76-986d-1724943cb777 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 608cc0ea-3f6b-4b4f-83c1-01aa50999d63] Destroyed the VM [ 826.604370] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-c3e5127a-ba4f-4e76-986d-1724943cb777 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 608cc0ea-3f6b-4b4f-83c1-01aa50999d63] Deleting Snapshot of the VM instance {{(pid=65726) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 826.604370] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-42501360-f33c-47fc-8e9c-9f8b5717f5d7 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.610497] env[65726]: DEBUG oslo_vmware.api [None req-81594943-9858-4765-a371-42f9ef0cbec9 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] Waiting for the task: (returnval){ [ 826.610497] env[65726]: value = "task-5115980" [ 826.610497] env[65726]: _type = "Task" [ 826.610497] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.611981] env[65726]: DEBUG oslo_vmware.api [None req-c3e5127a-ba4f-4e76-986d-1724943cb777 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Waiting for the task: (returnval){ [ 826.611981] env[65726]: value = "task-5115981" [ 826.611981] env[65726]: _type = "Task" [ 826.611981] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.625929] env[65726]: DEBUG oslo_vmware.api [None req-c3e5127a-ba4f-4e76-986d-1724943cb777 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5115981, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.629650] env[65726]: DEBUG oslo_vmware.api [None req-81594943-9858-4765-a371-42f9ef0cbec9 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] Task: {'id': task-5115980, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.650492] env[65726]: DEBUG oslo_vmware.api [None req-0b112b08-1593-41b5-b38e-86e5b4d22081 tempest-ServersV294TestFqdnHostnames-819811677 tempest-ServersV294TestFqdnHostnames-819811677-project-member] Task: {'id': task-5115979, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.688600] env[65726]: DEBUG nova.compute.manager [req-09754832-5ff0-4131-b192-a3966ed082fb req-a6c3e8af-0aaf-4c1c-9409-d02f6f8a3b27 service nova] [instance: 1bd26aef-995e-43b6-af9f-077e878c8c44] Received event network-vif-deleted-61bb4097-108f-4424-9e34-38ed6f561f0c {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 826.688600] env[65726]: DEBUG nova.compute.manager [req-09754832-5ff0-4131-b192-a3966ed082fb req-a6c3e8af-0aaf-4c1c-9409-d02f6f8a3b27 service nova] [instance: 49bcab1f-7b64-4999-abff-37771c58a271] Received event network-vif-deleted-45612b7e-c052-4346-a779-5f94f9adc084 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 826.695021] env[65726]: DEBUG nova.network.neutron [req-e8483eb8-e308-4a11-ab4b-e9a6d2595303 req-acf05380-44eb-416c-8e00-1e52695eb917 service nova] [instance: c8be64a3-1bfa-41a3-b9be-68a7e6e052e0] Updated VIF entry in instance network info cache for port e9d6f425-5466-4d5d-ae7f-80a433defe57. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 826.695021] env[65726]: DEBUG nova.network.neutron [req-e8483eb8-e308-4a11-ab4b-e9a6d2595303 req-acf05380-44eb-416c-8e00-1e52695eb917 service nova] [instance: c8be64a3-1bfa-41a3-b9be-68a7e6e052e0] Updating instance_info_cache with network_info: [{"id": "e9d6f425-5466-4d5d-ae7f-80a433defe57", "address": "fa:16:3e:7d:00:be", "network": {"id": "00391368-0b6f-45a0-9510-7b6ac2ab47ca", "bridge": "br-int", "label": "tempest-ServersV294TestFqdnHostnames-611600830-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b8855e0d30be40b49e32a3eb0b6eef30", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d6e940e5-e083-4238-973e-f1b4e2a3a5c7", "external-id": "nsx-vlan-transportzone-64", "segmentation_id": 64, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape9d6f425-54", "ovs_interfaceid": "e9d6f425-5466-4d5d-ae7f-80a433defe57", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 826.706581] env[65726]: INFO nova.compute.manager [-] [instance: 49bcab1f-7b64-4999-abff-37771c58a271] Took 1.59 seconds to deallocate network for instance. [ 826.734411] env[65726]: DEBUG nova.scheduler.client.report [None req-84a40308-06fb-49aa-8717-9d81681505a7 tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 826.918626] env[65726]: DEBUG nova.compute.manager [None req-645c8c71-cd4d-4583-8270-b88bb6120d10 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 01d86089-6b9b-4588-864e-ef91375a1eea] Start spawning the instance on the hypervisor. {{(pid=65726) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 826.957030] env[65726]: DEBUG nova.virt.hardware [None req-645c8c71-cd4d-4583-8270-b88bb6120d10 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T19:28:34Z,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1539f06fa3534e90acbc3a60c4b8bd3f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T19:28:35Z,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 826.957235] env[65726]: DEBUG nova.virt.hardware [None req-645c8c71-cd4d-4583-8270-b88bb6120d10 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 826.957414] env[65726]: DEBUG nova.virt.hardware [None req-645c8c71-cd4d-4583-8270-b88bb6120d10 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 826.958239] env[65726]: DEBUG nova.virt.hardware [None req-645c8c71-cd4d-4583-8270-b88bb6120d10 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 826.958239] env[65726]: DEBUG nova.virt.hardware [None req-645c8c71-cd4d-4583-8270-b88bb6120d10 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 826.958391] env[65726]: DEBUG nova.virt.hardware [None req-645c8c71-cd4d-4583-8270-b88bb6120d10 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 826.958638] env[65726]: DEBUG nova.virt.hardware [None req-645c8c71-cd4d-4583-8270-b88bb6120d10 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 826.958884] env[65726]: DEBUG nova.virt.hardware [None req-645c8c71-cd4d-4583-8270-b88bb6120d10 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 826.959250] env[65726]: DEBUG nova.virt.hardware [None req-645c8c71-cd4d-4583-8270-b88bb6120d10 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 826.959590] env[65726]: DEBUG nova.virt.hardware [None req-645c8c71-cd4d-4583-8270-b88bb6120d10 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 826.959898] env[65726]: DEBUG nova.virt.hardware [None req-645c8c71-cd4d-4583-8270-b88bb6120d10 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 826.960506] env[65726]: DEBUG nova.compute.manager [None req-49b35bf1-9410-49fb-9a5b-632c41ecd3a6 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] [instance: c4cc4f85-cb35-4edc-a58b-adfee0ce1265] Starting instance... {{(pid=65726) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 826.965505] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d957429-a0f8-4b7e-abaa-36be8978042d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.979387] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40e3b10b-9a0d-4b99-a59e-76ec301de5d0 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.024760] env[65726]: DEBUG oslo_concurrency.lockutils [req-dc7a0b44-a57c-49e4-b360-702b8cdb9e8a req-25c08c64-b367-4771-8f94-58f026904773 service nova] Releasing lock "refresh_cache-6c8db442-daaa-4eca-ae24-b9d4afaf3a71" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 827.123943] env[65726]: DEBUG oslo_vmware.api [None req-81594943-9858-4765-a371-42f9ef0cbec9 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] Task: {'id': task-5115980, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.129712] env[65726]: DEBUG oslo_vmware.api [None req-c3e5127a-ba4f-4e76-986d-1724943cb777 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5115981, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.146549] env[65726]: DEBUG oslo_vmware.api [None req-0b112b08-1593-41b5-b38e-86e5b4d22081 tempest-ServersV294TestFqdnHostnames-819811677 tempest-ServersV294TestFqdnHostnames-819811677-project-member] Task: {'id': task-5115979, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.86613} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 827.147088] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-0b112b08-1593-41b5-b38e-86e5b4d22081 tempest-ServersV294TestFqdnHostnames-819811677 tempest-ServersV294TestFqdnHostnames-819811677-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] c8be64a3-1bfa-41a3-b9be-68a7e6e052e0/c8be64a3-1bfa-41a3-b9be-68a7e6e052e0.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 827.147088] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-0b112b08-1593-41b5-b38e-86e5b4d22081 tempest-ServersV294TestFqdnHostnames-819811677 tempest-ServersV294TestFqdnHostnames-819811677-project-member] [instance: c8be64a3-1bfa-41a3-b9be-68a7e6e052e0] Extending root virtual disk to 1048576 {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 827.147452] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-db752ca5-94c8-437a-b727-d511edf77515 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.157022] env[65726]: DEBUG oslo_vmware.api [None req-0b112b08-1593-41b5-b38e-86e5b4d22081 tempest-ServersV294TestFqdnHostnames-819811677 tempest-ServersV294TestFqdnHostnames-819811677-project-member] Waiting for the task: (returnval){ [ 827.157022] env[65726]: value = "task-5115982" [ 827.157022] env[65726]: _type = "Task" [ 827.157022] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 827.170298] env[65726]: DEBUG oslo_vmware.api [None req-0b112b08-1593-41b5-b38e-86e5b4d22081 tempest-ServersV294TestFqdnHostnames-819811677 tempest-ServersV294TestFqdnHostnames-819811677-project-member] Task: {'id': task-5115982, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.197471] env[65726]: DEBUG oslo_concurrency.lockutils [req-e8483eb8-e308-4a11-ab4b-e9a6d2595303 req-acf05380-44eb-416c-8e00-1e52695eb917 service nova] Releasing lock "refresh_cache-c8be64a3-1bfa-41a3-b9be-68a7e6e052e0" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 827.214608] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e1137661-56aa-48f2-a91c-68a6f9993ed2 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 827.241295] env[65726]: DEBUG oslo_concurrency.lockutils [None req-84a40308-06fb-49aa-8717-9d81681505a7 tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.348s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 827.244102] env[65726]: DEBUG oslo_concurrency.lockutils [None req-ebfc0e41-4845-41ca-9f74-d2a2946626c0 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 41.910s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 827.245876] env[65726]: INFO nova.compute.claims [None req-ebfc0e41-4845-41ca-9f74-d2a2946626c0 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] [instance: 1cfb1bba-1e21-47eb-b50f-a86575846a65] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 827.273843] env[65726]: INFO nova.scheduler.client.report [None req-84a40308-06fb-49aa-8717-9d81681505a7 tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] Deleted allocations for instance 3827eed1-b7a1-4922-a4de-f0c038cb8566 [ 827.492044] env[65726]: DEBUG oslo_concurrency.lockutils [None req-49b35bf1-9410-49fb-9a5b-632c41ecd3a6 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 827.574887] env[65726]: DEBUG nova.network.neutron [None req-645c8c71-cd4d-4583-8270-b88bb6120d10 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 01d86089-6b9b-4588-864e-ef91375a1eea] Successfully updated port: beb674aa-cade-455d-a97a-4da9699c9a70 {{(pid=65726) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 827.637738] env[65726]: DEBUG oslo_vmware.api [None req-c3e5127a-ba4f-4e76-986d-1724943cb777 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5115981, 'name': RemoveSnapshot_Task, 'duration_secs': 0.599905} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 827.638151] env[65726]: DEBUG oslo_vmware.api [None req-81594943-9858-4765-a371-42f9ef0cbec9 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] Task: {'id': task-5115980, 'name': PowerOnVM_Task, 'duration_secs': 0.724524} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 827.638527] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-c3e5127a-ba4f-4e76-986d-1724943cb777 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 608cc0ea-3f6b-4b4f-83c1-01aa50999d63] Deleted Snapshot of the VM instance {{(pid=65726) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 827.638882] env[65726]: INFO nova.compute.manager [None req-c3e5127a-ba4f-4e76-986d-1724943cb777 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 608cc0ea-3f6b-4b4f-83c1-01aa50999d63] Took 14.40 seconds to snapshot the instance on the hypervisor. [ 827.642506] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-81594943-9858-4765-a371-42f9ef0cbec9 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] [instance: 3cbbe146-be3a-43d5-867d-4d669884758c] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 827.642832] env[65726]: INFO nova.compute.manager [None req-81594943-9858-4765-a371-42f9ef0cbec9 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] [instance: 3cbbe146-be3a-43d5-867d-4d669884758c] Took 9.17 seconds to spawn the instance on the hypervisor. [ 827.643152] env[65726]: DEBUG nova.compute.manager [None req-81594943-9858-4765-a371-42f9ef0cbec9 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] [instance: 3cbbe146-be3a-43d5-867d-4d669884758c] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 827.648855] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae8e00f4-3095-4d3f-b876-f7c30973d196 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.668748] env[65726]: DEBUG oslo_vmware.api [None req-0b112b08-1593-41b5-b38e-86e5b4d22081 tempest-ServersV294TestFqdnHostnames-819811677 tempest-ServersV294TestFqdnHostnames-819811677-project-member] Task: {'id': task-5115982, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.085452} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 827.669022] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-0b112b08-1593-41b5-b38e-86e5b4d22081 tempest-ServersV294TestFqdnHostnames-819811677 tempest-ServersV294TestFqdnHostnames-819811677-project-member] [instance: c8be64a3-1bfa-41a3-b9be-68a7e6e052e0] Extended root virtual disk {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 827.669839] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80dbf979-168b-48cf-bb75-af34b5132b94 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.696407] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-0b112b08-1593-41b5-b38e-86e5b4d22081 tempest-ServersV294TestFqdnHostnames-819811677 tempest-ServersV294TestFqdnHostnames-819811677-project-member] [instance: c8be64a3-1bfa-41a3-b9be-68a7e6e052e0] Reconfiguring VM instance instance-00000035 to attach disk [datastore1] c8be64a3-1bfa-41a3-b9be-68a7e6e052e0/c8be64a3-1bfa-41a3-b9be-68a7e6e052e0.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 827.696407] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5172af7a-3fda-4e3e-a18f-5713d23a32c6 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.719674] env[65726]: DEBUG oslo_vmware.api [None req-0b112b08-1593-41b5-b38e-86e5b4d22081 tempest-ServersV294TestFqdnHostnames-819811677 tempest-ServersV294TestFqdnHostnames-819811677-project-member] Waiting for the task: (returnval){ [ 827.719674] env[65726]: value = "task-5115983" [ 827.719674] env[65726]: _type = "Task" [ 827.719674] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 827.729973] env[65726]: DEBUG oslo_vmware.api [None req-0b112b08-1593-41b5-b38e-86e5b4d22081 tempest-ServersV294TestFqdnHostnames-819811677 tempest-ServersV294TestFqdnHostnames-819811677-project-member] Task: {'id': task-5115983, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.787170] env[65726]: DEBUG oslo_concurrency.lockutils [None req-84a40308-06fb-49aa-8717-9d81681505a7 tempest-ServersListShow2100Test-655751442 tempest-ServersListShow2100Test-655751442-project-member] Lock "3827eed1-b7a1-4922-a4de-f0c038cb8566" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 48.536s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 828.080732] env[65726]: DEBUG oslo_concurrency.lockutils [None req-645c8c71-cd4d-4583-8270-b88bb6120d10 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Acquiring lock "refresh_cache-01d86089-6b9b-4588-864e-ef91375a1eea" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 828.080996] env[65726]: DEBUG oslo_concurrency.lockutils [None req-645c8c71-cd4d-4583-8270-b88bb6120d10 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Acquired lock "refresh_cache-01d86089-6b9b-4588-864e-ef91375a1eea" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 828.081220] env[65726]: DEBUG nova.network.neutron [None req-645c8c71-cd4d-4583-8270-b88bb6120d10 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 01d86089-6b9b-4588-864e-ef91375a1eea] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 828.176647] env[65726]: INFO nova.compute.manager [None req-81594943-9858-4765-a371-42f9ef0cbec9 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] [instance: 3cbbe146-be3a-43d5-867d-4d669884758c] Took 52.39 seconds to build instance. [ 828.208803] env[65726]: DEBUG nova.compute.manager [None req-c3e5127a-ba4f-4e76-986d-1724943cb777 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 608cc0ea-3f6b-4b4f-83c1-01aa50999d63] Found 3 images (rotation: 2) {{(pid=65726) _rotate_backups /opt/stack/nova/nova/compute/manager.py:5074}} [ 828.208803] env[65726]: DEBUG nova.compute.manager [None req-c3e5127a-ba4f-4e76-986d-1724943cb777 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 608cc0ea-3f6b-4b4f-83c1-01aa50999d63] Rotating out 1 backups {{(pid=65726) _rotate_backups /opt/stack/nova/nova/compute/manager.py:5082}} [ 828.208803] env[65726]: DEBUG nova.compute.manager [None req-c3e5127a-ba4f-4e76-986d-1724943cb777 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 608cc0ea-3f6b-4b4f-83c1-01aa50999d63] Deleting image 5554ed4e-2d17-4d6f-946f-e0ce42b880fc {{(pid=65726) _rotate_backups /opt/stack/nova/nova/compute/manager.py:5087}} [ 828.234473] env[65726]: DEBUG oslo_vmware.api [None req-0b112b08-1593-41b5-b38e-86e5b4d22081 tempest-ServersV294TestFqdnHostnames-819811677 tempest-ServersV294TestFqdnHostnames-819811677-project-member] Task: {'id': task-5115983, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.586201] env[65726]: WARNING openstack [None req-645c8c71-cd4d-4583-8270-b88bb6120d10 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 828.586705] env[65726]: WARNING openstack [None req-645c8c71-cd4d-4583-8270-b88bb6120d10 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 828.592418] env[65726]: DEBUG nova.network.neutron [None req-645c8c71-cd4d-4583-8270-b88bb6120d10 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 01d86089-6b9b-4588-864e-ef91375a1eea] Instance cache missing network info. {{(pid=65726) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 828.674269] env[65726]: WARNING openstack [None req-645c8c71-cd4d-4583-8270-b88bb6120d10 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 828.674637] env[65726]: WARNING openstack [None req-645c8c71-cd4d-4583-8270-b88bb6120d10 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 828.681364] env[65726]: DEBUG oslo_concurrency.lockutils [None req-81594943-9858-4765-a371-42f9ef0cbec9 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] Lock "3cbbe146-be3a-43d5-867d-4d669884758c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 76.983s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 828.737550] env[65726]: DEBUG oslo_vmware.api [None req-0b112b08-1593-41b5-b38e-86e5b4d22081 tempest-ServersV294TestFqdnHostnames-819811677 tempest-ServersV294TestFqdnHostnames-819811677-project-member] Task: {'id': task-5115983, 'name': ReconfigVM_Task, 'duration_secs': 0.79575} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.737872] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-0b112b08-1593-41b5-b38e-86e5b4d22081 tempest-ServersV294TestFqdnHostnames-819811677 tempest-ServersV294TestFqdnHostnames-819811677-project-member] [instance: c8be64a3-1bfa-41a3-b9be-68a7e6e052e0] Reconfigured VM instance instance-00000035 to attach disk [datastore1] c8be64a3-1bfa-41a3-b9be-68a7e6e052e0/c8be64a3-1bfa-41a3-b9be-68a7e6e052e0.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 828.739186] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ada28bce-8a69-4332-89be-e09e62c00e28 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.747481] env[65726]: DEBUG oslo_vmware.api [None req-0b112b08-1593-41b5-b38e-86e5b4d22081 tempest-ServersV294TestFqdnHostnames-819811677 tempest-ServersV294TestFqdnHostnames-819811677-project-member] Waiting for the task: (returnval){ [ 828.747481] env[65726]: value = "task-5115984" [ 828.747481] env[65726]: _type = "Task" [ 828.747481] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.760251] env[65726]: WARNING openstack [None req-645c8c71-cd4d-4583-8270-b88bb6120d10 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 828.761197] env[65726]: WARNING openstack [None req-645c8c71-cd4d-4583-8270-b88bb6120d10 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 828.783133] env[65726]: DEBUG oslo_vmware.api [None req-0b112b08-1593-41b5-b38e-86e5b4d22081 tempest-ServersV294TestFqdnHostnames-819811677 tempest-ServersV294TestFqdnHostnames-819811677-project-member] Task: {'id': task-5115984, 'name': Rename_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.835126] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8eae94ec-6184-45d7-ad1d-0f5ca960e8e5 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.845832] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32b83514-3690-469e-b814-f15aff52f6b9 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.887261] env[65726]: DEBUG nova.network.neutron [None req-645c8c71-cd4d-4583-8270-b88bb6120d10 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 01d86089-6b9b-4588-864e-ef91375a1eea] Updating instance_info_cache with network_info: [{"id": "beb674aa-cade-455d-a97a-4da9699c9a70", "address": "fa:16:3e:f9:f6:08", "network": {"id": "e206b5b9-da36-4e6f-926f-914dec38fac2", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1053201655-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "39067912d091464192516c08eb576696", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7c80243e-93a7-4a95-bc8d-e9534bacd66e", "external-id": "nsx-vlan-transportzone-306", "segmentation_id": 306, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbeb674aa-ca", "ovs_interfaceid": "beb674aa-cade-455d-a97a-4da9699c9a70", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 828.891887] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8a24d91-6044-45ac-a2d4-5a55ab3efc35 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.901917] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36f8630f-a270-4264-9b89-a468c45dea17 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.921389] env[65726]: DEBUG nova.compute.provider_tree [None req-ebfc0e41-4845-41ca-9f74-d2a2946626c0 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Updating inventory in ProviderTree for provider 07c4692f-bdb4-4058-9173-ff9664830295 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 829.187365] env[65726]: DEBUG nova.compute.manager [None req-73b0a627-5d81-4bf2-833a-caff3534c1e3 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: cf950ec3-9914-4eb9-99db-048b3969bb21] Starting instance... {{(pid=65726) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 829.262225] env[65726]: DEBUG oslo_vmware.api [None req-0b112b08-1593-41b5-b38e-86e5b4d22081 tempest-ServersV294TestFqdnHostnames-819811677 tempest-ServersV294TestFqdnHostnames-819811677-project-member] Task: {'id': task-5115984, 'name': Rename_Task, 'duration_secs': 0.323224} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.262549] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-0b112b08-1593-41b5-b38e-86e5b4d22081 tempest-ServersV294TestFqdnHostnames-819811677 tempest-ServersV294TestFqdnHostnames-819811677-project-member] [instance: c8be64a3-1bfa-41a3-b9be-68a7e6e052e0] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 829.262975] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c8d3d2e4-a98e-4041-b11d-848a86e7d2b2 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.270795] env[65726]: DEBUG oslo_vmware.api [None req-0b112b08-1593-41b5-b38e-86e5b4d22081 tempest-ServersV294TestFqdnHostnames-819811677 tempest-ServersV294TestFqdnHostnames-819811677-project-member] Waiting for the task: (returnval){ [ 829.270795] env[65726]: value = "task-5115985" [ 829.270795] env[65726]: _type = "Task" [ 829.270795] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.280872] env[65726]: DEBUG oslo_vmware.api [None req-0b112b08-1593-41b5-b38e-86e5b4d22081 tempest-ServersV294TestFqdnHostnames-819811677 tempest-ServersV294TestFqdnHostnames-819811677-project-member] Task: {'id': task-5115985, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.334374] env[65726]: DEBUG nova.compute.manager [req-9b401ad2-6e37-4c56-9e36-20b866d91330 req-a4dc17a3-40d7-4ff7-8ae9-ebeeb6804f30 service nova] [instance: 01d86089-6b9b-4588-864e-ef91375a1eea] Received event network-vif-plugged-beb674aa-cade-455d-a97a-4da9699c9a70 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 829.334854] env[65726]: DEBUG oslo_concurrency.lockutils [req-9b401ad2-6e37-4c56-9e36-20b866d91330 req-a4dc17a3-40d7-4ff7-8ae9-ebeeb6804f30 service nova] Acquiring lock "01d86089-6b9b-4588-864e-ef91375a1eea-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 829.335901] env[65726]: DEBUG oslo_concurrency.lockutils [req-9b401ad2-6e37-4c56-9e36-20b866d91330 req-a4dc17a3-40d7-4ff7-8ae9-ebeeb6804f30 service nova] Lock "01d86089-6b9b-4588-864e-ef91375a1eea-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 829.336475] env[65726]: DEBUG oslo_concurrency.lockutils [req-9b401ad2-6e37-4c56-9e36-20b866d91330 req-a4dc17a3-40d7-4ff7-8ae9-ebeeb6804f30 service nova] Lock "01d86089-6b9b-4588-864e-ef91375a1eea-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 829.336475] env[65726]: DEBUG nova.compute.manager [req-9b401ad2-6e37-4c56-9e36-20b866d91330 req-a4dc17a3-40d7-4ff7-8ae9-ebeeb6804f30 service nova] [instance: 01d86089-6b9b-4588-864e-ef91375a1eea] No waiting events found dispatching network-vif-plugged-beb674aa-cade-455d-a97a-4da9699c9a70 {{(pid=65726) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 829.336475] env[65726]: WARNING nova.compute.manager [req-9b401ad2-6e37-4c56-9e36-20b866d91330 req-a4dc17a3-40d7-4ff7-8ae9-ebeeb6804f30 service nova] [instance: 01d86089-6b9b-4588-864e-ef91375a1eea] Received unexpected event network-vif-plugged-beb674aa-cade-455d-a97a-4da9699c9a70 for instance with vm_state building and task_state spawning. [ 829.336613] env[65726]: DEBUG nova.compute.manager [req-9b401ad2-6e37-4c56-9e36-20b866d91330 req-a4dc17a3-40d7-4ff7-8ae9-ebeeb6804f30 service nova] [instance: 01d86089-6b9b-4588-864e-ef91375a1eea] Received event network-changed-beb674aa-cade-455d-a97a-4da9699c9a70 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 829.336709] env[65726]: DEBUG nova.compute.manager [req-9b401ad2-6e37-4c56-9e36-20b866d91330 req-a4dc17a3-40d7-4ff7-8ae9-ebeeb6804f30 service nova] [instance: 01d86089-6b9b-4588-864e-ef91375a1eea] Refreshing instance network info cache due to event network-changed-beb674aa-cade-455d-a97a-4da9699c9a70. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 829.336873] env[65726]: DEBUG oslo_concurrency.lockutils [req-9b401ad2-6e37-4c56-9e36-20b866d91330 req-a4dc17a3-40d7-4ff7-8ae9-ebeeb6804f30 service nova] Acquiring lock "refresh_cache-01d86089-6b9b-4588-864e-ef91375a1eea" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 829.395527] env[65726]: DEBUG oslo_concurrency.lockutils [None req-645c8c71-cd4d-4583-8270-b88bb6120d10 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Releasing lock "refresh_cache-01d86089-6b9b-4588-864e-ef91375a1eea" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 829.395940] env[65726]: DEBUG nova.compute.manager [None req-645c8c71-cd4d-4583-8270-b88bb6120d10 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 01d86089-6b9b-4588-864e-ef91375a1eea] Instance network_info: |[{"id": "beb674aa-cade-455d-a97a-4da9699c9a70", "address": "fa:16:3e:f9:f6:08", "network": {"id": "e206b5b9-da36-4e6f-926f-914dec38fac2", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1053201655-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "39067912d091464192516c08eb576696", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7c80243e-93a7-4a95-bc8d-e9534bacd66e", "external-id": "nsx-vlan-transportzone-306", "segmentation_id": 306, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbeb674aa-ca", "ovs_interfaceid": "beb674aa-cade-455d-a97a-4da9699c9a70", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 829.396379] env[65726]: INFO nova.compute.manager [None req-76673fd1-cee8-44a1-9e95-0d5b0dfe1651 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] [instance: 3cbbe146-be3a-43d5-867d-4d669884758c] Rescuing [ 829.396623] env[65726]: DEBUG oslo_concurrency.lockutils [None req-76673fd1-cee8-44a1-9e95-0d5b0dfe1651 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] Acquiring lock "refresh_cache-3cbbe146-be3a-43d5-867d-4d669884758c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 829.396779] env[65726]: DEBUG oslo_concurrency.lockutils [None req-76673fd1-cee8-44a1-9e95-0d5b0dfe1651 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] Acquired lock "refresh_cache-3cbbe146-be3a-43d5-867d-4d669884758c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 829.396941] env[65726]: DEBUG nova.network.neutron [None req-76673fd1-cee8-44a1-9e95-0d5b0dfe1651 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] [instance: 3cbbe146-be3a-43d5-867d-4d669884758c] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 829.401027] env[65726]: DEBUG oslo_concurrency.lockutils [req-9b401ad2-6e37-4c56-9e36-20b866d91330 req-a4dc17a3-40d7-4ff7-8ae9-ebeeb6804f30 service nova] Acquired lock "refresh_cache-01d86089-6b9b-4588-864e-ef91375a1eea" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 829.401027] env[65726]: DEBUG nova.network.neutron [req-9b401ad2-6e37-4c56-9e36-20b866d91330 req-a4dc17a3-40d7-4ff7-8ae9-ebeeb6804f30 service nova] [instance: 01d86089-6b9b-4588-864e-ef91375a1eea] Refreshing network info cache for port beb674aa-cade-455d-a97a-4da9699c9a70 {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 829.401027] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-645c8c71-cd4d-4583-8270-b88bb6120d10 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 01d86089-6b9b-4588-864e-ef91375a1eea] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f9:f6:08', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '7c80243e-93a7-4a95-bc8d-e9534bacd66e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'beb674aa-cade-455d-a97a-4da9699c9a70', 'vif_model': 'vmxnet3'}] {{(pid=65726) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 829.407192] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-645c8c71-cd4d-4583-8270-b88bb6120d10 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 829.408909] env[65726]: WARNING openstack [None req-76673fd1-cee8-44a1-9e95-0d5b0dfe1651 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 829.409316] env[65726]: WARNING openstack [None req-76673fd1-cee8-44a1-9e95-0d5b0dfe1651 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 829.416242] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 01d86089-6b9b-4588-864e-ef91375a1eea] Creating VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 829.417065] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-04a76c31-1a74-4492-93a0-5a378f965df2 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.441638] env[65726]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 829.441638] env[65726]: value = "task-5115986" [ 829.441638] env[65726]: _type = "Task" [ 829.441638] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.450608] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115986, 'name': CreateVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.454402] env[65726]: ERROR nova.scheduler.client.report [None req-ebfc0e41-4845-41ca-9f74-d2a2946626c0 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] [req-3f47ad99-9353-438d-ae72-b082601c7395] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 07c4692f-bdb4-4058-9173-ff9664830295. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-3f47ad99-9353-438d-ae72-b082601c7395"}]} [ 829.473302] env[65726]: DEBUG nova.scheduler.client.report [None req-ebfc0e41-4845-41ca-9f74-d2a2946626c0 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Refreshing inventories for resource provider 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 829.497578] env[65726]: DEBUG nova.scheduler.client.report [None req-ebfc0e41-4845-41ca-9f74-d2a2946626c0 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Updating ProviderTree inventory for provider 07c4692f-bdb4-4058-9173-ff9664830295 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 829.497857] env[65726]: DEBUG nova.compute.provider_tree [None req-ebfc0e41-4845-41ca-9f74-d2a2946626c0 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Updating inventory in ProviderTree for provider 07c4692f-bdb4-4058-9173-ff9664830295 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 829.528903] env[65726]: DEBUG nova.scheduler.client.report [None req-ebfc0e41-4845-41ca-9f74-d2a2946626c0 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Refreshing aggregate associations for resource provider 07c4692f-bdb4-4058-9173-ff9664830295, aggregates: None {{(pid=65726) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 829.535290] env[65726]: WARNING openstack [None req-76673fd1-cee8-44a1-9e95-0d5b0dfe1651 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 829.535290] env[65726]: WARNING openstack [None req-76673fd1-cee8-44a1-9e95-0d5b0dfe1651 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 829.567127] env[65726]: DEBUG nova.scheduler.client.report [None req-ebfc0e41-4845-41ca-9f74-d2a2946626c0 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Refreshing trait associations for resource provider 07c4692f-bdb4-4058-9173-ff9664830295, traits: COMPUTE_IMAGE_TYPE_ISO,HW_ARCH_X86_64,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=65726) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 829.606169] env[65726]: WARNING openstack [None req-76673fd1-cee8-44a1-9e95-0d5b0dfe1651 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 829.606698] env[65726]: WARNING openstack [None req-76673fd1-cee8-44a1-9e95-0d5b0dfe1651 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 829.715632] env[65726]: DEBUG oslo_concurrency.lockutils [None req-73b0a627-5d81-4bf2-833a-caff3534c1e3 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 829.723343] env[65726]: DEBUG nova.network.neutron [None req-76673fd1-cee8-44a1-9e95-0d5b0dfe1651 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] [instance: 3cbbe146-be3a-43d5-867d-4d669884758c] Updating instance_info_cache with network_info: [{"id": "40240c4e-9f3e-46aa-bd75-c62c2776d9b6", "address": "fa:16:3e:8f:bd:09", "network": {"id": "2266eb6e-4c8f-4a3f-85d3-0c06022c0d5e", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-716412627-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "92c1cedbc08545d58fc7172ac95e2c23", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "24376631-ee89-4ff1-b8ac-f09911fc8329", "external-id": "nsx-vlan-transportzone-960", "segmentation_id": 960, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap40240c4e-9f", "ovs_interfaceid": "40240c4e-9f3e-46aa-bd75-c62c2776d9b6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 829.788465] env[65726]: DEBUG oslo_vmware.api [None req-0b112b08-1593-41b5-b38e-86e5b4d22081 tempest-ServersV294TestFqdnHostnames-819811677 tempest-ServersV294TestFqdnHostnames-819811677-project-member] Task: {'id': task-5115985, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.909602] env[65726]: WARNING openstack [req-9b401ad2-6e37-4c56-9e36-20b866d91330 req-a4dc17a3-40d7-4ff7-8ae9-ebeeb6804f30 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 829.910015] env[65726]: WARNING openstack [req-9b401ad2-6e37-4c56-9e36-20b866d91330 req-a4dc17a3-40d7-4ff7-8ae9-ebeeb6804f30 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 829.956165] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115986, 'name': CreateVM_Task, 'duration_secs': 0.451154} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.956433] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 01d86089-6b9b-4588-864e-ef91375a1eea] Created VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 829.957631] env[65726]: WARNING openstack [None req-645c8c71-cd4d-4583-8270-b88bb6120d10 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 829.958083] env[65726]: WARNING openstack [None req-645c8c71-cd4d-4583-8270-b88bb6120d10 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 829.964578] env[65726]: DEBUG oslo_concurrency.lockutils [None req-645c8c71-cd4d-4583-8270-b88bb6120d10 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 829.964727] env[65726]: DEBUG oslo_concurrency.lockutils [None req-645c8c71-cd4d-4583-8270-b88bb6120d10 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 829.965191] env[65726]: DEBUG oslo_concurrency.lockutils [None req-645c8c71-cd4d-4583-8270-b88bb6120d10 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 829.968508] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d297ae92-c042-4a93-95a9-9789db6300c8 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.979576] env[65726]: DEBUG oslo_vmware.api [None req-645c8c71-cd4d-4583-8270-b88bb6120d10 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Waiting for the task: (returnval){ [ 829.979576] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52327374-9d7b-203e-75fd-3c0701e327da" [ 829.979576] env[65726]: _type = "Task" [ 829.979576] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.996568] env[65726]: DEBUG oslo_vmware.api [None req-645c8c71-cd4d-4583-8270-b88bb6120d10 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52327374-9d7b-203e-75fd-3c0701e327da, 'name': SearchDatastore_Task, 'duration_secs': 0.012336} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.997059] env[65726]: DEBUG oslo_concurrency.lockutils [None req-645c8c71-cd4d-4583-8270-b88bb6120d10 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 829.997369] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-645c8c71-cd4d-4583-8270-b88bb6120d10 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 01d86089-6b9b-4588-864e-ef91375a1eea] Processing image b52362a3-ee8a-4cbf-b06f-513b0cc8f95c {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 829.997653] env[65726]: DEBUG oslo_concurrency.lockutils [None req-645c8c71-cd4d-4583-8270-b88bb6120d10 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 829.997867] env[65726]: DEBUG oslo_concurrency.lockutils [None req-645c8c71-cd4d-4583-8270-b88bb6120d10 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 829.998435] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-645c8c71-cd4d-4583-8270-b88bb6120d10 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 830.001658] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ea968a69-8833-403a-83bd-842376b3ed8f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.012769] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-645c8c71-cd4d-4583-8270-b88bb6120d10 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 830.012769] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-645c8c71-cd4d-4583-8270-b88bb6120d10 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65726) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 830.013501] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fb0727a6-30ad-4321-a180-6f5c0e305b9d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.023640] env[65726]: DEBUG oslo_vmware.api [None req-645c8c71-cd4d-4583-8270-b88bb6120d10 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Waiting for the task: (returnval){ [ 830.023640] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5201d819-76a4-6709-66fc-d8ee2bd6efc9" [ 830.023640] env[65726]: _type = "Task" [ 830.023640] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.037189] env[65726]: DEBUG oslo_vmware.api [None req-645c8c71-cd4d-4583-8270-b88bb6120d10 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5201d819-76a4-6709-66fc-d8ee2bd6efc9, 'name': SearchDatastore_Task, 'duration_secs': 0.012523} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.041922] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b8f044f2-1c7c-4f91-a89c-3858b8a8d0aa {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.047380] env[65726]: DEBUG oslo_vmware.api [None req-645c8c71-cd4d-4583-8270-b88bb6120d10 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Waiting for the task: (returnval){ [ 830.047380] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52b0e7ac-259b-4a37-dba5-7748292c3c6b" [ 830.047380] env[65726]: _type = "Task" [ 830.047380] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.054300] env[65726]: WARNING openstack [req-9b401ad2-6e37-4c56-9e36-20b866d91330 req-a4dc17a3-40d7-4ff7-8ae9-ebeeb6804f30 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 830.054941] env[65726]: WARNING openstack [req-9b401ad2-6e37-4c56-9e36-20b866d91330 req-a4dc17a3-40d7-4ff7-8ae9-ebeeb6804f30 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 830.069299] env[65726]: DEBUG oslo_vmware.api [None req-645c8c71-cd4d-4583-8270-b88bb6120d10 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52b0e7ac-259b-4a37-dba5-7748292c3c6b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.137323] env[65726]: WARNING openstack [req-9b401ad2-6e37-4c56-9e36-20b866d91330 req-a4dc17a3-40d7-4ff7-8ae9-ebeeb6804f30 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 830.137729] env[65726]: WARNING openstack [req-9b401ad2-6e37-4c56-9e36-20b866d91330 req-a4dc17a3-40d7-4ff7-8ae9-ebeeb6804f30 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 830.165873] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2313e91-bb46-45e4-993d-970e5fbd2f4b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.176352] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ca9533f-018f-4549-894d-4516728ce2a8 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.184599] env[65726]: DEBUG oslo_concurrency.lockutils [None req-34f1738e-744a-4a46-b434-5db18d6e964a tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Acquiring lock "4ea2b70b-0cfe-4fbf-817a-baa028b8372a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 830.184892] env[65726]: DEBUG oslo_concurrency.lockutils [None req-34f1738e-744a-4a46-b434-5db18d6e964a tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Lock "4ea2b70b-0cfe-4fbf-817a-baa028b8372a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 830.216436] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1a1cd76-2870-48fe-97dc-e91035f40dec {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.228179] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-698c3055-8b31-49fb-83db-1f5d014db531 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.230588] env[65726]: DEBUG oslo_concurrency.lockutils [None req-76673fd1-cee8-44a1-9e95-0d5b0dfe1651 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] Releasing lock "refresh_cache-3cbbe146-be3a-43d5-867d-4d669884758c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 830.245549] env[65726]: DEBUG nova.compute.provider_tree [None req-ebfc0e41-4845-41ca-9f74-d2a2946626c0 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Updating inventory in ProviderTree for provider 07c4692f-bdb4-4058-9173-ff9664830295 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 830.272067] env[65726]: DEBUG nova.network.neutron [req-9b401ad2-6e37-4c56-9e36-20b866d91330 req-a4dc17a3-40d7-4ff7-8ae9-ebeeb6804f30 service nova] [instance: 01d86089-6b9b-4588-864e-ef91375a1eea] Updated VIF entry in instance network info cache for port beb674aa-cade-455d-a97a-4da9699c9a70. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 830.272577] env[65726]: DEBUG nova.network.neutron [req-9b401ad2-6e37-4c56-9e36-20b866d91330 req-a4dc17a3-40d7-4ff7-8ae9-ebeeb6804f30 service nova] [instance: 01d86089-6b9b-4588-864e-ef91375a1eea] Updating instance_info_cache with network_info: [{"id": "beb674aa-cade-455d-a97a-4da9699c9a70", "address": "fa:16:3e:f9:f6:08", "network": {"id": "e206b5b9-da36-4e6f-926f-914dec38fac2", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1053201655-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "39067912d091464192516c08eb576696", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7c80243e-93a7-4a95-bc8d-e9534bacd66e", "external-id": "nsx-vlan-transportzone-306", "segmentation_id": 306, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbeb674aa-ca", "ovs_interfaceid": "beb674aa-cade-455d-a97a-4da9699c9a70", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 830.288605] env[65726]: DEBUG oslo_vmware.api [None req-0b112b08-1593-41b5-b38e-86e5b4d22081 tempest-ServersV294TestFqdnHostnames-819811677 tempest-ServersV294TestFqdnHostnames-819811677-project-member] Task: {'id': task-5115985, 'name': PowerOnVM_Task, 'duration_secs': 0.716502} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.289015] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-0b112b08-1593-41b5-b38e-86e5b4d22081 tempest-ServersV294TestFqdnHostnames-819811677 tempest-ServersV294TestFqdnHostnames-819811677-project-member] [instance: c8be64a3-1bfa-41a3-b9be-68a7e6e052e0] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 830.289332] env[65726]: INFO nova.compute.manager [None req-0b112b08-1593-41b5-b38e-86e5b4d22081 tempest-ServersV294TestFqdnHostnames-819811677 tempest-ServersV294TestFqdnHostnames-819811677-project-member] [instance: c8be64a3-1bfa-41a3-b9be-68a7e6e052e0] Took 9.55 seconds to spawn the instance on the hypervisor. [ 830.289614] env[65726]: DEBUG nova.compute.manager [None req-0b112b08-1593-41b5-b38e-86e5b4d22081 tempest-ServersV294TestFqdnHostnames-819811677 tempest-ServersV294TestFqdnHostnames-819811677-project-member] [instance: c8be64a3-1bfa-41a3-b9be-68a7e6e052e0] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 830.290164] env[65726]: DEBUG oslo_concurrency.lockutils [None req-398d7f82-4e73-43c8-8dc0-2fe897bf965f tempest-ServersNegativeTestMultiTenantJSON-944098855 tempest-ServersNegativeTestMultiTenantJSON-944098855-project-member] Acquiring lock "018c8ac5-66c4-4a9a-ab45-85e84f92b4fe" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 830.290494] env[65726]: DEBUG oslo_concurrency.lockutils [None req-398d7f82-4e73-43c8-8dc0-2fe897bf965f tempest-ServersNegativeTestMultiTenantJSON-944098855 tempest-ServersNegativeTestMultiTenantJSON-944098855-project-member] Lock "018c8ac5-66c4-4a9a-ab45-85e84f92b4fe" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 830.290792] env[65726]: DEBUG oslo_concurrency.lockutils [None req-398d7f82-4e73-43c8-8dc0-2fe897bf965f tempest-ServersNegativeTestMultiTenantJSON-944098855 tempest-ServersNegativeTestMultiTenantJSON-944098855-project-member] Acquiring lock "018c8ac5-66c4-4a9a-ab45-85e84f92b4fe-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 830.291099] env[65726]: DEBUG oslo_concurrency.lockutils [None req-398d7f82-4e73-43c8-8dc0-2fe897bf965f tempest-ServersNegativeTestMultiTenantJSON-944098855 tempest-ServersNegativeTestMultiTenantJSON-944098855-project-member] Lock "018c8ac5-66c4-4a9a-ab45-85e84f92b4fe-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 830.291370] env[65726]: DEBUG oslo_concurrency.lockutils [None req-398d7f82-4e73-43c8-8dc0-2fe897bf965f tempest-ServersNegativeTestMultiTenantJSON-944098855 tempest-ServersNegativeTestMultiTenantJSON-944098855-project-member] Lock "018c8ac5-66c4-4a9a-ab45-85e84f92b4fe-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 830.294270] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a84ed4e5-6aa6-4510-9780-a2fd64ba7a53 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.297669] env[65726]: INFO nova.compute.manager [None req-398d7f82-4e73-43c8-8dc0-2fe897bf965f tempest-ServersNegativeTestMultiTenantJSON-944098855 tempest-ServersNegativeTestMultiTenantJSON-944098855-project-member] [instance: 018c8ac5-66c4-4a9a-ab45-85e84f92b4fe] Terminating instance [ 830.559552] env[65726]: DEBUG oslo_vmware.api [None req-645c8c71-cd4d-4583-8270-b88bb6120d10 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52b0e7ac-259b-4a37-dba5-7748292c3c6b, 'name': SearchDatastore_Task, 'duration_secs': 0.011228} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.559552] env[65726]: DEBUG oslo_concurrency.lockutils [None req-645c8c71-cd4d-4583-8270-b88bb6120d10 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 830.559782] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-645c8c71-cd4d-4583-8270-b88bb6120d10 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] 01d86089-6b9b-4588-864e-ef91375a1eea/01d86089-6b9b-4588-864e-ef91375a1eea.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 830.560131] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-fe35a4b9-467c-4615-ba9f-eeaa11c21209 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.569724] env[65726]: DEBUG oslo_vmware.api [None req-645c8c71-cd4d-4583-8270-b88bb6120d10 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Waiting for the task: (returnval){ [ 830.569724] env[65726]: value = "task-5115987" [ 830.569724] env[65726]: _type = "Task" [ 830.569724] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.578799] env[65726]: DEBUG oslo_vmware.api [None req-645c8c71-cd4d-4583-8270-b88bb6120d10 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': task-5115987, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.770954] env[65726]: ERROR nova.scheduler.client.report [None req-ebfc0e41-4845-41ca-9f74-d2a2946626c0 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] [req-b4574963-8a8e-41ef-bab2-9341945e16cb] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 07c4692f-bdb4-4058-9173-ff9664830295. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-b4574963-8a8e-41ef-bab2-9341945e16cb"}]} [ 830.781851] env[65726]: DEBUG oslo_concurrency.lockutils [req-9b401ad2-6e37-4c56-9e36-20b866d91330 req-a4dc17a3-40d7-4ff7-8ae9-ebeeb6804f30 service nova] Releasing lock "refresh_cache-01d86089-6b9b-4588-864e-ef91375a1eea" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 830.790356] env[65726]: DEBUG nova.scheduler.client.report [None req-ebfc0e41-4845-41ca-9f74-d2a2946626c0 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Refreshing inventories for resource provider 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 830.804787] env[65726]: DEBUG nova.compute.manager [None req-398d7f82-4e73-43c8-8dc0-2fe897bf965f tempest-ServersNegativeTestMultiTenantJSON-944098855 tempest-ServersNegativeTestMultiTenantJSON-944098855-project-member] [instance: 018c8ac5-66c4-4a9a-ab45-85e84f92b4fe] Start destroying the instance on the hypervisor. {{(pid=65726) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 830.804922] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-398d7f82-4e73-43c8-8dc0-2fe897bf965f tempest-ServersNegativeTestMultiTenantJSON-944098855 tempest-ServersNegativeTestMultiTenantJSON-944098855-project-member] [instance: 018c8ac5-66c4-4a9a-ab45-85e84f92b4fe] Destroying instance {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 830.806292] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c823774-74fd-4894-af9a-e959932c7229 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.810358] env[65726]: DEBUG nova.scheduler.client.report [None req-ebfc0e41-4845-41ca-9f74-d2a2946626c0 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Updating ProviderTree inventory for provider 07c4692f-bdb4-4058-9173-ff9664830295 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 830.810576] env[65726]: DEBUG nova.compute.provider_tree [None req-ebfc0e41-4845-41ca-9f74-d2a2946626c0 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Updating inventory in ProviderTree for provider 07c4692f-bdb4-4058-9173-ff9664830295 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 830.834475] env[65726]: INFO nova.compute.manager [None req-0b112b08-1593-41b5-b38e-86e5b4d22081 tempest-ServersV294TestFqdnHostnames-819811677 tempest-ServersV294TestFqdnHostnames-819811677-project-member] [instance: c8be64a3-1bfa-41a3-b9be-68a7e6e052e0] Took 53.94 seconds to build instance. [ 830.835041] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-398d7f82-4e73-43c8-8dc0-2fe897bf965f tempest-ServersNegativeTestMultiTenantJSON-944098855 tempest-ServersNegativeTestMultiTenantJSON-944098855-project-member] [instance: 018c8ac5-66c4-4a9a-ab45-85e84f92b4fe] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 830.835041] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8477dd3a-d757-460c-84a9-c487b650b062 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.843769] env[65726]: DEBUG nova.scheduler.client.report [None req-ebfc0e41-4845-41ca-9f74-d2a2946626c0 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Refreshing aggregate associations for resource provider 07c4692f-bdb4-4058-9173-ff9664830295, aggregates: None {{(pid=65726) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 830.848481] env[65726]: DEBUG oslo_vmware.api [None req-398d7f82-4e73-43c8-8dc0-2fe897bf965f tempest-ServersNegativeTestMultiTenantJSON-944098855 tempest-ServersNegativeTestMultiTenantJSON-944098855-project-member] Waiting for the task: (returnval){ [ 830.848481] env[65726]: value = "task-5115988" [ 830.848481] env[65726]: _type = "Task" [ 830.848481] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.858936] env[65726]: DEBUG oslo_vmware.api [None req-398d7f82-4e73-43c8-8dc0-2fe897bf965f tempest-ServersNegativeTestMultiTenantJSON-944098855 tempest-ServersNegativeTestMultiTenantJSON-944098855-project-member] Task: {'id': task-5115988, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.874255] env[65726]: DEBUG nova.scheduler.client.report [None req-ebfc0e41-4845-41ca-9f74-d2a2946626c0 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Refreshing trait associations for resource provider 07c4692f-bdb4-4058-9173-ff9664830295, traits: COMPUTE_IMAGE_TYPE_ISO,HW_ARCH_X86_64,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=65726) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 831.087482] env[65726]: DEBUG oslo_vmware.api [None req-645c8c71-cd4d-4583-8270-b88bb6120d10 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': task-5115987, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.338056] env[65726]: DEBUG oslo_concurrency.lockutils [None req-0b112b08-1593-41b5-b38e-86e5b4d22081 tempest-ServersV294TestFqdnHostnames-819811677 tempest-ServersV294TestFqdnHostnames-819811677-project-member] Lock "c8be64a3-1bfa-41a3-b9be-68a7e6e052e0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 75.518s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 831.365045] env[65726]: DEBUG oslo_vmware.api [None req-398d7f82-4e73-43c8-8dc0-2fe897bf965f tempest-ServersNegativeTestMultiTenantJSON-944098855 tempest-ServersNegativeTestMultiTenantJSON-944098855-project-member] Task: {'id': task-5115988, 'name': PowerOffVM_Task, 'duration_secs': 0.278509} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 831.365045] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-398d7f82-4e73-43c8-8dc0-2fe897bf965f tempest-ServersNegativeTestMultiTenantJSON-944098855 tempest-ServersNegativeTestMultiTenantJSON-944098855-project-member] [instance: 018c8ac5-66c4-4a9a-ab45-85e84f92b4fe] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 831.365045] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-398d7f82-4e73-43c8-8dc0-2fe897bf965f tempest-ServersNegativeTestMultiTenantJSON-944098855 tempest-ServersNegativeTestMultiTenantJSON-944098855-project-member] [instance: 018c8ac5-66c4-4a9a-ab45-85e84f92b4fe] Unregistering the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 831.365246] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b55beb07-2def-400c-b765-1b4cfa370bcd {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.367642] env[65726]: DEBUG oslo_concurrency.lockutils [None req-34f2bf54-ef3a-4b87-a2a0-3305b3d54834 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Acquiring lock "608cc0ea-3f6b-4b4f-83c1-01aa50999d63" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 831.367936] env[65726]: DEBUG oslo_concurrency.lockutils [None req-34f2bf54-ef3a-4b87-a2a0-3305b3d54834 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Lock "608cc0ea-3f6b-4b4f-83c1-01aa50999d63" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 831.368295] env[65726]: DEBUG nova.compute.manager [None req-34f2bf54-ef3a-4b87-a2a0-3305b3d54834 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 608cc0ea-3f6b-4b4f-83c1-01aa50999d63] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 831.372429] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d4d7452-fa20-4858-b672-b941e615f9a1 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.381571] env[65726]: DEBUG nova.compute.manager [None req-34f2bf54-ef3a-4b87-a2a0-3305b3d54834 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 608cc0ea-3f6b-4b4f-83c1-01aa50999d63] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=65726) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3459}} [ 831.382925] env[65726]: DEBUG nova.objects.instance [None req-34f2bf54-ef3a-4b87-a2a0-3305b3d54834 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Lazy-loading 'flavor' on Instance uuid 608cc0ea-3f6b-4b4f-83c1-01aa50999d63 {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 831.394341] env[65726]: DEBUG nova.compute.manager [req-1c17cde7-0a51-42e2-b50d-4c0a77648ecd req-bffb8806-15d0-4ea3-9ae8-800f5970e40f service nova] [instance: c8be64a3-1bfa-41a3-b9be-68a7e6e052e0] Received event network-changed-e9d6f425-5466-4d5d-ae7f-80a433defe57 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 831.394539] env[65726]: DEBUG nova.compute.manager [req-1c17cde7-0a51-42e2-b50d-4c0a77648ecd req-bffb8806-15d0-4ea3-9ae8-800f5970e40f service nova] [instance: c8be64a3-1bfa-41a3-b9be-68a7e6e052e0] Refreshing instance network info cache due to event network-changed-e9d6f425-5466-4d5d-ae7f-80a433defe57. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 831.395813] env[65726]: DEBUG oslo_concurrency.lockutils [req-1c17cde7-0a51-42e2-b50d-4c0a77648ecd req-bffb8806-15d0-4ea3-9ae8-800f5970e40f service nova] Acquiring lock "refresh_cache-c8be64a3-1bfa-41a3-b9be-68a7e6e052e0" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 831.395813] env[65726]: DEBUG oslo_concurrency.lockutils [req-1c17cde7-0a51-42e2-b50d-4c0a77648ecd req-bffb8806-15d0-4ea3-9ae8-800f5970e40f service nova] Acquired lock "refresh_cache-c8be64a3-1bfa-41a3-b9be-68a7e6e052e0" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 831.395813] env[65726]: DEBUG nova.network.neutron [req-1c17cde7-0a51-42e2-b50d-4c0a77648ecd req-bffb8806-15d0-4ea3-9ae8-800f5970e40f service nova] [instance: c8be64a3-1bfa-41a3-b9be-68a7e6e052e0] Refreshing network info cache for port e9d6f425-5466-4d5d-ae7f-80a433defe57 {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 831.445251] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-398d7f82-4e73-43c8-8dc0-2fe897bf965f tempest-ServersNegativeTestMultiTenantJSON-944098855 tempest-ServersNegativeTestMultiTenantJSON-944098855-project-member] [instance: 018c8ac5-66c4-4a9a-ab45-85e84f92b4fe] Unregistered the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 831.445434] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-398d7f82-4e73-43c8-8dc0-2fe897bf965f tempest-ServersNegativeTestMultiTenantJSON-944098855 tempest-ServersNegativeTestMultiTenantJSON-944098855-project-member] [instance: 018c8ac5-66c4-4a9a-ab45-85e84f92b4fe] Deleting contents of the VM from datastore datastore2 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 831.445497] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-398d7f82-4e73-43c8-8dc0-2fe897bf965f tempest-ServersNegativeTestMultiTenantJSON-944098855 tempest-ServersNegativeTestMultiTenantJSON-944098855-project-member] Deleting the datastore file [datastore2] 018c8ac5-66c4-4a9a-ab45-85e84f92b4fe {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 831.445781] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e3b869eb-74b0-4a5c-a30c-be55e24adccf {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.456079] env[65726]: DEBUG oslo_vmware.api [None req-398d7f82-4e73-43c8-8dc0-2fe897bf965f tempest-ServersNegativeTestMultiTenantJSON-944098855 tempest-ServersNegativeTestMultiTenantJSON-944098855-project-member] Waiting for the task: (returnval){ [ 831.456079] env[65726]: value = "task-5115990" [ 831.456079] env[65726]: _type = "Task" [ 831.456079] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.466189] env[65726]: DEBUG oslo_vmware.api [None req-398d7f82-4e73-43c8-8dc0-2fe897bf965f tempest-ServersNegativeTestMultiTenantJSON-944098855 tempest-ServersNegativeTestMultiTenantJSON-944098855-project-member] Task: {'id': task-5115990, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.470712] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed656040-d405-46d1-bb1b-4224f4ec3162 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.479042] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e32889e6-d472-458a-a93a-ebfdfd4a2a88 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.510858] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18f83060-36c4-4c2c-a5b2-86155c96c75d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.519727] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba3fb0c2-2571-4755-a41c-b6fb31b8ebd0 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.535031] env[65726]: DEBUG nova.compute.provider_tree [None req-ebfc0e41-4845-41ca-9f74-d2a2946626c0 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Updating inventory in ProviderTree for provider 07c4692f-bdb4-4058-9173-ff9664830295 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 831.582845] env[65726]: DEBUG oslo_vmware.api [None req-645c8c71-cd4d-4583-8270-b88bb6120d10 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': task-5115987, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.545275} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 831.583140] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-645c8c71-cd4d-4583-8270-b88bb6120d10 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] 01d86089-6b9b-4588-864e-ef91375a1eea/01d86089-6b9b-4588-864e-ef91375a1eea.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 831.583848] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-645c8c71-cd4d-4583-8270-b88bb6120d10 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 01d86089-6b9b-4588-864e-ef91375a1eea] Extending root virtual disk to 1048576 {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 831.583848] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3bad72fd-7966-4ce6-a64a-465dcd98c45c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.591359] env[65726]: DEBUG oslo_vmware.api [None req-645c8c71-cd4d-4583-8270-b88bb6120d10 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Waiting for the task: (returnval){ [ 831.591359] env[65726]: value = "task-5115991" [ 831.591359] env[65726]: _type = "Task" [ 831.591359] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.601746] env[65726]: DEBUG oslo_vmware.api [None req-645c8c71-cd4d-4583-8270-b88bb6120d10 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': task-5115991, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.773863] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-76673fd1-cee8-44a1-9e95-0d5b0dfe1651 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] [instance: 3cbbe146-be3a-43d5-867d-4d669884758c] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 831.774321] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-00cc3fb1-9d2a-4c00-8e17-8335dfe79e6f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.783124] env[65726]: DEBUG oslo_vmware.api [None req-76673fd1-cee8-44a1-9e95-0d5b0dfe1651 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] Waiting for the task: (returnval){ [ 831.783124] env[65726]: value = "task-5115992" [ 831.783124] env[65726]: _type = "Task" [ 831.783124] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.794865] env[65726]: DEBUG oslo_vmware.api [None req-76673fd1-cee8-44a1-9e95-0d5b0dfe1651 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] Task: {'id': task-5115992, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.846107] env[65726]: DEBUG nova.compute.manager [None req-b20cc073-fafc-44fa-bcf9-9893dc98ebec tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04] Starting instance... {{(pid=65726) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 831.899618] env[65726]: WARNING openstack [req-1c17cde7-0a51-42e2-b50d-4c0a77648ecd req-bffb8806-15d0-4ea3-9ae8-800f5970e40f service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 831.899943] env[65726]: WARNING openstack [req-1c17cde7-0a51-42e2-b50d-4c0a77648ecd req-bffb8806-15d0-4ea3-9ae8-800f5970e40f service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 831.968950] env[65726]: DEBUG oslo_vmware.api [None req-398d7f82-4e73-43c8-8dc0-2fe897bf965f tempest-ServersNegativeTestMultiTenantJSON-944098855 tempest-ServersNegativeTestMultiTenantJSON-944098855-project-member] Task: {'id': task-5115990, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.15618} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 831.969315] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-398d7f82-4e73-43c8-8dc0-2fe897bf965f tempest-ServersNegativeTestMultiTenantJSON-944098855 tempest-ServersNegativeTestMultiTenantJSON-944098855-project-member] Deleted the datastore file {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 831.969505] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-398d7f82-4e73-43c8-8dc0-2fe897bf965f tempest-ServersNegativeTestMultiTenantJSON-944098855 tempest-ServersNegativeTestMultiTenantJSON-944098855-project-member] [instance: 018c8ac5-66c4-4a9a-ab45-85e84f92b4fe] Deleted contents of the VM from datastore datastore2 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 831.969682] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-398d7f82-4e73-43c8-8dc0-2fe897bf965f tempest-ServersNegativeTestMultiTenantJSON-944098855 tempest-ServersNegativeTestMultiTenantJSON-944098855-project-member] [instance: 018c8ac5-66c4-4a9a-ab45-85e84f92b4fe] Instance destroyed {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 831.969850] env[65726]: INFO nova.compute.manager [None req-398d7f82-4e73-43c8-8dc0-2fe897bf965f tempest-ServersNegativeTestMultiTenantJSON-944098855 tempest-ServersNegativeTestMultiTenantJSON-944098855-project-member] [instance: 018c8ac5-66c4-4a9a-ab45-85e84f92b4fe] Took 1.17 seconds to destroy the instance on the hypervisor. [ 831.970125] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-398d7f82-4e73-43c8-8dc0-2fe897bf965f tempest-ServersNegativeTestMultiTenantJSON-944098855 tempest-ServersNegativeTestMultiTenantJSON-944098855-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 831.970754] env[65726]: DEBUG nova.compute.manager [-] [instance: 018c8ac5-66c4-4a9a-ab45-85e84f92b4fe] Deallocating network for instance {{(pid=65726) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 831.970754] env[65726]: DEBUG nova.network.neutron [-] [instance: 018c8ac5-66c4-4a9a-ab45-85e84f92b4fe] deallocate_for_instance() {{(pid=65726) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 831.971027] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 831.972421] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 832.008123] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 832.008419] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 832.025352] env[65726]: WARNING openstack [req-1c17cde7-0a51-42e2-b50d-4c0a77648ecd req-bffb8806-15d0-4ea3-9ae8-800f5970e40f service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 832.025725] env[65726]: WARNING openstack [req-1c17cde7-0a51-42e2-b50d-4c0a77648ecd req-bffb8806-15d0-4ea3-9ae8-800f5970e40f service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 832.077808] env[65726]: DEBUG nova.scheduler.client.report [None req-ebfc0e41-4845-41ca-9f74-d2a2946626c0 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Updated inventory for provider 07c4692f-bdb4-4058-9173-ff9664830295 with generation 86 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 832.078128] env[65726]: DEBUG nova.compute.provider_tree [None req-ebfc0e41-4845-41ca-9f74-d2a2946626c0 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Updating resource provider 07c4692f-bdb4-4058-9173-ff9664830295 generation from 86 to 87 during operation: update_inventory {{(pid=65726) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 832.078311] env[65726]: DEBUG nova.compute.provider_tree [None req-ebfc0e41-4845-41ca-9f74-d2a2946626c0 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Updating inventory in ProviderTree for provider 07c4692f-bdb4-4058-9173-ff9664830295 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 832.099592] env[65726]: WARNING openstack [req-1c17cde7-0a51-42e2-b50d-4c0a77648ecd req-bffb8806-15d0-4ea3-9ae8-800f5970e40f service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 832.099972] env[65726]: WARNING openstack [req-1c17cde7-0a51-42e2-b50d-4c0a77648ecd req-bffb8806-15d0-4ea3-9ae8-800f5970e40f service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 832.115284] env[65726]: DEBUG oslo_vmware.api [None req-645c8c71-cd4d-4583-8270-b88bb6120d10 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': task-5115991, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.072877} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.115399] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-645c8c71-cd4d-4583-8270-b88bb6120d10 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 01d86089-6b9b-4588-864e-ef91375a1eea] Extended root virtual disk {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 832.116727] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9cb5f6b5-c4ee-44c0-9049-de5e6f0490c0 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.141905] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-645c8c71-cd4d-4583-8270-b88bb6120d10 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 01d86089-6b9b-4588-864e-ef91375a1eea] Reconfiguring VM instance instance-00000036 to attach disk [datastore1] 01d86089-6b9b-4588-864e-ef91375a1eea/01d86089-6b9b-4588-864e-ef91375a1eea.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 832.142326] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2e2850bc-a837-4255-8fb3-e07cc7b1a19c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.171431] env[65726]: DEBUG oslo_vmware.api [None req-645c8c71-cd4d-4583-8270-b88bb6120d10 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Waiting for the task: (returnval){ [ 832.171431] env[65726]: value = "task-5115993" [ 832.171431] env[65726]: _type = "Task" [ 832.171431] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.183856] env[65726]: DEBUG oslo_vmware.api [None req-645c8c71-cd4d-4583-8270-b88bb6120d10 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': task-5115993, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.273942] env[65726]: DEBUG nova.network.neutron [req-1c17cde7-0a51-42e2-b50d-4c0a77648ecd req-bffb8806-15d0-4ea3-9ae8-800f5970e40f service nova] [instance: c8be64a3-1bfa-41a3-b9be-68a7e6e052e0] Updated VIF entry in instance network info cache for port e9d6f425-5466-4d5d-ae7f-80a433defe57. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 832.274832] env[65726]: DEBUG nova.network.neutron [req-1c17cde7-0a51-42e2-b50d-4c0a77648ecd req-bffb8806-15d0-4ea3-9ae8-800f5970e40f service nova] [instance: c8be64a3-1bfa-41a3-b9be-68a7e6e052e0] Updating instance_info_cache with network_info: [{"id": "e9d6f425-5466-4d5d-ae7f-80a433defe57", "address": "fa:16:3e:7d:00:be", "network": {"id": "00391368-0b6f-45a0-9510-7b6ac2ab47ca", "bridge": "br-int", "label": "tempest-ServersV294TestFqdnHostnames-611600830-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.156", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b8855e0d30be40b49e32a3eb0b6eef30", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d6e940e5-e083-4238-973e-f1b4e2a3a5c7", "external-id": "nsx-vlan-transportzone-64", "segmentation_id": 64, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape9d6f425-54", "ovs_interfaceid": "e9d6f425-5466-4d5d-ae7f-80a433defe57", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 832.296633] env[65726]: DEBUG oslo_vmware.api [None req-76673fd1-cee8-44a1-9e95-0d5b0dfe1651 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] Task: {'id': task-5115992, 'name': PowerOffVM_Task, 'duration_secs': 0.481955} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.296633] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-76673fd1-cee8-44a1-9e95-0d5b0dfe1651 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] [instance: 3cbbe146-be3a-43d5-867d-4d669884758c] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 832.296633] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-743c7d7e-d259-4ee6-93b0-68c49957808c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.319425] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd064420-3f1b-4439-b738-51ec2c795c78 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.369107] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-76673fd1-cee8-44a1-9e95-0d5b0dfe1651 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] [instance: 3cbbe146-be3a-43d5-867d-4d669884758c] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 832.369405] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4ec4f07b-6d6a-4688-b874-1d10fda56311 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.377711] env[65726]: DEBUG oslo_vmware.api [None req-76673fd1-cee8-44a1-9e95-0d5b0dfe1651 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] Waiting for the task: (returnval){ [ 832.377711] env[65726]: value = "task-5115994" [ 832.377711] env[65726]: _type = "Task" [ 832.377711] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.378450] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b20cc073-fafc-44fa-bcf9-9893dc98ebec tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 832.388303] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-76673fd1-cee8-44a1-9e95-0d5b0dfe1651 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] [instance: 3cbbe146-be3a-43d5-867d-4d669884758c] VM already powered off {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 832.388486] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-76673fd1-cee8-44a1-9e95-0d5b0dfe1651 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] [instance: 3cbbe146-be3a-43d5-867d-4d669884758c] Processing image b52362a3-ee8a-4cbf-b06f-513b0cc8f95c {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 832.388758] env[65726]: DEBUG oslo_concurrency.lockutils [None req-76673fd1-cee8-44a1-9e95-0d5b0dfe1651 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 832.388904] env[65726]: DEBUG oslo_concurrency.lockutils [None req-76673fd1-cee8-44a1-9e95-0d5b0dfe1651 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 832.389993] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-76673fd1-cee8-44a1-9e95-0d5b0dfe1651 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 832.391030] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-34f2bf54-ef3a-4b87-a2a0-3305b3d54834 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 608cc0ea-3f6b-4b4f-83c1-01aa50999d63] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 832.391430] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d3761034-0d78-4030-b153-618b9f754459 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.395409] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5ee06040-7e93-4b94-9124-02cf07076b81 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.406175] env[65726]: DEBUG oslo_vmware.api [None req-34f2bf54-ef3a-4b87-a2a0-3305b3d54834 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Waiting for the task: (returnval){ [ 832.406175] env[65726]: value = "task-5115995" [ 832.406175] env[65726]: _type = "Task" [ 832.406175] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.412066] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-76673fd1-cee8-44a1-9e95-0d5b0dfe1651 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 832.412278] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-76673fd1-cee8-44a1-9e95-0d5b0dfe1651 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65726) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 832.413231] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-93cc634a-e881-4477-b6f2-050a7dfc9071 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.424375] env[65726]: DEBUG oslo_vmware.api [None req-34f2bf54-ef3a-4b87-a2a0-3305b3d54834 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5115995, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.424708] env[65726]: DEBUG oslo_vmware.api [None req-76673fd1-cee8-44a1-9e95-0d5b0dfe1651 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] Waiting for the task: (returnval){ [ 832.424708] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]520e1d4d-8f25-4786-fbd7-287fc08a52d2" [ 832.424708] env[65726]: _type = "Task" [ 832.424708] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.434410] env[65726]: DEBUG oslo_vmware.api [None req-76673fd1-cee8-44a1-9e95-0d5b0dfe1651 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]520e1d4d-8f25-4786-fbd7-287fc08a52d2, 'name': SearchDatastore_Task, 'duration_secs': 0.01143} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.435386] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c5d44d7c-8f59-4910-992a-617615a2f491 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.442482] env[65726]: DEBUG oslo_vmware.api [None req-76673fd1-cee8-44a1-9e95-0d5b0dfe1651 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] Waiting for the task: (returnval){ [ 832.442482] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52b93c47-b50c-c4a7-ea66-57cd7e83815e" [ 832.442482] env[65726]: _type = "Task" [ 832.442482] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.453751] env[65726]: DEBUG oslo_vmware.api [None req-76673fd1-cee8-44a1-9e95-0d5b0dfe1651 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52b93c47-b50c-c4a7-ea66-57cd7e83815e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.585206] env[65726]: DEBUG oslo_concurrency.lockutils [None req-ebfc0e41-4845-41ca-9f74-d2a2946626c0 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 5.341s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 832.585415] env[65726]: DEBUG nova.compute.manager [None req-ebfc0e41-4845-41ca-9f74-d2a2946626c0 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] [instance: 1cfb1bba-1e21-47eb-b50f-a86575846a65] Start building networks asynchronously for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 832.591579] env[65726]: DEBUG oslo_concurrency.lockutils [None req-447a3897-f4ca-48fd-a799-fe155c3fcf63 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 42.941s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 832.591579] env[65726]: DEBUG nova.objects.instance [None req-447a3897-f4ca-48fd-a799-fe155c3fcf63 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Lazy-loading 'resources' on Instance uuid 65aaa2eb-3ca5-4c2b-92f9-b524366c10dd {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 832.683648] env[65726]: DEBUG oslo_vmware.api [None req-645c8c71-cd4d-4583-8270-b88bb6120d10 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': task-5115993, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.778883] env[65726]: DEBUG oslo_concurrency.lockutils [req-1c17cde7-0a51-42e2-b50d-4c0a77648ecd req-bffb8806-15d0-4ea3-9ae8-800f5970e40f service nova] Releasing lock "refresh_cache-c8be64a3-1bfa-41a3-b9be-68a7e6e052e0" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 832.789859] env[65726]: DEBUG nova.network.neutron [-] [instance: 018c8ac5-66c4-4a9a-ab45-85e84f92b4fe] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 832.917589] env[65726]: DEBUG oslo_vmware.api [None req-34f2bf54-ef3a-4b87-a2a0-3305b3d54834 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5115995, 'name': PowerOffVM_Task, 'duration_secs': 0.21467} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.917869] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-34f2bf54-ef3a-4b87-a2a0-3305b3d54834 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 608cc0ea-3f6b-4b4f-83c1-01aa50999d63] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 832.918084] env[65726]: DEBUG nova.compute.manager [None req-34f2bf54-ef3a-4b87-a2a0-3305b3d54834 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 608cc0ea-3f6b-4b4f-83c1-01aa50999d63] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 832.918880] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d28b39ec-3c02-4ef2-b7e7-23be6825f6fb {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.960276] env[65726]: DEBUG oslo_vmware.api [None req-76673fd1-cee8-44a1-9e95-0d5b0dfe1651 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52b93c47-b50c-c4a7-ea66-57cd7e83815e, 'name': SearchDatastore_Task, 'duration_secs': 0.01288} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.960582] env[65726]: DEBUG oslo_concurrency.lockutils [None req-76673fd1-cee8-44a1-9e95-0d5b0dfe1651 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 832.960836] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-76673fd1-cee8-44a1-9e95-0d5b0dfe1651 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] Copying virtual disk from [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] 3cbbe146-be3a-43d5-867d-4d669884758c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c-rescue.vmdk. {{(pid=65726) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 832.961139] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5e08274b-e8a8-4b4a-8096-4082cc9f7969 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.971150] env[65726]: DEBUG oslo_vmware.api [None req-76673fd1-cee8-44a1-9e95-0d5b0dfe1651 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] Waiting for the task: (returnval){ [ 832.971150] env[65726]: value = "task-5115996" [ 832.971150] env[65726]: _type = "Task" [ 832.971150] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.982561] env[65726]: DEBUG oslo_vmware.api [None req-76673fd1-cee8-44a1-9e95-0d5b0dfe1651 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] Task: {'id': task-5115996, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.092413] env[65726]: DEBUG nova.compute.utils [None req-ebfc0e41-4845-41ca-9f74-d2a2946626c0 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Using /dev/sd instead of None {{(pid=65726) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 833.093980] env[65726]: DEBUG nova.compute.manager [None req-ebfc0e41-4845-41ca-9f74-d2a2946626c0 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] [instance: 1cfb1bba-1e21-47eb-b50f-a86575846a65] Not allocating networking since 'none' was specified. {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2015}} [ 833.187775] env[65726]: DEBUG oslo_vmware.api [None req-645c8c71-cd4d-4583-8270-b88bb6120d10 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': task-5115993, 'name': ReconfigVM_Task, 'duration_secs': 0.556903} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.188154] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-645c8c71-cd4d-4583-8270-b88bb6120d10 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 01d86089-6b9b-4588-864e-ef91375a1eea] Reconfigured VM instance instance-00000036 to attach disk [datastore1] 01d86089-6b9b-4588-864e-ef91375a1eea/01d86089-6b9b-4588-864e-ef91375a1eea.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 833.188878] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e3ba2308-389a-41af-a08e-63bf0fbcdad3 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.202249] env[65726]: DEBUG oslo_vmware.api [None req-645c8c71-cd4d-4583-8270-b88bb6120d10 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Waiting for the task: (returnval){ [ 833.202249] env[65726]: value = "task-5115997" [ 833.202249] env[65726]: _type = "Task" [ 833.202249] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.219800] env[65726]: DEBUG oslo_vmware.api [None req-645c8c71-cd4d-4583-8270-b88bb6120d10 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': task-5115997, 'name': Rename_Task} progress is 6%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.294503] env[65726]: INFO nova.compute.manager [-] [instance: 018c8ac5-66c4-4a9a-ab45-85e84f92b4fe] Took 1.32 seconds to deallocate network for instance. [ 833.435796] env[65726]: DEBUG oslo_concurrency.lockutils [None req-34f2bf54-ef3a-4b87-a2a0-3305b3d54834 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Lock "608cc0ea-3f6b-4b4f-83c1-01aa50999d63" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.068s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 833.491670] env[65726]: DEBUG oslo_vmware.api [None req-76673fd1-cee8-44a1-9e95-0d5b0dfe1651 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] Task: {'id': task-5115996, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.600789] env[65726]: DEBUG nova.compute.manager [None req-ebfc0e41-4845-41ca-9f74-d2a2946626c0 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] [instance: 1cfb1bba-1e21-47eb-b50f-a86575846a65] Start building block device mappings for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 833.672067] env[65726]: DEBUG nova.compute.manager [req-fe2d1509-347e-4ce1-b77e-dfc8a507927e req-37a68e6f-33fe-40c5-adb8-f4ce7e7d8856 service nova] [instance: 018c8ac5-66c4-4a9a-ab45-85e84f92b4fe] Received event network-vif-deleted-5a241c1a-c51d-49fc-8387-7dd3cf7f38af {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 833.714315] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed9f9d67-c321-447d-a5c4-e2014e75622f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.720796] env[65726]: DEBUG oslo_vmware.api [None req-645c8c71-cd4d-4583-8270-b88bb6120d10 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': task-5115997, 'name': Rename_Task, 'duration_secs': 0.254173} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.722052] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-645c8c71-cd4d-4583-8270-b88bb6120d10 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 01d86089-6b9b-4588-864e-ef91375a1eea] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 833.722052] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e1ebc8bd-d33b-44d9-9bc6-799306e298d7 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.728543] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6a73587-4244-4d09-b953-b8cb542e5c99 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.734575] env[65726]: DEBUG oslo_vmware.api [None req-645c8c71-cd4d-4583-8270-b88bb6120d10 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Waiting for the task: (returnval){ [ 833.734575] env[65726]: value = "task-5115998" [ 833.734575] env[65726]: _type = "Task" [ 833.734575] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.767584] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4efd8570-fa61-4688-9de7-d0c4244db867 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.775216] env[65726]: DEBUG oslo_vmware.api [None req-645c8c71-cd4d-4583-8270-b88bb6120d10 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': task-5115998, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.782756] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea4fdc9c-36e8-4dbe-afcb-935a5d7dcf0d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.800084] env[65726]: DEBUG nova.compute.provider_tree [None req-447a3897-f4ca-48fd-a799-fe155c3fcf63 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 833.806349] env[65726]: DEBUG oslo_concurrency.lockutils [None req-398d7f82-4e73-43c8-8dc0-2fe897bf965f tempest-ServersNegativeTestMultiTenantJSON-944098855 tempest-ServersNegativeTestMultiTenantJSON-944098855-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 833.983775] env[65726]: DEBUG oslo_vmware.api [None req-76673fd1-cee8-44a1-9e95-0d5b0dfe1651 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] Task: {'id': task-5115996, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.572708} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.983984] env[65726]: INFO nova.virt.vmwareapi.ds_util [None req-76673fd1-cee8-44a1-9e95-0d5b0dfe1651 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] Copied virtual disk from [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] 3cbbe146-be3a-43d5-867d-4d669884758c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c-rescue.vmdk. [ 833.984906] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a21d1a51-08fb-41ac-86de-3908e5a9e159 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.015258] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-76673fd1-cee8-44a1-9e95-0d5b0dfe1651 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] [instance: 3cbbe146-be3a-43d5-867d-4d669884758c] Reconfiguring VM instance instance-00000034 to attach disk [datastore1] 3cbbe146-be3a-43d5-867d-4d669884758c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c-rescue.vmdk or device None with type thin {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 834.015756] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-534fbca6-9e52-4446-9c28-f7155caadf61 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.039135] env[65726]: DEBUG oslo_vmware.api [None req-76673fd1-cee8-44a1-9e95-0d5b0dfe1651 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] Waiting for the task: (returnval){ [ 834.039135] env[65726]: value = "task-5115999" [ 834.039135] env[65726]: _type = "Task" [ 834.039135] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.053868] env[65726]: DEBUG oslo_vmware.api [None req-76673fd1-cee8-44a1-9e95-0d5b0dfe1651 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] Task: {'id': task-5115999, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.248742] env[65726]: DEBUG oslo_vmware.api [None req-645c8c71-cd4d-4583-8270-b88bb6120d10 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': task-5115998, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.253764] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a1a4c202-18b9-4cfc-8067-d677c8415fff tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Acquiring lock "815ee20b-cb07-4514-bd43-1b74972508a5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 834.253990] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a1a4c202-18b9-4cfc-8067-d677c8415fff tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Lock "815ee20b-cb07-4514-bd43-1b74972508a5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 834.303370] env[65726]: DEBUG nova.scheduler.client.report [None req-447a3897-f4ca-48fd-a799-fe155c3fcf63 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 834.551689] env[65726]: DEBUG oslo_vmware.api [None req-76673fd1-cee8-44a1-9e95-0d5b0dfe1651 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] Task: {'id': task-5115999, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.613362] env[65726]: DEBUG nova.compute.manager [None req-ebfc0e41-4845-41ca-9f74-d2a2946626c0 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] [instance: 1cfb1bba-1e21-47eb-b50f-a86575846a65] Start spawning the instance on the hypervisor. {{(pid=65726) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 834.650638] env[65726]: DEBUG nova.virt.hardware [None req-ebfc0e41-4845-41ca-9f74-d2a2946626c0 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T19:28:34Z,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1539f06fa3534e90acbc3a60c4b8bd3f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T19:28:35Z,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 834.650920] env[65726]: DEBUG nova.virt.hardware [None req-ebfc0e41-4845-41ca-9f74-d2a2946626c0 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 834.651088] env[65726]: DEBUG nova.virt.hardware [None req-ebfc0e41-4845-41ca-9f74-d2a2946626c0 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 834.651275] env[65726]: DEBUG nova.virt.hardware [None req-ebfc0e41-4845-41ca-9f74-d2a2946626c0 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 834.651412] env[65726]: DEBUG nova.virt.hardware [None req-ebfc0e41-4845-41ca-9f74-d2a2946626c0 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 834.651552] env[65726]: DEBUG nova.virt.hardware [None req-ebfc0e41-4845-41ca-9f74-d2a2946626c0 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 834.651833] env[65726]: DEBUG nova.virt.hardware [None req-ebfc0e41-4845-41ca-9f74-d2a2946626c0 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 834.652056] env[65726]: DEBUG nova.virt.hardware [None req-ebfc0e41-4845-41ca-9f74-d2a2946626c0 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 834.652163] env[65726]: DEBUG nova.virt.hardware [None req-ebfc0e41-4845-41ca-9f74-d2a2946626c0 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 834.652325] env[65726]: DEBUG nova.virt.hardware [None req-ebfc0e41-4845-41ca-9f74-d2a2946626c0 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 834.652491] env[65726]: DEBUG nova.virt.hardware [None req-ebfc0e41-4845-41ca-9f74-d2a2946626c0 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 834.653521] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-421cb5c7-9e06-47b4-bc0c-e657911b08ba {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.663543] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9b83804-0e84-4a6f-bf6a-6536ba8890f6 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.683680] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-ebfc0e41-4845-41ca-9f74-d2a2946626c0 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] [instance: 1cfb1bba-1e21-47eb-b50f-a86575846a65] Instance VIF info [] {{(pid=65726) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 834.691103] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-ebfc0e41-4845-41ca-9f74-d2a2946626c0 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Creating folder: Project (e6d4f419846a4ed785988d5de5a0aebb). Parent ref: group-v995008. {{(pid=65726) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 834.691103] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b09e7263-9a70-4287-b6e0-22e17ff37672 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.706671] env[65726]: INFO nova.virt.vmwareapi.vm_util [None req-ebfc0e41-4845-41ca-9f74-d2a2946626c0 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Created folder: Project (e6d4f419846a4ed785988d5de5a0aebb) in parent group-v995008. [ 834.707198] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-ebfc0e41-4845-41ca-9f74-d2a2946626c0 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Creating folder: Instances. Parent ref: group-v995158. {{(pid=65726) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 834.707198] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-2fb157e6-b800-4dc7-a97a-8cac4d9db51a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.721598] env[65726]: INFO nova.virt.vmwareapi.vm_util [None req-ebfc0e41-4845-41ca-9f74-d2a2946626c0 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Created folder: Instances in parent group-v995158. [ 834.721897] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-ebfc0e41-4845-41ca-9f74-d2a2946626c0 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 834.722151] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1cfb1bba-1e21-47eb-b50f-a86575846a65] Creating VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 834.722380] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ae85d2a2-9bca-4b53-9dc5-4e53c5bef9d5 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.746538] env[65726]: DEBUG oslo_vmware.api [None req-645c8c71-cd4d-4583-8270-b88bb6120d10 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': task-5115998, 'name': PowerOnVM_Task, 'duration_secs': 0.994617} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.748339] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-645c8c71-cd4d-4583-8270-b88bb6120d10 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 01d86089-6b9b-4588-864e-ef91375a1eea] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 834.748496] env[65726]: INFO nova.compute.manager [None req-645c8c71-cd4d-4583-8270-b88bb6120d10 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 01d86089-6b9b-4588-864e-ef91375a1eea] Took 7.83 seconds to spawn the instance on the hypervisor. [ 834.748691] env[65726]: DEBUG nova.compute.manager [None req-645c8c71-cd4d-4583-8270-b88bb6120d10 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 01d86089-6b9b-4588-864e-ef91375a1eea] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 834.749042] env[65726]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 834.749042] env[65726]: value = "task-5116002" [ 834.749042] env[65726]: _type = "Task" [ 834.749042] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.749857] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1224663-cb45-4ce6-9733-14b23c8bdecb {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.771475] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116002, 'name': CreateVM_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.811444] env[65726]: DEBUG oslo_concurrency.lockutils [None req-447a3897-f4ca-48fd-a799-fe155c3fcf63 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.219s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 834.812682] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3260fa2a-3a51-48a8-9aea-39966d4e575e tempest-ServersAaction247Test-712120018 tempest-ServersAaction247Test-712120018-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 42.545s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 834.813016] env[65726]: DEBUG nova.objects.instance [None req-3260fa2a-3a51-48a8-9aea-39966d4e575e tempest-ServersAaction247Test-712120018 tempest-ServersAaction247Test-712120018-project-member] Lazy-loading 'resources' on Instance uuid ad48cadd-9bb6-4191-a68a-5c8fdaeebd44 {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 834.862393] env[65726]: INFO nova.scheduler.client.report [None req-447a3897-f4ca-48fd-a799-fe155c3fcf63 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Deleted allocations for instance 65aaa2eb-3ca5-4c2b-92f9-b524366c10dd [ 835.053059] env[65726]: DEBUG oslo_vmware.api [None req-76673fd1-cee8-44a1-9e95-0d5b0dfe1651 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] Task: {'id': task-5115999, 'name': ReconfigVM_Task, 'duration_secs': 0.725476} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.053468] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-76673fd1-cee8-44a1-9e95-0d5b0dfe1651 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] [instance: 3cbbe146-be3a-43d5-867d-4d669884758c] Reconfigured VM instance instance-00000034 to attach disk [datastore1] 3cbbe146-be3a-43d5-867d-4d669884758c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c-rescue.vmdk or device None with type thin {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 835.054581] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7212f55b-38d0-4fd8-8da8-6c3518e9949a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.092617] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1b08bbd7-34c7-498b-83cf-013e609187cc {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.118941] env[65726]: DEBUG oslo_vmware.api [None req-76673fd1-cee8-44a1-9e95-0d5b0dfe1651 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] Waiting for the task: (returnval){ [ 835.118941] env[65726]: value = "task-5116003" [ 835.118941] env[65726]: _type = "Task" [ 835.118941] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.129801] env[65726]: DEBUG oslo_vmware.api [None req-76673fd1-cee8-44a1-9e95-0d5b0dfe1651 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] Task: {'id': task-5116003, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.265798] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116002, 'name': CreateVM_Task, 'duration_secs': 0.491271} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.266075] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1cfb1bba-1e21-47eb-b50f-a86575846a65] Created VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 835.266535] env[65726]: DEBUG oslo_concurrency.lockutils [None req-ebfc0e41-4845-41ca-9f74-d2a2946626c0 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 835.266708] env[65726]: DEBUG oslo_concurrency.lockutils [None req-ebfc0e41-4845-41ca-9f74-d2a2946626c0 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 835.267161] env[65726]: DEBUG oslo_concurrency.lockutils [None req-ebfc0e41-4845-41ca-9f74-d2a2946626c0 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 835.267415] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7f94ef7c-64ef-45e3-a293-3f6a93116a4b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.276799] env[65726]: DEBUG oslo_vmware.api [None req-ebfc0e41-4845-41ca-9f74-d2a2946626c0 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Waiting for the task: (returnval){ [ 835.276799] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]526f5398-da15-1054-2114-4f4b090285ad" [ 835.276799] env[65726]: _type = "Task" [ 835.276799] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.283020] env[65726]: INFO nova.compute.manager [None req-645c8c71-cd4d-4583-8270-b88bb6120d10 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 01d86089-6b9b-4588-864e-ef91375a1eea] Took 56.67 seconds to build instance. [ 835.289766] env[65726]: DEBUG oslo_vmware.api [None req-ebfc0e41-4845-41ca-9f74-d2a2946626c0 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]526f5398-da15-1054-2114-4f4b090285ad, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.384167] env[65726]: DEBUG oslo_concurrency.lockutils [None req-447a3897-f4ca-48fd-a799-fe155c3fcf63 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Lock "65aaa2eb-3ca5-4c2b-92f9-b524366c10dd" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 49.171s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 835.385342] env[65726]: DEBUG oslo_concurrency.lockutils [None req-4826faf0-9908-4f3c-a509-20fe2a708723 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Acquiring lock "29823bc7-3909-4ab7-8119-91ee59b289c5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 835.387076] env[65726]: DEBUG oslo_concurrency.lockutils [None req-4826faf0-9908-4f3c-a509-20fe2a708723 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Lock "29823bc7-3909-4ab7-8119-91ee59b289c5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 835.387076] env[65726]: DEBUG oslo_concurrency.lockutils [None req-6397aebb-7e67-4113-8134-9f602fab9ef7 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Lock "65aaa2eb-3ca5-4c2b-92f9-b524366c10dd" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 44.429s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 835.387076] env[65726]: DEBUG oslo_concurrency.lockutils [None req-6397aebb-7e67-4113-8134-9f602fab9ef7 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Acquiring lock "65aaa2eb-3ca5-4c2b-92f9-b524366c10dd-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 835.387076] env[65726]: DEBUG oslo_concurrency.lockutils [None req-6397aebb-7e67-4113-8134-9f602fab9ef7 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Lock "65aaa2eb-3ca5-4c2b-92f9-b524366c10dd-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 835.387076] env[65726]: DEBUG oslo_concurrency.lockutils [None req-6397aebb-7e67-4113-8134-9f602fab9ef7 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Lock "65aaa2eb-3ca5-4c2b-92f9-b524366c10dd-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.001s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 835.388646] env[65726]: INFO nova.compute.manager [None req-6397aebb-7e67-4113-8134-9f602fab9ef7 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] [instance: 65aaa2eb-3ca5-4c2b-92f9-b524366c10dd] Terminating instance [ 835.617138] env[65726]: DEBUG nova.compute.manager [None req-d6fa7de0-71fb-41d9-ba04-00cd57a3d822 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 608cc0ea-3f6b-4b4f-83c1-01aa50999d63] Stashing vm_state: stopped {{(pid=65726) _prep_resize /opt/stack/nova/nova/compute/manager.py:6193}} [ 835.634663] env[65726]: DEBUG oslo_vmware.api [None req-76673fd1-cee8-44a1-9e95-0d5b0dfe1651 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] Task: {'id': task-5116003, 'name': ReconfigVM_Task, 'duration_secs': 0.255014} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.634973] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-76673fd1-cee8-44a1-9e95-0d5b0dfe1651 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] [instance: 3cbbe146-be3a-43d5-867d-4d669884758c] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 835.635204] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-77743924-2e79-4e1f-8626-069b317773f9 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.645893] env[65726]: DEBUG oslo_vmware.api [None req-76673fd1-cee8-44a1-9e95-0d5b0dfe1651 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] Waiting for the task: (returnval){ [ 835.645893] env[65726]: value = "task-5116004" [ 835.645893] env[65726]: _type = "Task" [ 835.645893] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.655649] env[65726]: DEBUG oslo_vmware.api [None req-76673fd1-cee8-44a1-9e95-0d5b0dfe1651 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] Task: {'id': task-5116004, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.785670] env[65726]: DEBUG oslo_concurrency.lockutils [None req-645c8c71-cd4d-4583-8270-b88bb6120d10 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Lock "01d86089-6b9b-4588-864e-ef91375a1eea" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 77.253s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 835.790719] env[65726]: DEBUG oslo_vmware.api [None req-ebfc0e41-4845-41ca-9f74-d2a2946626c0 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]526f5398-da15-1054-2114-4f4b090285ad, 'name': SearchDatastore_Task, 'duration_secs': 0.018091} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.791269] env[65726]: DEBUG oslo_concurrency.lockutils [None req-ebfc0e41-4845-41ca-9f74-d2a2946626c0 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 835.791725] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-ebfc0e41-4845-41ca-9f74-d2a2946626c0 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] [instance: 1cfb1bba-1e21-47eb-b50f-a86575846a65] Processing image b52362a3-ee8a-4cbf-b06f-513b0cc8f95c {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 835.791977] env[65726]: DEBUG oslo_concurrency.lockutils [None req-ebfc0e41-4845-41ca-9f74-d2a2946626c0 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 835.792205] env[65726]: DEBUG oslo_concurrency.lockutils [None req-ebfc0e41-4845-41ca-9f74-d2a2946626c0 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 835.792709] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-ebfc0e41-4845-41ca-9f74-d2a2946626c0 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 835.792991] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b0309e97-d68d-4336-b9dd-6f2da9dba153 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.805050] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-ebfc0e41-4845-41ca-9f74-d2a2946626c0 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 835.805243] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-ebfc0e41-4845-41ca-9f74-d2a2946626c0 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65726) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 835.806198] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6bcdffbd-b0c3-4128-af3a-1bd4148dbb4d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.815911] env[65726]: DEBUG oslo_vmware.api [None req-ebfc0e41-4845-41ca-9f74-d2a2946626c0 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Waiting for the task: (returnval){ [ 835.815911] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52180271-a100-fa90-2bb2-4e5dd5534030" [ 835.815911] env[65726]: _type = "Task" [ 835.815911] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.825125] env[65726]: DEBUG oslo_vmware.api [None req-ebfc0e41-4845-41ca-9f74-d2a2946626c0 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52180271-a100-fa90-2bb2-4e5dd5534030, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.899435] env[65726]: DEBUG oslo_concurrency.lockutils [None req-6397aebb-7e67-4113-8134-9f602fab9ef7 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Acquiring lock "refresh_cache-65aaa2eb-3ca5-4c2b-92f9-b524366c10dd" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 835.899569] env[65726]: DEBUG oslo_concurrency.lockutils [None req-6397aebb-7e67-4113-8134-9f602fab9ef7 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Acquired lock "refresh_cache-65aaa2eb-3ca5-4c2b-92f9-b524366c10dd" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 835.899743] env[65726]: DEBUG nova.network.neutron [None req-6397aebb-7e67-4113-8134-9f602fab9ef7 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] [instance: 65aaa2eb-3ca5-4c2b-92f9-b524366c10dd] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 835.928239] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93354326-2649-401f-b2ef-d749703e6e0e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.937388] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-847a65c9-44bf-4268-932a-70cdc38ef0e7 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.969771] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e00bf0cf-f7c5-4c13-b261-b66ad9519e11 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.978239] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ada28f75-8102-4a3f-a9c0-30e4bd1aa2a3 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.996142] env[65726]: DEBUG nova.compute.provider_tree [None req-3260fa2a-3a51-48a8-9aea-39966d4e575e tempest-ServersAaction247Test-712120018 tempest-ServersAaction247Test-712120018-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 836.137814] env[65726]: DEBUG oslo_concurrency.lockutils [None req-d6fa7de0-71fb-41d9-ba04-00cd57a3d822 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 836.156842] env[65726]: DEBUG oslo_vmware.api [None req-76673fd1-cee8-44a1-9e95-0d5b0dfe1651 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] Task: {'id': task-5116004, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.292239] env[65726]: DEBUG nova.compute.manager [None req-34f1738e-744a-4a46-b434-5db18d6e964a tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] [instance: 4ea2b70b-0cfe-4fbf-817a-baa028b8372a] Starting instance... {{(pid=65726) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 836.330616] env[65726]: DEBUG oslo_vmware.api [None req-ebfc0e41-4845-41ca-9f74-d2a2946626c0 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52180271-a100-fa90-2bb2-4e5dd5534030, 'name': SearchDatastore_Task, 'duration_secs': 0.048791} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.331583] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-61bebf65-8541-48f1-b1be-b94209417467 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.338587] env[65726]: DEBUG oslo_vmware.api [None req-ebfc0e41-4845-41ca-9f74-d2a2946626c0 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Waiting for the task: (returnval){ [ 836.338587] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52b23801-a40d-e313-7211-23832dceeff4" [ 836.338587] env[65726]: _type = "Task" [ 836.338587] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.349555] env[65726]: DEBUG oslo_vmware.api [None req-ebfc0e41-4845-41ca-9f74-d2a2946626c0 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52b23801-a40d-e313-7211-23832dceeff4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.403834] env[65726]: DEBUG nova.compute.utils [None req-6397aebb-7e67-4113-8134-9f602fab9ef7 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] [instance: 65aaa2eb-3ca5-4c2b-92f9-b524366c10dd] Can not refresh info_cache because instance was not found {{(pid=65726) refresh_info_cache_for_instance /opt/stack/nova/nova/compute/utils.py:1056}} [ 836.404731] env[65726]: WARNING openstack [None req-6397aebb-7e67-4113-8134-9f602fab9ef7 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 836.405442] env[65726]: WARNING openstack [None req-6397aebb-7e67-4113-8134-9f602fab9ef7 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 836.413939] env[65726]: DEBUG nova.network.neutron [None req-6397aebb-7e67-4113-8134-9f602fab9ef7 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] [instance: 65aaa2eb-3ca5-4c2b-92f9-b524366c10dd] Instance cache missing network info. {{(pid=65726) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 836.499014] env[65726]: DEBUG nova.scheduler.client.report [None req-3260fa2a-3a51-48a8-9aea-39966d4e575e tempest-ServersAaction247Test-712120018 tempest-ServersAaction247Test-712120018-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 836.533492] env[65726]: DEBUG nova.network.neutron [None req-6397aebb-7e67-4113-8134-9f602fab9ef7 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] [instance: 65aaa2eb-3ca5-4c2b-92f9-b524366c10dd] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 836.657691] env[65726]: DEBUG oslo_vmware.api [None req-76673fd1-cee8-44a1-9e95-0d5b0dfe1651 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] Task: {'id': task-5116004, 'name': PowerOnVM_Task, 'duration_secs': 0.723951} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.658138] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-76673fd1-cee8-44a1-9e95-0d5b0dfe1651 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] [instance: 3cbbe146-be3a-43d5-867d-4d669884758c] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 836.661601] env[65726]: DEBUG nova.compute.manager [None req-76673fd1-cee8-44a1-9e95-0d5b0dfe1651 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] [instance: 3cbbe146-be3a-43d5-867d-4d669884758c] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 836.662443] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a513fd3-d2b7-4902-8e12-b3dc69d2bde4 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.714896] env[65726]: DEBUG oslo_concurrency.lockutils [None req-63a3f335-a9b3-4e24-b6da-ba230f1ce119 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Acquiring lock "d7905c4f-1ba3-43b3-b4b3-aebacd2f9cc3" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 836.715182] env[65726]: DEBUG oslo_concurrency.lockutils [None req-63a3f335-a9b3-4e24-b6da-ba230f1ce119 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Lock "d7905c4f-1ba3-43b3-b4b3-aebacd2f9cc3" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 836.715337] env[65726]: DEBUG oslo_concurrency.lockutils [None req-63a3f335-a9b3-4e24-b6da-ba230f1ce119 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Acquiring lock "d7905c4f-1ba3-43b3-b4b3-aebacd2f9cc3-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 836.715513] env[65726]: DEBUG oslo_concurrency.lockutils [None req-63a3f335-a9b3-4e24-b6da-ba230f1ce119 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Lock "d7905c4f-1ba3-43b3-b4b3-aebacd2f9cc3-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 836.715671] env[65726]: DEBUG oslo_concurrency.lockutils [None req-63a3f335-a9b3-4e24-b6da-ba230f1ce119 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Lock "d7905c4f-1ba3-43b3-b4b3-aebacd2f9cc3-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 836.718668] env[65726]: INFO nova.compute.manager [None req-63a3f335-a9b3-4e24-b6da-ba230f1ce119 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] [instance: d7905c4f-1ba3-43b3-b4b3-aebacd2f9cc3] Terminating instance [ 836.814792] env[65726]: DEBUG oslo_concurrency.lockutils [None req-34f1738e-744a-4a46-b434-5db18d6e964a tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 836.850572] env[65726]: DEBUG oslo_vmware.api [None req-ebfc0e41-4845-41ca-9f74-d2a2946626c0 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52b23801-a40d-e313-7211-23832dceeff4, 'name': SearchDatastore_Task, 'duration_secs': 0.012516} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.850983] env[65726]: DEBUG oslo_concurrency.lockutils [None req-ebfc0e41-4845-41ca-9f74-d2a2946626c0 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 836.851286] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-ebfc0e41-4845-41ca-9f74-d2a2946626c0 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore2] 1cfb1bba-1e21-47eb-b50f-a86575846a65/1cfb1bba-1e21-47eb-b50f-a86575846a65.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 836.851568] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d026ba68-4797-4150-88cf-3d2a946e94e6 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.860056] env[65726]: DEBUG oslo_vmware.api [None req-ebfc0e41-4845-41ca-9f74-d2a2946626c0 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Waiting for the task: (returnval){ [ 836.860056] env[65726]: value = "task-5116005" [ 836.860056] env[65726]: _type = "Task" [ 836.860056] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.872154] env[65726]: DEBUG oslo_vmware.api [None req-ebfc0e41-4845-41ca-9f74-d2a2946626c0 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Task: {'id': task-5116005, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.004712] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3260fa2a-3a51-48a8-9aea-39966d4e575e tempest-ServersAaction247Test-712120018 tempest-ServersAaction247Test-712120018-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.192s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 837.007171] env[65726]: DEBUG oslo_concurrency.lockutils [None req-d4c6362a-f5ab-42dd-9c7f-bfc3bee5f94c tempest-VolumesAssistedSnapshotsTest-1875841935 tempest-VolumesAssistedSnapshotsTest-1875841935-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 42.560s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 837.007433] env[65726]: DEBUG nova.objects.instance [None req-d4c6362a-f5ab-42dd-9c7f-bfc3bee5f94c tempest-VolumesAssistedSnapshotsTest-1875841935 tempest-VolumesAssistedSnapshotsTest-1875841935-project-member] Lazy-loading 'resources' on Instance uuid aa2716a9-2cf8-4e51-800d-1c8f98b2cf05 {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 837.030757] env[65726]: INFO nova.scheduler.client.report [None req-3260fa2a-3a51-48a8-9aea-39966d4e575e tempest-ServersAaction247Test-712120018 tempest-ServersAaction247Test-712120018-project-member] Deleted allocations for instance ad48cadd-9bb6-4191-a68a-5c8fdaeebd44 [ 837.037031] env[65726]: DEBUG oslo_concurrency.lockutils [None req-6397aebb-7e67-4113-8134-9f602fab9ef7 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Releasing lock "refresh_cache-65aaa2eb-3ca5-4c2b-92f9-b524366c10dd" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 837.037031] env[65726]: DEBUG nova.compute.manager [None req-6397aebb-7e67-4113-8134-9f602fab9ef7 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] [instance: 65aaa2eb-3ca5-4c2b-92f9-b524366c10dd] Start destroying the instance on the hypervisor. {{(pid=65726) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 837.037031] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-6397aebb-7e67-4113-8134-9f602fab9ef7 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] [instance: 65aaa2eb-3ca5-4c2b-92f9-b524366c10dd] Destroying instance {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 837.037277] env[65726]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-938b3031-c1d5-47a0-aa1c-dcbbe7b7c6d8 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.048337] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-352c1546-b63a-4f73-9622-c7720c3d6bdc {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.086060] env[65726]: WARNING nova.virt.vmwareapi.vmops [None req-6397aebb-7e67-4113-8134-9f602fab9ef7 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] [instance: 65aaa2eb-3ca5-4c2b-92f9-b524366c10dd] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 65aaa2eb-3ca5-4c2b-92f9-b524366c10dd could not be found. [ 837.087173] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-6397aebb-7e67-4113-8134-9f602fab9ef7 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] [instance: 65aaa2eb-3ca5-4c2b-92f9-b524366c10dd] Instance destroyed {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 837.087173] env[65726]: INFO nova.compute.manager [None req-6397aebb-7e67-4113-8134-9f602fab9ef7 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] [instance: 65aaa2eb-3ca5-4c2b-92f9-b524366c10dd] Took 0.05 seconds to destroy the instance on the hypervisor. [ 837.087173] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-6397aebb-7e67-4113-8134-9f602fab9ef7 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 837.087173] env[65726]: DEBUG nova.compute.manager [-] [instance: 65aaa2eb-3ca5-4c2b-92f9-b524366c10dd] Deallocating network for instance {{(pid=65726) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 837.087173] env[65726]: DEBUG nova.network.neutron [-] [instance: 65aaa2eb-3ca5-4c2b-92f9-b524366c10dd] deallocate_for_instance() {{(pid=65726) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 837.087826] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 837.087981] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 837.115020] env[65726]: DEBUG nova.network.neutron [-] [instance: 65aaa2eb-3ca5-4c2b-92f9-b524366c10dd] Instance cache missing network info. {{(pid=65726) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 837.115574] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 837.115876] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 837.224943] env[65726]: DEBUG nova.compute.manager [None req-63a3f335-a9b3-4e24-b6da-ba230f1ce119 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] [instance: d7905c4f-1ba3-43b3-b4b3-aebacd2f9cc3] Start destroying the instance on the hypervisor. {{(pid=65726) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 837.225283] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-63a3f335-a9b3-4e24-b6da-ba230f1ce119 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] [instance: d7905c4f-1ba3-43b3-b4b3-aebacd2f9cc3] Destroying instance {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 837.226619] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f656393a-9730-4ac3-a2dc-bd5ea8cab943 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.238112] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-63a3f335-a9b3-4e24-b6da-ba230f1ce119 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] [instance: d7905c4f-1ba3-43b3-b4b3-aebacd2f9cc3] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 837.238490] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fb14accc-b70f-46dd-bfc5-7f6293610581 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.247628] env[65726]: DEBUG oslo_vmware.api [None req-63a3f335-a9b3-4e24-b6da-ba230f1ce119 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Waiting for the task: (returnval){ [ 837.247628] env[65726]: value = "task-5116006" [ 837.247628] env[65726]: _type = "Task" [ 837.247628] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.258770] env[65726]: DEBUG oslo_vmware.api [None req-63a3f335-a9b3-4e24-b6da-ba230f1ce119 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Task: {'id': task-5116006, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.371224] env[65726]: DEBUG oslo_vmware.api [None req-ebfc0e41-4845-41ca-9f74-d2a2946626c0 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Task: {'id': task-5116005, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.543080] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3260fa2a-3a51-48a8-9aea-39966d4e575e tempest-ServersAaction247Test-712120018 tempest-ServersAaction247Test-712120018-project-member] Lock "ad48cadd-9bb6-4191-a68a-5c8fdaeebd44" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 49.757s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 837.623120] env[65726]: DEBUG nova.network.neutron [-] [instance: 65aaa2eb-3ca5-4c2b-92f9-b524366c10dd] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 837.728222] env[65726]: DEBUG nova.compute.manager [None req-6f5d9ea9-7d8d-4c8e-8d3c-edf0333fb67b tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 01d86089-6b9b-4588-864e-ef91375a1eea] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 837.729357] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14316cea-fcb5-41e0-aa03-10cda6cf786c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.758028] env[65726]: DEBUG oslo_vmware.api [None req-63a3f335-a9b3-4e24-b6da-ba230f1ce119 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Task: {'id': task-5116006, 'name': PowerOffVM_Task, 'duration_secs': 0.289919} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.758028] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-63a3f335-a9b3-4e24-b6da-ba230f1ce119 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] [instance: d7905c4f-1ba3-43b3-b4b3-aebacd2f9cc3] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 837.758028] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-63a3f335-a9b3-4e24-b6da-ba230f1ce119 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] [instance: d7905c4f-1ba3-43b3-b4b3-aebacd2f9cc3] Unregistering the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 837.758320] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f2a70efc-04e0-403c-917f-188940491313 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.835467] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-63a3f335-a9b3-4e24-b6da-ba230f1ce119 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] [instance: d7905c4f-1ba3-43b3-b4b3-aebacd2f9cc3] Unregistered the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 837.835467] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-63a3f335-a9b3-4e24-b6da-ba230f1ce119 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] [instance: d7905c4f-1ba3-43b3-b4b3-aebacd2f9cc3] Deleting contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 837.835467] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-63a3f335-a9b3-4e24-b6da-ba230f1ce119 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Deleting the datastore file [datastore1] d7905c4f-1ba3-43b3-b4b3-aebacd2f9cc3 {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 837.839662] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-466f98c4-7919-42a5-8514-1d2ecaf71b55 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.848018] env[65726]: DEBUG oslo_vmware.api [None req-63a3f335-a9b3-4e24-b6da-ba230f1ce119 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Waiting for the task: (returnval){ [ 837.848018] env[65726]: value = "task-5116008" [ 837.848018] env[65726]: _type = "Task" [ 837.848018] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.859298] env[65726]: DEBUG oslo_vmware.api [None req-63a3f335-a9b3-4e24-b6da-ba230f1ce119 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Task: {'id': task-5116008, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.871719] env[65726]: DEBUG oslo_vmware.api [None req-ebfc0e41-4845-41ca-9f74-d2a2946626c0 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Task: {'id': task-5116005, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.566729} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.874144] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-ebfc0e41-4845-41ca-9f74-d2a2946626c0 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore2] 1cfb1bba-1e21-47eb-b50f-a86575846a65/1cfb1bba-1e21-47eb-b50f-a86575846a65.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 837.874382] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-ebfc0e41-4845-41ca-9f74-d2a2946626c0 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] [instance: 1cfb1bba-1e21-47eb-b50f-a86575846a65] Extending root virtual disk to 1048576 {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 837.875574] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-569b68fe-10c5-4f57-b4fe-4b1b97902fac {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.883342] env[65726]: DEBUG oslo_vmware.api [None req-ebfc0e41-4845-41ca-9f74-d2a2946626c0 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Waiting for the task: (returnval){ [ 837.883342] env[65726]: value = "task-5116009" [ 837.883342] env[65726]: _type = "Task" [ 837.883342] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.894504] env[65726]: DEBUG oslo_vmware.api [None req-ebfc0e41-4845-41ca-9f74-d2a2946626c0 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Task: {'id': task-5116009, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.089508] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2af6fbe-a14b-4e96-b660-dcc41ce4ffa1 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.099883] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86e3ed84-de78-49df-ac62-b4010b406f31 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.134185] env[65726]: INFO nova.compute.manager [-] [instance: 65aaa2eb-3ca5-4c2b-92f9-b524366c10dd] Took 1.05 seconds to deallocate network for instance. [ 838.138450] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77f9d0b8-93d1-4f1b-89a5-40f28efe7dea {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.142490] env[65726]: DEBUG nova.compute.manager [req-ab004be9-5bba-423d-bf38-8a37b4dace1e req-7cf471e3-a807-4099-9e73-168486d39b38 service nova] [instance: 3cbbe146-be3a-43d5-867d-4d669884758c] Received event network-changed-40240c4e-9f3e-46aa-bd75-c62c2776d9b6 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 838.142672] env[65726]: DEBUG nova.compute.manager [req-ab004be9-5bba-423d-bf38-8a37b4dace1e req-7cf471e3-a807-4099-9e73-168486d39b38 service nova] [instance: 3cbbe146-be3a-43d5-867d-4d669884758c] Refreshing instance network info cache due to event network-changed-40240c4e-9f3e-46aa-bd75-c62c2776d9b6. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 838.142874] env[65726]: DEBUG oslo_concurrency.lockutils [req-ab004be9-5bba-423d-bf38-8a37b4dace1e req-7cf471e3-a807-4099-9e73-168486d39b38 service nova] Acquiring lock "refresh_cache-3cbbe146-be3a-43d5-867d-4d669884758c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 838.143014] env[65726]: DEBUG oslo_concurrency.lockutils [req-ab004be9-5bba-423d-bf38-8a37b4dace1e req-7cf471e3-a807-4099-9e73-168486d39b38 service nova] Acquired lock "refresh_cache-3cbbe146-be3a-43d5-867d-4d669884758c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 838.143216] env[65726]: DEBUG nova.network.neutron [req-ab004be9-5bba-423d-bf38-8a37b4dace1e req-7cf471e3-a807-4099-9e73-168486d39b38 service nova] [instance: 3cbbe146-be3a-43d5-867d-4d669884758c] Refreshing network info cache for port 40240c4e-9f3e-46aa-bd75-c62c2776d9b6 {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 838.158769] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2841f415-fa98-43a6-8438-5c40722bc9b3 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.183916] env[65726]: DEBUG nova.compute.provider_tree [None req-d4c6362a-f5ab-42dd-9c7f-bfc3bee5f94c tempest-VolumesAssistedSnapshotsTest-1875841935 tempest-VolumesAssistedSnapshotsTest-1875841935-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 838.252163] env[65726]: INFO nova.compute.manager [None req-6f5d9ea9-7d8d-4c8e-8d3c-edf0333fb67b tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 01d86089-6b9b-4588-864e-ef91375a1eea] instance snapshotting [ 838.255510] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e461f97-12e8-40fb-ad32-43a8f3b6e72a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.277280] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-faad03df-4220-42a1-98f5-1e1b16135a88 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.359864] env[65726]: DEBUG oslo_vmware.api [None req-63a3f335-a9b3-4e24-b6da-ba230f1ce119 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Task: {'id': task-5116008, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.168108} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.360207] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-63a3f335-a9b3-4e24-b6da-ba230f1ce119 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Deleted the datastore file {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 838.360496] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-63a3f335-a9b3-4e24-b6da-ba230f1ce119 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] [instance: d7905c4f-1ba3-43b3-b4b3-aebacd2f9cc3] Deleted contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 838.360651] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-63a3f335-a9b3-4e24-b6da-ba230f1ce119 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] [instance: d7905c4f-1ba3-43b3-b4b3-aebacd2f9cc3] Instance destroyed {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 838.360877] env[65726]: INFO nova.compute.manager [None req-63a3f335-a9b3-4e24-b6da-ba230f1ce119 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] [instance: d7905c4f-1ba3-43b3-b4b3-aebacd2f9cc3] Took 1.14 seconds to destroy the instance on the hypervisor. [ 838.361178] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-63a3f335-a9b3-4e24-b6da-ba230f1ce119 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 838.361800] env[65726]: DEBUG nova.compute.manager [-] [instance: d7905c4f-1ba3-43b3-b4b3-aebacd2f9cc3] Deallocating network for instance {{(pid=65726) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 838.361933] env[65726]: DEBUG nova.network.neutron [-] [instance: d7905c4f-1ba3-43b3-b4b3-aebacd2f9cc3] deallocate_for_instance() {{(pid=65726) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 838.362901] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 838.363345] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 838.399308] env[65726]: DEBUG oslo_vmware.api [None req-ebfc0e41-4845-41ca-9f74-d2a2946626c0 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Task: {'id': task-5116009, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.108612} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.399730] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-ebfc0e41-4845-41ca-9f74-d2a2946626c0 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] [instance: 1cfb1bba-1e21-47eb-b50f-a86575846a65] Extended root virtual disk {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 838.401376] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc14d33d-76ad-4c2e-8cb9-a2b280e0502c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.428490] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-ebfc0e41-4845-41ca-9f74-d2a2946626c0 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] [instance: 1cfb1bba-1e21-47eb-b50f-a86575846a65] Reconfiguring VM instance instance-00000037 to attach disk [datastore2] 1cfb1bba-1e21-47eb-b50f-a86575846a65/1cfb1bba-1e21-47eb-b50f-a86575846a65.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 838.432208] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 838.432632] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 838.443197] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ef886f9e-6e2e-4ba9-92b6-22bd0e0467c7 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.467839] env[65726]: DEBUG oslo_vmware.api [None req-ebfc0e41-4845-41ca-9f74-d2a2946626c0 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Waiting for the task: (returnval){ [ 838.467839] env[65726]: value = "task-5116010" [ 838.467839] env[65726]: _type = "Task" [ 838.467839] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.476270] env[65726]: DEBUG oslo_vmware.api [None req-ebfc0e41-4845-41ca-9f74-d2a2946626c0 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Task: {'id': task-5116010, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.648850] env[65726]: WARNING openstack [req-ab004be9-5bba-423d-bf38-8a37b4dace1e req-7cf471e3-a807-4099-9e73-168486d39b38 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 838.649273] env[65726]: WARNING openstack [req-ab004be9-5bba-423d-bf38-8a37b4dace1e req-7cf471e3-a807-4099-9e73-168486d39b38 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 838.657237] env[65726]: INFO nova.compute.manager [None req-6397aebb-7e67-4113-8134-9f602fab9ef7 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] [instance: 65aaa2eb-3ca5-4c2b-92f9-b524366c10dd] Instance disappeared during terminate [ 838.657513] env[65726]: DEBUG oslo_concurrency.lockutils [None req-6397aebb-7e67-4113-8134-9f602fab9ef7 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Lock "65aaa2eb-3ca5-4c2b-92f9-b524366c10dd" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 3.272s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 838.689230] env[65726]: DEBUG nova.scheduler.client.report [None req-d4c6362a-f5ab-42dd-9c7f-bfc3bee5f94c tempest-VolumesAssistedSnapshotsTest-1875841935 tempest-VolumesAssistedSnapshotsTest-1875841935-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 838.791354] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-6f5d9ea9-7d8d-4c8e-8d3c-edf0333fb67b tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 01d86089-6b9b-4588-864e-ef91375a1eea] Creating Snapshot of the VM instance {{(pid=65726) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 838.791711] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-c7a52a26-57fb-4b39-8629-16a4e18de6ed {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.806049] env[65726]: DEBUG oslo_vmware.api [None req-6f5d9ea9-7d8d-4c8e-8d3c-edf0333fb67b tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Waiting for the task: (returnval){ [ 838.806049] env[65726]: value = "task-5116011" [ 838.806049] env[65726]: _type = "Task" [ 838.806049] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.824815] env[65726]: DEBUG oslo_vmware.api [None req-6f5d9ea9-7d8d-4c8e-8d3c-edf0333fb67b tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': task-5116011, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.837951] env[65726]: WARNING openstack [req-ab004be9-5bba-423d-bf38-8a37b4dace1e req-7cf471e3-a807-4099-9e73-168486d39b38 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 838.838513] env[65726]: WARNING openstack [req-ab004be9-5bba-423d-bf38-8a37b4dace1e req-7cf471e3-a807-4099-9e73-168486d39b38 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 838.903172] env[65726]: WARNING openstack [req-ab004be9-5bba-423d-bf38-8a37b4dace1e req-7cf471e3-a807-4099-9e73-168486d39b38 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 838.903172] env[65726]: WARNING openstack [req-ab004be9-5bba-423d-bf38-8a37b4dace1e req-7cf471e3-a807-4099-9e73-168486d39b38 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 838.978355] env[65726]: DEBUG oslo_vmware.api [None req-ebfc0e41-4845-41ca-9f74-d2a2946626c0 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Task: {'id': task-5116010, 'name': ReconfigVM_Task, 'duration_secs': 0.388485} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.978651] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-ebfc0e41-4845-41ca-9f74-d2a2946626c0 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] [instance: 1cfb1bba-1e21-47eb-b50f-a86575846a65] Reconfigured VM instance instance-00000037 to attach disk [datastore2] 1cfb1bba-1e21-47eb-b50f-a86575846a65/1cfb1bba-1e21-47eb-b50f-a86575846a65.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 838.979664] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0f30300f-128b-4e7a-ae6d-aa6e6d45c587 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.992081] env[65726]: DEBUG oslo_vmware.api [None req-ebfc0e41-4845-41ca-9f74-d2a2946626c0 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Waiting for the task: (returnval){ [ 838.992081] env[65726]: value = "task-5116012" [ 838.992081] env[65726]: _type = "Task" [ 838.992081] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.000970] env[65726]: DEBUG oslo_vmware.api [None req-ebfc0e41-4845-41ca-9f74-d2a2946626c0 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Task: {'id': task-5116012, 'name': Rename_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.011629] env[65726]: DEBUG nova.network.neutron [req-ab004be9-5bba-423d-bf38-8a37b4dace1e req-7cf471e3-a807-4099-9e73-168486d39b38 service nova] [instance: 3cbbe146-be3a-43d5-867d-4d669884758c] Updated VIF entry in instance network info cache for port 40240c4e-9f3e-46aa-bd75-c62c2776d9b6. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 839.011996] env[65726]: DEBUG nova.network.neutron [req-ab004be9-5bba-423d-bf38-8a37b4dace1e req-7cf471e3-a807-4099-9e73-168486d39b38 service nova] [instance: 3cbbe146-be3a-43d5-867d-4d669884758c] Updating instance_info_cache with network_info: [{"id": "40240c4e-9f3e-46aa-bd75-c62c2776d9b6", "address": "fa:16:3e:8f:bd:09", "network": {"id": "2266eb6e-4c8f-4a3f-85d3-0c06022c0d5e", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-716412627-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "92c1cedbc08545d58fc7172ac95e2c23", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "24376631-ee89-4ff1-b8ac-f09911fc8329", "external-id": "nsx-vlan-transportzone-960", "segmentation_id": 960, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap40240c4e-9f", "ovs_interfaceid": "40240c4e-9f3e-46aa-bd75-c62c2776d9b6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 839.195603] env[65726]: DEBUG oslo_concurrency.lockutils [None req-d4c6362a-f5ab-42dd-9c7f-bfc3bee5f94c tempest-VolumesAssistedSnapshotsTest-1875841935 tempest-VolumesAssistedSnapshotsTest-1875841935-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.188s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 839.197880] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 41.329s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 839.198162] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 839.198243] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=65726) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:937}} [ 839.198508] env[65726]: DEBUG oslo_concurrency.lockutils [None req-debf7b46-904e-41a8-8c0e-839a5a6575cb tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 39.800s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 839.198708] env[65726]: DEBUG nova.objects.instance [None req-debf7b46-904e-41a8-8c0e-839a5a6575cb tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Lazy-loading 'resources' on Instance uuid 703fff2d-5aec-4ed1-b987-9e1cc3a25f67 {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 839.201152] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b5b3930-f50a-48de-a96f-afdc892055a7 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.204228] env[65726]: DEBUG nova.network.neutron [-] [instance: d7905c4f-1ba3-43b3-b4b3-aebacd2f9cc3] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 839.211888] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-716602eb-222f-450e-98bf-0d77f4767fb5 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.228458] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4bc668e-9f08-4d3a-bd28-14d43f34f0d6 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.232242] env[65726]: INFO nova.scheduler.client.report [None req-d4c6362a-f5ab-42dd-9c7f-bfc3bee5f94c tempest-VolumesAssistedSnapshotsTest-1875841935 tempest-VolumesAssistedSnapshotsTest-1875841935-project-member] Deleted allocations for instance aa2716a9-2cf8-4e51-800d-1c8f98b2cf05 [ 839.242995] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-605ac20b-88cd-41f2-ae9d-50cd75a0f962 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.273613] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=178357MB free_disk=94GB free_vcpus=48 pci_devices=None {{(pid=65726) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1136}} [ 839.273613] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 839.318196] env[65726]: DEBUG oslo_vmware.api [None req-6f5d9ea9-7d8d-4c8e-8d3c-edf0333fb67b tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': task-5116011, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.501056] env[65726]: DEBUG oslo_vmware.api [None req-ebfc0e41-4845-41ca-9f74-d2a2946626c0 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Task: {'id': task-5116012, 'name': Rename_Task, 'duration_secs': 0.157381} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.501193] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-ebfc0e41-4845-41ca-9f74-d2a2946626c0 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] [instance: 1cfb1bba-1e21-47eb-b50f-a86575846a65] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 839.501368] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d9e1bd7a-7f91-4eb1-8f1a-9ec7fcb7ec32 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.510046] env[65726]: DEBUG oslo_vmware.api [None req-ebfc0e41-4845-41ca-9f74-d2a2946626c0 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Waiting for the task: (returnval){ [ 839.510046] env[65726]: value = "task-5116013" [ 839.510046] env[65726]: _type = "Task" [ 839.510046] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.515496] env[65726]: DEBUG oslo_concurrency.lockutils [req-ab004be9-5bba-423d-bf38-8a37b4dace1e req-7cf471e3-a807-4099-9e73-168486d39b38 service nova] Releasing lock "refresh_cache-3cbbe146-be3a-43d5-867d-4d669884758c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 839.519017] env[65726]: DEBUG oslo_vmware.api [None req-ebfc0e41-4845-41ca-9f74-d2a2946626c0 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Task: {'id': task-5116013, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.709100] env[65726]: INFO nova.compute.manager [-] [instance: d7905c4f-1ba3-43b3-b4b3-aebacd2f9cc3] Took 1.35 seconds to deallocate network for instance. [ 839.741803] env[65726]: DEBUG oslo_concurrency.lockutils [None req-d4c6362a-f5ab-42dd-9c7f-bfc3bee5f94c tempest-VolumesAssistedSnapshotsTest-1875841935 tempest-VolumesAssistedSnapshotsTest-1875841935-project-member] Lock "aa2716a9-2cf8-4e51-800d-1c8f98b2cf05" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 48.995s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 839.818098] env[65726]: DEBUG oslo_vmware.api [None req-6f5d9ea9-7d8d-4c8e-8d3c-edf0333fb67b tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': task-5116011, 'name': CreateSnapshot_Task, 'duration_secs': 0.973203} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.818504] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-6f5d9ea9-7d8d-4c8e-8d3c-edf0333fb67b tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 01d86089-6b9b-4588-864e-ef91375a1eea] Created Snapshot of the VM instance {{(pid=65726) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 839.819315] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13f3ed0b-df27-4441-9be1-71d5b2ec94fe {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.028025] env[65726]: DEBUG oslo_vmware.api [None req-ebfc0e41-4845-41ca-9f74-d2a2946626c0 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Task: {'id': task-5116013, 'name': PowerOnVM_Task, 'duration_secs': 0.494267} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.032245] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-ebfc0e41-4845-41ca-9f74-d2a2946626c0 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] [instance: 1cfb1bba-1e21-47eb-b50f-a86575846a65] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 840.032245] env[65726]: INFO nova.compute.manager [None req-ebfc0e41-4845-41ca-9f74-d2a2946626c0 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] [instance: 1cfb1bba-1e21-47eb-b50f-a86575846a65] Took 5.42 seconds to spawn the instance on the hypervisor. [ 840.032245] env[65726]: DEBUG nova.compute.manager [None req-ebfc0e41-4845-41ca-9f74-d2a2946626c0 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] [instance: 1cfb1bba-1e21-47eb-b50f-a86575846a65] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 840.033547] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d69bf0a4-233c-4338-aceb-661fb747c20f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.217540] env[65726]: DEBUG oslo_concurrency.lockutils [None req-63a3f335-a9b3-4e24-b6da-ba230f1ce119 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 840.221437] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba89b13a-08c5-42a1-8f97-3b46823d90d6 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.230893] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6db7daff-f9fa-4d43-b852-560a3a4527f7 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.265345] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-968e631b-18bb-43d2-8d3b-cc3669e7854f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.274959] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a8f1786-0ce7-4f87-b54a-3a235db2bb39 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.294972] env[65726]: DEBUG nova.compute.provider_tree [None req-debf7b46-904e-41a8-8c0e-839a5a6575cb tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 840.344234] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-6f5d9ea9-7d8d-4c8e-8d3c-edf0333fb67b tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 01d86089-6b9b-4588-864e-ef91375a1eea] Creating linked-clone VM from snapshot {{(pid=65726) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 840.346056] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-843c84d1-97b2-4d4f-acf8-e7a96e027f37 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.356267] env[65726]: DEBUG oslo_vmware.api [None req-6f5d9ea9-7d8d-4c8e-8d3c-edf0333fb67b tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Waiting for the task: (returnval){ [ 840.356267] env[65726]: value = "task-5116014" [ 840.356267] env[65726]: _type = "Task" [ 840.356267] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.366706] env[65726]: DEBUG oslo_vmware.api [None req-6f5d9ea9-7d8d-4c8e-8d3c-edf0333fb67b tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': task-5116014, 'name': CloneVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.554949] env[65726]: INFO nova.compute.manager [None req-ebfc0e41-4845-41ca-9f74-d2a2946626c0 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] [instance: 1cfb1bba-1e21-47eb-b50f-a86575846a65] Took 55.26 seconds to build instance. [ 840.564181] env[65726]: DEBUG nova.compute.manager [req-5f1e77d4-1355-4d7e-bc7e-aa206c506fc1 req-33247233-1236-4f18-bcba-0cf3bec20cdf service nova] [instance: d7905c4f-1ba3-43b3-b4b3-aebacd2f9cc3] Received event network-vif-deleted-ee24bb04-7df3-4d06-9751-289b4bdeb753 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 840.633844] env[65726]: DEBUG nova.compute.manager [req-9479b7c7-8bfa-4ecf-837c-3e91b8555cb2 req-efa5a0bf-092b-41fd-8fb5-1f013e9300a5 service nova] [instance: 3cbbe146-be3a-43d5-867d-4d669884758c] Received event network-changed-40240c4e-9f3e-46aa-bd75-c62c2776d9b6 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 840.633844] env[65726]: DEBUG nova.compute.manager [req-9479b7c7-8bfa-4ecf-837c-3e91b8555cb2 req-efa5a0bf-092b-41fd-8fb5-1f013e9300a5 service nova] [instance: 3cbbe146-be3a-43d5-867d-4d669884758c] Refreshing instance network info cache due to event network-changed-40240c4e-9f3e-46aa-bd75-c62c2776d9b6. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 840.633844] env[65726]: DEBUG oslo_concurrency.lockutils [req-9479b7c7-8bfa-4ecf-837c-3e91b8555cb2 req-efa5a0bf-092b-41fd-8fb5-1f013e9300a5 service nova] Acquiring lock "refresh_cache-3cbbe146-be3a-43d5-867d-4d669884758c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 840.634061] env[65726]: DEBUG oslo_concurrency.lockutils [req-9479b7c7-8bfa-4ecf-837c-3e91b8555cb2 req-efa5a0bf-092b-41fd-8fb5-1f013e9300a5 service nova] Acquired lock "refresh_cache-3cbbe146-be3a-43d5-867d-4d669884758c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 840.634155] env[65726]: DEBUG nova.network.neutron [req-9479b7c7-8bfa-4ecf-837c-3e91b8555cb2 req-efa5a0bf-092b-41fd-8fb5-1f013e9300a5 service nova] [instance: 3cbbe146-be3a-43d5-867d-4d669884758c] Refreshing network info cache for port 40240c4e-9f3e-46aa-bd75-c62c2776d9b6 {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 840.798554] env[65726]: DEBUG nova.scheduler.client.report [None req-debf7b46-904e-41a8-8c0e-839a5a6575cb tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 840.868443] env[65726]: DEBUG oslo_vmware.api [None req-6f5d9ea9-7d8d-4c8e-8d3c-edf0333fb67b tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': task-5116014, 'name': CloneVM_Task} progress is 94%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.060732] env[65726]: DEBUG oslo_concurrency.lockutils [None req-ebfc0e41-4845-41ca-9f74-d2a2946626c0 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Lock "1cfb1bba-1e21-47eb-b50f-a86575846a65" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 76.059s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 841.137673] env[65726]: WARNING openstack [req-9479b7c7-8bfa-4ecf-837c-3e91b8555cb2 req-efa5a0bf-092b-41fd-8fb5-1f013e9300a5 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 841.139029] env[65726]: WARNING openstack [req-9479b7c7-8bfa-4ecf-837c-3e91b8555cb2 req-efa5a0bf-092b-41fd-8fb5-1f013e9300a5 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 841.305061] env[65726]: DEBUG oslo_concurrency.lockutils [None req-debf7b46-904e-41a8-8c0e-839a5a6575cb tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.106s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 841.307513] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a6bb6238-1018-4e0c-84f4-13a2ac492d77 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 41.907s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 841.309829] env[65726]: INFO nova.compute.claims [None req-a6bb6238-1018-4e0c-84f4-13a2ac492d77 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] [instance: 44d73b2d-2b6a-4501-9944-432da3c9330d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 841.331470] env[65726]: INFO nova.scheduler.client.report [None req-debf7b46-904e-41a8-8c0e-839a5a6575cb tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Deleted allocations for instance 703fff2d-5aec-4ed1-b987-9e1cc3a25f67 [ 841.372870] env[65726]: DEBUG oslo_vmware.api [None req-6f5d9ea9-7d8d-4c8e-8d3c-edf0333fb67b tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': task-5116014, 'name': CloneVM_Task} progress is 94%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.560267] env[65726]: WARNING openstack [req-9479b7c7-8bfa-4ecf-837c-3e91b8555cb2 req-efa5a0bf-092b-41fd-8fb5-1f013e9300a5 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 841.560781] env[65726]: WARNING openstack [req-9479b7c7-8bfa-4ecf-837c-3e91b8555cb2 req-efa5a0bf-092b-41fd-8fb5-1f013e9300a5 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 841.568574] env[65726]: DEBUG nova.compute.manager [None req-a1a4c202-18b9-4cfc-8067-d677c8415fff tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] [instance: 815ee20b-cb07-4514-bd43-1b74972508a5] Starting instance... {{(pid=65726) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 841.699731] env[65726]: WARNING openstack [req-9479b7c7-8bfa-4ecf-837c-3e91b8555cb2 req-efa5a0bf-092b-41fd-8fb5-1f013e9300a5 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 841.699920] env[65726]: WARNING openstack [req-9479b7c7-8bfa-4ecf-837c-3e91b8555cb2 req-efa5a0bf-092b-41fd-8fb5-1f013e9300a5 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 841.834209] env[65726]: DEBUG nova.network.neutron [req-9479b7c7-8bfa-4ecf-837c-3e91b8555cb2 req-efa5a0bf-092b-41fd-8fb5-1f013e9300a5 service nova] [instance: 3cbbe146-be3a-43d5-867d-4d669884758c] Updated VIF entry in instance network info cache for port 40240c4e-9f3e-46aa-bd75-c62c2776d9b6. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 841.834586] env[65726]: DEBUG nova.network.neutron [req-9479b7c7-8bfa-4ecf-837c-3e91b8555cb2 req-efa5a0bf-092b-41fd-8fb5-1f013e9300a5 service nova] [instance: 3cbbe146-be3a-43d5-867d-4d669884758c] Updating instance_info_cache with network_info: [{"id": "40240c4e-9f3e-46aa-bd75-c62c2776d9b6", "address": "fa:16:3e:8f:bd:09", "network": {"id": "2266eb6e-4c8f-4a3f-85d3-0c06022c0d5e", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-716412627-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "92c1cedbc08545d58fc7172ac95e2c23", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "24376631-ee89-4ff1-b8ac-f09911fc8329", "external-id": "nsx-vlan-transportzone-960", "segmentation_id": 960, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap40240c4e-9f", "ovs_interfaceid": "40240c4e-9f3e-46aa-bd75-c62c2776d9b6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 841.846754] env[65726]: DEBUG oslo_concurrency.lockutils [None req-debf7b46-904e-41a8-8c0e-839a5a6575cb tempest-ListServerFiltersTestJSON-1076632629 tempest-ListServerFiltersTestJSON-1076632629-project-member] Lock "703fff2d-5aec-4ed1-b987-9e1cc3a25f67" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 45.900s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 841.876865] env[65726]: DEBUG oslo_vmware.api [None req-6f5d9ea9-7d8d-4c8e-8d3c-edf0333fb67b tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': task-5116014, 'name': CloneVM_Task} progress is 95%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.095730] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a1a4c202-18b9-4cfc-8067-d677c8415fff tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 842.340072] env[65726]: DEBUG oslo_concurrency.lockutils [req-9479b7c7-8bfa-4ecf-837c-3e91b8555cb2 req-efa5a0bf-092b-41fd-8fb5-1f013e9300a5 service nova] Releasing lock "refresh_cache-3cbbe146-be3a-43d5-867d-4d669884758c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 842.373039] env[65726]: DEBUG oslo_vmware.api [None req-6f5d9ea9-7d8d-4c8e-8d3c-edf0333fb67b tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': task-5116014, 'name': CloneVM_Task, 'duration_secs': 1.84305} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 842.375282] env[65726]: INFO nova.virt.vmwareapi.vmops [None req-6f5d9ea9-7d8d-4c8e-8d3c-edf0333fb67b tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 01d86089-6b9b-4588-864e-ef91375a1eea] Created linked-clone VM from snapshot [ 842.375282] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-755fcb33-23e8-4030-a19d-4c804d4dd8d3 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.383902] env[65726]: DEBUG nova.virt.vmwareapi.images [None req-6f5d9ea9-7d8d-4c8e-8d3c-edf0333fb67b tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 01d86089-6b9b-4588-864e-ef91375a1eea] Uploading image 8ecee2b1-8703-4e91-8a32-dc17932b237e {{(pid=65726) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 842.426097] env[65726]: DEBUG oslo_vmware.rw_handles [None req-6f5d9ea9-7d8d-4c8e-8d3c-edf0333fb67b tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 842.426097] env[65726]: value = "vm-995162" [ 842.426097] env[65726]: _type = "VirtualMachine" [ 842.426097] env[65726]: }. {{(pid=65726) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 842.426097] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-978bf63c-96c9-4bab-9e60-912a355cf7ed {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.435785] env[65726]: DEBUG oslo_vmware.rw_handles [None req-6f5d9ea9-7d8d-4c8e-8d3c-edf0333fb67b tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Lease: (returnval){ [ 842.435785] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52268f5f-a823-eb62-2448-895d6697ca56" [ 842.435785] env[65726]: _type = "HttpNfcLease" [ 842.435785] env[65726]: } obtained for exporting VM: (result){ [ 842.435785] env[65726]: value = "vm-995162" [ 842.435785] env[65726]: _type = "VirtualMachine" [ 842.435785] env[65726]: }. {{(pid=65726) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 842.436193] env[65726]: DEBUG oslo_vmware.api [None req-6f5d9ea9-7d8d-4c8e-8d3c-edf0333fb67b tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Waiting for the lease: (returnval){ [ 842.436193] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52268f5f-a823-eb62-2448-895d6697ca56" [ 842.436193] env[65726]: _type = "HttpNfcLease" [ 842.436193] env[65726]: } to be ready. {{(pid=65726) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 842.446518] env[65726]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 842.446518] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52268f5f-a823-eb62-2448-895d6697ca56" [ 842.446518] env[65726]: _type = "HttpNfcLease" [ 842.446518] env[65726]: } is initializing. {{(pid=65726) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 842.864459] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67ed3995-5f39-4e1b-ab1d-558ff5ac09f0 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.874161] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9c4454e-7a5e-4adf-b4f7-8656f5759ca1 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.908932] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a86d38d3-b432-4bf7-b1ca-c417bebcc3d5 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.917809] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62beac83-c5a0-4b92-8870-b99be3c6498f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.932663] env[65726]: DEBUG nova.compute.provider_tree [None req-a6bb6238-1018-4e0c-84f4-13a2ac492d77 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 842.944512] env[65726]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 842.944512] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52268f5f-a823-eb62-2448-895d6697ca56" [ 842.944512] env[65726]: _type = "HttpNfcLease" [ 842.944512] env[65726]: } is ready. {{(pid=65726) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 842.945754] env[65726]: DEBUG oslo_vmware.rw_handles [None req-6f5d9ea9-7d8d-4c8e-8d3c-edf0333fb67b tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 842.945754] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52268f5f-a823-eb62-2448-895d6697ca56" [ 842.945754] env[65726]: _type = "HttpNfcLease" [ 842.945754] env[65726]: }. {{(pid=65726) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 842.946674] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d8e1f4a-265a-4694-981d-f89d51c3ffc0 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.957769] env[65726]: DEBUG oslo_vmware.rw_handles [None req-6f5d9ea9-7d8d-4c8e-8d3c-edf0333fb67b tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52390d2a-eea8-668b-aa2d-087de53fcaa9/disk-0.vmdk from lease info. {{(pid=65726) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 842.958013] env[65726]: DEBUG oslo_vmware.rw_handles [None req-6f5d9ea9-7d8d-4c8e-8d3c-edf0333fb67b tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52390d2a-eea8-668b-aa2d-087de53fcaa9/disk-0.vmdk for reading. {{(pid=65726) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 843.145593] env[65726]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-3fdf11c7-c91e-4661-bbe0-f07db6cfce5b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.295260] env[65726]: DEBUG nova.compute.manager [req-00b8808e-be75-487d-9a3d-9d07fc8aaf54 req-e80505fe-db70-4b8f-b475-82bec1d60c70 service nova] [instance: 3cbbe146-be3a-43d5-867d-4d669884758c] Received event network-changed-40240c4e-9f3e-46aa-bd75-c62c2776d9b6 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 843.295260] env[65726]: DEBUG nova.compute.manager [req-00b8808e-be75-487d-9a3d-9d07fc8aaf54 req-e80505fe-db70-4b8f-b475-82bec1d60c70 service nova] [instance: 3cbbe146-be3a-43d5-867d-4d669884758c] Refreshing instance network info cache due to event network-changed-40240c4e-9f3e-46aa-bd75-c62c2776d9b6. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 843.295260] env[65726]: DEBUG oslo_concurrency.lockutils [req-00b8808e-be75-487d-9a3d-9d07fc8aaf54 req-e80505fe-db70-4b8f-b475-82bec1d60c70 service nova] Acquiring lock "refresh_cache-3cbbe146-be3a-43d5-867d-4d669884758c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 843.295260] env[65726]: DEBUG oslo_concurrency.lockutils [req-00b8808e-be75-487d-9a3d-9d07fc8aaf54 req-e80505fe-db70-4b8f-b475-82bec1d60c70 service nova] Acquired lock "refresh_cache-3cbbe146-be3a-43d5-867d-4d669884758c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 843.295260] env[65726]: DEBUG nova.network.neutron [req-00b8808e-be75-487d-9a3d-9d07fc8aaf54 req-e80505fe-db70-4b8f-b475-82bec1d60c70 service nova] [instance: 3cbbe146-be3a-43d5-867d-4d669884758c] Refreshing network info cache for port 40240c4e-9f3e-46aa-bd75-c62c2776d9b6 {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 843.437630] env[65726]: DEBUG nova.scheduler.client.report [None req-a6bb6238-1018-4e0c-84f4-13a2ac492d77 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 843.798917] env[65726]: WARNING openstack [req-00b8808e-be75-487d-9a3d-9d07fc8aaf54 req-e80505fe-db70-4b8f-b475-82bec1d60c70 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 843.799349] env[65726]: WARNING openstack [req-00b8808e-be75-487d-9a3d-9d07fc8aaf54 req-e80505fe-db70-4b8f-b475-82bec1d60c70 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 843.944328] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a6bb6238-1018-4e0c-84f4-13a2ac492d77 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.637s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 843.945643] env[65726]: DEBUG nova.compute.manager [None req-a6bb6238-1018-4e0c-84f4-13a2ac492d77 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] [instance: 44d73b2d-2b6a-4501-9944-432da3c9330d] Start building networks asynchronously for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 843.950802] env[65726]: DEBUG oslo_concurrency.lockutils [None req-4562f28d-bad2-4bca-833c-c069e59f6cc6 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 44.192s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 844.408566] env[65726]: WARNING openstack [req-00b8808e-be75-487d-9a3d-9d07fc8aaf54 req-e80505fe-db70-4b8f-b475-82bec1d60c70 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 844.409867] env[65726]: WARNING openstack [req-00b8808e-be75-487d-9a3d-9d07fc8aaf54 req-e80505fe-db70-4b8f-b475-82bec1d60c70 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 844.451703] env[65726]: DEBUG nova.compute.utils [None req-a6bb6238-1018-4e0c-84f4-13a2ac492d77 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Using /dev/sd instead of None {{(pid=65726) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 844.453244] env[65726]: DEBUG nova.compute.manager [None req-a6bb6238-1018-4e0c-84f4-13a2ac492d77 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] [instance: 44d73b2d-2b6a-4501-9944-432da3c9330d] Not allocating networking since 'none' was specified. {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2015}} [ 844.456340] env[65726]: INFO nova.compute.claims [None req-4562f28d-bad2-4bca-833c-c069e59f6cc6 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] [instance: 76249623-6f83-46a3-b8c5-c001111aa698] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 844.961923] env[65726]: DEBUG nova.compute.manager [None req-a6bb6238-1018-4e0c-84f4-13a2ac492d77 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] [instance: 44d73b2d-2b6a-4501-9944-432da3c9330d] Start building block device mappings for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 844.967020] env[65726]: INFO nova.compute.resource_tracker [None req-4562f28d-bad2-4bca-833c-c069e59f6cc6 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] [instance: 76249623-6f83-46a3-b8c5-c001111aa698] Updating resource usage from migration 8f31ab13-33c9-4d52-9a2a-8c3409ce12a9 [ 845.171388] env[65726]: WARNING openstack [req-00b8808e-be75-487d-9a3d-9d07fc8aaf54 req-e80505fe-db70-4b8f-b475-82bec1d60c70 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 845.171882] env[65726]: WARNING openstack [req-00b8808e-be75-487d-9a3d-9d07fc8aaf54 req-e80505fe-db70-4b8f-b475-82bec1d60c70 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 845.546806] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76c6b204-11d5-493e-aa51-be1ad1694aab {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.560925] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd431ed2-c859-4722-b19b-3659dac231bf {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.601436] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c677c91b-3c85-4b56-b3aa-81c8a1599dbd {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.610580] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44ac97c5-ab82-416c-8884-717e69f6e702 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.625743] env[65726]: DEBUG nova.compute.provider_tree [None req-4562f28d-bad2-4bca-833c-c069e59f6cc6 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 845.738194] env[65726]: DEBUG nova.network.neutron [req-00b8808e-be75-487d-9a3d-9d07fc8aaf54 req-e80505fe-db70-4b8f-b475-82bec1d60c70 service nova] [instance: 3cbbe146-be3a-43d5-867d-4d669884758c] Updated VIF entry in instance network info cache for port 40240c4e-9f3e-46aa-bd75-c62c2776d9b6. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 845.738585] env[65726]: DEBUG nova.network.neutron [req-00b8808e-be75-487d-9a3d-9d07fc8aaf54 req-e80505fe-db70-4b8f-b475-82bec1d60c70 service nova] [instance: 3cbbe146-be3a-43d5-867d-4d669884758c] Updating instance_info_cache with network_info: [{"id": "40240c4e-9f3e-46aa-bd75-c62c2776d9b6", "address": "fa:16:3e:8f:bd:09", "network": {"id": "2266eb6e-4c8f-4a3f-85d3-0c06022c0d5e", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-716412627-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "92c1cedbc08545d58fc7172ac95e2c23", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "24376631-ee89-4ff1-b8ac-f09911fc8329", "external-id": "nsx-vlan-transportzone-960", "segmentation_id": 960, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap40240c4e-9f", "ovs_interfaceid": "40240c4e-9f3e-46aa-bd75-c62c2776d9b6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 845.975762] env[65726]: DEBUG nova.compute.manager [None req-a6bb6238-1018-4e0c-84f4-13a2ac492d77 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] [instance: 44d73b2d-2b6a-4501-9944-432da3c9330d] Start spawning the instance on the hypervisor. {{(pid=65726) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 846.000199] env[65726]: DEBUG nova.virt.hardware [None req-a6bb6238-1018-4e0c-84f4-13a2ac492d77 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T19:28:34Z,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1539f06fa3534e90acbc3a60c4b8bd3f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T19:28:35Z,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 846.000524] env[65726]: DEBUG nova.virt.hardware [None req-a6bb6238-1018-4e0c-84f4-13a2ac492d77 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 846.000688] env[65726]: DEBUG nova.virt.hardware [None req-a6bb6238-1018-4e0c-84f4-13a2ac492d77 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 846.000913] env[65726]: DEBUG nova.virt.hardware [None req-a6bb6238-1018-4e0c-84f4-13a2ac492d77 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 846.001109] env[65726]: DEBUG nova.virt.hardware [None req-a6bb6238-1018-4e0c-84f4-13a2ac492d77 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 846.001265] env[65726]: DEBUG nova.virt.hardware [None req-a6bb6238-1018-4e0c-84f4-13a2ac492d77 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 846.001628] env[65726]: DEBUG nova.virt.hardware [None req-a6bb6238-1018-4e0c-84f4-13a2ac492d77 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 846.001871] env[65726]: DEBUG nova.virt.hardware [None req-a6bb6238-1018-4e0c-84f4-13a2ac492d77 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 846.002079] env[65726]: DEBUG nova.virt.hardware [None req-a6bb6238-1018-4e0c-84f4-13a2ac492d77 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 846.002251] env[65726]: DEBUG nova.virt.hardware [None req-a6bb6238-1018-4e0c-84f4-13a2ac492d77 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 846.002464] env[65726]: DEBUG nova.virt.hardware [None req-a6bb6238-1018-4e0c-84f4-13a2ac492d77 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 846.003542] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ac016e2-ff29-4c88-aa24-d31e6e21e030 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.013077] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-769e2d90-8e63-4fd5-95ef-dc329235ad00 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.029174] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-a6bb6238-1018-4e0c-84f4-13a2ac492d77 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] [instance: 44d73b2d-2b6a-4501-9944-432da3c9330d] Instance VIF info [] {{(pid=65726) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 846.036603] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-a6bb6238-1018-4e0c-84f4-13a2ac492d77 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 846.036603] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 44d73b2d-2b6a-4501-9944-432da3c9330d] Creating VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 846.036603] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-4af550fb-dfcf-457d-8dac-ee823054d4fe {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.057347] env[65726]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 846.057347] env[65726]: value = "task-5116016" [ 846.057347] env[65726]: _type = "Task" [ 846.057347] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.067982] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116016, 'name': CreateVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.131616] env[65726]: DEBUG nova.scheduler.client.report [None req-4562f28d-bad2-4bca-833c-c069e59f6cc6 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 846.242606] env[65726]: DEBUG oslo_concurrency.lockutils [req-00b8808e-be75-487d-9a3d-9d07fc8aaf54 req-e80505fe-db70-4b8f-b475-82bec1d60c70 service nova] Releasing lock "refresh_cache-3cbbe146-be3a-43d5-867d-4d669884758c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 846.570846] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116016, 'name': CreateVM_Task, 'duration_secs': 0.39161} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.570958] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 44d73b2d-2b6a-4501-9944-432da3c9330d] Created VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 846.571479] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a6bb6238-1018-4e0c-84f4-13a2ac492d77 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 846.571630] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a6bb6238-1018-4e0c-84f4-13a2ac492d77 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 846.571972] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a6bb6238-1018-4e0c-84f4-13a2ac492d77 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 846.572196] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9b38b831-1a16-447a-ac0d-e513c1f88750 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.578657] env[65726]: DEBUG oslo_vmware.api [None req-a6bb6238-1018-4e0c-84f4-13a2ac492d77 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Waiting for the task: (returnval){ [ 846.578657] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5236ea74-d5ba-40d8-5f7a-467b851f7ad1" [ 846.578657] env[65726]: _type = "Task" [ 846.578657] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.588167] env[65726]: DEBUG oslo_vmware.api [None req-a6bb6238-1018-4e0c-84f4-13a2ac492d77 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5236ea74-d5ba-40d8-5f7a-467b851f7ad1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.637468] env[65726]: DEBUG oslo_concurrency.lockutils [None req-4562f28d-bad2-4bca-833c-c069e59f6cc6 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.687s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 846.637573] env[65726]: INFO nova.compute.manager [None req-4562f28d-bad2-4bca-833c-c069e59f6cc6 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] [instance: 76249623-6f83-46a3-b8c5-c001111aa698] Migrating [ 846.646048] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b0e194cf-16dc-40d0-b65c-88bd21cd63e3 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 43.431s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 846.647832] env[65726]: INFO nova.compute.claims [None req-b0e194cf-16dc-40d0-b65c-88bd21cd63e3 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: cb8855b6-589d-4863-b86f-7fb4e30fb29e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 846.891131] env[65726]: DEBUG oslo_concurrency.lockutils [None req-ba8f1a5b-5e31-4b5e-94c2-8c464832cc35 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] Acquiring lock "3cbbe146-be3a-43d5-867d-4d669884758c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 846.891131] env[65726]: DEBUG oslo_concurrency.lockutils [None req-ba8f1a5b-5e31-4b5e-94c2-8c464832cc35 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] Lock "3cbbe146-be3a-43d5-867d-4d669884758c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 846.891131] env[65726]: DEBUG oslo_concurrency.lockutils [None req-ba8f1a5b-5e31-4b5e-94c2-8c464832cc35 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] Acquiring lock "3cbbe146-be3a-43d5-867d-4d669884758c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 846.891368] env[65726]: DEBUG oslo_concurrency.lockutils [None req-ba8f1a5b-5e31-4b5e-94c2-8c464832cc35 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] Lock "3cbbe146-be3a-43d5-867d-4d669884758c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 846.891465] env[65726]: DEBUG oslo_concurrency.lockutils [None req-ba8f1a5b-5e31-4b5e-94c2-8c464832cc35 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] Lock "3cbbe146-be3a-43d5-867d-4d669884758c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 846.896847] env[65726]: INFO nova.compute.manager [None req-ba8f1a5b-5e31-4b5e-94c2-8c464832cc35 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] [instance: 3cbbe146-be3a-43d5-867d-4d669884758c] Terminating instance [ 847.092504] env[65726]: DEBUG oslo_vmware.api [None req-a6bb6238-1018-4e0c-84f4-13a2ac492d77 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5236ea74-d5ba-40d8-5f7a-467b851f7ad1, 'name': SearchDatastore_Task, 'duration_secs': 0.013306} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.092925] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a6bb6238-1018-4e0c-84f4-13a2ac492d77 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 847.093112] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-a6bb6238-1018-4e0c-84f4-13a2ac492d77 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] [instance: 44d73b2d-2b6a-4501-9944-432da3c9330d] Processing image b52362a3-ee8a-4cbf-b06f-513b0cc8f95c {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 847.093341] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a6bb6238-1018-4e0c-84f4-13a2ac492d77 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 847.093480] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a6bb6238-1018-4e0c-84f4-13a2ac492d77 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 847.093649] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-a6bb6238-1018-4e0c-84f4-13a2ac492d77 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 847.093923] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2b20d990-f212-4d43-ac68-a3b73a1fecc2 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.109329] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-a6bb6238-1018-4e0c-84f4-13a2ac492d77 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 847.109584] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-a6bb6238-1018-4e0c-84f4-13a2ac492d77 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65726) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 847.110303] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0bbda4ab-8ec7-4324-9739-6a22b82ca42d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.116819] env[65726]: DEBUG oslo_vmware.api [None req-a6bb6238-1018-4e0c-84f4-13a2ac492d77 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Waiting for the task: (returnval){ [ 847.116819] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52885343-fa1d-21ac-1458-e19504534335" [ 847.116819] env[65726]: _type = "Task" [ 847.116819] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.126553] env[65726]: DEBUG oslo_vmware.api [None req-a6bb6238-1018-4e0c-84f4-13a2ac492d77 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52885343-fa1d-21ac-1458-e19504534335, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.166636] env[65726]: DEBUG oslo_concurrency.lockutils [None req-4562f28d-bad2-4bca-833c-c069e59f6cc6 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Acquiring lock "refresh_cache-76249623-6f83-46a3-b8c5-c001111aa698" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 847.166636] env[65726]: DEBUG oslo_concurrency.lockutils [None req-4562f28d-bad2-4bca-833c-c069e59f6cc6 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Acquired lock "refresh_cache-76249623-6f83-46a3-b8c5-c001111aa698" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 847.166636] env[65726]: DEBUG nova.network.neutron [None req-4562f28d-bad2-4bca-833c-c069e59f6cc6 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] [instance: 76249623-6f83-46a3-b8c5-c001111aa698] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 847.405691] env[65726]: DEBUG nova.compute.manager [None req-ba8f1a5b-5e31-4b5e-94c2-8c464832cc35 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] [instance: 3cbbe146-be3a-43d5-867d-4d669884758c] Start destroying the instance on the hypervisor. {{(pid=65726) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 847.406210] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-ba8f1a5b-5e31-4b5e-94c2-8c464832cc35 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] [instance: 3cbbe146-be3a-43d5-867d-4d669884758c] Destroying instance {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 847.406886] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd10be70-95af-4f54-99b4-92bcd89fa52c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.415628] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-ba8f1a5b-5e31-4b5e-94c2-8c464832cc35 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] [instance: 3cbbe146-be3a-43d5-867d-4d669884758c] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 847.416075] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a72a4feb-733a-4fb2-aab4-f82a4fecc653 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.424298] env[65726]: DEBUG oslo_vmware.api [None req-ba8f1a5b-5e31-4b5e-94c2-8c464832cc35 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] Waiting for the task: (returnval){ [ 847.424298] env[65726]: value = "task-5116017" [ 847.424298] env[65726]: _type = "Task" [ 847.424298] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.433675] env[65726]: DEBUG oslo_vmware.api [None req-ba8f1a5b-5e31-4b5e-94c2-8c464832cc35 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] Task: {'id': task-5116017, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.634589] env[65726]: DEBUG oslo_vmware.api [None req-a6bb6238-1018-4e0c-84f4-13a2ac492d77 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52885343-fa1d-21ac-1458-e19504534335, 'name': SearchDatastore_Task, 'duration_secs': 0.011255} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.634589] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-36e80edb-309f-4bfb-b594-82a11bfc3210 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.642762] env[65726]: DEBUG oslo_vmware.api [None req-a6bb6238-1018-4e0c-84f4-13a2ac492d77 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Waiting for the task: (returnval){ [ 847.642762] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52016413-440f-efe9-6885-22b3546c842f" [ 847.642762] env[65726]: _type = "Task" [ 847.642762] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.659670] env[65726]: DEBUG oslo_vmware.api [None req-a6bb6238-1018-4e0c-84f4-13a2ac492d77 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52016413-440f-efe9-6885-22b3546c842f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.671236] env[65726]: WARNING openstack [None req-4562f28d-bad2-4bca-833c-c069e59f6cc6 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 847.672041] env[65726]: WARNING openstack [None req-4562f28d-bad2-4bca-833c-c069e59f6cc6 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 847.816532] env[65726]: WARNING openstack [None req-4562f28d-bad2-4bca-833c-c069e59f6cc6 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 847.817156] env[65726]: WARNING openstack [None req-4562f28d-bad2-4bca-833c-c069e59f6cc6 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 847.900166] env[65726]: WARNING openstack [None req-4562f28d-bad2-4bca-833c-c069e59f6cc6 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 847.900599] env[65726]: WARNING openstack [None req-4562f28d-bad2-4bca-833c-c069e59f6cc6 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 847.937495] env[65726]: DEBUG oslo_vmware.api [None req-ba8f1a5b-5e31-4b5e-94c2-8c464832cc35 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] Task: {'id': task-5116017, 'name': PowerOffVM_Task, 'duration_secs': 0.32432} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.937759] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-ba8f1a5b-5e31-4b5e-94c2-8c464832cc35 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] [instance: 3cbbe146-be3a-43d5-867d-4d669884758c] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 847.937918] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-ba8f1a5b-5e31-4b5e-94c2-8c464832cc35 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] [instance: 3cbbe146-be3a-43d5-867d-4d669884758c] Unregistering the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 847.938202] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6bfa9d58-5c2e-4af1-ba9e-90615eb8a740 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.011447] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-ba8f1a5b-5e31-4b5e-94c2-8c464832cc35 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] [instance: 3cbbe146-be3a-43d5-867d-4d669884758c] Unregistered the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 848.011447] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-ba8f1a5b-5e31-4b5e-94c2-8c464832cc35 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] [instance: 3cbbe146-be3a-43d5-867d-4d669884758c] Deleting contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 848.011787] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-ba8f1a5b-5e31-4b5e-94c2-8c464832cc35 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] Deleting the datastore file [datastore1] 3cbbe146-be3a-43d5-867d-4d669884758c {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 848.012054] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-cceecfb6-6f73-4a4d-9ef9-8e9771110334 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.022595] env[65726]: DEBUG nova.network.neutron [None req-4562f28d-bad2-4bca-833c-c069e59f6cc6 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] [instance: 76249623-6f83-46a3-b8c5-c001111aa698] Updating instance_info_cache with network_info: [{"id": "e97a5da4-5540-4c0c-a33d-be1182e82aa7", "address": "fa:16:3e:af:b6:0f", "network": {"id": "55d629b8-a541-4720-9289-927bbc6e6f22", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.253", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "1539f06fa3534e90acbc3a60c4b8bd3f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "35342bcb-8b06-472e-b3c0-43fd3d6c4b30", "external-id": "nsx-vlan-transportzone-524", "segmentation_id": 524, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape97a5da4-55", "ovs_interfaceid": "e97a5da4-5540-4c0c-a33d-be1182e82aa7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 848.033048] env[65726]: DEBUG oslo_vmware.api [None req-ba8f1a5b-5e31-4b5e-94c2-8c464832cc35 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] Waiting for the task: (returnval){ [ 848.033048] env[65726]: value = "task-5116019" [ 848.033048] env[65726]: _type = "Task" [ 848.033048] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 848.043666] env[65726]: DEBUG oslo_vmware.api [None req-ba8f1a5b-5e31-4b5e-94c2-8c464832cc35 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] Task: {'id': task-5116019, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.133328] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9c76632b-07ea-476e-92d3-ffb457aef721 tempest-ServerAddressesTestJSON-569100133 tempest-ServerAddressesTestJSON-569100133-project-member] Acquiring lock "70375746-06dd-4710-9136-95e9e5759a0e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 848.133862] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9c76632b-07ea-476e-92d3-ffb457aef721 tempest-ServerAddressesTestJSON-569100133 tempest-ServerAddressesTestJSON-569100133-project-member] Lock "70375746-06dd-4710-9136-95e9e5759a0e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 848.156283] env[65726]: DEBUG oslo_vmware.api [None req-a6bb6238-1018-4e0c-84f4-13a2ac492d77 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52016413-440f-efe9-6885-22b3546c842f, 'name': SearchDatastore_Task, 'duration_secs': 0.012667} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.156452] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a6bb6238-1018-4e0c-84f4-13a2ac492d77 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 848.156730] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-a6bb6238-1018-4e0c-84f4-13a2ac492d77 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore2] 44d73b2d-2b6a-4501-9944-432da3c9330d/44d73b2d-2b6a-4501-9944-432da3c9330d.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 848.158173] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-54c63df5-9f7a-4bd3-b91f-91555bfd1f72 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.167285] env[65726]: DEBUG oslo_vmware.api [None req-a6bb6238-1018-4e0c-84f4-13a2ac492d77 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Waiting for the task: (returnval){ [ 848.167285] env[65726]: value = "task-5116020" [ 848.167285] env[65726]: _type = "Task" [ 848.167285] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 848.183275] env[65726]: DEBUG oslo_vmware.api [None req-a6bb6238-1018-4e0c-84f4-13a2ac492d77 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Task: {'id': task-5116020, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.399804] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca7d59dc-f4c7-456f-93f6-bd8414c24721 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.409347] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31082277-40a4-4100-80f0-d16fd6b30f03 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.447694] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1a3b450-ade5-4363-86b0-1b7feb22d19f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.457919] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a924d21d-bbf4-4433-8d74-a1842f8758c3 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.474716] env[65726]: DEBUG nova.compute.provider_tree [None req-b0e194cf-16dc-40d0-b65c-88bd21cd63e3 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Updating inventory in ProviderTree for provider 07c4692f-bdb4-4058-9173-ff9664830295 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 848.526084] env[65726]: DEBUG oslo_concurrency.lockutils [None req-4562f28d-bad2-4bca-833c-c069e59f6cc6 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Releasing lock "refresh_cache-76249623-6f83-46a3-b8c5-c001111aa698" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 848.546330] env[65726]: DEBUG oslo_vmware.api [None req-ba8f1a5b-5e31-4b5e-94c2-8c464832cc35 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] Task: {'id': task-5116019, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.213214} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.547092] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-ba8f1a5b-5e31-4b5e-94c2-8c464832cc35 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] Deleted the datastore file {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 848.547354] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-ba8f1a5b-5e31-4b5e-94c2-8c464832cc35 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] [instance: 3cbbe146-be3a-43d5-867d-4d669884758c] Deleted contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 848.547574] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-ba8f1a5b-5e31-4b5e-94c2-8c464832cc35 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] [instance: 3cbbe146-be3a-43d5-867d-4d669884758c] Instance destroyed {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 848.547658] env[65726]: INFO nova.compute.manager [None req-ba8f1a5b-5e31-4b5e-94c2-8c464832cc35 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] [instance: 3cbbe146-be3a-43d5-867d-4d669884758c] Took 1.14 seconds to destroy the instance on the hypervisor. [ 848.548075] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-ba8f1a5b-5e31-4b5e-94c2-8c464832cc35 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 848.548075] env[65726]: DEBUG nova.compute.manager [-] [instance: 3cbbe146-be3a-43d5-867d-4d669884758c] Deallocating network for instance {{(pid=65726) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 848.548329] env[65726]: DEBUG nova.network.neutron [-] [instance: 3cbbe146-be3a-43d5-867d-4d669884758c] deallocate_for_instance() {{(pid=65726) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 848.549418] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 848.549826] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 848.618154] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 848.618154] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 848.684583] env[65726]: DEBUG oslo_vmware.api [None req-a6bb6238-1018-4e0c-84f4-13a2ac492d77 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Task: {'id': task-5116020, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.990447] env[65726]: DEBUG nova.compute.manager [req-1e0534df-b9c2-415f-8a0a-8c77b86cf224 req-325b314e-0018-4a3a-a935-e2f0629c09e0 service nova] [instance: 3cbbe146-be3a-43d5-867d-4d669884758c] Received event network-vif-deleted-40240c4e-9f3e-46aa-bd75-c62c2776d9b6 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 848.990703] env[65726]: INFO nova.compute.manager [req-1e0534df-b9c2-415f-8a0a-8c77b86cf224 req-325b314e-0018-4a3a-a935-e2f0629c09e0 service nova] [instance: 3cbbe146-be3a-43d5-867d-4d669884758c] Neutron deleted interface 40240c4e-9f3e-46aa-bd75-c62c2776d9b6; detaching it from the instance and deleting it from the info cache [ 848.990878] env[65726]: DEBUG nova.network.neutron [req-1e0534df-b9c2-415f-8a0a-8c77b86cf224 req-325b314e-0018-4a3a-a935-e2f0629c09e0 service nova] [instance: 3cbbe146-be3a-43d5-867d-4d669884758c] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 849.006948] env[65726]: ERROR nova.scheduler.client.report [None req-b0e194cf-16dc-40d0-b65c-88bd21cd63e3 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [req-30256910-5da2-4326-8338-2ddccfbedbb1] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 07c4692f-bdb4-4058-9173-ff9664830295. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-30256910-5da2-4326-8338-2ddccfbedbb1"}]} [ 849.034641] env[65726]: DEBUG nova.scheduler.client.report [None req-b0e194cf-16dc-40d0-b65c-88bd21cd63e3 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Refreshing inventories for resource provider 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 849.055644] env[65726]: DEBUG nova.scheduler.client.report [None req-b0e194cf-16dc-40d0-b65c-88bd21cd63e3 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Updating ProviderTree inventory for provider 07c4692f-bdb4-4058-9173-ff9664830295 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 849.055919] env[65726]: DEBUG nova.compute.provider_tree [None req-b0e194cf-16dc-40d0-b65c-88bd21cd63e3 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Updating inventory in ProviderTree for provider 07c4692f-bdb4-4058-9173-ff9664830295 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 849.074388] env[65726]: DEBUG nova.scheduler.client.report [None req-b0e194cf-16dc-40d0-b65c-88bd21cd63e3 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Refreshing aggregate associations for resource provider 07c4692f-bdb4-4058-9173-ff9664830295, aggregates: None {{(pid=65726) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 849.095031] env[65726]: DEBUG nova.scheduler.client.report [None req-b0e194cf-16dc-40d0-b65c-88bd21cd63e3 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Refreshing trait associations for resource provider 07c4692f-bdb4-4058-9173-ff9664830295, traits: COMPUTE_IMAGE_TYPE_ISO,HW_ARCH_X86_64,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=65726) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 849.192066] env[65726]: DEBUG oslo_vmware.api [None req-a6bb6238-1018-4e0c-84f4-13a2ac492d77 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Task: {'id': task-5116020, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.521288} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.192066] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-a6bb6238-1018-4e0c-84f4-13a2ac492d77 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore2] 44d73b2d-2b6a-4501-9944-432da3c9330d/44d73b2d-2b6a-4501-9944-432da3c9330d.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 849.192454] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-a6bb6238-1018-4e0c-84f4-13a2ac492d77 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] [instance: 44d73b2d-2b6a-4501-9944-432da3c9330d] Extending root virtual disk to 1048576 {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 849.192454] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-28d75b71-7ce9-4d73-94ec-0b411b14fb20 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.203239] env[65726]: DEBUG oslo_vmware.api [None req-a6bb6238-1018-4e0c-84f4-13a2ac492d77 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Waiting for the task: (returnval){ [ 849.203239] env[65726]: value = "task-5116021" [ 849.203239] env[65726]: _type = "Task" [ 849.203239] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.214588] env[65726]: DEBUG oslo_vmware.api [None req-a6bb6238-1018-4e0c-84f4-13a2ac492d77 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Task: {'id': task-5116021, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.365104] env[65726]: DEBUG nova.network.neutron [-] [instance: 3cbbe146-be3a-43d5-867d-4d669884758c] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 849.495115] env[65726]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1018a5a1-fe31-432c-a370-ae7ed685c1c8 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.508022] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65cdd7c0-b87f-48eb-b50f-caa47fd6df47 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.557524] env[65726]: DEBUG nova.compute.manager [req-1e0534df-b9c2-415f-8a0a-8c77b86cf224 req-325b314e-0018-4a3a-a935-e2f0629c09e0 service nova] [instance: 3cbbe146-be3a-43d5-867d-4d669884758c] Detach interface failed, port_id=40240c4e-9f3e-46aa-bd75-c62c2776d9b6, reason: Instance 3cbbe146-be3a-43d5-867d-4d669884758c could not be found. {{(pid=65726) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11669}} [ 849.665813] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82680613-95bd-4850-9ac8-47e664c4532e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.676858] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27785fff-1e71-4347-ba92-b8a820e8ca5b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.721103] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8bf2b287-0d10-44e4-82f4-70a56d3b5c47 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.731656] env[65726]: DEBUG oslo_vmware.api [None req-a6bb6238-1018-4e0c-84f4-13a2ac492d77 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Task: {'id': task-5116021, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.076926} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.734387] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-a6bb6238-1018-4e0c-84f4-13a2ac492d77 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] [instance: 44d73b2d-2b6a-4501-9944-432da3c9330d] Extended root virtual disk {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 849.735290] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db39678b-98ad-42d4-b5fc-dce62f94b8ad {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.739134] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec7f45d1-ef39-4ede-a5e7-e20ed460a873 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.761023] env[65726]: DEBUG nova.compute.provider_tree [None req-b0e194cf-16dc-40d0-b65c-88bd21cd63e3 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Updating inventory in ProviderTree for provider 07c4692f-bdb4-4058-9173-ff9664830295 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 849.779275] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-a6bb6238-1018-4e0c-84f4-13a2ac492d77 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] [instance: 44d73b2d-2b6a-4501-9944-432da3c9330d] Reconfiguring VM instance instance-00000038 to attach disk [datastore2] 44d73b2d-2b6a-4501-9944-432da3c9330d/44d73b2d-2b6a-4501-9944-432da3c9330d.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 849.783428] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7a9aad6b-e524-4955-bf60-9f0178609f02 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.805124] env[65726]: DEBUG oslo_vmware.api [None req-a6bb6238-1018-4e0c-84f4-13a2ac492d77 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Waiting for the task: (returnval){ [ 849.805124] env[65726]: value = "task-5116022" [ 849.805124] env[65726]: _type = "Task" [ 849.805124] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.815435] env[65726]: DEBUG oslo_vmware.api [None req-a6bb6238-1018-4e0c-84f4-13a2ac492d77 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Task: {'id': task-5116022, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.837994] env[65726]: DEBUG nova.scheduler.client.report [None req-b0e194cf-16dc-40d0-b65c-88bd21cd63e3 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Updated inventory for provider 07c4692f-bdb4-4058-9173-ff9664830295 with generation 92 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 849.838315] env[65726]: DEBUG nova.compute.provider_tree [None req-b0e194cf-16dc-40d0-b65c-88bd21cd63e3 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Updating resource provider 07c4692f-bdb4-4058-9173-ff9664830295 generation from 92 to 93 during operation: update_inventory {{(pid=65726) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 849.838539] env[65726]: DEBUG nova.compute.provider_tree [None req-b0e194cf-16dc-40d0-b65c-88bd21cd63e3 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Updating inventory in ProviderTree for provider 07c4692f-bdb4-4058-9173-ff9664830295 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 849.868854] env[65726]: INFO nova.compute.manager [-] [instance: 3cbbe146-be3a-43d5-867d-4d669884758c] Took 1.32 seconds to deallocate network for instance. [ 850.065761] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6389ae15-be3d-4bd1-b617-3c3710661094 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.086229] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-4562f28d-bad2-4bca-833c-c069e59f6cc6 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] [instance: 76249623-6f83-46a3-b8c5-c001111aa698] Updating instance '76249623-6f83-46a3-b8c5-c001111aa698' progress to 0 {{(pid=65726) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 850.317040] env[65726]: DEBUG oslo_vmware.api [None req-a6bb6238-1018-4e0c-84f4-13a2ac492d77 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Task: {'id': task-5116022, 'name': ReconfigVM_Task, 'duration_secs': 0.325142} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.317040] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-a6bb6238-1018-4e0c-84f4-13a2ac492d77 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] [instance: 44d73b2d-2b6a-4501-9944-432da3c9330d] Reconfigured VM instance instance-00000038 to attach disk [datastore2] 44d73b2d-2b6a-4501-9944-432da3c9330d/44d73b2d-2b6a-4501-9944-432da3c9330d.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 850.317698] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b3775596-913f-40ef-baae-4ae37a2e42d6 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.325264] env[65726]: DEBUG oslo_vmware.api [None req-a6bb6238-1018-4e0c-84f4-13a2ac492d77 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Waiting for the task: (returnval){ [ 850.325264] env[65726]: value = "task-5116023" [ 850.325264] env[65726]: _type = "Task" [ 850.325264] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.334510] env[65726]: DEBUG oslo_vmware.api [None req-a6bb6238-1018-4e0c-84f4-13a2ac492d77 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Task: {'id': task-5116023, 'name': Rename_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.347405] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b0e194cf-16dc-40d0-b65c-88bd21cd63e3 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.701s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 850.348625] env[65726]: DEBUG nova.compute.manager [None req-b0e194cf-16dc-40d0-b65c-88bd21cd63e3 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: cb8855b6-589d-4863-b86f-7fb4e30fb29e] Start building networks asynchronously for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 850.351767] env[65726]: DEBUG oslo_concurrency.lockutils [None req-861b1b19-8cf2-4cb5-a3b7-93ab69618cb7 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 44.209s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 850.351953] env[65726]: DEBUG nova.objects.instance [None req-861b1b19-8cf2-4cb5-a3b7-93ab69618cb7 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Lazy-loading 'resources' on Instance uuid 621b899a-e9ec-4fc0-a574-4a08cecff6d9 {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 850.379874] env[65726]: DEBUG oslo_concurrency.lockutils [None req-ba8f1a5b-5e31-4b5e-94c2-8c464832cc35 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 850.594952] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-4562f28d-bad2-4bca-833c-c069e59f6cc6 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] [instance: 76249623-6f83-46a3-b8c5-c001111aa698] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 850.595342] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9a8e237a-af2b-450b-b423-b0e022e7535c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.604858] env[65726]: DEBUG oslo_vmware.api [None req-4562f28d-bad2-4bca-833c-c069e59f6cc6 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Waiting for the task: (returnval){ [ 850.604858] env[65726]: value = "task-5116024" [ 850.604858] env[65726]: _type = "Task" [ 850.604858] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.616409] env[65726]: DEBUG oslo_vmware.api [None req-4562f28d-bad2-4bca-833c-c069e59f6cc6 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Task: {'id': task-5116024, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.837958] env[65726]: DEBUG oslo_vmware.api [None req-a6bb6238-1018-4e0c-84f4-13a2ac492d77 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Task: {'id': task-5116023, 'name': Rename_Task, 'duration_secs': 0.144526} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.837958] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-a6bb6238-1018-4e0c-84f4-13a2ac492d77 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] [instance: 44d73b2d-2b6a-4501-9944-432da3c9330d] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 850.837958] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-106b80a7-f4e7-4dea-be48-35324872e5d5 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.846138] env[65726]: DEBUG oslo_vmware.api [None req-a6bb6238-1018-4e0c-84f4-13a2ac492d77 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Waiting for the task: (returnval){ [ 850.846138] env[65726]: value = "task-5116025" [ 850.846138] env[65726]: _type = "Task" [ 850.846138] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.859109] env[65726]: DEBUG nova.compute.utils [None req-b0e194cf-16dc-40d0-b65c-88bd21cd63e3 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Using /dev/sd instead of None {{(pid=65726) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 850.862213] env[65726]: DEBUG oslo_vmware.api [None req-a6bb6238-1018-4e0c-84f4-13a2ac492d77 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Task: {'id': task-5116025, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.863606] env[65726]: DEBUG nova.compute.manager [None req-b0e194cf-16dc-40d0-b65c-88bd21cd63e3 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: cb8855b6-589d-4863-b86f-7fb4e30fb29e] Allocating IP information in the background. {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 850.863606] env[65726]: DEBUG nova.network.neutron [None req-b0e194cf-16dc-40d0-b65c-88bd21cd63e3 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: cb8855b6-589d-4863-b86f-7fb4e30fb29e] allocate_for_instance() {{(pid=65726) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 850.863606] env[65726]: WARNING neutronclient.v2_0.client [None req-b0e194cf-16dc-40d0-b65c-88bd21cd63e3 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 850.863606] env[65726]: WARNING neutronclient.v2_0.client [None req-b0e194cf-16dc-40d0-b65c-88bd21cd63e3 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 850.864136] env[65726]: WARNING openstack [None req-b0e194cf-16dc-40d0-b65c-88bd21cd63e3 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 850.864443] env[65726]: WARNING openstack [None req-b0e194cf-16dc-40d0-b65c-88bd21cd63e3 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 850.922981] env[65726]: DEBUG nova.policy [None req-b0e194cf-16dc-40d0-b65c-88bd21cd63e3 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4b6c9cc4ac3c487693240e375a83ea19', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b794c99309e94c1181fc9d8eeeb84702', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65726) authorize /opt/stack/nova/nova/policy.py:192}} [ 851.115963] env[65726]: DEBUG oslo_vmware.api [None req-4562f28d-bad2-4bca-833c-c069e59f6cc6 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Task: {'id': task-5116024, 'name': PowerOffVM_Task, 'duration_secs': 0.220079} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.118901] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-4562f28d-bad2-4bca-833c-c069e59f6cc6 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] [instance: 76249623-6f83-46a3-b8c5-c001111aa698] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 851.119105] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-4562f28d-bad2-4bca-833c-c069e59f6cc6 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] [instance: 76249623-6f83-46a3-b8c5-c001111aa698] Updating instance '76249623-6f83-46a3-b8c5-c001111aa698' progress to 17 {{(pid=65726) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 851.226447] env[65726]: DEBUG nova.network.neutron [None req-b0e194cf-16dc-40d0-b65c-88bd21cd63e3 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: cb8855b6-589d-4863-b86f-7fb4e30fb29e] Successfully created port: 136da8b5-d8fd-4cef-ae44-a9ecc98c28d2 {{(pid=65726) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 851.361396] env[65726]: DEBUG oslo_vmware.api [None req-a6bb6238-1018-4e0c-84f4-13a2ac492d77 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Task: {'id': task-5116025, 'name': PowerOnVM_Task, 'duration_secs': 0.487581} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.361755] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-a6bb6238-1018-4e0c-84f4-13a2ac492d77 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] [instance: 44d73b2d-2b6a-4501-9944-432da3c9330d] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 851.361931] env[65726]: INFO nova.compute.manager [None req-a6bb6238-1018-4e0c-84f4-13a2ac492d77 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] [instance: 44d73b2d-2b6a-4501-9944-432da3c9330d] Took 5.39 seconds to spawn the instance on the hypervisor. [ 851.362129] env[65726]: DEBUG nova.compute.manager [None req-a6bb6238-1018-4e0c-84f4-13a2ac492d77 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] [instance: 44d73b2d-2b6a-4501-9944-432da3c9330d] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 851.363179] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63e38ad5-f158-4e40-a722-0c8f534934f6 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.373570] env[65726]: DEBUG nova.compute.manager [None req-b0e194cf-16dc-40d0-b65c-88bd21cd63e3 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: cb8855b6-589d-4863-b86f-7fb4e30fb29e] Start building block device mappings for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 851.413185] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d216cbf4-5c5d-4909-87c8-ef0e53e4181c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.424040] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42fd8129-29ed-4507-8bbe-d463249ddbc0 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.456571] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9d7f934-01c9-422c-915a-0e12021b31e8 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.465806] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f107367e-2e36-4e82-a3b2-c66068a39641 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.482129] env[65726]: DEBUG nova.compute.provider_tree [None req-861b1b19-8cf2-4cb5-a3b7-93ab69618cb7 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 851.625791] env[65726]: DEBUG nova.virt.hardware [None req-4562f28d-bad2-4bca-833c-c069e59f6cc6 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 851.626096] env[65726]: DEBUG nova.virt.hardware [None req-4562f28d-bad2-4bca-833c-c069e59f6cc6 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 851.626264] env[65726]: DEBUG nova.virt.hardware [None req-4562f28d-bad2-4bca-833c-c069e59f6cc6 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 851.626442] env[65726]: DEBUG nova.virt.hardware [None req-4562f28d-bad2-4bca-833c-c069e59f6cc6 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 851.626582] env[65726]: DEBUG nova.virt.hardware [None req-4562f28d-bad2-4bca-833c-c069e59f6cc6 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 851.626719] env[65726]: DEBUG nova.virt.hardware [None req-4562f28d-bad2-4bca-833c-c069e59f6cc6 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 851.626949] env[65726]: DEBUG nova.virt.hardware [None req-4562f28d-bad2-4bca-833c-c069e59f6cc6 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 851.627118] env[65726]: DEBUG nova.virt.hardware [None req-4562f28d-bad2-4bca-833c-c069e59f6cc6 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 851.627279] env[65726]: DEBUG nova.virt.hardware [None req-4562f28d-bad2-4bca-833c-c069e59f6cc6 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 851.627441] env[65726]: DEBUG nova.virt.hardware [None req-4562f28d-bad2-4bca-833c-c069e59f6cc6 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 851.627641] env[65726]: DEBUG nova.virt.hardware [None req-4562f28d-bad2-4bca-833c-c069e59f6cc6 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 851.633548] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8cf257b1-3247-4be1-8f7e-703a1a7d7c7d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.653052] env[65726]: DEBUG oslo_vmware.api [None req-4562f28d-bad2-4bca-833c-c069e59f6cc6 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Waiting for the task: (returnval){ [ 851.653052] env[65726]: value = "task-5116026" [ 851.653052] env[65726]: _type = "Task" [ 851.653052] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.662497] env[65726]: DEBUG oslo_vmware.api [None req-4562f28d-bad2-4bca-833c-c069e59f6cc6 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Task: {'id': task-5116026, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.889254] env[65726]: INFO nova.compute.manager [None req-a6bb6238-1018-4e0c-84f4-13a2ac492d77 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] [instance: 44d73b2d-2b6a-4501-9944-432da3c9330d] Took 52.51 seconds to build instance. [ 851.985940] env[65726]: DEBUG nova.scheduler.client.report [None req-861b1b19-8cf2-4cb5-a3b7-93ab69618cb7 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 852.165100] env[65726]: DEBUG oslo_vmware.api [None req-4562f28d-bad2-4bca-833c-c069e59f6cc6 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Task: {'id': task-5116026, 'name': ReconfigVM_Task, 'duration_secs': 0.284435} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.165578] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-4562f28d-bad2-4bca-833c-c069e59f6cc6 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] [instance: 76249623-6f83-46a3-b8c5-c001111aa698] Updating instance '76249623-6f83-46a3-b8c5-c001111aa698' progress to 33 {{(pid=65726) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 852.390503] env[65726]: DEBUG nova.compute.manager [None req-b0e194cf-16dc-40d0-b65c-88bd21cd63e3 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: cb8855b6-589d-4863-b86f-7fb4e30fb29e] Start spawning the instance on the hypervisor. {{(pid=65726) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 852.394456] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a6bb6238-1018-4e0c-84f4-13a2ac492d77 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Lock "44d73b2d-2b6a-4501-9944-432da3c9330d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 84.681s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 852.423926] env[65726]: DEBUG nova.virt.hardware [None req-b0e194cf-16dc-40d0-b65c-88bd21cd63e3 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T19:28:34Z,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1539f06fa3534e90acbc3a60c4b8bd3f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T19:28:35Z,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 852.424193] env[65726]: DEBUG nova.virt.hardware [None req-b0e194cf-16dc-40d0-b65c-88bd21cd63e3 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 852.424345] env[65726]: DEBUG nova.virt.hardware [None req-b0e194cf-16dc-40d0-b65c-88bd21cd63e3 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 852.424610] env[65726]: DEBUG nova.virt.hardware [None req-b0e194cf-16dc-40d0-b65c-88bd21cd63e3 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 852.424788] env[65726]: DEBUG nova.virt.hardware [None req-b0e194cf-16dc-40d0-b65c-88bd21cd63e3 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 852.424934] env[65726]: DEBUG nova.virt.hardware [None req-b0e194cf-16dc-40d0-b65c-88bd21cd63e3 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 852.425156] env[65726]: DEBUG nova.virt.hardware [None req-b0e194cf-16dc-40d0-b65c-88bd21cd63e3 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 852.425310] env[65726]: DEBUG nova.virt.hardware [None req-b0e194cf-16dc-40d0-b65c-88bd21cd63e3 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 852.425480] env[65726]: DEBUG nova.virt.hardware [None req-b0e194cf-16dc-40d0-b65c-88bd21cd63e3 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 852.426684] env[65726]: DEBUG nova.virt.hardware [None req-b0e194cf-16dc-40d0-b65c-88bd21cd63e3 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 852.426684] env[65726]: DEBUG nova.virt.hardware [None req-b0e194cf-16dc-40d0-b65c-88bd21cd63e3 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 852.426905] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2eb684a7-20f2-4821-8fc0-8eb5904f66da {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.441614] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6e37dc3-c4b2-4bca-a6b1-aec3f652a299 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.474119] env[65726]: DEBUG oslo_vmware.rw_handles [None req-6f5d9ea9-7d8d-4c8e-8d3c-edf0333fb67b tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52390d2a-eea8-668b-aa2d-087de53fcaa9/disk-0.vmdk. {{(pid=65726) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 852.475277] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77790292-1116-4532-bc46-47eda8fbf1e1 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.482364] env[65726]: DEBUG oslo_vmware.rw_handles [None req-6f5d9ea9-7d8d-4c8e-8d3c-edf0333fb67b tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52390d2a-eea8-668b-aa2d-087de53fcaa9/disk-0.vmdk is in state: ready. {{(pid=65726) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 852.482533] env[65726]: ERROR oslo_vmware.rw_handles [None req-6f5d9ea9-7d8d-4c8e-8d3c-edf0333fb67b tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52390d2a-eea8-668b-aa2d-087de53fcaa9/disk-0.vmdk due to incomplete transfer. [ 852.482779] env[65726]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-79d5808d-1dab-4037-96b5-f514a6e38468 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.491206] env[65726]: DEBUG oslo_concurrency.lockutils [None req-861b1b19-8cf2-4cb5-a3b7-93ab69618cb7 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.140s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 852.493220] env[65726]: DEBUG oslo_vmware.rw_handles [None req-6f5d9ea9-7d8d-4c8e-8d3c-edf0333fb67b tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52390d2a-eea8-668b-aa2d-087de53fcaa9/disk-0.vmdk. {{(pid=65726) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 852.493404] env[65726]: DEBUG nova.virt.vmwareapi.images [None req-6f5d9ea9-7d8d-4c8e-8d3c-edf0333fb67b tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 01d86089-6b9b-4588-864e-ef91375a1eea] Uploaded image 8ecee2b1-8703-4e91-8a32-dc17932b237e to the Glance image server {{(pid=65726) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 852.495918] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-6f5d9ea9-7d8d-4c8e-8d3c-edf0333fb67b tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 01d86089-6b9b-4588-864e-ef91375a1eea] Destroying the VM {{(pid=65726) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 852.496650] env[65726]: DEBUG oslo_concurrency.lockutils [None req-55d74d77-063b-418b-b818-59d05c490f34 tempest-ImagesNegativeTestJSON-1015404931 tempest-ImagesNegativeTestJSON-1015404931-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 42.045s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 852.498116] env[65726]: INFO nova.compute.claims [None req-55d74d77-063b-418b-b818-59d05c490f34 tempest-ImagesNegativeTestJSON-1015404931 tempest-ImagesNegativeTestJSON-1015404931-project-member] [instance: c4177e20-b1bd-4b54-a275-c93582359a07] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 852.501190] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-c48f8b6b-eb21-48b9-88ad-5d2534ceb621 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.509770] env[65726]: DEBUG oslo_vmware.api [None req-6f5d9ea9-7d8d-4c8e-8d3c-edf0333fb67b tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Waiting for the task: (returnval){ [ 852.509770] env[65726]: value = "task-5116027" [ 852.509770] env[65726]: _type = "Task" [ 852.509770] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.520122] env[65726]: DEBUG oslo_vmware.api [None req-6f5d9ea9-7d8d-4c8e-8d3c-edf0333fb67b tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': task-5116027, 'name': Destroy_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.529445] env[65726]: INFO nova.scheduler.client.report [None req-861b1b19-8cf2-4cb5-a3b7-93ab69618cb7 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Deleted allocations for instance 621b899a-e9ec-4fc0-a574-4a08cecff6d9 [ 852.673696] env[65726]: DEBUG nova.virt.hardware [None req-4562f28d-bad2-4bca-833c-c069e59f6cc6 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:32:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='d5a398f3-3467-4467-9a69-ab6d1bafa43b',id=37,is_public=True,memory_mb=192,name='tempest-test_resize_flavor_-1568765588',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 852.673919] env[65726]: DEBUG nova.virt.hardware [None req-4562f28d-bad2-4bca-833c-c069e59f6cc6 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 852.674105] env[65726]: DEBUG nova.virt.hardware [None req-4562f28d-bad2-4bca-833c-c069e59f6cc6 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 852.674299] env[65726]: DEBUG nova.virt.hardware [None req-4562f28d-bad2-4bca-833c-c069e59f6cc6 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 852.674458] env[65726]: DEBUG nova.virt.hardware [None req-4562f28d-bad2-4bca-833c-c069e59f6cc6 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 852.674736] env[65726]: DEBUG nova.virt.hardware [None req-4562f28d-bad2-4bca-833c-c069e59f6cc6 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 852.674816] env[65726]: DEBUG nova.virt.hardware [None req-4562f28d-bad2-4bca-833c-c069e59f6cc6 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 852.674964] env[65726]: DEBUG nova.virt.hardware [None req-4562f28d-bad2-4bca-833c-c069e59f6cc6 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 852.675302] env[65726]: DEBUG nova.virt.hardware [None req-4562f28d-bad2-4bca-833c-c069e59f6cc6 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 852.675531] env[65726]: DEBUG nova.virt.hardware [None req-4562f28d-bad2-4bca-833c-c069e59f6cc6 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 852.675663] env[65726]: DEBUG nova.virt.hardware [None req-4562f28d-bad2-4bca-833c-c069e59f6cc6 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 852.680978] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-4562f28d-bad2-4bca-833c-c069e59f6cc6 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] [instance: 76249623-6f83-46a3-b8c5-c001111aa698] Reconfiguring VM instance instance-0000002e to detach disk 2000 {{(pid=65726) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 852.681301] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-129424cb-cbc3-469c-aa24-e3d16fb5acda {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.703273] env[65726]: DEBUG oslo_vmware.api [None req-4562f28d-bad2-4bca-833c-c069e59f6cc6 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Waiting for the task: (returnval){ [ 852.703273] env[65726]: value = "task-5116028" [ 852.703273] env[65726]: _type = "Task" [ 852.703273] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.713361] env[65726]: DEBUG oslo_vmware.api [None req-4562f28d-bad2-4bca-833c-c069e59f6cc6 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Task: {'id': task-5116028, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.841012] env[65726]: DEBUG nova.network.neutron [None req-b0e194cf-16dc-40d0-b65c-88bd21cd63e3 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: cb8855b6-589d-4863-b86f-7fb4e30fb29e] Successfully updated port: 136da8b5-d8fd-4cef-ae44-a9ecc98c28d2 {{(pid=65726) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 852.899543] env[65726]: DEBUG nova.compute.manager [None req-4826faf0-9908-4f3c-a509-20fe2a708723 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] [instance: 29823bc7-3909-4ab7-8119-91ee59b289c5] Starting instance... {{(pid=65726) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 853.019918] env[65726]: DEBUG oslo_vmware.api [None req-6f5d9ea9-7d8d-4c8e-8d3c-edf0333fb67b tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': task-5116027, 'name': Destroy_Task, 'duration_secs': 0.424641} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 853.020882] env[65726]: INFO nova.virt.vmwareapi.vm_util [None req-6f5d9ea9-7d8d-4c8e-8d3c-edf0333fb67b tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 01d86089-6b9b-4588-864e-ef91375a1eea] Destroyed the VM [ 853.021140] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-6f5d9ea9-7d8d-4c8e-8d3c-edf0333fb67b tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 01d86089-6b9b-4588-864e-ef91375a1eea] Deleting Snapshot of the VM instance {{(pid=65726) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 853.021407] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-4a33f1c5-4326-47b3-9026-9bb9ce551566 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.029302] env[65726]: DEBUG oslo_vmware.api [None req-6f5d9ea9-7d8d-4c8e-8d3c-edf0333fb67b tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Waiting for the task: (returnval){ [ 853.029302] env[65726]: value = "task-5116029" [ 853.029302] env[65726]: _type = "Task" [ 853.029302] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.043562] env[65726]: DEBUG oslo_vmware.api [None req-6f5d9ea9-7d8d-4c8e-8d3c-edf0333fb67b tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': task-5116029, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.044771] env[65726]: DEBUG oslo_concurrency.lockutils [None req-861b1b19-8cf2-4cb5-a3b7-93ab69618cb7 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Lock "621b899a-e9ec-4fc0-a574-4a08cecff6d9" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 50.374s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 853.224444] env[65726]: DEBUG oslo_vmware.api [None req-4562f28d-bad2-4bca-833c-c069e59f6cc6 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Task: {'id': task-5116028, 'name': ReconfigVM_Task, 'duration_secs': 0.422572} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 853.224636] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-4562f28d-bad2-4bca-833c-c069e59f6cc6 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] [instance: 76249623-6f83-46a3-b8c5-c001111aa698] Reconfigured VM instance instance-0000002e to detach disk 2000 {{(pid=65726) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 853.225631] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad3ccb9e-7e96-4e48-b239-bee6d8a04724 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.256558] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-4562f28d-bad2-4bca-833c-c069e59f6cc6 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] [instance: 76249623-6f83-46a3-b8c5-c001111aa698] Reconfiguring VM instance instance-0000002e to attach disk [datastore1] 76249623-6f83-46a3-b8c5-c001111aa698/76249623-6f83-46a3-b8c5-c001111aa698.vmdk or device None with type thin {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 853.257122] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0b84d769-42fe-44ff-a341-65fbb123e694 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.283189] env[65726]: DEBUG oslo_vmware.api [None req-4562f28d-bad2-4bca-833c-c069e59f6cc6 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Waiting for the task: (returnval){ [ 853.283189] env[65726]: value = "task-5116030" [ 853.283189] env[65726]: _type = "Task" [ 853.283189] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.296276] env[65726]: DEBUG oslo_vmware.api [None req-4562f28d-bad2-4bca-833c-c069e59f6cc6 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Task: {'id': task-5116030, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.346804] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b0e194cf-16dc-40d0-b65c-88bd21cd63e3 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Acquiring lock "refresh_cache-cb8855b6-589d-4863-b86f-7fb4e30fb29e" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 853.346979] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b0e194cf-16dc-40d0-b65c-88bd21cd63e3 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Acquired lock "refresh_cache-cb8855b6-589d-4863-b86f-7fb4e30fb29e" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 853.347176] env[65726]: DEBUG nova.network.neutron [None req-b0e194cf-16dc-40d0-b65c-88bd21cd63e3 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: cb8855b6-589d-4863-b86f-7fb4e30fb29e] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 853.425272] env[65726]: DEBUG oslo_concurrency.lockutils [None req-4826faf0-9908-4f3c-a509-20fe2a708723 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 853.543116] env[65726]: DEBUG oslo_vmware.api [None req-6f5d9ea9-7d8d-4c8e-8d3c-edf0333fb67b tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': task-5116029, 'name': RemoveSnapshot_Task} progress is 97%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.704341] env[65726]: DEBUG nova.compute.manager [req-088bbd09-6865-4b71-b8e9-12724d591aa3 req-8c6b1ea7-3e2d-4085-a9b9-2a3d008b8929 service nova] [instance: cb8855b6-589d-4863-b86f-7fb4e30fb29e] Received event network-vif-plugged-136da8b5-d8fd-4cef-ae44-a9ecc98c28d2 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 853.704929] env[65726]: DEBUG oslo_concurrency.lockutils [req-088bbd09-6865-4b71-b8e9-12724d591aa3 req-8c6b1ea7-3e2d-4085-a9b9-2a3d008b8929 service nova] Acquiring lock "cb8855b6-589d-4863-b86f-7fb4e30fb29e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 853.704929] env[65726]: DEBUG oslo_concurrency.lockutils [req-088bbd09-6865-4b71-b8e9-12724d591aa3 req-8c6b1ea7-3e2d-4085-a9b9-2a3d008b8929 service nova] Lock "cb8855b6-589d-4863-b86f-7fb4e30fb29e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 853.706387] env[65726]: DEBUG oslo_concurrency.lockutils [req-088bbd09-6865-4b71-b8e9-12724d591aa3 req-8c6b1ea7-3e2d-4085-a9b9-2a3d008b8929 service nova] Lock "cb8855b6-589d-4863-b86f-7fb4e30fb29e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 853.706387] env[65726]: DEBUG nova.compute.manager [req-088bbd09-6865-4b71-b8e9-12724d591aa3 req-8c6b1ea7-3e2d-4085-a9b9-2a3d008b8929 service nova] [instance: cb8855b6-589d-4863-b86f-7fb4e30fb29e] No waiting events found dispatching network-vif-plugged-136da8b5-d8fd-4cef-ae44-a9ecc98c28d2 {{(pid=65726) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 853.706387] env[65726]: WARNING nova.compute.manager [req-088bbd09-6865-4b71-b8e9-12724d591aa3 req-8c6b1ea7-3e2d-4085-a9b9-2a3d008b8929 service nova] [instance: cb8855b6-589d-4863-b86f-7fb4e30fb29e] Received unexpected event network-vif-plugged-136da8b5-d8fd-4cef-ae44-a9ecc98c28d2 for instance with vm_state building and task_state spawning. [ 853.796208] env[65726]: DEBUG oslo_vmware.api [None req-4562f28d-bad2-4bca-833c-c069e59f6cc6 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Task: {'id': task-5116030, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.852072] env[65726]: WARNING openstack [None req-b0e194cf-16dc-40d0-b65c-88bd21cd63e3 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 853.852449] env[65726]: WARNING openstack [None req-b0e194cf-16dc-40d0-b65c-88bd21cd63e3 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 853.858587] env[65726]: DEBUG nova.network.neutron [None req-b0e194cf-16dc-40d0-b65c-88bd21cd63e3 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: cb8855b6-589d-4863-b86f-7fb4e30fb29e] Instance cache missing network info. {{(pid=65726) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 853.928682] env[65726]: WARNING openstack [None req-b0e194cf-16dc-40d0-b65c-88bd21cd63e3 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 853.928794] env[65726]: WARNING openstack [None req-b0e194cf-16dc-40d0-b65c-88bd21cd63e3 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 854.043117] env[65726]: DEBUG oslo_vmware.api [None req-6f5d9ea9-7d8d-4c8e-8d3c-edf0333fb67b tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': task-5116029, 'name': RemoveSnapshot_Task, 'duration_secs': 0.670852} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.043414] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-6f5d9ea9-7d8d-4c8e-8d3c-edf0333fb67b tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 01d86089-6b9b-4588-864e-ef91375a1eea] Deleted Snapshot of the VM instance {{(pid=65726) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 854.044050] env[65726]: INFO nova.compute.manager [None req-6f5d9ea9-7d8d-4c8e-8d3c-edf0333fb67b tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 01d86089-6b9b-4588-864e-ef91375a1eea] Took 15.79 seconds to snapshot the instance on the hypervisor. [ 854.073164] env[65726]: WARNING openstack [None req-b0e194cf-16dc-40d0-b65c-88bd21cd63e3 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 854.073235] env[65726]: WARNING openstack [None req-b0e194cf-16dc-40d0-b65c-88bd21cd63e3 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 854.113744] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8061ae4c-8ed4-499a-874a-396de89ad516 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.124417] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be201647-1819-4e43-aaa2-72e3a3903ace {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.159579] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81b8c70d-3fa3-450f-b7fe-b25dd1c5b88f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.168093] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f19ba401-91fc-4cf5-aaaa-98db8cc47c67 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.183413] env[65726]: DEBUG nova.compute.provider_tree [None req-55d74d77-063b-418b-b818-59d05c490f34 tempest-ImagesNegativeTestJSON-1015404931 tempest-ImagesNegativeTestJSON-1015404931-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 854.295835] env[65726]: DEBUG oslo_vmware.api [None req-4562f28d-bad2-4bca-833c-c069e59f6cc6 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Task: {'id': task-5116030, 'name': ReconfigVM_Task, 'duration_secs': 0.631494} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.296147] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-4562f28d-bad2-4bca-833c-c069e59f6cc6 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] [instance: 76249623-6f83-46a3-b8c5-c001111aa698] Reconfigured VM instance instance-0000002e to attach disk [datastore1] 76249623-6f83-46a3-b8c5-c001111aa698/76249623-6f83-46a3-b8c5-c001111aa698.vmdk or device None with type thin {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 854.296929] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-4562f28d-bad2-4bca-833c-c069e59f6cc6 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] [instance: 76249623-6f83-46a3-b8c5-c001111aa698] Updating instance '76249623-6f83-46a3-b8c5-c001111aa698' progress to 50 {{(pid=65726) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 854.403362] env[65726]: DEBUG nova.network.neutron [None req-b0e194cf-16dc-40d0-b65c-88bd21cd63e3 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: cb8855b6-589d-4863-b86f-7fb4e30fb29e] Updating instance_info_cache with network_info: [{"id": "136da8b5-d8fd-4cef-ae44-a9ecc98c28d2", "address": "fa:16:3e:22:7a:70", "network": {"id": "3f7916f7-ee9c-4629-b3b3-1367de9f98ea", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-579713617-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b794c99309e94c1181fc9d8eeeb84702", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d0bdd11b-58af-4cc0-9d38-8322e1bb4e74", "external-id": "nsx-vlan-transportzone-398", "segmentation_id": 398, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap136da8b5-d8", "ovs_interfaceid": "136da8b5-d8fd-4cef-ae44-a9ecc98c28d2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 854.589333] env[65726]: INFO nova.compute.manager [None req-d7ed2ba1-3808-4f50-920c-ef8e5d179423 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] [instance: 44d73b2d-2b6a-4501-9944-432da3c9330d] Rebuilding instance [ 854.661019] env[65726]: DEBUG nova.compute.manager [None req-d7ed2ba1-3808-4f50-920c-ef8e5d179423 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] [instance: 44d73b2d-2b6a-4501-9944-432da3c9330d] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 854.661019] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8af5b581-0d61-4fcb-b142-2137d699ee71 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.685916] env[65726]: DEBUG nova.scheduler.client.report [None req-55d74d77-063b-418b-b818-59d05c490f34 tempest-ImagesNegativeTestJSON-1015404931 tempest-ImagesNegativeTestJSON-1015404931-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 854.803555] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33816abd-0a57-49f2-a1ad-b52ee5999e88 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.824029] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32d30d8d-a12d-4d3f-8041-b3017a944e42 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.843632] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-4562f28d-bad2-4bca-833c-c069e59f6cc6 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] [instance: 76249623-6f83-46a3-b8c5-c001111aa698] Updating instance '76249623-6f83-46a3-b8c5-c001111aa698' progress to 67 {{(pid=65726) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 854.906779] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b0e194cf-16dc-40d0-b65c-88bd21cd63e3 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Releasing lock "refresh_cache-cb8855b6-589d-4863-b86f-7fb4e30fb29e" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 854.907287] env[65726]: DEBUG nova.compute.manager [None req-b0e194cf-16dc-40d0-b65c-88bd21cd63e3 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: cb8855b6-589d-4863-b86f-7fb4e30fb29e] Instance network_info: |[{"id": "136da8b5-d8fd-4cef-ae44-a9ecc98c28d2", "address": "fa:16:3e:22:7a:70", "network": {"id": "3f7916f7-ee9c-4629-b3b3-1367de9f98ea", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-579713617-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b794c99309e94c1181fc9d8eeeb84702", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d0bdd11b-58af-4cc0-9d38-8322e1bb4e74", "external-id": "nsx-vlan-transportzone-398", "segmentation_id": 398, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap136da8b5-d8", "ovs_interfaceid": "136da8b5-d8fd-4cef-ae44-a9ecc98c28d2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 854.907857] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-b0e194cf-16dc-40d0-b65c-88bd21cd63e3 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: cb8855b6-589d-4863-b86f-7fb4e30fb29e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:22:7a:70', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd0bdd11b-58af-4cc0-9d38-8322e1bb4e74', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '136da8b5-d8fd-4cef-ae44-a9ecc98c28d2', 'vif_model': 'vmxnet3'}] {{(pid=65726) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 854.916146] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-b0e194cf-16dc-40d0-b65c-88bd21cd63e3 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 854.916387] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: cb8855b6-589d-4863-b86f-7fb4e30fb29e] Creating VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 854.916664] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e43459d5-04f3-4061-bac3-a0a279f55e34 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.939359] env[65726]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 854.939359] env[65726]: value = "task-5116031" [ 854.939359] env[65726]: _type = "Task" [ 854.939359] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.948243] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116031, 'name': CreateVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.192644] env[65726]: DEBUG oslo_concurrency.lockutils [None req-55d74d77-063b-418b-b818-59d05c490f34 tempest-ImagesNegativeTestJSON-1015404931 tempest-ImagesNegativeTestJSON-1015404931-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.696s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 855.194167] env[65726]: DEBUG nova.compute.manager [None req-55d74d77-063b-418b-b818-59d05c490f34 tempest-ImagesNegativeTestJSON-1015404931 tempest-ImagesNegativeTestJSON-1015404931-project-member] [instance: c4177e20-b1bd-4b54-a275-c93582359a07] Start building networks asynchronously for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 855.197016] env[65726]: DEBUG oslo_concurrency.lockutils [None req-cb2d094a-412f-492d-9ef1-aba8757b7146 tempest-ServerMetadataNegativeTestJSON-936475568 tempest-ServerMetadataNegativeTestJSON-936475568-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 36.572s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 855.199209] env[65726]: INFO nova.compute.claims [None req-cb2d094a-412f-492d-9ef1-aba8757b7146 tempest-ServerMetadataNegativeTestJSON-936475568 tempest-ServerMetadataNegativeTestJSON-936475568-project-member] [instance: d179030a-e5ba-45b1-ad11-4d2f71ed7bd4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 855.350024] env[65726]: WARNING neutronclient.v2_0.client [None req-4562f28d-bad2-4bca-833c-c069e59f6cc6 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 855.397058] env[65726]: DEBUG nova.network.neutron [None req-4562f28d-bad2-4bca-833c-c069e59f6cc6 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] [instance: 76249623-6f83-46a3-b8c5-c001111aa698] Port e97a5da4-5540-4c0c-a33d-be1182e82aa7 binding to destination host cpu-1 is already ACTIVE {{(pid=65726) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3236}} [ 855.450214] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116031, 'name': CreateVM_Task, 'duration_secs': 0.346599} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.450433] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: cb8855b6-589d-4863-b86f-7fb4e30fb29e] Created VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 855.451210] env[65726]: WARNING openstack [None req-b0e194cf-16dc-40d0-b65c-88bd21cd63e3 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 855.451596] env[65726]: WARNING openstack [None req-b0e194cf-16dc-40d0-b65c-88bd21cd63e3 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 855.456517] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b0e194cf-16dc-40d0-b65c-88bd21cd63e3 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 855.456704] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b0e194cf-16dc-40d0-b65c-88bd21cd63e3 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 855.458208] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b0e194cf-16dc-40d0-b65c-88bd21cd63e3 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 855.458208] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-eda78b46-61e3-4342-ad4d-44ccd3be6852 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.462583] env[65726]: DEBUG oslo_vmware.api [None req-b0e194cf-16dc-40d0-b65c-88bd21cd63e3 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Waiting for the task: (returnval){ [ 855.462583] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]526ba4a2-b867-9af3-e2ec-557a5f4a9a68" [ 855.462583] env[65726]: _type = "Task" [ 855.462583] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.473013] env[65726]: DEBUG oslo_vmware.api [None req-b0e194cf-16dc-40d0-b65c-88bd21cd63e3 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]526ba4a2-b867-9af3-e2ec-557a5f4a9a68, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.570650] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e5e88d5e-aa68-4ff5-9280-b2cfe1d14b27 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Acquiring lock "5a252ef1-93c9-4bff-842b-b64df2bc5d75" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 855.571102] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e5e88d5e-aa68-4ff5-9280-b2cfe1d14b27 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Lock "5a252ef1-93c9-4bff-842b-b64df2bc5d75" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 855.679049] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-d7ed2ba1-3808-4f50-920c-ef8e5d179423 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] [instance: 44d73b2d-2b6a-4501-9944-432da3c9330d] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 855.679049] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-499d8d95-3fdc-4c56-b582-e866e59840e1 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.687716] env[65726]: DEBUG oslo_vmware.api [None req-d7ed2ba1-3808-4f50-920c-ef8e5d179423 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Waiting for the task: (returnval){ [ 855.687716] env[65726]: value = "task-5116032" [ 855.687716] env[65726]: _type = "Task" [ 855.687716] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.697539] env[65726]: DEBUG oslo_vmware.api [None req-d7ed2ba1-3808-4f50-920c-ef8e5d179423 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Task: {'id': task-5116032, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.710807] env[65726]: DEBUG nova.compute.utils [None req-55d74d77-063b-418b-b818-59d05c490f34 tempest-ImagesNegativeTestJSON-1015404931 tempest-ImagesNegativeTestJSON-1015404931-project-member] Using /dev/sd instead of None {{(pid=65726) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 855.710807] env[65726]: DEBUG nova.compute.manager [None req-55d74d77-063b-418b-b818-59d05c490f34 tempest-ImagesNegativeTestJSON-1015404931 tempest-ImagesNegativeTestJSON-1015404931-project-member] [instance: c4177e20-b1bd-4b54-a275-c93582359a07] Allocating IP information in the background. {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 855.710807] env[65726]: DEBUG nova.network.neutron [None req-55d74d77-063b-418b-b818-59d05c490f34 tempest-ImagesNegativeTestJSON-1015404931 tempest-ImagesNegativeTestJSON-1015404931-project-member] [instance: c4177e20-b1bd-4b54-a275-c93582359a07] allocate_for_instance() {{(pid=65726) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 855.710807] env[65726]: WARNING neutronclient.v2_0.client [None req-55d74d77-063b-418b-b818-59d05c490f34 tempest-ImagesNegativeTestJSON-1015404931 tempest-ImagesNegativeTestJSON-1015404931-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 855.710807] env[65726]: WARNING neutronclient.v2_0.client [None req-55d74d77-063b-418b-b818-59d05c490f34 tempest-ImagesNegativeTestJSON-1015404931 tempest-ImagesNegativeTestJSON-1015404931-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 855.710999] env[65726]: WARNING openstack [None req-55d74d77-063b-418b-b818-59d05c490f34 tempest-ImagesNegativeTestJSON-1015404931 tempest-ImagesNegativeTestJSON-1015404931-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 855.711243] env[65726]: WARNING openstack [None req-55d74d77-063b-418b-b818-59d05c490f34 tempest-ImagesNegativeTestJSON-1015404931 tempest-ImagesNegativeTestJSON-1015404931-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 855.771261] env[65726]: DEBUG nova.policy [None req-55d74d77-063b-418b-b818-59d05c490f34 tempest-ImagesNegativeTestJSON-1015404931 tempest-ImagesNegativeTestJSON-1015404931-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0f2d1a7cc63c43bdbedfbe0c1c5bf768', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '28432176889e4680aa34fce120b47304', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65726) authorize /opt/stack/nova/nova/policy.py:192}} [ 855.782975] env[65726]: DEBUG oslo_concurrency.lockutils [None req-89fee262-eea7-4ed4-83cb-eb9998cc3059 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Acquiring lock "6f91b053-772a-4497-b29d-349b960c55eb" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 855.783445] env[65726]: DEBUG oslo_concurrency.lockutils [None req-89fee262-eea7-4ed4-83cb-eb9998cc3059 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Lock "6f91b053-772a-4497-b29d-349b960c55eb" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 855.975750] env[65726]: DEBUG oslo_vmware.api [None req-b0e194cf-16dc-40d0-b65c-88bd21cd63e3 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]526ba4a2-b867-9af3-e2ec-557a5f4a9a68, 'name': SearchDatastore_Task, 'duration_secs': 0.009553} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.975750] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b0e194cf-16dc-40d0-b65c-88bd21cd63e3 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 855.975750] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-b0e194cf-16dc-40d0-b65c-88bd21cd63e3 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: cb8855b6-589d-4863-b86f-7fb4e30fb29e] Processing image b52362a3-ee8a-4cbf-b06f-513b0cc8f95c {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 855.975750] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b0e194cf-16dc-40d0-b65c-88bd21cd63e3 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 855.975750] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b0e194cf-16dc-40d0-b65c-88bd21cd63e3 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 855.975750] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-b0e194cf-16dc-40d0-b65c-88bd21cd63e3 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 855.975750] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0a66cb58-5305-4304-a053-a86010d9ca50 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.988021] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-b0e194cf-16dc-40d0-b65c-88bd21cd63e3 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 855.988021] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-b0e194cf-16dc-40d0-b65c-88bd21cd63e3 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65726) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 855.988021] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-581dbdba-c6aa-4519-86c2-b181623f2b10 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.995364] env[65726]: DEBUG oslo_vmware.api [None req-b0e194cf-16dc-40d0-b65c-88bd21cd63e3 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Waiting for the task: (returnval){ [ 855.995364] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5247608f-182b-b209-dcaf-0e41fed32bbf" [ 855.995364] env[65726]: _type = "Task" [ 855.995364] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.005675] env[65726]: DEBUG oslo_vmware.api [None req-b0e194cf-16dc-40d0-b65c-88bd21cd63e3 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5247608f-182b-b209-dcaf-0e41fed32bbf, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.059257] env[65726]: DEBUG nova.network.neutron [None req-55d74d77-063b-418b-b818-59d05c490f34 tempest-ImagesNegativeTestJSON-1015404931 tempest-ImagesNegativeTestJSON-1015404931-project-member] [instance: c4177e20-b1bd-4b54-a275-c93582359a07] Successfully created port: da625869-f814-4733-b2ac-d50cc54a9858 {{(pid=65726) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 856.102133] env[65726]: DEBUG nova.compute.manager [req-fb1b322a-f0df-409c-81d6-39db4210390b req-894cb4a3-3f45-42bd-9a4e-507bb0474fda service nova] [instance: cb8855b6-589d-4863-b86f-7fb4e30fb29e] Received event network-changed-136da8b5-d8fd-4cef-ae44-a9ecc98c28d2 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 856.103032] env[65726]: DEBUG nova.compute.manager [req-fb1b322a-f0df-409c-81d6-39db4210390b req-894cb4a3-3f45-42bd-9a4e-507bb0474fda service nova] [instance: cb8855b6-589d-4863-b86f-7fb4e30fb29e] Refreshing instance network info cache due to event network-changed-136da8b5-d8fd-4cef-ae44-a9ecc98c28d2. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 856.103032] env[65726]: DEBUG oslo_concurrency.lockutils [req-fb1b322a-f0df-409c-81d6-39db4210390b req-894cb4a3-3f45-42bd-9a4e-507bb0474fda service nova] Acquiring lock "refresh_cache-cb8855b6-589d-4863-b86f-7fb4e30fb29e" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 856.103032] env[65726]: DEBUG oslo_concurrency.lockutils [req-fb1b322a-f0df-409c-81d6-39db4210390b req-894cb4a3-3f45-42bd-9a4e-507bb0474fda service nova] Acquired lock "refresh_cache-cb8855b6-589d-4863-b86f-7fb4e30fb29e" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 856.103032] env[65726]: DEBUG nova.network.neutron [req-fb1b322a-f0df-409c-81d6-39db4210390b req-894cb4a3-3f45-42bd-9a4e-507bb0474fda service nova] [instance: cb8855b6-589d-4863-b86f-7fb4e30fb29e] Refreshing network info cache for port 136da8b5-d8fd-4cef-ae44-a9ecc98c28d2 {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 856.199803] env[65726]: DEBUG oslo_vmware.api [None req-d7ed2ba1-3808-4f50-920c-ef8e5d179423 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Task: {'id': task-5116032, 'name': PowerOffVM_Task, 'duration_secs': 0.16654} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.200325] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-d7ed2ba1-3808-4f50-920c-ef8e5d179423 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] [instance: 44d73b2d-2b6a-4501-9944-432da3c9330d] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 856.200731] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-d7ed2ba1-3808-4f50-920c-ef8e5d179423 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] [instance: 44d73b2d-2b6a-4501-9944-432da3c9330d] Destroying instance {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 856.201947] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c77c4857-26cc-429a-9e9a-28b3b346b971 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.213815] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-d7ed2ba1-3808-4f50-920c-ef8e5d179423 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] [instance: 44d73b2d-2b6a-4501-9944-432da3c9330d] Unregistering the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 856.214666] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-bd1aae60-8b52-48ca-8255-c6734e4a6d6d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.220060] env[65726]: DEBUG nova.compute.manager [None req-55d74d77-063b-418b-b818-59d05c490f34 tempest-ImagesNegativeTestJSON-1015404931 tempest-ImagesNegativeTestJSON-1015404931-project-member] [instance: c4177e20-b1bd-4b54-a275-c93582359a07] Start building block device mappings for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 856.245023] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-d7ed2ba1-3808-4f50-920c-ef8e5d179423 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] [instance: 44d73b2d-2b6a-4501-9944-432da3c9330d] Unregistered the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 856.245023] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-d7ed2ba1-3808-4f50-920c-ef8e5d179423 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] [instance: 44d73b2d-2b6a-4501-9944-432da3c9330d] Deleting contents of the VM from datastore datastore2 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 856.245410] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-d7ed2ba1-3808-4f50-920c-ef8e5d179423 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Deleting the datastore file [datastore2] 44d73b2d-2b6a-4501-9944-432da3c9330d {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 856.246052] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b30c26e4-ffa8-4608-9a8d-b7fee6d06b47 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.263158] env[65726]: DEBUG oslo_vmware.api [None req-d7ed2ba1-3808-4f50-920c-ef8e5d179423 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Waiting for the task: (returnval){ [ 856.263158] env[65726]: value = "task-5116034" [ 856.263158] env[65726]: _type = "Task" [ 856.263158] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.277824] env[65726]: DEBUG oslo_vmware.api [None req-d7ed2ba1-3808-4f50-920c-ef8e5d179423 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Task: {'id': task-5116034, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.420294] env[65726]: DEBUG oslo_concurrency.lockutils [None req-4562f28d-bad2-4bca-833c-c069e59f6cc6 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Acquiring lock "76249623-6f83-46a3-b8c5-c001111aa698-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 856.420542] env[65726]: DEBUG oslo_concurrency.lockutils [None req-4562f28d-bad2-4bca-833c-c069e59f6cc6 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Lock "76249623-6f83-46a3-b8c5-c001111aa698-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 856.420749] env[65726]: DEBUG oslo_concurrency.lockutils [None req-4562f28d-bad2-4bca-833c-c069e59f6cc6 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Lock "76249623-6f83-46a3-b8c5-c001111aa698-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 856.506581] env[65726]: DEBUG oslo_vmware.api [None req-b0e194cf-16dc-40d0-b65c-88bd21cd63e3 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5247608f-182b-b209-dcaf-0e41fed32bbf, 'name': SearchDatastore_Task, 'duration_secs': 0.040819} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.509954] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-76bf392f-6bab-4354-bbc8-03d4d5cad896 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.515486] env[65726]: DEBUG oslo_vmware.api [None req-b0e194cf-16dc-40d0-b65c-88bd21cd63e3 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Waiting for the task: (returnval){ [ 856.515486] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]528c2a77-80df-e819-6d46-e3b7fe76ee0f" [ 856.515486] env[65726]: _type = "Task" [ 856.515486] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.524880] env[65726]: DEBUG oslo_vmware.api [None req-b0e194cf-16dc-40d0-b65c-88bd21cd63e3 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]528c2a77-80df-e819-6d46-e3b7fe76ee0f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.606284] env[65726]: WARNING openstack [req-fb1b322a-f0df-409c-81d6-39db4210390b req-894cb4a3-3f45-42bd-9a4e-507bb0474fda service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 856.606609] env[65726]: WARNING openstack [req-fb1b322a-f0df-409c-81d6-39db4210390b req-894cb4a3-3f45-42bd-9a4e-507bb0474fda service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 856.761984] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b6cc6f3-f496-4a77-8f1c-d494e7e00066 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.773737] env[65726]: DEBUG oslo_vmware.api [None req-d7ed2ba1-3808-4f50-920c-ef8e5d179423 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Task: {'id': task-5116034, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.116509} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.776239] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-d7ed2ba1-3808-4f50-920c-ef8e5d179423 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Deleted the datastore file {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 856.776480] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-d7ed2ba1-3808-4f50-920c-ef8e5d179423 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] [instance: 44d73b2d-2b6a-4501-9944-432da3c9330d] Deleted contents of the VM from datastore datastore2 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 856.776658] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-d7ed2ba1-3808-4f50-920c-ef8e5d179423 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] [instance: 44d73b2d-2b6a-4501-9944-432da3c9330d] Instance destroyed {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 856.780673] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ddea0862-3ca3-418b-a0fb-4fa992138639 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.819416] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1aaf10b4-fe95-4da5-98b3-2a93e9377ee9 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.828663] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88939c86-eb21-405e-8517-f10a2ad40d2f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.845980] env[65726]: DEBUG nova.compute.provider_tree [None req-cb2d094a-412f-492d-9ef1-aba8757b7146 tempest-ServerMetadataNegativeTestJSON-936475568 tempest-ServerMetadataNegativeTestJSON-936475568-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 856.848392] env[65726]: WARNING openstack [req-fb1b322a-f0df-409c-81d6-39db4210390b req-894cb4a3-3f45-42bd-9a4e-507bb0474fda service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 856.848777] env[65726]: WARNING openstack [req-fb1b322a-f0df-409c-81d6-39db4210390b req-894cb4a3-3f45-42bd-9a4e-507bb0474fda service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 856.976129] env[65726]: WARNING openstack [req-fb1b322a-f0df-409c-81d6-39db4210390b req-894cb4a3-3f45-42bd-9a4e-507bb0474fda service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 856.976492] env[65726]: WARNING openstack [req-fb1b322a-f0df-409c-81d6-39db4210390b req-894cb4a3-3f45-42bd-9a4e-507bb0474fda service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 857.027493] env[65726]: DEBUG oslo_vmware.api [None req-b0e194cf-16dc-40d0-b65c-88bd21cd63e3 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]528c2a77-80df-e819-6d46-e3b7fe76ee0f, 'name': SearchDatastore_Task, 'duration_secs': 0.009775} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.028128] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b0e194cf-16dc-40d0-b65c-88bd21cd63e3 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 857.028255] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-b0e194cf-16dc-40d0-b65c-88bd21cd63e3 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] cb8855b6-589d-4863-b86f-7fb4e30fb29e/cb8855b6-589d-4863-b86f-7fb4e30fb29e.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 857.028571] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e9dcd986-f615-43c2-b55e-1e1266376318 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.038848] env[65726]: DEBUG oslo_vmware.api [None req-b0e194cf-16dc-40d0-b65c-88bd21cd63e3 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Waiting for the task: (returnval){ [ 857.038848] env[65726]: value = "task-5116035" [ 857.038848] env[65726]: _type = "Task" [ 857.038848] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.049724] env[65726]: DEBUG oslo_vmware.api [None req-b0e194cf-16dc-40d0-b65c-88bd21cd63e3 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Task: {'id': task-5116035, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.082554] env[65726]: DEBUG nova.network.neutron [req-fb1b322a-f0df-409c-81d6-39db4210390b req-894cb4a3-3f45-42bd-9a4e-507bb0474fda service nova] [instance: cb8855b6-589d-4863-b86f-7fb4e30fb29e] Updated VIF entry in instance network info cache for port 136da8b5-d8fd-4cef-ae44-a9ecc98c28d2. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 857.082933] env[65726]: DEBUG nova.network.neutron [req-fb1b322a-f0df-409c-81d6-39db4210390b req-894cb4a3-3f45-42bd-9a4e-507bb0474fda service nova] [instance: cb8855b6-589d-4863-b86f-7fb4e30fb29e] Updating instance_info_cache with network_info: [{"id": "136da8b5-d8fd-4cef-ae44-a9ecc98c28d2", "address": "fa:16:3e:22:7a:70", "network": {"id": "3f7916f7-ee9c-4629-b3b3-1367de9f98ea", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-579713617-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b794c99309e94c1181fc9d8eeeb84702", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d0bdd11b-58af-4cc0-9d38-8322e1bb4e74", "external-id": "nsx-vlan-transportzone-398", "segmentation_id": 398, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap136da8b5-d8", "ovs_interfaceid": "136da8b5-d8fd-4cef-ae44-a9ecc98c28d2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 857.234990] env[65726]: DEBUG nova.compute.manager [None req-55d74d77-063b-418b-b818-59d05c490f34 tempest-ImagesNegativeTestJSON-1015404931 tempest-ImagesNegativeTestJSON-1015404931-project-member] [instance: c4177e20-b1bd-4b54-a275-c93582359a07] Start spawning the instance on the hypervisor. {{(pid=65726) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 857.262723] env[65726]: DEBUG nova.virt.hardware [None req-55d74d77-063b-418b-b818-59d05c490f34 tempest-ImagesNegativeTestJSON-1015404931 tempest-ImagesNegativeTestJSON-1015404931-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T19:28:34Z,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1539f06fa3534e90acbc3a60c4b8bd3f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T19:28:35Z,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 857.262984] env[65726]: DEBUG nova.virt.hardware [None req-55d74d77-063b-418b-b818-59d05c490f34 tempest-ImagesNegativeTestJSON-1015404931 tempest-ImagesNegativeTestJSON-1015404931-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 857.263129] env[65726]: DEBUG nova.virt.hardware [None req-55d74d77-063b-418b-b818-59d05c490f34 tempest-ImagesNegativeTestJSON-1015404931 tempest-ImagesNegativeTestJSON-1015404931-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 857.263312] env[65726]: DEBUG nova.virt.hardware [None req-55d74d77-063b-418b-b818-59d05c490f34 tempest-ImagesNegativeTestJSON-1015404931 tempest-ImagesNegativeTestJSON-1015404931-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 857.263452] env[65726]: DEBUG nova.virt.hardware [None req-55d74d77-063b-418b-b818-59d05c490f34 tempest-ImagesNegativeTestJSON-1015404931 tempest-ImagesNegativeTestJSON-1015404931-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 857.263689] env[65726]: DEBUG nova.virt.hardware [None req-55d74d77-063b-418b-b818-59d05c490f34 tempest-ImagesNegativeTestJSON-1015404931 tempest-ImagesNegativeTestJSON-1015404931-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 857.263812] env[65726]: DEBUG nova.virt.hardware [None req-55d74d77-063b-418b-b818-59d05c490f34 tempest-ImagesNegativeTestJSON-1015404931 tempest-ImagesNegativeTestJSON-1015404931-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 857.263949] env[65726]: DEBUG nova.virt.hardware [None req-55d74d77-063b-418b-b818-59d05c490f34 tempest-ImagesNegativeTestJSON-1015404931 tempest-ImagesNegativeTestJSON-1015404931-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 857.264568] env[65726]: DEBUG nova.virt.hardware [None req-55d74d77-063b-418b-b818-59d05c490f34 tempest-ImagesNegativeTestJSON-1015404931 tempest-ImagesNegativeTestJSON-1015404931-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 857.264775] env[65726]: DEBUG nova.virt.hardware [None req-55d74d77-063b-418b-b818-59d05c490f34 tempest-ImagesNegativeTestJSON-1015404931 tempest-ImagesNegativeTestJSON-1015404931-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 857.264953] env[65726]: DEBUG nova.virt.hardware [None req-55d74d77-063b-418b-b818-59d05c490f34 tempest-ImagesNegativeTestJSON-1015404931 tempest-ImagesNegativeTestJSON-1015404931-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 857.265938] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8413d90-3376-4d08-a77f-f7a18ee455f0 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.275763] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d206b782-3088-4331-9447-eaf4bea5d5b7 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.356637] env[65726]: DEBUG nova.scheduler.client.report [None req-cb2d094a-412f-492d-9ef1-aba8757b7146 tempest-ServerMetadataNegativeTestJSON-936475568 tempest-ServerMetadataNegativeTestJSON-936475568-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 857.427999] env[65726]: WARNING openstack [None req-4562f28d-bad2-4bca-833c-c069e59f6cc6 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 857.428514] env[65726]: WARNING openstack [None req-4562f28d-bad2-4bca-833c-c069e59f6cc6 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 857.499010] env[65726]: DEBUG oslo_concurrency.lockutils [None req-4562f28d-bad2-4bca-833c-c069e59f6cc6 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Acquiring lock "refresh_cache-76249623-6f83-46a3-b8c5-c001111aa698" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 857.499392] env[65726]: DEBUG oslo_concurrency.lockutils [None req-4562f28d-bad2-4bca-833c-c069e59f6cc6 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Acquired lock "refresh_cache-76249623-6f83-46a3-b8c5-c001111aa698" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 857.499696] env[65726]: DEBUG nova.network.neutron [None req-4562f28d-bad2-4bca-833c-c069e59f6cc6 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] [instance: 76249623-6f83-46a3-b8c5-c001111aa698] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 857.550033] env[65726]: DEBUG oslo_vmware.api [None req-b0e194cf-16dc-40d0-b65c-88bd21cd63e3 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Task: {'id': task-5116035, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.587340] env[65726]: DEBUG oslo_concurrency.lockutils [req-fb1b322a-f0df-409c-81d6-39db4210390b req-894cb4a3-3f45-42bd-9a4e-507bb0474fda service nova] Releasing lock "refresh_cache-cb8855b6-589d-4863-b86f-7fb4e30fb29e" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 857.682328] env[65726]: DEBUG oslo_concurrency.lockutils [None req-17b5a1c7-4bc8-41d1-988f-c933bb6421ad tempest-ServersTestManualDisk-1753452859 tempest-ServersTestManualDisk-1753452859-project-member] Acquiring lock "3088ee5e-0d27-4058-a94b-2e04c3b52add" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 857.682568] env[65726]: DEBUG oslo_concurrency.lockutils [None req-17b5a1c7-4bc8-41d1-988f-c933bb6421ad tempest-ServersTestManualDisk-1753452859 tempest-ServersTestManualDisk-1753452859-project-member] Lock "3088ee5e-0d27-4058-a94b-2e04c3b52add" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 857.711725] env[65726]: DEBUG nova.network.neutron [None req-55d74d77-063b-418b-b818-59d05c490f34 tempest-ImagesNegativeTestJSON-1015404931 tempest-ImagesNegativeTestJSON-1015404931-project-member] [instance: c4177e20-b1bd-4b54-a275-c93582359a07] Successfully updated port: da625869-f814-4733-b2ac-d50cc54a9858 {{(pid=65726) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 857.819576] env[65726]: DEBUG nova.compute.manager [req-eb6d5f6d-02a7-41b6-93d9-4eae03f2969e req-3f47babc-1193-4534-9458-7848ae19eae5 service nova] [instance: c4177e20-b1bd-4b54-a275-c93582359a07] Received event network-vif-plugged-da625869-f814-4733-b2ac-d50cc54a9858 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 857.819576] env[65726]: DEBUG oslo_concurrency.lockutils [req-eb6d5f6d-02a7-41b6-93d9-4eae03f2969e req-3f47babc-1193-4534-9458-7848ae19eae5 service nova] Acquiring lock "c4177e20-b1bd-4b54-a275-c93582359a07-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 857.819576] env[65726]: DEBUG oslo_concurrency.lockutils [req-eb6d5f6d-02a7-41b6-93d9-4eae03f2969e req-3f47babc-1193-4534-9458-7848ae19eae5 service nova] Lock "c4177e20-b1bd-4b54-a275-c93582359a07-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 857.819576] env[65726]: DEBUG oslo_concurrency.lockutils [req-eb6d5f6d-02a7-41b6-93d9-4eae03f2969e req-3f47babc-1193-4534-9458-7848ae19eae5 service nova] Lock "c4177e20-b1bd-4b54-a275-c93582359a07-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 857.819576] env[65726]: DEBUG nova.compute.manager [req-eb6d5f6d-02a7-41b6-93d9-4eae03f2969e req-3f47babc-1193-4534-9458-7848ae19eae5 service nova] [instance: c4177e20-b1bd-4b54-a275-c93582359a07] No waiting events found dispatching network-vif-plugged-da625869-f814-4733-b2ac-d50cc54a9858 {{(pid=65726) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 857.820539] env[65726]: WARNING nova.compute.manager [req-eb6d5f6d-02a7-41b6-93d9-4eae03f2969e req-3f47babc-1193-4534-9458-7848ae19eae5 service nova] [instance: c4177e20-b1bd-4b54-a275-c93582359a07] Received unexpected event network-vif-plugged-da625869-f814-4733-b2ac-d50cc54a9858 for instance with vm_state building and task_state spawning. [ 857.837949] env[65726]: DEBUG nova.virt.hardware [None req-d7ed2ba1-3808-4f50-920c-ef8e5d179423 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T19:28:34Z,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1539f06fa3534e90acbc3a60c4b8bd3f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T19:28:35Z,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 857.838966] env[65726]: DEBUG nova.virt.hardware [None req-d7ed2ba1-3808-4f50-920c-ef8e5d179423 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 857.839063] env[65726]: DEBUG nova.virt.hardware [None req-d7ed2ba1-3808-4f50-920c-ef8e5d179423 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 857.839190] env[65726]: DEBUG nova.virt.hardware [None req-d7ed2ba1-3808-4f50-920c-ef8e5d179423 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 857.839350] env[65726]: DEBUG nova.virt.hardware [None req-d7ed2ba1-3808-4f50-920c-ef8e5d179423 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 857.839495] env[65726]: DEBUG nova.virt.hardware [None req-d7ed2ba1-3808-4f50-920c-ef8e5d179423 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 857.839737] env[65726]: DEBUG nova.virt.hardware [None req-d7ed2ba1-3808-4f50-920c-ef8e5d179423 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 857.839919] env[65726]: DEBUG nova.virt.hardware [None req-d7ed2ba1-3808-4f50-920c-ef8e5d179423 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 857.840105] env[65726]: DEBUG nova.virt.hardware [None req-d7ed2ba1-3808-4f50-920c-ef8e5d179423 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 857.840271] env[65726]: DEBUG nova.virt.hardware [None req-d7ed2ba1-3808-4f50-920c-ef8e5d179423 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 857.840490] env[65726]: DEBUG nova.virt.hardware [None req-d7ed2ba1-3808-4f50-920c-ef8e5d179423 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 857.841637] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52d8689a-41aa-46a8-92a2-bcfc3f1173a4 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.850131] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49925242-831e-4605-977f-c007b4e7267f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.865593] env[65726]: DEBUG oslo_concurrency.lockutils [None req-cb2d094a-412f-492d-9ef1-aba8757b7146 tempest-ServerMetadataNegativeTestJSON-936475568 tempest-ServerMetadataNegativeTestJSON-936475568-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.669s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 857.866151] env[65726]: DEBUG nova.compute.manager [None req-cb2d094a-412f-492d-9ef1-aba8757b7146 tempest-ServerMetadataNegativeTestJSON-936475568 tempest-ServerMetadataNegativeTestJSON-936475568-project-member] [instance: d179030a-e5ba-45b1-ad11-4d2f71ed7bd4] Start building networks asynchronously for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 857.869048] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-d7ed2ba1-3808-4f50-920c-ef8e5d179423 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] [instance: 44d73b2d-2b6a-4501-9944-432da3c9330d] Instance VIF info [] {{(pid=65726) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 857.875246] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-d7ed2ba1-3808-4f50-920c-ef8e5d179423 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 857.876747] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b3ea747b-af8b-4592-b662-6643bc7f39d3 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 34.989s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 857.878120] env[65726]: INFO nova.compute.claims [None req-b3ea747b-af8b-4592-b662-6643bc7f39d3 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] [instance: 2cbee359-a1fc-45c0-bcc5-a2c24ded9c69] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 857.880108] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 44d73b2d-2b6a-4501-9944-432da3c9330d] Creating VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 857.880651] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-04b7359a-f72c-4dd7-a7d3-8cf10ad1e3e2 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.901248] env[65726]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 857.901248] env[65726]: value = "task-5116036" [ 857.901248] env[65726]: _type = "Task" [ 857.901248] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.908329] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116036, 'name': CreateVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.004058] env[65726]: WARNING openstack [None req-4562f28d-bad2-4bca-833c-c069e59f6cc6 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 858.004058] env[65726]: WARNING openstack [None req-4562f28d-bad2-4bca-833c-c069e59f6cc6 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 858.055058] env[65726]: DEBUG oslo_vmware.api [None req-b0e194cf-16dc-40d0-b65c-88bd21cd63e3 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Task: {'id': task-5116035, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.105270] env[65726]: WARNING openstack [None req-4562f28d-bad2-4bca-833c-c069e59f6cc6 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 858.105746] env[65726]: WARNING openstack [None req-4562f28d-bad2-4bca-833c-c069e59f6cc6 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 858.188273] env[65726]: WARNING openstack [None req-4562f28d-bad2-4bca-833c-c069e59f6cc6 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 858.189599] env[65726]: WARNING openstack [None req-4562f28d-bad2-4bca-833c-c069e59f6cc6 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 858.214964] env[65726]: DEBUG oslo_concurrency.lockutils [None req-55d74d77-063b-418b-b818-59d05c490f34 tempest-ImagesNegativeTestJSON-1015404931 tempest-ImagesNegativeTestJSON-1015404931-project-member] Acquiring lock "refresh_cache-c4177e20-b1bd-4b54-a275-c93582359a07" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 858.215276] env[65726]: DEBUG oslo_concurrency.lockutils [None req-55d74d77-063b-418b-b818-59d05c490f34 tempest-ImagesNegativeTestJSON-1015404931 tempest-ImagesNegativeTestJSON-1015404931-project-member] Acquired lock "refresh_cache-c4177e20-b1bd-4b54-a275-c93582359a07" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 858.215408] env[65726]: DEBUG nova.network.neutron [None req-55d74d77-063b-418b-b818-59d05c490f34 tempest-ImagesNegativeTestJSON-1015404931 tempest-ImagesNegativeTestJSON-1015404931-project-member] [instance: c4177e20-b1bd-4b54-a275-c93582359a07] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 858.296450] env[65726]: DEBUG nova.network.neutron [None req-4562f28d-bad2-4bca-833c-c069e59f6cc6 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] [instance: 76249623-6f83-46a3-b8c5-c001111aa698] Updating instance_info_cache with network_info: [{"id": "e97a5da4-5540-4c0c-a33d-be1182e82aa7", "address": "fa:16:3e:af:b6:0f", "network": {"id": "55d629b8-a541-4720-9289-927bbc6e6f22", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.253", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "1539f06fa3534e90acbc3a60c4b8bd3f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "35342bcb-8b06-472e-b3c0-43fd3d6c4b30", "external-id": "nsx-vlan-transportzone-524", "segmentation_id": 524, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape97a5da4-55", "ovs_interfaceid": "e97a5da4-5540-4c0c-a33d-be1182e82aa7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 858.377268] env[65726]: DEBUG nova.compute.utils [None req-cb2d094a-412f-492d-9ef1-aba8757b7146 tempest-ServerMetadataNegativeTestJSON-936475568 tempest-ServerMetadataNegativeTestJSON-936475568-project-member] Using /dev/sd instead of None {{(pid=65726) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 858.379296] env[65726]: DEBUG nova.compute.manager [None req-cb2d094a-412f-492d-9ef1-aba8757b7146 tempest-ServerMetadataNegativeTestJSON-936475568 tempest-ServerMetadataNegativeTestJSON-936475568-project-member] [instance: d179030a-e5ba-45b1-ad11-4d2f71ed7bd4] Allocating IP information in the background. {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 858.379499] env[65726]: DEBUG nova.network.neutron [None req-cb2d094a-412f-492d-9ef1-aba8757b7146 tempest-ServerMetadataNegativeTestJSON-936475568 tempest-ServerMetadataNegativeTestJSON-936475568-project-member] [instance: d179030a-e5ba-45b1-ad11-4d2f71ed7bd4] allocate_for_instance() {{(pid=65726) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 858.379826] env[65726]: WARNING neutronclient.v2_0.client [None req-cb2d094a-412f-492d-9ef1-aba8757b7146 tempest-ServerMetadataNegativeTestJSON-936475568 tempest-ServerMetadataNegativeTestJSON-936475568-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 858.380234] env[65726]: WARNING neutronclient.v2_0.client [None req-cb2d094a-412f-492d-9ef1-aba8757b7146 tempest-ServerMetadataNegativeTestJSON-936475568 tempest-ServerMetadataNegativeTestJSON-936475568-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 858.381117] env[65726]: WARNING openstack [None req-cb2d094a-412f-492d-9ef1-aba8757b7146 tempest-ServerMetadataNegativeTestJSON-936475568 tempest-ServerMetadataNegativeTestJSON-936475568-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 858.381472] env[65726]: WARNING openstack [None req-cb2d094a-412f-492d-9ef1-aba8757b7146 tempest-ServerMetadataNegativeTestJSON-936475568 tempest-ServerMetadataNegativeTestJSON-936475568-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 858.410874] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116036, 'name': CreateVM_Task, 'duration_secs': 0.340255} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 858.411033] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 44d73b2d-2b6a-4501-9944-432da3c9330d] Created VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 858.411463] env[65726]: DEBUG oslo_concurrency.lockutils [None req-d7ed2ba1-3808-4f50-920c-ef8e5d179423 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 858.411628] env[65726]: DEBUG oslo_concurrency.lockutils [None req-d7ed2ba1-3808-4f50-920c-ef8e5d179423 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 858.411942] env[65726]: DEBUG oslo_concurrency.lockutils [None req-d7ed2ba1-3808-4f50-920c-ef8e5d179423 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 858.412442] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-79754e50-cbba-4f0a-9ee3-03062901496c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.417921] env[65726]: DEBUG oslo_vmware.api [None req-d7ed2ba1-3808-4f50-920c-ef8e5d179423 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Waiting for the task: (returnval){ [ 858.417921] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]528aea06-9094-8a75-cbf9-cc19dffaeaf7" [ 858.417921] env[65726]: _type = "Task" [ 858.417921] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.426517] env[65726]: DEBUG oslo_vmware.api [None req-d7ed2ba1-3808-4f50-920c-ef8e5d179423 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]528aea06-9094-8a75-cbf9-cc19dffaeaf7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.431539] env[65726]: DEBUG nova.policy [None req-cb2d094a-412f-492d-9ef1-aba8757b7146 tempest-ServerMetadataNegativeTestJSON-936475568 tempest-ServerMetadataNegativeTestJSON-936475568-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '345b2074c2eb49009e9d90a01032c1b5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '921d46aee1824fd990022c7f851018b2', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65726) authorize /opt/stack/nova/nova/policy.py:192}} [ 858.555372] env[65726]: DEBUG oslo_vmware.api [None req-b0e194cf-16dc-40d0-b65c-88bd21cd63e3 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Task: {'id': task-5116035, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.670864] env[65726]: DEBUG nova.network.neutron [None req-cb2d094a-412f-492d-9ef1-aba8757b7146 tempest-ServerMetadataNegativeTestJSON-936475568 tempest-ServerMetadataNegativeTestJSON-936475568-project-member] [instance: d179030a-e5ba-45b1-ad11-4d2f71ed7bd4] Successfully created port: 51e75557-8c53-4ba4-982a-42e90b5decef {{(pid=65726) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 858.722178] env[65726]: WARNING openstack [None req-55d74d77-063b-418b-b818-59d05c490f34 tempest-ImagesNegativeTestJSON-1015404931 tempest-ImagesNegativeTestJSON-1015404931-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 858.722178] env[65726]: WARNING openstack [None req-55d74d77-063b-418b-b818-59d05c490f34 tempest-ImagesNegativeTestJSON-1015404931 tempest-ImagesNegativeTestJSON-1015404931-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 858.726570] env[65726]: DEBUG nova.network.neutron [None req-55d74d77-063b-418b-b818-59d05c490f34 tempest-ImagesNegativeTestJSON-1015404931 tempest-ImagesNegativeTestJSON-1015404931-project-member] [instance: c4177e20-b1bd-4b54-a275-c93582359a07] Instance cache missing network info. {{(pid=65726) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 858.799241] env[65726]: DEBUG oslo_concurrency.lockutils [None req-4562f28d-bad2-4bca-833c-c069e59f6cc6 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Releasing lock "refresh_cache-76249623-6f83-46a3-b8c5-c001111aa698" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 858.840317] env[65726]: WARNING openstack [None req-55d74d77-063b-418b-b818-59d05c490f34 tempest-ImagesNegativeTestJSON-1015404931 tempest-ImagesNegativeTestJSON-1015404931-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 858.840593] env[65726]: WARNING openstack [None req-55d74d77-063b-418b-b818-59d05c490f34 tempest-ImagesNegativeTestJSON-1015404931 tempest-ImagesNegativeTestJSON-1015404931-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 858.890011] env[65726]: DEBUG nova.compute.manager [None req-cb2d094a-412f-492d-9ef1-aba8757b7146 tempest-ServerMetadataNegativeTestJSON-936475568 tempest-ServerMetadataNegativeTestJSON-936475568-project-member] [instance: d179030a-e5ba-45b1-ad11-4d2f71ed7bd4] Start building block device mappings for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 858.938447] env[65726]: DEBUG oslo_vmware.api [None req-d7ed2ba1-3808-4f50-920c-ef8e5d179423 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]528aea06-9094-8a75-cbf9-cc19dffaeaf7, 'name': SearchDatastore_Task, 'duration_secs': 0.013308} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 858.942158] env[65726]: WARNING openstack [None req-55d74d77-063b-418b-b818-59d05c490f34 tempest-ImagesNegativeTestJSON-1015404931 tempest-ImagesNegativeTestJSON-1015404931-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 858.943034] env[65726]: WARNING openstack [None req-55d74d77-063b-418b-b818-59d05c490f34 tempest-ImagesNegativeTestJSON-1015404931 tempest-ImagesNegativeTestJSON-1015404931-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 858.951086] env[65726]: DEBUG oslo_concurrency.lockutils [None req-d7ed2ba1-3808-4f50-920c-ef8e5d179423 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 858.951370] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-d7ed2ba1-3808-4f50-920c-ef8e5d179423 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] [instance: 44d73b2d-2b6a-4501-9944-432da3c9330d] Processing image b52362a3-ee8a-4cbf-b06f-513b0cc8f95c {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 858.951618] env[65726]: DEBUG oslo_concurrency.lockutils [None req-d7ed2ba1-3808-4f50-920c-ef8e5d179423 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 858.951760] env[65726]: DEBUG oslo_concurrency.lockutils [None req-d7ed2ba1-3808-4f50-920c-ef8e5d179423 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 858.951932] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-d7ed2ba1-3808-4f50-920c-ef8e5d179423 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 858.955265] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-dfcda548-7e31-4189-a194-da793e13a137 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.965839] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-d7ed2ba1-3808-4f50-920c-ef8e5d179423 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 858.966042] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-d7ed2ba1-3808-4f50-920c-ef8e5d179423 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65726) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 858.967226] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-703fa062-699d-4a32-90f7-55f626707d8d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.977103] env[65726]: DEBUG oslo_vmware.api [None req-d7ed2ba1-3808-4f50-920c-ef8e5d179423 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Waiting for the task: (returnval){ [ 858.977103] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5258e9f3-5858-61df-fad4-2c4f68a5f365" [ 858.977103] env[65726]: _type = "Task" [ 858.977103] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.987900] env[65726]: DEBUG oslo_vmware.api [None req-d7ed2ba1-3808-4f50-920c-ef8e5d179423 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5258e9f3-5858-61df-fad4-2c4f68a5f365, 'name': SearchDatastore_Task, 'duration_secs': 0.010526} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 858.991717] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f6f36c63-dc98-45b8-bcc0-cf681fcd9902 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.997848] env[65726]: DEBUG oslo_vmware.api [None req-d7ed2ba1-3808-4f50-920c-ef8e5d179423 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Waiting for the task: (returnval){ [ 858.997848] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52401b29-f031-1db2-6b71-bc57006e6807" [ 858.997848] env[65726]: _type = "Task" [ 858.997848] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.015468] env[65726]: DEBUG oslo_vmware.api [None req-d7ed2ba1-3808-4f50-920c-ef8e5d179423 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52401b29-f031-1db2-6b71-bc57006e6807, 'name': SearchDatastore_Task, 'duration_secs': 0.010107} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.015934] env[65726]: DEBUG oslo_concurrency.lockutils [None req-d7ed2ba1-3808-4f50-920c-ef8e5d179423 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 859.016101] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-d7ed2ba1-3808-4f50-920c-ef8e5d179423 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] 44d73b2d-2b6a-4501-9944-432da3c9330d/44d73b2d-2b6a-4501-9944-432da3c9330d.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 859.016407] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-120d6e43-f704-4128-a0c8-5706cf4f9fa5 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.027796] env[65726]: DEBUG oslo_vmware.api [None req-d7ed2ba1-3808-4f50-920c-ef8e5d179423 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Waiting for the task: (returnval){ [ 859.027796] env[65726]: value = "task-5116037" [ 859.027796] env[65726]: _type = "Task" [ 859.027796] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.037893] env[65726]: DEBUG oslo_vmware.api [None req-d7ed2ba1-3808-4f50-920c-ef8e5d179423 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Task: {'id': task-5116037, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.054959] env[65726]: DEBUG oslo_vmware.api [None req-b0e194cf-16dc-40d0-b65c-88bd21cd63e3 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Task: {'id': task-5116035, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.52874} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.057636] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-b0e194cf-16dc-40d0-b65c-88bd21cd63e3 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] cb8855b6-589d-4863-b86f-7fb4e30fb29e/cb8855b6-589d-4863-b86f-7fb4e30fb29e.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 859.057848] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-b0e194cf-16dc-40d0-b65c-88bd21cd63e3 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: cb8855b6-589d-4863-b86f-7fb4e30fb29e] Extending root virtual disk to 1048576 {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 859.058602] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-08740685-cb22-4139-a283-9b8c7fec9535 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.065775] env[65726]: DEBUG oslo_vmware.api [None req-b0e194cf-16dc-40d0-b65c-88bd21cd63e3 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Waiting for the task: (returnval){ [ 859.065775] env[65726]: value = "task-5116038" [ 859.065775] env[65726]: _type = "Task" [ 859.065775] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.077969] env[65726]: DEBUG nova.network.neutron [None req-55d74d77-063b-418b-b818-59d05c490f34 tempest-ImagesNegativeTestJSON-1015404931 tempest-ImagesNegativeTestJSON-1015404931-project-member] [instance: c4177e20-b1bd-4b54-a275-c93582359a07] Updating instance_info_cache with network_info: [{"id": "da625869-f814-4733-b2ac-d50cc54a9858", "address": "fa:16:3e:69:88:8c", "network": {"id": "436a3921-05b0-4f20-b1a7-bbaea8c39407", "bridge": "br-int", "label": "tempest-ImagesNegativeTestJSON-344395661-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "28432176889e4680aa34fce120b47304", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bab6a6c3-1c5c-4776-b21b-dec21196d702", "external-id": "nsx-vlan-transportzone-634", "segmentation_id": 634, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapda625869-f8", "ovs_interfaceid": "da625869-f814-4733-b2ac-d50cc54a9858", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 859.081892] env[65726]: DEBUG oslo_vmware.api [None req-b0e194cf-16dc-40d0-b65c-88bd21cd63e3 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Task: {'id': task-5116038, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.330040] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-409bdca9-f552-40cb-9fd2-6c57c6a12d97 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.354722] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08424211-8bc3-4803-a0c8-e46df0a46b42 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.364408] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-4562f28d-bad2-4bca-833c-c069e59f6cc6 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] [instance: 76249623-6f83-46a3-b8c5-c001111aa698] Updating instance '76249623-6f83-46a3-b8c5-c001111aa698' progress to 83 {{(pid=65726) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 859.502267] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b178abc6-b846-42e9-b934-e47148253d2d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.518194] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a34bb764-a446-4cfe-9aa5-c3918fad7bb1 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.575913] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-337623e8-7c00-4e3a-abc5-5f746374cf26 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.585381] env[65726]: DEBUG oslo_concurrency.lockutils [None req-55d74d77-063b-418b-b818-59d05c490f34 tempest-ImagesNegativeTestJSON-1015404931 tempest-ImagesNegativeTestJSON-1015404931-project-member] Releasing lock "refresh_cache-c4177e20-b1bd-4b54-a275-c93582359a07" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 859.585741] env[65726]: DEBUG nova.compute.manager [None req-55d74d77-063b-418b-b818-59d05c490f34 tempest-ImagesNegativeTestJSON-1015404931 tempest-ImagesNegativeTestJSON-1015404931-project-member] [instance: c4177e20-b1bd-4b54-a275-c93582359a07] Instance network_info: |[{"id": "da625869-f814-4733-b2ac-d50cc54a9858", "address": "fa:16:3e:69:88:8c", "network": {"id": "436a3921-05b0-4f20-b1a7-bbaea8c39407", "bridge": "br-int", "label": "tempest-ImagesNegativeTestJSON-344395661-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "28432176889e4680aa34fce120b47304", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bab6a6c3-1c5c-4776-b21b-dec21196d702", "external-id": "nsx-vlan-transportzone-634", "segmentation_id": 634, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapda625869-f8", "ovs_interfaceid": "da625869-f814-4733-b2ac-d50cc54a9858", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 859.586087] env[65726]: DEBUG oslo_vmware.api [None req-d7ed2ba1-3808-4f50-920c-ef8e5d179423 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Task: {'id': task-5116037, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.591491] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-55d74d77-063b-418b-b818-59d05c490f34 tempest-ImagesNegativeTestJSON-1015404931 tempest-ImagesNegativeTestJSON-1015404931-project-member] [instance: c4177e20-b1bd-4b54-a275-c93582359a07] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:69:88:8c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'bab6a6c3-1c5c-4776-b21b-dec21196d702', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'da625869-f814-4733-b2ac-d50cc54a9858', 'vif_model': 'vmxnet3'}] {{(pid=65726) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 859.603079] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-55d74d77-063b-418b-b818-59d05c490f34 tempest-ImagesNegativeTestJSON-1015404931 tempest-ImagesNegativeTestJSON-1015404931-project-member] Creating folder: Project (28432176889e4680aa34fce120b47304). Parent ref: group-v995008. {{(pid=65726) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 859.603079] env[65726]: DEBUG oslo_vmware.api [None req-b0e194cf-16dc-40d0-b65c-88bd21cd63e3 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Task: {'id': task-5116038, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.076123} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.603079] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c99d9b9e-c32a-451c-ba4e-5a24b7304287 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.603079] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-b0e194cf-16dc-40d0-b65c-88bd21cd63e3 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: cb8855b6-589d-4863-b86f-7fb4e30fb29e] Extended root virtual disk {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 859.604104] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e1d0d17-df3e-4640-8cee-9ae0e7fd02eb {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.609599] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b506b6f-24a6-4445-a1fa-aaec71cf7eed {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.623392] env[65726]: DEBUG nova.compute.provider_tree [None req-b3ea747b-af8b-4592-b662-6643bc7f39d3 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 859.644057] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-b0e194cf-16dc-40d0-b65c-88bd21cd63e3 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: cb8855b6-589d-4863-b86f-7fb4e30fb29e] Reconfiguring VM instance instance-00000039 to attach disk [datastore1] cb8855b6-589d-4863-b86f-7fb4e30fb29e/cb8855b6-589d-4863-b86f-7fb4e30fb29e.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 859.644411] env[65726]: INFO nova.virt.vmwareapi.vm_util [None req-55d74d77-063b-418b-b818-59d05c490f34 tempest-ImagesNegativeTestJSON-1015404931 tempest-ImagesNegativeTestJSON-1015404931-project-member] Created folder: Project (28432176889e4680aa34fce120b47304) in parent group-v995008. [ 859.644577] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-55d74d77-063b-418b-b818-59d05c490f34 tempest-ImagesNegativeTestJSON-1015404931 tempest-ImagesNegativeTestJSON-1015404931-project-member] Creating folder: Instances. Parent ref: group-v995166. {{(pid=65726) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 859.646271] env[65726]: DEBUG nova.scheduler.client.report [None req-b3ea747b-af8b-4592-b662-6643bc7f39d3 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 859.649643] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-81ed486f-4224-445e-a976-71035fb34b46 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.664534] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-8ac0dd59-79c2-4fb3-91a9-81b98d5e74d1 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.667684] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b3ea747b-af8b-4592-b662-6643bc7f39d3 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 1.792s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 859.668153] env[65726]: DEBUG nova.compute.manager [None req-b3ea747b-af8b-4592-b662-6643bc7f39d3 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] [instance: 2cbee359-a1fc-45c0-bcc5-a2c24ded9c69] Start building networks asynchronously for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 859.670872] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c4393951-32f6-44d6-b9c7-1ad480073dbe tempest-InstanceActionsNegativeTestJSON-844675881 tempest-InstanceActionsNegativeTestJSON-844675881-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 33.394s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 859.671089] env[65726]: DEBUG nova.objects.instance [None req-c4393951-32f6-44d6-b9c7-1ad480073dbe tempest-InstanceActionsNegativeTestJSON-844675881 tempest-InstanceActionsNegativeTestJSON-844675881-project-member] Lazy-loading 'resources' on Instance uuid 1bd26aef-995e-43b6-af9f-077e878c8c44 {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 859.679365] env[65726]: DEBUG oslo_vmware.api [None req-b0e194cf-16dc-40d0-b65c-88bd21cd63e3 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Waiting for the task: (returnval){ [ 859.679365] env[65726]: value = "task-5116041" [ 859.679365] env[65726]: _type = "Task" [ 859.679365] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.686977] env[65726]: INFO nova.virt.vmwareapi.vm_util [None req-55d74d77-063b-418b-b818-59d05c490f34 tempest-ImagesNegativeTestJSON-1015404931 tempest-ImagesNegativeTestJSON-1015404931-project-member] Created folder: Instances in parent group-v995166. [ 859.687260] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-55d74d77-063b-418b-b818-59d05c490f34 tempest-ImagesNegativeTestJSON-1015404931 tempest-ImagesNegativeTestJSON-1015404931-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 859.688164] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c4177e20-b1bd-4b54-a275-c93582359a07] Creating VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 859.688460] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-0366e37c-d3e1-4d4f-81ed-595573888710 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.709025] env[65726]: DEBUG oslo_vmware.api [None req-b0e194cf-16dc-40d0-b65c-88bd21cd63e3 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Task: {'id': task-5116041, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.714550] env[65726]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 859.714550] env[65726]: value = "task-5116042" [ 859.714550] env[65726]: _type = "Task" [ 859.714550] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.724398] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116042, 'name': CreateVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.873017] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-4562f28d-bad2-4bca-833c-c069e59f6cc6 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] [instance: 76249623-6f83-46a3-b8c5-c001111aa698] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 859.873017] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7062a7bb-0834-4a93-8916-68480b2d328b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.879639] env[65726]: DEBUG oslo_vmware.api [None req-4562f28d-bad2-4bca-833c-c069e59f6cc6 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Waiting for the task: (returnval){ [ 859.879639] env[65726]: value = "task-5116043" [ 859.879639] env[65726]: _type = "Task" [ 859.879639] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.889270] env[65726]: DEBUG oslo_vmware.api [None req-4562f28d-bad2-4bca-833c-c069e59f6cc6 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Task: {'id': task-5116043, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.892317] env[65726]: DEBUG nova.compute.manager [req-8bfa2496-066d-487d-80e9-5f111e4f8f60 req-7539c948-2224-4f38-80ba-6f77b90a51da service nova] [instance: c4177e20-b1bd-4b54-a275-c93582359a07] Received event network-changed-da625869-f814-4733-b2ac-d50cc54a9858 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 859.892533] env[65726]: DEBUG nova.compute.manager [req-8bfa2496-066d-487d-80e9-5f111e4f8f60 req-7539c948-2224-4f38-80ba-6f77b90a51da service nova] [instance: c4177e20-b1bd-4b54-a275-c93582359a07] Refreshing instance network info cache due to event network-changed-da625869-f814-4733-b2ac-d50cc54a9858. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 859.892709] env[65726]: DEBUG oslo_concurrency.lockutils [req-8bfa2496-066d-487d-80e9-5f111e4f8f60 req-7539c948-2224-4f38-80ba-6f77b90a51da service nova] Acquiring lock "refresh_cache-c4177e20-b1bd-4b54-a275-c93582359a07" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 859.892845] env[65726]: DEBUG oslo_concurrency.lockutils [req-8bfa2496-066d-487d-80e9-5f111e4f8f60 req-7539c948-2224-4f38-80ba-6f77b90a51da service nova] Acquired lock "refresh_cache-c4177e20-b1bd-4b54-a275-c93582359a07" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 859.892993] env[65726]: DEBUG nova.network.neutron [req-8bfa2496-066d-487d-80e9-5f111e4f8f60 req-7539c948-2224-4f38-80ba-6f77b90a51da service nova] [instance: c4177e20-b1bd-4b54-a275-c93582359a07] Refreshing network info cache for port da625869-f814-4733-b2ac-d50cc54a9858 {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 859.903828] env[65726]: DEBUG nova.compute.manager [None req-cb2d094a-412f-492d-9ef1-aba8757b7146 tempest-ServerMetadataNegativeTestJSON-936475568 tempest-ServerMetadataNegativeTestJSON-936475568-project-member] [instance: d179030a-e5ba-45b1-ad11-4d2f71ed7bd4] Start spawning the instance on the hypervisor. {{(pid=65726) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 859.936740] env[65726]: DEBUG nova.virt.hardware [None req-cb2d094a-412f-492d-9ef1-aba8757b7146 tempest-ServerMetadataNegativeTestJSON-936475568 tempest-ServerMetadataNegativeTestJSON-936475568-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T19:28:34Z,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1539f06fa3534e90acbc3a60c4b8bd3f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T19:28:35Z,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 859.937083] env[65726]: DEBUG nova.virt.hardware [None req-cb2d094a-412f-492d-9ef1-aba8757b7146 tempest-ServerMetadataNegativeTestJSON-936475568 tempest-ServerMetadataNegativeTestJSON-936475568-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 859.937144] env[65726]: DEBUG nova.virt.hardware [None req-cb2d094a-412f-492d-9ef1-aba8757b7146 tempest-ServerMetadataNegativeTestJSON-936475568 tempest-ServerMetadataNegativeTestJSON-936475568-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 859.937340] env[65726]: DEBUG nova.virt.hardware [None req-cb2d094a-412f-492d-9ef1-aba8757b7146 tempest-ServerMetadataNegativeTestJSON-936475568 tempest-ServerMetadataNegativeTestJSON-936475568-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 859.937501] env[65726]: DEBUG nova.virt.hardware [None req-cb2d094a-412f-492d-9ef1-aba8757b7146 tempest-ServerMetadataNegativeTestJSON-936475568 tempest-ServerMetadataNegativeTestJSON-936475568-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 859.937657] env[65726]: DEBUG nova.virt.hardware [None req-cb2d094a-412f-492d-9ef1-aba8757b7146 tempest-ServerMetadataNegativeTestJSON-936475568 tempest-ServerMetadataNegativeTestJSON-936475568-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 859.937782] env[65726]: DEBUG nova.virt.hardware [None req-cb2d094a-412f-492d-9ef1-aba8757b7146 tempest-ServerMetadataNegativeTestJSON-936475568 tempest-ServerMetadataNegativeTestJSON-936475568-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 859.937955] env[65726]: DEBUG nova.virt.hardware [None req-cb2d094a-412f-492d-9ef1-aba8757b7146 tempest-ServerMetadataNegativeTestJSON-936475568 tempest-ServerMetadataNegativeTestJSON-936475568-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 859.938664] env[65726]: DEBUG nova.virt.hardware [None req-cb2d094a-412f-492d-9ef1-aba8757b7146 tempest-ServerMetadataNegativeTestJSON-936475568 tempest-ServerMetadataNegativeTestJSON-936475568-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 859.939690] env[65726]: DEBUG nova.virt.hardware [None req-cb2d094a-412f-492d-9ef1-aba8757b7146 tempest-ServerMetadataNegativeTestJSON-936475568 tempest-ServerMetadataNegativeTestJSON-936475568-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 859.939690] env[65726]: DEBUG nova.virt.hardware [None req-cb2d094a-412f-492d-9ef1-aba8757b7146 tempest-ServerMetadataNegativeTestJSON-936475568 tempest-ServerMetadataNegativeTestJSON-936475568-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 859.940509] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b88eac1-3562-4380-866a-8bb0af91a03f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.949595] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1ca0f90-dbdd-4a78-b3c9-7306d31f09a2 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.080661] env[65726]: DEBUG oslo_vmware.api [None req-d7ed2ba1-3808-4f50-920c-ef8e5d179423 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Task: {'id': task-5116037, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.177846] env[65726]: DEBUG nova.compute.utils [None req-b3ea747b-af8b-4592-b662-6643bc7f39d3 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] Using /dev/sd instead of None {{(pid=65726) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 860.182021] env[65726]: DEBUG nova.compute.manager [None req-b3ea747b-af8b-4592-b662-6643bc7f39d3 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] [instance: 2cbee359-a1fc-45c0-bcc5-a2c24ded9c69] Not allocating networking since 'none' was specified. {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2015}} [ 860.193187] env[65726]: DEBUG oslo_vmware.api [None req-b0e194cf-16dc-40d0-b65c-88bd21cd63e3 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Task: {'id': task-5116041, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.228228] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116042, 'name': CreateVM_Task} progress is 25%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.254964] env[65726]: DEBUG nova.network.neutron [None req-cb2d094a-412f-492d-9ef1-aba8757b7146 tempest-ServerMetadataNegativeTestJSON-936475568 tempest-ServerMetadataNegativeTestJSON-936475568-project-member] [instance: d179030a-e5ba-45b1-ad11-4d2f71ed7bd4] Successfully updated port: 51e75557-8c53-4ba4-982a-42e90b5decef {{(pid=65726) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 860.393724] env[65726]: DEBUG oslo_vmware.api [None req-4562f28d-bad2-4bca-833c-c069e59f6cc6 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Task: {'id': task-5116043, 'name': PowerOnVM_Task, 'duration_secs': 0.485903} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.393724] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-4562f28d-bad2-4bca-833c-c069e59f6cc6 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] [instance: 76249623-6f83-46a3-b8c5-c001111aa698] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 860.393859] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-4562f28d-bad2-4bca-833c-c069e59f6cc6 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] [instance: 76249623-6f83-46a3-b8c5-c001111aa698] Updating instance '76249623-6f83-46a3-b8c5-c001111aa698' progress to 100 {{(pid=65726) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 860.400151] env[65726]: WARNING openstack [req-8bfa2496-066d-487d-80e9-5f111e4f8f60 req-7539c948-2224-4f38-80ba-6f77b90a51da service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 860.400540] env[65726]: WARNING openstack [req-8bfa2496-066d-487d-80e9-5f111e4f8f60 req-7539c948-2224-4f38-80ba-6f77b90a51da service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 860.511368] env[65726]: WARNING openstack [req-8bfa2496-066d-487d-80e9-5f111e4f8f60 req-7539c948-2224-4f38-80ba-6f77b90a51da service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 860.511844] env[65726]: WARNING openstack [req-8bfa2496-066d-487d-80e9-5f111e4f8f60 req-7539c948-2224-4f38-80ba-6f77b90a51da service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 860.585178] env[65726]: DEBUG oslo_vmware.api [None req-d7ed2ba1-3808-4f50-920c-ef8e5d179423 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Task: {'id': task-5116037, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.528242} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.585466] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-d7ed2ba1-3808-4f50-920c-ef8e5d179423 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] 44d73b2d-2b6a-4501-9944-432da3c9330d/44d73b2d-2b6a-4501-9944-432da3c9330d.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 860.585685] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-d7ed2ba1-3808-4f50-920c-ef8e5d179423 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] [instance: 44d73b2d-2b6a-4501-9944-432da3c9330d] Extending root virtual disk to 1048576 {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 860.585951] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4fe271d5-9ac2-4dea-90c2-f03df514cf6f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.597134] env[65726]: WARNING openstack [req-8bfa2496-066d-487d-80e9-5f111e4f8f60 req-7539c948-2224-4f38-80ba-6f77b90a51da service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 860.597724] env[65726]: WARNING openstack [req-8bfa2496-066d-487d-80e9-5f111e4f8f60 req-7539c948-2224-4f38-80ba-6f77b90a51da service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 860.610517] env[65726]: DEBUG oslo_vmware.api [None req-d7ed2ba1-3808-4f50-920c-ef8e5d179423 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Waiting for the task: (returnval){ [ 860.610517] env[65726]: value = "task-5116044" [ 860.610517] env[65726]: _type = "Task" [ 860.610517] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.621193] env[65726]: DEBUG oslo_vmware.api [None req-d7ed2ba1-3808-4f50-920c-ef8e5d179423 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Task: {'id': task-5116044, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.681633] env[65726]: DEBUG nova.compute.manager [None req-b3ea747b-af8b-4592-b662-6643bc7f39d3 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] [instance: 2cbee359-a1fc-45c0-bcc5-a2c24ded9c69] Start building block device mappings for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 860.689919] env[65726]: DEBUG nova.network.neutron [req-8bfa2496-066d-487d-80e9-5f111e4f8f60 req-7539c948-2224-4f38-80ba-6f77b90a51da service nova] [instance: c4177e20-b1bd-4b54-a275-c93582359a07] Updated VIF entry in instance network info cache for port da625869-f814-4733-b2ac-d50cc54a9858. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 860.690293] env[65726]: DEBUG nova.network.neutron [req-8bfa2496-066d-487d-80e9-5f111e4f8f60 req-7539c948-2224-4f38-80ba-6f77b90a51da service nova] [instance: c4177e20-b1bd-4b54-a275-c93582359a07] Updating instance_info_cache with network_info: [{"id": "da625869-f814-4733-b2ac-d50cc54a9858", "address": "fa:16:3e:69:88:8c", "network": {"id": "436a3921-05b0-4f20-b1a7-bbaea8c39407", "bridge": "br-int", "label": "tempest-ImagesNegativeTestJSON-344395661-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "28432176889e4680aa34fce120b47304", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bab6a6c3-1c5c-4776-b21b-dec21196d702", "external-id": "nsx-vlan-transportzone-634", "segmentation_id": 634, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapda625869-f8", "ovs_interfaceid": "da625869-f814-4733-b2ac-d50cc54a9858", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 860.701409] env[65726]: DEBUG oslo_vmware.api [None req-b0e194cf-16dc-40d0-b65c-88bd21cd63e3 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Task: {'id': task-5116041, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.725724] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116042, 'name': CreateVM_Task} progress is 25%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.751705] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12cbba99-1fff-40f5-a107-e6e0e4c6eaf2 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.761224] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e74a6e83-f005-476c-9861-766248565794 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.765268] env[65726]: DEBUG oslo_concurrency.lockutils [None req-cb2d094a-412f-492d-9ef1-aba8757b7146 tempest-ServerMetadataNegativeTestJSON-936475568 tempest-ServerMetadataNegativeTestJSON-936475568-project-member] Acquiring lock "refresh_cache-d179030a-e5ba-45b1-ad11-4d2f71ed7bd4" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 860.765469] env[65726]: DEBUG oslo_concurrency.lockutils [None req-cb2d094a-412f-492d-9ef1-aba8757b7146 tempest-ServerMetadataNegativeTestJSON-936475568 tempest-ServerMetadataNegativeTestJSON-936475568-project-member] Acquired lock "refresh_cache-d179030a-e5ba-45b1-ad11-4d2f71ed7bd4" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 860.765605] env[65726]: DEBUG nova.network.neutron [None req-cb2d094a-412f-492d-9ef1-aba8757b7146 tempest-ServerMetadataNegativeTestJSON-936475568 tempest-ServerMetadataNegativeTestJSON-936475568-project-member] [instance: d179030a-e5ba-45b1-ad11-4d2f71ed7bd4] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 860.810149] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c762ef9c-a462-4e6d-b457-844787d4c25a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.822066] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6e2bfa6-ea6a-4702-a9be-a200c3bf4b24 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.839238] env[65726]: DEBUG nova.compute.provider_tree [None req-c4393951-32f6-44d6-b9c7-1ad480073dbe tempest-InstanceActionsNegativeTestJSON-844675881 tempest-InstanceActionsNegativeTestJSON-844675881-project-member] Updating inventory in ProviderTree for provider 07c4692f-bdb4-4058-9173-ff9664830295 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 861.122656] env[65726]: DEBUG oslo_vmware.api [None req-d7ed2ba1-3808-4f50-920c-ef8e5d179423 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Task: {'id': task-5116044, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066048} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.122656] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-d7ed2ba1-3808-4f50-920c-ef8e5d179423 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] [instance: 44d73b2d-2b6a-4501-9944-432da3c9330d] Extended root virtual disk {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 861.123288] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-268baea0-8fda-46f1-8a3a-a927fea41907 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.144660] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-d7ed2ba1-3808-4f50-920c-ef8e5d179423 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] [instance: 44d73b2d-2b6a-4501-9944-432da3c9330d] Reconfiguring VM instance instance-00000038 to attach disk [datastore1] 44d73b2d-2b6a-4501-9944-432da3c9330d/44d73b2d-2b6a-4501-9944-432da3c9330d.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 861.144983] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2a465733-a290-459d-88f0-e7cc09c07c03 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.167092] env[65726]: DEBUG oslo_vmware.api [None req-d7ed2ba1-3808-4f50-920c-ef8e5d179423 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Waiting for the task: (returnval){ [ 861.167092] env[65726]: value = "task-5116045" [ 861.167092] env[65726]: _type = "Task" [ 861.167092] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.176884] env[65726]: DEBUG oslo_vmware.api [None req-d7ed2ba1-3808-4f50-920c-ef8e5d179423 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Task: {'id': task-5116045, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.202011] env[65726]: DEBUG oslo_concurrency.lockutils [req-8bfa2496-066d-487d-80e9-5f111e4f8f60 req-7539c948-2224-4f38-80ba-6f77b90a51da service nova] Releasing lock "refresh_cache-c4177e20-b1bd-4b54-a275-c93582359a07" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 861.202570] env[65726]: DEBUG oslo_vmware.api [None req-b0e194cf-16dc-40d0-b65c-88bd21cd63e3 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Task: {'id': task-5116041, 'name': ReconfigVM_Task, 'duration_secs': 1.179435} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.203271] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-b0e194cf-16dc-40d0-b65c-88bd21cd63e3 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: cb8855b6-589d-4863-b86f-7fb4e30fb29e] Reconfigured VM instance instance-00000039 to attach disk [datastore1] cb8855b6-589d-4863-b86f-7fb4e30fb29e/cb8855b6-589d-4863-b86f-7fb4e30fb29e.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 861.204087] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-aaae7fad-b883-40dd-9026-f70f52aa07e2 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.212522] env[65726]: DEBUG oslo_vmware.api [None req-b0e194cf-16dc-40d0-b65c-88bd21cd63e3 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Waiting for the task: (returnval){ [ 861.212522] env[65726]: value = "task-5116046" [ 861.212522] env[65726]: _type = "Task" [ 861.212522] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.224943] env[65726]: DEBUG oslo_vmware.api [None req-b0e194cf-16dc-40d0-b65c-88bd21cd63e3 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Task: {'id': task-5116046, 'name': Rename_Task} progress is 6%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.228065] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116042, 'name': CreateVM_Task, 'duration_secs': 1.124011} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.228214] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c4177e20-b1bd-4b54-a275-c93582359a07] Created VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 861.228918] env[65726]: WARNING openstack [None req-55d74d77-063b-418b-b818-59d05c490f34 tempest-ImagesNegativeTestJSON-1015404931 tempest-ImagesNegativeTestJSON-1015404931-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 861.229317] env[65726]: WARNING openstack [None req-55d74d77-063b-418b-b818-59d05c490f34 tempest-ImagesNegativeTestJSON-1015404931 tempest-ImagesNegativeTestJSON-1015404931-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 861.235012] env[65726]: DEBUG oslo_concurrency.lockutils [None req-55d74d77-063b-418b-b818-59d05c490f34 tempest-ImagesNegativeTestJSON-1015404931 tempest-ImagesNegativeTestJSON-1015404931-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 861.235175] env[65726]: DEBUG oslo_concurrency.lockutils [None req-55d74d77-063b-418b-b818-59d05c490f34 tempest-ImagesNegativeTestJSON-1015404931 tempest-ImagesNegativeTestJSON-1015404931-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 861.235495] env[65726]: DEBUG oslo_concurrency.lockutils [None req-55d74d77-063b-418b-b818-59d05c490f34 tempest-ImagesNegativeTestJSON-1015404931 tempest-ImagesNegativeTestJSON-1015404931-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 861.235851] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a173f2ec-7cd5-49f0-a2cf-e2ce02a57bc1 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.241601] env[65726]: DEBUG oslo_vmware.api [None req-55d74d77-063b-418b-b818-59d05c490f34 tempest-ImagesNegativeTestJSON-1015404931 tempest-ImagesNegativeTestJSON-1015404931-project-member] Waiting for the task: (returnval){ [ 861.241601] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]523eee74-9b46-d006-e880-ce34017e1cd6" [ 861.241601] env[65726]: _type = "Task" [ 861.241601] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.254075] env[65726]: DEBUG oslo_vmware.api [None req-55d74d77-063b-418b-b818-59d05c490f34 tempest-ImagesNegativeTestJSON-1015404931 tempest-ImagesNegativeTestJSON-1015404931-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]523eee74-9b46-d006-e880-ce34017e1cd6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.270618] env[65726]: WARNING openstack [None req-cb2d094a-412f-492d-9ef1-aba8757b7146 tempest-ServerMetadataNegativeTestJSON-936475568 tempest-ServerMetadataNegativeTestJSON-936475568-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 861.271050] env[65726]: WARNING openstack [None req-cb2d094a-412f-492d-9ef1-aba8757b7146 tempest-ServerMetadataNegativeTestJSON-936475568 tempest-ServerMetadataNegativeTestJSON-936475568-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 861.276197] env[65726]: DEBUG nova.network.neutron [None req-cb2d094a-412f-492d-9ef1-aba8757b7146 tempest-ServerMetadataNegativeTestJSON-936475568 tempest-ServerMetadataNegativeTestJSON-936475568-project-member] [instance: d179030a-e5ba-45b1-ad11-4d2f71ed7bd4] Instance cache missing network info. {{(pid=65726) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 861.341988] env[65726]: WARNING openstack [None req-cb2d094a-412f-492d-9ef1-aba8757b7146 tempest-ServerMetadataNegativeTestJSON-936475568 tempest-ServerMetadataNegativeTestJSON-936475568-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 861.342426] env[65726]: WARNING openstack [None req-cb2d094a-412f-492d-9ef1-aba8757b7146 tempest-ServerMetadataNegativeTestJSON-936475568 tempest-ServerMetadataNegativeTestJSON-936475568-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 861.370311] env[65726]: ERROR nova.scheduler.client.report [None req-c4393951-32f6-44d6-b9c7-1ad480073dbe tempest-InstanceActionsNegativeTestJSON-844675881 tempest-InstanceActionsNegativeTestJSON-844675881-project-member] [req-14e9435f-3a05-4639-afd8-a301693f6e6d] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 07c4692f-bdb4-4058-9173-ff9664830295. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-14e9435f-3a05-4639-afd8-a301693f6e6d"}]} [ 861.389031] env[65726]: DEBUG nova.scheduler.client.report [None req-c4393951-32f6-44d6-b9c7-1ad480073dbe tempest-InstanceActionsNegativeTestJSON-844675881 tempest-InstanceActionsNegativeTestJSON-844675881-project-member] Refreshing inventories for resource provider 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 861.404955] env[65726]: DEBUG nova.scheduler.client.report [None req-c4393951-32f6-44d6-b9c7-1ad480073dbe tempest-InstanceActionsNegativeTestJSON-844675881 tempest-InstanceActionsNegativeTestJSON-844675881-project-member] Updating ProviderTree inventory for provider 07c4692f-bdb4-4058-9173-ff9664830295 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 861.405270] env[65726]: DEBUG nova.compute.provider_tree [None req-c4393951-32f6-44d6-b9c7-1ad480073dbe tempest-InstanceActionsNegativeTestJSON-844675881 tempest-InstanceActionsNegativeTestJSON-844675881-project-member] Updating inventory in ProviderTree for provider 07c4692f-bdb4-4058-9173-ff9664830295 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 861.418376] env[65726]: DEBUG nova.scheduler.client.report [None req-c4393951-32f6-44d6-b9c7-1ad480073dbe tempest-InstanceActionsNegativeTestJSON-844675881 tempest-InstanceActionsNegativeTestJSON-844675881-project-member] Refreshing aggregate associations for resource provider 07c4692f-bdb4-4058-9173-ff9664830295, aggregates: None {{(pid=65726) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 861.424483] env[65726]: WARNING openstack [None req-cb2d094a-412f-492d-9ef1-aba8757b7146 tempest-ServerMetadataNegativeTestJSON-936475568 tempest-ServerMetadataNegativeTestJSON-936475568-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 861.424930] env[65726]: WARNING openstack [None req-cb2d094a-412f-492d-9ef1-aba8757b7146 tempest-ServerMetadataNegativeTestJSON-936475568 tempest-ServerMetadataNegativeTestJSON-936475568-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 861.441707] env[65726]: DEBUG nova.scheduler.client.report [None req-c4393951-32f6-44d6-b9c7-1ad480073dbe tempest-InstanceActionsNegativeTestJSON-844675881 tempest-InstanceActionsNegativeTestJSON-844675881-project-member] Refreshing trait associations for resource provider 07c4692f-bdb4-4058-9173-ff9664830295, traits: COMPUTE_IMAGE_TYPE_ISO,HW_ARCH_X86_64,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=65726) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 861.524590] env[65726]: DEBUG nova.network.neutron [None req-cb2d094a-412f-492d-9ef1-aba8757b7146 tempest-ServerMetadataNegativeTestJSON-936475568 tempest-ServerMetadataNegativeTestJSON-936475568-project-member] [instance: d179030a-e5ba-45b1-ad11-4d2f71ed7bd4] Updating instance_info_cache with network_info: [{"id": "51e75557-8c53-4ba4-982a-42e90b5decef", "address": "fa:16:3e:d9:a4:69", "network": {"id": "c9aee0fe-f17c-4207-9bae-b92c4f507fb3", "bridge": "br-int", "label": "tempest-ServerMetadataNegativeTestJSON-1399638803-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "921d46aee1824fd990022c7f851018b2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1a8c8175-1197-4f12-baac-ef6aba95f585", "external-id": "nsx-vlan-transportzone-832", "segmentation_id": 832, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap51e75557-8c", "ovs_interfaceid": "51e75557-8c53-4ba4-982a-42e90b5decef", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 861.679798] env[65726]: DEBUG oslo_vmware.api [None req-d7ed2ba1-3808-4f50-920c-ef8e5d179423 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Task: {'id': task-5116045, 'name': ReconfigVM_Task, 'duration_secs': 0.472803} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.680279] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-d7ed2ba1-3808-4f50-920c-ef8e5d179423 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] [instance: 44d73b2d-2b6a-4501-9944-432da3c9330d] Reconfigured VM instance instance-00000038 to attach disk [datastore1] 44d73b2d-2b6a-4501-9944-432da3c9330d/44d73b2d-2b6a-4501-9944-432da3c9330d.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 861.681042] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d2a734f7-1a67-48a6-a9cf-4d0b8caca42b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.687895] env[65726]: DEBUG oslo_vmware.api [None req-d7ed2ba1-3808-4f50-920c-ef8e5d179423 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Waiting for the task: (returnval){ [ 861.687895] env[65726]: value = "task-5116047" [ 861.687895] env[65726]: _type = "Task" [ 861.687895] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.698759] env[65726]: DEBUG oslo_vmware.api [None req-d7ed2ba1-3808-4f50-920c-ef8e5d179423 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Task: {'id': task-5116047, 'name': Rename_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.704279] env[65726]: DEBUG nova.compute.manager [None req-b3ea747b-af8b-4592-b662-6643bc7f39d3 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] [instance: 2cbee359-a1fc-45c0-bcc5-a2c24ded9c69] Start spawning the instance on the hypervisor. {{(pid=65726) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 861.725603] env[65726]: DEBUG oslo_vmware.api [None req-b0e194cf-16dc-40d0-b65c-88bd21cd63e3 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Task: {'id': task-5116046, 'name': Rename_Task, 'duration_secs': 0.21326} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.727657] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-b0e194cf-16dc-40d0-b65c-88bd21cd63e3 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: cb8855b6-589d-4863-b86f-7fb4e30fb29e] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 861.728304] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1dbcd853-8fbd-4714-b9bf-832bb6c372b8 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.735431] env[65726]: DEBUG nova.virt.hardware [None req-b3ea747b-af8b-4592-b662-6643bc7f39d3 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T19:28:34Z,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1539f06fa3534e90acbc3a60c4b8bd3f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T19:28:35Z,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 861.735857] env[65726]: DEBUG nova.virt.hardware [None req-b3ea747b-af8b-4592-b662-6643bc7f39d3 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 861.736088] env[65726]: DEBUG nova.virt.hardware [None req-b3ea747b-af8b-4592-b662-6643bc7f39d3 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 861.736342] env[65726]: DEBUG nova.virt.hardware [None req-b3ea747b-af8b-4592-b662-6643bc7f39d3 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 861.736543] env[65726]: DEBUG nova.virt.hardware [None req-b3ea747b-af8b-4592-b662-6643bc7f39d3 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 861.736682] env[65726]: DEBUG nova.virt.hardware [None req-b3ea747b-af8b-4592-b662-6643bc7f39d3 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 861.736944] env[65726]: DEBUG nova.virt.hardware [None req-b3ea747b-af8b-4592-b662-6643bc7f39d3 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 861.737151] env[65726]: DEBUG nova.virt.hardware [None req-b3ea747b-af8b-4592-b662-6643bc7f39d3 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 861.737339] env[65726]: DEBUG nova.virt.hardware [None req-b3ea747b-af8b-4592-b662-6643bc7f39d3 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 861.737501] env[65726]: DEBUG nova.virt.hardware [None req-b3ea747b-af8b-4592-b662-6643bc7f39d3 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 861.737671] env[65726]: DEBUG nova.virt.hardware [None req-b3ea747b-af8b-4592-b662-6643bc7f39d3 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 861.738554] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5b47cdc-2269-42fb-ae83-b12aa0bffae3 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.745015] env[65726]: DEBUG oslo_vmware.api [None req-b0e194cf-16dc-40d0-b65c-88bd21cd63e3 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Waiting for the task: (returnval){ [ 861.745015] env[65726]: value = "task-5116048" [ 861.745015] env[65726]: _type = "Task" [ 861.745015] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.755664] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86ebe1cf-aef0-42f6-a700-d6be7ab5e146 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.766230] env[65726]: DEBUG oslo_vmware.api [None req-55d74d77-063b-418b-b818-59d05c490f34 tempest-ImagesNegativeTestJSON-1015404931 tempest-ImagesNegativeTestJSON-1015404931-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]523eee74-9b46-d006-e880-ce34017e1cd6, 'name': SearchDatastore_Task, 'duration_secs': 0.011133} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.767160] env[65726]: DEBUG oslo_concurrency.lockutils [None req-55d74d77-063b-418b-b818-59d05c490f34 tempest-ImagesNegativeTestJSON-1015404931 tempest-ImagesNegativeTestJSON-1015404931-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 861.767395] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-55d74d77-063b-418b-b818-59d05c490f34 tempest-ImagesNegativeTestJSON-1015404931 tempest-ImagesNegativeTestJSON-1015404931-project-member] [instance: c4177e20-b1bd-4b54-a275-c93582359a07] Processing image b52362a3-ee8a-4cbf-b06f-513b0cc8f95c {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 861.767726] env[65726]: DEBUG oslo_concurrency.lockutils [None req-55d74d77-063b-418b-b818-59d05c490f34 tempest-ImagesNegativeTestJSON-1015404931 tempest-ImagesNegativeTestJSON-1015404931-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 861.767987] env[65726]: DEBUG oslo_concurrency.lockutils [None req-55d74d77-063b-418b-b818-59d05c490f34 tempest-ImagesNegativeTestJSON-1015404931 tempest-ImagesNegativeTestJSON-1015404931-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 861.768302] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-55d74d77-063b-418b-b818-59d05c490f34 tempest-ImagesNegativeTestJSON-1015404931 tempest-ImagesNegativeTestJSON-1015404931-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 861.772919] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-20edb5be-f6dd-4f0d-83d8-1c7742792ee3 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.774618] env[65726]: DEBUG oslo_vmware.api [None req-b0e194cf-16dc-40d0-b65c-88bd21cd63e3 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Task: {'id': task-5116048, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.784951] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-b3ea747b-af8b-4592-b662-6643bc7f39d3 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] [instance: 2cbee359-a1fc-45c0-bcc5-a2c24ded9c69] Instance VIF info [] {{(pid=65726) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 861.790977] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-b3ea747b-af8b-4592-b662-6643bc7f39d3 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] Creating folder: Project (e3c48189fcee4ff29b2527fe1ad4fe0b). Parent ref: group-v995008. {{(pid=65726) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 861.794611] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-78afe6f1-b5e7-491b-87da-29c137385eed {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.807027] env[65726]: INFO nova.virt.vmwareapi.vm_util [None req-b3ea747b-af8b-4592-b662-6643bc7f39d3 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] Created folder: Project (e3c48189fcee4ff29b2527fe1ad4fe0b) in parent group-v995008. [ 861.807188] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-b3ea747b-af8b-4592-b662-6643bc7f39d3 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] Creating folder: Instances. Parent ref: group-v995169. {{(pid=65726) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 861.811357] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-48ed2c7a-4af4-4ddf-913f-9516c3ef34dc {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.813196] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-55d74d77-063b-418b-b818-59d05c490f34 tempest-ImagesNegativeTestJSON-1015404931 tempest-ImagesNegativeTestJSON-1015404931-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 861.813386] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-55d74d77-063b-418b-b818-59d05c490f34 tempest-ImagesNegativeTestJSON-1015404931 tempest-ImagesNegativeTestJSON-1015404931-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65726) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 861.814336] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9d968224-32fe-4beb-a869-6c9eaf3d5c38 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.820550] env[65726]: DEBUG oslo_vmware.api [None req-55d74d77-063b-418b-b818-59d05c490f34 tempest-ImagesNegativeTestJSON-1015404931 tempest-ImagesNegativeTestJSON-1015404931-project-member] Waiting for the task: (returnval){ [ 861.820550] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5252cf09-964a-cce3-1208-42ac163bd58d" [ 861.820550] env[65726]: _type = "Task" [ 861.820550] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.828659] env[65726]: INFO nova.virt.vmwareapi.vm_util [None req-b3ea747b-af8b-4592-b662-6643bc7f39d3 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] Created folder: Instances in parent group-v995169. [ 861.829064] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-b3ea747b-af8b-4592-b662-6643bc7f39d3 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 861.829388] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2cbee359-a1fc-45c0-bcc5-a2c24ded9c69] Creating VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 861.829563] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-4a7b805d-8fff-4665-89b7-b102e0966842 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.849028] env[65726]: DEBUG oslo_vmware.api [None req-55d74d77-063b-418b-b818-59d05c490f34 tempest-ImagesNegativeTestJSON-1015404931 tempest-ImagesNegativeTestJSON-1015404931-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5252cf09-964a-cce3-1208-42ac163bd58d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.859996] env[65726]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 861.859996] env[65726]: value = "task-5116051" [ 861.859996] env[65726]: _type = "Task" [ 861.859996] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.872721] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116051, 'name': CreateVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.939402] env[65726]: DEBUG nova.compute.manager [req-0d5d700d-ae02-452d-bad7-37d024b6ab53 req-0e204bef-dd48-4319-a628-536f757be00d service nova] [instance: d179030a-e5ba-45b1-ad11-4d2f71ed7bd4] Received event network-vif-plugged-51e75557-8c53-4ba4-982a-42e90b5decef {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 861.939402] env[65726]: DEBUG oslo_concurrency.lockutils [req-0d5d700d-ae02-452d-bad7-37d024b6ab53 req-0e204bef-dd48-4319-a628-536f757be00d service nova] Acquiring lock "d179030a-e5ba-45b1-ad11-4d2f71ed7bd4-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 861.939402] env[65726]: DEBUG oslo_concurrency.lockutils [req-0d5d700d-ae02-452d-bad7-37d024b6ab53 req-0e204bef-dd48-4319-a628-536f757be00d service nova] Lock "d179030a-e5ba-45b1-ad11-4d2f71ed7bd4-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 861.939402] env[65726]: DEBUG oslo_concurrency.lockutils [req-0d5d700d-ae02-452d-bad7-37d024b6ab53 req-0e204bef-dd48-4319-a628-536f757be00d service nova] Lock "d179030a-e5ba-45b1-ad11-4d2f71ed7bd4-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 861.939402] env[65726]: DEBUG nova.compute.manager [req-0d5d700d-ae02-452d-bad7-37d024b6ab53 req-0e204bef-dd48-4319-a628-536f757be00d service nova] [instance: d179030a-e5ba-45b1-ad11-4d2f71ed7bd4] No waiting events found dispatching network-vif-plugged-51e75557-8c53-4ba4-982a-42e90b5decef {{(pid=65726) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 861.939402] env[65726]: WARNING nova.compute.manager [req-0d5d700d-ae02-452d-bad7-37d024b6ab53 req-0e204bef-dd48-4319-a628-536f757be00d service nova] [instance: d179030a-e5ba-45b1-ad11-4d2f71ed7bd4] Received unexpected event network-vif-plugged-51e75557-8c53-4ba4-982a-42e90b5decef for instance with vm_state building and task_state spawning. [ 861.939632] env[65726]: DEBUG nova.compute.manager [req-0d5d700d-ae02-452d-bad7-37d024b6ab53 req-0e204bef-dd48-4319-a628-536f757be00d service nova] [instance: d179030a-e5ba-45b1-ad11-4d2f71ed7bd4] Received event network-changed-51e75557-8c53-4ba4-982a-42e90b5decef {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 861.939786] env[65726]: DEBUG nova.compute.manager [req-0d5d700d-ae02-452d-bad7-37d024b6ab53 req-0e204bef-dd48-4319-a628-536f757be00d service nova] [instance: d179030a-e5ba-45b1-ad11-4d2f71ed7bd4] Refreshing instance network info cache due to event network-changed-51e75557-8c53-4ba4-982a-42e90b5decef. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 861.939917] env[65726]: DEBUG oslo_concurrency.lockutils [req-0d5d700d-ae02-452d-bad7-37d024b6ab53 req-0e204bef-dd48-4319-a628-536f757be00d service nova] Acquiring lock "refresh_cache-d179030a-e5ba-45b1-ad11-4d2f71ed7bd4" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 861.996180] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46f9dc79-087f-4a6d-8d0b-da5e2f076712 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.005445] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77cb28a2-75a6-4518-94b0-205f90e29c32 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.043450] env[65726]: DEBUG oslo_concurrency.lockutils [None req-cb2d094a-412f-492d-9ef1-aba8757b7146 tempest-ServerMetadataNegativeTestJSON-936475568 tempest-ServerMetadataNegativeTestJSON-936475568-project-member] Releasing lock "refresh_cache-d179030a-e5ba-45b1-ad11-4d2f71ed7bd4" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 862.043450] env[65726]: DEBUG nova.compute.manager [None req-cb2d094a-412f-492d-9ef1-aba8757b7146 tempest-ServerMetadataNegativeTestJSON-936475568 tempest-ServerMetadataNegativeTestJSON-936475568-project-member] [instance: d179030a-e5ba-45b1-ad11-4d2f71ed7bd4] Instance network_info: |[{"id": "51e75557-8c53-4ba4-982a-42e90b5decef", "address": "fa:16:3e:d9:a4:69", "network": {"id": "c9aee0fe-f17c-4207-9bae-b92c4f507fb3", "bridge": "br-int", "label": "tempest-ServerMetadataNegativeTestJSON-1399638803-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "921d46aee1824fd990022c7f851018b2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1a8c8175-1197-4f12-baac-ef6aba95f585", "external-id": "nsx-vlan-transportzone-832", "segmentation_id": 832, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap51e75557-8c", "ovs_interfaceid": "51e75557-8c53-4ba4-982a-42e90b5decef", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 862.043450] env[65726]: DEBUG oslo_concurrency.lockutils [req-0d5d700d-ae02-452d-bad7-37d024b6ab53 req-0e204bef-dd48-4319-a628-536f757be00d service nova] Acquired lock "refresh_cache-d179030a-e5ba-45b1-ad11-4d2f71ed7bd4" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 862.043450] env[65726]: DEBUG nova.network.neutron [req-0d5d700d-ae02-452d-bad7-37d024b6ab53 req-0e204bef-dd48-4319-a628-536f757be00d service nova] [instance: d179030a-e5ba-45b1-ad11-4d2f71ed7bd4] Refreshing network info cache for port 51e75557-8c53-4ba4-982a-42e90b5decef {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 862.043450] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-cb2d094a-412f-492d-9ef1-aba8757b7146 tempest-ServerMetadataNegativeTestJSON-936475568 tempest-ServerMetadataNegativeTestJSON-936475568-project-member] [instance: d179030a-e5ba-45b1-ad11-4d2f71ed7bd4] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d9:a4:69', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '1a8c8175-1197-4f12-baac-ef6aba95f585', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '51e75557-8c53-4ba4-982a-42e90b5decef', 'vif_model': 'vmxnet3'}] {{(pid=65726) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 862.101897] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-cb2d094a-412f-492d-9ef1-aba8757b7146 tempest-ServerMetadataNegativeTestJSON-936475568 tempest-ServerMetadataNegativeTestJSON-936475568-project-member] Creating folder: Project (921d46aee1824fd990022c7f851018b2). Parent ref: group-v995008. {{(pid=65726) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 862.101897] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a773218c-9e6b-4c1d-8ee3-a561701457f9 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.101897] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-114fce54-77db-4ded-b77e-3c7ade027b90 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.101897] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc04ab7c-e8cd-4296-9ee2-752638676446 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.101897] env[65726]: DEBUG nova.compute.provider_tree [None req-c4393951-32f6-44d6-b9c7-1ad480073dbe tempest-InstanceActionsNegativeTestJSON-844675881 tempest-InstanceActionsNegativeTestJSON-844675881-project-member] Updating inventory in ProviderTree for provider 07c4692f-bdb4-4058-9173-ff9664830295 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 862.101897] env[65726]: INFO nova.virt.vmwareapi.vm_util [None req-cb2d094a-412f-492d-9ef1-aba8757b7146 tempest-ServerMetadataNegativeTestJSON-936475568 tempest-ServerMetadataNegativeTestJSON-936475568-project-member] Created folder: Project (921d46aee1824fd990022c7f851018b2) in parent group-v995008. [ 862.101897] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-cb2d094a-412f-492d-9ef1-aba8757b7146 tempest-ServerMetadataNegativeTestJSON-936475568 tempest-ServerMetadataNegativeTestJSON-936475568-project-member] Creating folder: Instances. Parent ref: group-v995172. {{(pid=65726) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 862.101897] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-2d926d21-5a53-4d6e-83cd-78a95d9e69f3 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.103069] env[65726]: INFO nova.virt.vmwareapi.vm_util [None req-cb2d094a-412f-492d-9ef1-aba8757b7146 tempest-ServerMetadataNegativeTestJSON-936475568 tempest-ServerMetadataNegativeTestJSON-936475568-project-member] Created folder: Instances in parent group-v995172. [ 862.103069] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-cb2d094a-412f-492d-9ef1-aba8757b7146 tempest-ServerMetadataNegativeTestJSON-936475568 tempest-ServerMetadataNegativeTestJSON-936475568-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 862.103548] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d179030a-e5ba-45b1-ad11-4d2f71ed7bd4] Creating VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 862.103816] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-adf56230-8c0c-40cb-a110-8ce3506ded21 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.127934] env[65726]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 862.127934] env[65726]: value = "task-5116054" [ 862.127934] env[65726]: _type = "Task" [ 862.127934] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.141816] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116054, 'name': CreateVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.200725] env[65726]: DEBUG oslo_vmware.api [None req-d7ed2ba1-3808-4f50-920c-ef8e5d179423 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Task: {'id': task-5116047, 'name': Rename_Task, 'duration_secs': 0.220664} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.201204] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-d7ed2ba1-3808-4f50-920c-ef8e5d179423 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] [instance: 44d73b2d-2b6a-4501-9944-432da3c9330d] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 862.201592] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e9ef6e4b-84d0-488c-8c81-2082baac51d7 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.212613] env[65726]: DEBUG oslo_vmware.api [None req-d7ed2ba1-3808-4f50-920c-ef8e5d179423 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Waiting for the task: (returnval){ [ 862.212613] env[65726]: value = "task-5116055" [ 862.212613] env[65726]: _type = "Task" [ 862.212613] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.234451] env[65726]: DEBUG oslo_vmware.api [None req-d7ed2ba1-3808-4f50-920c-ef8e5d179423 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Task: {'id': task-5116055, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.262064] env[65726]: DEBUG oslo_vmware.api [None req-b0e194cf-16dc-40d0-b65c-88bd21cd63e3 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Task: {'id': task-5116048, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.332372] env[65726]: DEBUG oslo_vmware.api [None req-55d74d77-063b-418b-b818-59d05c490f34 tempest-ImagesNegativeTestJSON-1015404931 tempest-ImagesNegativeTestJSON-1015404931-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5252cf09-964a-cce3-1208-42ac163bd58d, 'name': SearchDatastore_Task, 'duration_secs': 0.012999} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.333213] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9bf38cb8-29dd-4df3-bfb2-c90d6e1a52ac {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.340625] env[65726]: DEBUG oslo_vmware.api [None req-55d74d77-063b-418b-b818-59d05c490f34 tempest-ImagesNegativeTestJSON-1015404931 tempest-ImagesNegativeTestJSON-1015404931-project-member] Waiting for the task: (returnval){ [ 862.340625] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]522b3e01-475d-d7b1-9121-4478ee29aae2" [ 862.340625] env[65726]: _type = "Task" [ 862.340625] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.350758] env[65726]: DEBUG oslo_vmware.api [None req-55d74d77-063b-418b-b818-59d05c490f34 tempest-ImagesNegativeTestJSON-1015404931 tempest-ImagesNegativeTestJSON-1015404931-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]522b3e01-475d-d7b1-9121-4478ee29aae2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.370428] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116051, 'name': CreateVM_Task} progress is 99%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.557741] env[65726]: WARNING openstack [req-0d5d700d-ae02-452d-bad7-37d024b6ab53 req-0e204bef-dd48-4319-a628-536f757be00d service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 862.559220] env[65726]: WARNING openstack [req-0d5d700d-ae02-452d-bad7-37d024b6ab53 req-0e204bef-dd48-4319-a628-536f757be00d service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 862.625393] env[65726]: DEBUG nova.scheduler.client.report [None req-c4393951-32f6-44d6-b9c7-1ad480073dbe tempest-InstanceActionsNegativeTestJSON-844675881 tempest-InstanceActionsNegativeTestJSON-844675881-project-member] Updated inventory for provider 07c4692f-bdb4-4058-9173-ff9664830295 with generation 96 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 862.625952] env[65726]: DEBUG nova.compute.provider_tree [None req-c4393951-32f6-44d6-b9c7-1ad480073dbe tempest-InstanceActionsNegativeTestJSON-844675881 tempest-InstanceActionsNegativeTestJSON-844675881-project-member] Updating resource provider 07c4692f-bdb4-4058-9173-ff9664830295 generation from 96 to 97 during operation: update_inventory {{(pid=65726) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 862.625952] env[65726]: DEBUG nova.compute.provider_tree [None req-c4393951-32f6-44d6-b9c7-1ad480073dbe tempest-InstanceActionsNegativeTestJSON-844675881 tempest-InstanceActionsNegativeTestJSON-844675881-project-member] Updating inventory in ProviderTree for provider 07c4692f-bdb4-4058-9173-ff9664830295 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 862.640020] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116054, 'name': CreateVM_Task, 'duration_secs': 0.399209} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.640020] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d179030a-e5ba-45b1-ad11-4d2f71ed7bd4] Created VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 862.640656] env[65726]: WARNING openstack [None req-cb2d094a-412f-492d-9ef1-aba8757b7146 tempest-ServerMetadataNegativeTestJSON-936475568 tempest-ServerMetadataNegativeTestJSON-936475568-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 862.641028] env[65726]: WARNING openstack [None req-cb2d094a-412f-492d-9ef1-aba8757b7146 tempest-ServerMetadataNegativeTestJSON-936475568 tempest-ServerMetadataNegativeTestJSON-936475568-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 862.645969] env[65726]: DEBUG oslo_concurrency.lockutils [None req-cb2d094a-412f-492d-9ef1-aba8757b7146 tempest-ServerMetadataNegativeTestJSON-936475568 tempest-ServerMetadataNegativeTestJSON-936475568-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 862.646146] env[65726]: DEBUG oslo_concurrency.lockutils [None req-cb2d094a-412f-492d-9ef1-aba8757b7146 tempest-ServerMetadataNegativeTestJSON-936475568 tempest-ServerMetadataNegativeTestJSON-936475568-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 862.646436] env[65726]: DEBUG oslo_concurrency.lockutils [None req-cb2d094a-412f-492d-9ef1-aba8757b7146 tempest-ServerMetadataNegativeTestJSON-936475568 tempest-ServerMetadataNegativeTestJSON-936475568-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 862.647346] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f219afe9-eacb-44b9-837a-b1a2f93a1832 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.657545] env[65726]: DEBUG oslo_vmware.api [None req-cb2d094a-412f-492d-9ef1-aba8757b7146 tempest-ServerMetadataNegativeTestJSON-936475568 tempest-ServerMetadataNegativeTestJSON-936475568-project-member] Waiting for the task: (returnval){ [ 862.657545] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52fa9c9a-be38-5848-df40-05c830f22c12" [ 862.657545] env[65726]: _type = "Task" [ 862.657545] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.668443] env[65726]: DEBUG oslo_vmware.api [None req-cb2d094a-412f-492d-9ef1-aba8757b7146 tempest-ServerMetadataNegativeTestJSON-936475568 tempest-ServerMetadataNegativeTestJSON-936475568-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52fa9c9a-be38-5848-df40-05c830f22c12, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.675016] env[65726]: WARNING openstack [req-0d5d700d-ae02-452d-bad7-37d024b6ab53 req-0e204bef-dd48-4319-a628-536f757be00d service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 862.675403] env[65726]: WARNING openstack [req-0d5d700d-ae02-452d-bad7-37d024b6ab53 req-0e204bef-dd48-4319-a628-536f757be00d service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 862.725032] env[65726]: DEBUG oslo_vmware.api [None req-d7ed2ba1-3808-4f50-920c-ef8e5d179423 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Task: {'id': task-5116055, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.739905] env[65726]: WARNING openstack [req-0d5d700d-ae02-452d-bad7-37d024b6ab53 req-0e204bef-dd48-4319-a628-536f757be00d service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 862.740130] env[65726]: WARNING openstack [req-0d5d700d-ae02-452d-bad7-37d024b6ab53 req-0e204bef-dd48-4319-a628-536f757be00d service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 862.760684] env[65726]: DEBUG oslo_vmware.api [None req-b0e194cf-16dc-40d0-b65c-88bd21cd63e3 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Task: {'id': task-5116048, 'name': PowerOnVM_Task, 'duration_secs': 0.53683} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.760966] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-b0e194cf-16dc-40d0-b65c-88bd21cd63e3 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: cb8855b6-589d-4863-b86f-7fb4e30fb29e] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 862.761181] env[65726]: INFO nova.compute.manager [None req-b0e194cf-16dc-40d0-b65c-88bd21cd63e3 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: cb8855b6-589d-4863-b86f-7fb4e30fb29e] Took 10.37 seconds to spawn the instance on the hypervisor. [ 862.761373] env[65726]: DEBUG nova.compute.manager [None req-b0e194cf-16dc-40d0-b65c-88bd21cd63e3 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: cb8855b6-589d-4863-b86f-7fb4e30fb29e] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 862.762169] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73031dc1-55ba-44bc-8935-44f1b89bbdac {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.827042] env[65726]: DEBUG nova.network.neutron [req-0d5d700d-ae02-452d-bad7-37d024b6ab53 req-0e204bef-dd48-4319-a628-536f757be00d service nova] [instance: d179030a-e5ba-45b1-ad11-4d2f71ed7bd4] Updated VIF entry in instance network info cache for port 51e75557-8c53-4ba4-982a-42e90b5decef. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 862.827398] env[65726]: DEBUG nova.network.neutron [req-0d5d700d-ae02-452d-bad7-37d024b6ab53 req-0e204bef-dd48-4319-a628-536f757be00d service nova] [instance: d179030a-e5ba-45b1-ad11-4d2f71ed7bd4] Updating instance_info_cache with network_info: [{"id": "51e75557-8c53-4ba4-982a-42e90b5decef", "address": "fa:16:3e:d9:a4:69", "network": {"id": "c9aee0fe-f17c-4207-9bae-b92c4f507fb3", "bridge": "br-int", "label": "tempest-ServerMetadataNegativeTestJSON-1399638803-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "921d46aee1824fd990022c7f851018b2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1a8c8175-1197-4f12-baac-ef6aba95f585", "external-id": "nsx-vlan-transportzone-832", "segmentation_id": 832, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap51e75557-8c", "ovs_interfaceid": "51e75557-8c53-4ba4-982a-42e90b5decef", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 862.851607] env[65726]: DEBUG oslo_vmware.api [None req-55d74d77-063b-418b-b818-59d05c490f34 tempest-ImagesNegativeTestJSON-1015404931 tempest-ImagesNegativeTestJSON-1015404931-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]522b3e01-475d-d7b1-9121-4478ee29aae2, 'name': SearchDatastore_Task, 'duration_secs': 0.013973} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.851890] env[65726]: DEBUG oslo_concurrency.lockutils [None req-55d74d77-063b-418b-b818-59d05c490f34 tempest-ImagesNegativeTestJSON-1015404931 tempest-ImagesNegativeTestJSON-1015404931-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 862.852165] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-55d74d77-063b-418b-b818-59d05c490f34 tempest-ImagesNegativeTestJSON-1015404931 tempest-ImagesNegativeTestJSON-1015404931-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] c4177e20-b1bd-4b54-a275-c93582359a07/c4177e20-b1bd-4b54-a275-c93582359a07.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 862.852433] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c0a2bb35-2128-4929-98b1-dd22edc011b9 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.859481] env[65726]: DEBUG oslo_vmware.api [None req-55d74d77-063b-418b-b818-59d05c490f34 tempest-ImagesNegativeTestJSON-1015404931 tempest-ImagesNegativeTestJSON-1015404931-project-member] Waiting for the task: (returnval){ [ 862.859481] env[65726]: value = "task-5116056" [ 862.859481] env[65726]: _type = "Task" [ 862.859481] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.871273] env[65726]: DEBUG oslo_vmware.api [None req-55d74d77-063b-418b-b818-59d05c490f34 tempest-ImagesNegativeTestJSON-1015404931 tempest-ImagesNegativeTestJSON-1015404931-project-member] Task: {'id': task-5116056, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.874964] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116051, 'name': CreateVM_Task} progress is 99%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.923800] env[65726]: WARNING openstack [None req-253a2ea4-406a-415c-8211-e66a1f71d581 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 862.924758] env[65726]: WARNING openstack [None req-253a2ea4-406a-415c-8211-e66a1f71d581 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 862.963633] env[65726]: WARNING openstack [None req-253a2ea4-406a-415c-8211-e66a1f71d581 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 862.964040] env[65726]: WARNING openstack [None req-253a2ea4-406a-415c-8211-e66a1f71d581 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 862.968865] env[65726]: WARNING neutronclient.v2_0.client [None req-253a2ea4-406a-415c-8211-e66a1f71d581 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 863.006081] env[65726]: DEBUG nova.network.neutron [None req-253a2ea4-406a-415c-8211-e66a1f71d581 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] [instance: 76249623-6f83-46a3-b8c5-c001111aa698] Port e97a5da4-5540-4c0c-a33d-be1182e82aa7 binding to destination host cpu-1 is already ACTIVE {{(pid=65726) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3236}} [ 863.006373] env[65726]: DEBUG oslo_concurrency.lockutils [None req-253a2ea4-406a-415c-8211-e66a1f71d581 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Acquiring lock "refresh_cache-76249623-6f83-46a3-b8c5-c001111aa698" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 863.006521] env[65726]: DEBUG oslo_concurrency.lockutils [None req-253a2ea4-406a-415c-8211-e66a1f71d581 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Acquired lock "refresh_cache-76249623-6f83-46a3-b8c5-c001111aa698" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 863.006691] env[65726]: DEBUG nova.network.neutron [None req-253a2ea4-406a-415c-8211-e66a1f71d581 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] [instance: 76249623-6f83-46a3-b8c5-c001111aa698] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 863.134942] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c4393951-32f6-44d6-b9c7-1ad480073dbe tempest-InstanceActionsNegativeTestJSON-844675881 tempest-InstanceActionsNegativeTestJSON-844675881-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 3.464s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 863.137733] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e1137661-56aa-48f2-a91c-68a6f9993ed2 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 35.923s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 863.137964] env[65726]: DEBUG nova.objects.instance [None req-e1137661-56aa-48f2-a91c-68a6f9993ed2 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Lazy-loading 'resources' on Instance uuid 49bcab1f-7b64-4999-abff-37771c58a271 {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 863.157285] env[65726]: INFO nova.scheduler.client.report [None req-c4393951-32f6-44d6-b9c7-1ad480073dbe tempest-InstanceActionsNegativeTestJSON-844675881 tempest-InstanceActionsNegativeTestJSON-844675881-project-member] Deleted allocations for instance 1bd26aef-995e-43b6-af9f-077e878c8c44 [ 863.172498] env[65726]: DEBUG oslo_vmware.api [None req-cb2d094a-412f-492d-9ef1-aba8757b7146 tempest-ServerMetadataNegativeTestJSON-936475568 tempest-ServerMetadataNegativeTestJSON-936475568-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52fa9c9a-be38-5848-df40-05c830f22c12, 'name': SearchDatastore_Task, 'duration_secs': 0.01386} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.173085] env[65726]: DEBUG oslo_concurrency.lockutils [None req-cb2d094a-412f-492d-9ef1-aba8757b7146 tempest-ServerMetadataNegativeTestJSON-936475568 tempest-ServerMetadataNegativeTestJSON-936475568-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 863.173085] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-cb2d094a-412f-492d-9ef1-aba8757b7146 tempest-ServerMetadataNegativeTestJSON-936475568 tempest-ServerMetadataNegativeTestJSON-936475568-project-member] [instance: d179030a-e5ba-45b1-ad11-4d2f71ed7bd4] Processing image b52362a3-ee8a-4cbf-b06f-513b0cc8f95c {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 863.173319] env[65726]: DEBUG oslo_concurrency.lockutils [None req-cb2d094a-412f-492d-9ef1-aba8757b7146 tempest-ServerMetadataNegativeTestJSON-936475568 tempest-ServerMetadataNegativeTestJSON-936475568-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 863.173719] env[65726]: DEBUG oslo_concurrency.lockutils [None req-cb2d094a-412f-492d-9ef1-aba8757b7146 tempest-ServerMetadataNegativeTestJSON-936475568 tempest-ServerMetadataNegativeTestJSON-936475568-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 863.173719] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-cb2d094a-412f-492d-9ef1-aba8757b7146 tempest-ServerMetadataNegativeTestJSON-936475568 tempest-ServerMetadataNegativeTestJSON-936475568-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 863.173887] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7b10974f-e706-4a04-aa75-c2cfb0653c10 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.184772] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-cb2d094a-412f-492d-9ef1-aba8757b7146 tempest-ServerMetadataNegativeTestJSON-936475568 tempest-ServerMetadataNegativeTestJSON-936475568-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 863.185177] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-cb2d094a-412f-492d-9ef1-aba8757b7146 tempest-ServerMetadataNegativeTestJSON-936475568 tempest-ServerMetadataNegativeTestJSON-936475568-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65726) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 863.185743] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9e8d5b2b-36a9-4aca-94a7-b98dbc1e99ec {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.194176] env[65726]: DEBUG oslo_vmware.api [None req-cb2d094a-412f-492d-9ef1-aba8757b7146 tempest-ServerMetadataNegativeTestJSON-936475568 tempest-ServerMetadataNegativeTestJSON-936475568-project-member] Waiting for the task: (returnval){ [ 863.194176] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5217f590-c367-65f1-87f8-5b777e0ca6c4" [ 863.194176] env[65726]: _type = "Task" [ 863.194176] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.202960] env[65726]: DEBUG oslo_vmware.api [None req-cb2d094a-412f-492d-9ef1-aba8757b7146 tempest-ServerMetadataNegativeTestJSON-936475568 tempest-ServerMetadataNegativeTestJSON-936475568-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5217f590-c367-65f1-87f8-5b777e0ca6c4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.224486] env[65726]: DEBUG oslo_vmware.api [None req-d7ed2ba1-3808-4f50-920c-ef8e5d179423 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Task: {'id': task-5116055, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.280860] env[65726]: INFO nova.compute.manager [None req-b0e194cf-16dc-40d0-b65c-88bd21cd63e3 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: cb8855b6-589d-4863-b86f-7fb4e30fb29e] Took 60.09 seconds to build instance. [ 863.331145] env[65726]: DEBUG oslo_concurrency.lockutils [req-0d5d700d-ae02-452d-bad7-37d024b6ab53 req-0e204bef-dd48-4319-a628-536f757be00d service nova] Releasing lock "refresh_cache-d179030a-e5ba-45b1-ad11-4d2f71ed7bd4" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 863.372816] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116051, 'name': CreateVM_Task, 'duration_secs': 1.394954} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.376263] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2cbee359-a1fc-45c0-bcc5-a2c24ded9c69] Created VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 863.376566] env[65726]: DEBUG oslo_vmware.api [None req-55d74d77-063b-418b-b818-59d05c490f34 tempest-ImagesNegativeTestJSON-1015404931 tempest-ImagesNegativeTestJSON-1015404931-project-member] Task: {'id': task-5116056, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.376959] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b3ea747b-af8b-4592-b662-6643bc7f39d3 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 863.377175] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b3ea747b-af8b-4592-b662-6643bc7f39d3 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 863.377437] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b3ea747b-af8b-4592-b662-6643bc7f39d3 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 863.377689] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-08099bb2-c06a-4118-a70d-763ba1f62fa1 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.383112] env[65726]: DEBUG oslo_vmware.api [None req-b3ea747b-af8b-4592-b662-6643bc7f39d3 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] Waiting for the task: (returnval){ [ 863.383112] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52d3c82c-91e6-163e-57d2-0ef9244b15fd" [ 863.383112] env[65726]: _type = "Task" [ 863.383112] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.391874] env[65726]: DEBUG oslo_vmware.api [None req-b3ea747b-af8b-4592-b662-6643bc7f39d3 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52d3c82c-91e6-163e-57d2-0ef9244b15fd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.476844] env[65726]: DEBUG oslo_concurrency.lockutils [None req-8644ddd8-ab30-4b52-b188-a0427e248610 tempest-ServersAdminTestJSON-1259723601 tempest-ServersAdminTestJSON-1259723601-project-admin] Acquiring lock "refresh_cache-cb8855b6-589d-4863-b86f-7fb4e30fb29e" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 863.477098] env[65726]: DEBUG oslo_concurrency.lockutils [None req-8644ddd8-ab30-4b52-b188-a0427e248610 tempest-ServersAdminTestJSON-1259723601 tempest-ServersAdminTestJSON-1259723601-project-admin] Acquired lock "refresh_cache-cb8855b6-589d-4863-b86f-7fb4e30fb29e" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 863.477371] env[65726]: DEBUG nova.network.neutron [None req-8644ddd8-ab30-4b52-b188-a0427e248610 tempest-ServersAdminTestJSON-1259723601 tempest-ServersAdminTestJSON-1259723601-project-admin] [instance: cb8855b6-589d-4863-b86f-7fb4e30fb29e] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 863.510125] env[65726]: WARNING openstack [None req-253a2ea4-406a-415c-8211-e66a1f71d581 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 863.510598] env[65726]: WARNING openstack [None req-253a2ea4-406a-415c-8211-e66a1f71d581 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 863.662074] env[65726]: WARNING openstack [None req-253a2ea4-406a-415c-8211-e66a1f71d581 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 863.662508] env[65726]: WARNING openstack [None req-253a2ea4-406a-415c-8211-e66a1f71d581 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 863.673886] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c4393951-32f6-44d6-b9c7-1ad480073dbe tempest-InstanceActionsNegativeTestJSON-844675881 tempest-InstanceActionsNegativeTestJSON-844675881-project-member] Lock "1bd26aef-995e-43b6-af9f-077e878c8c44" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 40.968s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 863.719433] env[65726]: DEBUG oslo_vmware.api [None req-cb2d094a-412f-492d-9ef1-aba8757b7146 tempest-ServerMetadataNegativeTestJSON-936475568 tempest-ServerMetadataNegativeTestJSON-936475568-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5217f590-c367-65f1-87f8-5b777e0ca6c4, 'name': SearchDatastore_Task, 'duration_secs': 0.013047} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.727894] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f9e9c8c7-e618-4b4f-8ae4-211ea56fd0e9 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.738293] env[65726]: DEBUG oslo_vmware.api [None req-d7ed2ba1-3808-4f50-920c-ef8e5d179423 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Task: {'id': task-5116055, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.739517] env[65726]: DEBUG oslo_vmware.api [None req-cb2d094a-412f-492d-9ef1-aba8757b7146 tempest-ServerMetadataNegativeTestJSON-936475568 tempest-ServerMetadataNegativeTestJSON-936475568-project-member] Waiting for the task: (returnval){ [ 863.739517] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52e00500-1fdc-88af-99b7-51de43b2a0f4" [ 863.739517] env[65726]: _type = "Task" [ 863.739517] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.749850] env[65726]: WARNING openstack [None req-253a2ea4-406a-415c-8211-e66a1f71d581 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 863.750284] env[65726]: WARNING openstack [None req-253a2ea4-406a-415c-8211-e66a1f71d581 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 863.765568] env[65726]: DEBUG oslo_vmware.api [None req-cb2d094a-412f-492d-9ef1-aba8757b7146 tempest-ServerMetadataNegativeTestJSON-936475568 tempest-ServerMetadataNegativeTestJSON-936475568-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52e00500-1fdc-88af-99b7-51de43b2a0f4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.783746] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b0e194cf-16dc-40d0-b65c-88bd21cd63e3 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Lock "cb8855b6-589d-4863-b86f-7fb4e30fb29e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 93.952s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 863.861969] env[65726]: DEBUG nova.network.neutron [None req-253a2ea4-406a-415c-8211-e66a1f71d581 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] [instance: 76249623-6f83-46a3-b8c5-c001111aa698] Updating instance_info_cache with network_info: [{"id": "e97a5da4-5540-4c0c-a33d-be1182e82aa7", "address": "fa:16:3e:af:b6:0f", "network": {"id": "55d629b8-a541-4720-9289-927bbc6e6f22", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.253", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "1539f06fa3534e90acbc3a60c4b8bd3f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "35342bcb-8b06-472e-b3c0-43fd3d6c4b30", "external-id": "nsx-vlan-transportzone-524", "segmentation_id": 524, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape97a5da4-55", "ovs_interfaceid": "e97a5da4-5540-4c0c-a33d-be1182e82aa7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 863.875041] env[65726]: DEBUG oslo_vmware.api [None req-55d74d77-063b-418b-b818-59d05c490f34 tempest-ImagesNegativeTestJSON-1015404931 tempest-ImagesNegativeTestJSON-1015404931-project-member] Task: {'id': task-5116056, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.956263} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.876068] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-55d74d77-063b-418b-b818-59d05c490f34 tempest-ImagesNegativeTestJSON-1015404931 tempest-ImagesNegativeTestJSON-1015404931-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] c4177e20-b1bd-4b54-a275-c93582359a07/c4177e20-b1bd-4b54-a275-c93582359a07.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 863.876290] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-55d74d77-063b-418b-b818-59d05c490f34 tempest-ImagesNegativeTestJSON-1015404931 tempest-ImagesNegativeTestJSON-1015404931-project-member] [instance: c4177e20-b1bd-4b54-a275-c93582359a07] Extending root virtual disk to 1048576 {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 863.876542] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-41a9a647-47e0-45a2-bd87-9809194dbcdd {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.890057] env[65726]: DEBUG oslo_vmware.api [None req-55d74d77-063b-418b-b818-59d05c490f34 tempest-ImagesNegativeTestJSON-1015404931 tempest-ImagesNegativeTestJSON-1015404931-project-member] Waiting for the task: (returnval){ [ 863.890057] env[65726]: value = "task-5116057" [ 863.890057] env[65726]: _type = "Task" [ 863.890057] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.897389] env[65726]: DEBUG oslo_vmware.api [None req-b3ea747b-af8b-4592-b662-6643bc7f39d3 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52d3c82c-91e6-163e-57d2-0ef9244b15fd, 'name': SearchDatastore_Task, 'duration_secs': 0.031376} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.898849] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b3ea747b-af8b-4592-b662-6643bc7f39d3 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 863.898929] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-b3ea747b-af8b-4592-b662-6643bc7f39d3 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] [instance: 2cbee359-a1fc-45c0-bcc5-a2c24ded9c69] Processing image b52362a3-ee8a-4cbf-b06f-513b0cc8f95c {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 863.899217] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b3ea747b-af8b-4592-b662-6643bc7f39d3 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 863.899439] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b3ea747b-af8b-4592-b662-6643bc7f39d3 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 863.899654] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-b3ea747b-af8b-4592-b662-6643bc7f39d3 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 863.905250] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e2df1ca8-9d22-4a8c-85c1-0fbb952c6e0e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.907123] env[65726]: DEBUG oslo_vmware.api [None req-55d74d77-063b-418b-b818-59d05c490f34 tempest-ImagesNegativeTestJSON-1015404931 tempest-ImagesNegativeTestJSON-1015404931-project-member] Task: {'id': task-5116057, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.916770] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-b3ea747b-af8b-4592-b662-6643bc7f39d3 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 863.917023] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-b3ea747b-af8b-4592-b662-6643bc7f39d3 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65726) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 863.917758] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a0051e67-b926-4bbc-bafa-da300cf49bc2 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.924920] env[65726]: DEBUG oslo_vmware.api [None req-b3ea747b-af8b-4592-b662-6643bc7f39d3 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] Waiting for the task: (returnval){ [ 863.924920] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]526ca8c1-bc95-3c7d-c110-146e069564bc" [ 863.924920] env[65726]: _type = "Task" [ 863.924920] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.933236] env[65726]: DEBUG oslo_vmware.api [None req-b3ea747b-af8b-4592-b662-6643bc7f39d3 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]526ca8c1-bc95-3c7d-c110-146e069564bc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.981478] env[65726]: WARNING openstack [None req-8644ddd8-ab30-4b52-b188-a0427e248610 tempest-ServersAdminTestJSON-1259723601 tempest-ServersAdminTestJSON-1259723601-project-admin] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 863.981896] env[65726]: WARNING openstack [None req-8644ddd8-ab30-4b52-b188-a0427e248610 tempest-ServersAdminTestJSON-1259723601 tempest-ServersAdminTestJSON-1259723601-project-admin] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 864.154596] env[65726]: WARNING openstack [None req-8644ddd8-ab30-4b52-b188-a0427e248610 tempest-ServersAdminTestJSON-1259723601 tempest-ServersAdminTestJSON-1259723601-project-admin] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 864.154950] env[65726]: WARNING openstack [None req-8644ddd8-ab30-4b52-b188-a0427e248610 tempest-ServersAdminTestJSON-1259723601 tempest-ServersAdminTestJSON-1259723601-project-admin] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 864.231119] env[65726]: DEBUG oslo_vmware.api [None req-d7ed2ba1-3808-4f50-920c-ef8e5d179423 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Task: {'id': task-5116055, 'name': PowerOnVM_Task, 'duration_secs': 1.644846} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.231403] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-d7ed2ba1-3808-4f50-920c-ef8e5d179423 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] [instance: 44d73b2d-2b6a-4501-9944-432da3c9330d] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 864.231696] env[65726]: DEBUG nova.compute.manager [None req-d7ed2ba1-3808-4f50-920c-ef8e5d179423 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] [instance: 44d73b2d-2b6a-4501-9944-432da3c9330d] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 864.232601] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8d74794-f6aa-4083-9a4f-7daa8690695f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.255655] env[65726]: DEBUG oslo_vmware.api [None req-cb2d094a-412f-492d-9ef1-aba8757b7146 tempest-ServerMetadataNegativeTestJSON-936475568 tempest-ServerMetadataNegativeTestJSON-936475568-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52e00500-1fdc-88af-99b7-51de43b2a0f4, 'name': SearchDatastore_Task, 'duration_secs': 0.037675} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.255898] env[65726]: DEBUG oslo_concurrency.lockutils [None req-cb2d094a-412f-492d-9ef1-aba8757b7146 tempest-ServerMetadataNegativeTestJSON-936475568 tempest-ServerMetadataNegativeTestJSON-936475568-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 864.256176] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-cb2d094a-412f-492d-9ef1-aba8757b7146 tempest-ServerMetadataNegativeTestJSON-936475568 tempest-ServerMetadataNegativeTestJSON-936475568-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] d179030a-e5ba-45b1-ad11-4d2f71ed7bd4/d179030a-e5ba-45b1-ad11-4d2f71ed7bd4.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 864.256447] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-61a7aac3-fdb9-4367-92f1-8b6b53ca9ada {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.260363] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa5e0a75-33f5-4425-8806-3d765c948be4 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.270199] env[65726]: WARNING openstack [None req-8644ddd8-ab30-4b52-b188-a0427e248610 tempest-ServersAdminTestJSON-1259723601 tempest-ServersAdminTestJSON-1259723601-project-admin] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 864.270471] env[65726]: WARNING openstack [None req-8644ddd8-ab30-4b52-b188-a0427e248610 tempest-ServersAdminTestJSON-1259723601 tempest-ServersAdminTestJSON-1259723601-project-admin] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 864.291249] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df370eb1-6ba9-4749-91f6-69e256d0442b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.298949] env[65726]: DEBUG nova.compute.manager [None req-9c76632b-07ea-476e-92d3-ffb457aef721 tempest-ServerAddressesTestJSON-569100133 tempest-ServerAddressesTestJSON-569100133-project-member] [instance: 70375746-06dd-4710-9136-95e9e5759a0e] Starting instance... {{(pid=65726) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 864.301514] env[65726]: DEBUG oslo_vmware.api [None req-cb2d094a-412f-492d-9ef1-aba8757b7146 tempest-ServerMetadataNegativeTestJSON-936475568 tempest-ServerMetadataNegativeTestJSON-936475568-project-member] Waiting for the task: (returnval){ [ 864.301514] env[65726]: value = "task-5116058" [ 864.301514] env[65726]: _type = "Task" [ 864.301514] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.342785] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5adce485-1292-48d1-aed1-b696ea60c771 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.351445] env[65726]: DEBUG oslo_vmware.api [None req-cb2d094a-412f-492d-9ef1-aba8757b7146 tempest-ServerMetadataNegativeTestJSON-936475568 tempest-ServerMetadataNegativeTestJSON-936475568-project-member] Task: {'id': task-5116058, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.357446] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ae4c14b-ed2d-4712-a4be-9547940570de {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.371855] env[65726]: DEBUG oslo_concurrency.lockutils [None req-253a2ea4-406a-415c-8211-e66a1f71d581 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Releasing lock "refresh_cache-76249623-6f83-46a3-b8c5-c001111aa698" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 864.374518] env[65726]: DEBUG nova.compute.provider_tree [None req-e1137661-56aa-48f2-a91c-68a6f9993ed2 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 864.400979] env[65726]: DEBUG nova.network.neutron [None req-8644ddd8-ab30-4b52-b188-a0427e248610 tempest-ServersAdminTestJSON-1259723601 tempest-ServersAdminTestJSON-1259723601-project-admin] [instance: cb8855b6-589d-4863-b86f-7fb4e30fb29e] Updating instance_info_cache with network_info: [{"id": "136da8b5-d8fd-4cef-ae44-a9ecc98c28d2", "address": "fa:16:3e:22:7a:70", "network": {"id": "3f7916f7-ee9c-4629-b3b3-1367de9f98ea", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-579713617-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b794c99309e94c1181fc9d8eeeb84702", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d0bdd11b-58af-4cc0-9d38-8322e1bb4e74", "external-id": "nsx-vlan-transportzone-398", "segmentation_id": 398, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap136da8b5-d8", "ovs_interfaceid": "136da8b5-d8fd-4cef-ae44-a9ecc98c28d2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 864.405549] env[65726]: DEBUG oslo_vmware.api [None req-55d74d77-063b-418b-b818-59d05c490f34 tempest-ImagesNegativeTestJSON-1015404931 tempest-ImagesNegativeTestJSON-1015404931-project-member] Task: {'id': task-5116057, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.077423} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.405891] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-55d74d77-063b-418b-b818-59d05c490f34 tempest-ImagesNegativeTestJSON-1015404931 tempest-ImagesNegativeTestJSON-1015404931-project-member] [instance: c4177e20-b1bd-4b54-a275-c93582359a07] Extended root virtual disk {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 864.406700] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-640cb770-6353-425c-85b0-4ecf47eef4bd {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.432617] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-55d74d77-063b-418b-b818-59d05c490f34 tempest-ImagesNegativeTestJSON-1015404931 tempest-ImagesNegativeTestJSON-1015404931-project-member] [instance: c4177e20-b1bd-4b54-a275-c93582359a07] Reconfiguring VM instance instance-0000003a to attach disk [datastore1] c4177e20-b1bd-4b54-a275-c93582359a07/c4177e20-b1bd-4b54-a275-c93582359a07.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 864.432994] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4725da8f-2f11-41a3-b98e-b3b90c82ca22 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.457703] env[65726]: DEBUG oslo_vmware.api [None req-b3ea747b-af8b-4592-b662-6643bc7f39d3 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]526ca8c1-bc95-3c7d-c110-146e069564bc, 'name': SearchDatastore_Task, 'duration_secs': 0.016753} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.459526] env[65726]: DEBUG oslo_vmware.api [None req-55d74d77-063b-418b-b818-59d05c490f34 tempest-ImagesNegativeTestJSON-1015404931 tempest-ImagesNegativeTestJSON-1015404931-project-member] Waiting for the task: (returnval){ [ 864.459526] env[65726]: value = "task-5116059" [ 864.459526] env[65726]: _type = "Task" [ 864.459526] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.460136] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ee92157f-3848-47c4-bd47-59459dd06d6a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.472222] env[65726]: DEBUG oslo_vmware.api [None req-55d74d77-063b-418b-b818-59d05c490f34 tempest-ImagesNegativeTestJSON-1015404931 tempest-ImagesNegativeTestJSON-1015404931-project-member] Task: {'id': task-5116059, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.473413] env[65726]: DEBUG oslo_vmware.api [None req-b3ea747b-af8b-4592-b662-6643bc7f39d3 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] Waiting for the task: (returnval){ [ 864.473413] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52b0bae1-e7c2-e8a3-4b06-daf4b7c7856c" [ 864.473413] env[65726]: _type = "Task" [ 864.473413] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.482956] env[65726]: DEBUG oslo_vmware.api [None req-b3ea747b-af8b-4592-b662-6643bc7f39d3 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52b0bae1-e7c2-e8a3-4b06-daf4b7c7856c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.760815] env[65726]: DEBUG oslo_concurrency.lockutils [None req-d7ed2ba1-3808-4f50-920c-ef8e5d179423 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 864.816572] env[65726]: DEBUG oslo_vmware.api [None req-cb2d094a-412f-492d-9ef1-aba8757b7146 tempest-ServerMetadataNegativeTestJSON-936475568 tempest-ServerMetadataNegativeTestJSON-936475568-project-member] Task: {'id': task-5116058, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.827738] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9c76632b-07ea-476e-92d3-ffb457aef721 tempest-ServerAddressesTestJSON-569100133 tempest-ServerAddressesTestJSON-569100133-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 864.879045] env[65726]: DEBUG nova.compute.manager [None req-253a2ea4-406a-415c-8211-e66a1f71d581 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] [instance: 76249623-6f83-46a3-b8c5-c001111aa698] Hypervisor driver does not support instance shared storage check, assuming it's not on shared storage {{(pid=65726) _is_instance_storage_shared /opt/stack/nova/nova/compute/manager.py:924}} [ 864.879045] env[65726]: DEBUG oslo_concurrency.lockutils [None req-253a2ea4-406a-415c-8211-e66a1f71d581 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 864.879852] env[65726]: DEBUG nova.scheduler.client.report [None req-e1137661-56aa-48f2-a91c-68a6f9993ed2 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 864.906815] env[65726]: DEBUG oslo_concurrency.lockutils [None req-8644ddd8-ab30-4b52-b188-a0427e248610 tempest-ServersAdminTestJSON-1259723601 tempest-ServersAdminTestJSON-1259723601-project-admin] Releasing lock "refresh_cache-cb8855b6-589d-4863-b86f-7fb4e30fb29e" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 864.907072] env[65726]: DEBUG nova.compute.manager [None req-8644ddd8-ab30-4b52-b188-a0427e248610 tempest-ServersAdminTestJSON-1259723601 tempest-ServersAdminTestJSON-1259723601-project-admin] [instance: cb8855b6-589d-4863-b86f-7fb4e30fb29e] Inject network info {{(pid=65726) _inject_network_info /opt/stack/nova/nova/compute/manager.py:7794}} [ 864.907331] env[65726]: DEBUG nova.compute.manager [None req-8644ddd8-ab30-4b52-b188-a0427e248610 tempest-ServersAdminTestJSON-1259723601 tempest-ServersAdminTestJSON-1259723601-project-admin] [instance: cb8855b6-589d-4863-b86f-7fb4e30fb29e] network_info to inject: |[{"id": "136da8b5-d8fd-4cef-ae44-a9ecc98c28d2", "address": "fa:16:3e:22:7a:70", "network": {"id": "3f7916f7-ee9c-4629-b3b3-1367de9f98ea", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-579713617-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b794c99309e94c1181fc9d8eeeb84702", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d0bdd11b-58af-4cc0-9d38-8322e1bb4e74", "external-id": "nsx-vlan-transportzone-398", "segmentation_id": 398, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap136da8b5-d8", "ovs_interfaceid": "136da8b5-d8fd-4cef-ae44-a9ecc98c28d2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65726) _inject_network_info /opt/stack/nova/nova/compute/manager.py:7795}} [ 864.912898] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-8644ddd8-ab30-4b52-b188-a0427e248610 tempest-ServersAdminTestJSON-1259723601 tempest-ServersAdminTestJSON-1259723601-project-admin] [instance: cb8855b6-589d-4863-b86f-7fb4e30fb29e] Reconfiguring VM instance to set the machine id {{(pid=65726) _set_machine_id /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1796}} [ 864.913857] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a79f5ac6-b34f-4376-9182-c38d795fb44f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.935485] env[65726]: DEBUG oslo_vmware.api [None req-8644ddd8-ab30-4b52-b188-a0427e248610 tempest-ServersAdminTestJSON-1259723601 tempest-ServersAdminTestJSON-1259723601-project-admin] Waiting for the task: (returnval){ [ 864.935485] env[65726]: value = "task-5116060" [ 864.935485] env[65726]: _type = "Task" [ 864.935485] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.948197] env[65726]: DEBUG oslo_vmware.api [None req-8644ddd8-ab30-4b52-b188-a0427e248610 tempest-ServersAdminTestJSON-1259723601 tempest-ServersAdminTestJSON-1259723601-project-admin] Task: {'id': task-5116060, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.977172] env[65726]: DEBUG oslo_vmware.api [None req-55d74d77-063b-418b-b818-59d05c490f34 tempest-ImagesNegativeTestJSON-1015404931 tempest-ImagesNegativeTestJSON-1015404931-project-member] Task: {'id': task-5116059, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.987588] env[65726]: DEBUG oslo_vmware.api [None req-b3ea747b-af8b-4592-b662-6643bc7f39d3 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52b0bae1-e7c2-e8a3-4b06-daf4b7c7856c, 'name': SearchDatastore_Task, 'duration_secs': 0.019964} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.988071] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b3ea747b-af8b-4592-b662-6643bc7f39d3 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 864.988644] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-b3ea747b-af8b-4592-b662-6643bc7f39d3 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore2] 2cbee359-a1fc-45c0-bcc5-a2c24ded9c69/2cbee359-a1fc-45c0-bcc5-a2c24ded9c69.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 864.989010] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0c7053f6-20a3-448e-8dc4-1c0cd2bf849f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.005718] env[65726]: DEBUG oslo_vmware.api [None req-b3ea747b-af8b-4592-b662-6643bc7f39d3 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] Waiting for the task: (returnval){ [ 865.005718] env[65726]: value = "task-5116061" [ 865.005718] env[65726]: _type = "Task" [ 865.005718] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.019593] env[65726]: DEBUG oslo_vmware.api [None req-b3ea747b-af8b-4592-b662-6643bc7f39d3 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] Task: {'id': task-5116061, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.316085] env[65726]: DEBUG oslo_vmware.api [None req-cb2d094a-412f-492d-9ef1-aba8757b7146 tempest-ServerMetadataNegativeTestJSON-936475568 tempest-ServerMetadataNegativeTestJSON-936475568-project-member] Task: {'id': task-5116058, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.941111} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.316509] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-cb2d094a-412f-492d-9ef1-aba8757b7146 tempest-ServerMetadataNegativeTestJSON-936475568 tempest-ServerMetadataNegativeTestJSON-936475568-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] d179030a-e5ba-45b1-ad11-4d2f71ed7bd4/d179030a-e5ba-45b1-ad11-4d2f71ed7bd4.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 865.316837] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-cb2d094a-412f-492d-9ef1-aba8757b7146 tempest-ServerMetadataNegativeTestJSON-936475568 tempest-ServerMetadataNegativeTestJSON-936475568-project-member] [instance: d179030a-e5ba-45b1-ad11-4d2f71ed7bd4] Extending root virtual disk to 1048576 {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 865.317184] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1bdbd8a8-84a7-4d3e-9814-b456aef0b28d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.325267] env[65726]: DEBUG oslo_vmware.api [None req-cb2d094a-412f-492d-9ef1-aba8757b7146 tempest-ServerMetadataNegativeTestJSON-936475568 tempest-ServerMetadataNegativeTestJSON-936475568-project-member] Waiting for the task: (returnval){ [ 865.325267] env[65726]: value = "task-5116062" [ 865.325267] env[65726]: _type = "Task" [ 865.325267] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.337888] env[65726]: DEBUG oslo_vmware.api [None req-cb2d094a-412f-492d-9ef1-aba8757b7146 tempest-ServerMetadataNegativeTestJSON-936475568 tempest-ServerMetadataNegativeTestJSON-936475568-project-member] Task: {'id': task-5116062, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.388893] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e1137661-56aa-48f2-a91c-68a6f9993ed2 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.249s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 865.391244] env[65726]: DEBUG oslo_concurrency.lockutils [None req-49b35bf1-9410-49fb-9a5b-632c41ecd3a6 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 37.899s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 865.392416] env[65726]: INFO nova.compute.claims [None req-49b35bf1-9410-49fb-9a5b-632c41ecd3a6 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] [instance: c4cc4f85-cb35-4edc-a58b-adfee0ce1265] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 865.419790] env[65726]: INFO nova.scheduler.client.report [None req-e1137661-56aa-48f2-a91c-68a6f9993ed2 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Deleted allocations for instance 49bcab1f-7b64-4999-abff-37771c58a271 [ 865.448621] env[65726]: DEBUG oslo_vmware.api [None req-8644ddd8-ab30-4b52-b188-a0427e248610 tempest-ServersAdminTestJSON-1259723601 tempest-ServersAdminTestJSON-1259723601-project-admin] Task: {'id': task-5116060, 'name': ReconfigVM_Task, 'duration_secs': 0.407389} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.450017] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-8644ddd8-ab30-4b52-b188-a0427e248610 tempest-ServersAdminTestJSON-1259723601 tempest-ServersAdminTestJSON-1259723601-project-admin] [instance: cb8855b6-589d-4863-b86f-7fb4e30fb29e] Reconfigured VM instance to set the machine id {{(pid=65726) _set_machine_id /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1799}} [ 865.479967] env[65726]: DEBUG oslo_vmware.api [None req-55d74d77-063b-418b-b818-59d05c490f34 tempest-ImagesNegativeTestJSON-1015404931 tempest-ImagesNegativeTestJSON-1015404931-project-member] Task: {'id': task-5116059, 'name': ReconfigVM_Task, 'duration_secs': 0.924581} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.481514] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-55d74d77-063b-418b-b818-59d05c490f34 tempest-ImagesNegativeTestJSON-1015404931 tempest-ImagesNegativeTestJSON-1015404931-project-member] [instance: c4177e20-b1bd-4b54-a275-c93582359a07] Reconfigured VM instance instance-0000003a to attach disk [datastore1] c4177e20-b1bd-4b54-a275-c93582359a07/c4177e20-b1bd-4b54-a275-c93582359a07.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 865.482977] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-5a9a55ef-226a-4288-b8ca-9cc1a8ed9a10 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.494022] env[65726]: DEBUG oslo_vmware.api [None req-55d74d77-063b-418b-b818-59d05c490f34 tempest-ImagesNegativeTestJSON-1015404931 tempest-ImagesNegativeTestJSON-1015404931-project-member] Waiting for the task: (returnval){ [ 865.494022] env[65726]: value = "task-5116063" [ 865.494022] env[65726]: _type = "Task" [ 865.494022] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.504290] env[65726]: DEBUG oslo_vmware.api [None req-55d74d77-063b-418b-b818-59d05c490f34 tempest-ImagesNegativeTestJSON-1015404931 tempest-ImagesNegativeTestJSON-1015404931-project-member] Task: {'id': task-5116063, 'name': Rename_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.518991] env[65726]: DEBUG oslo_vmware.api [None req-b3ea747b-af8b-4592-b662-6643bc7f39d3 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] Task: {'id': task-5116061, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.835215] env[65726]: DEBUG oslo_vmware.api [None req-cb2d094a-412f-492d-9ef1-aba8757b7146 tempest-ServerMetadataNegativeTestJSON-936475568 tempest-ServerMetadataNegativeTestJSON-936475568-project-member] Task: {'id': task-5116062, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.111076} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.835631] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-cb2d094a-412f-492d-9ef1-aba8757b7146 tempest-ServerMetadataNegativeTestJSON-936475568 tempest-ServerMetadataNegativeTestJSON-936475568-project-member] [instance: d179030a-e5ba-45b1-ad11-4d2f71ed7bd4] Extended root virtual disk {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 865.836336] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17d2a600-59d4-41e4-8d3d-2b908dfb3a42 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.859221] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-cb2d094a-412f-492d-9ef1-aba8757b7146 tempest-ServerMetadataNegativeTestJSON-936475568 tempest-ServerMetadataNegativeTestJSON-936475568-project-member] [instance: d179030a-e5ba-45b1-ad11-4d2f71ed7bd4] Reconfiguring VM instance instance-0000003b to attach disk [datastore1] d179030a-e5ba-45b1-ad11-4d2f71ed7bd4/d179030a-e5ba-45b1-ad11-4d2f71ed7bd4.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 865.859524] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f0e007b7-03d0-4c69-80ce-7258e5a6524c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.880212] env[65726]: DEBUG oslo_vmware.api [None req-cb2d094a-412f-492d-9ef1-aba8757b7146 tempest-ServerMetadataNegativeTestJSON-936475568 tempest-ServerMetadataNegativeTestJSON-936475568-project-member] Waiting for the task: (returnval){ [ 865.880212] env[65726]: value = "task-5116064" [ 865.880212] env[65726]: _type = "Task" [ 865.880212] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.889375] env[65726]: DEBUG oslo_vmware.api [None req-cb2d094a-412f-492d-9ef1-aba8757b7146 tempest-ServerMetadataNegativeTestJSON-936475568 tempest-ServerMetadataNegativeTestJSON-936475568-project-member] Task: {'id': task-5116064, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.928842] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e1137661-56aa-48f2-a91c-68a6f9993ed2 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Lock "49bcab1f-7b64-4999-abff-37771c58a271" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 42.473s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 866.013427] env[65726]: DEBUG oslo_vmware.api [None req-55d74d77-063b-418b-b818-59d05c490f34 tempest-ImagesNegativeTestJSON-1015404931 tempest-ImagesNegativeTestJSON-1015404931-project-member] Task: {'id': task-5116063, 'name': Rename_Task, 'duration_secs': 0.48235} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.017071] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-55d74d77-063b-418b-b818-59d05c490f34 tempest-ImagesNegativeTestJSON-1015404931 tempest-ImagesNegativeTestJSON-1015404931-project-member] [instance: c4177e20-b1bd-4b54-a275-c93582359a07] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 866.017467] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b1402fae-0d20-4521-9eda-076774e1e678 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.024488] env[65726]: DEBUG oslo_vmware.api [None req-b3ea747b-af8b-4592-b662-6643bc7f39d3 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] Task: {'id': task-5116061, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.786096} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.024840] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-b3ea747b-af8b-4592-b662-6643bc7f39d3 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore2] 2cbee359-a1fc-45c0-bcc5-a2c24ded9c69/2cbee359-a1fc-45c0-bcc5-a2c24ded9c69.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 866.024937] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-b3ea747b-af8b-4592-b662-6643bc7f39d3 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] [instance: 2cbee359-a1fc-45c0-bcc5-a2c24ded9c69] Extending root virtual disk to 1048576 {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 866.026338] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-09439bf4-3cf3-43dd-918d-fcbd58a970f7 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.028829] env[65726]: DEBUG oslo_vmware.api [None req-55d74d77-063b-418b-b818-59d05c490f34 tempest-ImagesNegativeTestJSON-1015404931 tempest-ImagesNegativeTestJSON-1015404931-project-member] Waiting for the task: (returnval){ [ 866.028829] env[65726]: value = "task-5116065" [ 866.028829] env[65726]: _type = "Task" [ 866.028829] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.035767] env[65726]: DEBUG oslo_vmware.api [None req-b3ea747b-af8b-4592-b662-6643bc7f39d3 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] Waiting for the task: (returnval){ [ 866.035767] env[65726]: value = "task-5116066" [ 866.035767] env[65726]: _type = "Task" [ 866.035767] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.044289] env[65726]: DEBUG oslo_vmware.api [None req-55d74d77-063b-418b-b818-59d05c490f34 tempest-ImagesNegativeTestJSON-1015404931 tempest-ImagesNegativeTestJSON-1015404931-project-member] Task: {'id': task-5116065, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.050106] env[65726]: DEBUG oslo_vmware.api [None req-b3ea747b-af8b-4592-b662-6643bc7f39d3 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] Task: {'id': task-5116066, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.391188] env[65726]: DEBUG oslo_vmware.api [None req-cb2d094a-412f-492d-9ef1-aba8757b7146 tempest-ServerMetadataNegativeTestJSON-936475568 tempest-ServerMetadataNegativeTestJSON-936475568-project-member] Task: {'id': task-5116064, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.443733] env[65726]: DEBUG oslo_concurrency.lockutils [None req-7734f530-a2df-4d41-b25a-6f45fb216a4e tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Acquiring lock "44d73b2d-2b6a-4501-9944-432da3c9330d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 866.444042] env[65726]: DEBUG oslo_concurrency.lockutils [None req-7734f530-a2df-4d41-b25a-6f45fb216a4e tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Lock "44d73b2d-2b6a-4501-9944-432da3c9330d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 866.444268] env[65726]: DEBUG oslo_concurrency.lockutils [None req-7734f530-a2df-4d41-b25a-6f45fb216a4e tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Acquiring lock "44d73b2d-2b6a-4501-9944-432da3c9330d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 866.444459] env[65726]: DEBUG oslo_concurrency.lockutils [None req-7734f530-a2df-4d41-b25a-6f45fb216a4e tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Lock "44d73b2d-2b6a-4501-9944-432da3c9330d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 866.444692] env[65726]: DEBUG oslo_concurrency.lockutils [None req-7734f530-a2df-4d41-b25a-6f45fb216a4e tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Lock "44d73b2d-2b6a-4501-9944-432da3c9330d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 866.447155] env[65726]: INFO nova.compute.manager [None req-7734f530-a2df-4d41-b25a-6f45fb216a4e tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] [instance: 44d73b2d-2b6a-4501-9944-432da3c9330d] Terminating instance [ 866.544492] env[65726]: DEBUG oslo_vmware.api [None req-55d74d77-063b-418b-b818-59d05c490f34 tempest-ImagesNegativeTestJSON-1015404931 tempest-ImagesNegativeTestJSON-1015404931-project-member] Task: {'id': task-5116065, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.550863] env[65726]: DEBUG oslo_vmware.api [None req-b3ea747b-af8b-4592-b662-6643bc7f39d3 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] Task: {'id': task-5116066, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.168401} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.550863] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-b3ea747b-af8b-4592-b662-6643bc7f39d3 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] [instance: 2cbee359-a1fc-45c0-bcc5-a2c24ded9c69] Extended root virtual disk {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 866.554442] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9666ba6d-6924-44c2-b488-1f10db66662b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.584419] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-b3ea747b-af8b-4592-b662-6643bc7f39d3 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] [instance: 2cbee359-a1fc-45c0-bcc5-a2c24ded9c69] Reconfiguring VM instance instance-0000003c to attach disk [datastore2] 2cbee359-a1fc-45c0-bcc5-a2c24ded9c69/2cbee359-a1fc-45c0-bcc5-a2c24ded9c69.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 866.587535] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4f485b3e-b7e4-4723-8ae7-6d44065b6021 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.618855] env[65726]: DEBUG oslo_vmware.api [None req-b3ea747b-af8b-4592-b662-6643bc7f39d3 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] Waiting for the task: (returnval){ [ 866.618855] env[65726]: value = "task-5116067" [ 866.618855] env[65726]: _type = "Task" [ 866.618855] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.638302] env[65726]: DEBUG oslo_vmware.api [None req-b3ea747b-af8b-4592-b662-6643bc7f39d3 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] Task: {'id': task-5116067, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.896596] env[65726]: DEBUG oslo_vmware.api [None req-cb2d094a-412f-492d-9ef1-aba8757b7146 tempest-ServerMetadataNegativeTestJSON-936475568 tempest-ServerMetadataNegativeTestJSON-936475568-project-member] Task: {'id': task-5116064, 'name': ReconfigVM_Task, 'duration_secs': 0.739295} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.896998] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-cb2d094a-412f-492d-9ef1-aba8757b7146 tempest-ServerMetadataNegativeTestJSON-936475568 tempest-ServerMetadataNegativeTestJSON-936475568-project-member] [instance: d179030a-e5ba-45b1-ad11-4d2f71ed7bd4] Reconfigured VM instance instance-0000003b to attach disk [datastore1] d179030a-e5ba-45b1-ad11-4d2f71ed7bd4/d179030a-e5ba-45b1-ad11-4d2f71ed7bd4.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 866.897761] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-35a8a571-9fcf-4ffb-ba75-af261ec2a62b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.910887] env[65726]: DEBUG oslo_vmware.api [None req-cb2d094a-412f-492d-9ef1-aba8757b7146 tempest-ServerMetadataNegativeTestJSON-936475568 tempest-ServerMetadataNegativeTestJSON-936475568-project-member] Waiting for the task: (returnval){ [ 866.910887] env[65726]: value = "task-5116068" [ 866.910887] env[65726]: _type = "Task" [ 866.910887] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.922915] env[65726]: DEBUG oslo_vmware.api [None req-cb2d094a-412f-492d-9ef1-aba8757b7146 tempest-ServerMetadataNegativeTestJSON-936475568 tempest-ServerMetadataNegativeTestJSON-936475568-project-member] Task: {'id': task-5116068, 'name': Rename_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.954117] env[65726]: DEBUG oslo_concurrency.lockutils [None req-7734f530-a2df-4d41-b25a-6f45fb216a4e tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Acquiring lock "refresh_cache-44d73b2d-2b6a-4501-9944-432da3c9330d" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 866.954300] env[65726]: DEBUG oslo_concurrency.lockutils [None req-7734f530-a2df-4d41-b25a-6f45fb216a4e tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Acquired lock "refresh_cache-44d73b2d-2b6a-4501-9944-432da3c9330d" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 866.954567] env[65726]: DEBUG nova.network.neutron [None req-7734f530-a2df-4d41-b25a-6f45fb216a4e tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] [instance: 44d73b2d-2b6a-4501-9944-432da3c9330d] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 867.027544] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-383fab3b-83e5-4b3e-b25d-1f07674c552d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.043108] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-794f6765-1ba0-4c06-9a33-e261af85c9ba {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.047266] env[65726]: DEBUG oslo_vmware.api [None req-55d74d77-063b-418b-b818-59d05c490f34 tempest-ImagesNegativeTestJSON-1015404931 tempest-ImagesNegativeTestJSON-1015404931-project-member] Task: {'id': task-5116065, 'name': PowerOnVM_Task, 'duration_secs': 0.830354} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.047539] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-55d74d77-063b-418b-b818-59d05c490f34 tempest-ImagesNegativeTestJSON-1015404931 tempest-ImagesNegativeTestJSON-1015404931-project-member] [instance: c4177e20-b1bd-4b54-a275-c93582359a07] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 867.047539] env[65726]: INFO nova.compute.manager [None req-55d74d77-063b-418b-b818-59d05c490f34 tempest-ImagesNegativeTestJSON-1015404931 tempest-ImagesNegativeTestJSON-1015404931-project-member] [instance: c4177e20-b1bd-4b54-a275-c93582359a07] Took 9.81 seconds to spawn the instance on the hypervisor. [ 867.047942] env[65726]: DEBUG nova.compute.manager [None req-55d74d77-063b-418b-b818-59d05c490f34 tempest-ImagesNegativeTestJSON-1015404931 tempest-ImagesNegativeTestJSON-1015404931-project-member] [instance: c4177e20-b1bd-4b54-a275-c93582359a07] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 867.049169] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-696ca97d-7698-4efa-86f1-2d2721aeb7c1 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.080248] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15b8f5f0-ff38-485f-b120-bfe8980b6981 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.094534] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b3b7cbf-aa2b-4d4a-882e-7fd3391f4adc {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.112178] env[65726]: DEBUG nova.compute.provider_tree [None req-49b35bf1-9410-49fb-9a5b-632c41ecd3a6 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 867.113825] env[65726]: INFO nova.compute.manager [None req-8333c704-ca27-426e-acee-8cfd63213196 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: f2d9090c-988f-43f4-9c81-7aa718a3438a] Rebuilding instance [ 867.130334] env[65726]: DEBUG oslo_vmware.api [None req-b3ea747b-af8b-4592-b662-6643bc7f39d3 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] Task: {'id': task-5116067, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.182839] env[65726]: DEBUG nova.compute.manager [None req-8333c704-ca27-426e-acee-8cfd63213196 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: f2d9090c-988f-43f4-9c81-7aa718a3438a] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 867.183738] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65d6b940-50a1-4f62-a4db-77e3a000f2db {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.428317] env[65726]: DEBUG oslo_vmware.api [None req-cb2d094a-412f-492d-9ef1-aba8757b7146 tempest-ServerMetadataNegativeTestJSON-936475568 tempest-ServerMetadataNegativeTestJSON-936475568-project-member] Task: {'id': task-5116068, 'name': Rename_Task, 'duration_secs': 0.222874} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.428768] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-cb2d094a-412f-492d-9ef1-aba8757b7146 tempest-ServerMetadataNegativeTestJSON-936475568 tempest-ServerMetadataNegativeTestJSON-936475568-project-member] [instance: d179030a-e5ba-45b1-ad11-4d2f71ed7bd4] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 867.429135] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b8acf157-a295-42dc-8b8c-0c03f61a7896 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.437833] env[65726]: DEBUG oslo_vmware.api [None req-cb2d094a-412f-492d-9ef1-aba8757b7146 tempest-ServerMetadataNegativeTestJSON-936475568 tempest-ServerMetadataNegativeTestJSON-936475568-project-member] Waiting for the task: (returnval){ [ 867.437833] env[65726]: value = "task-5116069" [ 867.437833] env[65726]: _type = "Task" [ 867.437833] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.446891] env[65726]: DEBUG oslo_vmware.api [None req-cb2d094a-412f-492d-9ef1-aba8757b7146 tempest-ServerMetadataNegativeTestJSON-936475568 tempest-ServerMetadataNegativeTestJSON-936475568-project-member] Task: {'id': task-5116069, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.462103] env[65726]: WARNING openstack [None req-7734f530-a2df-4d41-b25a-6f45fb216a4e tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 867.462509] env[65726]: WARNING openstack [None req-7734f530-a2df-4d41-b25a-6f45fb216a4e tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 867.468414] env[65726]: DEBUG nova.network.neutron [None req-7734f530-a2df-4d41-b25a-6f45fb216a4e tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] [instance: 44d73b2d-2b6a-4501-9944-432da3c9330d] Instance cache missing network info. {{(pid=65726) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 867.584563] env[65726]: DEBUG nova.network.neutron [None req-7734f530-a2df-4d41-b25a-6f45fb216a4e tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] [instance: 44d73b2d-2b6a-4501-9944-432da3c9330d] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 867.597624] env[65726]: INFO nova.compute.manager [None req-55d74d77-063b-418b-b818-59d05c490f34 tempest-ImagesNegativeTestJSON-1015404931 tempest-ImagesNegativeTestJSON-1015404931-project-member] [instance: c4177e20-b1bd-4b54-a275-c93582359a07] Took 57.17 seconds to build instance. [ 867.622357] env[65726]: DEBUG nova.scheduler.client.report [None req-49b35bf1-9410-49fb-9a5b-632c41ecd3a6 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 867.636662] env[65726]: DEBUG oslo_vmware.api [None req-b3ea747b-af8b-4592-b662-6643bc7f39d3 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] Task: {'id': task-5116067, 'name': ReconfigVM_Task, 'duration_secs': 0.755467} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.638118] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-b3ea747b-af8b-4592-b662-6643bc7f39d3 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] [instance: 2cbee359-a1fc-45c0-bcc5-a2c24ded9c69] Reconfigured VM instance instance-0000003c to attach disk [datastore2] 2cbee359-a1fc-45c0-bcc5-a2c24ded9c69/2cbee359-a1fc-45c0-bcc5-a2c24ded9c69.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 867.638927] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ecc1d305-339c-4346-a496-1f8dd26484a7 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.649073] env[65726]: DEBUG oslo_vmware.api [None req-b3ea747b-af8b-4592-b662-6643bc7f39d3 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] Waiting for the task: (returnval){ [ 867.649073] env[65726]: value = "task-5116070" [ 867.649073] env[65726]: _type = "Task" [ 867.649073] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.658185] env[65726]: DEBUG oslo_vmware.api [None req-b3ea747b-af8b-4592-b662-6643bc7f39d3 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] Task: {'id': task-5116070, 'name': Rename_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.949542] env[65726]: DEBUG oslo_vmware.api [None req-cb2d094a-412f-492d-9ef1-aba8757b7146 tempest-ServerMetadataNegativeTestJSON-936475568 tempest-ServerMetadataNegativeTestJSON-936475568-project-member] Task: {'id': task-5116069, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.087812] env[65726]: DEBUG oslo_concurrency.lockutils [None req-7734f530-a2df-4d41-b25a-6f45fb216a4e tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Releasing lock "refresh_cache-44d73b2d-2b6a-4501-9944-432da3c9330d" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 868.088325] env[65726]: DEBUG nova.compute.manager [None req-7734f530-a2df-4d41-b25a-6f45fb216a4e tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] [instance: 44d73b2d-2b6a-4501-9944-432da3c9330d] Start destroying the instance on the hypervisor. {{(pid=65726) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 868.088946] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-7734f530-a2df-4d41-b25a-6f45fb216a4e tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] [instance: 44d73b2d-2b6a-4501-9944-432da3c9330d] Destroying instance {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 868.089531] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0bb3f606-6e91-4c0a-9cd3-1ca726336ace {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.098851] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-7734f530-a2df-4d41-b25a-6f45fb216a4e tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] [instance: 44d73b2d-2b6a-4501-9944-432da3c9330d] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 868.099184] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-635a6d9c-fe7b-461f-9601-96c5ec64a9d6 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.101194] env[65726]: DEBUG oslo_concurrency.lockutils [None req-55d74d77-063b-418b-b818-59d05c490f34 tempest-ImagesNegativeTestJSON-1015404931 tempest-ImagesNegativeTestJSON-1015404931-project-member] Lock "c4177e20-b1bd-4b54-a275-c93582359a07" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 91.991s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 868.109062] env[65726]: DEBUG oslo_vmware.api [None req-7734f530-a2df-4d41-b25a-6f45fb216a4e tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Waiting for the task: (returnval){ [ 868.109062] env[65726]: value = "task-5116071" [ 868.109062] env[65726]: _type = "Task" [ 868.109062] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.120316] env[65726]: DEBUG oslo_vmware.api [None req-7734f530-a2df-4d41-b25a-6f45fb216a4e tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Task: {'id': task-5116071, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.133393] env[65726]: DEBUG oslo_concurrency.lockutils [None req-49b35bf1-9410-49fb-9a5b-632c41ecd3a6 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.740s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 868.133393] env[65726]: DEBUG nova.compute.manager [None req-49b35bf1-9410-49fb-9a5b-632c41ecd3a6 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] [instance: c4cc4f85-cb35-4edc-a58b-adfee0ce1265] Start building networks asynchronously for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 868.134495] env[65726]: DEBUG oslo_concurrency.lockutils [None req-73b0a627-5d81-4bf2-833a-caff3534c1e3 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 38.419s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 868.136713] env[65726]: INFO nova.compute.claims [None req-73b0a627-5d81-4bf2-833a-caff3534c1e3 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: cf950ec3-9914-4eb9-99db-048b3969bb21] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 868.163189] env[65726]: DEBUG oslo_vmware.api [None req-b3ea747b-af8b-4592-b662-6643bc7f39d3 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] Task: {'id': task-5116070, 'name': Rename_Task, 'duration_secs': 0.215029} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.163921] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-b3ea747b-af8b-4592-b662-6643bc7f39d3 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] [instance: 2cbee359-a1fc-45c0-bcc5-a2c24ded9c69] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 868.164133] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2dfd7cfa-f0dd-49f8-a5bb-899597395ed6 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.175187] env[65726]: DEBUG oslo_vmware.api [None req-b3ea747b-af8b-4592-b662-6643bc7f39d3 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] Waiting for the task: (returnval){ [ 868.175187] env[65726]: value = "task-5116072" [ 868.175187] env[65726]: _type = "Task" [ 868.175187] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.188807] env[65726]: DEBUG oslo_vmware.api [None req-b3ea747b-af8b-4592-b662-6643bc7f39d3 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] Task: {'id': task-5116072, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.202998] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-8333c704-ca27-426e-acee-8cfd63213196 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: f2d9090c-988f-43f4-9c81-7aa718a3438a] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 868.204237] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3dd55af5-aa01-4ef9-a9d3-b2004834fee9 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.214312] env[65726]: DEBUG oslo_vmware.api [None req-8333c704-ca27-426e-acee-8cfd63213196 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Waiting for the task: (returnval){ [ 868.214312] env[65726]: value = "task-5116073" [ 868.214312] env[65726]: _type = "Task" [ 868.214312] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.225611] env[65726]: DEBUG oslo_vmware.api [None req-8333c704-ca27-426e-acee-8cfd63213196 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Task: {'id': task-5116073, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.451718] env[65726]: DEBUG oslo_vmware.api [None req-cb2d094a-412f-492d-9ef1-aba8757b7146 tempest-ServerMetadataNegativeTestJSON-936475568 tempest-ServerMetadataNegativeTestJSON-936475568-project-member] Task: {'id': task-5116069, 'name': PowerOnVM_Task, 'duration_secs': 0.799465} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.452076] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-cb2d094a-412f-492d-9ef1-aba8757b7146 tempest-ServerMetadataNegativeTestJSON-936475568 tempest-ServerMetadataNegativeTestJSON-936475568-project-member] [instance: d179030a-e5ba-45b1-ad11-4d2f71ed7bd4] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 868.452278] env[65726]: INFO nova.compute.manager [None req-cb2d094a-412f-492d-9ef1-aba8757b7146 tempest-ServerMetadataNegativeTestJSON-936475568 tempest-ServerMetadataNegativeTestJSON-936475568-project-member] [instance: d179030a-e5ba-45b1-ad11-4d2f71ed7bd4] Took 8.55 seconds to spawn the instance on the hypervisor. [ 868.452460] env[65726]: DEBUG nova.compute.manager [None req-cb2d094a-412f-492d-9ef1-aba8757b7146 tempest-ServerMetadataNegativeTestJSON-936475568 tempest-ServerMetadataNegativeTestJSON-936475568-project-member] [instance: d179030a-e5ba-45b1-ad11-4d2f71ed7bd4] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 868.454057] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78c5afaa-a8db-45e3-86c6-f109a4d8775f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.605354] env[65726]: DEBUG nova.compute.manager [None req-e5e88d5e-aa68-4ff5-9280-b2cfe1d14b27 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: 5a252ef1-93c9-4bff-842b-b64df2bc5d75] Starting instance... {{(pid=65726) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 868.613023] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a4b79740-ccaa-4794-bd78-55e7c8646033 tempest-ImagesNegativeTestJSON-1015404931 tempest-ImagesNegativeTestJSON-1015404931-project-member] Acquiring lock "c4177e20-b1bd-4b54-a275-c93582359a07" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 868.613023] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a4b79740-ccaa-4794-bd78-55e7c8646033 tempest-ImagesNegativeTestJSON-1015404931 tempest-ImagesNegativeTestJSON-1015404931-project-member] Lock "c4177e20-b1bd-4b54-a275-c93582359a07" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 868.613023] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a4b79740-ccaa-4794-bd78-55e7c8646033 tempest-ImagesNegativeTestJSON-1015404931 tempest-ImagesNegativeTestJSON-1015404931-project-member] Acquiring lock "c4177e20-b1bd-4b54-a275-c93582359a07-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 868.613023] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a4b79740-ccaa-4794-bd78-55e7c8646033 tempest-ImagesNegativeTestJSON-1015404931 tempest-ImagesNegativeTestJSON-1015404931-project-member] Lock "c4177e20-b1bd-4b54-a275-c93582359a07-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 868.613023] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a4b79740-ccaa-4794-bd78-55e7c8646033 tempest-ImagesNegativeTestJSON-1015404931 tempest-ImagesNegativeTestJSON-1015404931-project-member] Lock "c4177e20-b1bd-4b54-a275-c93582359a07-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 868.617908] env[65726]: INFO nova.compute.manager [None req-a4b79740-ccaa-4794-bd78-55e7c8646033 tempest-ImagesNegativeTestJSON-1015404931 tempest-ImagesNegativeTestJSON-1015404931-project-member] [instance: c4177e20-b1bd-4b54-a275-c93582359a07] Terminating instance [ 868.626437] env[65726]: DEBUG oslo_vmware.api [None req-7734f530-a2df-4d41-b25a-6f45fb216a4e tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Task: {'id': task-5116071, 'name': PowerOffVM_Task, 'duration_secs': 0.163817} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.626931] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-7734f530-a2df-4d41-b25a-6f45fb216a4e tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] [instance: 44d73b2d-2b6a-4501-9944-432da3c9330d] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 868.627107] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-7734f530-a2df-4d41-b25a-6f45fb216a4e tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] [instance: 44d73b2d-2b6a-4501-9944-432da3c9330d] Unregistering the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 868.627368] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4fb522cc-b081-4de1-8db3-b05b880c9150 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.643175] env[65726]: DEBUG nova.compute.utils [None req-49b35bf1-9410-49fb-9a5b-632c41ecd3a6 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Using /dev/sd instead of None {{(pid=65726) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 868.649465] env[65726]: DEBUG nova.compute.manager [None req-49b35bf1-9410-49fb-9a5b-632c41ecd3a6 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] [instance: c4cc4f85-cb35-4edc-a58b-adfee0ce1265] Allocating IP information in the background. {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 868.649465] env[65726]: DEBUG nova.network.neutron [None req-49b35bf1-9410-49fb-9a5b-632c41ecd3a6 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] [instance: c4cc4f85-cb35-4edc-a58b-adfee0ce1265] allocate_for_instance() {{(pid=65726) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 868.651405] env[65726]: WARNING neutronclient.v2_0.client [None req-49b35bf1-9410-49fb-9a5b-632c41ecd3a6 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 868.651405] env[65726]: WARNING neutronclient.v2_0.client [None req-49b35bf1-9410-49fb-9a5b-632c41ecd3a6 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 868.651405] env[65726]: WARNING openstack [None req-49b35bf1-9410-49fb-9a5b-632c41ecd3a6 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 868.651405] env[65726]: WARNING openstack [None req-49b35bf1-9410-49fb-9a5b-632c41ecd3a6 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 868.661718] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-7734f530-a2df-4d41-b25a-6f45fb216a4e tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] [instance: 44d73b2d-2b6a-4501-9944-432da3c9330d] Unregistered the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 868.662078] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-7734f530-a2df-4d41-b25a-6f45fb216a4e tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] [instance: 44d73b2d-2b6a-4501-9944-432da3c9330d] Deleting contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 868.665100] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-7734f530-a2df-4d41-b25a-6f45fb216a4e tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Deleting the datastore file [datastore1] 44d73b2d-2b6a-4501-9944-432da3c9330d {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 868.665100] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-cc79a98b-792c-4514-b42f-395943fdcea6 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.672875] env[65726]: DEBUG oslo_vmware.api [None req-7734f530-a2df-4d41-b25a-6f45fb216a4e tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Waiting for the task: (returnval){ [ 868.672875] env[65726]: value = "task-5116075" [ 868.672875] env[65726]: _type = "Task" [ 868.672875] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.688904] env[65726]: DEBUG oslo_vmware.api [None req-b3ea747b-af8b-4592-b662-6643bc7f39d3 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] Task: {'id': task-5116072, 'name': PowerOnVM_Task} progress is 88%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.689301] env[65726]: DEBUG oslo_vmware.api [None req-7734f530-a2df-4d41-b25a-6f45fb216a4e tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Task: {'id': task-5116075, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.702916] env[65726]: DEBUG nova.policy [None req-49b35bf1-9410-49fb-9a5b-632c41ecd3a6 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ce619055926b4fad8ca6869354f864eb', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2772e93581124978b55c8054d2116232', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65726) authorize /opt/stack/nova/nova/policy.py:192}} [ 868.724409] env[65726]: DEBUG oslo_vmware.api [None req-8333c704-ca27-426e-acee-8cfd63213196 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Task: {'id': task-5116073, 'name': PowerOffVM_Task, 'duration_secs': 0.256913} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.724775] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-8333c704-ca27-426e-acee-8cfd63213196 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: f2d9090c-988f-43f4-9c81-7aa718a3438a] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 868.725103] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-8333c704-ca27-426e-acee-8cfd63213196 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: f2d9090c-988f-43f4-9c81-7aa718a3438a] Destroying instance {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 868.725817] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f62d617d-7397-46aa-b736-a921491e9ac1 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.734034] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-8333c704-ca27-426e-acee-8cfd63213196 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: f2d9090c-988f-43f4-9c81-7aa718a3438a] Unregistering the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 868.734447] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8ee868fa-81d3-4324-82a4-04a428f75fdc {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.813892] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-8333c704-ca27-426e-acee-8cfd63213196 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: f2d9090c-988f-43f4-9c81-7aa718a3438a] Unregistered the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 868.814277] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-8333c704-ca27-426e-acee-8cfd63213196 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: f2d9090c-988f-43f4-9c81-7aa718a3438a] Deleting contents of the VM from datastore datastore2 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 868.814563] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-8333c704-ca27-426e-acee-8cfd63213196 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Deleting the datastore file [datastore2] f2d9090c-988f-43f4-9c81-7aa718a3438a {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 868.814964] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ae04b9e5-a0e5-48bd-b8ff-35b87842d8d2 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.823122] env[65726]: DEBUG oslo_vmware.api [None req-8333c704-ca27-426e-acee-8cfd63213196 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Waiting for the task: (returnval){ [ 868.823122] env[65726]: value = "task-5116077" [ 868.823122] env[65726]: _type = "Task" [ 868.823122] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.834232] env[65726]: DEBUG oslo_vmware.api [None req-8333c704-ca27-426e-acee-8cfd63213196 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Task: {'id': task-5116077, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.972234] env[65726]: DEBUG nova.network.neutron [None req-49b35bf1-9410-49fb-9a5b-632c41ecd3a6 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] [instance: c4cc4f85-cb35-4edc-a58b-adfee0ce1265] Successfully created port: 01aa3d90-3b27-4f98-a3ef-eb316f36f0cc {{(pid=65726) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 868.978320] env[65726]: INFO nova.compute.manager [None req-cb2d094a-412f-492d-9ef1-aba8757b7146 tempest-ServerMetadataNegativeTestJSON-936475568 tempest-ServerMetadataNegativeTestJSON-936475568-project-member] [instance: d179030a-e5ba-45b1-ad11-4d2f71ed7bd4] Took 50.38 seconds to build instance. [ 869.128046] env[65726]: DEBUG nova.compute.manager [None req-a4b79740-ccaa-4794-bd78-55e7c8646033 tempest-ImagesNegativeTestJSON-1015404931 tempest-ImagesNegativeTestJSON-1015404931-project-member] [instance: c4177e20-b1bd-4b54-a275-c93582359a07] Start destroying the instance on the hypervisor. {{(pid=65726) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 869.128435] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-a4b79740-ccaa-4794-bd78-55e7c8646033 tempest-ImagesNegativeTestJSON-1015404931 tempest-ImagesNegativeTestJSON-1015404931-project-member] [instance: c4177e20-b1bd-4b54-a275-c93582359a07] Destroying instance {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 869.129649] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2dd0774e-4a0d-4667-9595-f83e7c7aafc9 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.134356] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e5e88d5e-aa68-4ff5-9280-b2cfe1d14b27 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 869.144935] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-a4b79740-ccaa-4794-bd78-55e7c8646033 tempest-ImagesNegativeTestJSON-1015404931 tempest-ImagesNegativeTestJSON-1015404931-project-member] [instance: c4177e20-b1bd-4b54-a275-c93582359a07] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 869.145881] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8207dcfa-9587-4711-ad1c-57e232583adf {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.150780] env[65726]: DEBUG nova.compute.manager [None req-49b35bf1-9410-49fb-9a5b-632c41ecd3a6 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] [instance: c4cc4f85-cb35-4edc-a58b-adfee0ce1265] Start building block device mappings for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 869.157790] env[65726]: DEBUG oslo_vmware.api [None req-a4b79740-ccaa-4794-bd78-55e7c8646033 tempest-ImagesNegativeTestJSON-1015404931 tempest-ImagesNegativeTestJSON-1015404931-project-member] Waiting for the task: (returnval){ [ 869.157790] env[65726]: value = "task-5116078" [ 869.157790] env[65726]: _type = "Task" [ 869.157790] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.177624] env[65726]: DEBUG oslo_vmware.api [None req-a4b79740-ccaa-4794-bd78-55e7c8646033 tempest-ImagesNegativeTestJSON-1015404931 tempest-ImagesNegativeTestJSON-1015404931-project-member] Task: {'id': task-5116078, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.191427] env[65726]: DEBUG oslo_vmware.api [None req-b3ea747b-af8b-4592-b662-6643bc7f39d3 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] Task: {'id': task-5116072, 'name': PowerOnVM_Task, 'duration_secs': 0.716125} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.195519] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-b3ea747b-af8b-4592-b662-6643bc7f39d3 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] [instance: 2cbee359-a1fc-45c0-bcc5-a2c24ded9c69] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 869.195757] env[65726]: INFO nova.compute.manager [None req-b3ea747b-af8b-4592-b662-6643bc7f39d3 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] [instance: 2cbee359-a1fc-45c0-bcc5-a2c24ded9c69] Took 7.49 seconds to spawn the instance on the hypervisor. [ 869.195951] env[65726]: DEBUG nova.compute.manager [None req-b3ea747b-af8b-4592-b662-6643bc7f39d3 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] [instance: 2cbee359-a1fc-45c0-bcc5-a2c24ded9c69] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 869.196279] env[65726]: DEBUG oslo_vmware.api [None req-7734f530-a2df-4d41-b25a-6f45fb216a4e tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Task: {'id': task-5116075, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.222317} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.197166] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1b781e5-54bf-4197-b42d-46f2796a376b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.203025] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-7734f530-a2df-4d41-b25a-6f45fb216a4e tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Deleted the datastore file {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 869.203025] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-7734f530-a2df-4d41-b25a-6f45fb216a4e tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] [instance: 44d73b2d-2b6a-4501-9944-432da3c9330d] Deleted contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 869.203025] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-7734f530-a2df-4d41-b25a-6f45fb216a4e tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] [instance: 44d73b2d-2b6a-4501-9944-432da3c9330d] Instance destroyed {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 869.203025] env[65726]: INFO nova.compute.manager [None req-7734f530-a2df-4d41-b25a-6f45fb216a4e tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] [instance: 44d73b2d-2b6a-4501-9944-432da3c9330d] Took 1.11 seconds to destroy the instance on the hypervisor. [ 869.203025] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-7734f530-a2df-4d41-b25a-6f45fb216a4e tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 869.203025] env[65726]: DEBUG nova.compute.manager [-] [instance: 44d73b2d-2b6a-4501-9944-432da3c9330d] Deallocating network for instance {{(pid=65726) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 869.203025] env[65726]: DEBUG nova.network.neutron [-] [instance: 44d73b2d-2b6a-4501-9944-432da3c9330d] deallocate_for_instance() {{(pid=65726) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 869.203025] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 869.203025] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 869.236380] env[65726]: DEBUG nova.network.neutron [-] [instance: 44d73b2d-2b6a-4501-9944-432da3c9330d] Instance cache missing network info. {{(pid=65726) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 869.236927] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 869.237192] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 869.334358] env[65726]: DEBUG oslo_vmware.api [None req-8333c704-ca27-426e-acee-8cfd63213196 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Task: {'id': task-5116077, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.389507} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.334761] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-8333c704-ca27-426e-acee-8cfd63213196 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Deleted the datastore file {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 869.334953] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-8333c704-ca27-426e-acee-8cfd63213196 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: f2d9090c-988f-43f4-9c81-7aa718a3438a] Deleted contents of the VM from datastore datastore2 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 869.335139] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-8333c704-ca27-426e-acee-8cfd63213196 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: f2d9090c-988f-43f4-9c81-7aa718a3438a] Instance destroyed {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 869.479889] env[65726]: DEBUG oslo_concurrency.lockutils [None req-cb2d094a-412f-492d-9ef1-aba8757b7146 tempest-ServerMetadataNegativeTestJSON-936475568 tempest-ServerMetadataNegativeTestJSON-936475568-project-member] Lock "d179030a-e5ba-45b1-ad11-4d2f71ed7bd4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 84.491s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 869.679869] env[65726]: DEBUG oslo_vmware.api [None req-a4b79740-ccaa-4794-bd78-55e7c8646033 tempest-ImagesNegativeTestJSON-1015404931 tempest-ImagesNegativeTestJSON-1015404931-project-member] Task: {'id': task-5116078, 'name': PowerOffVM_Task, 'duration_secs': 0.342325} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.682736] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-a4b79740-ccaa-4794-bd78-55e7c8646033 tempest-ImagesNegativeTestJSON-1015404931 tempest-ImagesNegativeTestJSON-1015404931-project-member] [instance: c4177e20-b1bd-4b54-a275-c93582359a07] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 869.682905] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-a4b79740-ccaa-4794-bd78-55e7c8646033 tempest-ImagesNegativeTestJSON-1015404931 tempest-ImagesNegativeTestJSON-1015404931-project-member] [instance: c4177e20-b1bd-4b54-a275-c93582359a07] Unregistering the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 869.684198] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5f11c2b7-37e8-4d84-86b6-b00a67c0299e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.738158] env[65726]: INFO nova.compute.manager [None req-b3ea747b-af8b-4592-b662-6643bc7f39d3 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] [instance: 2cbee359-a1fc-45c0-bcc5-a2c24ded9c69] Took 46.87 seconds to build instance. [ 869.746515] env[65726]: DEBUG nova.network.neutron [-] [instance: 44d73b2d-2b6a-4501-9944-432da3c9330d] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 869.771871] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-a4b79740-ccaa-4794-bd78-55e7c8646033 tempest-ImagesNegativeTestJSON-1015404931 tempest-ImagesNegativeTestJSON-1015404931-project-member] [instance: c4177e20-b1bd-4b54-a275-c93582359a07] Unregistered the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 869.771871] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-a4b79740-ccaa-4794-bd78-55e7c8646033 tempest-ImagesNegativeTestJSON-1015404931 tempest-ImagesNegativeTestJSON-1015404931-project-member] [instance: c4177e20-b1bd-4b54-a275-c93582359a07] Deleting contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 869.771871] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-a4b79740-ccaa-4794-bd78-55e7c8646033 tempest-ImagesNegativeTestJSON-1015404931 tempest-ImagesNegativeTestJSON-1015404931-project-member] Deleting the datastore file [datastore1] c4177e20-b1bd-4b54-a275-c93582359a07 {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 869.772154] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0140aa48-25a5-4219-9240-2a19bdbecc09 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.780439] env[65726]: DEBUG oslo_vmware.api [None req-a4b79740-ccaa-4794-bd78-55e7c8646033 tempest-ImagesNegativeTestJSON-1015404931 tempest-ImagesNegativeTestJSON-1015404931-project-member] Waiting for the task: (returnval){ [ 869.780439] env[65726]: value = "task-5116080" [ 869.780439] env[65726]: _type = "Task" [ 869.780439] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.791921] env[65726]: DEBUG oslo_vmware.api [None req-a4b79740-ccaa-4794-bd78-55e7c8646033 tempest-ImagesNegativeTestJSON-1015404931 tempest-ImagesNegativeTestJSON-1015404931-project-member] Task: {'id': task-5116080, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.813951] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81ad068e-c6de-412b-b140-b8f8bda92085 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.826981] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c32078e8-f128-41cb-9e6a-c268f2c2b9bf {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.863129] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db923f6a-ab0f-4729-81ea-c5d930798b73 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.871420] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c44e4eb7-9834-4085-ae81-18e7b833dacb {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.885704] env[65726]: DEBUG nova.compute.provider_tree [None req-73b0a627-5d81-4bf2-833a-caff3534c1e3 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 869.983113] env[65726]: DEBUG nova.compute.manager [None req-89fee262-eea7-4ed4-83cb-eb9998cc3059 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 6f91b053-772a-4497-b29d-349b960c55eb] Starting instance... {{(pid=65726) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 870.175273] env[65726]: DEBUG nova.compute.manager [None req-49b35bf1-9410-49fb-9a5b-632c41ecd3a6 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] [instance: c4cc4f85-cb35-4edc-a58b-adfee0ce1265] Start spawning the instance on the hypervisor. {{(pid=65726) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 870.214347] env[65726]: DEBUG nova.virt.hardware [None req-49b35bf1-9410-49fb-9a5b-632c41ecd3a6 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T19:28:34Z,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1539f06fa3534e90acbc3a60c4b8bd3f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T19:28:35Z,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 870.214654] env[65726]: DEBUG nova.virt.hardware [None req-49b35bf1-9410-49fb-9a5b-632c41ecd3a6 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 870.214837] env[65726]: DEBUG nova.virt.hardware [None req-49b35bf1-9410-49fb-9a5b-632c41ecd3a6 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 870.215134] env[65726]: DEBUG nova.virt.hardware [None req-49b35bf1-9410-49fb-9a5b-632c41ecd3a6 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 870.215346] env[65726]: DEBUG nova.virt.hardware [None req-49b35bf1-9410-49fb-9a5b-632c41ecd3a6 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 870.215509] env[65726]: DEBUG nova.virt.hardware [None req-49b35bf1-9410-49fb-9a5b-632c41ecd3a6 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 870.215720] env[65726]: DEBUG nova.virt.hardware [None req-49b35bf1-9410-49fb-9a5b-632c41ecd3a6 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 870.215875] env[65726]: DEBUG nova.virt.hardware [None req-49b35bf1-9410-49fb-9a5b-632c41ecd3a6 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 870.216055] env[65726]: DEBUG nova.virt.hardware [None req-49b35bf1-9410-49fb-9a5b-632c41ecd3a6 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 870.216219] env[65726]: DEBUG nova.virt.hardware [None req-49b35bf1-9410-49fb-9a5b-632c41ecd3a6 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 870.216383] env[65726]: DEBUG nova.virt.hardware [None req-49b35bf1-9410-49fb-9a5b-632c41ecd3a6 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 870.217321] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11b8c6c1-e6ed-4689-ac25-30f36e5988ac {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.227419] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b78398a4-34dc-4698-9386-bbd1bf963dc9 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.244815] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b3ea747b-af8b-4592-b662-6643bc7f39d3 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] Lock "2cbee359-a1fc-45c0-bcc5-a2c24ded9c69" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 63.398s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 870.250465] env[65726]: INFO nova.compute.manager [-] [instance: 44d73b2d-2b6a-4501-9944-432da3c9330d] Took 1.05 seconds to deallocate network for instance. [ 870.294327] env[65726]: DEBUG oslo_vmware.api [None req-a4b79740-ccaa-4794-bd78-55e7c8646033 tempest-ImagesNegativeTestJSON-1015404931 tempest-ImagesNegativeTestJSON-1015404931-project-member] Task: {'id': task-5116080, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.452185} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.295122] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-a4b79740-ccaa-4794-bd78-55e7c8646033 tempest-ImagesNegativeTestJSON-1015404931 tempest-ImagesNegativeTestJSON-1015404931-project-member] Deleted the datastore file {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 870.295560] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-a4b79740-ccaa-4794-bd78-55e7c8646033 tempest-ImagesNegativeTestJSON-1015404931 tempest-ImagesNegativeTestJSON-1015404931-project-member] [instance: c4177e20-b1bd-4b54-a275-c93582359a07] Deleted contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 870.295745] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-a4b79740-ccaa-4794-bd78-55e7c8646033 tempest-ImagesNegativeTestJSON-1015404931 tempest-ImagesNegativeTestJSON-1015404931-project-member] [instance: c4177e20-b1bd-4b54-a275-c93582359a07] Instance destroyed {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 870.295926] env[65726]: INFO nova.compute.manager [None req-a4b79740-ccaa-4794-bd78-55e7c8646033 tempest-ImagesNegativeTestJSON-1015404931 tempest-ImagesNegativeTestJSON-1015404931-project-member] [instance: c4177e20-b1bd-4b54-a275-c93582359a07] Took 1.17 seconds to destroy the instance on the hypervisor. [ 870.296207] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-a4b79740-ccaa-4794-bd78-55e7c8646033 tempest-ImagesNegativeTestJSON-1015404931 tempest-ImagesNegativeTestJSON-1015404931-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 870.297039] env[65726]: DEBUG nova.compute.manager [-] [instance: c4177e20-b1bd-4b54-a275-c93582359a07] Deallocating network for instance {{(pid=65726) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 870.297142] env[65726]: DEBUG nova.network.neutron [-] [instance: c4177e20-b1bd-4b54-a275-c93582359a07] deallocate_for_instance() {{(pid=65726) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 870.297638] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 870.297894] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 870.388844] env[65726]: DEBUG nova.virt.hardware [None req-8333c704-ca27-426e-acee-8cfd63213196 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T19:28:34Z,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1539f06fa3534e90acbc3a60c4b8bd3f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T19:28:35Z,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 870.389125] env[65726]: DEBUG nova.virt.hardware [None req-8333c704-ca27-426e-acee-8cfd63213196 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 870.389279] env[65726]: DEBUG nova.virt.hardware [None req-8333c704-ca27-426e-acee-8cfd63213196 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 870.389451] env[65726]: DEBUG nova.virt.hardware [None req-8333c704-ca27-426e-acee-8cfd63213196 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 870.389586] env[65726]: DEBUG nova.virt.hardware [None req-8333c704-ca27-426e-acee-8cfd63213196 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 870.389726] env[65726]: DEBUG nova.virt.hardware [None req-8333c704-ca27-426e-acee-8cfd63213196 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 870.389929] env[65726]: DEBUG nova.virt.hardware [None req-8333c704-ca27-426e-acee-8cfd63213196 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 870.390155] env[65726]: DEBUG nova.virt.hardware [None req-8333c704-ca27-426e-acee-8cfd63213196 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 870.390392] env[65726]: DEBUG nova.virt.hardware [None req-8333c704-ca27-426e-acee-8cfd63213196 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 870.390575] env[65726]: DEBUG nova.virt.hardware [None req-8333c704-ca27-426e-acee-8cfd63213196 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 870.390877] env[65726]: DEBUG nova.virt.hardware [None req-8333c704-ca27-426e-acee-8cfd63213196 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 870.391891] env[65726]: DEBUG nova.scheduler.client.report [None req-73b0a627-5d81-4bf2-833a-caff3534c1e3 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 870.395876] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a8afda9-41c5-45ba-b064-78097cd55747 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.405185] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f47db8cf-60b9-4247-a991-05bd7609f122 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.420585] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-8333c704-ca27-426e-acee-8cfd63213196 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: f2d9090c-988f-43f4-9c81-7aa718a3438a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:10:2a:24', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd0bdd11b-58af-4cc0-9d38-8322e1bb4e74', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7117cd21-ee2a-497d-b789-65e43b068258', 'vif_model': 'vmxnet3'}] {{(pid=65726) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 870.428312] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-8333c704-ca27-426e-acee-8cfd63213196 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 870.428995] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f2d9090c-988f-43f4-9c81-7aa718a3438a] Creating VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 870.429261] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d87a7f37-a11d-4665-979d-5dd90f8a6ac5 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.451761] env[65726]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 870.451761] env[65726]: value = "task-5116081" [ 870.451761] env[65726]: _type = "Task" [ 870.451761] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.462799] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116081, 'name': CreateVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.507458] env[65726]: DEBUG oslo_concurrency.lockutils [None req-89fee262-eea7-4ed4-83cb-eb9998cc3059 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 870.605357] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 870.606042] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 870.672820] env[65726]: DEBUG nova.network.neutron [None req-49b35bf1-9410-49fb-9a5b-632c41ecd3a6 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] [instance: c4cc4f85-cb35-4edc-a58b-adfee0ce1265] Successfully updated port: 01aa3d90-3b27-4f98-a3ef-eb316f36f0cc {{(pid=65726) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 870.750246] env[65726]: DEBUG nova.compute.manager [None req-17b5a1c7-4bc8-41d1-988f-c933bb6421ad tempest-ServersTestManualDisk-1753452859 tempest-ServersTestManualDisk-1753452859-project-member] [instance: 3088ee5e-0d27-4058-a94b-2e04c3b52add] Starting instance... {{(pid=65726) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 870.759710] env[65726]: DEBUG oslo_concurrency.lockutils [None req-7734f530-a2df-4d41-b25a-6f45fb216a4e tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 870.899870] env[65726]: DEBUG oslo_concurrency.lockutils [None req-73b0a627-5d81-4bf2-833a-caff3534c1e3 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.765s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 870.900432] env[65726]: DEBUG nova.compute.manager [None req-73b0a627-5d81-4bf2-833a-caff3534c1e3 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: cf950ec3-9914-4eb9-99db-048b3969bb21] Start building networks asynchronously for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 870.906618] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b20cc073-fafc-44fa-bcf9-9893dc98ebec tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 38.528s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 870.910034] env[65726]: INFO nova.compute.claims [None req-b20cc073-fafc-44fa-bcf9-9893dc98ebec tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 870.964645] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116081, 'name': CreateVM_Task, 'duration_secs': 0.408666} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.965066] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f2d9090c-988f-43f4-9c81-7aa718a3438a] Created VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 870.967121] env[65726]: WARNING openstack [None req-8333c704-ca27-426e-acee-8cfd63213196 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 870.967653] env[65726]: WARNING openstack [None req-8333c704-ca27-426e-acee-8cfd63213196 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 870.975375] env[65726]: DEBUG oslo_concurrency.lockutils [None req-8333c704-ca27-426e-acee-8cfd63213196 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 870.975889] env[65726]: DEBUG oslo_concurrency.lockutils [None req-8333c704-ca27-426e-acee-8cfd63213196 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 870.975889] env[65726]: DEBUG oslo_concurrency.lockutils [None req-8333c704-ca27-426e-acee-8cfd63213196 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 870.976958] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-06c8a134-dc5e-4fc2-b31a-f2ff942a7ae1 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.984342] env[65726]: DEBUG oslo_vmware.api [None req-8333c704-ca27-426e-acee-8cfd63213196 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Waiting for the task: (returnval){ [ 870.984342] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5202c109-7927-88f6-2a1d-1258cf9b3eba" [ 870.984342] env[65726]: _type = "Task" [ 870.984342] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.995394] env[65726]: DEBUG oslo_vmware.api [None req-8333c704-ca27-426e-acee-8cfd63213196 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5202c109-7927-88f6-2a1d-1258cf9b3eba, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.177402] env[65726]: DEBUG oslo_concurrency.lockutils [None req-49b35bf1-9410-49fb-9a5b-632c41ecd3a6 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Acquiring lock "refresh_cache-c4cc4f85-cb35-4edc-a58b-adfee0ce1265" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 871.177621] env[65726]: DEBUG oslo_concurrency.lockutils [None req-49b35bf1-9410-49fb-9a5b-632c41ecd3a6 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Acquired lock "refresh_cache-c4cc4f85-cb35-4edc-a58b-adfee0ce1265" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 871.177924] env[65726]: DEBUG nova.network.neutron [None req-49b35bf1-9410-49fb-9a5b-632c41ecd3a6 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] [instance: c4cc4f85-cb35-4edc-a58b-adfee0ce1265] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 871.280227] env[65726]: DEBUG oslo_concurrency.lockutils [None req-17b5a1c7-4bc8-41d1-988f-c933bb6421ad tempest-ServersTestManualDisk-1753452859 tempest-ServersTestManualDisk-1753452859-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 871.384233] env[65726]: DEBUG nova.network.neutron [-] [instance: c4177e20-b1bd-4b54-a275-c93582359a07] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 871.408248] env[65726]: DEBUG nova.compute.utils [None req-73b0a627-5d81-4bf2-833a-caff3534c1e3 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Using /dev/sd instead of None {{(pid=65726) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 871.410026] env[65726]: DEBUG nova.compute.manager [None req-73b0a627-5d81-4bf2-833a-caff3534c1e3 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: cf950ec3-9914-4eb9-99db-048b3969bb21] Allocating IP information in the background. {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 871.410256] env[65726]: DEBUG nova.network.neutron [None req-73b0a627-5d81-4bf2-833a-caff3534c1e3 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: cf950ec3-9914-4eb9-99db-048b3969bb21] allocate_for_instance() {{(pid=65726) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 871.410617] env[65726]: WARNING neutronclient.v2_0.client [None req-73b0a627-5d81-4bf2-833a-caff3534c1e3 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 871.410985] env[65726]: WARNING neutronclient.v2_0.client [None req-73b0a627-5d81-4bf2-833a-caff3534c1e3 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 871.412555] env[65726]: WARNING openstack [None req-73b0a627-5d81-4bf2-833a-caff3534c1e3 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 871.412555] env[65726]: WARNING openstack [None req-73b0a627-5d81-4bf2-833a-caff3534c1e3 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 871.432664] env[65726]: DEBUG oslo_concurrency.lockutils [None req-2cc737a8-ad7d-449c-b270-9db9e18d971d tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Acquiring lock "dcde25df-87a5-47fb-94ce-334f68894e04" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 871.432795] env[65726]: DEBUG oslo_concurrency.lockutils [None req-2cc737a8-ad7d-449c-b270-9db9e18d971d tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Lock "dcde25df-87a5-47fb-94ce-334f68894e04" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 871.484295] env[65726]: DEBUG nova.policy [None req-73b0a627-5d81-4bf2-833a-caff3534c1e3 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'faf82c34c4884dc890edf0909b9f510d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a8362295a62f4b51bae719a7ef5a4656', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65726) authorize /opt/stack/nova/nova/policy.py:192}} [ 871.497175] env[65726]: DEBUG oslo_vmware.api [None req-8333c704-ca27-426e-acee-8cfd63213196 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5202c109-7927-88f6-2a1d-1258cf9b3eba, 'name': SearchDatastore_Task, 'duration_secs': 0.010808} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 871.497623] env[65726]: DEBUG oslo_concurrency.lockutils [None req-8333c704-ca27-426e-acee-8cfd63213196 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 871.497872] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-8333c704-ca27-426e-acee-8cfd63213196 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: f2d9090c-988f-43f4-9c81-7aa718a3438a] Processing image b52362a3-ee8a-4cbf-b06f-513b0cc8f95c {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 871.498812] env[65726]: DEBUG oslo_concurrency.lockutils [None req-8333c704-ca27-426e-acee-8cfd63213196 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 871.498812] env[65726]: DEBUG oslo_concurrency.lockutils [None req-8333c704-ca27-426e-acee-8cfd63213196 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 871.498812] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-8333c704-ca27-426e-acee-8cfd63213196 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 871.499706] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b502f866-a4fa-44fc-9acb-2098e2f2c136 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.509936] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-8333c704-ca27-426e-acee-8cfd63213196 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 871.510274] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-8333c704-ca27-426e-acee-8cfd63213196 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65726) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 871.510789] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-51e7e803-2d03-4fb7-8285-7f5f129699fd {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.517553] env[65726]: DEBUG oslo_vmware.api [None req-8333c704-ca27-426e-acee-8cfd63213196 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Waiting for the task: (returnval){ [ 871.517553] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52809fe0-bd41-2853-3672-7942a1b96467" [ 871.517553] env[65726]: _type = "Task" [ 871.517553] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 871.531244] env[65726]: DEBUG oslo_vmware.api [None req-8333c704-ca27-426e-acee-8cfd63213196 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52809fe0-bd41-2853-3672-7942a1b96467, 'name': SearchDatastore_Task, 'duration_secs': 0.011098} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 871.531904] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5de2d763-ea0e-4fe4-94b2-795fe071c259 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.537882] env[65726]: DEBUG oslo_vmware.api [None req-8333c704-ca27-426e-acee-8cfd63213196 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Waiting for the task: (returnval){ [ 871.537882] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]525632f9-b73a-4ae1-ef25-2254d2b44da2" [ 871.537882] env[65726]: _type = "Task" [ 871.537882] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 871.546931] env[65726]: DEBUG oslo_vmware.api [None req-8333c704-ca27-426e-acee-8cfd63213196 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]525632f9-b73a-4ae1-ef25-2254d2b44da2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.552264] env[65726]: DEBUG oslo_concurrency.lockutils [None req-801b9b62-b5bd-4b5f-80fa-32388dd4cb88 tempest-ServerTagsTestJSON-59142761 tempest-ServerTagsTestJSON-59142761-project-member] Acquiring lock "9029549c-1914-4bae-91e2-8812b79051ec" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 871.552478] env[65726]: DEBUG oslo_concurrency.lockutils [None req-801b9b62-b5bd-4b5f-80fa-32388dd4cb88 tempest-ServerTagsTestJSON-59142761 tempest-ServerTagsTestJSON-59142761-project-member] Lock "9029549c-1914-4bae-91e2-8812b79051ec" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 871.612634] env[65726]: DEBUG nova.compute.manager [req-827eb3f7-7e07-4a63-a440-3efd95ae85f8 req-84c7c1f7-1600-43a2-8519-639980cccc75 service nova] [instance: c4cc4f85-cb35-4edc-a58b-adfee0ce1265] Received event network-vif-plugged-01aa3d90-3b27-4f98-a3ef-eb316f36f0cc {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 871.612728] env[65726]: DEBUG oslo_concurrency.lockutils [req-827eb3f7-7e07-4a63-a440-3efd95ae85f8 req-84c7c1f7-1600-43a2-8519-639980cccc75 service nova] Acquiring lock "c4cc4f85-cb35-4edc-a58b-adfee0ce1265-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 871.612900] env[65726]: DEBUG oslo_concurrency.lockutils [req-827eb3f7-7e07-4a63-a440-3efd95ae85f8 req-84c7c1f7-1600-43a2-8519-639980cccc75 service nova] Lock "c4cc4f85-cb35-4edc-a58b-adfee0ce1265-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 871.613178] env[65726]: DEBUG oslo_concurrency.lockutils [req-827eb3f7-7e07-4a63-a440-3efd95ae85f8 req-84c7c1f7-1600-43a2-8519-639980cccc75 service nova] Lock "c4cc4f85-cb35-4edc-a58b-adfee0ce1265-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 871.613399] env[65726]: DEBUG nova.compute.manager [req-827eb3f7-7e07-4a63-a440-3efd95ae85f8 req-84c7c1f7-1600-43a2-8519-639980cccc75 service nova] [instance: c4cc4f85-cb35-4edc-a58b-adfee0ce1265] No waiting events found dispatching network-vif-plugged-01aa3d90-3b27-4f98-a3ef-eb316f36f0cc {{(pid=65726) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 871.613602] env[65726]: WARNING nova.compute.manager [req-827eb3f7-7e07-4a63-a440-3efd95ae85f8 req-84c7c1f7-1600-43a2-8519-639980cccc75 service nova] [instance: c4cc4f85-cb35-4edc-a58b-adfee0ce1265] Received unexpected event network-vif-plugged-01aa3d90-3b27-4f98-a3ef-eb316f36f0cc for instance with vm_state building and task_state spawning. [ 871.630836] env[65726]: DEBUG nova.compute.manager [req-97b04017-9131-4098-9a49-be0f223bb24b req-8f6056a9-acea-4e23-8e9d-88207ed89bd3 service nova] [instance: c4177e20-b1bd-4b54-a275-c93582359a07] Received event network-vif-deleted-da625869-f814-4733-b2ac-d50cc54a9858 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 871.682073] env[65726]: WARNING openstack [None req-49b35bf1-9410-49fb-9a5b-632c41ecd3a6 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 871.682575] env[65726]: WARNING openstack [None req-49b35bf1-9410-49fb-9a5b-632c41ecd3a6 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 871.691753] env[65726]: DEBUG nova.network.neutron [None req-49b35bf1-9410-49fb-9a5b-632c41ecd3a6 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] [instance: c4cc4f85-cb35-4edc-a58b-adfee0ce1265] Instance cache missing network info. {{(pid=65726) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 871.755634] env[65726]: WARNING openstack [None req-49b35bf1-9410-49fb-9a5b-632c41ecd3a6 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 871.756371] env[65726]: WARNING openstack [None req-49b35bf1-9410-49fb-9a5b-632c41ecd3a6 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 871.772815] env[65726]: DEBUG nova.network.neutron [None req-73b0a627-5d81-4bf2-833a-caff3534c1e3 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: cf950ec3-9914-4eb9-99db-048b3969bb21] Successfully created port: 58accf41-587f-4367-854e-fc571a6a3424 {{(pid=65726) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 871.840847] env[65726]: WARNING openstack [None req-49b35bf1-9410-49fb-9a5b-632c41ecd3a6 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 871.840847] env[65726]: WARNING openstack [None req-49b35bf1-9410-49fb-9a5b-632c41ecd3a6 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 871.885692] env[65726]: INFO nova.compute.manager [-] [instance: c4177e20-b1bd-4b54-a275-c93582359a07] Took 1.59 seconds to deallocate network for instance. [ 871.922023] env[65726]: DEBUG nova.compute.manager [None req-73b0a627-5d81-4bf2-833a-caff3534c1e3 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: cf950ec3-9914-4eb9-99db-048b3969bb21] Start building block device mappings for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 871.947922] env[65726]: DEBUG nova.network.neutron [None req-49b35bf1-9410-49fb-9a5b-632c41ecd3a6 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] [instance: c4cc4f85-cb35-4edc-a58b-adfee0ce1265] Updating instance_info_cache with network_info: [{"id": "01aa3d90-3b27-4f98-a3ef-eb316f36f0cc", "address": "fa:16:3e:df:b2:05", "network": {"id": "53875783-ff10-40b4-a7a0-81e6e7b607e1", "bridge": "br-int", "label": "tempest-ImagesOneServerTestJSON-1511040822-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2772e93581124978b55c8054d2116232", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f5fe645c-e088-401e-ab53-4ae2981dea72", "external-id": "nsx-vlan-transportzone-219", "segmentation_id": 219, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap01aa3d90-3b", "ovs_interfaceid": "01aa3d90-3b27-4f98-a3ef-eb316f36f0cc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 872.049719] env[65726]: DEBUG oslo_vmware.api [None req-8333c704-ca27-426e-acee-8cfd63213196 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]525632f9-b73a-4ae1-ef25-2254d2b44da2, 'name': SearchDatastore_Task, 'duration_secs': 0.011577} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.053390] env[65726]: DEBUG oslo_concurrency.lockutils [None req-8333c704-ca27-426e-acee-8cfd63213196 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 872.053390] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-8333c704-ca27-426e-acee-8cfd63213196 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore2] f2d9090c-988f-43f4-9c81-7aa718a3438a/f2d9090c-988f-43f4-9c81-7aa718a3438a.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 872.053647] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ce8e904a-9b74-4cc7-82d6-56a7619b37ad {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.062069] env[65726]: DEBUG oslo_vmware.api [None req-8333c704-ca27-426e-acee-8cfd63213196 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Waiting for the task: (returnval){ [ 872.062069] env[65726]: value = "task-5116082" [ 872.062069] env[65726]: _type = "Task" [ 872.062069] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.076717] env[65726]: DEBUG oslo_vmware.api [None req-8333c704-ca27-426e-acee-8cfd63213196 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Task: {'id': task-5116082, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.394379] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a4b79740-ccaa-4794-bd78-55e7c8646033 tempest-ImagesNegativeTestJSON-1015404931 tempest-ImagesNegativeTestJSON-1015404931-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 872.411879] env[65726]: INFO nova.compute.manager [None req-c85b0a62-807d-4793-b1be-86b28d9a6745 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] [instance: 2cbee359-a1fc-45c0-bcc5-a2c24ded9c69] Rebuilding instance [ 872.452918] env[65726]: DEBUG oslo_concurrency.lockutils [None req-49b35bf1-9410-49fb-9a5b-632c41ecd3a6 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Releasing lock "refresh_cache-c4cc4f85-cb35-4edc-a58b-adfee0ce1265" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 872.453531] env[65726]: DEBUG nova.compute.manager [None req-49b35bf1-9410-49fb-9a5b-632c41ecd3a6 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] [instance: c4cc4f85-cb35-4edc-a58b-adfee0ce1265] Instance network_info: |[{"id": "01aa3d90-3b27-4f98-a3ef-eb316f36f0cc", "address": "fa:16:3e:df:b2:05", "network": {"id": "53875783-ff10-40b4-a7a0-81e6e7b607e1", "bridge": "br-int", "label": "tempest-ImagesOneServerTestJSON-1511040822-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2772e93581124978b55c8054d2116232", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f5fe645c-e088-401e-ab53-4ae2981dea72", "external-id": "nsx-vlan-transportzone-219", "segmentation_id": 219, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap01aa3d90-3b", "ovs_interfaceid": "01aa3d90-3b27-4f98-a3ef-eb316f36f0cc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 872.459079] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-49b35bf1-9410-49fb-9a5b-632c41ecd3a6 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] [instance: c4cc4f85-cb35-4edc-a58b-adfee0ce1265] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:df:b2:05', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f5fe645c-e088-401e-ab53-4ae2981dea72', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '01aa3d90-3b27-4f98-a3ef-eb316f36f0cc', 'vif_model': 'vmxnet3'}] {{(pid=65726) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 872.467693] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-49b35bf1-9410-49fb-9a5b-632c41ecd3a6 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Creating folder: Project (2772e93581124978b55c8054d2116232). Parent ref: group-v995008. {{(pid=65726) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 872.470986] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ba604f55-1cb9-433e-8763-72e58ba6ea0a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.485511] env[65726]: INFO nova.virt.vmwareapi.vm_util [None req-49b35bf1-9410-49fb-9a5b-632c41ecd3a6 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Created folder: Project (2772e93581124978b55c8054d2116232) in parent group-v995008. [ 872.485714] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-49b35bf1-9410-49fb-9a5b-632c41ecd3a6 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Creating folder: Instances. Parent ref: group-v995176. {{(pid=65726) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 872.486269] env[65726]: DEBUG nova.compute.manager [None req-c85b0a62-807d-4793-b1be-86b28d9a6745 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] [instance: 2cbee359-a1fc-45c0-bcc5-a2c24ded9c69] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 872.489115] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-258bec58-e978-4538-a311-d23a217a61c8 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.492046] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ba5e93c-7f7e-4c6c-b6d9-a8b3f622871e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.506293] env[65726]: INFO nova.virt.vmwareapi.vm_util [None req-49b35bf1-9410-49fb-9a5b-632c41ecd3a6 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Created folder: Instances in parent group-v995176. [ 872.506559] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-49b35bf1-9410-49fb-9a5b-632c41ecd3a6 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 872.506814] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c4cc4f85-cb35-4edc-a58b-adfee0ce1265] Creating VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 872.507119] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6964c88a-e4c6-4818-aada-3a5be713442d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.533678] env[65726]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 872.533678] env[65726]: value = "task-5116085" [ 872.533678] env[65726]: _type = "Task" [ 872.533678] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.550641] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116085, 'name': CreateVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.580943] env[65726]: DEBUG oslo_vmware.api [None req-8333c704-ca27-426e-acee-8cfd63213196 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Task: {'id': task-5116082, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.583297] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c44adfbc-554f-4bf6-a8ff-37d2b545e3e6 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.593874] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e64602fc-2681-4008-9ed8-89e5be11a71d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.634526] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-554410c0-d6e4-4c9c-bf3d-65a78295ac8d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.645672] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46de0b72-7a43-4aa9-87dd-5a827cc954bc {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.664042] env[65726]: DEBUG nova.compute.provider_tree [None req-b20cc073-fafc-44fa-bcf9-9893dc98ebec tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 872.946216] env[65726]: DEBUG nova.compute.manager [None req-73b0a627-5d81-4bf2-833a-caff3534c1e3 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: cf950ec3-9914-4eb9-99db-048b3969bb21] Start spawning the instance on the hypervisor. {{(pid=65726) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 872.967219] env[65726]: DEBUG nova.virt.hardware [None req-73b0a627-5d81-4bf2-833a-caff3534c1e3 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T19:28:34Z,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1539f06fa3534e90acbc3a60c4b8bd3f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T19:28:35Z,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 872.967475] env[65726]: DEBUG nova.virt.hardware [None req-73b0a627-5d81-4bf2-833a-caff3534c1e3 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 872.967630] env[65726]: DEBUG nova.virt.hardware [None req-73b0a627-5d81-4bf2-833a-caff3534c1e3 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 872.967826] env[65726]: DEBUG nova.virt.hardware [None req-73b0a627-5d81-4bf2-833a-caff3534c1e3 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 872.967964] env[65726]: DEBUG nova.virt.hardware [None req-73b0a627-5d81-4bf2-833a-caff3534c1e3 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 872.968117] env[65726]: DEBUG nova.virt.hardware [None req-73b0a627-5d81-4bf2-833a-caff3534c1e3 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 872.968319] env[65726]: DEBUG nova.virt.hardware [None req-73b0a627-5d81-4bf2-833a-caff3534c1e3 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 872.968502] env[65726]: DEBUG nova.virt.hardware [None req-73b0a627-5d81-4bf2-833a-caff3534c1e3 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 872.968745] env[65726]: DEBUG nova.virt.hardware [None req-73b0a627-5d81-4bf2-833a-caff3534c1e3 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 872.968920] env[65726]: DEBUG nova.virt.hardware [None req-73b0a627-5d81-4bf2-833a-caff3534c1e3 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 872.969128] env[65726]: DEBUG nova.virt.hardware [None req-73b0a627-5d81-4bf2-833a-caff3534c1e3 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 872.969979] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89e75ae6-f95e-4fa4-bde7-e38144b7a3de {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.979220] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59b3f776-ed2e-48c0-860d-b35388658f7e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.043529] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116085, 'name': CreateVM_Task} progress is 99%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.078816] env[65726]: DEBUG oslo_vmware.api [None req-8333c704-ca27-426e-acee-8cfd63213196 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Task: {'id': task-5116082, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.803662} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.079097] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-8333c704-ca27-426e-acee-8cfd63213196 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore2] f2d9090c-988f-43f4-9c81-7aa718a3438a/f2d9090c-988f-43f4-9c81-7aa718a3438a.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 873.079315] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-8333c704-ca27-426e-acee-8cfd63213196 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: f2d9090c-988f-43f4-9c81-7aa718a3438a] Extending root virtual disk to 1048576 {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 873.079856] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-9ebfbae4-dc51-47b8-a3f7-315669615342 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.087405] env[65726]: DEBUG oslo_vmware.api [None req-8333c704-ca27-426e-acee-8cfd63213196 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Waiting for the task: (returnval){ [ 873.087405] env[65726]: value = "task-5116086" [ 873.087405] env[65726]: _type = "Task" [ 873.087405] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.095485] env[65726]: DEBUG oslo_vmware.api [None req-8333c704-ca27-426e-acee-8cfd63213196 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Task: {'id': task-5116086, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.167968] env[65726]: DEBUG nova.scheduler.client.report [None req-b20cc073-fafc-44fa-bcf9-9893dc98ebec tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 873.429132] env[65726]: DEBUG nova.network.neutron [None req-73b0a627-5d81-4bf2-833a-caff3534c1e3 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: cf950ec3-9914-4eb9-99db-048b3969bb21] Successfully updated port: 58accf41-587f-4367-854e-fc571a6a3424 {{(pid=65726) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 873.488260] env[65726]: DEBUG oslo_concurrency.lockutils [None req-037b42a1-7b2a-4493-8518-be35c6d71062 tempest-ServerMetadataNegativeTestJSON-936475568 tempest-ServerMetadataNegativeTestJSON-936475568-project-member] Acquiring lock "d179030a-e5ba-45b1-ad11-4d2f71ed7bd4" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 873.488556] env[65726]: DEBUG oslo_concurrency.lockutils [None req-037b42a1-7b2a-4493-8518-be35c6d71062 tempest-ServerMetadataNegativeTestJSON-936475568 tempest-ServerMetadataNegativeTestJSON-936475568-project-member] Lock "d179030a-e5ba-45b1-ad11-4d2f71ed7bd4" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 873.488787] env[65726]: DEBUG oslo_concurrency.lockutils [None req-037b42a1-7b2a-4493-8518-be35c6d71062 tempest-ServerMetadataNegativeTestJSON-936475568 tempest-ServerMetadataNegativeTestJSON-936475568-project-member] Acquiring lock "d179030a-e5ba-45b1-ad11-4d2f71ed7bd4-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 873.488974] env[65726]: DEBUG oslo_concurrency.lockutils [None req-037b42a1-7b2a-4493-8518-be35c6d71062 tempest-ServerMetadataNegativeTestJSON-936475568 tempest-ServerMetadataNegativeTestJSON-936475568-project-member] Lock "d179030a-e5ba-45b1-ad11-4d2f71ed7bd4-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 873.489232] env[65726]: DEBUG oslo_concurrency.lockutils [None req-037b42a1-7b2a-4493-8518-be35c6d71062 tempest-ServerMetadataNegativeTestJSON-936475568 tempest-ServerMetadataNegativeTestJSON-936475568-project-member] Lock "d179030a-e5ba-45b1-ad11-4d2f71ed7bd4-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 873.491605] env[65726]: INFO nova.compute.manager [None req-037b42a1-7b2a-4493-8518-be35c6d71062 tempest-ServerMetadataNegativeTestJSON-936475568 tempest-ServerMetadataNegativeTestJSON-936475568-project-member] [instance: d179030a-e5ba-45b1-ad11-4d2f71ed7bd4] Terminating instance [ 873.507038] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-c85b0a62-807d-4793-b1be-86b28d9a6745 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] [instance: 2cbee359-a1fc-45c0-bcc5-a2c24ded9c69] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 873.507325] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c11b605c-fd0a-4071-8d7c-f69e732372e6 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.514991] env[65726]: DEBUG oslo_vmware.api [None req-c85b0a62-807d-4793-b1be-86b28d9a6745 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] Waiting for the task: (returnval){ [ 873.514991] env[65726]: value = "task-5116087" [ 873.514991] env[65726]: _type = "Task" [ 873.514991] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.523809] env[65726]: DEBUG oslo_vmware.api [None req-c85b0a62-807d-4793-b1be-86b28d9a6745 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] Task: {'id': task-5116087, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.544817] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116085, 'name': CreateVM_Task} progress is 99%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.597357] env[65726]: DEBUG oslo_vmware.api [None req-8333c704-ca27-426e-acee-8cfd63213196 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Task: {'id': task-5116086, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.071342} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.597598] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-8333c704-ca27-426e-acee-8cfd63213196 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: f2d9090c-988f-43f4-9c81-7aa718a3438a] Extended root virtual disk {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 873.598353] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1929c60-09a8-486c-a2bf-bcceea0929ba {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.622189] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-8333c704-ca27-426e-acee-8cfd63213196 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: f2d9090c-988f-43f4-9c81-7aa718a3438a] Reconfiguring VM instance instance-00000011 to attach disk [datastore2] f2d9090c-988f-43f4-9c81-7aa718a3438a/f2d9090c-988f-43f4-9c81-7aa718a3438a.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 873.622526] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-35b896a4-f107-4e25-8970-5fa0c3e11ca6 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.643195] env[65726]: DEBUG oslo_vmware.api [None req-8333c704-ca27-426e-acee-8cfd63213196 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Waiting for the task: (returnval){ [ 873.643195] env[65726]: value = "task-5116088" [ 873.643195] env[65726]: _type = "Task" [ 873.643195] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.651639] env[65726]: DEBUG oslo_vmware.api [None req-8333c704-ca27-426e-acee-8cfd63213196 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Task: {'id': task-5116088, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.661607] env[65726]: DEBUG nova.compute.manager [req-47c21cae-4c82-4bfa-87ef-33c0c45669c3 req-0a024ec1-973d-47cb-9652-2eba67a9a450 service nova] [instance: c4cc4f85-cb35-4edc-a58b-adfee0ce1265] Received event network-changed-01aa3d90-3b27-4f98-a3ef-eb316f36f0cc {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 873.661874] env[65726]: DEBUG nova.compute.manager [req-47c21cae-4c82-4bfa-87ef-33c0c45669c3 req-0a024ec1-973d-47cb-9652-2eba67a9a450 service nova] [instance: c4cc4f85-cb35-4edc-a58b-adfee0ce1265] Refreshing instance network info cache due to event network-changed-01aa3d90-3b27-4f98-a3ef-eb316f36f0cc. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 873.662226] env[65726]: DEBUG oslo_concurrency.lockutils [req-47c21cae-4c82-4bfa-87ef-33c0c45669c3 req-0a024ec1-973d-47cb-9652-2eba67a9a450 service nova] Acquiring lock "refresh_cache-c4cc4f85-cb35-4edc-a58b-adfee0ce1265" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 873.662267] env[65726]: DEBUG oslo_concurrency.lockutils [req-47c21cae-4c82-4bfa-87ef-33c0c45669c3 req-0a024ec1-973d-47cb-9652-2eba67a9a450 service nova] Acquired lock "refresh_cache-c4cc4f85-cb35-4edc-a58b-adfee0ce1265" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 873.662466] env[65726]: DEBUG nova.network.neutron [req-47c21cae-4c82-4bfa-87ef-33c0c45669c3 req-0a024ec1-973d-47cb-9652-2eba67a9a450 service nova] [instance: c4cc4f85-cb35-4edc-a58b-adfee0ce1265] Refreshing network info cache for port 01aa3d90-3b27-4f98-a3ef-eb316f36f0cc {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 873.673804] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b20cc073-fafc-44fa-bcf9-9893dc98ebec tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.767s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 873.674431] env[65726]: DEBUG nova.compute.manager [None req-b20cc073-fafc-44fa-bcf9-9893dc98ebec tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04] Start building networks asynchronously for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 873.677016] env[65726]: DEBUG oslo_concurrency.lockutils [None req-398d7f82-4e73-43c8-8dc0-2fe897bf965f tempest-ServersNegativeTestMultiTenantJSON-944098855 tempest-ServersNegativeTestMultiTenantJSON-944098855-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 39.872s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 873.677511] env[65726]: DEBUG nova.objects.instance [None req-398d7f82-4e73-43c8-8dc0-2fe897bf965f tempest-ServersNegativeTestMultiTenantJSON-944098855 tempest-ServersNegativeTestMultiTenantJSON-944098855-project-member] Lazy-loading 'resources' on Instance uuid 018c8ac5-66c4-4a9a-ab45-85e84f92b4fe {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 873.743830] env[65726]: DEBUG nova.compute.manager [req-518cb5bf-8235-4ca9-a600-563380cd1dd4 req-1eff9706-c375-4786-8450-a6677735b12a service nova] [instance: cf950ec3-9914-4eb9-99db-048b3969bb21] Received event network-vif-plugged-58accf41-587f-4367-854e-fc571a6a3424 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 873.744073] env[65726]: DEBUG oslo_concurrency.lockutils [req-518cb5bf-8235-4ca9-a600-563380cd1dd4 req-1eff9706-c375-4786-8450-a6677735b12a service nova] Acquiring lock "cf950ec3-9914-4eb9-99db-048b3969bb21-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 873.744310] env[65726]: DEBUG oslo_concurrency.lockutils [req-518cb5bf-8235-4ca9-a600-563380cd1dd4 req-1eff9706-c375-4786-8450-a6677735b12a service nova] Lock "cf950ec3-9914-4eb9-99db-048b3969bb21-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 873.744476] env[65726]: DEBUG oslo_concurrency.lockutils [req-518cb5bf-8235-4ca9-a600-563380cd1dd4 req-1eff9706-c375-4786-8450-a6677735b12a service nova] Lock "cf950ec3-9914-4eb9-99db-048b3969bb21-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 873.744641] env[65726]: DEBUG nova.compute.manager [req-518cb5bf-8235-4ca9-a600-563380cd1dd4 req-1eff9706-c375-4786-8450-a6677735b12a service nova] [instance: cf950ec3-9914-4eb9-99db-048b3969bb21] No waiting events found dispatching network-vif-plugged-58accf41-587f-4367-854e-fc571a6a3424 {{(pid=65726) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 873.744801] env[65726]: WARNING nova.compute.manager [req-518cb5bf-8235-4ca9-a600-563380cd1dd4 req-1eff9706-c375-4786-8450-a6677735b12a service nova] [instance: cf950ec3-9914-4eb9-99db-048b3969bb21] Received unexpected event network-vif-plugged-58accf41-587f-4367-854e-fc571a6a3424 for instance with vm_state building and task_state spawning. [ 873.744956] env[65726]: DEBUG nova.compute.manager [req-518cb5bf-8235-4ca9-a600-563380cd1dd4 req-1eff9706-c375-4786-8450-a6677735b12a service nova] [instance: cf950ec3-9914-4eb9-99db-048b3969bb21] Received event network-changed-58accf41-587f-4367-854e-fc571a6a3424 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 873.745119] env[65726]: DEBUG nova.compute.manager [req-518cb5bf-8235-4ca9-a600-563380cd1dd4 req-1eff9706-c375-4786-8450-a6677735b12a service nova] [instance: cf950ec3-9914-4eb9-99db-048b3969bb21] Refreshing instance network info cache due to event network-changed-58accf41-587f-4367-854e-fc571a6a3424. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 873.745292] env[65726]: DEBUG oslo_concurrency.lockutils [req-518cb5bf-8235-4ca9-a600-563380cd1dd4 req-1eff9706-c375-4786-8450-a6677735b12a service nova] Acquiring lock "refresh_cache-cf950ec3-9914-4eb9-99db-048b3969bb21" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 873.745420] env[65726]: DEBUG oslo_concurrency.lockutils [req-518cb5bf-8235-4ca9-a600-563380cd1dd4 req-1eff9706-c375-4786-8450-a6677735b12a service nova] Acquired lock "refresh_cache-cf950ec3-9914-4eb9-99db-048b3969bb21" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 873.745611] env[65726]: DEBUG nova.network.neutron [req-518cb5bf-8235-4ca9-a600-563380cd1dd4 req-1eff9706-c375-4786-8450-a6677735b12a service nova] [instance: cf950ec3-9914-4eb9-99db-048b3969bb21] Refreshing network info cache for port 58accf41-587f-4367-854e-fc571a6a3424 {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 873.931371] env[65726]: DEBUG oslo_concurrency.lockutils [None req-73b0a627-5d81-4bf2-833a-caff3534c1e3 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Acquiring lock "refresh_cache-cf950ec3-9914-4eb9-99db-048b3969bb21" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 873.995839] env[65726]: DEBUG nova.compute.manager [None req-037b42a1-7b2a-4493-8518-be35c6d71062 tempest-ServerMetadataNegativeTestJSON-936475568 tempest-ServerMetadataNegativeTestJSON-936475568-project-member] [instance: d179030a-e5ba-45b1-ad11-4d2f71ed7bd4] Start destroying the instance on the hypervisor. {{(pid=65726) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 873.996224] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-037b42a1-7b2a-4493-8518-be35c6d71062 tempest-ServerMetadataNegativeTestJSON-936475568 tempest-ServerMetadataNegativeTestJSON-936475568-project-member] [instance: d179030a-e5ba-45b1-ad11-4d2f71ed7bd4] Destroying instance {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 873.997567] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5e6e4ac-0008-4339-8c42-356d9062ba1e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.008307] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-037b42a1-7b2a-4493-8518-be35c6d71062 tempest-ServerMetadataNegativeTestJSON-936475568 tempest-ServerMetadataNegativeTestJSON-936475568-project-member] [instance: d179030a-e5ba-45b1-ad11-4d2f71ed7bd4] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 874.008650] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9af487e5-2dc5-4704-8d81-e15b22617dcc {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.021318] env[65726]: DEBUG oslo_vmware.api [None req-037b42a1-7b2a-4493-8518-be35c6d71062 tempest-ServerMetadataNegativeTestJSON-936475568 tempest-ServerMetadataNegativeTestJSON-936475568-project-member] Waiting for the task: (returnval){ [ 874.021318] env[65726]: value = "task-5116089" [ 874.021318] env[65726]: _type = "Task" [ 874.021318] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.029616] env[65726]: DEBUG oslo_vmware.api [None req-c85b0a62-807d-4793-b1be-86b28d9a6745 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] Task: {'id': task-5116087, 'name': PowerOffVM_Task} progress is 100%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.036453] env[65726]: DEBUG oslo_vmware.api [None req-037b42a1-7b2a-4493-8518-be35c6d71062 tempest-ServerMetadataNegativeTestJSON-936475568 tempest-ServerMetadataNegativeTestJSON-936475568-project-member] Task: {'id': task-5116089, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.046123] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116085, 'name': CreateVM_Task, 'duration_secs': 1.491803} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.046312] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c4cc4f85-cb35-4edc-a58b-adfee0ce1265] Created VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 874.047207] env[65726]: WARNING openstack [None req-49b35bf1-9410-49fb-9a5b-632c41ecd3a6 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 874.047607] env[65726]: WARNING openstack [None req-49b35bf1-9410-49fb-9a5b-632c41ecd3a6 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 874.053365] env[65726]: DEBUG oslo_concurrency.lockutils [None req-49b35bf1-9410-49fb-9a5b-632c41ecd3a6 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 874.053566] env[65726]: DEBUG oslo_concurrency.lockutils [None req-49b35bf1-9410-49fb-9a5b-632c41ecd3a6 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 874.053946] env[65726]: DEBUG oslo_concurrency.lockutils [None req-49b35bf1-9410-49fb-9a5b-632c41ecd3a6 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 874.054331] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fb31a516-f977-4feb-b749-e9125792bd8c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.061384] env[65726]: DEBUG oslo_vmware.api [None req-49b35bf1-9410-49fb-9a5b-632c41ecd3a6 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Waiting for the task: (returnval){ [ 874.061384] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52b38a1f-499b-9294-9269-15127de12937" [ 874.061384] env[65726]: _type = "Task" [ 874.061384] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.070479] env[65726]: DEBUG oslo_vmware.api [None req-49b35bf1-9410-49fb-9a5b-632c41ecd3a6 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52b38a1f-499b-9294-9269-15127de12937, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.154818] env[65726]: DEBUG oslo_vmware.api [None req-8333c704-ca27-426e-acee-8cfd63213196 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Task: {'id': task-5116088, 'name': ReconfigVM_Task, 'duration_secs': 0.290469} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.155188] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-8333c704-ca27-426e-acee-8cfd63213196 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: f2d9090c-988f-43f4-9c81-7aa718a3438a] Reconfigured VM instance instance-00000011 to attach disk [datastore2] f2d9090c-988f-43f4-9c81-7aa718a3438a/f2d9090c-988f-43f4-9c81-7aa718a3438a.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 874.155878] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1f49d0c0-ff97-4468-ac03-6ebb548f74c8 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.163911] env[65726]: DEBUG oslo_vmware.api [None req-8333c704-ca27-426e-acee-8cfd63213196 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Waiting for the task: (returnval){ [ 874.163911] env[65726]: value = "task-5116090" [ 874.163911] env[65726]: _type = "Task" [ 874.163911] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.169225] env[65726]: WARNING openstack [req-47c21cae-4c82-4bfa-87ef-33c0c45669c3 req-0a024ec1-973d-47cb-9652-2eba67a9a450 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 874.169846] env[65726]: WARNING openstack [req-47c21cae-4c82-4bfa-87ef-33c0c45669c3 req-0a024ec1-973d-47cb-9652-2eba67a9a450 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 874.180932] env[65726]: DEBUG nova.compute.utils [None req-b20cc073-fafc-44fa-bcf9-9893dc98ebec tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Using /dev/sd instead of None {{(pid=65726) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 874.190395] env[65726]: DEBUG nova.compute.manager [None req-b20cc073-fafc-44fa-bcf9-9893dc98ebec tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04] Allocating IP information in the background. {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 874.191773] env[65726]: DEBUG nova.network.neutron [None req-b20cc073-fafc-44fa-bcf9-9893dc98ebec tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04] allocate_for_instance() {{(pid=65726) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 874.191773] env[65726]: WARNING neutronclient.v2_0.client [None req-b20cc073-fafc-44fa-bcf9-9893dc98ebec tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 874.191773] env[65726]: WARNING neutronclient.v2_0.client [None req-b20cc073-fafc-44fa-bcf9-9893dc98ebec tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 874.192319] env[65726]: WARNING openstack [None req-b20cc073-fafc-44fa-bcf9-9893dc98ebec tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 874.192685] env[65726]: WARNING openstack [None req-b20cc073-fafc-44fa-bcf9-9893dc98ebec tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 874.200305] env[65726]: DEBUG oslo_vmware.api [None req-8333c704-ca27-426e-acee-8cfd63213196 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Task: {'id': task-5116090, 'name': Rename_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.249422] env[65726]: WARNING openstack [req-518cb5bf-8235-4ca9-a600-563380cd1dd4 req-1eff9706-c375-4786-8450-a6677735b12a service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 874.249904] env[65726]: WARNING openstack [req-518cb5bf-8235-4ca9-a600-563380cd1dd4 req-1eff9706-c375-4786-8450-a6677735b12a service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 874.255696] env[65726]: DEBUG nova.network.neutron [req-518cb5bf-8235-4ca9-a600-563380cd1dd4 req-1eff9706-c375-4786-8450-a6677735b12a service nova] [instance: cf950ec3-9914-4eb9-99db-048b3969bb21] Instance cache missing network info. {{(pid=65726) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 874.280968] env[65726]: DEBUG nova.policy [None req-b20cc073-fafc-44fa-bcf9-9893dc98ebec tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e6593c9d94ad4425ab0aede7e76ebb31', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '96149159e18e44f9bf3453e67681f224', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65726) authorize /opt/stack/nova/nova/policy.py:192}} [ 874.305308] env[65726]: WARNING openstack [req-47c21cae-4c82-4bfa-87ef-33c0c45669c3 req-0a024ec1-973d-47cb-9652-2eba67a9a450 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 874.305795] env[65726]: WARNING openstack [req-47c21cae-4c82-4bfa-87ef-33c0c45669c3 req-0a024ec1-973d-47cb-9652-2eba67a9a450 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 874.433675] env[65726]: WARNING openstack [req-47c21cae-4c82-4bfa-87ef-33c0c45669c3 req-0a024ec1-973d-47cb-9652-2eba67a9a450 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 874.434977] env[65726]: WARNING openstack [req-47c21cae-4c82-4bfa-87ef-33c0c45669c3 req-0a024ec1-973d-47cb-9652-2eba67a9a450 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 874.487014] env[65726]: DEBUG nova.network.neutron [req-518cb5bf-8235-4ca9-a600-563380cd1dd4 req-1eff9706-c375-4786-8450-a6677735b12a service nova] [instance: cf950ec3-9914-4eb9-99db-048b3969bb21] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 874.532445] env[65726]: DEBUG oslo_vmware.api [None req-c85b0a62-807d-4793-b1be-86b28d9a6745 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] Task: {'id': task-5116087, 'name': PowerOffVM_Task, 'duration_secs': 0.556842} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.533201] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-c85b0a62-807d-4793-b1be-86b28d9a6745 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] [instance: 2cbee359-a1fc-45c0-bcc5-a2c24ded9c69] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 874.533498] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-c85b0a62-807d-4793-b1be-86b28d9a6745 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] [instance: 2cbee359-a1fc-45c0-bcc5-a2c24ded9c69] Destroying instance {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 874.534294] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36160760-49e5-4d1c-8223-06a6f9f3c285 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.543337] env[65726]: DEBUG oslo_vmware.api [None req-037b42a1-7b2a-4493-8518-be35c6d71062 tempest-ServerMetadataNegativeTestJSON-936475568 tempest-ServerMetadataNegativeTestJSON-936475568-project-member] Task: {'id': task-5116089, 'name': PowerOffVM_Task, 'duration_secs': 0.250142} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.544208] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-037b42a1-7b2a-4493-8518-be35c6d71062 tempest-ServerMetadataNegativeTestJSON-936475568 tempest-ServerMetadataNegativeTestJSON-936475568-project-member] [instance: d179030a-e5ba-45b1-ad11-4d2f71ed7bd4] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 874.545584] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-037b42a1-7b2a-4493-8518-be35c6d71062 tempest-ServerMetadataNegativeTestJSON-936475568 tempest-ServerMetadataNegativeTestJSON-936475568-project-member] [instance: d179030a-e5ba-45b1-ad11-4d2f71ed7bd4] Unregistering the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 874.550079] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-75a1550d-3ad4-41d4-9474-97cb15f030e4 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.554505] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-c85b0a62-807d-4793-b1be-86b28d9a6745 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] [instance: 2cbee359-a1fc-45c0-bcc5-a2c24ded9c69] Unregistering the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 874.557407] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1095ba91-7c91-4ac9-80c1-da095bcc6298 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.574876] env[65726]: DEBUG oslo_vmware.api [None req-49b35bf1-9410-49fb-9a5b-632c41ecd3a6 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52b38a1f-499b-9294-9269-15127de12937, 'name': SearchDatastore_Task, 'duration_secs': 0.011771} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.577975] env[65726]: DEBUG oslo_concurrency.lockutils [None req-49b35bf1-9410-49fb-9a5b-632c41ecd3a6 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 874.578282] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-49b35bf1-9410-49fb-9a5b-632c41ecd3a6 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] [instance: c4cc4f85-cb35-4edc-a58b-adfee0ce1265] Processing image b52362a3-ee8a-4cbf-b06f-513b0cc8f95c {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 874.578571] env[65726]: DEBUG oslo_concurrency.lockutils [None req-49b35bf1-9410-49fb-9a5b-632c41ecd3a6 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 874.578728] env[65726]: DEBUG oslo_concurrency.lockutils [None req-49b35bf1-9410-49fb-9a5b-632c41ecd3a6 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 874.578910] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-49b35bf1-9410-49fb-9a5b-632c41ecd3a6 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 874.580609] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7cc00471-98d3-47a1-a915-543d351e9939 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.586625] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-c85b0a62-807d-4793-b1be-86b28d9a6745 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] [instance: 2cbee359-a1fc-45c0-bcc5-a2c24ded9c69] Unregistered the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 874.586897] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-c85b0a62-807d-4793-b1be-86b28d9a6745 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] [instance: 2cbee359-a1fc-45c0-bcc5-a2c24ded9c69] Deleting contents of the VM from datastore datastore2 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 874.587097] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-c85b0a62-807d-4793-b1be-86b28d9a6745 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] Deleting the datastore file [datastore2] 2cbee359-a1fc-45c0-bcc5-a2c24ded9c69 {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 874.587371] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-aab786a9-8390-45fc-8bf3-0c860d5c4e7e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.603478] env[65726]: DEBUG nova.network.neutron [None req-b20cc073-fafc-44fa-bcf9-9893dc98ebec tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04] Successfully created port: 25b9d108-0b43-4459-b9db-7bd90a495bb3 {{(pid=65726) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 874.609951] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-49b35bf1-9410-49fb-9a5b-632c41ecd3a6 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 874.610292] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-49b35bf1-9410-49fb-9a5b-632c41ecd3a6 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65726) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 874.611366] env[65726]: DEBUG oslo_vmware.api [None req-c85b0a62-807d-4793-b1be-86b28d9a6745 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] Waiting for the task: (returnval){ [ 874.611366] env[65726]: value = "task-5116093" [ 874.611366] env[65726]: _type = "Task" [ 874.611366] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.611656] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8b1b1063-17d6-4520-8fbb-8ec239402ba4 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.618899] env[65726]: DEBUG nova.network.neutron [req-47c21cae-4c82-4bfa-87ef-33c0c45669c3 req-0a024ec1-973d-47cb-9652-2eba67a9a450 service nova] [instance: c4cc4f85-cb35-4edc-a58b-adfee0ce1265] Updated VIF entry in instance network info cache for port 01aa3d90-3b27-4f98-a3ef-eb316f36f0cc. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 874.619280] env[65726]: DEBUG nova.network.neutron [req-47c21cae-4c82-4bfa-87ef-33c0c45669c3 req-0a024ec1-973d-47cb-9652-2eba67a9a450 service nova] [instance: c4cc4f85-cb35-4edc-a58b-adfee0ce1265] Updating instance_info_cache with network_info: [{"id": "01aa3d90-3b27-4f98-a3ef-eb316f36f0cc", "address": "fa:16:3e:df:b2:05", "network": {"id": "53875783-ff10-40b4-a7a0-81e6e7b607e1", "bridge": "br-int", "label": "tempest-ImagesOneServerTestJSON-1511040822-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2772e93581124978b55c8054d2116232", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f5fe645c-e088-401e-ab53-4ae2981dea72", "external-id": "nsx-vlan-transportzone-219", "segmentation_id": 219, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap01aa3d90-3b", "ovs_interfaceid": "01aa3d90-3b27-4f98-a3ef-eb316f36f0cc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 874.631431] env[65726]: DEBUG oslo_vmware.api [None req-49b35bf1-9410-49fb-9a5b-632c41ecd3a6 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Waiting for the task: (returnval){ [ 874.631431] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52bc150d-bd34-3b00-edef-340ee1d90a4d" [ 874.631431] env[65726]: _type = "Task" [ 874.631431] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.635571] env[65726]: DEBUG oslo_vmware.api [None req-c85b0a62-807d-4793-b1be-86b28d9a6745 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] Task: {'id': task-5116093, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.641392] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-037b42a1-7b2a-4493-8518-be35c6d71062 tempest-ServerMetadataNegativeTestJSON-936475568 tempest-ServerMetadataNegativeTestJSON-936475568-project-member] [instance: d179030a-e5ba-45b1-ad11-4d2f71ed7bd4] Unregistered the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 874.641614] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-037b42a1-7b2a-4493-8518-be35c6d71062 tempest-ServerMetadataNegativeTestJSON-936475568 tempest-ServerMetadataNegativeTestJSON-936475568-project-member] [instance: d179030a-e5ba-45b1-ad11-4d2f71ed7bd4] Deleting contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 874.641801] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-037b42a1-7b2a-4493-8518-be35c6d71062 tempest-ServerMetadataNegativeTestJSON-936475568 tempest-ServerMetadataNegativeTestJSON-936475568-project-member] Deleting the datastore file [datastore1] d179030a-e5ba-45b1-ad11-4d2f71ed7bd4 {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 874.645056] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c621805d-96f2-45b7-9bd4-111027e03e8a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.653824] env[65726]: DEBUG oslo_vmware.api [None req-49b35bf1-9410-49fb-9a5b-632c41ecd3a6 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52bc150d-bd34-3b00-edef-340ee1d90a4d, 'name': SearchDatastore_Task, 'duration_secs': 0.010632} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.656374] env[65726]: DEBUG oslo_vmware.api [None req-037b42a1-7b2a-4493-8518-be35c6d71062 tempest-ServerMetadataNegativeTestJSON-936475568 tempest-ServerMetadataNegativeTestJSON-936475568-project-member] Waiting for the task: (returnval){ [ 874.656374] env[65726]: value = "task-5116094" [ 874.656374] env[65726]: _type = "Task" [ 874.656374] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.658015] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4c214c75-b76a-4d9d-b0fd-39229a22ace5 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.670331] env[65726]: DEBUG oslo_vmware.api [None req-49b35bf1-9410-49fb-9a5b-632c41ecd3a6 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Waiting for the task: (returnval){ [ 874.670331] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52f28aeb-1ca2-7c32-1444-d691370101e8" [ 874.670331] env[65726]: _type = "Task" [ 874.670331] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.677159] env[65726]: DEBUG oslo_vmware.api [None req-037b42a1-7b2a-4493-8518-be35c6d71062 tempest-ServerMetadataNegativeTestJSON-936475568 tempest-ServerMetadataNegativeTestJSON-936475568-project-member] Task: {'id': task-5116094, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.680717] env[65726]: DEBUG oslo_vmware.api [None req-8333c704-ca27-426e-acee-8cfd63213196 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Task: {'id': task-5116090, 'name': Rename_Task, 'duration_secs': 0.254184} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.683940] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-8333c704-ca27-426e-acee-8cfd63213196 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: f2d9090c-988f-43f4-9c81-7aa718a3438a] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 874.684865] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8cf4ab29-5215-42aa-bea2-f501dff68de6 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.689843] env[65726]: DEBUG oslo_vmware.api [None req-49b35bf1-9410-49fb-9a5b-632c41ecd3a6 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52f28aeb-1ca2-7c32-1444-d691370101e8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.690867] env[65726]: DEBUG nova.compute.manager [None req-b20cc073-fafc-44fa-bcf9-9893dc98ebec tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04] Start building block device mappings for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 874.698201] env[65726]: DEBUG oslo_vmware.api [None req-8333c704-ca27-426e-acee-8cfd63213196 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Waiting for the task: (returnval){ [ 874.698201] env[65726]: value = "task-5116095" [ 874.698201] env[65726]: _type = "Task" [ 874.698201] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.708294] env[65726]: DEBUG oslo_vmware.api [None req-8333c704-ca27-426e-acee-8cfd63213196 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Task: {'id': task-5116095, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.802913] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48f11d20-62c1-45a8-9755-e2c026ad428c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.811854] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ca70af1-6544-41fb-9fc7-3bd1743c12ed {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.847451] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dde6f586-c4b6-47bb-addc-1f819050280b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.858148] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5182fcd-4fb8-4afd-8f3c-f3e01745a590 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.870973] env[65726]: DEBUG nova.compute.provider_tree [None req-398d7f82-4e73-43c8-8dc0-2fe897bf965f tempest-ServersNegativeTestMultiTenantJSON-944098855 tempest-ServersNegativeTestMultiTenantJSON-944098855-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 874.990745] env[65726]: DEBUG oslo_concurrency.lockutils [req-518cb5bf-8235-4ca9-a600-563380cd1dd4 req-1eff9706-c375-4786-8450-a6677735b12a service nova] Releasing lock "refresh_cache-cf950ec3-9914-4eb9-99db-048b3969bb21" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 874.991290] env[65726]: DEBUG oslo_concurrency.lockutils [None req-73b0a627-5d81-4bf2-833a-caff3534c1e3 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Acquired lock "refresh_cache-cf950ec3-9914-4eb9-99db-048b3969bb21" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 874.991589] env[65726]: DEBUG nova.network.neutron [None req-73b0a627-5d81-4bf2-833a-caff3534c1e3 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: cf950ec3-9914-4eb9-99db-048b3969bb21] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 875.128703] env[65726]: DEBUG oslo_concurrency.lockutils [req-47c21cae-4c82-4bfa-87ef-33c0c45669c3 req-0a024ec1-973d-47cb-9652-2eba67a9a450 service nova] Releasing lock "refresh_cache-c4cc4f85-cb35-4edc-a58b-adfee0ce1265" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 875.128703] env[65726]: DEBUG oslo_vmware.api [None req-c85b0a62-807d-4793-b1be-86b28d9a6745 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] Task: {'id': task-5116093, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.1708} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.128703] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-c85b0a62-807d-4793-b1be-86b28d9a6745 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] Deleted the datastore file {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 875.128703] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-c85b0a62-807d-4793-b1be-86b28d9a6745 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] [instance: 2cbee359-a1fc-45c0-bcc5-a2c24ded9c69] Deleted contents of the VM from datastore datastore2 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 875.128703] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-c85b0a62-807d-4793-b1be-86b28d9a6745 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] [instance: 2cbee359-a1fc-45c0-bcc5-a2c24ded9c69] Instance destroyed {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 875.171106] env[65726]: DEBUG oslo_vmware.api [None req-037b42a1-7b2a-4493-8518-be35c6d71062 tempest-ServerMetadataNegativeTestJSON-936475568 tempest-ServerMetadataNegativeTestJSON-936475568-project-member] Task: {'id': task-5116094, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.321861} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.171479] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-037b42a1-7b2a-4493-8518-be35c6d71062 tempest-ServerMetadataNegativeTestJSON-936475568 tempest-ServerMetadataNegativeTestJSON-936475568-project-member] Deleted the datastore file {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 875.171714] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-037b42a1-7b2a-4493-8518-be35c6d71062 tempest-ServerMetadataNegativeTestJSON-936475568 tempest-ServerMetadataNegativeTestJSON-936475568-project-member] [instance: d179030a-e5ba-45b1-ad11-4d2f71ed7bd4] Deleted contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 875.171941] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-037b42a1-7b2a-4493-8518-be35c6d71062 tempest-ServerMetadataNegativeTestJSON-936475568 tempest-ServerMetadataNegativeTestJSON-936475568-project-member] [instance: d179030a-e5ba-45b1-ad11-4d2f71ed7bd4] Instance destroyed {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 875.172161] env[65726]: INFO nova.compute.manager [None req-037b42a1-7b2a-4493-8518-be35c6d71062 tempest-ServerMetadataNegativeTestJSON-936475568 tempest-ServerMetadataNegativeTestJSON-936475568-project-member] [instance: d179030a-e5ba-45b1-ad11-4d2f71ed7bd4] Took 1.18 seconds to destroy the instance on the hypervisor. [ 875.172454] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-037b42a1-7b2a-4493-8518-be35c6d71062 tempest-ServerMetadataNegativeTestJSON-936475568 tempest-ServerMetadataNegativeTestJSON-936475568-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 875.176597] env[65726]: DEBUG nova.compute.manager [-] [instance: d179030a-e5ba-45b1-ad11-4d2f71ed7bd4] Deallocating network for instance {{(pid=65726) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 875.176749] env[65726]: DEBUG nova.network.neutron [-] [instance: d179030a-e5ba-45b1-ad11-4d2f71ed7bd4] deallocate_for_instance() {{(pid=65726) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 875.177422] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 875.177729] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 875.199173] env[65726]: DEBUG oslo_vmware.api [None req-49b35bf1-9410-49fb-9a5b-632c41ecd3a6 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52f28aeb-1ca2-7c32-1444-d691370101e8, 'name': SearchDatastore_Task, 'duration_secs': 0.019629} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.199650] env[65726]: DEBUG oslo_concurrency.lockutils [None req-49b35bf1-9410-49fb-9a5b-632c41ecd3a6 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 875.199928] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-49b35bf1-9410-49fb-9a5b-632c41ecd3a6 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore2] c4cc4f85-cb35-4edc-a58b-adfee0ce1265/c4cc4f85-cb35-4edc-a58b-adfee0ce1265.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 875.200281] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-45ed4bc0-8ea0-4c4b-b18e-5befca837e61 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.214394] env[65726]: DEBUG oslo_vmware.api [None req-8333c704-ca27-426e-acee-8cfd63213196 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Task: {'id': task-5116095, 'name': PowerOnVM_Task} progress is 88%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.216450] env[65726]: DEBUG oslo_vmware.api [None req-49b35bf1-9410-49fb-9a5b-632c41ecd3a6 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Waiting for the task: (returnval){ [ 875.216450] env[65726]: value = "task-5116096" [ 875.216450] env[65726]: _type = "Task" [ 875.216450] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.227884] env[65726]: DEBUG oslo_vmware.api [None req-49b35bf1-9410-49fb-9a5b-632c41ecd3a6 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Task: {'id': task-5116096, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.231251] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 875.231681] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 875.376033] env[65726]: DEBUG nova.scheduler.client.report [None req-398d7f82-4e73-43c8-8dc0-2fe897bf965f tempest-ServersNegativeTestMultiTenantJSON-944098855 tempest-ServersNegativeTestMultiTenantJSON-944098855-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 875.494835] env[65726]: WARNING openstack [None req-73b0a627-5d81-4bf2-833a-caff3534c1e3 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 875.495278] env[65726]: WARNING openstack [None req-73b0a627-5d81-4bf2-833a-caff3534c1e3 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 875.502214] env[65726]: DEBUG nova.network.neutron [None req-73b0a627-5d81-4bf2-833a-caff3534c1e3 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: cf950ec3-9914-4eb9-99db-048b3969bb21] Instance cache missing network info. {{(pid=65726) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 875.643264] env[65726]: WARNING openstack [None req-73b0a627-5d81-4bf2-833a-caff3534c1e3 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 875.643930] env[65726]: WARNING openstack [None req-73b0a627-5d81-4bf2-833a-caff3534c1e3 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 875.701434] env[65726]: DEBUG nova.compute.manager [None req-b20cc073-fafc-44fa-bcf9-9893dc98ebec tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04] Start spawning the instance on the hypervisor. {{(pid=65726) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 875.715444] env[65726]: DEBUG oslo_vmware.api [None req-8333c704-ca27-426e-acee-8cfd63213196 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Task: {'id': task-5116095, 'name': PowerOnVM_Task, 'duration_secs': 0.59462} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.715737] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-8333c704-ca27-426e-acee-8cfd63213196 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: f2d9090c-988f-43f4-9c81-7aa718a3438a] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 875.715946] env[65726]: DEBUG nova.compute.manager [None req-8333c704-ca27-426e-acee-8cfd63213196 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: f2d9090c-988f-43f4-9c81-7aa718a3438a] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 875.716963] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d9c555c-b313-4870-ba8c-79678000e8d5 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.734752] env[65726]: DEBUG oslo_vmware.api [None req-49b35bf1-9410-49fb-9a5b-632c41ecd3a6 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Task: {'id': task-5116096, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.742126] env[65726]: DEBUG nova.virt.hardware [None req-b20cc073-fafc-44fa-bcf9-9893dc98ebec tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T19:28:34Z,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1539f06fa3534e90acbc3a60c4b8bd3f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T19:28:35Z,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 875.742317] env[65726]: DEBUG nova.virt.hardware [None req-b20cc073-fafc-44fa-bcf9-9893dc98ebec tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 875.742392] env[65726]: DEBUG nova.virt.hardware [None req-b20cc073-fafc-44fa-bcf9-9893dc98ebec tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 875.742599] env[65726]: DEBUG nova.virt.hardware [None req-b20cc073-fafc-44fa-bcf9-9893dc98ebec tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 875.742743] env[65726]: DEBUG nova.virt.hardware [None req-b20cc073-fafc-44fa-bcf9-9893dc98ebec tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 875.742906] env[65726]: DEBUG nova.virt.hardware [None req-b20cc073-fafc-44fa-bcf9-9893dc98ebec tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 875.743180] env[65726]: DEBUG nova.virt.hardware [None req-b20cc073-fafc-44fa-bcf9-9893dc98ebec tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 875.743315] env[65726]: DEBUG nova.virt.hardware [None req-b20cc073-fafc-44fa-bcf9-9893dc98ebec tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 875.743479] env[65726]: DEBUG nova.virt.hardware [None req-b20cc073-fafc-44fa-bcf9-9893dc98ebec tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 875.743652] env[65726]: DEBUG nova.virt.hardware [None req-b20cc073-fafc-44fa-bcf9-9893dc98ebec tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 875.743846] env[65726]: DEBUG nova.virt.hardware [None req-b20cc073-fafc-44fa-bcf9-9893dc98ebec tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 875.744807] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-faf07dde-0bc2-483d-97a2-3052fb08716b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.754568] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86385a10-1f99-402b-a189-cea8e54d5a6f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.834358] env[65726]: WARNING openstack [None req-73b0a627-5d81-4bf2-833a-caff3534c1e3 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 875.834912] env[65726]: WARNING openstack [None req-73b0a627-5d81-4bf2-833a-caff3534c1e3 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 875.882121] env[65726]: DEBUG oslo_concurrency.lockutils [None req-398d7f82-4e73-43c8-8dc0-2fe897bf965f tempest-ServersNegativeTestMultiTenantJSON-944098855 tempest-ServersNegativeTestMultiTenantJSON-944098855-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.205s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 875.884816] env[65726]: DEBUG oslo_concurrency.lockutils [None req-d6fa7de0-71fb-41d9-ba04-00cd57a3d822 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 39.747s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 875.894420] env[65726]: DEBUG nova.compute.manager [req-93155672-087e-461b-a80d-332a32d8894e req-1899b31a-efa6-459a-85c8-a3282ebf443c service nova] [instance: d179030a-e5ba-45b1-ad11-4d2f71ed7bd4] Received event network-vif-deleted-51e75557-8c53-4ba4-982a-42e90b5decef {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 875.894630] env[65726]: INFO nova.compute.manager [req-93155672-087e-461b-a80d-332a32d8894e req-1899b31a-efa6-459a-85c8-a3282ebf443c service nova] [instance: d179030a-e5ba-45b1-ad11-4d2f71ed7bd4] Neutron deleted interface 51e75557-8c53-4ba4-982a-42e90b5decef; detaching it from the instance and deleting it from the info cache [ 875.894775] env[65726]: DEBUG nova.network.neutron [req-93155672-087e-461b-a80d-332a32d8894e req-1899b31a-efa6-459a-85c8-a3282ebf443c service nova] [instance: d179030a-e5ba-45b1-ad11-4d2f71ed7bd4] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 875.922843] env[65726]: INFO nova.scheduler.client.report [None req-398d7f82-4e73-43c8-8dc0-2fe897bf965f tempest-ServersNegativeTestMultiTenantJSON-944098855 tempest-ServersNegativeTestMultiTenantJSON-944098855-project-member] Deleted allocations for instance 018c8ac5-66c4-4a9a-ab45-85e84f92b4fe [ 875.982794] env[65726]: DEBUG nova.network.neutron [None req-73b0a627-5d81-4bf2-833a-caff3534c1e3 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: cf950ec3-9914-4eb9-99db-048b3969bb21] Updating instance_info_cache with network_info: [{"id": "58accf41-587f-4367-854e-fc571a6a3424", "address": "fa:16:3e:83:9d:0c", "network": {"id": "9d5ad167-99fb-4208-ab30-7053f35774e4", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-2082516762-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a8362295a62f4b51bae719a7ef5a4656", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dabbac20-1723-40ad-9da0-e53b28073651", "external-id": "nsx-vlan-transportzone-790", "segmentation_id": 790, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap58accf41-58", "ovs_interfaceid": "58accf41-587f-4367-854e-fc571a6a3424", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 876.036936] env[65726]: DEBUG nova.network.neutron [-] [instance: d179030a-e5ba-45b1-ad11-4d2f71ed7bd4] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 876.184779] env[65726]: DEBUG nova.virt.hardware [None req-c85b0a62-807d-4793-b1be-86b28d9a6745 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T19:28:34Z,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1539f06fa3534e90acbc3a60c4b8bd3f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T19:28:35Z,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 876.185483] env[65726]: DEBUG nova.virt.hardware [None req-c85b0a62-807d-4793-b1be-86b28d9a6745 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 876.185483] env[65726]: DEBUG nova.virt.hardware [None req-c85b0a62-807d-4793-b1be-86b28d9a6745 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 876.185483] env[65726]: DEBUG nova.virt.hardware [None req-c85b0a62-807d-4793-b1be-86b28d9a6745 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 876.185648] env[65726]: DEBUG nova.virt.hardware [None req-c85b0a62-807d-4793-b1be-86b28d9a6745 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 876.185709] env[65726]: DEBUG nova.virt.hardware [None req-c85b0a62-807d-4793-b1be-86b28d9a6745 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 876.185925] env[65726]: DEBUG nova.virt.hardware [None req-c85b0a62-807d-4793-b1be-86b28d9a6745 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 876.186107] env[65726]: DEBUG nova.virt.hardware [None req-c85b0a62-807d-4793-b1be-86b28d9a6745 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 876.186277] env[65726]: DEBUG nova.virt.hardware [None req-c85b0a62-807d-4793-b1be-86b28d9a6745 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 876.186451] env[65726]: DEBUG nova.virt.hardware [None req-c85b0a62-807d-4793-b1be-86b28d9a6745 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 876.186628] env[65726]: DEBUG nova.virt.hardware [None req-c85b0a62-807d-4793-b1be-86b28d9a6745 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 876.188217] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6df82b3a-6bf0-4b15-b321-b0b4c4fb7e06 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.197442] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-089ad2d1-d8cc-4e4b-8656-fc7892b4e1a9 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.213143] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-c85b0a62-807d-4793-b1be-86b28d9a6745 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] [instance: 2cbee359-a1fc-45c0-bcc5-a2c24ded9c69] Instance VIF info [] {{(pid=65726) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 876.220991] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-c85b0a62-807d-4793-b1be-86b28d9a6745 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 876.223164] env[65726]: DEBUG nova.network.neutron [None req-b20cc073-fafc-44fa-bcf9-9893dc98ebec tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04] Successfully updated port: 25b9d108-0b43-4459-b9db-7bd90a495bb3 {{(pid=65726) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 876.226449] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2cbee359-a1fc-45c0-bcc5-a2c24ded9c69] Creating VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 876.226907] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-193a1f41-093a-4e08-ae2e-d8013970b16f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.256666] env[65726]: DEBUG oslo_concurrency.lockutils [None req-8333c704-ca27-426e-acee-8cfd63213196 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 876.266297] env[65726]: DEBUG oslo_vmware.api [None req-49b35bf1-9410-49fb-9a5b-632c41ecd3a6 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Task: {'id': task-5116096, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.553718} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.268418] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-49b35bf1-9410-49fb-9a5b-632c41ecd3a6 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore2] c4cc4f85-cb35-4edc-a58b-adfee0ce1265/c4cc4f85-cb35-4edc-a58b-adfee0ce1265.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 876.268720] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-49b35bf1-9410-49fb-9a5b-632c41ecd3a6 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] [instance: c4cc4f85-cb35-4edc-a58b-adfee0ce1265] Extending root virtual disk to 1048576 {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 876.269021] env[65726]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 876.269021] env[65726]: value = "task-5116097" [ 876.269021] env[65726]: _type = "Task" [ 876.269021] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.269391] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ad2b8cb1-b0f5-426a-a9d3-500966cee757 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.283756] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116097, 'name': CreateVM_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.285434] env[65726]: DEBUG oslo_vmware.api [None req-49b35bf1-9410-49fb-9a5b-632c41ecd3a6 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Waiting for the task: (returnval){ [ 876.285434] env[65726]: value = "task-5116098" [ 876.285434] env[65726]: _type = "Task" [ 876.285434] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.296756] env[65726]: DEBUG oslo_vmware.api [None req-49b35bf1-9410-49fb-9a5b-632c41ecd3a6 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Task: {'id': task-5116098, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.390739] env[65726]: INFO nova.compute.claims [None req-d6fa7de0-71fb-41d9-ba04-00cd57a3d822 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 608cc0ea-3f6b-4b4f-83c1-01aa50999d63] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 876.397758] env[65726]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-81c2dc43-b711-4b88-96e2-f6f284a81091 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.414064] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7732dff2-f258-4f7d-8040-730525facba1 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.432248] env[65726]: DEBUG nova.compute.manager [req-e21da23f-1a83-48d9-a986-ffd6db448f8c req-ba15d2c1-7809-4314-9951-e448627053e0 service nova] [instance: 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04] Received event network-vif-plugged-25b9d108-0b43-4459-b9db-7bd90a495bb3 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 876.432780] env[65726]: DEBUG oslo_concurrency.lockutils [req-e21da23f-1a83-48d9-a986-ffd6db448f8c req-ba15d2c1-7809-4314-9951-e448627053e0 service nova] Acquiring lock "2fe6fc5a-f5c7-4f8e-96df-4e621a252f04-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 876.432780] env[65726]: DEBUG oslo_concurrency.lockutils [req-e21da23f-1a83-48d9-a986-ffd6db448f8c req-ba15d2c1-7809-4314-9951-e448627053e0 service nova] Lock "2fe6fc5a-f5c7-4f8e-96df-4e621a252f04-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 876.432954] env[65726]: DEBUG oslo_concurrency.lockutils [req-e21da23f-1a83-48d9-a986-ffd6db448f8c req-ba15d2c1-7809-4314-9951-e448627053e0 service nova] Lock "2fe6fc5a-f5c7-4f8e-96df-4e621a252f04-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 876.433058] env[65726]: DEBUG nova.compute.manager [req-e21da23f-1a83-48d9-a986-ffd6db448f8c req-ba15d2c1-7809-4314-9951-e448627053e0 service nova] [instance: 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04] No waiting events found dispatching network-vif-plugged-25b9d108-0b43-4459-b9db-7bd90a495bb3 {{(pid=65726) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 876.433326] env[65726]: WARNING nova.compute.manager [req-e21da23f-1a83-48d9-a986-ffd6db448f8c req-ba15d2c1-7809-4314-9951-e448627053e0 service nova] [instance: 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04] Received unexpected event network-vif-plugged-25b9d108-0b43-4459-b9db-7bd90a495bb3 for instance with vm_state building and task_state spawning. [ 876.436836] env[65726]: DEBUG oslo_concurrency.lockutils [None req-398d7f82-4e73-43c8-8dc0-2fe897bf965f tempest-ServersNegativeTestMultiTenantJSON-944098855 tempest-ServersNegativeTestMultiTenantJSON-944098855-project-member] Lock "018c8ac5-66c4-4a9a-ab45-85e84f92b4fe" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 46.146s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 876.454844] env[65726]: DEBUG nova.compute.manager [req-93155672-087e-461b-a80d-332a32d8894e req-1899b31a-efa6-459a-85c8-a3282ebf443c service nova] [instance: d179030a-e5ba-45b1-ad11-4d2f71ed7bd4] Detach interface failed, port_id=51e75557-8c53-4ba4-982a-42e90b5decef, reason: Instance d179030a-e5ba-45b1-ad11-4d2f71ed7bd4 could not be found. {{(pid=65726) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11669}} [ 876.485648] env[65726]: DEBUG oslo_concurrency.lockutils [None req-73b0a627-5d81-4bf2-833a-caff3534c1e3 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Releasing lock "refresh_cache-cf950ec3-9914-4eb9-99db-048b3969bb21" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 876.486218] env[65726]: DEBUG nova.compute.manager [None req-73b0a627-5d81-4bf2-833a-caff3534c1e3 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: cf950ec3-9914-4eb9-99db-048b3969bb21] Instance network_info: |[{"id": "58accf41-587f-4367-854e-fc571a6a3424", "address": "fa:16:3e:83:9d:0c", "network": {"id": "9d5ad167-99fb-4208-ab30-7053f35774e4", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-2082516762-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a8362295a62f4b51bae719a7ef5a4656", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dabbac20-1723-40ad-9da0-e53b28073651", "external-id": "nsx-vlan-transportzone-790", "segmentation_id": 790, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap58accf41-58", "ovs_interfaceid": "58accf41-587f-4367-854e-fc571a6a3424", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 876.486997] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-73b0a627-5d81-4bf2-833a-caff3534c1e3 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: cf950ec3-9914-4eb9-99db-048b3969bb21] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:83:9d:0c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'dabbac20-1723-40ad-9da0-e53b28073651', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '58accf41-587f-4367-854e-fc571a6a3424', 'vif_model': 'vmxnet3'}] {{(pid=65726) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 876.496432] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-73b0a627-5d81-4bf2-833a-caff3534c1e3 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 876.497038] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: cf950ec3-9914-4eb9-99db-048b3969bb21] Creating VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 876.497309] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-bdf39c89-60d9-4cc1-8a45-2a21ffa7112c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.520624] env[65726]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 876.520624] env[65726]: value = "task-5116099" [ 876.520624] env[65726]: _type = "Task" [ 876.520624] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.532497] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116099, 'name': CreateVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.540151] env[65726]: INFO nova.compute.manager [-] [instance: d179030a-e5ba-45b1-ad11-4d2f71ed7bd4] Took 1.36 seconds to deallocate network for instance. [ 876.726190] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b20cc073-fafc-44fa-bcf9-9893dc98ebec tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Acquiring lock "refresh_cache-2fe6fc5a-f5c7-4f8e-96df-4e621a252f04" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 876.726456] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b20cc073-fafc-44fa-bcf9-9893dc98ebec tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Acquired lock "refresh_cache-2fe6fc5a-f5c7-4f8e-96df-4e621a252f04" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 876.726756] env[65726]: DEBUG nova.network.neutron [None req-b20cc073-fafc-44fa-bcf9-9893dc98ebec tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 876.784293] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116097, 'name': CreateVM_Task, 'duration_secs': 0.349943} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.784527] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2cbee359-a1fc-45c0-bcc5-a2c24ded9c69] Created VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 876.785041] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c85b0a62-807d-4793-b1be-86b28d9a6745 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 876.785199] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c85b0a62-807d-4793-b1be-86b28d9a6745 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 876.785667] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c85b0a62-807d-4793-b1be-86b28d9a6745 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 876.786091] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-79852f73-7466-4b1e-b498-b4d61c324b12 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.801893] env[65726]: DEBUG oslo_vmware.api [None req-c85b0a62-807d-4793-b1be-86b28d9a6745 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] Waiting for the task: (returnval){ [ 876.801893] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]527c327c-f0e2-cef3-327e-0e3838b2cdc0" [ 876.801893] env[65726]: _type = "Task" [ 876.801893] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.802286] env[65726]: DEBUG oslo_vmware.api [None req-49b35bf1-9410-49fb-9a5b-632c41ecd3a6 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Task: {'id': task-5116098, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.118435} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.802796] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-49b35bf1-9410-49fb-9a5b-632c41ecd3a6 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] [instance: c4cc4f85-cb35-4edc-a58b-adfee0ce1265] Extended root virtual disk {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 876.811648] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9fbf0d48-ac92-4856-8285-cd9dab5f8e34 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.838606] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-49b35bf1-9410-49fb-9a5b-632c41ecd3a6 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] [instance: c4cc4f85-cb35-4edc-a58b-adfee0ce1265] Reconfiguring VM instance instance-0000003d to attach disk [datastore2] c4cc4f85-cb35-4edc-a58b-adfee0ce1265/c4cc4f85-cb35-4edc-a58b-adfee0ce1265.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 876.842611] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c89ed1a4-4393-4c12-a18c-eba82e114420 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.857775] env[65726]: DEBUG oslo_vmware.api [None req-c85b0a62-807d-4793-b1be-86b28d9a6745 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]527c327c-f0e2-cef3-327e-0e3838b2cdc0, 'name': SearchDatastore_Task, 'duration_secs': 0.018679} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.858121] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c85b0a62-807d-4793-b1be-86b28d9a6745 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 876.858367] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-c85b0a62-807d-4793-b1be-86b28d9a6745 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] [instance: 2cbee359-a1fc-45c0-bcc5-a2c24ded9c69] Processing image b52362a3-ee8a-4cbf-b06f-513b0cc8f95c {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 876.858763] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c85b0a62-807d-4793-b1be-86b28d9a6745 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 876.858763] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c85b0a62-807d-4793-b1be-86b28d9a6745 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 876.858906] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-c85b0a62-807d-4793-b1be-86b28d9a6745 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 876.859662] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2eb80683-d795-4b99-aa4c-852d60cf84b0 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.720249] env[65726]: INFO nova.compute.resource_tracker [None req-d6fa7de0-71fb-41d9-ba04-00cd57a3d822 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 608cc0ea-3f6b-4b4f-83c1-01aa50999d63] Updating resource usage from migration 3a215dd8-f815-44fc-8620-fb9bc12739f7 [ 877.724281] env[65726]: DEBUG oslo_concurrency.lockutils [None req-037b42a1-7b2a-4493-8518-be35c6d71062 tempest-ServerMetadataNegativeTestJSON-936475568 tempest-ServerMetadataNegativeTestJSON-936475568-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 877.725094] env[65726]: WARNING openstack [None req-b20cc073-fafc-44fa-bcf9-9893dc98ebec tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 877.725436] env[65726]: WARNING openstack [None req-b20cc073-fafc-44fa-bcf9-9893dc98ebec tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 877.730532] env[65726]: DEBUG nova.network.neutron [None req-b20cc073-fafc-44fa-bcf9-9893dc98ebec tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04] Instance cache missing network info. {{(pid=65726) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 877.738257] env[65726]: DEBUG oslo_vmware.api [None req-49b35bf1-9410-49fb-9a5b-632c41ecd3a6 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Waiting for the task: (returnval){ [ 877.738257] env[65726]: value = "task-5116100" [ 877.738257] env[65726]: _type = "Task" [ 877.738257] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.748101] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-c85b0a62-807d-4793-b1be-86b28d9a6745 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 877.748101] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-c85b0a62-807d-4793-b1be-86b28d9a6745 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65726) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 877.749073] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-39abbbbb-9c5d-48a1-b2e6-7dd3b2eb02c3 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.760210] env[65726]: DEBUG oslo_vmware.api [None req-49b35bf1-9410-49fb-9a5b-632c41ecd3a6 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Task: {'id': task-5116100, 'name': ReconfigVM_Task, 'duration_secs': 0.358702} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.760447] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116099, 'name': CreateVM_Task, 'duration_secs': 0.456348} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.761839] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-49b35bf1-9410-49fb-9a5b-632c41ecd3a6 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] [instance: c4cc4f85-cb35-4edc-a58b-adfee0ce1265] Reconfigured VM instance instance-0000003d to attach disk [datastore2] c4cc4f85-cb35-4edc-a58b-adfee0ce1265/c4cc4f85-cb35-4edc-a58b-adfee0ce1265.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 877.762463] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: cf950ec3-9914-4eb9-99db-048b3969bb21] Created VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 877.763014] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2613787e-c933-42f4-8bd3-029628b694a3 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.765565] env[65726]: WARNING openstack [None req-73b0a627-5d81-4bf2-833a-caff3534c1e3 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 877.765909] env[65726]: WARNING openstack [None req-73b0a627-5d81-4bf2-833a-caff3534c1e3 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 877.771146] env[65726]: DEBUG oslo_concurrency.lockutils [None req-73b0a627-5d81-4bf2-833a-caff3534c1e3 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 877.771392] env[65726]: DEBUG oslo_concurrency.lockutils [None req-73b0a627-5d81-4bf2-833a-caff3534c1e3 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 877.772902] env[65726]: DEBUG oslo_concurrency.lockutils [None req-73b0a627-5d81-4bf2-833a-caff3534c1e3 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 877.776215] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e57f93d0-1e56-4ba9-9ee9-b0bfbfb2b12c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.778306] env[65726]: DEBUG oslo_vmware.api [None req-c85b0a62-807d-4793-b1be-86b28d9a6745 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] Waiting for the task: (returnval){ [ 877.778306] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]527585c3-9d40-278b-8ce1-ac7bda29c586" [ 877.778306] env[65726]: _type = "Task" [ 877.778306] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.785304] env[65726]: DEBUG oslo_vmware.api [None req-73b0a627-5d81-4bf2-833a-caff3534c1e3 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Waiting for the task: (returnval){ [ 877.785304] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]529335d7-4080-58bc-416b-3861ba7fba4c" [ 877.785304] env[65726]: _type = "Task" [ 877.785304] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.787080] env[65726]: DEBUG oslo_vmware.api [None req-49b35bf1-9410-49fb-9a5b-632c41ecd3a6 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Waiting for the task: (returnval){ [ 877.787080] env[65726]: value = "task-5116101" [ 877.787080] env[65726]: _type = "Task" [ 877.787080] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.808671] env[65726]: DEBUG oslo_vmware.api [None req-c85b0a62-807d-4793-b1be-86b28d9a6745 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]527585c3-9d40-278b-8ce1-ac7bda29c586, 'name': SearchDatastore_Task, 'duration_secs': 0.01905} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.810528] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-42b7bad2-b37c-4139-b1d8-29e6b8f9477e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.817453] env[65726]: DEBUG oslo_vmware.api [None req-73b0a627-5d81-4bf2-833a-caff3534c1e3 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]529335d7-4080-58bc-416b-3861ba7fba4c, 'name': SearchDatastore_Task, 'duration_secs': 0.010691} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.821491] env[65726]: DEBUG oslo_concurrency.lockutils [None req-73b0a627-5d81-4bf2-833a-caff3534c1e3 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 877.821791] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-73b0a627-5d81-4bf2-833a-caff3534c1e3 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: cf950ec3-9914-4eb9-99db-048b3969bb21] Processing image b52362a3-ee8a-4cbf-b06f-513b0cc8f95c {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 877.822038] env[65726]: DEBUG oslo_concurrency.lockutils [None req-73b0a627-5d81-4bf2-833a-caff3534c1e3 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 877.822624] env[65726]: DEBUG oslo_concurrency.lockutils [None req-73b0a627-5d81-4bf2-833a-caff3534c1e3 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 877.822624] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-73b0a627-5d81-4bf2-833a-caff3534c1e3 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 877.823099] env[65726]: DEBUG oslo_vmware.api [None req-49b35bf1-9410-49fb-9a5b-632c41ecd3a6 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Task: {'id': task-5116101, 'name': Rename_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.824684] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c1dbbab0-e4b6-4ede-9d6d-b7209f525ed4 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.826873] env[65726]: DEBUG oslo_vmware.api [None req-c85b0a62-807d-4793-b1be-86b28d9a6745 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] Waiting for the task: (returnval){ [ 877.826873] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]526f55ca-983a-0cdf-ec99-55c1ad08262a" [ 877.826873] env[65726]: _type = "Task" [ 877.826873] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.835800] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-73b0a627-5d81-4bf2-833a-caff3534c1e3 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 877.835800] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-73b0a627-5d81-4bf2-833a-caff3534c1e3 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65726) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 877.840298] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3de36ffa-497e-4e30-b8b3-39ae428785cb {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.843706] env[65726]: DEBUG oslo_vmware.api [None req-c85b0a62-807d-4793-b1be-86b28d9a6745 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]526f55ca-983a-0cdf-ec99-55c1ad08262a, 'name': SearchDatastore_Task, 'duration_secs': 0.011986} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.846218] env[65726]: WARNING openstack [None req-b20cc073-fafc-44fa-bcf9-9893dc98ebec tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 877.846218] env[65726]: WARNING openstack [None req-b20cc073-fafc-44fa-bcf9-9893dc98ebec tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 877.852391] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c85b0a62-807d-4793-b1be-86b28d9a6745 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 877.852391] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-c85b0a62-807d-4793-b1be-86b28d9a6745 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore2] 2cbee359-a1fc-45c0-bcc5-a2c24ded9c69/2cbee359-a1fc-45c0-bcc5-a2c24ded9c69.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 877.855739] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7073dc33-d1c3-4982-ab42-2471e862e508 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.861411] env[65726]: DEBUG oslo_vmware.api [None req-73b0a627-5d81-4bf2-833a-caff3534c1e3 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Waiting for the task: (returnval){ [ 877.861411] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52322503-7747-97ea-8e5d-340736f45858" [ 877.861411] env[65726]: _type = "Task" [ 877.861411] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.867569] env[65726]: DEBUG oslo_vmware.api [None req-c85b0a62-807d-4793-b1be-86b28d9a6745 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] Waiting for the task: (returnval){ [ 877.867569] env[65726]: value = "task-5116102" [ 877.867569] env[65726]: _type = "Task" [ 877.867569] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.871254] env[65726]: DEBUG oslo_vmware.api [None req-73b0a627-5d81-4bf2-833a-caff3534c1e3 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52322503-7747-97ea-8e5d-340736f45858, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.885654] env[65726]: DEBUG oslo_vmware.api [None req-c85b0a62-807d-4793-b1be-86b28d9a6745 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] Task: {'id': task-5116102, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.955748] env[65726]: WARNING openstack [None req-b20cc073-fafc-44fa-bcf9-9893dc98ebec tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 877.956178] env[65726]: WARNING openstack [None req-b20cc073-fafc-44fa-bcf9-9893dc98ebec tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 878.011691] env[65726]: DEBUG oslo_concurrency.lockutils [None req-172c2bf1-c043-4217-8650-efc42870269c tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Acquiring lock "895797e4-2941-44cd-aab1-67afa4fac02d" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 878.012140] env[65726]: DEBUG oslo_concurrency.lockutils [None req-172c2bf1-c043-4217-8650-efc42870269c tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Lock "895797e4-2941-44cd-aab1-67afa4fac02d" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.001s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 878.080884] env[65726]: DEBUG nova.network.neutron [None req-b20cc073-fafc-44fa-bcf9-9893dc98ebec tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04] Updating instance_info_cache with network_info: [{"id": "25b9d108-0b43-4459-b9db-7bd90a495bb3", "address": "fa:16:3e:5a:1e:6b", "network": {"id": "eb3278b7-c190-441b-84d2-305f9b896382", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1754571022-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "96149159e18e44f9bf3453e67681f224", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f77ff7a1-209c-4f3f-b2a0-fd817741e739", "external-id": "nsx-vlan-transportzone-935", "segmentation_id": 935, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap25b9d108-0b", "ovs_interfaceid": "25b9d108-0b43-4459-b9db-7bd90a495bb3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 878.238608] env[65726]: INFO nova.compute.manager [None req-5782ba22-8056-4347-9a97-4b1e5a35ed3b tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: f2d9090c-988f-43f4-9c81-7aa718a3438a] Rebuilding instance [ 878.309114] env[65726]: DEBUG nova.compute.manager [None req-5782ba22-8056-4347-9a97-4b1e5a35ed3b tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: f2d9090c-988f-43f4-9c81-7aa718a3438a] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 878.310382] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d951c094-4388-4b94-9f8f-3fd473465f79 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.317333] env[65726]: DEBUG oslo_vmware.api [None req-49b35bf1-9410-49fb-9a5b-632c41ecd3a6 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Task: {'id': task-5116101, 'name': Rename_Task, 'duration_secs': 0.174837} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.318336] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-49b35bf1-9410-49fb-9a5b-632c41ecd3a6 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] [instance: c4cc4f85-cb35-4edc-a58b-adfee0ce1265] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 878.318816] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a2326c6a-1fab-43bc-a335-298fc0c7bb9e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.339028] env[65726]: DEBUG oslo_vmware.api [None req-49b35bf1-9410-49fb-9a5b-632c41ecd3a6 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Waiting for the task: (returnval){ [ 878.339028] env[65726]: value = "task-5116103" [ 878.339028] env[65726]: _type = "Task" [ 878.339028] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.351223] env[65726]: DEBUG oslo_vmware.api [None req-49b35bf1-9410-49fb-9a5b-632c41ecd3a6 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Task: {'id': task-5116103, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.381784] env[65726]: DEBUG oslo_vmware.api [None req-73b0a627-5d81-4bf2-833a-caff3534c1e3 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52322503-7747-97ea-8e5d-340736f45858, 'name': SearchDatastore_Task, 'duration_secs': 0.011653} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.386952] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cdfbb93e-ed07-436c-a108-ae189735e820 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.396522] env[65726]: DEBUG oslo_vmware.api [None req-c85b0a62-807d-4793-b1be-86b28d9a6745 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] Task: {'id': task-5116102, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.404482] env[65726]: DEBUG oslo_vmware.api [None req-73b0a627-5d81-4bf2-833a-caff3534c1e3 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Waiting for the task: (returnval){ [ 878.404482] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]527d987f-569e-ad9e-bdae-74008d4228f8" [ 878.404482] env[65726]: _type = "Task" [ 878.404482] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.420728] env[65726]: DEBUG oslo_vmware.api [None req-73b0a627-5d81-4bf2-833a-caff3534c1e3 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]527d987f-569e-ad9e-bdae-74008d4228f8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.425805] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8cc43f6e-ac75-4b37-a990-172a139760a2 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.435434] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c18d0ca-352c-4280-b8cb-e468af3fe5c7 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.473840] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5796aed9-f2c5-4e2f-972a-680e1d07cf57 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.482679] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ebffd87-b30f-45d0-8942-be3dee4f0016 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.500893] env[65726]: DEBUG nova.compute.provider_tree [None req-d6fa7de0-71fb-41d9-ba04-00cd57a3d822 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 878.519303] env[65726]: DEBUG nova.compute.utils [None req-172c2bf1-c043-4217-8650-efc42870269c tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Using /dev/sd instead of None {{(pid=65726) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 878.584764] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b20cc073-fafc-44fa-bcf9-9893dc98ebec tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Releasing lock "refresh_cache-2fe6fc5a-f5c7-4f8e-96df-4e621a252f04" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 878.585624] env[65726]: DEBUG nova.compute.manager [None req-b20cc073-fafc-44fa-bcf9-9893dc98ebec tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04] Instance network_info: |[{"id": "25b9d108-0b43-4459-b9db-7bd90a495bb3", "address": "fa:16:3e:5a:1e:6b", "network": {"id": "eb3278b7-c190-441b-84d2-305f9b896382", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1754571022-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "96149159e18e44f9bf3453e67681f224", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f77ff7a1-209c-4f3f-b2a0-fd817741e739", "external-id": "nsx-vlan-transportzone-935", "segmentation_id": 935, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap25b9d108-0b", "ovs_interfaceid": "25b9d108-0b43-4459-b9db-7bd90a495bb3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 878.586202] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-b20cc073-fafc-44fa-bcf9-9893dc98ebec tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:5a:1e:6b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f77ff7a1-209c-4f3f-b2a0-fd817741e739', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '25b9d108-0b43-4459-b9db-7bd90a495bb3', 'vif_model': 'vmxnet3'}] {{(pid=65726) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 878.595636] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-b20cc073-fafc-44fa-bcf9-9893dc98ebec tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Creating folder: Project (96149159e18e44f9bf3453e67681f224). Parent ref: group-v995008. {{(pid=65726) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 878.596062] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b0b92c50-39e9-4ca0-8a1f-135c5565c92f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.608535] env[65726]: INFO nova.virt.vmwareapi.vm_util [None req-b20cc073-fafc-44fa-bcf9-9893dc98ebec tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Created folder: Project (96149159e18e44f9bf3453e67681f224) in parent group-v995008. [ 878.608796] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-b20cc073-fafc-44fa-bcf9-9893dc98ebec tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Creating folder: Instances. Parent ref: group-v995181. {{(pid=65726) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 878.609262] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a0603fcd-4b03-4da7-9a9e-fe8c9ac573c4 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.620912] env[65726]: INFO nova.virt.vmwareapi.vm_util [None req-b20cc073-fafc-44fa-bcf9-9893dc98ebec tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Created folder: Instances in parent group-v995181. [ 878.621358] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-b20cc073-fafc-44fa-bcf9-9893dc98ebec tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 878.621583] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04] Creating VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 878.621831] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b38e08de-eaaa-415a-9c96-65f1c3f33efe {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.642553] env[65726]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 878.642553] env[65726]: value = "task-5116106" [ 878.642553] env[65726]: _type = "Task" [ 878.642553] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.652074] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116106, 'name': CreateVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.748344] env[65726]: DEBUG nova.compute.manager [req-da42b144-72ca-4f16-b99f-64a0613e87ae req-d325fa89-eede-4b44-9dc4-b6fc26d574ad service nova] [instance: 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04] Received event network-changed-25b9d108-0b43-4459-b9db-7bd90a495bb3 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 878.748608] env[65726]: DEBUG nova.compute.manager [req-da42b144-72ca-4f16-b99f-64a0613e87ae req-d325fa89-eede-4b44-9dc4-b6fc26d574ad service nova] [instance: 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04] Refreshing instance network info cache due to event network-changed-25b9d108-0b43-4459-b9db-7bd90a495bb3. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 878.748780] env[65726]: DEBUG oslo_concurrency.lockutils [req-da42b144-72ca-4f16-b99f-64a0613e87ae req-d325fa89-eede-4b44-9dc4-b6fc26d574ad service nova] Acquiring lock "refresh_cache-2fe6fc5a-f5c7-4f8e-96df-4e621a252f04" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 878.748920] env[65726]: DEBUG oslo_concurrency.lockutils [req-da42b144-72ca-4f16-b99f-64a0613e87ae req-d325fa89-eede-4b44-9dc4-b6fc26d574ad service nova] Acquired lock "refresh_cache-2fe6fc5a-f5c7-4f8e-96df-4e621a252f04" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 878.749068] env[65726]: DEBUG nova.network.neutron [req-da42b144-72ca-4f16-b99f-64a0613e87ae req-d325fa89-eede-4b44-9dc4-b6fc26d574ad service nova] [instance: 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04] Refreshing network info cache for port 25b9d108-0b43-4459-b9db-7bd90a495bb3 {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 878.847886] env[65726]: DEBUG oslo_vmware.api [None req-49b35bf1-9410-49fb-9a5b-632c41ecd3a6 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Task: {'id': task-5116103, 'name': PowerOnVM_Task} progress is 90%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.883375] env[65726]: DEBUG oslo_vmware.api [None req-c85b0a62-807d-4793-b1be-86b28d9a6745 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] Task: {'id': task-5116102, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.569165} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.883557] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-c85b0a62-807d-4793-b1be-86b28d9a6745 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore2] 2cbee359-a1fc-45c0-bcc5-a2c24ded9c69/2cbee359-a1fc-45c0-bcc5-a2c24ded9c69.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 878.883820] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-c85b0a62-807d-4793-b1be-86b28d9a6745 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] [instance: 2cbee359-a1fc-45c0-bcc5-a2c24ded9c69] Extending root virtual disk to 1048576 {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 878.884126] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d657b7d6-37c3-4c34-a2ed-d355b42560e2 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.892671] env[65726]: DEBUG oslo_vmware.api [None req-c85b0a62-807d-4793-b1be-86b28d9a6745 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] Waiting for the task: (returnval){ [ 878.892671] env[65726]: value = "task-5116107" [ 878.892671] env[65726]: _type = "Task" [ 878.892671] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.903104] env[65726]: DEBUG oslo_vmware.api [None req-c85b0a62-807d-4793-b1be-86b28d9a6745 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] Task: {'id': task-5116107, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.917033] env[65726]: DEBUG oslo_vmware.api [None req-73b0a627-5d81-4bf2-833a-caff3534c1e3 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]527d987f-569e-ad9e-bdae-74008d4228f8, 'name': SearchDatastore_Task, 'duration_secs': 0.023828} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.917033] env[65726]: DEBUG oslo_concurrency.lockutils [None req-73b0a627-5d81-4bf2-833a-caff3534c1e3 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 878.917033] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-73b0a627-5d81-4bf2-833a-caff3534c1e3 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] cf950ec3-9914-4eb9-99db-048b3969bb21/cf950ec3-9914-4eb9-99db-048b3969bb21.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 878.917033] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6545320b-e22f-4b64-b8a4-e9adc3815303 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.924697] env[65726]: DEBUG oslo_vmware.api [None req-73b0a627-5d81-4bf2-833a-caff3534c1e3 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Waiting for the task: (returnval){ [ 878.924697] env[65726]: value = "task-5116108" [ 878.924697] env[65726]: _type = "Task" [ 878.924697] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.934337] env[65726]: DEBUG oslo_vmware.api [None req-73b0a627-5d81-4bf2-833a-caff3534c1e3 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Task: {'id': task-5116108, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.004746] env[65726]: DEBUG nova.scheduler.client.report [None req-d6fa7de0-71fb-41d9-ba04-00cd57a3d822 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 879.024583] env[65726]: DEBUG oslo_concurrency.lockutils [None req-172c2bf1-c043-4217-8650-efc42870269c tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Lock "895797e4-2941-44cd-aab1-67afa4fac02d" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.010s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 879.158489] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116106, 'name': CreateVM_Task, 'duration_secs': 0.478418} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.158786] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04] Created VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 879.159761] env[65726]: WARNING openstack [None req-b20cc073-fafc-44fa-bcf9-9893dc98ebec tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 879.160111] env[65726]: WARNING openstack [None req-b20cc073-fafc-44fa-bcf9-9893dc98ebec tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 879.167046] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b20cc073-fafc-44fa-bcf9-9893dc98ebec tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 879.167275] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b20cc073-fafc-44fa-bcf9-9893dc98ebec tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 879.167722] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b20cc073-fafc-44fa-bcf9-9893dc98ebec tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 879.168120] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e7547181-92e9-4deb-80c5-9b7b085ae2bd {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.178783] env[65726]: DEBUG oslo_vmware.api [None req-b20cc073-fafc-44fa-bcf9-9893dc98ebec tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Waiting for the task: (returnval){ [ 879.178783] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52d3b37e-be70-4743-1fa5-9f65378076bd" [ 879.178783] env[65726]: _type = "Task" [ 879.178783] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.194513] env[65726]: DEBUG oslo_vmware.api [None req-b20cc073-fafc-44fa-bcf9-9893dc98ebec tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52d3b37e-be70-4743-1fa5-9f65378076bd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.252436] env[65726]: WARNING openstack [req-da42b144-72ca-4f16-b99f-64a0613e87ae req-d325fa89-eede-4b44-9dc4-b6fc26d574ad service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 879.252610] env[65726]: WARNING openstack [req-da42b144-72ca-4f16-b99f-64a0613e87ae req-d325fa89-eede-4b44-9dc4-b6fc26d574ad service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 879.338153] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-5782ba22-8056-4347-9a97-4b1e5a35ed3b tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: f2d9090c-988f-43f4-9c81-7aa718a3438a] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 879.339101] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4c23e4b3-e61d-4135-86da-e5f262f7e006 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.351767] env[65726]: DEBUG oslo_vmware.api [None req-49b35bf1-9410-49fb-9a5b-632c41ecd3a6 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Task: {'id': task-5116103, 'name': PowerOnVM_Task, 'duration_secs': 0.570594} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.352182] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-49b35bf1-9410-49fb-9a5b-632c41ecd3a6 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] [instance: c4cc4f85-cb35-4edc-a58b-adfee0ce1265] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 879.352501] env[65726]: INFO nova.compute.manager [None req-49b35bf1-9410-49fb-9a5b-632c41ecd3a6 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] [instance: c4cc4f85-cb35-4edc-a58b-adfee0ce1265] Took 9.18 seconds to spawn the instance on the hypervisor. [ 879.352746] env[65726]: DEBUG nova.compute.manager [None req-49b35bf1-9410-49fb-9a5b-632c41ecd3a6 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] [instance: c4cc4f85-cb35-4edc-a58b-adfee0ce1265] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 879.353673] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7170c451-1024-45dc-ad6b-81a5134eda13 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.358337] env[65726]: DEBUG oslo_vmware.api [None req-5782ba22-8056-4347-9a97-4b1e5a35ed3b tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Waiting for the task: (returnval){ [ 879.358337] env[65726]: value = "task-5116109" [ 879.358337] env[65726]: _type = "Task" [ 879.358337] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.380395] env[65726]: DEBUG oslo_vmware.api [None req-5782ba22-8056-4347-9a97-4b1e5a35ed3b tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Task: {'id': task-5116109, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.412502] env[65726]: DEBUG oslo_vmware.api [None req-c85b0a62-807d-4793-b1be-86b28d9a6745 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] Task: {'id': task-5116107, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.076104} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.413205] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-c85b0a62-807d-4793-b1be-86b28d9a6745 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] [instance: 2cbee359-a1fc-45c0-bcc5-a2c24ded9c69] Extended root virtual disk {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 879.415164] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f64be0c1-7da5-448b-9194-3818b4698ba0 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.447268] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-c85b0a62-807d-4793-b1be-86b28d9a6745 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] [instance: 2cbee359-a1fc-45c0-bcc5-a2c24ded9c69] Reconfiguring VM instance instance-0000003c to attach disk [datastore2] 2cbee359-a1fc-45c0-bcc5-a2c24ded9c69/2cbee359-a1fc-45c0-bcc5-a2c24ded9c69.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 879.453232] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-37a7a01a-78ea-463e-9669-266b8df4675c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.486366] env[65726]: DEBUG oslo_vmware.api [None req-73b0a627-5d81-4bf2-833a-caff3534c1e3 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Task: {'id': task-5116108, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.488081] env[65726]: DEBUG oslo_vmware.api [None req-c85b0a62-807d-4793-b1be-86b28d9a6745 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] Waiting for the task: (returnval){ [ 879.488081] env[65726]: value = "task-5116110" [ 879.488081] env[65726]: _type = "Task" [ 879.488081] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.499222] env[65726]: DEBUG oslo_vmware.api [None req-c85b0a62-807d-4793-b1be-86b28d9a6745 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] Task: {'id': task-5116110, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.503135] env[65726]: WARNING openstack [req-da42b144-72ca-4f16-b99f-64a0613e87ae req-d325fa89-eede-4b44-9dc4-b6fc26d574ad service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 879.503865] env[65726]: WARNING openstack [req-da42b144-72ca-4f16-b99f-64a0613e87ae req-d325fa89-eede-4b44-9dc4-b6fc26d574ad service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 879.513142] env[65726]: DEBUG oslo_concurrency.lockutils [None req-d6fa7de0-71fb-41d9-ba04-00cd57a3d822 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 3.628s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 879.513299] env[65726]: INFO nova.compute.manager [None req-d6fa7de0-71fb-41d9-ba04-00cd57a3d822 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 608cc0ea-3f6b-4b4f-83c1-01aa50999d63] Migrating [ 879.523047] env[65726]: DEBUG oslo_concurrency.lockutils [None req-34f1738e-744a-4a46-b434-5db18d6e964a tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 42.708s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 879.525376] env[65726]: INFO nova.compute.claims [None req-34f1738e-744a-4a46-b434-5db18d6e964a tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] [instance: 4ea2b70b-0cfe-4fbf-817a-baa028b8372a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 879.603956] env[65726]: WARNING openstack [req-da42b144-72ca-4f16-b99f-64a0613e87ae req-d325fa89-eede-4b44-9dc4-b6fc26d574ad service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 879.604449] env[65726]: WARNING openstack [req-da42b144-72ca-4f16-b99f-64a0613e87ae req-d325fa89-eede-4b44-9dc4-b6fc26d574ad service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 879.690710] env[65726]: DEBUG oslo_vmware.api [None req-b20cc073-fafc-44fa-bcf9-9893dc98ebec tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52d3b37e-be70-4743-1fa5-9f65378076bd, 'name': SearchDatastore_Task, 'duration_secs': 0.026496} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.691056] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b20cc073-fafc-44fa-bcf9-9893dc98ebec tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 879.691683] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-b20cc073-fafc-44fa-bcf9-9893dc98ebec tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04] Processing image b52362a3-ee8a-4cbf-b06f-513b0cc8f95c {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 879.691683] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b20cc073-fafc-44fa-bcf9-9893dc98ebec tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 879.691683] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b20cc073-fafc-44fa-bcf9-9893dc98ebec tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 879.691884] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-b20cc073-fafc-44fa-bcf9-9893dc98ebec tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 879.692205] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4316d422-9d10-4439-98a5-cc9943f3ab6d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.709291] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-b20cc073-fafc-44fa-bcf9-9893dc98ebec tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 879.709469] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-b20cc073-fafc-44fa-bcf9-9893dc98ebec tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65726) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 879.713434] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ea9175b9-3e31-45c2-9857-05acadd67ddd {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.720705] env[65726]: DEBUG oslo_vmware.api [None req-b20cc073-fafc-44fa-bcf9-9893dc98ebec tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Waiting for the task: (returnval){ [ 879.720705] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]528b2e0b-6981-921d-2032-40d002db08fb" [ 879.720705] env[65726]: _type = "Task" [ 879.720705] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.729763] env[65726]: DEBUG oslo_vmware.api [None req-b20cc073-fafc-44fa-bcf9-9893dc98ebec tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]528b2e0b-6981-921d-2032-40d002db08fb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.733225] env[65726]: DEBUG nova.network.neutron [req-da42b144-72ca-4f16-b99f-64a0613e87ae req-d325fa89-eede-4b44-9dc4-b6fc26d574ad service nova] [instance: 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04] Updated VIF entry in instance network info cache for port 25b9d108-0b43-4459-b9db-7bd90a495bb3. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 879.733577] env[65726]: DEBUG nova.network.neutron [req-da42b144-72ca-4f16-b99f-64a0613e87ae req-d325fa89-eede-4b44-9dc4-b6fc26d574ad service nova] [instance: 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04] Updating instance_info_cache with network_info: [{"id": "25b9d108-0b43-4459-b9db-7bd90a495bb3", "address": "fa:16:3e:5a:1e:6b", "network": {"id": "eb3278b7-c190-441b-84d2-305f9b896382", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1754571022-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "96149159e18e44f9bf3453e67681f224", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f77ff7a1-209c-4f3f-b2a0-fd817741e739", "external-id": "nsx-vlan-transportzone-935", "segmentation_id": 935, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap25b9d108-0b", "ovs_interfaceid": "25b9d108-0b43-4459-b9db-7bd90a495bb3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 879.876025] env[65726]: DEBUG oslo_vmware.api [None req-5782ba22-8056-4347-9a97-4b1e5a35ed3b tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Task: {'id': task-5116109, 'name': PowerOffVM_Task, 'duration_secs': 0.236825} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.876333] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-5782ba22-8056-4347-9a97-4b1e5a35ed3b tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: f2d9090c-988f-43f4-9c81-7aa718a3438a] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 879.876333] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-5782ba22-8056-4347-9a97-4b1e5a35ed3b tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: f2d9090c-988f-43f4-9c81-7aa718a3438a] Destroying instance {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 879.877381] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-584943c0-cee2-4dc8-b1d1-d822b232fbcf {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.884790] env[65726]: INFO nova.compute.manager [None req-49b35bf1-9410-49fb-9a5b-632c41ecd3a6 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] [instance: c4cc4f85-cb35-4edc-a58b-adfee0ce1265] Took 52.41 seconds to build instance. [ 879.891032] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-5782ba22-8056-4347-9a97-4b1e5a35ed3b tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: f2d9090c-988f-43f4-9c81-7aa718a3438a] Unregistering the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 879.891804] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-12dae619-0503-4e17-8bb0-8c6f32adbfb7 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.946785] env[65726]: DEBUG oslo_vmware.api [None req-73b0a627-5d81-4bf2-833a-caff3534c1e3 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Task: {'id': task-5116108, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.570449} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.947302] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-73b0a627-5d81-4bf2-833a-caff3534c1e3 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] cf950ec3-9914-4eb9-99db-048b3969bb21/cf950ec3-9914-4eb9-99db-048b3969bb21.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 879.948035] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-73b0a627-5d81-4bf2-833a-caff3534c1e3 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: cf950ec3-9914-4eb9-99db-048b3969bb21] Extending root virtual disk to 1048576 {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 879.948035] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4fea94a2-7bf6-456f-89ca-93a4be1314df {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.957411] env[65726]: DEBUG oslo_vmware.api [None req-73b0a627-5d81-4bf2-833a-caff3534c1e3 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Waiting for the task: (returnval){ [ 879.957411] env[65726]: value = "task-5116112" [ 879.957411] env[65726]: _type = "Task" [ 879.957411] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.965866] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-5782ba22-8056-4347-9a97-4b1e5a35ed3b tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: f2d9090c-988f-43f4-9c81-7aa718a3438a] Unregistered the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 879.966116] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-5782ba22-8056-4347-9a97-4b1e5a35ed3b tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: f2d9090c-988f-43f4-9c81-7aa718a3438a] Deleting contents of the VM from datastore datastore2 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 879.966358] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-5782ba22-8056-4347-9a97-4b1e5a35ed3b tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Deleting the datastore file [datastore2] f2d9090c-988f-43f4-9c81-7aa718a3438a {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 879.969814] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-bf6fcb21-7edd-499e-b7c5-c5bfaa8ddfd7 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.972384] env[65726]: DEBUG oslo_vmware.api [None req-73b0a627-5d81-4bf2-833a-caff3534c1e3 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Task: {'id': task-5116112, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.979397] env[65726]: DEBUG oslo_vmware.api [None req-5782ba22-8056-4347-9a97-4b1e5a35ed3b tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Waiting for the task: (returnval){ [ 879.979397] env[65726]: value = "task-5116113" [ 879.979397] env[65726]: _type = "Task" [ 879.979397] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.990230] env[65726]: DEBUG oslo_vmware.api [None req-5782ba22-8056-4347-9a97-4b1e5a35ed3b tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Task: {'id': task-5116113, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.001681] env[65726]: DEBUG oslo_vmware.api [None req-c85b0a62-807d-4793-b1be-86b28d9a6745 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] Task: {'id': task-5116110, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.042055] env[65726]: DEBUG oslo_concurrency.lockutils [None req-d6fa7de0-71fb-41d9-ba04-00cd57a3d822 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Acquiring lock "refresh_cache-608cc0ea-3f6b-4b4f-83c1-01aa50999d63" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 880.042055] env[65726]: DEBUG oslo_concurrency.lockutils [None req-d6fa7de0-71fb-41d9-ba04-00cd57a3d822 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Acquired lock "refresh_cache-608cc0ea-3f6b-4b4f-83c1-01aa50999d63" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 880.042055] env[65726]: DEBUG nova.network.neutron [None req-d6fa7de0-71fb-41d9-ba04-00cd57a3d822 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 608cc0ea-3f6b-4b4f-83c1-01aa50999d63] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 880.121344] env[65726]: DEBUG oslo_concurrency.lockutils [None req-172c2bf1-c043-4217-8650-efc42870269c tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Acquiring lock "895797e4-2941-44cd-aab1-67afa4fac02d" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 880.121480] env[65726]: DEBUG oslo_concurrency.lockutils [None req-172c2bf1-c043-4217-8650-efc42870269c tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Lock "895797e4-2941-44cd-aab1-67afa4fac02d" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 880.121731] env[65726]: INFO nova.compute.manager [None req-172c2bf1-c043-4217-8650-efc42870269c tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: 895797e4-2941-44cd-aab1-67afa4fac02d] Attaching volume 3c728bbf-472e-4a76-86f7-6fd547117357 to /dev/sdb [ 880.177513] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3cf2e7b-ed75-4c0f-ab7d-9a3121468b07 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.186951] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5af3ac41-a07d-436a-b505-497497916f47 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.206381] env[65726]: DEBUG nova.virt.block_device [None req-172c2bf1-c043-4217-8650-efc42870269c tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: 895797e4-2941-44cd-aab1-67afa4fac02d] Updating existing volume attachment record: 3b4aabe4-bbd7-48fa-8cea-6bb6b65f75ba {{(pid=65726) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 880.233476] env[65726]: DEBUG oslo_vmware.api [None req-b20cc073-fafc-44fa-bcf9-9893dc98ebec tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]528b2e0b-6981-921d-2032-40d002db08fb, 'name': SearchDatastore_Task, 'duration_secs': 0.031311} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.234369] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-51be098d-996f-4b4c-9c51-0ea0d16c8c15 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.242055] env[65726]: DEBUG oslo_concurrency.lockutils [req-da42b144-72ca-4f16-b99f-64a0613e87ae req-d325fa89-eede-4b44-9dc4-b6fc26d574ad service nova] Releasing lock "refresh_cache-2fe6fc5a-f5c7-4f8e-96df-4e621a252f04" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 880.242870] env[65726]: DEBUG oslo_vmware.api [None req-b20cc073-fafc-44fa-bcf9-9893dc98ebec tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Waiting for the task: (returnval){ [ 880.242870] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]520845ac-00a4-d863-6ddb-eee51bd6a8ee" [ 880.242870] env[65726]: _type = "Task" [ 880.242870] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.258270] env[65726]: DEBUG oslo_vmware.api [None req-b20cc073-fafc-44fa-bcf9-9893dc98ebec tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]520845ac-00a4-d863-6ddb-eee51bd6a8ee, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.387850] env[65726]: DEBUG oslo_concurrency.lockutils [None req-49b35bf1-9410-49fb-9a5b-632c41ecd3a6 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Lock "c4cc4f85-cb35-4edc-a58b-adfee0ce1265" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 68.492s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 880.469091] env[65726]: DEBUG oslo_vmware.api [None req-73b0a627-5d81-4bf2-833a-caff3534c1e3 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Task: {'id': task-5116112, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.073907} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.469381] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-73b0a627-5d81-4bf2-833a-caff3534c1e3 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: cf950ec3-9914-4eb9-99db-048b3969bb21] Extended root virtual disk {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 880.470359] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-658b0c64-ad9b-4e1a-a4e8-2a5e8983a884 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.498635] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-73b0a627-5d81-4bf2-833a-caff3534c1e3 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: cf950ec3-9914-4eb9-99db-048b3969bb21] Reconfiguring VM instance instance-0000003e to attach disk [datastore1] cf950ec3-9914-4eb9-99db-048b3969bb21/cf950ec3-9914-4eb9-99db-048b3969bb21.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 880.502380] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6163edc3-8917-46c4-84c3-4201803fbc68 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.530991] env[65726]: DEBUG oslo_vmware.api [None req-5782ba22-8056-4347-9a97-4b1e5a35ed3b tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Task: {'id': task-5116113, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.343499} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.533670] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-5782ba22-8056-4347-9a97-4b1e5a35ed3b tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Deleted the datastore file {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 880.533829] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-5782ba22-8056-4347-9a97-4b1e5a35ed3b tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: f2d9090c-988f-43f4-9c81-7aa718a3438a] Deleted contents of the VM from datastore datastore2 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 880.534075] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-5782ba22-8056-4347-9a97-4b1e5a35ed3b tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: f2d9090c-988f-43f4-9c81-7aa718a3438a] Instance destroyed {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 880.537730] env[65726]: DEBUG oslo_vmware.api [None req-c85b0a62-807d-4793-b1be-86b28d9a6745 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] Task: {'id': task-5116110, 'name': ReconfigVM_Task, 'duration_secs': 0.820269} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.537730] env[65726]: DEBUG oslo_vmware.api [None req-73b0a627-5d81-4bf2-833a-caff3534c1e3 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Waiting for the task: (returnval){ [ 880.537730] env[65726]: value = "task-5116116" [ 880.537730] env[65726]: _type = "Task" [ 880.537730] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.542114] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-c85b0a62-807d-4793-b1be-86b28d9a6745 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] [instance: 2cbee359-a1fc-45c0-bcc5-a2c24ded9c69] Reconfigured VM instance instance-0000003c to attach disk [datastore2] 2cbee359-a1fc-45c0-bcc5-a2c24ded9c69/2cbee359-a1fc-45c0-bcc5-a2c24ded9c69.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 880.542114] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6cba4fcf-b09d-496e-8576-e2b09a928e16 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.549356] env[65726]: WARNING openstack [None req-d6fa7de0-71fb-41d9-ba04-00cd57a3d822 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 880.550079] env[65726]: WARNING openstack [None req-d6fa7de0-71fb-41d9-ba04-00cd57a3d822 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 880.565546] env[65726]: DEBUG oslo_vmware.api [None req-73b0a627-5d81-4bf2-833a-caff3534c1e3 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Task: {'id': task-5116116, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.567170] env[65726]: DEBUG oslo_vmware.api [None req-c85b0a62-807d-4793-b1be-86b28d9a6745 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] Waiting for the task: (returnval){ [ 880.567170] env[65726]: value = "task-5116118" [ 880.567170] env[65726]: _type = "Task" [ 880.567170] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.579891] env[65726]: DEBUG oslo_vmware.api [None req-c85b0a62-807d-4793-b1be-86b28d9a6745 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] Task: {'id': task-5116118, 'name': Rename_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.737377] env[65726]: WARNING openstack [None req-d6fa7de0-71fb-41d9-ba04-00cd57a3d822 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 880.738035] env[65726]: WARNING openstack [None req-d6fa7de0-71fb-41d9-ba04-00cd57a3d822 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 880.759626] env[65726]: DEBUG oslo_vmware.api [None req-b20cc073-fafc-44fa-bcf9-9893dc98ebec tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]520845ac-00a4-d863-6ddb-eee51bd6a8ee, 'name': SearchDatastore_Task, 'duration_secs': 0.03739} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.759951] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b20cc073-fafc-44fa-bcf9-9893dc98ebec tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 880.760261] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-b20cc073-fafc-44fa-bcf9-9893dc98ebec tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore2] 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04/2fe6fc5a-f5c7-4f8e-96df-4e621a252f04.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 880.760857] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1b64146b-72fc-4428-8c26-a7746473c367 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.768732] env[65726]: DEBUG oslo_vmware.api [None req-b20cc073-fafc-44fa-bcf9-9893dc98ebec tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Waiting for the task: (returnval){ [ 880.768732] env[65726]: value = "task-5116119" [ 880.768732] env[65726]: _type = "Task" [ 880.768732] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.785111] env[65726]: DEBUG oslo_vmware.api [None req-b20cc073-fafc-44fa-bcf9-9893dc98ebec tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Task: {'id': task-5116119, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.812542] env[65726]: WARNING openstack [None req-d6fa7de0-71fb-41d9-ba04-00cd57a3d822 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 880.813043] env[65726]: WARNING openstack [None req-d6fa7de0-71fb-41d9-ba04-00cd57a3d822 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 880.894354] env[65726]: DEBUG nova.compute.manager [None req-2cc737a8-ad7d-449c-b270-9db9e18d971d tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: dcde25df-87a5-47fb-94ce-334f68894e04] Starting instance... {{(pid=65726) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 880.942146] env[65726]: DEBUG nova.network.neutron [None req-d6fa7de0-71fb-41d9-ba04-00cd57a3d822 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 608cc0ea-3f6b-4b4f-83c1-01aa50999d63] Updating instance_info_cache with network_info: [{"id": "e9f7132a-8d5a-47fe-afc6-d45bf3687ea4", "address": "fa:16:3e:fe:4d:df", "network": {"id": "0d0f561d-ac41-4ef5-9926-c707409799f7", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1134321217-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.160", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4eece77569624f90bf64e5c51974173f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2ce62383-8e84-4e26-955b-74c11392f4c9", "external-id": "nsx-vlan-transportzone-215", "segmentation_id": 215, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape9f7132a-8d", "ovs_interfaceid": "e9f7132a-8d5a-47fe-afc6-d45bf3687ea4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 881.060685] env[65726]: DEBUG oslo_vmware.api [None req-73b0a627-5d81-4bf2-833a-caff3534c1e3 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Task: {'id': task-5116116, 'name': ReconfigVM_Task, 'duration_secs': 0.500054} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.061041] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-73b0a627-5d81-4bf2-833a-caff3534c1e3 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: cf950ec3-9914-4eb9-99db-048b3969bb21] Reconfigured VM instance instance-0000003e to attach disk [datastore1] cf950ec3-9914-4eb9-99db-048b3969bb21/cf950ec3-9914-4eb9-99db-048b3969bb21.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 881.062372] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a92da638-740c-4751-9d00-4e88f28b9553 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.076844] env[65726]: DEBUG oslo_vmware.api [None req-73b0a627-5d81-4bf2-833a-caff3534c1e3 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Waiting for the task: (returnval){ [ 881.076844] env[65726]: value = "task-5116120" [ 881.076844] env[65726]: _type = "Task" [ 881.076844] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.084942] env[65726]: DEBUG oslo_vmware.api [None req-c85b0a62-807d-4793-b1be-86b28d9a6745 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] Task: {'id': task-5116118, 'name': Rename_Task, 'duration_secs': 0.233198} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.089503] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-c85b0a62-807d-4793-b1be-86b28d9a6745 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] [instance: 2cbee359-a1fc-45c0-bcc5-a2c24ded9c69] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 881.089503] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ad434b89-b84d-4683-a51e-f0184df6b189 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.095482] env[65726]: DEBUG oslo_vmware.api [None req-73b0a627-5d81-4bf2-833a-caff3534c1e3 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Task: {'id': task-5116120, 'name': Rename_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.107055] env[65726]: DEBUG oslo_vmware.api [None req-c85b0a62-807d-4793-b1be-86b28d9a6745 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] Waiting for the task: (returnval){ [ 881.107055] env[65726]: value = "task-5116121" [ 881.107055] env[65726]: _type = "Task" [ 881.107055] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.118784] env[65726]: DEBUG oslo_vmware.api [None req-c85b0a62-807d-4793-b1be-86b28d9a6745 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] Task: {'id': task-5116121, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.232493] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7f9bed9-faea-4531-a64b-637ad5dcffdd {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.241362] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6242c0a2-182e-42cb-b02f-39e9ef598cb7 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.282767] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aab4be0a-f8f3-490e-8f19-54c794c6c846 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.291047] env[65726]: DEBUG oslo_vmware.api [None req-b20cc073-fafc-44fa-bcf9-9893dc98ebec tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Task: {'id': task-5116119, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.294851] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ceed9d0c-ba01-49ad-9489-3e266d31768c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.311331] env[65726]: DEBUG nova.compute.provider_tree [None req-34f1738e-744a-4a46-b434-5db18d6e964a tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 881.423843] env[65726]: DEBUG oslo_concurrency.lockutils [None req-2cc737a8-ad7d-449c-b270-9db9e18d971d tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 881.445363] env[65726]: DEBUG oslo_concurrency.lockutils [None req-d6fa7de0-71fb-41d9-ba04-00cd57a3d822 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Releasing lock "refresh_cache-608cc0ea-3f6b-4b4f-83c1-01aa50999d63" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 881.585394] env[65726]: DEBUG nova.virt.hardware [None req-5782ba22-8056-4347-9a97-4b1e5a35ed3b tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T19:28:34Z,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1539f06fa3534e90acbc3a60c4b8bd3f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T19:28:35Z,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 881.586592] env[65726]: DEBUG nova.virt.hardware [None req-5782ba22-8056-4347-9a97-4b1e5a35ed3b tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 881.586592] env[65726]: DEBUG nova.virt.hardware [None req-5782ba22-8056-4347-9a97-4b1e5a35ed3b tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 881.586592] env[65726]: DEBUG nova.virt.hardware [None req-5782ba22-8056-4347-9a97-4b1e5a35ed3b tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 881.586592] env[65726]: DEBUG nova.virt.hardware [None req-5782ba22-8056-4347-9a97-4b1e5a35ed3b tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 881.586592] env[65726]: DEBUG nova.virt.hardware [None req-5782ba22-8056-4347-9a97-4b1e5a35ed3b tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 881.586764] env[65726]: DEBUG nova.virt.hardware [None req-5782ba22-8056-4347-9a97-4b1e5a35ed3b tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 881.587071] env[65726]: DEBUG nova.virt.hardware [None req-5782ba22-8056-4347-9a97-4b1e5a35ed3b tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 881.587178] env[65726]: DEBUG nova.virt.hardware [None req-5782ba22-8056-4347-9a97-4b1e5a35ed3b tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 881.587350] env[65726]: DEBUG nova.virt.hardware [None req-5782ba22-8056-4347-9a97-4b1e5a35ed3b tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 881.587597] env[65726]: DEBUG nova.virt.hardware [None req-5782ba22-8056-4347-9a97-4b1e5a35ed3b tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 881.588493] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-096531b8-944e-4d16-9f6c-1d1965f88eb2 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.597720] env[65726]: DEBUG oslo_vmware.api [None req-73b0a627-5d81-4bf2-833a-caff3534c1e3 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Task: {'id': task-5116120, 'name': Rename_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.601240] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f761cfb5-5497-43ac-aca8-85b112203ffa {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.619296] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-5782ba22-8056-4347-9a97-4b1e5a35ed3b tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: f2d9090c-988f-43f4-9c81-7aa718a3438a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:10:2a:24', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd0bdd11b-58af-4cc0-9d38-8322e1bb4e74', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7117cd21-ee2a-497d-b789-65e43b068258', 'vif_model': 'vmxnet3'}] {{(pid=65726) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 881.627018] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-5782ba22-8056-4347-9a97-4b1e5a35ed3b tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 881.637245] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f2d9090c-988f-43f4-9c81-7aa718a3438a] Creating VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 881.637245] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-bf9fa353-86df-4383-8527-b94c489a9926 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.656434] env[65726]: DEBUG oslo_vmware.api [None req-c85b0a62-807d-4793-b1be-86b28d9a6745 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] Task: {'id': task-5116121, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.658218] env[65726]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 881.658218] env[65726]: value = "task-5116122" [ 881.658218] env[65726]: _type = "Task" [ 881.658218] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.666978] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116122, 'name': CreateVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.788993] env[65726]: DEBUG oslo_vmware.api [None req-b20cc073-fafc-44fa-bcf9-9893dc98ebec tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Task: {'id': task-5116119, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.531875} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.789366] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-b20cc073-fafc-44fa-bcf9-9893dc98ebec tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore2] 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04/2fe6fc5a-f5c7-4f8e-96df-4e621a252f04.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 881.789597] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-b20cc073-fafc-44fa-bcf9-9893dc98ebec tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04] Extending root virtual disk to 1048576 {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 881.789932] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ee9a86c7-af1e-45ff-ace9-9cbe01e8de07 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.797623] env[65726]: DEBUG oslo_vmware.api [None req-b20cc073-fafc-44fa-bcf9-9893dc98ebec tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Waiting for the task: (returnval){ [ 881.797623] env[65726]: value = "task-5116123" [ 881.797623] env[65726]: _type = "Task" [ 881.797623] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.808256] env[65726]: DEBUG oslo_vmware.api [None req-b20cc073-fafc-44fa-bcf9-9893dc98ebec tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Task: {'id': task-5116123, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.814501] env[65726]: DEBUG nova.scheduler.client.report [None req-34f1738e-744a-4a46-b434-5db18d6e964a tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 882.089735] env[65726]: DEBUG oslo_vmware.api [None req-73b0a627-5d81-4bf2-833a-caff3534c1e3 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Task: {'id': task-5116120, 'name': Rename_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.129013] env[65726]: DEBUG oslo_vmware.api [None req-c85b0a62-807d-4793-b1be-86b28d9a6745 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] Task: {'id': task-5116121, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.135509] env[65726]: DEBUG nova.compute.manager [None req-fcafe3da-7a8e-4b08-9577-72f88a634bf5 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] [instance: c4cc4f85-cb35-4edc-a58b-adfee0ce1265] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 882.136345] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c233c1aa-e556-4754-8486-b65019289272 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.170764] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116122, 'name': CreateVM_Task} progress is 25%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.308248] env[65726]: DEBUG oslo_vmware.api [None req-b20cc073-fafc-44fa-bcf9-9893dc98ebec tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Task: {'id': task-5116123, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.143155} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.308536] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-b20cc073-fafc-44fa-bcf9-9893dc98ebec tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04] Extended root virtual disk {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 882.309358] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98ddfcdd-398c-4d60-9e56-d19a033509f0 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.326513] env[65726]: DEBUG oslo_concurrency.lockutils [None req-34f1738e-744a-4a46-b434-5db18d6e964a tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.803s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 882.327283] env[65726]: DEBUG nova.compute.manager [None req-34f1738e-744a-4a46-b434-5db18d6e964a tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] [instance: 4ea2b70b-0cfe-4fbf-817a-baa028b8372a] Start building networks asynchronously for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 882.339102] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-b20cc073-fafc-44fa-bcf9-9893dc98ebec tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04] Reconfiguring VM instance instance-0000003f to attach disk [datastore2] 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04/2fe6fc5a-f5c7-4f8e-96df-4e621a252f04.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 882.339572] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 43.066s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 882.341462] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a2452268-dbf9-4366-a221-4c7845160744 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.364045] env[65726]: DEBUG oslo_vmware.api [None req-b20cc073-fafc-44fa-bcf9-9893dc98ebec tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Waiting for the task: (returnval){ [ 882.364045] env[65726]: value = "task-5116124" [ 882.364045] env[65726]: _type = "Task" [ 882.364045] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.373742] env[65726]: DEBUG oslo_vmware.api [None req-b20cc073-fafc-44fa-bcf9-9893dc98ebec tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Task: {'id': task-5116124, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.594352] env[65726]: DEBUG oslo_vmware.api [None req-73b0a627-5d81-4bf2-833a-caff3534c1e3 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Task: {'id': task-5116120, 'name': Rename_Task, 'duration_secs': 1.182406} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.594352] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-73b0a627-5d81-4bf2-833a-caff3534c1e3 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: cf950ec3-9914-4eb9-99db-048b3969bb21] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 882.594352] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-256988f3-f28c-4c14-86f5-ec6d5988538e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.601836] env[65726]: DEBUG oslo_vmware.api [None req-73b0a627-5d81-4bf2-833a-caff3534c1e3 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Waiting for the task: (returnval){ [ 882.601836] env[65726]: value = "task-5116125" [ 882.601836] env[65726]: _type = "Task" [ 882.601836] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.611899] env[65726]: DEBUG oslo_vmware.api [None req-73b0a627-5d81-4bf2-833a-caff3534c1e3 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Task: {'id': task-5116125, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.631755] env[65726]: DEBUG oslo_vmware.api [None req-c85b0a62-807d-4793-b1be-86b28d9a6745 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] Task: {'id': task-5116121, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.651391] env[65726]: INFO nova.compute.manager [None req-fcafe3da-7a8e-4b08-9577-72f88a634bf5 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] [instance: c4cc4f85-cb35-4edc-a58b-adfee0ce1265] instance snapshotting [ 882.655862] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45e7202c-564c-458c-8b53-e1403b7ef194 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.679827] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d291549-fd14-4d02-aaa9-46447a9efe38 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.685934] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116122, 'name': CreateVM_Task, 'duration_secs': 0.8081} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.686573] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f2d9090c-988f-43f4-9c81-7aa718a3438a] Created VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 882.687452] env[65726]: WARNING openstack [None req-5782ba22-8056-4347-9a97-4b1e5a35ed3b tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 882.687868] env[65726]: WARNING openstack [None req-5782ba22-8056-4347-9a97-4b1e5a35ed3b tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 882.693334] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5782ba22-8056-4347-9a97-4b1e5a35ed3b tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 882.693509] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5782ba22-8056-4347-9a97-4b1e5a35ed3b tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 882.694428] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5782ba22-8056-4347-9a97-4b1e5a35ed3b tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 882.696104] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dd170283-dbc2-406b-874f-12f941b01ab3 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.705158] env[65726]: DEBUG oslo_vmware.api [None req-5782ba22-8056-4347-9a97-4b1e5a35ed3b tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Waiting for the task: (returnval){ [ 882.705158] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52d8fb4c-a514-6754-2c8c-accc53648682" [ 882.705158] env[65726]: _type = "Task" [ 882.705158] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.715835] env[65726]: DEBUG oslo_vmware.api [None req-5782ba22-8056-4347-9a97-4b1e5a35ed3b tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52d8fb4c-a514-6754-2c8c-accc53648682, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.841586] env[65726]: DEBUG nova.compute.utils [None req-34f1738e-744a-4a46-b434-5db18d6e964a tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Using /dev/sd instead of None {{(pid=65726) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 882.843870] env[65726]: DEBUG nova.compute.manager [None req-34f1738e-744a-4a46-b434-5db18d6e964a tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] [instance: 4ea2b70b-0cfe-4fbf-817a-baa028b8372a] Allocating IP information in the background. {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 882.844249] env[65726]: DEBUG nova.network.neutron [None req-34f1738e-744a-4a46-b434-5db18d6e964a tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] [instance: 4ea2b70b-0cfe-4fbf-817a-baa028b8372a] allocate_for_instance() {{(pid=65726) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 882.844458] env[65726]: WARNING neutronclient.v2_0.client [None req-34f1738e-744a-4a46-b434-5db18d6e964a tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 882.844885] env[65726]: WARNING neutronclient.v2_0.client [None req-34f1738e-744a-4a46-b434-5db18d6e964a tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 882.845611] env[65726]: WARNING openstack [None req-34f1738e-744a-4a46-b434-5db18d6e964a tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 882.845975] env[65726]: WARNING openstack [None req-34f1738e-744a-4a46-b434-5db18d6e964a tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 882.886544] env[65726]: DEBUG oslo_vmware.api [None req-b20cc073-fafc-44fa-bcf9-9893dc98ebec tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Task: {'id': task-5116124, 'name': ReconfigVM_Task, 'duration_secs': 0.399912} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.886967] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-b20cc073-fafc-44fa-bcf9-9893dc98ebec tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04] Reconfigured VM instance instance-0000003f to attach disk [datastore2] 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04/2fe6fc5a-f5c7-4f8e-96df-4e621a252f04.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 882.887832] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7a8b16cc-f4e1-49b2-aeec-8d19d6a454f9 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.896471] env[65726]: DEBUG oslo_vmware.api [None req-b20cc073-fafc-44fa-bcf9-9893dc98ebec tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Waiting for the task: (returnval){ [ 882.896471] env[65726]: value = "task-5116127" [ 882.896471] env[65726]: _type = "Task" [ 882.896471] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.912674] env[65726]: DEBUG oslo_vmware.api [None req-b20cc073-fafc-44fa-bcf9-9893dc98ebec tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Task: {'id': task-5116127, 'name': Rename_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.917159] env[65726]: DEBUG nova.policy [None req-34f1738e-744a-4a46-b434-5db18d6e964a tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e5c535534cca4b6f9e55c2ada589dc05', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '55f93cdf51e84cf683879668c2acecf4', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65726) authorize /opt/stack/nova/nova/policy.py:192}} [ 882.960487] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34b4d3dd-209f-4d52-8e21-d346c0817ffa {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.986166] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-d6fa7de0-71fb-41d9-ba04-00cd57a3d822 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 608cc0ea-3f6b-4b4f-83c1-01aa50999d63] Updating instance '608cc0ea-3f6b-4b4f-83c1-01aa50999d63' progress to 0 {{(pid=65726) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 883.113449] env[65726]: DEBUG oslo_vmware.api [None req-73b0a627-5d81-4bf2-833a-caff3534c1e3 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Task: {'id': task-5116125, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.131438] env[65726]: DEBUG oslo_vmware.api [None req-c85b0a62-807d-4793-b1be-86b28d9a6745 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] Task: {'id': task-5116121, 'name': PowerOnVM_Task, 'duration_secs': 1.547028} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.131749] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-c85b0a62-807d-4793-b1be-86b28d9a6745 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] [instance: 2cbee359-a1fc-45c0-bcc5-a2c24ded9c69] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 883.131952] env[65726]: DEBUG nova.compute.manager [None req-c85b0a62-807d-4793-b1be-86b28d9a6745 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] [instance: 2cbee359-a1fc-45c0-bcc5-a2c24ded9c69] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 883.132815] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a6c106f-79b9-4052-8f8c-f1e2fbfbea5e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.202339] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-fcafe3da-7a8e-4b08-9577-72f88a634bf5 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] [instance: c4cc4f85-cb35-4edc-a58b-adfee0ce1265] Creating Snapshot of the VM instance {{(pid=65726) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 883.202661] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-2e5906f5-13f0-4631-b3b2-c137be62eb05 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.217402] env[65726]: DEBUG oslo_vmware.api [None req-5782ba22-8056-4347-9a97-4b1e5a35ed3b tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52d8fb4c-a514-6754-2c8c-accc53648682, 'name': SearchDatastore_Task, 'duration_secs': 0.01213} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.219219] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5782ba22-8056-4347-9a97-4b1e5a35ed3b tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 883.219475] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-5782ba22-8056-4347-9a97-4b1e5a35ed3b tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: f2d9090c-988f-43f4-9c81-7aa718a3438a] Processing image b52362a3-ee8a-4cbf-b06f-513b0cc8f95c {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 883.219780] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5782ba22-8056-4347-9a97-4b1e5a35ed3b tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 883.220022] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5782ba22-8056-4347-9a97-4b1e5a35ed3b tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 883.220309] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-5782ba22-8056-4347-9a97-4b1e5a35ed3b tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 883.220766] env[65726]: DEBUG oslo_vmware.api [None req-fcafe3da-7a8e-4b08-9577-72f88a634bf5 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Waiting for the task: (returnval){ [ 883.220766] env[65726]: value = "task-5116128" [ 883.220766] env[65726]: _type = "Task" [ 883.220766] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.220976] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f0f1a089-5e64-4f22-b3d1-6564fe0c14e8 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.230342] env[65726]: DEBUG nova.network.neutron [None req-34f1738e-744a-4a46-b434-5db18d6e964a tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] [instance: 4ea2b70b-0cfe-4fbf-817a-baa028b8372a] Successfully created port: a97348ff-93e0-442b-8a41-b56fa04e8cb6 {{(pid=65726) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 883.240454] env[65726]: DEBUG oslo_vmware.api [None req-fcafe3da-7a8e-4b08-9577-72f88a634bf5 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Task: {'id': task-5116128, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.243274] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-5782ba22-8056-4347-9a97-4b1e5a35ed3b tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 883.243614] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-5782ba22-8056-4347-9a97-4b1e5a35ed3b tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65726) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 883.244495] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-124630d0-590e-4689-8ec5-d6bb8d4f9adc {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.253233] env[65726]: DEBUG oslo_vmware.api [None req-5782ba22-8056-4347-9a97-4b1e5a35ed3b tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Waiting for the task: (returnval){ [ 883.253233] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52a9fcd0-8691-6e02-b601-3a25cef8dea9" [ 883.253233] env[65726]: _type = "Task" [ 883.253233] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.263039] env[65726]: DEBUG oslo_vmware.api [None req-5782ba22-8056-4347-9a97-4b1e5a35ed3b tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52a9fcd0-8691-6e02-b601-3a25cef8dea9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.354621] env[65726]: DEBUG nova.compute.manager [None req-34f1738e-744a-4a46-b434-5db18d6e964a tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] [instance: 4ea2b70b-0cfe-4fbf-817a-baa028b8372a] Start building block device mappings for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 883.380208] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Applying migration context for instance 76249623-6f83-46a3-b8c5-c001111aa698 as it has an incoming, in-progress migration 8f31ab13-33c9-4d52-9a2a-8c3409ce12a9. Migration status is reverting {{(pid=65726) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1046}} [ 883.380208] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Applying migration context for instance 608cc0ea-3f6b-4b4f-83c1-01aa50999d63 as it has an incoming, in-progress migration 3a215dd8-f815-44fc-8620-fb9bc12739f7. Migration status is migrating {{(pid=65726) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1046}} [ 883.383033] env[65726]: INFO nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] [instance: 76249623-6f83-46a3-b8c5-c001111aa698] Updating resource usage from migration 8f31ab13-33c9-4d52-9a2a-8c3409ce12a9 [ 883.383033] env[65726]: INFO nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] [instance: 608cc0ea-3f6b-4b4f-83c1-01aa50999d63] Updating resource usage from migration 3a215dd8-f815-44fc-8620-fb9bc12739f7 [ 883.408716] env[65726]: DEBUG oslo_vmware.api [None req-b20cc073-fafc-44fa-bcf9-9893dc98ebec tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Task: {'id': task-5116127, 'name': Rename_Task, 'duration_secs': 0.187768} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.409587] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-b20cc073-fafc-44fa-bcf9-9893dc98ebec tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 883.409828] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3319b263-f459-40a0-a259-38ef5811f8d8 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.413680] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Instance f2d9090c-988f-43f4-9c81-7aa718a3438a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65726) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 883.413821] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Instance 274ab469-61a9-4b7e-852c-074c871e3abf actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65726) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 883.413937] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Instance 0e064341-4e4a-407b-8c26-3eb04b409029 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=65726) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 883.414083] env[65726]: WARNING nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Instance d7905c4f-1ba3-43b3-b4b3-aebacd2f9cc3 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 883.414195] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Instance c108f2a5-031b-47ee-9a5b-d62c2f42c26b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65726) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 883.414329] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Instance 895797e4-2941-44cd-aab1-67afa4fac02d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65726) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 883.414486] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Instance 6d2b86c2-0a44-433f-8b3f-2e9c7f693baa actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65726) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 883.414766] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Instance 9fc0c9ee-9dc7-4059-b8d2-a5db965e00c2 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65726) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 883.414766] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Instance bc2c12e0-0d06-432f-b42f-be468e3b6ee4 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65726) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 883.414870] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Instance 6c8db442-daaa-4eca-ae24-b9d4afaf3a71 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 2, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65726) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 883.414976] env[65726]: WARNING nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Instance 3cbbe146-be3a-43d5-867d-4d669884758c is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 883.415864] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Instance c8be64a3-1bfa-41a3-b9be-68a7e6e052e0 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65726) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 883.415864] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Instance 01d86089-6b9b-4588-864e-ef91375a1eea actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65726) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 883.415864] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Instance 1cfb1bba-1e21-47eb-b50f-a86575846a65 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65726) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 883.415864] env[65726]: WARNING nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Instance 44d73b2d-2b6a-4501-9944-432da3c9330d is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 883.415864] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Instance cb8855b6-589d-4863-b86f-7fb4e30fb29e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65726) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 883.415864] env[65726]: WARNING nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Instance c4177e20-b1bd-4b54-a275-c93582359a07 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 883.415864] env[65726]: WARNING nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Instance d179030a-e5ba-45b1-ad11-4d2f71ed7bd4 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 883.416273] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Migration 8f31ab13-33c9-4d52-9a2a-8c3409ce12a9 is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65726) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1745}} [ 883.416273] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Instance 76249623-6f83-46a3-b8c5-c001111aa698 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=65726) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 883.416273] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Instance 2cbee359-a1fc-45c0-bcc5-a2c24ded9c69 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65726) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 883.416460] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Instance c4cc4f85-cb35-4edc-a58b-adfee0ce1265 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65726) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 883.416460] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Instance cf950ec3-9914-4eb9-99db-048b3969bb21 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65726) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 883.416551] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Instance 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65726) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 883.416595] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Instance 4ea2b70b-0cfe-4fbf-817a-baa028b8372a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65726) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 883.420392] env[65726]: DEBUG oslo_vmware.api [None req-b20cc073-fafc-44fa-bcf9-9893dc98ebec tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Waiting for the task: (returnval){ [ 883.420392] env[65726]: value = "task-5116129" [ 883.420392] env[65726]: _type = "Task" [ 883.420392] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.429880] env[65726]: DEBUG oslo_vmware.api [None req-b20cc073-fafc-44fa-bcf9-9893dc98ebec tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Task: {'id': task-5116129, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.494304] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-d6fa7de0-71fb-41d9-ba04-00cd57a3d822 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 608cc0ea-3f6b-4b4f-83c1-01aa50999d63] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 883.495042] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a44edac5-74a7-4dff-8e74-fadfa644b57c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.504341] env[65726]: DEBUG oslo_vmware.api [None req-d6fa7de0-71fb-41d9-ba04-00cd57a3d822 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Waiting for the task: (returnval){ [ 883.504341] env[65726]: value = "task-5116130" [ 883.504341] env[65726]: _type = "Task" [ 883.504341] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.513600] env[65726]: DEBUG oslo_vmware.api [None req-d6fa7de0-71fb-41d9-ba04-00cd57a3d822 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5116130, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.613271] env[65726]: DEBUG oslo_vmware.api [None req-73b0a627-5d81-4bf2-833a-caff3534c1e3 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Task: {'id': task-5116125, 'name': PowerOnVM_Task, 'duration_secs': 0.66456} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.613571] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-73b0a627-5d81-4bf2-833a-caff3534c1e3 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: cf950ec3-9914-4eb9-99db-048b3969bb21] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 883.613770] env[65726]: INFO nova.compute.manager [None req-73b0a627-5d81-4bf2-833a-caff3534c1e3 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: cf950ec3-9914-4eb9-99db-048b3969bb21] Took 10.67 seconds to spawn the instance on the hypervisor. [ 883.613943] env[65726]: DEBUG nova.compute.manager [None req-73b0a627-5d81-4bf2-833a-caff3534c1e3 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: cf950ec3-9914-4eb9-99db-048b3969bb21] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 883.614739] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eec9026e-fd00-4fe0-bb1a-7db43183b9cd {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.653352] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c85b0a62-807d-4793-b1be-86b28d9a6745 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 883.733377] env[65726]: DEBUG oslo_vmware.api [None req-fcafe3da-7a8e-4b08-9577-72f88a634bf5 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Task: {'id': task-5116128, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.763660] env[65726]: DEBUG oslo_vmware.api [None req-5782ba22-8056-4347-9a97-4b1e5a35ed3b tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52a9fcd0-8691-6e02-b601-3a25cef8dea9, 'name': SearchDatastore_Task, 'duration_secs': 0.012501} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.764479] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d654d9bb-59ab-4a94-a4ab-36485a5e6121 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.771817] env[65726]: DEBUG oslo_vmware.api [None req-5782ba22-8056-4347-9a97-4b1e5a35ed3b tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Waiting for the task: (returnval){ [ 883.771817] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]528960d0-d8c1-e5ef-006c-a73511ede52a" [ 883.771817] env[65726]: _type = "Task" [ 883.771817] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.781408] env[65726]: DEBUG oslo_vmware.api [None req-5782ba22-8056-4347-9a97-4b1e5a35ed3b tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]528960d0-d8c1-e5ef-006c-a73511ede52a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.921315] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Instance 815ee20b-cb07-4514-bd43-1b74972508a5 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65726) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 883.921528] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Migration 3a215dd8-f815-44fc-8620-fb9bc12739f7 is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65726) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1745}} [ 883.921528] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Instance 608cc0ea-3f6b-4b4f-83c1-01aa50999d63 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=65726) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 883.932736] env[65726]: DEBUG oslo_vmware.api [None req-b20cc073-fafc-44fa-bcf9-9893dc98ebec tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Task: {'id': task-5116129, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.016259] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-d6fa7de0-71fb-41d9-ba04-00cd57a3d822 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 608cc0ea-3f6b-4b4f-83c1-01aa50999d63] VM already powered off {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 884.016473] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-d6fa7de0-71fb-41d9-ba04-00cd57a3d822 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 608cc0ea-3f6b-4b4f-83c1-01aa50999d63] Updating instance '608cc0ea-3f6b-4b4f-83c1-01aa50999d63' progress to 17 {{(pid=65726) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 884.133906] env[65726]: INFO nova.compute.manager [None req-73b0a627-5d81-4bf2-833a-caff3534c1e3 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: cf950ec3-9914-4eb9-99db-048b3969bb21] Took 54.44 seconds to build instance. [ 884.240584] env[65726]: DEBUG oslo_vmware.api [None req-fcafe3da-7a8e-4b08-9577-72f88a634bf5 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Task: {'id': task-5116128, 'name': CreateSnapshot_Task, 'duration_secs': 0.909332} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.240931] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-fcafe3da-7a8e-4b08-9577-72f88a634bf5 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] [instance: c4cc4f85-cb35-4edc-a58b-adfee0ce1265] Created Snapshot of the VM instance {{(pid=65726) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 884.241880] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ea62cc0-99db-4d1c-a9b5-c81cd4675449 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.283860] env[65726]: DEBUG oslo_vmware.api [None req-5782ba22-8056-4347-9a97-4b1e5a35ed3b tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]528960d0-d8c1-e5ef-006c-a73511ede52a, 'name': SearchDatastore_Task, 'duration_secs': 0.010345} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.284218] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5782ba22-8056-4347-9a97-4b1e5a35ed3b tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 884.284519] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-5782ba22-8056-4347-9a97-4b1e5a35ed3b tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] f2d9090c-988f-43f4-9c81-7aa718a3438a/f2d9090c-988f-43f4-9c81-7aa718a3438a.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 884.284821] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-cab46876-bf7a-481d-9575-fd26d5223934 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.294378] env[65726]: DEBUG oslo_vmware.api [None req-5782ba22-8056-4347-9a97-4b1e5a35ed3b tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Waiting for the task: (returnval){ [ 884.294378] env[65726]: value = "task-5116131" [ 884.294378] env[65726]: _type = "Task" [ 884.294378] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.305645] env[65726]: DEBUG oslo_vmware.api [None req-5782ba22-8056-4347-9a97-4b1e5a35ed3b tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Task: {'id': task-5116131, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.365533] env[65726]: DEBUG nova.compute.manager [None req-34f1738e-744a-4a46-b434-5db18d6e964a tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] [instance: 4ea2b70b-0cfe-4fbf-817a-baa028b8372a] Start spawning the instance on the hypervisor. {{(pid=65726) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 884.395936] env[65726]: DEBUG nova.virt.hardware [None req-34f1738e-744a-4a46-b434-5db18d6e964a tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T19:28:34Z,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1539f06fa3534e90acbc3a60c4b8bd3f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T19:28:35Z,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 884.396203] env[65726]: DEBUG nova.virt.hardware [None req-34f1738e-744a-4a46-b434-5db18d6e964a tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 884.396357] env[65726]: DEBUG nova.virt.hardware [None req-34f1738e-744a-4a46-b434-5db18d6e964a tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 884.396537] env[65726]: DEBUG nova.virt.hardware [None req-34f1738e-744a-4a46-b434-5db18d6e964a tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 884.396691] env[65726]: DEBUG nova.virt.hardware [None req-34f1738e-744a-4a46-b434-5db18d6e964a tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 884.396878] env[65726]: DEBUG nova.virt.hardware [None req-34f1738e-744a-4a46-b434-5db18d6e964a tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 884.397560] env[65726]: DEBUG nova.virt.hardware [None req-34f1738e-744a-4a46-b434-5db18d6e964a tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 884.397560] env[65726]: DEBUG nova.virt.hardware [None req-34f1738e-744a-4a46-b434-5db18d6e964a tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 884.397560] env[65726]: DEBUG nova.virt.hardware [None req-34f1738e-744a-4a46-b434-5db18d6e964a tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 884.397797] env[65726]: DEBUG nova.virt.hardware [None req-34f1738e-744a-4a46-b434-5db18d6e964a tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 884.397878] env[65726]: DEBUG nova.virt.hardware [None req-34f1738e-744a-4a46-b434-5db18d6e964a tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 884.398926] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4cb1f44f-236b-4f9c-8e9c-a93a8f7d19fd {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.410220] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6285392a-8057-499b-8535-8b518d550693 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.430081] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Instance 29823bc7-3909-4ab7-8119-91ee59b289c5 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65726) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 884.437659] env[65726]: DEBUG oslo_vmware.api [None req-b20cc073-fafc-44fa-bcf9-9893dc98ebec tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Task: {'id': task-5116129, 'name': PowerOnVM_Task, 'duration_secs': 0.786347} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.438012] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-b20cc073-fafc-44fa-bcf9-9893dc98ebec tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 884.438245] env[65726]: INFO nova.compute.manager [None req-b20cc073-fafc-44fa-bcf9-9893dc98ebec tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04] Took 8.74 seconds to spawn the instance on the hypervisor. [ 884.438443] env[65726]: DEBUG nova.compute.manager [None req-b20cc073-fafc-44fa-bcf9-9893dc98ebec tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 884.439309] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1ff0bc3-354c-4d87-8bc7-e3a4a8424cfe {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.523824] env[65726]: DEBUG nova.virt.hardware [None req-d6fa7de0-71fb-41d9-ba04-00cd57a3d822 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 884.524076] env[65726]: DEBUG nova.virt.hardware [None req-d6fa7de0-71fb-41d9-ba04-00cd57a3d822 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 884.524234] env[65726]: DEBUG nova.virt.hardware [None req-d6fa7de0-71fb-41d9-ba04-00cd57a3d822 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 884.524410] env[65726]: DEBUG nova.virt.hardware [None req-d6fa7de0-71fb-41d9-ba04-00cd57a3d822 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 884.524551] env[65726]: DEBUG nova.virt.hardware [None req-d6fa7de0-71fb-41d9-ba04-00cd57a3d822 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 884.524686] env[65726]: DEBUG nova.virt.hardware [None req-d6fa7de0-71fb-41d9-ba04-00cd57a3d822 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 884.524975] env[65726]: DEBUG nova.virt.hardware [None req-d6fa7de0-71fb-41d9-ba04-00cd57a3d822 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 884.525165] env[65726]: DEBUG nova.virt.hardware [None req-d6fa7de0-71fb-41d9-ba04-00cd57a3d822 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 884.525333] env[65726]: DEBUG nova.virt.hardware [None req-d6fa7de0-71fb-41d9-ba04-00cd57a3d822 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 884.525493] env[65726]: DEBUG nova.virt.hardware [None req-d6fa7de0-71fb-41d9-ba04-00cd57a3d822 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 884.525672] env[65726]: DEBUG nova.virt.hardware [None req-d6fa7de0-71fb-41d9-ba04-00cd57a3d822 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 884.531261] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e95ebf5c-6b1b-4967-a508-26215c1057aa {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.547834] env[65726]: DEBUG oslo_vmware.api [None req-d6fa7de0-71fb-41d9-ba04-00cd57a3d822 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Waiting for the task: (returnval){ [ 884.547834] env[65726]: value = "task-5116132" [ 884.547834] env[65726]: _type = "Task" [ 884.547834] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.557187] env[65726]: DEBUG oslo_vmware.api [None req-d6fa7de0-71fb-41d9-ba04-00cd57a3d822 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5116132, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.637248] env[65726]: DEBUG oslo_concurrency.lockutils [None req-73b0a627-5d81-4bf2-833a-caff3534c1e3 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Lock "cf950ec3-9914-4eb9-99db-048b3969bb21" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 63.602s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 884.765175] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-fcafe3da-7a8e-4b08-9577-72f88a634bf5 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] [instance: c4cc4f85-cb35-4edc-a58b-adfee0ce1265] Creating linked-clone VM from snapshot {{(pid=65726) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 884.765558] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-5d6f7f0d-8281-41a8-9f16-7a6e743a0408 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.776721] env[65726]: DEBUG oslo_vmware.api [None req-fcafe3da-7a8e-4b08-9577-72f88a634bf5 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Waiting for the task: (returnval){ [ 884.776721] env[65726]: value = "task-5116133" [ 884.776721] env[65726]: _type = "Task" [ 884.776721] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.778281] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-172c2bf1-c043-4217-8650-efc42870269c tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: 895797e4-2941-44cd-aab1-67afa4fac02d] Volume attach. Driver type: vmdk {{(pid=65726) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 884.778632] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-172c2bf1-c043-4217-8650-efc42870269c tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: 895797e4-2941-44cd-aab1-67afa4fac02d] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-995185', 'volume_id': '3c728bbf-472e-4a76-86f7-6fd547117357', 'name': 'volume-3c728bbf-472e-4a76-86f7-6fd547117357', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '895797e4-2941-44cd-aab1-67afa4fac02d', 'attached_at': '', 'detached_at': '', 'volume_id': '3c728bbf-472e-4a76-86f7-6fd547117357', 'serial': '3c728bbf-472e-4a76-86f7-6fd547117357'} {{(pid=65726) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 884.779731] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99ff5732-fc0c-479c-a6d6-8c021357ce79 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.794528] env[65726]: DEBUG oslo_vmware.api [None req-fcafe3da-7a8e-4b08-9577-72f88a634bf5 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Task: {'id': task-5116133, 'name': CloneVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.813178] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3375d29b-64db-47f0-801b-361384839aeb {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.822851] env[65726]: DEBUG oslo_vmware.api [None req-5782ba22-8056-4347-9a97-4b1e5a35ed3b tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Task: {'id': task-5116131, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.844957] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-172c2bf1-c043-4217-8650-efc42870269c tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: 895797e4-2941-44cd-aab1-67afa4fac02d] Reconfiguring VM instance instance-00000028 to attach disk [datastore1] volume-3c728bbf-472e-4a76-86f7-6fd547117357/volume-3c728bbf-472e-4a76-86f7-6fd547117357.vmdk or device None with type thin {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 884.845398] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2563e6b2-5857-4261-8604-5fb0d6ac1029 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.868897] env[65726]: DEBUG oslo_vmware.api [None req-172c2bf1-c043-4217-8650-efc42870269c tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Waiting for the task: (returnval){ [ 884.868897] env[65726]: value = "task-5116134" [ 884.868897] env[65726]: _type = "Task" [ 884.868897] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.872245] env[65726]: DEBUG nova.network.neutron [None req-34f1738e-744a-4a46-b434-5db18d6e964a tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] [instance: 4ea2b70b-0cfe-4fbf-817a-baa028b8372a] Successfully updated port: a97348ff-93e0-442b-8a41-b56fa04e8cb6 {{(pid=65726) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 884.885792] env[65726]: DEBUG oslo_vmware.api [None req-172c2bf1-c043-4217-8650-efc42870269c tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Task: {'id': task-5116134, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.934068] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Instance 70375746-06dd-4710-9136-95e9e5759a0e has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65726) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 884.964786] env[65726]: INFO nova.compute.manager [None req-b20cc073-fafc-44fa-bcf9-9893dc98ebec tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04] Took 52.61 seconds to build instance. [ 885.059311] env[65726]: DEBUG oslo_vmware.api [None req-d6fa7de0-71fb-41d9-ba04-00cd57a3d822 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5116132, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.087185] env[65726]: DEBUG nova.compute.manager [req-448018b8-5d97-4f28-b838-d9829eb03a66 req-ecced4d6-3518-498c-87e2-d3a9b9a2801f service nova] [instance: 4ea2b70b-0cfe-4fbf-817a-baa028b8372a] Received event network-vif-plugged-a97348ff-93e0-442b-8a41-b56fa04e8cb6 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 885.087185] env[65726]: DEBUG oslo_concurrency.lockutils [req-448018b8-5d97-4f28-b838-d9829eb03a66 req-ecced4d6-3518-498c-87e2-d3a9b9a2801f service nova] Acquiring lock "4ea2b70b-0cfe-4fbf-817a-baa028b8372a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 885.087185] env[65726]: DEBUG oslo_concurrency.lockutils [req-448018b8-5d97-4f28-b838-d9829eb03a66 req-ecced4d6-3518-498c-87e2-d3a9b9a2801f service nova] Lock "4ea2b70b-0cfe-4fbf-817a-baa028b8372a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 885.087185] env[65726]: DEBUG oslo_concurrency.lockutils [req-448018b8-5d97-4f28-b838-d9829eb03a66 req-ecced4d6-3518-498c-87e2-d3a9b9a2801f service nova] Lock "4ea2b70b-0cfe-4fbf-817a-baa028b8372a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 885.087185] env[65726]: DEBUG nova.compute.manager [req-448018b8-5d97-4f28-b838-d9829eb03a66 req-ecced4d6-3518-498c-87e2-d3a9b9a2801f service nova] [instance: 4ea2b70b-0cfe-4fbf-817a-baa028b8372a] No waiting events found dispatching network-vif-plugged-a97348ff-93e0-442b-8a41-b56fa04e8cb6 {{(pid=65726) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 885.087486] env[65726]: WARNING nova.compute.manager [req-448018b8-5d97-4f28-b838-d9829eb03a66 req-ecced4d6-3518-498c-87e2-d3a9b9a2801f service nova] [instance: 4ea2b70b-0cfe-4fbf-817a-baa028b8372a] Received unexpected event network-vif-plugged-a97348ff-93e0-442b-8a41-b56fa04e8cb6 for instance with vm_state building and task_state spawning. [ 885.139987] env[65726]: DEBUG nova.compute.manager [None req-801b9b62-b5bd-4b5f-80fa-32388dd4cb88 tempest-ServerTagsTestJSON-59142761 tempest-ServerTagsTestJSON-59142761-project-member] [instance: 9029549c-1914-4bae-91e2-8812b79051ec] Starting instance... {{(pid=65726) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 885.291788] env[65726]: DEBUG oslo_vmware.api [None req-fcafe3da-7a8e-4b08-9577-72f88a634bf5 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Task: {'id': task-5116133, 'name': CloneVM_Task} progress is 93%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.320131] env[65726]: DEBUG oslo_vmware.api [None req-5782ba22-8056-4347-9a97-4b1e5a35ed3b tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Task: {'id': task-5116131, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.654519} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.320412] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-5782ba22-8056-4347-9a97-4b1e5a35ed3b tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] f2d9090c-988f-43f4-9c81-7aa718a3438a/f2d9090c-988f-43f4-9c81-7aa718a3438a.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 885.320689] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-5782ba22-8056-4347-9a97-4b1e5a35ed3b tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: f2d9090c-988f-43f4-9c81-7aa718a3438a] Extending root virtual disk to 1048576 {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 885.320962] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-79f21ea7-f761-4e4d-a69e-d7e126b663e1 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.329525] env[65726]: DEBUG oslo_vmware.api [None req-5782ba22-8056-4347-9a97-4b1e5a35ed3b tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Waiting for the task: (returnval){ [ 885.329525] env[65726]: value = "task-5116135" [ 885.329525] env[65726]: _type = "Task" [ 885.329525] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.339714] env[65726]: DEBUG oslo_vmware.api [None req-5782ba22-8056-4347-9a97-4b1e5a35ed3b tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Task: {'id': task-5116135, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.380493] env[65726]: DEBUG oslo_concurrency.lockutils [None req-34f1738e-744a-4a46-b434-5db18d6e964a tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Acquiring lock "refresh_cache-4ea2b70b-0cfe-4fbf-817a-baa028b8372a" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 885.381368] env[65726]: DEBUG oslo_concurrency.lockutils [None req-34f1738e-744a-4a46-b434-5db18d6e964a tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Acquired lock "refresh_cache-4ea2b70b-0cfe-4fbf-817a-baa028b8372a" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 885.381368] env[65726]: DEBUG nova.network.neutron [None req-34f1738e-744a-4a46-b434-5db18d6e964a tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] [instance: 4ea2b70b-0cfe-4fbf-817a-baa028b8372a] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 885.386398] env[65726]: DEBUG oslo_vmware.api [None req-172c2bf1-c043-4217-8650-efc42870269c tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Task: {'id': task-5116134, 'name': ReconfigVM_Task, 'duration_secs': 0.418638} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.387230] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-172c2bf1-c043-4217-8650-efc42870269c tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: 895797e4-2941-44cd-aab1-67afa4fac02d] Reconfigured VM instance instance-00000028 to attach disk [datastore1] volume-3c728bbf-472e-4a76-86f7-6fd547117357/volume-3c728bbf-472e-4a76-86f7-6fd547117357.vmdk or device None with type thin {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 885.393669] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f9fcc5b2-9380-408c-89dc-d3c8735adc85 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.412259] env[65726]: DEBUG oslo_vmware.api [None req-172c2bf1-c043-4217-8650-efc42870269c tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Waiting for the task: (returnval){ [ 885.412259] env[65726]: value = "task-5116136" [ 885.412259] env[65726]: _type = "Task" [ 885.412259] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.422457] env[65726]: DEBUG oslo_vmware.api [None req-172c2bf1-c043-4217-8650-efc42870269c tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Task: {'id': task-5116136, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.438104] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Instance 5a252ef1-93c9-4bff-842b-b64df2bc5d75 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65726) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 885.465978] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b20cc073-fafc-44fa-bcf9-9893dc98ebec tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Lock "2fe6fc5a-f5c7-4f8e-96df-4e621a252f04" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 62.362s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 885.470914] env[65726]: DEBUG nova.compute.manager [req-726a11a1-f2f3-4039-8d58-4fb1da33c40a req-7e85ebc6-106b-4e6a-a58d-5aeb24c6c82a service nova] [instance: cf950ec3-9914-4eb9-99db-048b3969bb21] Received event network-changed-58accf41-587f-4367-854e-fc571a6a3424 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 885.471269] env[65726]: DEBUG nova.compute.manager [req-726a11a1-f2f3-4039-8d58-4fb1da33c40a req-7e85ebc6-106b-4e6a-a58d-5aeb24c6c82a service nova] [instance: cf950ec3-9914-4eb9-99db-048b3969bb21] Refreshing instance network info cache due to event network-changed-58accf41-587f-4367-854e-fc571a6a3424. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 885.471759] env[65726]: DEBUG oslo_concurrency.lockutils [req-726a11a1-f2f3-4039-8d58-4fb1da33c40a req-7e85ebc6-106b-4e6a-a58d-5aeb24c6c82a service nova] Acquiring lock "refresh_cache-cf950ec3-9914-4eb9-99db-048b3969bb21" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 885.472105] env[65726]: DEBUG oslo_concurrency.lockutils [req-726a11a1-f2f3-4039-8d58-4fb1da33c40a req-7e85ebc6-106b-4e6a-a58d-5aeb24c6c82a service nova] Acquired lock "refresh_cache-cf950ec3-9914-4eb9-99db-048b3969bb21" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 885.472359] env[65726]: DEBUG nova.network.neutron [req-726a11a1-f2f3-4039-8d58-4fb1da33c40a req-7e85ebc6-106b-4e6a-a58d-5aeb24c6c82a service nova] [instance: cf950ec3-9914-4eb9-99db-048b3969bb21] Refreshing network info cache for port 58accf41-587f-4367-854e-fc571a6a3424 {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 885.562476] env[65726]: DEBUG oslo_vmware.api [None req-d6fa7de0-71fb-41d9-ba04-00cd57a3d822 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5116132, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.673140] env[65726]: DEBUG oslo_concurrency.lockutils [None req-801b9b62-b5bd-4b5f-80fa-32388dd4cb88 tempest-ServerTagsTestJSON-59142761 tempest-ServerTagsTestJSON-59142761-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 885.794022] env[65726]: DEBUG oslo_vmware.api [None req-fcafe3da-7a8e-4b08-9577-72f88a634bf5 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Task: {'id': task-5116133, 'name': CloneVM_Task} progress is 94%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.840597] env[65726]: DEBUG oslo_vmware.api [None req-5782ba22-8056-4347-9a97-4b1e5a35ed3b tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Task: {'id': task-5116135, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.087983} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.841701] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-5782ba22-8056-4347-9a97-4b1e5a35ed3b tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: f2d9090c-988f-43f4-9c81-7aa718a3438a] Extended root virtual disk {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 885.842275] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-136e64b0-2664-4bf5-bff3-5709bc87deec {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.860310] env[65726]: DEBUG oslo_concurrency.lockutils [None req-1cae8219-3b3b-4d11-bdce-863de7bcd2d6 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] Acquiring lock "2cbee359-a1fc-45c0-bcc5-a2c24ded9c69" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 885.860601] env[65726]: DEBUG oslo_concurrency.lockutils [None req-1cae8219-3b3b-4d11-bdce-863de7bcd2d6 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] Lock "2cbee359-a1fc-45c0-bcc5-a2c24ded9c69" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 885.860848] env[65726]: DEBUG oslo_concurrency.lockutils [None req-1cae8219-3b3b-4d11-bdce-863de7bcd2d6 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] Acquiring lock "2cbee359-a1fc-45c0-bcc5-a2c24ded9c69-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 885.861056] env[65726]: DEBUG oslo_concurrency.lockutils [None req-1cae8219-3b3b-4d11-bdce-863de7bcd2d6 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] Lock "2cbee359-a1fc-45c0-bcc5-a2c24ded9c69-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 885.861258] env[65726]: DEBUG oslo_concurrency.lockutils [None req-1cae8219-3b3b-4d11-bdce-863de7bcd2d6 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] Lock "2cbee359-a1fc-45c0-bcc5-a2c24ded9c69-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 885.872665] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-5782ba22-8056-4347-9a97-4b1e5a35ed3b tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: f2d9090c-988f-43f4-9c81-7aa718a3438a] Reconfiguring VM instance instance-00000011 to attach disk [datastore1] f2d9090c-988f-43f4-9c81-7aa718a3438a/f2d9090c-988f-43f4-9c81-7aa718a3438a.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 885.873566] env[65726]: INFO nova.compute.manager [None req-1cae8219-3b3b-4d11-bdce-863de7bcd2d6 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] [instance: 2cbee359-a1fc-45c0-bcc5-a2c24ded9c69] Terminating instance [ 885.875081] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bd4e7f86-c9c6-4e57-a484-f2565c852ef9 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.892960] env[65726]: WARNING openstack [None req-34f1738e-744a-4a46-b434-5db18d6e964a tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 885.893318] env[65726]: WARNING openstack [None req-34f1738e-744a-4a46-b434-5db18d6e964a tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 885.898827] env[65726]: DEBUG nova.network.neutron [None req-34f1738e-744a-4a46-b434-5db18d6e964a tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] [instance: 4ea2b70b-0cfe-4fbf-817a-baa028b8372a] Instance cache missing network info. {{(pid=65726) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 885.908327] env[65726]: DEBUG oslo_vmware.api [None req-5782ba22-8056-4347-9a97-4b1e5a35ed3b tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Waiting for the task: (returnval){ [ 885.908327] env[65726]: value = "task-5116137" [ 885.908327] env[65726]: _type = "Task" [ 885.908327] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.927916] env[65726]: DEBUG oslo_vmware.api [None req-172c2bf1-c043-4217-8650-efc42870269c tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Task: {'id': task-5116136, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.942854] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Instance 6f91b053-772a-4497-b29d-349b960c55eb has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65726) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 885.963836] env[65726]: WARNING openstack [None req-34f1738e-744a-4a46-b434-5db18d6e964a tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 885.964466] env[65726]: WARNING openstack [None req-34f1738e-744a-4a46-b434-5db18d6e964a tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 885.976259] env[65726]: WARNING openstack [req-726a11a1-f2f3-4039-8d58-4fb1da33c40a req-7e85ebc6-106b-4e6a-a58d-5aeb24c6c82a service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 885.976648] env[65726]: WARNING openstack [req-726a11a1-f2f3-4039-8d58-4fb1da33c40a req-7e85ebc6-106b-4e6a-a58d-5aeb24c6c82a service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 886.041451] env[65726]: WARNING openstack [None req-34f1738e-744a-4a46-b434-5db18d6e964a tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 886.042323] env[65726]: WARNING openstack [None req-34f1738e-744a-4a46-b434-5db18d6e964a tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 886.060404] env[65726]: DEBUG oslo_vmware.api [None req-d6fa7de0-71fb-41d9-ba04-00cd57a3d822 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5116132, 'name': ReconfigVM_Task, 'duration_secs': 1.359241} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.060755] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-d6fa7de0-71fb-41d9-ba04-00cd57a3d822 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 608cc0ea-3f6b-4b4f-83c1-01aa50999d63] Updating instance '608cc0ea-3f6b-4b4f-83c1-01aa50999d63' progress to 33 {{(pid=65726) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 886.118249] env[65726]: WARNING openstack [req-726a11a1-f2f3-4039-8d58-4fb1da33c40a req-7e85ebc6-106b-4e6a-a58d-5aeb24c6c82a service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 886.118703] env[65726]: WARNING openstack [req-726a11a1-f2f3-4039-8d58-4fb1da33c40a req-7e85ebc6-106b-4e6a-a58d-5aeb24c6c82a service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 886.128708] env[65726]: DEBUG nova.network.neutron [None req-34f1738e-744a-4a46-b434-5db18d6e964a tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] [instance: 4ea2b70b-0cfe-4fbf-817a-baa028b8372a] Updating instance_info_cache with network_info: [{"id": "a97348ff-93e0-442b-8a41-b56fa04e8cb6", "address": "fa:16:3e:f5:96:db", "network": {"id": "dd4bbaef-0a1e-4e00-b65a-f19e7a837cc1", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-255385230-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "55f93cdf51e84cf683879668c2acecf4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "04ccbc7a-cf8d-4ea2-8411-291a1e27df7b", "external-id": "nsx-vlan-transportzone-998", "segmentation_id": 998, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa97348ff-93", "ovs_interfaceid": "a97348ff-93e0-442b-8a41-b56fa04e8cb6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 886.206333] env[65726]: WARNING openstack [req-726a11a1-f2f3-4039-8d58-4fb1da33c40a req-7e85ebc6-106b-4e6a-a58d-5aeb24c6c82a service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 886.206832] env[65726]: WARNING openstack [req-726a11a1-f2f3-4039-8d58-4fb1da33c40a req-7e85ebc6-106b-4e6a-a58d-5aeb24c6c82a service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 886.292187] env[65726]: DEBUG oslo_vmware.api [None req-fcafe3da-7a8e-4b08-9577-72f88a634bf5 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Task: {'id': task-5116133, 'name': CloneVM_Task} progress is 94%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.321600] env[65726]: DEBUG nova.network.neutron [req-726a11a1-f2f3-4039-8d58-4fb1da33c40a req-7e85ebc6-106b-4e6a-a58d-5aeb24c6c82a service nova] [instance: cf950ec3-9914-4eb9-99db-048b3969bb21] Updated VIF entry in instance network info cache for port 58accf41-587f-4367-854e-fc571a6a3424. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 886.321600] env[65726]: DEBUG nova.network.neutron [req-726a11a1-f2f3-4039-8d58-4fb1da33c40a req-7e85ebc6-106b-4e6a-a58d-5aeb24c6c82a service nova] [instance: cf950ec3-9914-4eb9-99db-048b3969bb21] Updating instance_info_cache with network_info: [{"id": "58accf41-587f-4367-854e-fc571a6a3424", "address": "fa:16:3e:83:9d:0c", "network": {"id": "9d5ad167-99fb-4208-ab30-7053f35774e4", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-2082516762-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.132", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a8362295a62f4b51bae719a7ef5a4656", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dabbac20-1723-40ad-9da0-e53b28073651", "external-id": "nsx-vlan-transportzone-790", "segmentation_id": 790, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap58accf41-58", "ovs_interfaceid": "58accf41-587f-4367-854e-fc571a6a3424", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 886.402516] env[65726]: DEBUG oslo_concurrency.lockutils [None req-1cae8219-3b3b-4d11-bdce-863de7bcd2d6 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] Acquiring lock "refresh_cache-2cbee359-a1fc-45c0-bcc5-a2c24ded9c69" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 886.402758] env[65726]: DEBUG oslo_concurrency.lockutils [None req-1cae8219-3b3b-4d11-bdce-863de7bcd2d6 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] Acquired lock "refresh_cache-2cbee359-a1fc-45c0-bcc5-a2c24ded9c69" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 886.403036] env[65726]: DEBUG nova.network.neutron [None req-1cae8219-3b3b-4d11-bdce-863de7bcd2d6 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] [instance: 2cbee359-a1fc-45c0-bcc5-a2c24ded9c69] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 886.420960] env[65726]: DEBUG oslo_vmware.api [None req-5782ba22-8056-4347-9a97-4b1e5a35ed3b tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Task: {'id': task-5116137, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.427991] env[65726]: DEBUG oslo_vmware.api [None req-172c2bf1-c043-4217-8650-efc42870269c tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Task: {'id': task-5116136, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.447448] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Instance 3088ee5e-0d27-4058-a94b-2e04c3b52add has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65726) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 886.569129] env[65726]: DEBUG nova.virt.hardware [None req-d6fa7de0-71fb-41d9-ba04-00cd57a3d822 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 886.569129] env[65726]: DEBUG nova.virt.hardware [None req-d6fa7de0-71fb-41d9-ba04-00cd57a3d822 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 886.569129] env[65726]: DEBUG nova.virt.hardware [None req-d6fa7de0-71fb-41d9-ba04-00cd57a3d822 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 886.569129] env[65726]: DEBUG nova.virt.hardware [None req-d6fa7de0-71fb-41d9-ba04-00cd57a3d822 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 886.569488] env[65726]: DEBUG nova.virt.hardware [None req-d6fa7de0-71fb-41d9-ba04-00cd57a3d822 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 886.569488] env[65726]: DEBUG nova.virt.hardware [None req-d6fa7de0-71fb-41d9-ba04-00cd57a3d822 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 886.573265] env[65726]: DEBUG nova.virt.hardware [None req-d6fa7de0-71fb-41d9-ba04-00cd57a3d822 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 886.573468] env[65726]: DEBUG nova.virt.hardware [None req-d6fa7de0-71fb-41d9-ba04-00cd57a3d822 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 886.573587] env[65726]: DEBUG nova.virt.hardware [None req-d6fa7de0-71fb-41d9-ba04-00cd57a3d822 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 886.573790] env[65726]: DEBUG nova.virt.hardware [None req-d6fa7de0-71fb-41d9-ba04-00cd57a3d822 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 886.574026] env[65726]: DEBUG nova.virt.hardware [None req-d6fa7de0-71fb-41d9-ba04-00cd57a3d822 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 886.581429] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-d6fa7de0-71fb-41d9-ba04-00cd57a3d822 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 608cc0ea-3f6b-4b4f-83c1-01aa50999d63] Reconfiguring VM instance instance-0000001f to detach disk 2000 {{(pid=65726) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 886.581833] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e313b50a-40c0-4e60-ad60-bba6a6f3890c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.602843] env[65726]: DEBUG oslo_vmware.api [None req-d6fa7de0-71fb-41d9-ba04-00cd57a3d822 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Waiting for the task: (returnval){ [ 886.602843] env[65726]: value = "task-5116138" [ 886.602843] env[65726]: _type = "Task" [ 886.602843] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.612241] env[65726]: DEBUG oslo_vmware.api [None req-d6fa7de0-71fb-41d9-ba04-00cd57a3d822 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5116138, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.635061] env[65726]: DEBUG oslo_concurrency.lockutils [None req-34f1738e-744a-4a46-b434-5db18d6e964a tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Releasing lock "refresh_cache-4ea2b70b-0cfe-4fbf-817a-baa028b8372a" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 886.635061] env[65726]: DEBUG nova.compute.manager [None req-34f1738e-744a-4a46-b434-5db18d6e964a tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] [instance: 4ea2b70b-0cfe-4fbf-817a-baa028b8372a] Instance network_info: |[{"id": "a97348ff-93e0-442b-8a41-b56fa04e8cb6", "address": "fa:16:3e:f5:96:db", "network": {"id": "dd4bbaef-0a1e-4e00-b65a-f19e7a837cc1", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-255385230-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "55f93cdf51e84cf683879668c2acecf4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "04ccbc7a-cf8d-4ea2-8411-291a1e27df7b", "external-id": "nsx-vlan-transportzone-998", "segmentation_id": 998, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa97348ff-93", "ovs_interfaceid": "a97348ff-93e0-442b-8a41-b56fa04e8cb6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 886.635061] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-34f1738e-744a-4a46-b434-5db18d6e964a tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] [instance: 4ea2b70b-0cfe-4fbf-817a-baa028b8372a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f5:96:db', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '04ccbc7a-cf8d-4ea2-8411-291a1e27df7b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a97348ff-93e0-442b-8a41-b56fa04e8cb6', 'vif_model': 'vmxnet3'}] {{(pid=65726) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 886.641495] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-34f1738e-744a-4a46-b434-5db18d6e964a tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 886.641968] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4ea2b70b-0cfe-4fbf-817a-baa028b8372a] Creating VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 886.642429] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5f22fc4d-94ed-43f0-ae17-d67a7f5a65a0 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.665331] env[65726]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 886.665331] env[65726]: value = "task-5116139" [ 886.665331] env[65726]: _type = "Task" [ 886.665331] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.675476] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116139, 'name': CreateVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.792649] env[65726]: DEBUG oslo_vmware.api [None req-fcafe3da-7a8e-4b08-9577-72f88a634bf5 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Task: {'id': task-5116133, 'name': CloneVM_Task} progress is 94%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.827918] env[65726]: DEBUG oslo_concurrency.lockutils [req-726a11a1-f2f3-4039-8d58-4fb1da33c40a req-7e85ebc6-106b-4e6a-a58d-5aeb24c6c82a service nova] Releasing lock "refresh_cache-cf950ec3-9914-4eb9-99db-048b3969bb21" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 886.906283] env[65726]: WARNING openstack [None req-1cae8219-3b3b-4d11-bdce-863de7bcd2d6 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 886.906839] env[65726]: WARNING openstack [None req-1cae8219-3b3b-4d11-bdce-863de7bcd2d6 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 886.915601] env[65726]: DEBUG nova.network.neutron [None req-1cae8219-3b3b-4d11-bdce-863de7bcd2d6 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] [instance: 2cbee359-a1fc-45c0-bcc5-a2c24ded9c69] Instance cache missing network info. {{(pid=65726) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 886.935526] env[65726]: DEBUG oslo_vmware.api [None req-5782ba22-8056-4347-9a97-4b1e5a35ed3b tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Task: {'id': task-5116137, 'name': ReconfigVM_Task, 'duration_secs': 0.572916} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.941227] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-5782ba22-8056-4347-9a97-4b1e5a35ed3b tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: f2d9090c-988f-43f4-9c81-7aa718a3438a] Reconfigured VM instance instance-00000011 to attach disk [datastore1] f2d9090c-988f-43f4-9c81-7aa718a3438a/f2d9090c-988f-43f4-9c81-7aa718a3438a.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 886.942293] env[65726]: DEBUG oslo_vmware.api [None req-172c2bf1-c043-4217-8650-efc42870269c tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Task: {'id': task-5116136, 'name': ReconfigVM_Task, 'duration_secs': 1.190196} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.946092] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-de51d5ed-8aaf-48ca-ad05-7ee43306b0ae {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.950022] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-172c2bf1-c043-4217-8650-efc42870269c tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: 895797e4-2941-44cd-aab1-67afa4fac02d] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-995185', 'volume_id': '3c728bbf-472e-4a76-86f7-6fd547117357', 'name': 'volume-3c728bbf-472e-4a76-86f7-6fd547117357', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '895797e4-2941-44cd-aab1-67afa4fac02d', 'attached_at': '', 'detached_at': '', 'volume_id': '3c728bbf-472e-4a76-86f7-6fd547117357', 'serial': '3c728bbf-472e-4a76-86f7-6fd547117357'} {{(pid=65726) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 886.952887] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Instance dcde25df-87a5-47fb-94ce-334f68894e04 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65726) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 886.961572] env[65726]: DEBUG oslo_vmware.api [None req-5782ba22-8056-4347-9a97-4b1e5a35ed3b tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Waiting for the task: (returnval){ [ 886.961572] env[65726]: value = "task-5116140" [ 886.961572] env[65726]: _type = "Task" [ 886.961572] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.972753] env[65726]: DEBUG oslo_vmware.api [None req-5782ba22-8056-4347-9a97-4b1e5a35ed3b tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Task: {'id': task-5116140, 'name': Rename_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.034339] env[65726]: DEBUG nova.network.neutron [None req-1cae8219-3b3b-4d11-bdce-863de7bcd2d6 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] [instance: 2cbee359-a1fc-45c0-bcc5-a2c24ded9c69] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 887.114015] env[65726]: DEBUG oslo_vmware.api [None req-d6fa7de0-71fb-41d9-ba04-00cd57a3d822 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5116138, 'name': ReconfigVM_Task, 'duration_secs': 0.269812} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.115362] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-d6fa7de0-71fb-41d9-ba04-00cd57a3d822 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 608cc0ea-3f6b-4b4f-83c1-01aa50999d63] Reconfigured VM instance instance-0000001f to detach disk 2000 {{(pid=65726) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 887.116528] env[65726]: DEBUG nova.compute.manager [req-4b4bf7a2-1237-495e-bf17-b3e5292823c7 req-63177a1f-0f92-402d-9a55-489a99204283 service nova] [instance: 4ea2b70b-0cfe-4fbf-817a-baa028b8372a] Received event network-changed-a97348ff-93e0-442b-8a41-b56fa04e8cb6 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 887.116711] env[65726]: DEBUG nova.compute.manager [req-4b4bf7a2-1237-495e-bf17-b3e5292823c7 req-63177a1f-0f92-402d-9a55-489a99204283 service nova] [instance: 4ea2b70b-0cfe-4fbf-817a-baa028b8372a] Refreshing instance network info cache due to event network-changed-a97348ff-93e0-442b-8a41-b56fa04e8cb6. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 887.116916] env[65726]: DEBUG oslo_concurrency.lockutils [req-4b4bf7a2-1237-495e-bf17-b3e5292823c7 req-63177a1f-0f92-402d-9a55-489a99204283 service nova] Acquiring lock "refresh_cache-4ea2b70b-0cfe-4fbf-817a-baa028b8372a" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 887.117064] env[65726]: DEBUG oslo_concurrency.lockutils [req-4b4bf7a2-1237-495e-bf17-b3e5292823c7 req-63177a1f-0f92-402d-9a55-489a99204283 service nova] Acquired lock "refresh_cache-4ea2b70b-0cfe-4fbf-817a-baa028b8372a" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 887.117224] env[65726]: DEBUG nova.network.neutron [req-4b4bf7a2-1237-495e-bf17-b3e5292823c7 req-63177a1f-0f92-402d-9a55-489a99204283 service nova] [instance: 4ea2b70b-0cfe-4fbf-817a-baa028b8372a] Refreshing network info cache for port a97348ff-93e0-442b-8a41-b56fa04e8cb6 {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 887.118812] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61c8b1d1-735f-43ff-9959-44300913d6fb {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.143992] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-d6fa7de0-71fb-41d9-ba04-00cd57a3d822 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 608cc0ea-3f6b-4b4f-83c1-01aa50999d63] Reconfiguring VM instance instance-0000001f to attach disk [datastore1] 608cc0ea-3f6b-4b4f-83c1-01aa50999d63/608cc0ea-3f6b-4b4f-83c1-01aa50999d63.vmdk or device None with type thin {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 887.144624] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-617b147d-6cb4-44c0-a615-99336cad25eb {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.172028] env[65726]: DEBUG oslo_vmware.api [None req-d6fa7de0-71fb-41d9-ba04-00cd57a3d822 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Waiting for the task: (returnval){ [ 887.172028] env[65726]: value = "task-5116141" [ 887.172028] env[65726]: _type = "Task" [ 887.172028] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.179748] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116139, 'name': CreateVM_Task} progress is 99%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.186014] env[65726]: DEBUG oslo_vmware.api [None req-d6fa7de0-71fb-41d9-ba04-00cd57a3d822 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5116141, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.294690] env[65726]: DEBUG oslo_vmware.api [None req-fcafe3da-7a8e-4b08-9577-72f88a634bf5 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Task: {'id': task-5116133, 'name': CloneVM_Task, 'duration_secs': 2.259627} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.295055] env[65726]: INFO nova.virt.vmwareapi.vmops [None req-fcafe3da-7a8e-4b08-9577-72f88a634bf5 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] [instance: c4cc4f85-cb35-4edc-a58b-adfee0ce1265] Created linked-clone VM from snapshot [ 887.295855] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b05de866-e47d-4b04-bda4-b11b31de6949 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.303943] env[65726]: DEBUG nova.virt.vmwareapi.images [None req-fcafe3da-7a8e-4b08-9577-72f88a634bf5 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] [instance: c4cc4f85-cb35-4edc-a58b-adfee0ce1265] Uploading image 776f8431-e63e-4251-a79b-9d7edff7e53f {{(pid=65726) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 887.331724] env[65726]: DEBUG oslo_vmware.rw_handles [None req-fcafe3da-7a8e-4b08-9577-72f88a634bf5 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 887.331724] env[65726]: value = "vm-995188" [ 887.331724] env[65726]: _type = "VirtualMachine" [ 887.331724] env[65726]: }. {{(pid=65726) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 887.332286] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-8d5fc576-7559-4994-86d0-cdd42e16d9e0 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.341031] env[65726]: DEBUG oslo_vmware.rw_handles [None req-fcafe3da-7a8e-4b08-9577-72f88a634bf5 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Lease: (returnval){ [ 887.341031] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]527c882b-fad9-1b00-a1a1-145ee17181a3" [ 887.341031] env[65726]: _type = "HttpNfcLease" [ 887.341031] env[65726]: } obtained for exporting VM: (result){ [ 887.341031] env[65726]: value = "vm-995188" [ 887.341031] env[65726]: _type = "VirtualMachine" [ 887.341031] env[65726]: }. {{(pid=65726) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 887.341031] env[65726]: DEBUG oslo_vmware.api [None req-fcafe3da-7a8e-4b08-9577-72f88a634bf5 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Waiting for the lease: (returnval){ [ 887.341031] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]527c882b-fad9-1b00-a1a1-145ee17181a3" [ 887.341031] env[65726]: _type = "HttpNfcLease" [ 887.341031] env[65726]: } to be ready. {{(pid=65726) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 887.348682] env[65726]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 887.348682] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]527c882b-fad9-1b00-a1a1-145ee17181a3" [ 887.348682] env[65726]: _type = "HttpNfcLease" [ 887.348682] env[65726]: } is initializing. {{(pid=65726) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 887.458751] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Instance 9029549c-1914-4bae-91e2-8812b79051ec has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65726) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 887.459223] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Total usable vcpus: 48, total allocated vcpus: 22 {{(pid=65726) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1159}} [ 887.459264] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=4928MB phys_disk=100GB used_disk=23GB total_vcpus=48 used_vcpus=22 pci_stats=[] stats={'failed_builds': '0', 'num_instances': '20', 'num_vm_active': '14', 'num_task_None': '13', 'num_os_type_None': '20', 'num_proj_fae9ef0198a2468a9793374b76378abf': '2', 'io_workload': '5', 'num_task_rebuild_spawning': '2', 'num_proj_b794c99309e94c1181fc9d8eeeb84702': '4', 'num_vm_stopped': '1', 'num_task_resize_migrating': '1', 'num_proj_4eece77569624f90bf64e5c51974173f': '1', 'num_proj_a090d553766847e58e3231b966c92565': '1', 'num_proj_a8362295a62f4b51bae719a7ef5a4656': '3', 'num_vm_rescued': '1', 'num_vm_resized': '1', 'num_task_resize_reverting': '1', 'num_proj_55f93cdf51e84cf683879668c2acecf4': '2', 'num_proj_79d3e960f190439fae7c61d693e9356e': '1', 'num_proj_b8855e0d30be40b49e32a3eb0b6eef30': '1', 'num_proj_39067912d091464192516c08eb576696': '1', 'num_proj_e6d4f419846a4ed785988d5de5a0aebb': '1', 'num_proj_e3c48189fcee4ff29b2527fe1ad4fe0b': '1', 'num_task_image_snapshot': '1', 'num_proj_2772e93581124978b55c8054d2116232': '1', 'num_vm_building': '3', 'num_task_spawning': '2', 'num_proj_96149159e18e44f9bf3453e67681f224': '1'} {{(pid=65726) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1168}} [ 887.473940] env[65726]: DEBUG oslo_vmware.api [None req-5782ba22-8056-4347-9a97-4b1e5a35ed3b tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Task: {'id': task-5116140, 'name': Rename_Task, 'duration_secs': 0.218774} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.473940] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-5782ba22-8056-4347-9a97-4b1e5a35ed3b tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: f2d9090c-988f-43f4-9c81-7aa718a3438a] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 887.474171] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6eb87f82-cf2f-43af-9c7d-781eb1ce751e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.483851] env[65726]: DEBUG oslo_vmware.api [None req-5782ba22-8056-4347-9a97-4b1e5a35ed3b tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Waiting for the task: (returnval){ [ 887.483851] env[65726]: value = "task-5116143" [ 887.483851] env[65726]: _type = "Task" [ 887.483851] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.503415] env[65726]: DEBUG oslo_vmware.api [None req-5782ba22-8056-4347-9a97-4b1e5a35ed3b tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Task: {'id': task-5116143, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.512849] env[65726]: DEBUG nova.compute.manager [req-40f8f01e-dc74-4970-b289-61c136bd09bf req-69ac2d0b-0988-4692-bb91-a5c925053120 service nova] [instance: 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04] Received event network-changed-25b9d108-0b43-4459-b9db-7bd90a495bb3 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 887.513115] env[65726]: DEBUG nova.compute.manager [req-40f8f01e-dc74-4970-b289-61c136bd09bf req-69ac2d0b-0988-4692-bb91-a5c925053120 service nova] [instance: 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04] Refreshing instance network info cache due to event network-changed-25b9d108-0b43-4459-b9db-7bd90a495bb3. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 887.513471] env[65726]: DEBUG oslo_concurrency.lockutils [req-40f8f01e-dc74-4970-b289-61c136bd09bf req-69ac2d0b-0988-4692-bb91-a5c925053120 service nova] Acquiring lock "refresh_cache-2fe6fc5a-f5c7-4f8e-96df-4e621a252f04" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 887.513626] env[65726]: DEBUG oslo_concurrency.lockutils [req-40f8f01e-dc74-4970-b289-61c136bd09bf req-69ac2d0b-0988-4692-bb91-a5c925053120 service nova] Acquired lock "refresh_cache-2fe6fc5a-f5c7-4f8e-96df-4e621a252f04" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 887.514519] env[65726]: DEBUG nova.network.neutron [req-40f8f01e-dc74-4970-b289-61c136bd09bf req-69ac2d0b-0988-4692-bb91-a5c925053120 service nova] [instance: 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04] Refreshing network info cache for port 25b9d108-0b43-4459-b9db-7bd90a495bb3 {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 887.536144] env[65726]: DEBUG oslo_concurrency.lockutils [None req-1cae8219-3b3b-4d11-bdce-863de7bcd2d6 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] Releasing lock "refresh_cache-2cbee359-a1fc-45c0-bcc5-a2c24ded9c69" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 887.536977] env[65726]: DEBUG nova.compute.manager [None req-1cae8219-3b3b-4d11-bdce-863de7bcd2d6 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] [instance: 2cbee359-a1fc-45c0-bcc5-a2c24ded9c69] Start destroying the instance on the hypervisor. {{(pid=65726) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 887.537366] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-1cae8219-3b3b-4d11-bdce-863de7bcd2d6 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] [instance: 2cbee359-a1fc-45c0-bcc5-a2c24ded9c69] Destroying instance {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 887.539855] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa8e655f-cef6-4d14-8e8b-50b151ae0eb0 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.552461] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-1cae8219-3b3b-4d11-bdce-863de7bcd2d6 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] [instance: 2cbee359-a1fc-45c0-bcc5-a2c24ded9c69] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 887.553644] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e6ce8cdd-ef65-4f25-aa7d-86eaa74af844 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.562781] env[65726]: DEBUG oslo_vmware.api [None req-1cae8219-3b3b-4d11-bdce-863de7bcd2d6 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] Waiting for the task: (returnval){ [ 887.562781] env[65726]: value = "task-5116144" [ 887.562781] env[65726]: _type = "Task" [ 887.562781] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.575216] env[65726]: DEBUG oslo_vmware.api [None req-1cae8219-3b3b-4d11-bdce-863de7bcd2d6 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] Task: {'id': task-5116144, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.623172] env[65726]: WARNING openstack [req-4b4bf7a2-1237-495e-bf17-b3e5292823c7 req-63177a1f-0f92-402d-9a55-489a99204283 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 887.623394] env[65726]: WARNING openstack [req-4b4bf7a2-1237-495e-bf17-b3e5292823c7 req-63177a1f-0f92-402d-9a55-489a99204283 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 887.685590] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116139, 'name': CreateVM_Task} progress is 99%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.694296] env[65726]: DEBUG oslo_vmware.api [None req-d6fa7de0-71fb-41d9-ba04-00cd57a3d822 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5116141, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.769371] env[65726]: WARNING openstack [req-4b4bf7a2-1237-495e-bf17-b3e5292823c7 req-63177a1f-0f92-402d-9a55-489a99204283 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 887.769815] env[65726]: WARNING openstack [req-4b4bf7a2-1237-495e-bf17-b3e5292823c7 req-63177a1f-0f92-402d-9a55-489a99204283 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 887.845382] env[65726]: WARNING openstack [req-4b4bf7a2-1237-495e-bf17-b3e5292823c7 req-63177a1f-0f92-402d-9a55-489a99204283 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 887.846059] env[65726]: WARNING openstack [req-4b4bf7a2-1237-495e-bf17-b3e5292823c7 req-63177a1f-0f92-402d-9a55-489a99204283 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 887.864387] env[65726]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 887.864387] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]527c882b-fad9-1b00-a1a1-145ee17181a3" [ 887.864387] env[65726]: _type = "HttpNfcLease" [ 887.864387] env[65726]: } is ready. {{(pid=65726) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 887.864687] env[65726]: DEBUG oslo_vmware.rw_handles [None req-fcafe3da-7a8e-4b08-9577-72f88a634bf5 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 887.864687] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]527c882b-fad9-1b00-a1a1-145ee17181a3" [ 887.864687] env[65726]: _type = "HttpNfcLease" [ 887.864687] env[65726]: }. {{(pid=65726) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 887.865434] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1f79189-fae7-4888-bda4-b393358d8abe {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.875642] env[65726]: DEBUG oslo_vmware.rw_handles [None req-fcafe3da-7a8e-4b08-9577-72f88a634bf5 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52aff181-bf03-3dc8-d23e-16139f026815/disk-0.vmdk from lease info. {{(pid=65726) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 887.875859] env[65726]: DEBUG oslo_vmware.rw_handles [None req-fcafe3da-7a8e-4b08-9577-72f88a634bf5 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52aff181-bf03-3dc8-d23e-16139f026815/disk-0.vmdk for reading. {{(pid=65726) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 887.975742] env[65726]: DEBUG nova.network.neutron [req-4b4bf7a2-1237-495e-bf17-b3e5292823c7 req-63177a1f-0f92-402d-9a55-489a99204283 service nova] [instance: 4ea2b70b-0cfe-4fbf-817a-baa028b8372a] Updated VIF entry in instance network info cache for port a97348ff-93e0-442b-8a41-b56fa04e8cb6. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 887.976270] env[65726]: DEBUG nova.network.neutron [req-4b4bf7a2-1237-495e-bf17-b3e5292823c7 req-63177a1f-0f92-402d-9a55-489a99204283 service nova] [instance: 4ea2b70b-0cfe-4fbf-817a-baa028b8372a] Updating instance_info_cache with network_info: [{"id": "a97348ff-93e0-442b-8a41-b56fa04e8cb6", "address": "fa:16:3e:f5:96:db", "network": {"id": "dd4bbaef-0a1e-4e00-b65a-f19e7a837cc1", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-255385230-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "55f93cdf51e84cf683879668c2acecf4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "04ccbc7a-cf8d-4ea2-8411-291a1e27df7b", "external-id": "nsx-vlan-transportzone-998", "segmentation_id": 998, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa97348ff-93", "ovs_interfaceid": "a97348ff-93e0-442b-8a41-b56fa04e8cb6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 887.994306] env[65726]: DEBUG oslo_vmware.api [None req-5782ba22-8056-4347-9a97-4b1e5a35ed3b tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Task: {'id': task-5116143, 'name': PowerOnVM_Task} progress is 81%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.013114] env[65726]: DEBUG nova.objects.instance [None req-172c2bf1-c043-4217-8650-efc42870269c tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Lazy-loading 'flavor' on Instance uuid 895797e4-2941-44cd-aab1-67afa4fac02d {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 888.018843] env[65726]: WARNING openstack [req-40f8f01e-dc74-4970-b289-61c136bd09bf req-69ac2d0b-0988-4692-bb91-a5c925053120 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 888.018843] env[65726]: WARNING openstack [req-40f8f01e-dc74-4970-b289-61c136bd09bf req-69ac2d0b-0988-4692-bb91-a5c925053120 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 888.026906] env[65726]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-c166a8a6-8341-4d02-bb1a-5a7313e30d28 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.081714] env[65726]: DEBUG oslo_vmware.api [None req-1cae8219-3b3b-4d11-bdce-863de7bcd2d6 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] Task: {'id': task-5116144, 'name': PowerOffVM_Task, 'duration_secs': 0.218918} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.082100] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-1cae8219-3b3b-4d11-bdce-863de7bcd2d6 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] [instance: 2cbee359-a1fc-45c0-bcc5-a2c24ded9c69] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 888.082315] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-1cae8219-3b3b-4d11-bdce-863de7bcd2d6 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] [instance: 2cbee359-a1fc-45c0-bcc5-a2c24ded9c69] Unregistering the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 888.082579] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-376d6694-0e44-4d15-b8b9-494dcf702b5d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.115361] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c67103f-82b0-489f-852d-ec4febdf84d6 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.124956] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-1cae8219-3b3b-4d11-bdce-863de7bcd2d6 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] [instance: 2cbee359-a1fc-45c0-bcc5-a2c24ded9c69] Unregistered the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 888.125213] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-1cae8219-3b3b-4d11-bdce-863de7bcd2d6 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] [instance: 2cbee359-a1fc-45c0-bcc5-a2c24ded9c69] Deleting contents of the VM from datastore datastore2 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 888.125427] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-1cae8219-3b3b-4d11-bdce-863de7bcd2d6 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] Deleting the datastore file [datastore2] 2cbee359-a1fc-45c0-bcc5-a2c24ded9c69 {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 888.127096] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0ab1fde9-c0fa-4396-bcd8-76b44f01bb53 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.131820] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de1720f0-b0f3-498c-84ed-29f402cf99de {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.137065] env[65726]: DEBUG oslo_vmware.api [None req-1cae8219-3b3b-4d11-bdce-863de7bcd2d6 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] Waiting for the task: (returnval){ [ 888.137065] env[65726]: value = "task-5116146" [ 888.137065] env[65726]: _type = "Task" [ 888.137065] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.172986] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18ca9ac9-b5f2-45d0-b7d4-ce7a400f1f68 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.178810] env[65726]: WARNING openstack [req-40f8f01e-dc74-4970-b289-61c136bd09bf req-69ac2d0b-0988-4692-bb91-a5c925053120 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 888.179202] env[65726]: WARNING openstack [req-40f8f01e-dc74-4970-b289-61c136bd09bf req-69ac2d0b-0988-4692-bb91-a5c925053120 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 888.190068] env[65726]: DEBUG oslo_vmware.api [None req-1cae8219-3b3b-4d11-bdce-863de7bcd2d6 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] Task: {'id': task-5116146, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.198543] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116139, 'name': CreateVM_Task, 'duration_secs': 1.385736} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.203024] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4ea2b70b-0cfe-4fbf-817a-baa028b8372a] Created VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 888.203024] env[65726]: WARNING openstack [None req-34f1738e-744a-4a46-b434-5db18d6e964a tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 888.203024] env[65726]: WARNING openstack [None req-34f1738e-744a-4a46-b434-5db18d6e964a tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 888.208254] env[65726]: DEBUG oslo_concurrency.lockutils [None req-34f1738e-744a-4a46-b434-5db18d6e964a tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 888.209746] env[65726]: DEBUG oslo_concurrency.lockutils [None req-34f1738e-744a-4a46-b434-5db18d6e964a tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 888.209746] env[65726]: DEBUG oslo_concurrency.lockutils [None req-34f1738e-744a-4a46-b434-5db18d6e964a tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 888.210224] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4495541-0a9f-44a4-9957-f76d8575910c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.218299] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9e31a3ec-e623-4c0b-851c-55f1fc420f3f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.220795] env[65726]: DEBUG oslo_vmware.api [None req-d6fa7de0-71fb-41d9-ba04-00cd57a3d822 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5116141, 'name': ReconfigVM_Task, 'duration_secs': 0.572317} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.221505] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-d6fa7de0-71fb-41d9-ba04-00cd57a3d822 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 608cc0ea-3f6b-4b4f-83c1-01aa50999d63] Reconfigured VM instance instance-0000001f to attach disk [datastore1] 608cc0ea-3f6b-4b4f-83c1-01aa50999d63/608cc0ea-3f6b-4b4f-83c1-01aa50999d63.vmdk or device None with type thin {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 888.221963] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-d6fa7de0-71fb-41d9-ba04-00cd57a3d822 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 608cc0ea-3f6b-4b4f-83c1-01aa50999d63] Updating instance '608cc0ea-3f6b-4b4f-83c1-01aa50999d63' progress to 50 {{(pid=65726) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 888.238269] env[65726]: DEBUG nova.compute.provider_tree [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 888.242950] env[65726]: DEBUG oslo_vmware.api [None req-34f1738e-744a-4a46-b434-5db18d6e964a tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Waiting for the task: (returnval){ [ 888.242950] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52976bf4-a9eb-d85e-f419-cc325d88f4f9" [ 888.242950] env[65726]: _type = "Task" [ 888.242950] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.257238] env[65726]: DEBUG oslo_vmware.api [None req-34f1738e-744a-4a46-b434-5db18d6e964a tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52976bf4-a9eb-d85e-f419-cc325d88f4f9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.309970] env[65726]: WARNING openstack [req-40f8f01e-dc74-4970-b289-61c136bd09bf req-69ac2d0b-0988-4692-bb91-a5c925053120 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 888.310517] env[65726]: WARNING openstack [req-40f8f01e-dc74-4970-b289-61c136bd09bf req-69ac2d0b-0988-4692-bb91-a5c925053120 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 888.413321] env[65726]: DEBUG nova.network.neutron [req-40f8f01e-dc74-4970-b289-61c136bd09bf req-69ac2d0b-0988-4692-bb91-a5c925053120 service nova] [instance: 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04] Updated VIF entry in instance network info cache for port 25b9d108-0b43-4459-b9db-7bd90a495bb3. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 888.413763] env[65726]: DEBUG nova.network.neutron [req-40f8f01e-dc74-4970-b289-61c136bd09bf req-69ac2d0b-0988-4692-bb91-a5c925053120 service nova] [instance: 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04] Updating instance_info_cache with network_info: [{"id": "25b9d108-0b43-4459-b9db-7bd90a495bb3", "address": "fa:16:3e:5a:1e:6b", "network": {"id": "eb3278b7-c190-441b-84d2-305f9b896382", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1754571022-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.149", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "96149159e18e44f9bf3453e67681f224", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f77ff7a1-209c-4f3f-b2a0-fd817741e739", "external-id": "nsx-vlan-transportzone-935", "segmentation_id": 935, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap25b9d108-0b", "ovs_interfaceid": "25b9d108-0b43-4459-b9db-7bd90a495bb3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 888.483349] env[65726]: DEBUG oslo_concurrency.lockutils [req-4b4bf7a2-1237-495e-bf17-b3e5292823c7 req-63177a1f-0f92-402d-9a55-489a99204283 service nova] Releasing lock "refresh_cache-4ea2b70b-0cfe-4fbf-817a-baa028b8372a" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 888.499721] env[65726]: DEBUG oslo_vmware.api [None req-5782ba22-8056-4347-9a97-4b1e5a35ed3b tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Task: {'id': task-5116143, 'name': PowerOnVM_Task, 'duration_secs': 0.852741} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.499721] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-5782ba22-8056-4347-9a97-4b1e5a35ed3b tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: f2d9090c-988f-43f4-9c81-7aa718a3438a] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 888.499875] env[65726]: DEBUG nova.compute.manager [None req-5782ba22-8056-4347-9a97-4b1e5a35ed3b tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: f2d9090c-988f-43f4-9c81-7aa718a3438a] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 888.500873] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a24ec42-f773-49f3-813d-e2ad1a596bd3 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.530190] env[65726]: DEBUG oslo_concurrency.lockutils [None req-172c2bf1-c043-4217-8650-efc42870269c tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Lock "895797e4-2941-44cd-aab1-67afa4fac02d" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 8.409s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 888.653886] env[65726]: DEBUG oslo_vmware.api [None req-1cae8219-3b3b-4d11-bdce-863de7bcd2d6 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] Task: {'id': task-5116146, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.104758} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.656071] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-1cae8219-3b3b-4d11-bdce-863de7bcd2d6 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] Deleted the datastore file {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 888.656071] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-1cae8219-3b3b-4d11-bdce-863de7bcd2d6 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] [instance: 2cbee359-a1fc-45c0-bcc5-a2c24ded9c69] Deleted contents of the VM from datastore datastore2 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 888.656071] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-1cae8219-3b3b-4d11-bdce-863de7bcd2d6 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] [instance: 2cbee359-a1fc-45c0-bcc5-a2c24ded9c69] Instance destroyed {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 888.656071] env[65726]: INFO nova.compute.manager [None req-1cae8219-3b3b-4d11-bdce-863de7bcd2d6 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] [instance: 2cbee359-a1fc-45c0-bcc5-a2c24ded9c69] Took 1.12 seconds to destroy the instance on the hypervisor. [ 888.656071] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-1cae8219-3b3b-4d11-bdce-863de7bcd2d6 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 888.656071] env[65726]: DEBUG nova.compute.manager [-] [instance: 2cbee359-a1fc-45c0-bcc5-a2c24ded9c69] Deallocating network for instance {{(pid=65726) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 888.656071] env[65726]: DEBUG nova.network.neutron [-] [instance: 2cbee359-a1fc-45c0-bcc5-a2c24ded9c69] deallocate_for_instance() {{(pid=65726) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 888.657377] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 888.658314] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 888.686844] env[65726]: DEBUG nova.network.neutron [-] [instance: 2cbee359-a1fc-45c0-bcc5-a2c24ded9c69] Instance cache missing network info. {{(pid=65726) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 888.687614] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 888.688492] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 888.744407] env[65726]: DEBUG nova.scheduler.client.report [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 888.749045] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb2afc90-7d86-48b0-ad7c-28ac19fb3884 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.772857] env[65726]: DEBUG oslo_vmware.api [None req-34f1738e-744a-4a46-b434-5db18d6e964a tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52976bf4-a9eb-d85e-f419-cc325d88f4f9, 'name': SearchDatastore_Task, 'duration_secs': 0.015048} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.787159] env[65726]: DEBUG oslo_concurrency.lockutils [None req-34f1738e-744a-4a46-b434-5db18d6e964a tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 888.787518] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-34f1738e-744a-4a46-b434-5db18d6e964a tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] [instance: 4ea2b70b-0cfe-4fbf-817a-baa028b8372a] Processing image b52362a3-ee8a-4cbf-b06f-513b0cc8f95c {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 888.787828] env[65726]: DEBUG oslo_concurrency.lockutils [None req-34f1738e-744a-4a46-b434-5db18d6e964a tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 888.788042] env[65726]: DEBUG oslo_concurrency.lockutils [None req-34f1738e-744a-4a46-b434-5db18d6e964a tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 888.788266] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-34f1738e-744a-4a46-b434-5db18d6e964a tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 888.789145] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-cc651bcf-0548-44b4-aa92-8ec4ac807857 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.792120] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94d6f1d2-e62e-4fbc-9f52-6f299f03bcc4 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.816295] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-d6fa7de0-71fb-41d9-ba04-00cd57a3d822 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 608cc0ea-3f6b-4b4f-83c1-01aa50999d63] Updating instance '608cc0ea-3f6b-4b4f-83c1-01aa50999d63' progress to 67 {{(pid=65726) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 888.821817] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-34f1738e-744a-4a46-b434-5db18d6e964a tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 888.821961] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-34f1738e-744a-4a46-b434-5db18d6e964a tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65726) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 888.823134] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1b6cbe79-8c77-45b0-ba82-661c49dae16d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.829965] env[65726]: DEBUG oslo_vmware.api [None req-34f1738e-744a-4a46-b434-5db18d6e964a tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Waiting for the task: (returnval){ [ 888.829965] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]521ce9c7-b1ac-bf76-2ae6-51f1a09d66ce" [ 888.829965] env[65726]: _type = "Task" [ 888.829965] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.840133] env[65726]: DEBUG oslo_vmware.api [None req-34f1738e-744a-4a46-b434-5db18d6e964a tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]521ce9c7-b1ac-bf76-2ae6-51f1a09d66ce, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.916941] env[65726]: DEBUG oslo_concurrency.lockutils [req-40f8f01e-dc74-4970-b289-61c136bd09bf req-69ac2d0b-0988-4692-bb91-a5c925053120 service nova] Releasing lock "refresh_cache-2fe6fc5a-f5c7-4f8e-96df-4e621a252f04" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 889.019519] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5782ba22-8056-4347-9a97-4b1e5a35ed3b tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 889.197216] env[65726]: DEBUG nova.network.neutron [-] [instance: 2cbee359-a1fc-45c0-bcc5-a2c24ded9c69] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 889.257357] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=65726) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1097}} [ 889.260890] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 6.918s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 889.261590] env[65726]: DEBUG oslo_concurrency.lockutils [None req-63a3f335-a9b3-4e24-b6da-ba230f1ce119 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 49.044s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 889.261889] env[65726]: DEBUG oslo_concurrency.lockutils [None req-63a3f335-a9b3-4e24-b6da-ba230f1ce119 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 889.266303] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a1a4c202-18b9-4cfc-8067-d677c8415fff tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 47.170s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 889.268035] env[65726]: INFO nova.compute.claims [None req-a1a4c202-18b9-4cfc-8067-d677c8415fff tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] [instance: 815ee20b-cb07-4514-bd43-1b74972508a5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 889.274329] env[65726]: DEBUG oslo_service.periodic_task [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=65726) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 889.274329] env[65726]: DEBUG nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Cleaning up deleted instances {{(pid=65726) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11932}} [ 889.306476] env[65726]: INFO nova.scheduler.client.report [None req-63a3f335-a9b3-4e24-b6da-ba230f1ce119 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Deleted allocations for instance d7905c4f-1ba3-43b3-b4b3-aebacd2f9cc3 [ 889.325378] env[65726]: WARNING neutronclient.v2_0.client [None req-d6fa7de0-71fb-41d9-ba04-00cd57a3d822 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 889.345800] env[65726]: DEBUG oslo_vmware.api [None req-34f1738e-744a-4a46-b434-5db18d6e964a tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]521ce9c7-b1ac-bf76-2ae6-51f1a09d66ce, 'name': SearchDatastore_Task, 'duration_secs': 0.011629} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.346923] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cbf530ec-f77d-4498-bcbf-e97c06af8e4c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.354382] env[65726]: DEBUG oslo_vmware.api [None req-34f1738e-744a-4a46-b434-5db18d6e964a tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Waiting for the task: (returnval){ [ 889.354382] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]529b5a63-c682-f21d-0b28-92fc9ba82acc" [ 889.354382] env[65726]: _type = "Task" [ 889.354382] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.373046] env[65726]: DEBUG oslo_vmware.api [None req-34f1738e-744a-4a46-b434-5db18d6e964a tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]529b5a63-c682-f21d-0b28-92fc9ba82acc, 'name': SearchDatastore_Task, 'duration_secs': 0.012276} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.374952] env[65726]: DEBUG oslo_concurrency.lockutils [None req-34f1738e-744a-4a46-b434-5db18d6e964a tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 889.375309] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-34f1738e-744a-4a46-b434-5db18d6e964a tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] 4ea2b70b-0cfe-4fbf-817a-baa028b8372a/4ea2b70b-0cfe-4fbf-817a-baa028b8372a.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 889.375933] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5fffaa25-e734-4cbb-a798-1de43c8bacd1 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.385692] env[65726]: DEBUG oslo_vmware.api [None req-34f1738e-744a-4a46-b434-5db18d6e964a tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Waiting for the task: (returnval){ [ 889.385692] env[65726]: value = "task-5116147" [ 889.385692] env[65726]: _type = "Task" [ 889.385692] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.398185] env[65726]: DEBUG oslo_vmware.api [None req-34f1738e-744a-4a46-b434-5db18d6e964a tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Task: {'id': task-5116147, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.400394] env[65726]: DEBUG nova.network.neutron [None req-d6fa7de0-71fb-41d9-ba04-00cd57a3d822 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 608cc0ea-3f6b-4b4f-83c1-01aa50999d63] Port e9f7132a-8d5a-47fe-afc6-d45bf3687ea4 binding to destination host cpu-1 is already ACTIVE {{(pid=65726) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3236}} [ 889.700584] env[65726]: INFO nova.compute.manager [-] [instance: 2cbee359-a1fc-45c0-bcc5-a2c24ded9c69] Took 1.04 seconds to deallocate network for instance. [ 889.792759] env[65726]: DEBUG nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] There are 39 instances to clean {{(pid=65726) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11941}} [ 889.793254] env[65726]: DEBUG nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] [instance: 018c8ac5-66c4-4a9a-ab45-85e84f92b4fe] Instance has had 0 of 5 cleanup attempts {{(pid=65726) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 889.818939] env[65726]: DEBUG oslo_concurrency.lockutils [None req-63a3f335-a9b3-4e24-b6da-ba230f1ce119 tempest-FloatingIPsAssociationTestJSON-1410455798 tempest-FloatingIPsAssociationTestJSON-1410455798-project-member] Lock "d7905c4f-1ba3-43b3-b4b3-aebacd2f9cc3" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 53.103s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 889.899355] env[65726]: DEBUG oslo_vmware.api [None req-34f1738e-744a-4a46-b434-5db18d6e964a tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Task: {'id': task-5116147, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.057999] env[65726]: DEBUG oslo_concurrency.lockutils [None req-294000af-a13c-4aa5-b362-e66461f19340 tempest-ServersV294TestFqdnHostnames-819811677 tempest-ServersV294TestFqdnHostnames-819811677-project-member] Acquiring lock "c8be64a3-1bfa-41a3-b9be-68a7e6e052e0" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 890.058407] env[65726]: DEBUG oslo_concurrency.lockutils [None req-294000af-a13c-4aa5-b362-e66461f19340 tempest-ServersV294TestFqdnHostnames-819811677 tempest-ServersV294TestFqdnHostnames-819811677-project-member] Lock "c8be64a3-1bfa-41a3-b9be-68a7e6e052e0" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 890.058700] env[65726]: DEBUG oslo_concurrency.lockutils [None req-294000af-a13c-4aa5-b362-e66461f19340 tempest-ServersV294TestFqdnHostnames-819811677 tempest-ServersV294TestFqdnHostnames-819811677-project-member] Acquiring lock "c8be64a3-1bfa-41a3-b9be-68a7e6e052e0-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 890.059010] env[65726]: DEBUG oslo_concurrency.lockutils [None req-294000af-a13c-4aa5-b362-e66461f19340 tempest-ServersV294TestFqdnHostnames-819811677 tempest-ServersV294TestFqdnHostnames-819811677-project-member] Lock "c8be64a3-1bfa-41a3-b9be-68a7e6e052e0-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 890.059243] env[65726]: DEBUG oslo_concurrency.lockutils [None req-294000af-a13c-4aa5-b362-e66461f19340 tempest-ServersV294TestFqdnHostnames-819811677 tempest-ServersV294TestFqdnHostnames-819811677-project-member] Lock "c8be64a3-1bfa-41a3-b9be-68a7e6e052e0-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 890.063882] env[65726]: INFO nova.compute.manager [None req-294000af-a13c-4aa5-b362-e66461f19340 tempest-ServersV294TestFqdnHostnames-819811677 tempest-ServersV294TestFqdnHostnames-819811677-project-member] [instance: c8be64a3-1bfa-41a3-b9be-68a7e6e052e0] Terminating instance [ 890.209761] env[65726]: DEBUG oslo_concurrency.lockutils [None req-1cae8219-3b3b-4d11-bdce-863de7bcd2d6 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 890.303344] env[65726]: DEBUG nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] [instance: 1bd26aef-995e-43b6-af9f-077e878c8c44] Instance has had 0 of 5 cleanup attempts {{(pid=65726) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 890.400743] env[65726]: DEBUG oslo_vmware.api [None req-34f1738e-744a-4a46-b434-5db18d6e964a tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Task: {'id': task-5116147, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.52102} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.401099] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-34f1738e-744a-4a46-b434-5db18d6e964a tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] 4ea2b70b-0cfe-4fbf-817a-baa028b8372a/4ea2b70b-0cfe-4fbf-817a-baa028b8372a.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 890.401328] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-34f1738e-744a-4a46-b434-5db18d6e964a tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] [instance: 4ea2b70b-0cfe-4fbf-817a-baa028b8372a] Extending root virtual disk to 1048576 {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 890.401608] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-8fee5a44-6b57-4caf-8acd-bd33642999c7 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.410887] env[65726]: DEBUG oslo_vmware.api [None req-34f1738e-744a-4a46-b434-5db18d6e964a tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Waiting for the task: (returnval){ [ 890.410887] env[65726]: value = "task-5116148" [ 890.410887] env[65726]: _type = "Task" [ 890.410887] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.438728] env[65726]: DEBUG oslo_concurrency.lockutils [None req-d6fa7de0-71fb-41d9-ba04-00cd57a3d822 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Acquiring lock "608cc0ea-3f6b-4b4f-83c1-01aa50999d63-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 890.439021] env[65726]: DEBUG oslo_concurrency.lockutils [None req-d6fa7de0-71fb-41d9-ba04-00cd57a3d822 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Lock "608cc0ea-3f6b-4b4f-83c1-01aa50999d63-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 890.439237] env[65726]: DEBUG oslo_concurrency.lockutils [None req-d6fa7de0-71fb-41d9-ba04-00cd57a3d822 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Lock "608cc0ea-3f6b-4b4f-83c1-01aa50999d63-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 890.449076] env[65726]: DEBUG oslo_vmware.api [None req-34f1738e-744a-4a46-b434-5db18d6e964a tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Task: {'id': task-5116148, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.571314] env[65726]: DEBUG nova.compute.manager [None req-294000af-a13c-4aa5-b362-e66461f19340 tempest-ServersV294TestFqdnHostnames-819811677 tempest-ServersV294TestFqdnHostnames-819811677-project-member] [instance: c8be64a3-1bfa-41a3-b9be-68a7e6e052e0] Start destroying the instance on the hypervisor. {{(pid=65726) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 890.571552] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-294000af-a13c-4aa5-b362-e66461f19340 tempest-ServersV294TestFqdnHostnames-819811677 tempest-ServersV294TestFqdnHostnames-819811677-project-member] [instance: c8be64a3-1bfa-41a3-b9be-68a7e6e052e0] Destroying instance {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 890.572824] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c54a61d5-4cd6-4357-acd5-1eaad30b6147 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.582252] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-294000af-a13c-4aa5-b362-e66461f19340 tempest-ServersV294TestFqdnHostnames-819811677 tempest-ServersV294TestFqdnHostnames-819811677-project-member] [instance: c8be64a3-1bfa-41a3-b9be-68a7e6e052e0] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 890.582562] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-bd991b4f-f0f5-45b2-aa18-31cc1364ae04 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.594445] env[65726]: DEBUG oslo_vmware.api [None req-294000af-a13c-4aa5-b362-e66461f19340 tempest-ServersV294TestFqdnHostnames-819811677 tempest-ServersV294TestFqdnHostnames-819811677-project-member] Waiting for the task: (returnval){ [ 890.594445] env[65726]: value = "task-5116149" [ 890.594445] env[65726]: _type = "Task" [ 890.594445] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.608223] env[65726]: DEBUG oslo_vmware.api [None req-294000af-a13c-4aa5-b362-e66461f19340 tempest-ServersV294TestFqdnHostnames-819811677 tempest-ServersV294TestFqdnHostnames-819811677-project-member] Task: {'id': task-5116149, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.809637] env[65726]: DEBUG nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] [instance: 621b899a-e9ec-4fc0-a574-4a08cecff6d9] Instance has had 0 of 5 cleanup attempts {{(pid=65726) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 890.831973] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-360fb480-890f-4291-bfb5-693b3b72098f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.843070] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d1f3e8f-2a7d-4d5f-ac9b-0167b05b0410 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.888115] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b2e1bee-a91f-47ea-8819-1ce7629d939f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.898762] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1f08305-5908-4479-9f00-9807f003a386 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.917164] env[65726]: DEBUG nova.compute.provider_tree [None req-a1a4c202-18b9-4cfc-8067-d677c8415fff tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 890.927361] env[65726]: DEBUG oslo_vmware.api [None req-34f1738e-744a-4a46-b434-5db18d6e964a tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Task: {'id': task-5116148, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.073469} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.927664] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-34f1738e-744a-4a46-b434-5db18d6e964a tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] [instance: 4ea2b70b-0cfe-4fbf-817a-baa028b8372a] Extended root virtual disk {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 890.928511] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4915344b-f277-4735-9a5b-f212d59830d3 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.958183] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-34f1738e-744a-4a46-b434-5db18d6e964a tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] [instance: 4ea2b70b-0cfe-4fbf-817a-baa028b8372a] Reconfiguring VM instance instance-00000040 to attach disk [datastore1] 4ea2b70b-0cfe-4fbf-817a-baa028b8372a/4ea2b70b-0cfe-4fbf-817a-baa028b8372a.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 890.959389] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bcdc2763-a582-4057-888d-322d8fa0f409 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.987231] env[65726]: DEBUG oslo_vmware.api [None req-34f1738e-744a-4a46-b434-5db18d6e964a tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Waiting for the task: (returnval){ [ 890.987231] env[65726]: value = "task-5116150" [ 890.987231] env[65726]: _type = "Task" [ 890.987231] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.999011] env[65726]: DEBUG oslo_vmware.api [None req-34f1738e-744a-4a46-b434-5db18d6e964a tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Task: {'id': task-5116150, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.010623] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9a0984e5-cc6d-419e-a851-17965236c39e tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Acquiring lock "060646cc-f0fb-4e6c-9f67-7b2daa6b6fd6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 891.010903] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9a0984e5-cc6d-419e-a851-17965236c39e tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Lock "060646cc-f0fb-4e6c-9f67-7b2daa6b6fd6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 891.108414] env[65726]: DEBUG oslo_vmware.api [None req-294000af-a13c-4aa5-b362-e66461f19340 tempest-ServersV294TestFqdnHostnames-819811677 tempest-ServersV294TestFqdnHostnames-819811677-project-member] Task: {'id': task-5116149, 'name': PowerOffVM_Task, 'duration_secs': 0.312559} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.108696] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-294000af-a13c-4aa5-b362-e66461f19340 tempest-ServersV294TestFqdnHostnames-819811677 tempest-ServersV294TestFqdnHostnames-819811677-project-member] [instance: c8be64a3-1bfa-41a3-b9be-68a7e6e052e0] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 891.111206] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-294000af-a13c-4aa5-b362-e66461f19340 tempest-ServersV294TestFqdnHostnames-819811677 tempest-ServersV294TestFqdnHostnames-819811677-project-member] [instance: c8be64a3-1bfa-41a3-b9be-68a7e6e052e0] Unregistering the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 891.111206] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8d8b594e-d6ef-41dd-bf37-779057bb504d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.187972] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-294000af-a13c-4aa5-b362-e66461f19340 tempest-ServersV294TestFqdnHostnames-819811677 tempest-ServersV294TestFqdnHostnames-819811677-project-member] [instance: c8be64a3-1bfa-41a3-b9be-68a7e6e052e0] Unregistered the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 891.188320] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-294000af-a13c-4aa5-b362-e66461f19340 tempest-ServersV294TestFqdnHostnames-819811677 tempest-ServersV294TestFqdnHostnames-819811677-project-member] [instance: c8be64a3-1bfa-41a3-b9be-68a7e6e052e0] Deleting contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 891.188588] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-294000af-a13c-4aa5-b362-e66461f19340 tempest-ServersV294TestFqdnHostnames-819811677 tempest-ServersV294TestFqdnHostnames-819811677-project-member] Deleting the datastore file [datastore1] c8be64a3-1bfa-41a3-b9be-68a7e6e052e0 {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 891.188977] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b15f74ed-49e8-4819-92e6-169318b14698 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.199644] env[65726]: DEBUG oslo_vmware.api [None req-294000af-a13c-4aa5-b362-e66461f19340 tempest-ServersV294TestFqdnHostnames-819811677 tempest-ServersV294TestFqdnHostnames-819811677-project-member] Waiting for the task: (returnval){ [ 891.199644] env[65726]: value = "task-5116152" [ 891.199644] env[65726]: _type = "Task" [ 891.199644] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.212952] env[65726]: DEBUG oslo_vmware.api [None req-294000af-a13c-4aa5-b362-e66461f19340 tempest-ServersV294TestFqdnHostnames-819811677 tempest-ServersV294TestFqdnHostnames-819811677-project-member] Task: {'id': task-5116152, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.274251] env[65726]: DEBUG oslo_concurrency.lockutils [None req-73ce0e73-1afb-4566-995f-f51b4749346e tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Acquiring lock "cb8855b6-589d-4863-b86f-7fb4e30fb29e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 891.274664] env[65726]: DEBUG oslo_concurrency.lockutils [None req-73ce0e73-1afb-4566-995f-f51b4749346e tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Lock "cb8855b6-589d-4863-b86f-7fb4e30fb29e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 891.274924] env[65726]: DEBUG oslo_concurrency.lockutils [None req-73ce0e73-1afb-4566-995f-f51b4749346e tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Acquiring lock "cb8855b6-589d-4863-b86f-7fb4e30fb29e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 891.275149] env[65726]: DEBUG oslo_concurrency.lockutils [None req-73ce0e73-1afb-4566-995f-f51b4749346e tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Lock "cb8855b6-589d-4863-b86f-7fb4e30fb29e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 891.275343] env[65726]: DEBUG oslo_concurrency.lockutils [None req-73ce0e73-1afb-4566-995f-f51b4749346e tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Lock "cb8855b6-589d-4863-b86f-7fb4e30fb29e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 891.278085] env[65726]: INFO nova.compute.manager [None req-73ce0e73-1afb-4566-995f-f51b4749346e tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: cb8855b6-589d-4863-b86f-7fb4e30fb29e] Terminating instance [ 891.314543] env[65726]: DEBUG nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] [instance: ad48cadd-9bb6-4191-a68a-5c8fdaeebd44] Instance has had 0 of 5 cleanup attempts {{(pid=65726) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 891.423934] env[65726]: DEBUG nova.scheduler.client.report [None req-a1a4c202-18b9-4cfc-8067-d677c8415fff tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 891.460957] env[65726]: WARNING openstack [None req-d6fa7de0-71fb-41d9-ba04-00cd57a3d822 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 891.461504] env[65726]: WARNING openstack [None req-d6fa7de0-71fb-41d9-ba04-00cd57a3d822 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 891.500759] env[65726]: DEBUG oslo_vmware.api [None req-34f1738e-744a-4a46-b434-5db18d6e964a tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Task: {'id': task-5116150, 'name': ReconfigVM_Task, 'duration_secs': 0.349409} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.500759] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-34f1738e-744a-4a46-b434-5db18d6e964a tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] [instance: 4ea2b70b-0cfe-4fbf-817a-baa028b8372a] Reconfigured VM instance instance-00000040 to attach disk [datastore1] 4ea2b70b-0cfe-4fbf-817a-baa028b8372a/4ea2b70b-0cfe-4fbf-817a-baa028b8372a.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 891.501043] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f14536d2-dab0-4842-9656-2875c3730483 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.506722] env[65726]: DEBUG oslo_concurrency.lockutils [None req-d6fa7de0-71fb-41d9-ba04-00cd57a3d822 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Acquiring lock "refresh_cache-608cc0ea-3f6b-4b4f-83c1-01aa50999d63" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 891.508027] env[65726]: DEBUG oslo_concurrency.lockutils [None req-d6fa7de0-71fb-41d9-ba04-00cd57a3d822 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Acquired lock "refresh_cache-608cc0ea-3f6b-4b4f-83c1-01aa50999d63" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 891.508027] env[65726]: DEBUG nova.network.neutron [None req-d6fa7de0-71fb-41d9-ba04-00cd57a3d822 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 608cc0ea-3f6b-4b4f-83c1-01aa50999d63] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 891.514102] env[65726]: DEBUG nova.compute.manager [None req-9a0984e5-cc6d-419e-a851-17965236c39e tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: 060646cc-f0fb-4e6c-9f67-7b2daa6b6fd6] Starting instance... {{(pid=65726) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 891.522036] env[65726]: DEBUG oslo_vmware.api [None req-34f1738e-744a-4a46-b434-5db18d6e964a tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Waiting for the task: (returnval){ [ 891.522036] env[65726]: value = "task-5116153" [ 891.522036] env[65726]: _type = "Task" [ 891.522036] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.540366] env[65726]: DEBUG oslo_vmware.api [None req-34f1738e-744a-4a46-b434-5db18d6e964a tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Task: {'id': task-5116153, 'name': Rename_Task} progress is 10%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.715032] env[65726]: DEBUG oslo_vmware.api [None req-294000af-a13c-4aa5-b362-e66461f19340 tempest-ServersV294TestFqdnHostnames-819811677 tempest-ServersV294TestFqdnHostnames-819811677-project-member] Task: {'id': task-5116152, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.183555} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.715032] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-294000af-a13c-4aa5-b362-e66461f19340 tempest-ServersV294TestFqdnHostnames-819811677 tempest-ServersV294TestFqdnHostnames-819811677-project-member] Deleted the datastore file {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 891.715032] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-294000af-a13c-4aa5-b362-e66461f19340 tempest-ServersV294TestFqdnHostnames-819811677 tempest-ServersV294TestFqdnHostnames-819811677-project-member] [instance: c8be64a3-1bfa-41a3-b9be-68a7e6e052e0] Deleted contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 891.715032] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-294000af-a13c-4aa5-b362-e66461f19340 tempest-ServersV294TestFqdnHostnames-819811677 tempest-ServersV294TestFqdnHostnames-819811677-project-member] [instance: c8be64a3-1bfa-41a3-b9be-68a7e6e052e0] Instance destroyed {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 891.715032] env[65726]: INFO nova.compute.manager [None req-294000af-a13c-4aa5-b362-e66461f19340 tempest-ServersV294TestFqdnHostnames-819811677 tempest-ServersV294TestFqdnHostnames-819811677-project-member] [instance: c8be64a3-1bfa-41a3-b9be-68a7e6e052e0] Took 1.14 seconds to destroy the instance on the hypervisor. [ 891.715032] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-294000af-a13c-4aa5-b362-e66461f19340 tempest-ServersV294TestFqdnHostnames-819811677 tempest-ServersV294TestFqdnHostnames-819811677-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 891.715032] env[65726]: DEBUG nova.compute.manager [-] [instance: c8be64a3-1bfa-41a3-b9be-68a7e6e052e0] Deallocating network for instance {{(pid=65726) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 891.715032] env[65726]: DEBUG nova.network.neutron [-] [instance: c8be64a3-1bfa-41a3-b9be-68a7e6e052e0] deallocate_for_instance() {{(pid=65726) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 891.715032] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 891.715736] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 891.762899] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 891.763379] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 891.782644] env[65726]: DEBUG nova.compute.manager [None req-73ce0e73-1afb-4566-995f-f51b4749346e tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: cb8855b6-589d-4863-b86f-7fb4e30fb29e] Start destroying the instance on the hypervisor. {{(pid=65726) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 891.782880] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-73ce0e73-1afb-4566-995f-f51b4749346e tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: cb8855b6-589d-4863-b86f-7fb4e30fb29e] Destroying instance {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 891.783831] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c326cc17-304f-4c7c-93ae-3c0b1d9c1323 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.793052] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-73ce0e73-1afb-4566-995f-f51b4749346e tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: cb8855b6-589d-4863-b86f-7fb4e30fb29e] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 891.793281] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-69d4ceaa-25d3-4aa4-8060-795788aa461f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.800787] env[65726]: DEBUG oslo_vmware.api [None req-73ce0e73-1afb-4566-995f-f51b4749346e tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Waiting for the task: (returnval){ [ 891.800787] env[65726]: value = "task-5116154" [ 891.800787] env[65726]: _type = "Task" [ 891.800787] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.813961] env[65726]: DEBUG oslo_vmware.api [None req-73ce0e73-1afb-4566-995f-f51b4749346e tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Task: {'id': task-5116154, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.818731] env[65726]: DEBUG nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] [instance: 65aaa2eb-3ca5-4c2b-92f9-b524366c10dd] Instance has had 0 of 5 cleanup attempts {{(pid=65726) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 891.929346] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a1a4c202-18b9-4cfc-8067-d677c8415fff tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.663s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 891.929929] env[65726]: DEBUG nova.compute.manager [None req-a1a4c202-18b9-4cfc-8067-d677c8415fff tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] [instance: 815ee20b-cb07-4514-bd43-1b74972508a5] Start building networks asynchronously for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 891.933110] env[65726]: DEBUG oslo_concurrency.lockutils [None req-ba8f1a5b-5e31-4b5e-94c2-8c464832cc35 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 41.553s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 891.933420] env[65726]: DEBUG oslo_concurrency.lockutils [None req-ba8f1a5b-5e31-4b5e-94c2-8c464832cc35 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 891.936071] env[65726]: DEBUG oslo_concurrency.lockutils [None req-4826faf0-9908-4f3c-a509-20fe2a708723 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 38.511s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 891.937858] env[65726]: INFO nova.compute.claims [None req-4826faf0-9908-4f3c-a509-20fe2a708723 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] [instance: 29823bc7-3909-4ab7-8119-91ee59b289c5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 891.990865] env[65726]: INFO nova.scheduler.client.report [None req-ba8f1a5b-5e31-4b5e-94c2-8c464832cc35 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] Deleted allocations for instance 3cbbe146-be3a-43d5-867d-4d669884758c [ 892.011396] env[65726]: WARNING openstack [None req-d6fa7de0-71fb-41d9-ba04-00cd57a3d822 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 892.011957] env[65726]: WARNING openstack [None req-d6fa7de0-71fb-41d9-ba04-00cd57a3d822 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 892.040231] env[65726]: DEBUG oslo_vmware.api [None req-34f1738e-744a-4a46-b434-5db18d6e964a tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Task: {'id': task-5116153, 'name': Rename_Task} progress is 99%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.051635] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9a0984e5-cc6d-419e-a851-17965236c39e tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 892.316200] env[65726]: DEBUG oslo_vmware.api [None req-73ce0e73-1afb-4566-995f-f51b4749346e tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Task: {'id': task-5116154, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.322042] env[65726]: DEBUG nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] [instance: 0d6f14c5-5770-476f-a186-4cf1b2d4fbe8] Instance has had 0 of 5 cleanup attempts {{(pid=65726) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 892.443938] env[65726]: DEBUG nova.compute.utils [None req-a1a4c202-18b9-4cfc-8067-d677c8415fff tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Using /dev/sd instead of None {{(pid=65726) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 892.453868] env[65726]: DEBUG nova.compute.manager [None req-a1a4c202-18b9-4cfc-8067-d677c8415fff tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] [instance: 815ee20b-cb07-4514-bd43-1b74972508a5] Allocating IP information in the background. {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 892.453868] env[65726]: DEBUG nova.network.neutron [None req-a1a4c202-18b9-4cfc-8067-d677c8415fff tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] [instance: 815ee20b-cb07-4514-bd43-1b74972508a5] allocate_for_instance() {{(pid=65726) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 892.453868] env[65726]: WARNING neutronclient.v2_0.client [None req-a1a4c202-18b9-4cfc-8067-d677c8415fff tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 892.453868] env[65726]: WARNING neutronclient.v2_0.client [None req-a1a4c202-18b9-4cfc-8067-d677c8415fff tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 892.453868] env[65726]: WARNING openstack [None req-a1a4c202-18b9-4cfc-8067-d677c8415fff tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 892.454223] env[65726]: WARNING openstack [None req-a1a4c202-18b9-4cfc-8067-d677c8415fff tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 892.473712] env[65726]: WARNING openstack [None req-d6fa7de0-71fb-41d9-ba04-00cd57a3d822 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 892.474329] env[65726]: WARNING openstack [None req-d6fa7de0-71fb-41d9-ba04-00cd57a3d822 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 892.513701] env[65726]: DEBUG oslo_concurrency.lockutils [None req-ba8f1a5b-5e31-4b5e-94c2-8c464832cc35 tempest-ServerRescueTestJSONUnderV235-1762320888 tempest-ServerRescueTestJSONUnderV235-1762320888-project-member] Lock "3cbbe146-be3a-43d5-867d-4d669884758c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 45.623s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 892.545023] env[65726]: DEBUG oslo_vmware.api [None req-34f1738e-744a-4a46-b434-5db18d6e964a tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Task: {'id': task-5116153, 'name': Rename_Task} progress is 99%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.730881] env[65726]: WARNING openstack [None req-d6fa7de0-71fb-41d9-ba04-00cd57a3d822 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 892.731197] env[65726]: WARNING openstack [None req-d6fa7de0-71fb-41d9-ba04-00cd57a3d822 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 892.798967] env[65726]: DEBUG nova.network.neutron [-] [instance: c8be64a3-1bfa-41a3-b9be-68a7e6e052e0] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 892.815250] env[65726]: DEBUG oslo_vmware.api [None req-73ce0e73-1afb-4566-995f-f51b4749346e tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Task: {'id': task-5116154, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.825920] env[65726]: DEBUG nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] [instance: 00746508-d0d6-4dfa-9026-772398d004ee] Instance has had 0 of 5 cleanup attempts {{(pid=65726) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 892.952197] env[65726]: DEBUG nova.compute.manager [None req-a1a4c202-18b9-4cfc-8067-d677c8415fff tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] [instance: 815ee20b-cb07-4514-bd43-1b74972508a5] Start building block device mappings for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 892.960504] env[65726]: DEBUG nova.compute.manager [req-5a61c26f-e1f8-4087-88d0-293dad5ab872 req-ed9aaf6a-d3d7-461b-9977-3c1fe204a2dc service nova] [instance: c8be64a3-1bfa-41a3-b9be-68a7e6e052e0] Received event network-vif-deleted-e9d6f425-5466-4d5d-ae7f-80a433defe57 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 892.971413] env[65726]: DEBUG nova.network.neutron [None req-d6fa7de0-71fb-41d9-ba04-00cd57a3d822 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 608cc0ea-3f6b-4b4f-83c1-01aa50999d63] Updating instance_info_cache with network_info: [{"id": "e9f7132a-8d5a-47fe-afc6-d45bf3687ea4", "address": "fa:16:3e:fe:4d:df", "network": {"id": "0d0f561d-ac41-4ef5-9926-c707409799f7", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1134321217-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.160", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4eece77569624f90bf64e5c51974173f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2ce62383-8e84-4e26-955b-74c11392f4c9", "external-id": "nsx-vlan-transportzone-215", "segmentation_id": 215, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape9f7132a-8d", "ovs_interfaceid": "e9f7132a-8d5a-47fe-afc6-d45bf3687ea4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 892.996758] env[65726]: DEBUG nova.policy [None req-a1a4c202-18b9-4cfc-8067-d677c8415fff tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c826f93f56784f908bb2c08c694e5a8f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0ce9448f3ef44790af4ca94f6293d4ae', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65726) authorize /opt/stack/nova/nova/policy.py:192}} [ 893.041233] env[65726]: DEBUG oslo_vmware.api [None req-34f1738e-744a-4a46-b434-5db18d6e964a tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Task: {'id': task-5116153, 'name': Rename_Task, 'duration_secs': 1.195905} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.044452] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-34f1738e-744a-4a46-b434-5db18d6e964a tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] [instance: 4ea2b70b-0cfe-4fbf-817a-baa028b8372a] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 893.045808] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-65f03900-b2fb-44ca-b6b8-d97779c91974 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.056161] env[65726]: DEBUG oslo_vmware.api [None req-34f1738e-744a-4a46-b434-5db18d6e964a tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Waiting for the task: (returnval){ [ 893.056161] env[65726]: value = "task-5116155" [ 893.056161] env[65726]: _type = "Task" [ 893.056161] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.066775] env[65726]: DEBUG oslo_vmware.api [None req-34f1738e-744a-4a46-b434-5db18d6e964a tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Task: {'id': task-5116155, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.185527] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f8b40e1-1fef-4365-b0a5-cd67fe2fffed {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.195653] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73e0ae0d-170a-4f82-b160-c7cc7f34dff9 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.238127] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fc76160-9ed7-4221-82dd-14d6ef7fd260 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.249141] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3e46b6f-9b06-4b6f-9718-6e528fd03b3e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.267035] env[65726]: DEBUG nova.compute.provider_tree [None req-4826faf0-9908-4f3c-a509-20fe2a708723 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 893.301691] env[65726]: INFO nova.compute.manager [-] [instance: c8be64a3-1bfa-41a3-b9be-68a7e6e052e0] Took 1.59 seconds to deallocate network for instance. [ 893.317512] env[65726]: DEBUG oslo_vmware.api [None req-73ce0e73-1afb-4566-995f-f51b4749346e tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Task: {'id': task-5116154, 'name': PowerOffVM_Task, 'duration_secs': 1.061572} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.317948] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-73ce0e73-1afb-4566-995f-f51b4749346e tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: cb8855b6-589d-4863-b86f-7fb4e30fb29e] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 893.318090] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-73ce0e73-1afb-4566-995f-f51b4749346e tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: cb8855b6-589d-4863-b86f-7fb4e30fb29e] Unregistering the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 893.318364] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d8dd3f78-5fca-499c-9ec0-9e04708c3620 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.331335] env[65726]: DEBUG nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] [instance: 3827eed1-b7a1-4922-a4de-f0c038cb8566] Instance has had 0 of 5 cleanup attempts {{(pid=65726) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 893.411424] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-73ce0e73-1afb-4566-995f-f51b4749346e tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: cb8855b6-589d-4863-b86f-7fb4e30fb29e] Unregistered the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 893.411424] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-73ce0e73-1afb-4566-995f-f51b4749346e tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: cb8855b6-589d-4863-b86f-7fb4e30fb29e] Deleting contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 893.411634] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-73ce0e73-1afb-4566-995f-f51b4749346e tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Deleting the datastore file [datastore1] cb8855b6-589d-4863-b86f-7fb4e30fb29e {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 893.411894] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-76616b71-d9fc-423b-9e0b-e82aed866c0a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.421434] env[65726]: DEBUG oslo_vmware.api [None req-73ce0e73-1afb-4566-995f-f51b4749346e tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Waiting for the task: (returnval){ [ 893.421434] env[65726]: value = "task-5116157" [ 893.421434] env[65726]: _type = "Task" [ 893.421434] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.424305] env[65726]: DEBUG nova.network.neutron [None req-a1a4c202-18b9-4cfc-8067-d677c8415fff tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] [instance: 815ee20b-cb07-4514-bd43-1b74972508a5] Successfully created port: c5f4ccbb-2fbd-43ef-979c-d7538b388e9a {{(pid=65726) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 893.439263] env[65726]: DEBUG oslo_vmware.api [None req-73ce0e73-1afb-4566-995f-f51b4749346e tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Task: {'id': task-5116157, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.473664] env[65726]: DEBUG oslo_concurrency.lockutils [None req-d6fa7de0-71fb-41d9-ba04-00cd57a3d822 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Releasing lock "refresh_cache-608cc0ea-3f6b-4b4f-83c1-01aa50999d63" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 893.569094] env[65726]: DEBUG oslo_vmware.api [None req-34f1738e-744a-4a46-b434-5db18d6e964a tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Task: {'id': task-5116155, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.770829] env[65726]: DEBUG nova.scheduler.client.report [None req-4826faf0-9908-4f3c-a509-20fe2a708723 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 893.813991] env[65726]: DEBUG oslo_concurrency.lockutils [None req-294000af-a13c-4aa5-b362-e66461f19340 tempest-ServersV294TestFqdnHostnames-819811677 tempest-ServersV294TestFqdnHostnames-819811677-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 893.837250] env[65726]: DEBUG nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] [instance: ca67d1a4-d304-4d06-a436-f257b60ca080] Instance has had 0 of 5 cleanup attempts {{(pid=65726) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 893.941592] env[65726]: DEBUG oslo_vmware.api [None req-73ce0e73-1afb-4566-995f-f51b4749346e tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Task: {'id': task-5116157, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.288658} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.942358] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-73ce0e73-1afb-4566-995f-f51b4749346e tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Deleted the datastore file {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 893.942358] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-73ce0e73-1afb-4566-995f-f51b4749346e tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: cb8855b6-589d-4863-b86f-7fb4e30fb29e] Deleted contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 893.942554] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-73ce0e73-1afb-4566-995f-f51b4749346e tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: cb8855b6-589d-4863-b86f-7fb4e30fb29e] Instance destroyed {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 893.942756] env[65726]: INFO nova.compute.manager [None req-73ce0e73-1afb-4566-995f-f51b4749346e tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: cb8855b6-589d-4863-b86f-7fb4e30fb29e] Took 2.16 seconds to destroy the instance on the hypervisor. [ 893.943154] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-73ce0e73-1afb-4566-995f-f51b4749346e tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 893.943414] env[65726]: DEBUG nova.compute.manager [-] [instance: cb8855b6-589d-4863-b86f-7fb4e30fb29e] Deallocating network for instance {{(pid=65726) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 893.943568] env[65726]: DEBUG nova.network.neutron [-] [instance: cb8855b6-589d-4863-b86f-7fb4e30fb29e] deallocate_for_instance() {{(pid=65726) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 893.944164] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 893.944502] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 893.966383] env[65726]: DEBUG nova.compute.manager [None req-a1a4c202-18b9-4cfc-8067-d677c8415fff tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] [instance: 815ee20b-cb07-4514-bd43-1b74972508a5] Start spawning the instance on the hypervisor. {{(pid=65726) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 893.997032] env[65726]: DEBUG nova.virt.hardware [None req-a1a4c202-18b9-4cfc-8067-d677c8415fff tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T19:28:34Z,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1539f06fa3534e90acbc3a60c4b8bd3f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T19:28:35Z,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 893.997318] env[65726]: DEBUG nova.virt.hardware [None req-a1a4c202-18b9-4cfc-8067-d677c8415fff tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 893.997476] env[65726]: DEBUG nova.virt.hardware [None req-a1a4c202-18b9-4cfc-8067-d677c8415fff tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 893.997656] env[65726]: DEBUG nova.virt.hardware [None req-a1a4c202-18b9-4cfc-8067-d677c8415fff tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 893.997798] env[65726]: DEBUG nova.virt.hardware [None req-a1a4c202-18b9-4cfc-8067-d677c8415fff tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 893.997941] env[65726]: DEBUG nova.virt.hardware [None req-a1a4c202-18b9-4cfc-8067-d677c8415fff tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 893.998278] env[65726]: DEBUG nova.virt.hardware [None req-a1a4c202-18b9-4cfc-8067-d677c8415fff tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 893.998462] env[65726]: DEBUG nova.virt.hardware [None req-a1a4c202-18b9-4cfc-8067-d677c8415fff tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 893.998654] env[65726]: DEBUG nova.virt.hardware [None req-a1a4c202-18b9-4cfc-8067-d677c8415fff tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 893.998864] env[65726]: DEBUG nova.virt.hardware [None req-a1a4c202-18b9-4cfc-8067-d677c8415fff tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 893.999053] env[65726]: DEBUG nova.virt.hardware [None req-a1a4c202-18b9-4cfc-8067-d677c8415fff tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 893.999992] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84e88b93-3cbf-4854-9894-b295c291d1c1 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.008326] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4062b47d-255c-481c-b3a1-6ecfb858ccca {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.017467] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbcc19f1-b25e-4807-be04-4921b18f236f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.045159] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9245f2d-0de2-4e56-baf6-8e4ae2d6615e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.064489] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-d6fa7de0-71fb-41d9-ba04-00cd57a3d822 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 608cc0ea-3f6b-4b4f-83c1-01aa50999d63] Updating instance '608cc0ea-3f6b-4b4f-83c1-01aa50999d63' progress to 83 {{(pid=65726) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 894.074595] env[65726]: DEBUG oslo_vmware.api [None req-34f1738e-744a-4a46-b434-5db18d6e964a tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Task: {'id': task-5116155, 'name': PowerOnVM_Task, 'duration_secs': 0.736825} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.074885] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-34f1738e-744a-4a46-b434-5db18d6e964a tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] [instance: 4ea2b70b-0cfe-4fbf-817a-baa028b8372a] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 894.075105] env[65726]: INFO nova.compute.manager [None req-34f1738e-744a-4a46-b434-5db18d6e964a tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] [instance: 4ea2b70b-0cfe-4fbf-817a-baa028b8372a] Took 9.71 seconds to spawn the instance on the hypervisor. [ 894.075278] env[65726]: DEBUG nova.compute.manager [None req-34f1738e-744a-4a46-b434-5db18d6e964a tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] [instance: 4ea2b70b-0cfe-4fbf-817a-baa028b8372a] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 894.076256] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0c0fca4-a2d4-4572-8639-963f3dbdb170 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.280080] env[65726]: DEBUG oslo_concurrency.lockutils [None req-4826faf0-9908-4f3c-a509-20fe2a708723 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.343s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 894.280080] env[65726]: DEBUG nova.compute.manager [None req-4826faf0-9908-4f3c-a509-20fe2a708723 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] [instance: 29823bc7-3909-4ab7-8119-91ee59b289c5] Start building networks asynchronously for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 894.283834] env[65726]: DEBUG oslo_concurrency.lockutils [None req-d7ed2ba1-3808-4f50-920c-ef8e5d179423 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 29.523s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 894.284021] env[65726]: DEBUG nova.objects.instance [None req-d7ed2ba1-3808-4f50-920c-ef8e5d179423 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] [instance: 44d73b2d-2b6a-4501-9944-432da3c9330d] Trying to apply a migration context that does not seem to be set for this instance {{(pid=65726) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 894.300660] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 894.300969] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 894.338931] env[65726]: DEBUG nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] [instance: 02351120-8385-4403-8464-a154f3da9380] Instance has had 0 of 5 cleanup attempts {{(pid=65726) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 894.572215] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-d6fa7de0-71fb-41d9-ba04-00cd57a3d822 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 608cc0ea-3f6b-4b4f-83c1-01aa50999d63] Updating instance '608cc0ea-3f6b-4b4f-83c1-01aa50999d63' progress to 100 {{(pid=65726) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 894.600519] env[65726]: INFO nova.compute.manager [None req-34f1738e-744a-4a46-b434-5db18d6e964a tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] [instance: 4ea2b70b-0cfe-4fbf-817a-baa028b8372a] Took 57.80 seconds to build instance. [ 894.789276] env[65726]: DEBUG nova.compute.utils [None req-4826faf0-9908-4f3c-a509-20fe2a708723 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Using /dev/sd instead of None {{(pid=65726) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 894.794102] env[65726]: DEBUG nova.compute.manager [None req-4826faf0-9908-4f3c-a509-20fe2a708723 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] [instance: 29823bc7-3909-4ab7-8119-91ee59b289c5] Allocating IP information in the background. {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 894.794447] env[65726]: DEBUG nova.network.neutron [None req-4826faf0-9908-4f3c-a509-20fe2a708723 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] [instance: 29823bc7-3909-4ab7-8119-91ee59b289c5] allocate_for_instance() {{(pid=65726) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 894.794823] env[65726]: WARNING neutronclient.v2_0.client [None req-4826faf0-9908-4f3c-a509-20fe2a708723 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 894.795227] env[65726]: WARNING neutronclient.v2_0.client [None req-4826faf0-9908-4f3c-a509-20fe2a708723 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 894.795841] env[65726]: WARNING openstack [None req-4826faf0-9908-4f3c-a509-20fe2a708723 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 894.796266] env[65726]: WARNING openstack [None req-4826faf0-9908-4f3c-a509-20fe2a708723 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 894.806661] env[65726]: DEBUG oslo_concurrency.lockutils [None req-d7ed2ba1-3808-4f50-920c-ef8e5d179423 tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 0.523s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 894.811643] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9c76632b-07ea-476e-92d3-ffb457aef721 tempest-ServerAddressesTestJSON-569100133 tempest-ServerAddressesTestJSON-569100133-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 29.980s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 894.813401] env[65726]: INFO nova.compute.claims [None req-9c76632b-07ea-476e-92d3-ffb457aef721 tempest-ServerAddressesTestJSON-569100133 tempest-ServerAddressesTestJSON-569100133-project-member] [instance: 70375746-06dd-4710-9136-95e9e5759a0e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 894.826869] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e9f6f3ba-cd10-4f45-bec2-5e0b6cfdfe85 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Acquiring lock "6c8db442-daaa-4eca-ae24-b9d4afaf3a71" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 894.826869] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e9f6f3ba-cd10-4f45-bec2-5e0b6cfdfe85 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Lock "6c8db442-daaa-4eca-ae24-b9d4afaf3a71" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 894.826869] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e9f6f3ba-cd10-4f45-bec2-5e0b6cfdfe85 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Acquiring lock "6c8db442-daaa-4eca-ae24-b9d4afaf3a71-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 894.827245] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e9f6f3ba-cd10-4f45-bec2-5e0b6cfdfe85 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Lock "6c8db442-daaa-4eca-ae24-b9d4afaf3a71-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 894.827326] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e9f6f3ba-cd10-4f45-bec2-5e0b6cfdfe85 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Lock "6c8db442-daaa-4eca-ae24-b9d4afaf3a71-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 894.829871] env[65726]: INFO nova.compute.manager [None req-e9f6f3ba-cd10-4f45-bec2-5e0b6cfdfe85 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] [instance: 6c8db442-daaa-4eca-ae24-b9d4afaf3a71] Terminating instance [ 894.843536] env[65726]: DEBUG nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] [instance: 83573c2b-9448-456f-8fd2-b19661dd6cc4] Instance has had 0 of 5 cleanup attempts {{(pid=65726) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 895.104056] env[65726]: DEBUG oslo_concurrency.lockutils [None req-34f1738e-744a-4a46-b434-5db18d6e964a tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Lock "4ea2b70b-0cfe-4fbf-817a-baa028b8372a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 64.919s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 895.297254] env[65726]: DEBUG nova.compute.manager [None req-4826faf0-9908-4f3c-a509-20fe2a708723 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] [instance: 29823bc7-3909-4ab7-8119-91ee59b289c5] Start building block device mappings for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 895.337375] env[65726]: DEBUG nova.compute.manager [None req-e9f6f3ba-cd10-4f45-bec2-5e0b6cfdfe85 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] [instance: 6c8db442-daaa-4eca-ae24-b9d4afaf3a71] Start destroying the instance on the hypervisor. {{(pid=65726) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 895.337375] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-e9f6f3ba-cd10-4f45-bec2-5e0b6cfdfe85 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] [instance: 6c8db442-daaa-4eca-ae24-b9d4afaf3a71] Destroying instance {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 895.337375] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2138625-1ee8-4e24-a801-532c075a8614 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.348145] env[65726]: DEBUG nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] [instance: bd839fe1-8801-4ba1-9c23-288453258b66] Instance has had 0 of 5 cleanup attempts {{(pid=65726) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 895.350248] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-e9f6f3ba-cd10-4f45-bec2-5e0b6cfdfe85 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] [instance: 6c8db442-daaa-4eca-ae24-b9d4afaf3a71] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 895.350524] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ee3f4676-2d87-4713-a0ab-84e0d142e99b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.361670] env[65726]: DEBUG oslo_vmware.api [None req-e9f6f3ba-cd10-4f45-bec2-5e0b6cfdfe85 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Waiting for the task: (returnval){ [ 895.361670] env[65726]: value = "task-5116158" [ 895.361670] env[65726]: _type = "Task" [ 895.361670] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.371698] env[65726]: DEBUG oslo_vmware.api [None req-e9f6f3ba-cd10-4f45-bec2-5e0b6cfdfe85 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Task: {'id': task-5116158, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.434873] env[65726]: DEBUG nova.network.neutron [None req-a1a4c202-18b9-4cfc-8067-d677c8415fff tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] [instance: 815ee20b-cb07-4514-bd43-1b74972508a5] Successfully updated port: c5f4ccbb-2fbd-43ef-979c-d7538b388e9a {{(pid=65726) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 895.776210] env[65726]: DEBUG nova.policy [None req-4826faf0-9908-4f3c-a509-20fe2a708723 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c826f93f56784f908bb2c08c694e5a8f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0ce9448f3ef44790af4ca94f6293d4ae', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65726) authorize /opt/stack/nova/nova/policy.py:192}} [ 895.856871] env[65726]: DEBUG nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] [instance: 96acc5f8-b411-408e-929d-d1035cfd50db] Instance has had 0 of 5 cleanup attempts {{(pid=65726) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 895.872719] env[65726]: DEBUG oslo_vmware.api [None req-e9f6f3ba-cd10-4f45-bec2-5e0b6cfdfe85 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Task: {'id': task-5116158, 'name': PowerOffVM_Task, 'duration_secs': 0.315698} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.875710] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-e9f6f3ba-cd10-4f45-bec2-5e0b6cfdfe85 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] [instance: 6c8db442-daaa-4eca-ae24-b9d4afaf3a71] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 895.875915] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-e9f6f3ba-cd10-4f45-bec2-5e0b6cfdfe85 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] [instance: 6c8db442-daaa-4eca-ae24-b9d4afaf3a71] Unregistering the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 895.876466] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e2a95577-cbc3-4216-9ccc-b59bedba6be4 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.943519] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a1a4c202-18b9-4cfc-8067-d677c8415fff tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Acquiring lock "refresh_cache-815ee20b-cb07-4514-bd43-1b74972508a5" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 895.943519] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a1a4c202-18b9-4cfc-8067-d677c8415fff tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Acquired lock "refresh_cache-815ee20b-cb07-4514-bd43-1b74972508a5" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 895.943519] env[65726]: DEBUG nova.network.neutron [None req-a1a4c202-18b9-4cfc-8067-d677c8415fff tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] [instance: 815ee20b-cb07-4514-bd43-1b74972508a5] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 895.958026] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-e9f6f3ba-cd10-4f45-bec2-5e0b6cfdfe85 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] [instance: 6c8db442-daaa-4eca-ae24-b9d4afaf3a71] Unregistered the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 895.958317] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-e9f6f3ba-cd10-4f45-bec2-5e0b6cfdfe85 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] [instance: 6c8db442-daaa-4eca-ae24-b9d4afaf3a71] Deleting contents of the VM from datastore datastore2 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 895.958529] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-e9f6f3ba-cd10-4f45-bec2-5e0b6cfdfe85 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Deleting the datastore file [datastore2] 6c8db442-daaa-4eca-ae24-b9d4afaf3a71 {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 895.967304] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3becd91f-6b24-4095-b9f7-b6860b510269 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.976046] env[65726]: DEBUG oslo_vmware.api [None req-e9f6f3ba-cd10-4f45-bec2-5e0b6cfdfe85 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Waiting for the task: (returnval){ [ 895.976046] env[65726]: value = "task-5116160" [ 895.976046] env[65726]: _type = "Task" [ 895.976046] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.991859] env[65726]: DEBUG oslo_vmware.api [None req-e9f6f3ba-cd10-4f45-bec2-5e0b6cfdfe85 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Task: {'id': task-5116160, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.185343] env[65726]: DEBUG nova.network.neutron [-] [instance: cb8855b6-589d-4863-b86f-7fb4e30fb29e] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 896.256754] env[65726]: DEBUG nova.network.neutron [None req-4826faf0-9908-4f3c-a509-20fe2a708723 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] [instance: 29823bc7-3909-4ab7-8119-91ee59b289c5] Successfully created port: ef57c09b-e512-44ad-adeb-8aee81e9dc7b {{(pid=65726) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 896.309226] env[65726]: DEBUG nova.compute.manager [None req-4826faf0-9908-4f3c-a509-20fe2a708723 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] [instance: 29823bc7-3909-4ab7-8119-91ee59b289c5] Start spawning the instance on the hypervisor. {{(pid=65726) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 896.343150] env[65726]: DEBUG nova.virt.hardware [None req-4826faf0-9908-4f3c-a509-20fe2a708723 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T19:28:34Z,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1539f06fa3534e90acbc3a60c4b8bd3f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T19:28:35Z,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 896.343282] env[65726]: DEBUG nova.virt.hardware [None req-4826faf0-9908-4f3c-a509-20fe2a708723 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 896.343419] env[65726]: DEBUG nova.virt.hardware [None req-4826faf0-9908-4f3c-a509-20fe2a708723 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 896.343556] env[65726]: DEBUG nova.virt.hardware [None req-4826faf0-9908-4f3c-a509-20fe2a708723 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 896.343700] env[65726]: DEBUG nova.virt.hardware [None req-4826faf0-9908-4f3c-a509-20fe2a708723 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 896.343844] env[65726]: DEBUG nova.virt.hardware [None req-4826faf0-9908-4f3c-a509-20fe2a708723 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 896.344066] env[65726]: DEBUG nova.virt.hardware [None req-4826faf0-9908-4f3c-a509-20fe2a708723 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 896.344225] env[65726]: DEBUG nova.virt.hardware [None req-4826faf0-9908-4f3c-a509-20fe2a708723 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 896.344394] env[65726]: DEBUG nova.virt.hardware [None req-4826faf0-9908-4f3c-a509-20fe2a708723 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 896.344554] env[65726]: DEBUG nova.virt.hardware [None req-4826faf0-9908-4f3c-a509-20fe2a708723 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 896.344752] env[65726]: DEBUG nova.virt.hardware [None req-4826faf0-9908-4f3c-a509-20fe2a708723 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 896.345772] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27849e1a-e41e-407f-839c-8b587f1954ab {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.357182] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee7b075f-2dc9-483a-beb3-865eee4fc35b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.361529] env[65726]: DEBUG nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] [instance: 815f056b-2d88-40dd-9ef7-a891ec8d92cf] Instance has had 0 of 5 cleanup attempts {{(pid=65726) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 896.447255] env[65726]: WARNING openstack [None req-a1a4c202-18b9-4cfc-8067-d677c8415fff tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 896.447255] env[65726]: WARNING openstack [None req-a1a4c202-18b9-4cfc-8067-d677c8415fff tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 896.453967] env[65726]: DEBUG nova.network.neutron [None req-a1a4c202-18b9-4cfc-8067-d677c8415fff tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] [instance: 815ee20b-cb07-4514-bd43-1b74972508a5] Instance cache missing network info. {{(pid=65726) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 896.483691] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-134cb44f-070c-4ab1-861f-e6674805b80c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.498598] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19d7fa73-4d49-4463-bf7d-8f213a0958db {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.503847] env[65726]: DEBUG oslo_vmware.api [None req-e9f6f3ba-cd10-4f45-bec2-5e0b6cfdfe85 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Task: {'id': task-5116160, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.348487} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.504230] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-e9f6f3ba-cd10-4f45-bec2-5e0b6cfdfe85 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Deleted the datastore file {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 896.504437] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-e9f6f3ba-cd10-4f45-bec2-5e0b6cfdfe85 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] [instance: 6c8db442-daaa-4eca-ae24-b9d4afaf3a71] Deleted contents of the VM from datastore datastore2 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 896.504598] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-e9f6f3ba-cd10-4f45-bec2-5e0b6cfdfe85 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] [instance: 6c8db442-daaa-4eca-ae24-b9d4afaf3a71] Instance destroyed {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 896.504776] env[65726]: INFO nova.compute.manager [None req-e9f6f3ba-cd10-4f45-bec2-5e0b6cfdfe85 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] [instance: 6c8db442-daaa-4eca-ae24-b9d4afaf3a71] Took 1.17 seconds to destroy the instance on the hypervisor. [ 896.505075] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-e9f6f3ba-cd10-4f45-bec2-5e0b6cfdfe85 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 896.505779] env[65726]: DEBUG nova.compute.manager [-] [instance: 6c8db442-daaa-4eca-ae24-b9d4afaf3a71] Deallocating network for instance {{(pid=65726) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 896.505896] env[65726]: DEBUG nova.network.neutron [-] [instance: 6c8db442-daaa-4eca-ae24-b9d4afaf3a71] deallocate_for_instance() {{(pid=65726) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 896.506460] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 896.506744] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 896.543517] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a2f8697-347c-4c7a-bdd3-c10c6ee49e48 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.555657] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8166b850-c360-465d-865d-4962c5dd6ea5 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.574916] env[65726]: DEBUG nova.compute.provider_tree [None req-9c76632b-07ea-476e-92d3-ffb457aef721 tempest-ServerAddressesTestJSON-569100133 tempest-ServerAddressesTestJSON-569100133-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 896.688739] env[65726]: INFO nova.compute.manager [-] [instance: cb8855b6-589d-4863-b86f-7fb4e30fb29e] Took 2.74 seconds to deallocate network for instance. [ 896.868862] env[65726]: DEBUG nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] [instance: 423af936-a383-4652-8887-9c8b8f507909] Instance has had 0 of 5 cleanup attempts {{(pid=65726) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 897.000531] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 897.003246] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 897.033816] env[65726]: WARNING openstack [None req-a1a4c202-18b9-4cfc-8067-d677c8415fff tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 897.034217] env[65726]: WARNING openstack [None req-a1a4c202-18b9-4cfc-8067-d677c8415fff tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 897.080168] env[65726]: DEBUG nova.scheduler.client.report [None req-9c76632b-07ea-476e-92d3-ffb457aef721 tempest-ServerAddressesTestJSON-569100133 tempest-ServerAddressesTestJSON-569100133-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 897.151521] env[65726]: DEBUG oslo_vmware.rw_handles [None req-fcafe3da-7a8e-4b08-9577-72f88a634bf5 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52aff181-bf03-3dc8-d23e-16139f026815/disk-0.vmdk. {{(pid=65726) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 897.151521] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a30a856-2ecf-4604-8a47-f5c39df5e6b4 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.158058] env[65726]: DEBUG oslo_vmware.rw_handles [None req-fcafe3da-7a8e-4b08-9577-72f88a634bf5 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52aff181-bf03-3dc8-d23e-16139f026815/disk-0.vmdk is in state: ready. {{(pid=65726) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 897.158235] env[65726]: ERROR oslo_vmware.rw_handles [None req-fcafe3da-7a8e-4b08-9577-72f88a634bf5 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52aff181-bf03-3dc8-d23e-16139f026815/disk-0.vmdk due to incomplete transfer. [ 897.158476] env[65726]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-40685b50-2bb3-420e-88e3-82fff0f5f9a8 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.169304] env[65726]: DEBUG oslo_vmware.rw_handles [None req-fcafe3da-7a8e-4b08-9577-72f88a634bf5 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52aff181-bf03-3dc8-d23e-16139f026815/disk-0.vmdk. {{(pid=65726) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 897.169517] env[65726]: DEBUG nova.virt.vmwareapi.images [None req-fcafe3da-7a8e-4b08-9577-72f88a634bf5 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] [instance: c4cc4f85-cb35-4edc-a58b-adfee0ce1265] Uploaded image 776f8431-e63e-4251-a79b-9d7edff7e53f to the Glance image server {{(pid=65726) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 897.172618] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-fcafe3da-7a8e-4b08-9577-72f88a634bf5 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] [instance: c4cc4f85-cb35-4edc-a58b-adfee0ce1265] Destroying the VM {{(pid=65726) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 897.172944] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-4c8b650c-75a9-4bb1-88df-c76b79ef46dc {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.181117] env[65726]: DEBUG oslo_vmware.api [None req-fcafe3da-7a8e-4b08-9577-72f88a634bf5 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Waiting for the task: (returnval){ [ 897.181117] env[65726]: value = "task-5116161" [ 897.181117] env[65726]: _type = "Task" [ 897.181117] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.193129] env[65726]: DEBUG oslo_vmware.api [None req-fcafe3da-7a8e-4b08-9577-72f88a634bf5 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Task: {'id': task-5116161, 'name': Destroy_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.198854] env[65726]: DEBUG oslo_concurrency.lockutils [None req-73ce0e73-1afb-4566-995f-f51b4749346e tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 897.374445] env[65726]: DEBUG nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] [instance: 49bcab1f-7b64-4999-abff-37771c58a271] Instance has had 0 of 5 cleanup attempts {{(pid=65726) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 897.586225] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9c76632b-07ea-476e-92d3-ffb457aef721 tempest-ServerAddressesTestJSON-569100133 tempest-ServerAddressesTestJSON-569100133-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.778s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 897.586470] env[65726]: DEBUG nova.compute.manager [None req-9c76632b-07ea-476e-92d3-ffb457aef721 tempest-ServerAddressesTestJSON-569100133 tempest-ServerAddressesTestJSON-569100133-project-member] [instance: 70375746-06dd-4710-9136-95e9e5759a0e] Start building networks asynchronously for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 897.590532] env[65726]: DEBUG oslo_concurrency.lockutils [None req-253a2ea4-406a-415c-8211-e66a1f71d581 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: waited 32.712s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 897.679787] env[65726]: WARNING openstack [None req-a1a4c202-18b9-4cfc-8067-d677c8415fff tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 897.680241] env[65726]: WARNING openstack [None req-a1a4c202-18b9-4cfc-8067-d677c8415fff tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 897.700687] env[65726]: DEBUG oslo_vmware.api [None req-fcafe3da-7a8e-4b08-9577-72f88a634bf5 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Task: {'id': task-5116161, 'name': Destroy_Task} progress is 33%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.840569] env[65726]: DEBUG nova.network.neutron [None req-a1a4c202-18b9-4cfc-8067-d677c8415fff tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] [instance: 815ee20b-cb07-4514-bd43-1b74972508a5] Updating instance_info_cache with network_info: [{"id": "c5f4ccbb-2fbd-43ef-979c-d7538b388e9a", "address": "fa:16:3e:c4:a5:f9", "network": {"id": "55d629b8-a541-4720-9289-927bbc6e6f22", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.68", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "1539f06fa3534e90acbc3a60c4b8bd3f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "35342bcb-8b06-472e-b3c0-43fd3d6c4b30", "external-id": "nsx-vlan-transportzone-524", "segmentation_id": 524, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc5f4ccbb-2f", "ovs_interfaceid": "c5f4ccbb-2fbd-43ef-979c-d7538b388e9a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 897.878057] env[65726]: DEBUG nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] [instance: 30b17a17-229c-4ca7-9ae6-c67ccdcd3963] Instance has had 0 of 5 cleanup attempts {{(pid=65726) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 898.012421] env[65726]: DEBUG nova.compute.manager [req-2b6605c8-b97e-4dd1-a3ce-e64b608c5225 req-276d97b0-8af0-41a1-a94d-be55c5da32b5 service nova] [instance: 815ee20b-cb07-4514-bd43-1b74972508a5] Received event network-vif-plugged-c5f4ccbb-2fbd-43ef-979c-d7538b388e9a {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 898.012421] env[65726]: DEBUG oslo_concurrency.lockutils [req-2b6605c8-b97e-4dd1-a3ce-e64b608c5225 req-276d97b0-8af0-41a1-a94d-be55c5da32b5 service nova] Acquiring lock "815ee20b-cb07-4514-bd43-1b74972508a5-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 898.012675] env[65726]: DEBUG oslo_concurrency.lockutils [req-2b6605c8-b97e-4dd1-a3ce-e64b608c5225 req-276d97b0-8af0-41a1-a94d-be55c5da32b5 service nova] Lock "815ee20b-cb07-4514-bd43-1b74972508a5-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 898.012675] env[65726]: DEBUG oslo_concurrency.lockutils [req-2b6605c8-b97e-4dd1-a3ce-e64b608c5225 req-276d97b0-8af0-41a1-a94d-be55c5da32b5 service nova] Lock "815ee20b-cb07-4514-bd43-1b74972508a5-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 898.012910] env[65726]: DEBUG nova.compute.manager [req-2b6605c8-b97e-4dd1-a3ce-e64b608c5225 req-276d97b0-8af0-41a1-a94d-be55c5da32b5 service nova] [instance: 815ee20b-cb07-4514-bd43-1b74972508a5] No waiting events found dispatching network-vif-plugged-c5f4ccbb-2fbd-43ef-979c-d7538b388e9a {{(pid=65726) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 898.013316] env[65726]: WARNING nova.compute.manager [req-2b6605c8-b97e-4dd1-a3ce-e64b608c5225 req-276d97b0-8af0-41a1-a94d-be55c5da32b5 service nova] [instance: 815ee20b-cb07-4514-bd43-1b74972508a5] Received unexpected event network-vif-plugged-c5f4ccbb-2fbd-43ef-979c-d7538b388e9a for instance with vm_state building and task_state spawning. [ 898.079853] env[65726]: DEBUG nova.network.neutron [-] [instance: 6c8db442-daaa-4eca-ae24-b9d4afaf3a71] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 898.096345] env[65726]: DEBUG nova.compute.utils [None req-9c76632b-07ea-476e-92d3-ffb457aef721 tempest-ServerAddressesTestJSON-569100133 tempest-ServerAddressesTestJSON-569100133-project-member] Using /dev/sd instead of None {{(pid=65726) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 898.096345] env[65726]: DEBUG nova.objects.instance [None req-253a2ea4-406a-415c-8211-e66a1f71d581 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Lazy-loading 'migration_context' on Instance uuid 76249623-6f83-46a3-b8c5-c001111aa698 {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 898.096952] env[65726]: DEBUG nova.compute.manager [None req-9c76632b-07ea-476e-92d3-ffb457aef721 tempest-ServerAddressesTestJSON-569100133 tempest-ServerAddressesTestJSON-569100133-project-member] [instance: 70375746-06dd-4710-9136-95e9e5759a0e] Allocating IP information in the background. {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 898.097161] env[65726]: DEBUG nova.network.neutron [None req-9c76632b-07ea-476e-92d3-ffb457aef721 tempest-ServerAddressesTestJSON-569100133 tempest-ServerAddressesTestJSON-569100133-project-member] [instance: 70375746-06dd-4710-9136-95e9e5759a0e] allocate_for_instance() {{(pid=65726) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 898.097491] env[65726]: WARNING neutronclient.v2_0.client [None req-9c76632b-07ea-476e-92d3-ffb457aef721 tempest-ServerAddressesTestJSON-569100133 tempest-ServerAddressesTestJSON-569100133-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 898.097841] env[65726]: WARNING neutronclient.v2_0.client [None req-9c76632b-07ea-476e-92d3-ffb457aef721 tempest-ServerAddressesTestJSON-569100133 tempest-ServerAddressesTestJSON-569100133-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 898.098438] env[65726]: WARNING openstack [None req-9c76632b-07ea-476e-92d3-ffb457aef721 tempest-ServerAddressesTestJSON-569100133 tempest-ServerAddressesTestJSON-569100133-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 898.098806] env[65726]: WARNING openstack [None req-9c76632b-07ea-476e-92d3-ffb457aef721 tempest-ServerAddressesTestJSON-569100133 tempest-ServerAddressesTestJSON-569100133-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 898.131880] env[65726]: DEBUG nova.network.neutron [None req-4826faf0-9908-4f3c-a509-20fe2a708723 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] [instance: 29823bc7-3909-4ab7-8119-91ee59b289c5] Successfully updated port: ef57c09b-e512-44ad-adeb-8aee81e9dc7b {{(pid=65726) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 898.168089] env[65726]: DEBUG nova.policy [None req-9c76632b-07ea-476e-92d3-ffb457aef721 tempest-ServerAddressesTestJSON-569100133 tempest-ServerAddressesTestJSON-569100133-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5b06400c1cd847c1b198021400fd30b0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '04fbc0c66f094d73b87bccb72aa6f37a', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65726) authorize /opt/stack/nova/nova/policy.py:192}} [ 898.211644] env[65726]: DEBUG oslo_vmware.api [None req-fcafe3da-7a8e-4b08-9577-72f88a634bf5 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Task: {'id': task-5116161, 'name': Destroy_Task} progress is 100%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.346409] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a1a4c202-18b9-4cfc-8067-d677c8415fff tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Releasing lock "refresh_cache-815ee20b-cb07-4514-bd43-1b74972508a5" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 898.347059] env[65726]: DEBUG nova.compute.manager [None req-a1a4c202-18b9-4cfc-8067-d677c8415fff tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] [instance: 815ee20b-cb07-4514-bd43-1b74972508a5] Instance network_info: |[{"id": "c5f4ccbb-2fbd-43ef-979c-d7538b388e9a", "address": "fa:16:3e:c4:a5:f9", "network": {"id": "55d629b8-a541-4720-9289-927bbc6e6f22", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.68", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "1539f06fa3534e90acbc3a60c4b8bd3f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "35342bcb-8b06-472e-b3c0-43fd3d6c4b30", "external-id": "nsx-vlan-transportzone-524", "segmentation_id": 524, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc5f4ccbb-2f", "ovs_interfaceid": "c5f4ccbb-2fbd-43ef-979c-d7538b388e9a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 898.347386] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-a1a4c202-18b9-4cfc-8067-d677c8415fff tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] [instance: 815ee20b-cb07-4514-bd43-1b74972508a5] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c4:a5:f9', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '35342bcb-8b06-472e-b3c0-43fd3d6c4b30', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c5f4ccbb-2fbd-43ef-979c-d7538b388e9a', 'vif_model': 'vmxnet3'}] {{(pid=65726) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 898.355353] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-a1a4c202-18b9-4cfc-8067-d677c8415fff tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Creating folder: Project (0ce9448f3ef44790af4ca94f6293d4ae). Parent ref: group-v995008. {{(pid=65726) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 898.355773] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b540ee12-b4ef-4cba-be82-858d36db6ed4 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.369500] env[65726]: INFO nova.virt.vmwareapi.vm_util [None req-a1a4c202-18b9-4cfc-8067-d677c8415fff tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Created folder: Project (0ce9448f3ef44790af4ca94f6293d4ae) in parent group-v995008. [ 898.369500] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-a1a4c202-18b9-4cfc-8067-d677c8415fff tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Creating folder: Instances. Parent ref: group-v995190. {{(pid=65726) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 898.369961] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f81d6e5b-67b4-4656-a5d4-4c84bfab0fd3 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.381601] env[65726]: INFO nova.virt.vmwareapi.vm_util [None req-a1a4c202-18b9-4cfc-8067-d677c8415fff tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Created folder: Instances in parent group-v995190. [ 898.381890] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-a1a4c202-18b9-4cfc-8067-d677c8415fff tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 898.383434] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 815ee20b-cb07-4514-bd43-1b74972508a5] Creating VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 898.383434] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-4ac23879-2912-4a30-a732-f4aadbe7d1e9 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.403118] env[65726]: DEBUG nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] [instance: c977257e-dce0-4ecc-b42b-0f6ebab38797] Instance has had 0 of 5 cleanup attempts {{(pid=65726) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 898.411543] env[65726]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 898.411543] env[65726]: value = "task-5116164" [ 898.411543] env[65726]: _type = "Task" [ 898.411543] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.420698] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116164, 'name': CreateVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.485139] env[65726]: DEBUG nova.network.neutron [None req-9c76632b-07ea-476e-92d3-ffb457aef721 tempest-ServerAddressesTestJSON-569100133 tempest-ServerAddressesTestJSON-569100133-project-member] [instance: 70375746-06dd-4710-9136-95e9e5759a0e] Successfully created port: 7818914d-352b-408a-9d0f-4c2bb1a25f2e {{(pid=65726) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 898.585607] env[65726]: INFO nova.compute.manager [-] [instance: 6c8db442-daaa-4eca-ae24-b9d4afaf3a71] Took 2.08 seconds to deallocate network for instance. [ 898.611089] env[65726]: DEBUG nova.compute.manager [None req-9c76632b-07ea-476e-92d3-ffb457aef721 tempest-ServerAddressesTestJSON-569100133 tempest-ServerAddressesTestJSON-569100133-project-member] [instance: 70375746-06dd-4710-9136-95e9e5759a0e] Start building block device mappings for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 898.633302] env[65726]: DEBUG oslo_concurrency.lockutils [None req-4826faf0-9908-4f3c-a509-20fe2a708723 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Acquiring lock "refresh_cache-29823bc7-3909-4ab7-8119-91ee59b289c5" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 898.633584] env[65726]: DEBUG oslo_concurrency.lockutils [None req-4826faf0-9908-4f3c-a509-20fe2a708723 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Acquired lock "refresh_cache-29823bc7-3909-4ab7-8119-91ee59b289c5" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 898.633670] env[65726]: DEBUG nova.network.neutron [None req-4826faf0-9908-4f3c-a509-20fe2a708723 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] [instance: 29823bc7-3909-4ab7-8119-91ee59b289c5] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 898.704886] env[65726]: DEBUG oslo_vmware.api [None req-fcafe3da-7a8e-4b08-9577-72f88a634bf5 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Task: {'id': task-5116161, 'name': Destroy_Task} progress is 100%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.907742] env[65726]: DEBUG nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] [instance: ddd36db0-28ad-4582-994b-7eef8b08ed46] Instance has had 0 of 5 cleanup attempts {{(pid=65726) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 898.927070] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116164, 'name': CreateVM_Task, 'duration_secs': 0.403693} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.927278] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 815ee20b-cb07-4514-bd43-1b74972508a5] Created VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 898.928076] env[65726]: WARNING openstack [None req-a1a4c202-18b9-4cfc-8067-d677c8415fff tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 898.928493] env[65726]: WARNING openstack [None req-a1a4c202-18b9-4cfc-8067-d677c8415fff tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 898.933627] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a1a4c202-18b9-4cfc-8067-d677c8415fff tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 898.933789] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a1a4c202-18b9-4cfc-8067-d677c8415fff tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 898.934129] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a1a4c202-18b9-4cfc-8067-d677c8415fff tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 898.937070] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ee358425-bf20-4d57-a41d-2de1d481bc61 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.943775] env[65726]: DEBUG oslo_vmware.api [None req-a1a4c202-18b9-4cfc-8067-d677c8415fff tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Waiting for the task: (returnval){ [ 898.943775] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52c405f4-7afb-9bef-f811-099daf73bb60" [ 898.943775] env[65726]: _type = "Task" [ 898.943775] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.952651] env[65726]: DEBUG oslo_vmware.api [None req-a1a4c202-18b9-4cfc-8067-d677c8415fff tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52c405f4-7afb-9bef-f811-099daf73bb60, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.959860] env[65726]: DEBUG oslo_concurrency.lockutils [None req-21960d76-f281-41f6-8236-c885b049da01 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Acquiring lock "608cc0ea-3f6b-4b4f-83c1-01aa50999d63" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 898.960112] env[65726]: DEBUG oslo_concurrency.lockutils [None req-21960d76-f281-41f6-8236-c885b049da01 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Lock "608cc0ea-3f6b-4b4f-83c1-01aa50999d63" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 898.960295] env[65726]: DEBUG nova.compute.manager [None req-21960d76-f281-41f6-8236-c885b049da01 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 608cc0ea-3f6b-4b4f-83c1-01aa50999d63] Going to confirm migration 3 {{(pid=65726) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:5307}} [ 899.096373] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e9f6f3ba-cd10-4f45-bec2-5e0b6cfdfe85 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 899.138433] env[65726]: WARNING openstack [None req-4826faf0-9908-4f3c-a509-20fe2a708723 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 899.138922] env[65726]: WARNING openstack [None req-4826faf0-9908-4f3c-a509-20fe2a708723 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 899.144455] env[65726]: DEBUG nova.network.neutron [None req-4826faf0-9908-4f3c-a509-20fe2a708723 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] [instance: 29823bc7-3909-4ab7-8119-91ee59b289c5] Instance cache missing network info. {{(pid=65726) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 899.167303] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d3c738c-2d2d-4c11-9ae2-dfd90fec03b9 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.177220] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd953352-cb91-4df9-9dc9-e9e109ada987 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.214862] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-080c44a1-e1d9-44f4-b710-eb953bfe95cf {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.226529] env[65726]: DEBUG oslo_vmware.api [None req-fcafe3da-7a8e-4b08-9577-72f88a634bf5 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Task: {'id': task-5116161, 'name': Destroy_Task, 'duration_secs': 1.823821} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.227698] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4fbe9ee-9fc3-4b26-b427-8569932c055e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.232047] env[65726]: INFO nova.virt.vmwareapi.vm_util [None req-fcafe3da-7a8e-4b08-9577-72f88a634bf5 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] [instance: c4cc4f85-cb35-4edc-a58b-adfee0ce1265] Destroyed the VM [ 899.232323] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-fcafe3da-7a8e-4b08-9577-72f88a634bf5 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] [instance: c4cc4f85-cb35-4edc-a58b-adfee0ce1265] Deleting Snapshot of the VM instance {{(pid=65726) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 899.232607] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-117b900d-834f-46d8-8539-9d7f21bc0034 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.246292] env[65726]: DEBUG nova.compute.provider_tree [None req-253a2ea4-406a-415c-8211-e66a1f71d581 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 899.249340] env[65726]: DEBUG oslo_vmware.api [None req-fcafe3da-7a8e-4b08-9577-72f88a634bf5 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Waiting for the task: (returnval){ [ 899.249340] env[65726]: value = "task-5116165" [ 899.249340] env[65726]: _type = "Task" [ 899.249340] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.258639] env[65726]: DEBUG oslo_vmware.api [None req-fcafe3da-7a8e-4b08-9577-72f88a634bf5 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Task: {'id': task-5116165, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.411227] env[65726]: DEBUG nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] [instance: c2929727-a7ea-4f63-997d-474c29a305fc] Instance has had 0 of 5 cleanup attempts {{(pid=65726) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 899.458401] env[65726]: DEBUG oslo_vmware.api [None req-a1a4c202-18b9-4cfc-8067-d677c8415fff tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52c405f4-7afb-9bef-f811-099daf73bb60, 'name': SearchDatastore_Task, 'duration_secs': 0.030806} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.460342] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a1a4c202-18b9-4cfc-8067-d677c8415fff tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 899.460342] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-a1a4c202-18b9-4cfc-8067-d677c8415fff tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] [instance: 815ee20b-cb07-4514-bd43-1b74972508a5] Processing image b52362a3-ee8a-4cbf-b06f-513b0cc8f95c {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 899.460342] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a1a4c202-18b9-4cfc-8067-d677c8415fff tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 899.460342] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a1a4c202-18b9-4cfc-8067-d677c8415fff tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 899.460342] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-a1a4c202-18b9-4cfc-8067-d677c8415fff tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 899.461659] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b67605c9-5198-4462-bb1c-0aa5247f9ad9 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.468453] env[65726]: WARNING openstack [None req-21960d76-f281-41f6-8236-c885b049da01 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 899.468930] env[65726]: WARNING openstack [None req-21960d76-f281-41f6-8236-c885b049da01 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 899.478400] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-a1a4c202-18b9-4cfc-8067-d677c8415fff tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 899.478824] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-a1a4c202-18b9-4cfc-8067-d677c8415fff tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65726) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 899.480028] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-983e88e1-13dd-4206-94e4-8e863e04550e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.486571] env[65726]: DEBUG oslo_vmware.api [None req-a1a4c202-18b9-4cfc-8067-d677c8415fff tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Waiting for the task: (returnval){ [ 899.486571] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52be2fa9-751d-3346-95fd-9909b2b4c101" [ 899.486571] env[65726]: _type = "Task" [ 899.486571] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.496508] env[65726]: DEBUG oslo_vmware.api [None req-a1a4c202-18b9-4cfc-8067-d677c8415fff tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52be2fa9-751d-3346-95fd-9909b2b4c101, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.623256] env[65726]: DEBUG nova.compute.manager [None req-9c76632b-07ea-476e-92d3-ffb457aef721 tempest-ServerAddressesTestJSON-569100133 tempest-ServerAddressesTestJSON-569100133-project-member] [instance: 70375746-06dd-4710-9136-95e9e5759a0e] Start spawning the instance on the hypervisor. {{(pid=65726) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 899.651656] env[65726]: DEBUG nova.virt.hardware [None req-9c76632b-07ea-476e-92d3-ffb457aef721 tempest-ServerAddressesTestJSON-569100133 tempest-ServerAddressesTestJSON-569100133-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T19:28:34Z,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1539f06fa3534e90acbc3a60c4b8bd3f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T19:28:35Z,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 899.652100] env[65726]: DEBUG nova.virt.hardware [None req-9c76632b-07ea-476e-92d3-ffb457aef721 tempest-ServerAddressesTestJSON-569100133 tempest-ServerAddressesTestJSON-569100133-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 899.652616] env[65726]: DEBUG nova.virt.hardware [None req-9c76632b-07ea-476e-92d3-ffb457aef721 tempest-ServerAddressesTestJSON-569100133 tempest-ServerAddressesTestJSON-569100133-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 899.652616] env[65726]: DEBUG nova.virt.hardware [None req-9c76632b-07ea-476e-92d3-ffb457aef721 tempest-ServerAddressesTestJSON-569100133 tempest-ServerAddressesTestJSON-569100133-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 899.652616] env[65726]: DEBUG nova.virt.hardware [None req-9c76632b-07ea-476e-92d3-ffb457aef721 tempest-ServerAddressesTestJSON-569100133 tempest-ServerAddressesTestJSON-569100133-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 899.652908] env[65726]: DEBUG nova.virt.hardware [None req-9c76632b-07ea-476e-92d3-ffb457aef721 tempest-ServerAddressesTestJSON-569100133 tempest-ServerAddressesTestJSON-569100133-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 899.652990] env[65726]: DEBUG nova.virt.hardware [None req-9c76632b-07ea-476e-92d3-ffb457aef721 tempest-ServerAddressesTestJSON-569100133 tempest-ServerAddressesTestJSON-569100133-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 899.653136] env[65726]: DEBUG nova.virt.hardware [None req-9c76632b-07ea-476e-92d3-ffb457aef721 tempest-ServerAddressesTestJSON-569100133 tempest-ServerAddressesTestJSON-569100133-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 899.653309] env[65726]: DEBUG nova.virt.hardware [None req-9c76632b-07ea-476e-92d3-ffb457aef721 tempest-ServerAddressesTestJSON-569100133 tempest-ServerAddressesTestJSON-569100133-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 899.653470] env[65726]: DEBUG nova.virt.hardware [None req-9c76632b-07ea-476e-92d3-ffb457aef721 tempest-ServerAddressesTestJSON-569100133 tempest-ServerAddressesTestJSON-569100133-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 899.653636] env[65726]: DEBUG nova.virt.hardware [None req-9c76632b-07ea-476e-92d3-ffb457aef721 tempest-ServerAddressesTestJSON-569100133 tempest-ServerAddressesTestJSON-569100133-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 899.654687] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b3f9a46-648a-4a28-8701-1842ef45ab00 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.666289] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b0e418d-cae4-40b5-87e3-ff14816f0535 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.754327] env[65726]: DEBUG nova.scheduler.client.report [None req-253a2ea4-406a-415c-8211-e66a1f71d581 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 899.764126] env[65726]: DEBUG oslo_vmware.api [None req-fcafe3da-7a8e-4b08-9577-72f88a634bf5 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Task: {'id': task-5116165, 'name': RemoveSnapshot_Task} progress is 80%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.914770] env[65726]: DEBUG nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] [instance: fd481728-3ef0-4a96-affd-ab2dd3f596bb] Instance has had 0 of 5 cleanup attempts {{(pid=65726) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 899.942489] env[65726]: DEBUG nova.compute.manager [req-9dad6168-c35a-4131-b068-8a9b3eac895e req-b16a26b0-2280-4509-b1b2-9a696bf017b2 service nova] [instance: 29823bc7-3909-4ab7-8119-91ee59b289c5] Received event network-vif-plugged-ef57c09b-e512-44ad-adeb-8aee81e9dc7b {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 899.942489] env[65726]: DEBUG oslo_concurrency.lockutils [req-9dad6168-c35a-4131-b068-8a9b3eac895e req-b16a26b0-2280-4509-b1b2-9a696bf017b2 service nova] Acquiring lock "29823bc7-3909-4ab7-8119-91ee59b289c5-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 899.942741] env[65726]: DEBUG oslo_concurrency.lockutils [req-9dad6168-c35a-4131-b068-8a9b3eac895e req-b16a26b0-2280-4509-b1b2-9a696bf017b2 service nova] Lock "29823bc7-3909-4ab7-8119-91ee59b289c5-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 899.942930] env[65726]: DEBUG oslo_concurrency.lockutils [req-9dad6168-c35a-4131-b068-8a9b3eac895e req-b16a26b0-2280-4509-b1b2-9a696bf017b2 service nova] Lock "29823bc7-3909-4ab7-8119-91ee59b289c5-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 899.944311] env[65726]: DEBUG nova.compute.manager [req-9dad6168-c35a-4131-b068-8a9b3eac895e req-b16a26b0-2280-4509-b1b2-9a696bf017b2 service nova] [instance: 29823bc7-3909-4ab7-8119-91ee59b289c5] No waiting events found dispatching network-vif-plugged-ef57c09b-e512-44ad-adeb-8aee81e9dc7b {{(pid=65726) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 899.944311] env[65726]: WARNING nova.compute.manager [req-9dad6168-c35a-4131-b068-8a9b3eac895e req-b16a26b0-2280-4509-b1b2-9a696bf017b2 service nova] [instance: 29823bc7-3909-4ab7-8119-91ee59b289c5] Received unexpected event network-vif-plugged-ef57c09b-e512-44ad-adeb-8aee81e9dc7b for instance with vm_state building and task_state spawning. [ 899.985649] env[65726]: WARNING openstack [None req-21960d76-f281-41f6-8236-c885b049da01 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 899.986157] env[65726]: WARNING openstack [None req-21960d76-f281-41f6-8236-c885b049da01 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 899.992852] env[65726]: DEBUG oslo_concurrency.lockutils [None req-21960d76-f281-41f6-8236-c885b049da01 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Acquiring lock "refresh_cache-608cc0ea-3f6b-4b4f-83c1-01aa50999d63" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 899.993054] env[65726]: DEBUG oslo_concurrency.lockutils [None req-21960d76-f281-41f6-8236-c885b049da01 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Acquired lock "refresh_cache-608cc0ea-3f6b-4b4f-83c1-01aa50999d63" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 899.993212] env[65726]: DEBUG nova.network.neutron [None req-21960d76-f281-41f6-8236-c885b049da01 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 608cc0ea-3f6b-4b4f-83c1-01aa50999d63] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 899.993388] env[65726]: DEBUG nova.objects.instance [None req-21960d76-f281-41f6-8236-c885b049da01 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Lazy-loading 'info_cache' on Instance uuid 608cc0ea-3f6b-4b4f-83c1-01aa50999d63 {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 900.007545] env[65726]: DEBUG oslo_vmware.api [None req-a1a4c202-18b9-4cfc-8067-d677c8415fff tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52be2fa9-751d-3346-95fd-9909b2b4c101, 'name': SearchDatastore_Task, 'duration_secs': 0.012214} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.009599] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7bcaf80a-14e4-4ed7-a102-c2b3cf887cb1 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.018574] env[65726]: DEBUG oslo_vmware.api [None req-a1a4c202-18b9-4cfc-8067-d677c8415fff tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Waiting for the task: (returnval){ [ 900.018574] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52496a41-84ed-7187-b53c-3a0f249a6384" [ 900.018574] env[65726]: _type = "Task" [ 900.018574] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.028651] env[65726]: DEBUG oslo_vmware.api [None req-a1a4c202-18b9-4cfc-8067-d677c8415fff tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52496a41-84ed-7187-b53c-3a0f249a6384, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.097514] env[65726]: WARNING openstack [None req-4826faf0-9908-4f3c-a509-20fe2a708723 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 900.097514] env[65726]: WARNING openstack [None req-4826faf0-9908-4f3c-a509-20fe2a708723 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 900.113165] env[65726]: DEBUG nova.network.neutron [None req-9c76632b-07ea-476e-92d3-ffb457aef721 tempest-ServerAddressesTestJSON-569100133 tempest-ServerAddressesTestJSON-569100133-project-member] [instance: 70375746-06dd-4710-9136-95e9e5759a0e] Successfully updated port: 7818914d-352b-408a-9d0f-4c2bb1a25f2e {{(pid=65726) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 900.263502] env[65726]: DEBUG oslo_vmware.api [None req-fcafe3da-7a8e-4b08-9577-72f88a634bf5 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Task: {'id': task-5116165, 'name': RemoveSnapshot_Task, 'duration_secs': 0.935577} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.264834] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-fcafe3da-7a8e-4b08-9577-72f88a634bf5 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] [instance: c4cc4f85-cb35-4edc-a58b-adfee0ce1265] Deleted Snapshot of the VM instance {{(pid=65726) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 900.265144] env[65726]: INFO nova.compute.manager [None req-fcafe3da-7a8e-4b08-9577-72f88a634bf5 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] [instance: c4cc4f85-cb35-4edc-a58b-adfee0ce1265] Took 17.61 seconds to snapshot the instance on the hypervisor. [ 900.419229] env[65726]: DEBUG nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] [instance: 5a8d4357-4b44-4a19-b1da-42d188c38adc] Instance has had 0 of 5 cleanup attempts {{(pid=65726) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 900.530709] env[65726]: DEBUG oslo_vmware.api [None req-a1a4c202-18b9-4cfc-8067-d677c8415fff tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52496a41-84ed-7187-b53c-3a0f249a6384, 'name': SearchDatastore_Task, 'duration_secs': 0.012443} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.531078] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a1a4c202-18b9-4cfc-8067-d677c8415fff tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 900.531351] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-a1a4c202-18b9-4cfc-8067-d677c8415fff tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore2] 815ee20b-cb07-4514-bd43-1b74972508a5/815ee20b-cb07-4514-bd43-1b74972508a5.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 900.531634] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-26518581-fd6a-4744-b3cf-e1e692ac6b2a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.539827] env[65726]: DEBUG oslo_vmware.api [None req-a1a4c202-18b9-4cfc-8067-d677c8415fff tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Waiting for the task: (returnval){ [ 900.539827] env[65726]: value = "task-5116166" [ 900.539827] env[65726]: _type = "Task" [ 900.539827] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.553965] env[65726]: DEBUG oslo_vmware.api [None req-a1a4c202-18b9-4cfc-8067-d677c8415fff tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Task: {'id': task-5116166, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.622920] env[65726]: WARNING openstack [None req-4826faf0-9908-4f3c-a509-20fe2a708723 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 900.623755] env[65726]: WARNING openstack [None req-4826faf0-9908-4f3c-a509-20fe2a708723 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 900.634054] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9c76632b-07ea-476e-92d3-ffb457aef721 tempest-ServerAddressesTestJSON-569100133 tempest-ServerAddressesTestJSON-569100133-project-member] Acquiring lock "refresh_cache-70375746-06dd-4710-9136-95e9e5759a0e" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 900.634223] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9c76632b-07ea-476e-92d3-ffb457aef721 tempest-ServerAddressesTestJSON-569100133 tempest-ServerAddressesTestJSON-569100133-project-member] Acquired lock "refresh_cache-70375746-06dd-4710-9136-95e9e5759a0e" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 900.634389] env[65726]: DEBUG nova.network.neutron [None req-9c76632b-07ea-476e-92d3-ffb457aef721 tempest-ServerAddressesTestJSON-569100133 tempest-ServerAddressesTestJSON-569100133-project-member] [instance: 70375746-06dd-4710-9136-95e9e5759a0e] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 900.760832] env[65726]: DEBUG nova.network.neutron [None req-4826faf0-9908-4f3c-a509-20fe2a708723 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] [instance: 29823bc7-3909-4ab7-8119-91ee59b289c5] Updating instance_info_cache with network_info: [{"id": "ef57c09b-e512-44ad-adeb-8aee81e9dc7b", "address": "fa:16:3e:d9:3e:40", "network": {"id": "55d629b8-a541-4720-9289-927bbc6e6f22", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.50", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "1539f06fa3534e90acbc3a60c4b8bd3f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "35342bcb-8b06-472e-b3c0-43fd3d6c4b30", "external-id": "nsx-vlan-transportzone-524", "segmentation_id": 524, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapef57c09b-e5", "ovs_interfaceid": "ef57c09b-e512-44ad-adeb-8aee81e9dc7b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 900.766385] env[65726]: DEBUG oslo_concurrency.lockutils [None req-253a2ea4-406a-415c-8211-e66a1f71d581 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: held 3.176s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 900.774252] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e5e88d5e-aa68-4ff5-9280-b2cfe1d14b27 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 31.639s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 900.774974] env[65726]: INFO nova.compute.claims [None req-e5e88d5e-aa68-4ff5-9280-b2cfe1d14b27 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: 5a252ef1-93c9-4bff-842b-b64df2bc5d75] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 900.926424] env[65726]: DEBUG nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] [instance: 3b933684-ac19-44b0-a49d-6af45501e166] Instance has had 0 of 5 cleanup attempts {{(pid=65726) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 901.000283] env[65726]: WARNING openstack [None req-21960d76-f281-41f6-8236-c885b049da01 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 901.000681] env[65726]: WARNING openstack [None req-21960d76-f281-41f6-8236-c885b049da01 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 901.054253] env[65726]: DEBUG oslo_vmware.api [None req-a1a4c202-18b9-4cfc-8067-d677c8415fff tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Task: {'id': task-5116166, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.140862] env[65726]: WARNING openstack [None req-9c76632b-07ea-476e-92d3-ffb457aef721 tempest-ServerAddressesTestJSON-569100133 tempest-ServerAddressesTestJSON-569100133-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 901.141534] env[65726]: WARNING openstack [None req-9c76632b-07ea-476e-92d3-ffb457aef721 tempest-ServerAddressesTestJSON-569100133 tempest-ServerAddressesTestJSON-569100133-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 901.148487] env[65726]: DEBUG nova.network.neutron [None req-9c76632b-07ea-476e-92d3-ffb457aef721 tempest-ServerAddressesTestJSON-569100133 tempest-ServerAddressesTestJSON-569100133-project-member] [instance: 70375746-06dd-4710-9136-95e9e5759a0e] Instance cache missing network info. {{(pid=65726) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 901.155471] env[65726]: DEBUG nova.compute.manager [req-951802b2-3216-407b-b531-3d7188ed02a1 req-8cfe1f61-93a9-4f86-b842-ab19b679f3ba service nova] [instance: 815ee20b-cb07-4514-bd43-1b74972508a5] Received event network-changed-c5f4ccbb-2fbd-43ef-979c-d7538b388e9a {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 901.156141] env[65726]: DEBUG nova.compute.manager [req-951802b2-3216-407b-b531-3d7188ed02a1 req-8cfe1f61-93a9-4f86-b842-ab19b679f3ba service nova] [instance: 815ee20b-cb07-4514-bd43-1b74972508a5] Refreshing instance network info cache due to event network-changed-c5f4ccbb-2fbd-43ef-979c-d7538b388e9a. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 901.156141] env[65726]: DEBUG oslo_concurrency.lockutils [req-951802b2-3216-407b-b531-3d7188ed02a1 req-8cfe1f61-93a9-4f86-b842-ab19b679f3ba service nova] Acquiring lock "refresh_cache-815ee20b-cb07-4514-bd43-1b74972508a5" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 901.156366] env[65726]: DEBUG oslo_concurrency.lockutils [req-951802b2-3216-407b-b531-3d7188ed02a1 req-8cfe1f61-93a9-4f86-b842-ab19b679f3ba service nova] Acquired lock "refresh_cache-815ee20b-cb07-4514-bd43-1b74972508a5" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 901.156638] env[65726]: DEBUG nova.network.neutron [req-951802b2-3216-407b-b531-3d7188ed02a1 req-8cfe1f61-93a9-4f86-b842-ab19b679f3ba service nova] [instance: 815ee20b-cb07-4514-bd43-1b74972508a5] Refreshing network info cache for port c5f4ccbb-2fbd-43ef-979c-d7538b388e9a {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 901.165238] env[65726]: WARNING openstack [None req-21960d76-f281-41f6-8236-c885b049da01 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 901.165657] env[65726]: WARNING openstack [None req-21960d76-f281-41f6-8236-c885b049da01 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 901.227414] env[65726]: WARNING openstack [None req-9c76632b-07ea-476e-92d3-ffb457aef721 tempest-ServerAddressesTestJSON-569100133 tempest-ServerAddressesTestJSON-569100133-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 901.227823] env[65726]: WARNING openstack [None req-9c76632b-07ea-476e-92d3-ffb457aef721 tempest-ServerAddressesTestJSON-569100133 tempest-ServerAddressesTestJSON-569100133-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 901.265770] env[65726]: DEBUG oslo_concurrency.lockutils [None req-4826faf0-9908-4f3c-a509-20fe2a708723 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Releasing lock "refresh_cache-29823bc7-3909-4ab7-8119-91ee59b289c5" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 901.266175] env[65726]: DEBUG nova.compute.manager [None req-4826faf0-9908-4f3c-a509-20fe2a708723 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] [instance: 29823bc7-3909-4ab7-8119-91ee59b289c5] Instance network_info: |[{"id": "ef57c09b-e512-44ad-adeb-8aee81e9dc7b", "address": "fa:16:3e:d9:3e:40", "network": {"id": "55d629b8-a541-4720-9289-927bbc6e6f22", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.50", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "1539f06fa3534e90acbc3a60c4b8bd3f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "35342bcb-8b06-472e-b3c0-43fd3d6c4b30", "external-id": "nsx-vlan-transportzone-524", "segmentation_id": 524, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapef57c09b-e5", "ovs_interfaceid": "ef57c09b-e512-44ad-adeb-8aee81e9dc7b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 901.266655] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-4826faf0-9908-4f3c-a509-20fe2a708723 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] [instance: 29823bc7-3909-4ab7-8119-91ee59b289c5] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d9:3e:40', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '35342bcb-8b06-472e-b3c0-43fd3d6c4b30', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ef57c09b-e512-44ad-adeb-8aee81e9dc7b', 'vif_model': 'vmxnet3'}] {{(pid=65726) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 901.275867] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-4826faf0-9908-4f3c-a509-20fe2a708723 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 901.276170] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 29823bc7-3909-4ab7-8119-91ee59b289c5] Creating VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 901.276513] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-bef5416a-01c9-4c24-982b-84e9be56d092 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.310654] env[65726]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 901.310654] env[65726]: value = "task-5116167" [ 901.310654] env[65726]: _type = "Task" [ 901.310654] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.323965] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116167, 'name': CreateVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.334597] env[65726]: WARNING openstack [None req-21960d76-f281-41f6-8236-c885b049da01 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 901.335283] env[65726]: WARNING openstack [None req-21960d76-f281-41f6-8236-c885b049da01 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 901.432078] env[65726]: DEBUG nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] [instance: 7748c23f-9ea9-4172-8ab7-187678272bb3] Instance has had 0 of 5 cleanup attempts {{(pid=65726) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 901.530350] env[65726]: WARNING openstack [None req-9c76632b-07ea-476e-92d3-ffb457aef721 tempest-ServerAddressesTestJSON-569100133 tempest-ServerAddressesTestJSON-569100133-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 901.531781] env[65726]: WARNING openstack [None req-9c76632b-07ea-476e-92d3-ffb457aef721 tempest-ServerAddressesTestJSON-569100133 tempest-ServerAddressesTestJSON-569100133-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 901.554218] env[65726]: DEBUG oslo_vmware.api [None req-a1a4c202-18b9-4cfc-8067-d677c8415fff tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Task: {'id': task-5116166, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.921871} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.554500] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-a1a4c202-18b9-4cfc-8067-d677c8415fff tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore2] 815ee20b-cb07-4514-bd43-1b74972508a5/815ee20b-cb07-4514-bd43-1b74972508a5.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 901.554719] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-a1a4c202-18b9-4cfc-8067-d677c8415fff tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] [instance: 815ee20b-cb07-4514-bd43-1b74972508a5] Extending root virtual disk to 1048576 {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 901.554986] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ebb46b83-68cf-47e9-92a0-9de398aa3a93 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.563877] env[65726]: DEBUG nova.network.neutron [None req-21960d76-f281-41f6-8236-c885b049da01 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 608cc0ea-3f6b-4b4f-83c1-01aa50999d63] Updating instance_info_cache with network_info: [{"id": "e9f7132a-8d5a-47fe-afc6-d45bf3687ea4", "address": "fa:16:3e:fe:4d:df", "network": {"id": "0d0f561d-ac41-4ef5-9926-c707409799f7", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1134321217-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.160", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4eece77569624f90bf64e5c51974173f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2ce62383-8e84-4e26-955b-74c11392f4c9", "external-id": "nsx-vlan-transportzone-215", "segmentation_id": 215, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape9f7132a-8d", "ovs_interfaceid": "e9f7132a-8d5a-47fe-afc6-d45bf3687ea4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 901.567546] env[65726]: DEBUG oslo_vmware.api [None req-a1a4c202-18b9-4cfc-8067-d677c8415fff tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Waiting for the task: (returnval){ [ 901.567546] env[65726]: value = "task-5116168" [ 901.567546] env[65726]: _type = "Task" [ 901.567546] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.586662] env[65726]: DEBUG oslo_vmware.api [None req-a1a4c202-18b9-4cfc-8067-d677c8415fff tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Task: {'id': task-5116168, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.673511] env[65726]: DEBUG nova.network.neutron [None req-9c76632b-07ea-476e-92d3-ffb457aef721 tempest-ServerAddressesTestJSON-569100133 tempest-ServerAddressesTestJSON-569100133-project-member] [instance: 70375746-06dd-4710-9136-95e9e5759a0e] Updating instance_info_cache with network_info: [{"id": "7818914d-352b-408a-9d0f-4c2bb1a25f2e", "address": "fa:16:3e:8a:3a:b3", "network": {"id": "56a899e2-6ddb-4f9f-bcf6-0f1905f60e6f", "bridge": "br-int", "label": "tempest-ServerAddressesTestJSON-445774747-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "04fbc0c66f094d73b87bccb72aa6f37a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8c979f78-8597-41f8-b1de-995014032689", "external-id": "nsx-vlan-transportzone-477", "segmentation_id": 477, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7818914d-35", "ovs_interfaceid": "7818914d-352b-408a-9d0f-4c2bb1a25f2e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 901.676789] env[65726]: WARNING openstack [req-951802b2-3216-407b-b531-3d7188ed02a1 req-8cfe1f61-93a9-4f86-b842-ab19b679f3ba service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 901.677178] env[65726]: WARNING openstack [req-951802b2-3216-407b-b531-3d7188ed02a1 req-8cfe1f61-93a9-4f86-b842-ab19b679f3ba service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 901.824320] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116167, 'name': CreateVM_Task, 'duration_secs': 0.396831} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.826093] env[65726]: WARNING openstack [req-951802b2-3216-407b-b531-3d7188ed02a1 req-8cfe1f61-93a9-4f86-b842-ab19b679f3ba service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 901.828708] env[65726]: WARNING openstack [req-951802b2-3216-407b-b531-3d7188ed02a1 req-8cfe1f61-93a9-4f86-b842-ab19b679f3ba service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 901.835154] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 29823bc7-3909-4ab7-8119-91ee59b289c5] Created VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 901.836366] env[65726]: WARNING openstack [None req-4826faf0-9908-4f3c-a509-20fe2a708723 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 901.836614] env[65726]: WARNING openstack [None req-4826faf0-9908-4f3c-a509-20fe2a708723 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 901.841482] env[65726]: DEBUG oslo_concurrency.lockutils [None req-4826faf0-9908-4f3c-a509-20fe2a708723 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 901.841638] env[65726]: DEBUG oslo_concurrency.lockutils [None req-4826faf0-9908-4f3c-a509-20fe2a708723 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 901.841970] env[65726]: DEBUG oslo_concurrency.lockutils [None req-4826faf0-9908-4f3c-a509-20fe2a708723 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 901.844907] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dd64b8cd-7f34-4e87-9b73-0d10a13907f8 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.852047] env[65726]: DEBUG oslo_vmware.api [None req-4826faf0-9908-4f3c-a509-20fe2a708723 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Waiting for the task: (returnval){ [ 901.852047] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]525b6618-be3a-f68f-656a-cc8f285617e1" [ 901.852047] env[65726]: _type = "Task" [ 901.852047] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.860788] env[65726]: DEBUG oslo_vmware.api [None req-4826faf0-9908-4f3c-a509-20fe2a708723 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]525b6618-be3a-f68f-656a-cc8f285617e1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.936174] env[65726]: DEBUG nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] [instance: f5151062-57b1-4e4e-93f4-aab0e7f504d4] Instance has had 0 of 5 cleanup attempts {{(pid=65726) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 901.949164] env[65726]: WARNING openstack [req-951802b2-3216-407b-b531-3d7188ed02a1 req-8cfe1f61-93a9-4f86-b842-ab19b679f3ba service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 901.949807] env[65726]: WARNING openstack [req-951802b2-3216-407b-b531-3d7188ed02a1 req-8cfe1f61-93a9-4f86-b842-ab19b679f3ba service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 902.066910] env[65726]: DEBUG oslo_concurrency.lockutils [None req-21960d76-f281-41f6-8236-c885b049da01 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Releasing lock "refresh_cache-608cc0ea-3f6b-4b4f-83c1-01aa50999d63" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 902.067085] env[65726]: DEBUG nova.objects.instance [None req-21960d76-f281-41f6-8236-c885b049da01 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Lazy-loading 'migration_context' on Instance uuid 608cc0ea-3f6b-4b4f-83c1-01aa50999d63 {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 902.079999] env[65726]: DEBUG oslo_vmware.api [None req-a1a4c202-18b9-4cfc-8067-d677c8415fff tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Task: {'id': task-5116168, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.085546} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.080308] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-a1a4c202-18b9-4cfc-8067-d677c8415fff tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] [instance: 815ee20b-cb07-4514-bd43-1b74972508a5] Extended root virtual disk {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 902.081670] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-657de7cf-1c0f-4445-bdcc-6da17dbd6bb2 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.118897] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-a1a4c202-18b9-4cfc-8067-d677c8415fff tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] [instance: 815ee20b-cb07-4514-bd43-1b74972508a5] Reconfiguring VM instance instance-00000041 to attach disk [datastore2] 815ee20b-cb07-4514-bd43-1b74972508a5/815ee20b-cb07-4514-bd43-1b74972508a5.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 902.124054] env[65726]: DEBUG nova.network.neutron [req-951802b2-3216-407b-b531-3d7188ed02a1 req-8cfe1f61-93a9-4f86-b842-ab19b679f3ba service nova] [instance: 815ee20b-cb07-4514-bd43-1b74972508a5] Updated VIF entry in instance network info cache for port c5f4ccbb-2fbd-43ef-979c-d7538b388e9a. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 902.124410] env[65726]: DEBUG nova.network.neutron [req-951802b2-3216-407b-b531-3d7188ed02a1 req-8cfe1f61-93a9-4f86-b842-ab19b679f3ba service nova] [instance: 815ee20b-cb07-4514-bd43-1b74972508a5] Updating instance_info_cache with network_info: [{"id": "c5f4ccbb-2fbd-43ef-979c-d7538b388e9a", "address": "fa:16:3e:c4:a5:f9", "network": {"id": "55d629b8-a541-4720-9289-927bbc6e6f22", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.68", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "1539f06fa3534e90acbc3a60c4b8bd3f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "35342bcb-8b06-472e-b3c0-43fd3d6c4b30", "external-id": "nsx-vlan-transportzone-524", "segmentation_id": 524, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc5f4ccbb-2f", "ovs_interfaceid": "c5f4ccbb-2fbd-43ef-979c-d7538b388e9a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 902.126902] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-065f122a-b539-4a2c-8885-89650aea4870 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.146733] env[65726]: DEBUG oslo_concurrency.lockutils [None req-140eec77-ac75-4bcd-8eca-c7b64f5e10aa tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Acquiring lock "4ea2b70b-0cfe-4fbf-817a-baa028b8372a" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 902.147257] env[65726]: DEBUG oslo_concurrency.lockutils [None req-140eec77-ac75-4bcd-8eca-c7b64f5e10aa tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Lock "4ea2b70b-0cfe-4fbf-817a-baa028b8372a" acquired by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 902.147257] env[65726]: INFO nova.compute.manager [None req-140eec77-ac75-4bcd-8eca-c7b64f5e10aa tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] [instance: 4ea2b70b-0cfe-4fbf-817a-baa028b8372a] Rebooting instance [ 902.151019] env[65726]: DEBUG oslo_vmware.api [None req-a1a4c202-18b9-4cfc-8067-d677c8415fff tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Waiting for the task: (returnval){ [ 902.151019] env[65726]: value = "task-5116169" [ 902.151019] env[65726]: _type = "Task" [ 902.151019] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.161773] env[65726]: DEBUG oslo_vmware.api [None req-a1a4c202-18b9-4cfc-8067-d677c8415fff tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Task: {'id': task-5116169, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.176372] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9c76632b-07ea-476e-92d3-ffb457aef721 tempest-ServerAddressesTestJSON-569100133 tempest-ServerAddressesTestJSON-569100133-project-member] Releasing lock "refresh_cache-70375746-06dd-4710-9136-95e9e5759a0e" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 902.176548] env[65726]: DEBUG nova.compute.manager [None req-9c76632b-07ea-476e-92d3-ffb457aef721 tempest-ServerAddressesTestJSON-569100133 tempest-ServerAddressesTestJSON-569100133-project-member] [instance: 70375746-06dd-4710-9136-95e9e5759a0e] Instance network_info: |[{"id": "7818914d-352b-408a-9d0f-4c2bb1a25f2e", "address": "fa:16:3e:8a:3a:b3", "network": {"id": "56a899e2-6ddb-4f9f-bcf6-0f1905f60e6f", "bridge": "br-int", "label": "tempest-ServerAddressesTestJSON-445774747-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "04fbc0c66f094d73b87bccb72aa6f37a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8c979f78-8597-41f8-b1de-995014032689", "external-id": "nsx-vlan-transportzone-477", "segmentation_id": 477, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7818914d-35", "ovs_interfaceid": "7818914d-352b-408a-9d0f-4c2bb1a25f2e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 902.177026] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-9c76632b-07ea-476e-92d3-ffb457aef721 tempest-ServerAddressesTestJSON-569100133 tempest-ServerAddressesTestJSON-569100133-project-member] [instance: 70375746-06dd-4710-9136-95e9e5759a0e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:8a:3a:b3', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '8c979f78-8597-41f8-b1de-995014032689', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7818914d-352b-408a-9d0f-4c2bb1a25f2e', 'vif_model': 'vmxnet3'}] {{(pid=65726) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 902.185955] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-9c76632b-07ea-476e-92d3-ffb457aef721 tempest-ServerAddressesTestJSON-569100133 tempest-ServerAddressesTestJSON-569100133-project-member] Creating folder: Project (04fbc0c66f094d73b87bccb72aa6f37a). Parent ref: group-v995008. {{(pid=65726) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 902.186821] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-fdae2665-8c42-473a-a787-17266eea2790 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.201463] env[65726]: INFO nova.virt.vmwareapi.vm_util [None req-9c76632b-07ea-476e-92d3-ffb457aef721 tempest-ServerAddressesTestJSON-569100133 tempest-ServerAddressesTestJSON-569100133-project-member] Created folder: Project (04fbc0c66f094d73b87bccb72aa6f37a) in parent group-v995008. [ 902.201712] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-9c76632b-07ea-476e-92d3-ffb457aef721 tempest-ServerAddressesTestJSON-569100133 tempest-ServerAddressesTestJSON-569100133-project-member] Creating folder: Instances. Parent ref: group-v995194. {{(pid=65726) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 902.201939] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-6dfc3845-caf9-4d57-83b2-2e3eb5be9d28 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.219432] env[65726]: INFO nova.virt.vmwareapi.vm_util [None req-9c76632b-07ea-476e-92d3-ffb457aef721 tempest-ServerAddressesTestJSON-569100133 tempest-ServerAddressesTestJSON-569100133-project-member] Created folder: Instances in parent group-v995194. [ 902.220144] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-9c76632b-07ea-476e-92d3-ffb457aef721 tempest-ServerAddressesTestJSON-569100133 tempest-ServerAddressesTestJSON-569100133-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 902.220144] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 70375746-06dd-4710-9136-95e9e5759a0e] Creating VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 902.220290] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8ed6dba7-0717-4473-b7d0-950e94207247 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.249877] env[65726]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 902.249877] env[65726]: value = "task-5116172" [ 902.249877] env[65726]: _type = "Task" [ 902.249877] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.261683] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116172, 'name': CreateVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.321772] env[65726]: DEBUG nova.compute.manager [req-344d1f3c-8710-4831-9fe1-b8d63b07998e req-15bc8f76-833a-4878-958b-5d5cc85edb88 service nova] [instance: 29823bc7-3909-4ab7-8119-91ee59b289c5] Received event network-changed-ef57c09b-e512-44ad-adeb-8aee81e9dc7b {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 902.322170] env[65726]: DEBUG nova.compute.manager [req-344d1f3c-8710-4831-9fe1-b8d63b07998e req-15bc8f76-833a-4878-958b-5d5cc85edb88 service nova] [instance: 29823bc7-3909-4ab7-8119-91ee59b289c5] Refreshing instance network info cache due to event network-changed-ef57c09b-e512-44ad-adeb-8aee81e9dc7b. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 902.322708] env[65726]: DEBUG oslo_concurrency.lockutils [req-344d1f3c-8710-4831-9fe1-b8d63b07998e req-15bc8f76-833a-4878-958b-5d5cc85edb88 service nova] Acquiring lock "refresh_cache-29823bc7-3909-4ab7-8119-91ee59b289c5" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 902.322868] env[65726]: DEBUG oslo_concurrency.lockutils [req-344d1f3c-8710-4831-9fe1-b8d63b07998e req-15bc8f76-833a-4878-958b-5d5cc85edb88 service nova] Acquired lock "refresh_cache-29823bc7-3909-4ab7-8119-91ee59b289c5" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 902.323051] env[65726]: DEBUG nova.network.neutron [req-344d1f3c-8710-4831-9fe1-b8d63b07998e req-15bc8f76-833a-4878-958b-5d5cc85edb88 service nova] [instance: 29823bc7-3909-4ab7-8119-91ee59b289c5] Refreshing network info cache for port ef57c09b-e512-44ad-adeb-8aee81e9dc7b {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 902.335089] env[65726]: INFO nova.compute.manager [None req-253a2ea4-406a-415c-8211-e66a1f71d581 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] [instance: 76249623-6f83-46a3-b8c5-c001111aa698] Swapping old allocation on dict_keys(['07c4692f-bdb4-4058-9173-ff9664830295']) held by migration 8f31ab13-33c9-4d52-9a2a-8c3409ce12a9 for instance [ 902.367539] env[65726]: DEBUG oslo_vmware.api [None req-4826faf0-9908-4f3c-a509-20fe2a708723 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]525b6618-be3a-f68f-656a-cc8f285617e1, 'name': SearchDatastore_Task, 'duration_secs': 0.01852} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.371227] env[65726]: DEBUG oslo_concurrency.lockutils [None req-4826faf0-9908-4f3c-a509-20fe2a708723 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 902.371227] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-4826faf0-9908-4f3c-a509-20fe2a708723 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] [instance: 29823bc7-3909-4ab7-8119-91ee59b289c5] Processing image b52362a3-ee8a-4cbf-b06f-513b0cc8f95c {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 902.371413] env[65726]: DEBUG oslo_concurrency.lockutils [None req-4826faf0-9908-4f3c-a509-20fe2a708723 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 902.371515] env[65726]: DEBUG oslo_concurrency.lockutils [None req-4826faf0-9908-4f3c-a509-20fe2a708723 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 902.371779] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-4826faf0-9908-4f3c-a509-20fe2a708723 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 902.372775] env[65726]: DEBUG nova.scheduler.client.report [None req-253a2ea4-406a-415c-8211-e66a1f71d581 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Overwriting current allocation {'allocations': {'07c4692f-bdb4-4058-9173-ff9664830295': {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}, 'generation': 100}}, 'project_id': 'fae9ef0198a2468a9793374b76378abf', 'user_id': '58e5e1469d2241d3ad3178f085a7bee6', 'consumer_generation': 1} on consumer 76249623-6f83-46a3-b8c5-c001111aa698 {{(pid=65726) move_allocations /opt/stack/nova/nova/scheduler/client/report.py:2036}} [ 902.374900] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-63c6ca56-0097-4b22-adef-34dee89b4031 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.388622] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-4826faf0-9908-4f3c-a509-20fe2a708723 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 902.388622] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-4826faf0-9908-4f3c-a509-20fe2a708723 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65726) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 902.388622] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a49dd257-0723-4aea-81b0-456125277316 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.397415] env[65726]: DEBUG oslo_vmware.api [None req-4826faf0-9908-4f3c-a509-20fe2a708723 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Waiting for the task: (returnval){ [ 902.397415] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52b1f46b-2de3-2ab1-6a46-3cc40a1268de" [ 902.397415] env[65726]: _type = "Task" [ 902.397415] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.407200] env[65726]: DEBUG oslo_vmware.api [None req-4826faf0-9908-4f3c-a509-20fe2a708723 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52b1f46b-2de3-2ab1-6a46-3cc40a1268de, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.440526] env[65726]: DEBUG nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] [instance: 3a60b719-9db7-436e-9908-25f6a233c465] Instance has had 0 of 5 cleanup attempts {{(pid=65726) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 902.463347] env[65726]: WARNING openstack [None req-253a2ea4-406a-415c-8211-e66a1f71d581 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 902.463744] env[65726]: WARNING openstack [None req-253a2ea4-406a-415c-8211-e66a1f71d581 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 902.514098] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fab20cc-46c5-4a36-b77f-6e7b254d4adf {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.523634] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a842a8dd-23d2-49b0-bf88-9ed610f07f2e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.529680] env[65726]: DEBUG oslo_concurrency.lockutils [None req-253a2ea4-406a-415c-8211-e66a1f71d581 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Acquiring lock "refresh_cache-76249623-6f83-46a3-b8c5-c001111aa698" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 902.529853] env[65726]: DEBUG oslo_concurrency.lockutils [None req-253a2ea4-406a-415c-8211-e66a1f71d581 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Acquired lock "refresh_cache-76249623-6f83-46a3-b8c5-c001111aa698" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 902.530031] env[65726]: DEBUG nova.network.neutron [None req-253a2ea4-406a-415c-8211-e66a1f71d581 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] [instance: 76249623-6f83-46a3-b8c5-c001111aa698] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 902.562694] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d81d7159-3348-4a04-bfc9-502e44dc6b62 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.571706] env[65726]: DEBUG nova.objects.base [None req-21960d76-f281-41f6-8236-c885b049da01 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Object Instance<608cc0ea-3f6b-4b4f-83c1-01aa50999d63> lazy-loaded attributes: info_cache,migration_context {{(pid=65726) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 902.572733] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb27eda6-d7bd-4fa8-9f82-9180c9960a4b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.576927] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45751466-1220-4679-8870-402327b69791 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.615523] env[65726]: DEBUG nova.compute.provider_tree [None req-e5e88d5e-aa68-4ff5-9280-b2cfe1d14b27 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 902.618599] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b8494057-9208-4dfb-9d03-8246fc6ce621 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.624784] env[65726]: DEBUG oslo_vmware.api [None req-21960d76-f281-41f6-8236-c885b049da01 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Waiting for the task: (returnval){ [ 902.624784] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]523b1e2f-7676-cd7e-f273-d9de86fdb2f9" [ 902.624784] env[65726]: _type = "Task" [ 902.624784] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.628709] env[65726]: DEBUG oslo_concurrency.lockutils [req-951802b2-3216-407b-b531-3d7188ed02a1 req-8cfe1f61-93a9-4f86-b842-ab19b679f3ba service nova] Releasing lock "refresh_cache-815ee20b-cb07-4514-bd43-1b74972508a5" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 902.629010] env[65726]: DEBUG nova.compute.manager [req-951802b2-3216-407b-b531-3d7188ed02a1 req-8cfe1f61-93a9-4f86-b842-ab19b679f3ba service nova] [instance: cb8855b6-589d-4863-b86f-7fb4e30fb29e] Received event network-vif-deleted-136da8b5-d8fd-4cef-ae44-a9ecc98c28d2 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 902.629259] env[65726]: DEBUG nova.compute.manager [req-951802b2-3216-407b-b531-3d7188ed02a1 req-8cfe1f61-93a9-4f86-b842-ab19b679f3ba service nova] [instance: 6c8db442-daaa-4eca-ae24-b9d4afaf3a71] Received event network-vif-deleted-6f422d37-8262-454d-aa2d-1569bacee1b7 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 902.629360] env[65726]: DEBUG nova.compute.manager [req-951802b2-3216-407b-b531-3d7188ed02a1 req-8cfe1f61-93a9-4f86-b842-ab19b679f3ba service nova] [instance: 4ea2b70b-0cfe-4fbf-817a-baa028b8372a] Received event network-changed-a97348ff-93e0-442b-8a41-b56fa04e8cb6 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 902.629509] env[65726]: DEBUG nova.compute.manager [req-951802b2-3216-407b-b531-3d7188ed02a1 req-8cfe1f61-93a9-4f86-b842-ab19b679f3ba service nova] [instance: 4ea2b70b-0cfe-4fbf-817a-baa028b8372a] Refreshing instance network info cache due to event network-changed-a97348ff-93e0-442b-8a41-b56fa04e8cb6. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 902.629711] env[65726]: DEBUG oslo_concurrency.lockutils [req-951802b2-3216-407b-b531-3d7188ed02a1 req-8cfe1f61-93a9-4f86-b842-ab19b679f3ba service nova] Acquiring lock "refresh_cache-4ea2b70b-0cfe-4fbf-817a-baa028b8372a" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 902.629873] env[65726]: DEBUG oslo_concurrency.lockutils [req-951802b2-3216-407b-b531-3d7188ed02a1 req-8cfe1f61-93a9-4f86-b842-ab19b679f3ba service nova] Acquired lock "refresh_cache-4ea2b70b-0cfe-4fbf-817a-baa028b8372a" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 902.630096] env[65726]: DEBUG nova.network.neutron [req-951802b2-3216-407b-b531-3d7188ed02a1 req-8cfe1f61-93a9-4f86-b842-ab19b679f3ba service nova] [instance: 4ea2b70b-0cfe-4fbf-817a-baa028b8372a] Refreshing network info cache for port a97348ff-93e0-442b-8a41-b56fa04e8cb6 {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 902.639786] env[65726]: DEBUG oslo_vmware.api [None req-21960d76-f281-41f6-8236-c885b049da01 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]523b1e2f-7676-cd7e-f273-d9de86fdb2f9, 'name': SearchDatastore_Task, 'duration_secs': 0.011878} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.640109] env[65726]: DEBUG oslo_concurrency.lockutils [None req-21960d76-f281-41f6-8236-c885b049da01 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 902.664621] env[65726]: DEBUG oslo_vmware.api [None req-a1a4c202-18b9-4cfc-8067-d677c8415fff tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Task: {'id': task-5116169, 'name': ReconfigVM_Task, 'duration_secs': 0.312181} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.665128] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-a1a4c202-18b9-4cfc-8067-d677c8415fff tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] [instance: 815ee20b-cb07-4514-bd43-1b74972508a5] Reconfigured VM instance instance-00000041 to attach disk [datastore2] 815ee20b-cb07-4514-bd43-1b74972508a5/815ee20b-cb07-4514-bd43-1b74972508a5.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 902.665893] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-787c0a27-4047-4069-8b53-8a0341ccb0f5 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.675155] env[65726]: DEBUG oslo_vmware.api [None req-a1a4c202-18b9-4cfc-8067-d677c8415fff tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Waiting for the task: (returnval){ [ 902.675155] env[65726]: value = "task-5116173" [ 902.675155] env[65726]: _type = "Task" [ 902.675155] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.676445] env[65726]: DEBUG oslo_concurrency.lockutils [None req-140eec77-ac75-4bcd-8eca-c7b64f5e10aa tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Acquiring lock "refresh_cache-4ea2b70b-0cfe-4fbf-817a-baa028b8372a" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 902.686823] env[65726]: DEBUG oslo_vmware.api [None req-a1a4c202-18b9-4cfc-8067-d677c8415fff tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Task: {'id': task-5116173, 'name': Rename_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.760528] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116172, 'name': CreateVM_Task, 'duration_secs': 0.408954} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.760883] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 70375746-06dd-4710-9136-95e9e5759a0e] Created VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 902.762784] env[65726]: WARNING openstack [None req-9c76632b-07ea-476e-92d3-ffb457aef721 tempest-ServerAddressesTestJSON-569100133 tempest-ServerAddressesTestJSON-569100133-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 902.763476] env[65726]: WARNING openstack [None req-9c76632b-07ea-476e-92d3-ffb457aef721 tempest-ServerAddressesTestJSON-569100133 tempest-ServerAddressesTestJSON-569100133-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 902.772507] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9c76632b-07ea-476e-92d3-ffb457aef721 tempest-ServerAddressesTestJSON-569100133 tempest-ServerAddressesTestJSON-569100133-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 902.772840] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9c76632b-07ea-476e-92d3-ffb457aef721 tempest-ServerAddressesTestJSON-569100133 tempest-ServerAddressesTestJSON-569100133-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 902.773327] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9c76632b-07ea-476e-92d3-ffb457aef721 tempest-ServerAddressesTestJSON-569100133 tempest-ServerAddressesTestJSON-569100133-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 902.773761] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-28be5cb8-2a2e-4d90-a694-1f597df7a0d8 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.780666] env[65726]: DEBUG oslo_vmware.api [None req-9c76632b-07ea-476e-92d3-ffb457aef721 tempest-ServerAddressesTestJSON-569100133 tempest-ServerAddressesTestJSON-569100133-project-member] Waiting for the task: (returnval){ [ 902.780666] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52885344-4468-8161-f32c-3d72caa53624" [ 902.780666] env[65726]: _type = "Task" [ 902.780666] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.791613] env[65726]: DEBUG oslo_vmware.api [None req-9c76632b-07ea-476e-92d3-ffb457aef721 tempest-ServerAddressesTestJSON-569100133 tempest-ServerAddressesTestJSON-569100133-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52885344-4468-8161-f32c-3d72caa53624, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.826834] env[65726]: WARNING openstack [req-344d1f3c-8710-4831-9fe1-b8d63b07998e req-15bc8f76-833a-4878-958b-5d5cc85edb88 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 902.827219] env[65726]: WARNING openstack [req-344d1f3c-8710-4831-9fe1-b8d63b07998e req-15bc8f76-833a-4878-958b-5d5cc85edb88 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 902.914600] env[65726]: DEBUG oslo_vmware.api [None req-4826faf0-9908-4f3c-a509-20fe2a708723 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52b1f46b-2de3-2ab1-6a46-3cc40a1268de, 'name': SearchDatastore_Task, 'duration_secs': 0.014418} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.920271] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d5908a9b-15d4-47e2-960c-b832377872ed {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.927731] env[65726]: DEBUG oslo_vmware.api [None req-4826faf0-9908-4f3c-a509-20fe2a708723 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Waiting for the task: (returnval){ [ 902.927731] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]524eb83d-75a4-758a-c780-c2dd2a36a10b" [ 902.927731] env[65726]: _type = "Task" [ 902.927731] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.937528] env[65726]: DEBUG oslo_vmware.api [None req-4826faf0-9908-4f3c-a509-20fe2a708723 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]524eb83d-75a4-758a-c780-c2dd2a36a10b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.943410] env[65726]: WARNING openstack [req-344d1f3c-8710-4831-9fe1-b8d63b07998e req-15bc8f76-833a-4878-958b-5d5cc85edb88 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 902.943627] env[65726]: WARNING openstack [req-344d1f3c-8710-4831-9fe1-b8d63b07998e req-15bc8f76-833a-4878-958b-5d5cc85edb88 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 902.951606] env[65726]: DEBUG nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] [instance: 811c1bbd-259c-4f5e-9a61-581bfba0ec4d] Instance has had 0 of 5 cleanup attempts {{(pid=65726) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 903.034194] env[65726]: WARNING openstack [None req-253a2ea4-406a-415c-8211-e66a1f71d581 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 903.036617] env[65726]: WARNING openstack [None req-253a2ea4-406a-415c-8211-e66a1f71d581 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 903.056515] env[65726]: WARNING openstack [req-344d1f3c-8710-4831-9fe1-b8d63b07998e req-15bc8f76-833a-4878-958b-5d5cc85edb88 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 903.056826] env[65726]: WARNING openstack [req-344d1f3c-8710-4831-9fe1-b8d63b07998e req-15bc8f76-833a-4878-958b-5d5cc85edb88 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 903.122690] env[65726]: DEBUG nova.scheduler.client.report [None req-e5e88d5e-aa68-4ff5-9280-b2cfe1d14b27 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 903.136522] env[65726]: WARNING openstack [req-951802b2-3216-407b-b531-3d7188ed02a1 req-8cfe1f61-93a9-4f86-b842-ab19b679f3ba service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 903.136522] env[65726]: WARNING openstack [req-951802b2-3216-407b-b531-3d7188ed02a1 req-8cfe1f61-93a9-4f86-b842-ab19b679f3ba service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 903.191230] env[65726]: DEBUG oslo_vmware.api [None req-a1a4c202-18b9-4cfc-8067-d677c8415fff tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Task: {'id': task-5116173, 'name': Rename_Task, 'duration_secs': 0.153596} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.191559] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-a1a4c202-18b9-4cfc-8067-d677c8415fff tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] [instance: 815ee20b-cb07-4514-bd43-1b74972508a5] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 903.192439] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4d1e069a-f8c2-4242-9522-63636a46caee {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.201193] env[65726]: DEBUG oslo_vmware.api [None req-a1a4c202-18b9-4cfc-8067-d677c8415fff tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Waiting for the task: (returnval){ [ 903.201193] env[65726]: value = "task-5116174" [ 903.201193] env[65726]: _type = "Task" [ 903.201193] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.217874] env[65726]: DEBUG oslo_vmware.api [None req-a1a4c202-18b9-4cfc-8067-d677c8415fff tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Task: {'id': task-5116174, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.287705] env[65726]: DEBUG nova.network.neutron [req-344d1f3c-8710-4831-9fe1-b8d63b07998e req-15bc8f76-833a-4878-958b-5d5cc85edb88 service nova] [instance: 29823bc7-3909-4ab7-8119-91ee59b289c5] Updated VIF entry in instance network info cache for port ef57c09b-e512-44ad-adeb-8aee81e9dc7b. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 903.288149] env[65726]: DEBUG nova.network.neutron [req-344d1f3c-8710-4831-9fe1-b8d63b07998e req-15bc8f76-833a-4878-958b-5d5cc85edb88 service nova] [instance: 29823bc7-3909-4ab7-8119-91ee59b289c5] Updating instance_info_cache with network_info: [{"id": "ef57c09b-e512-44ad-adeb-8aee81e9dc7b", "address": "fa:16:3e:d9:3e:40", "network": {"id": "55d629b8-a541-4720-9289-927bbc6e6f22", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.50", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "1539f06fa3534e90acbc3a60c4b8bd3f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "35342bcb-8b06-472e-b3c0-43fd3d6c4b30", "external-id": "nsx-vlan-transportzone-524", "segmentation_id": 524, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapef57c09b-e5", "ovs_interfaceid": "ef57c09b-e512-44ad-adeb-8aee81e9dc7b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 903.297683] env[65726]: DEBUG oslo_vmware.api [None req-9c76632b-07ea-476e-92d3-ffb457aef721 tempest-ServerAddressesTestJSON-569100133 tempest-ServerAddressesTestJSON-569100133-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52885344-4468-8161-f32c-3d72caa53624, 'name': SearchDatastore_Task, 'duration_secs': 0.011424} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.298098] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9c76632b-07ea-476e-92d3-ffb457aef721 tempest-ServerAddressesTestJSON-569100133 tempest-ServerAddressesTestJSON-569100133-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 903.299312] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-9c76632b-07ea-476e-92d3-ffb457aef721 tempest-ServerAddressesTestJSON-569100133 tempest-ServerAddressesTestJSON-569100133-project-member] [instance: 70375746-06dd-4710-9136-95e9e5759a0e] Processing image b52362a3-ee8a-4cbf-b06f-513b0cc8f95c {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 903.299312] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9c76632b-07ea-476e-92d3-ffb457aef721 tempest-ServerAddressesTestJSON-569100133 tempest-ServerAddressesTestJSON-569100133-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 903.301472] env[65726]: WARNING openstack [None req-253a2ea4-406a-415c-8211-e66a1f71d581 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 903.301887] env[65726]: WARNING openstack [None req-253a2ea4-406a-415c-8211-e66a1f71d581 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 903.405827] env[65726]: WARNING openstack [req-951802b2-3216-407b-b531-3d7188ed02a1 req-8cfe1f61-93a9-4f86-b842-ab19b679f3ba service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 903.406984] env[65726]: WARNING openstack [req-951802b2-3216-407b-b531-3d7188ed02a1 req-8cfe1f61-93a9-4f86-b842-ab19b679f3ba service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 903.440226] env[65726]: DEBUG oslo_vmware.api [None req-4826faf0-9908-4f3c-a509-20fe2a708723 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]524eb83d-75a4-758a-c780-c2dd2a36a10b, 'name': SearchDatastore_Task, 'duration_secs': 0.010324} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.440483] env[65726]: DEBUG oslo_concurrency.lockutils [None req-4826faf0-9908-4f3c-a509-20fe2a708723 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 903.441039] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-4826faf0-9908-4f3c-a509-20fe2a708723 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore2] 29823bc7-3909-4ab7-8119-91ee59b289c5/29823bc7-3909-4ab7-8119-91ee59b289c5.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 903.441544] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9c76632b-07ea-476e-92d3-ffb457aef721 tempest-ServerAddressesTestJSON-569100133 tempest-ServerAddressesTestJSON-569100133-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 903.441544] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-9c76632b-07ea-476e-92d3-ffb457aef721 tempest-ServerAddressesTestJSON-569100133 tempest-ServerAddressesTestJSON-569100133-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 903.441794] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-90903b36-7dc2-41c0-9ec6-e9f0ea3a9adf {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.444437] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b980d221-233b-4381-9b10-585eaf6cd917 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.454253] env[65726]: DEBUG oslo_vmware.api [None req-4826faf0-9908-4f3c-a509-20fe2a708723 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Waiting for the task: (returnval){ [ 903.454253] env[65726]: value = "task-5116175" [ 903.454253] env[65726]: _type = "Task" [ 903.454253] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.461257] env[65726]: WARNING openstack [None req-253a2ea4-406a-415c-8211-e66a1f71d581 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 903.462174] env[65726]: WARNING openstack [None req-253a2ea4-406a-415c-8211-e66a1f71d581 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 903.477696] env[65726]: DEBUG nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] [instance: 04ffdf80-c54c-4587-9bb4-d520dc440501] Instance has had 0 of 5 cleanup attempts {{(pid=65726) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 903.480979] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-9c76632b-07ea-476e-92d3-ffb457aef721 tempest-ServerAddressesTestJSON-569100133 tempest-ServerAddressesTestJSON-569100133-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 903.481630] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-9c76632b-07ea-476e-92d3-ffb457aef721 tempest-ServerAddressesTestJSON-569100133 tempest-ServerAddressesTestJSON-569100133-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65726) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 903.490155] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-63c095c2-22d2-4f4c-9438-38d429c1f0a6 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.505184] env[65726]: DEBUG oslo_vmware.api [None req-4826faf0-9908-4f3c-a509-20fe2a708723 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Task: {'id': task-5116175, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.507785] env[65726]: DEBUG oslo_vmware.api [None req-9c76632b-07ea-476e-92d3-ffb457aef721 tempest-ServerAddressesTestJSON-569100133 tempest-ServerAddressesTestJSON-569100133-project-member] Waiting for the task: (returnval){ [ 903.507785] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]521384b6-9ae9-b6a3-e252-9df6997b50c0" [ 903.507785] env[65726]: _type = "Task" [ 903.507785] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.529577] env[65726]: DEBUG oslo_vmware.api [None req-9c76632b-07ea-476e-92d3-ffb457aef721 tempest-ServerAddressesTestJSON-569100133 tempest-ServerAddressesTestJSON-569100133-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]521384b6-9ae9-b6a3-e252-9df6997b50c0, 'name': SearchDatastore_Task, 'duration_secs': 0.011758} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.531662] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bbf753d3-a45b-43b4-956b-fc7ff559c211 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.541804] env[65726]: DEBUG oslo_vmware.api [None req-9c76632b-07ea-476e-92d3-ffb457aef721 tempest-ServerAddressesTestJSON-569100133 tempest-ServerAddressesTestJSON-569100133-project-member] Waiting for the task: (returnval){ [ 903.541804] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52d0e126-c901-f96f-add0-6bc573b172e5" [ 903.541804] env[65726]: _type = "Task" [ 903.541804] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.556407] env[65726]: DEBUG oslo_vmware.api [None req-9c76632b-07ea-476e-92d3-ffb457aef721 tempest-ServerAddressesTestJSON-569100133 tempest-ServerAddressesTestJSON-569100133-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52d0e126-c901-f96f-add0-6bc573b172e5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.608837] env[65726]: WARNING openstack [req-951802b2-3216-407b-b531-3d7188ed02a1 req-8cfe1f61-93a9-4f86-b842-ab19b679f3ba service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 903.608837] env[65726]: WARNING openstack [req-951802b2-3216-407b-b531-3d7188ed02a1 req-8cfe1f61-93a9-4f86-b842-ab19b679f3ba service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 903.619727] env[65726]: DEBUG nova.network.neutron [None req-253a2ea4-406a-415c-8211-e66a1f71d581 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] [instance: 76249623-6f83-46a3-b8c5-c001111aa698] Updating instance_info_cache with network_info: [{"id": "e97a5da4-5540-4c0c-a33d-be1182e82aa7", "address": "fa:16:3e:af:b6:0f", "network": {"id": "55d629b8-a541-4720-9289-927bbc6e6f22", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.253", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "1539f06fa3534e90acbc3a60c4b8bd3f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "35342bcb-8b06-472e-b3c0-43fd3d6c4b30", "external-id": "nsx-vlan-transportzone-524", "segmentation_id": 524, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape97a5da4-55", "ovs_interfaceid": "e97a5da4-5540-4c0c-a33d-be1182e82aa7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 903.628543] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e5e88d5e-aa68-4ff5-9280-b2cfe1d14b27 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.855s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 903.629416] env[65726]: DEBUG nova.compute.manager [None req-e5e88d5e-aa68-4ff5-9280-b2cfe1d14b27 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: 5a252ef1-93c9-4bff-842b-b64df2bc5d75] Start building networks asynchronously for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 903.633116] env[65726]: DEBUG oslo_concurrency.lockutils [None req-89fee262-eea7-4ed4-83cb-eb9998cc3059 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 33.126s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 903.634819] env[65726]: INFO nova.compute.claims [None req-89fee262-eea7-4ed4-83cb-eb9998cc3059 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 6f91b053-772a-4497-b29d-349b960c55eb] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 903.715065] env[65726]: DEBUG oslo_vmware.api [None req-a1a4c202-18b9-4cfc-8067-d677c8415fff tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Task: {'id': task-5116174, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.737066] env[65726]: DEBUG nova.network.neutron [req-951802b2-3216-407b-b531-3d7188ed02a1 req-8cfe1f61-93a9-4f86-b842-ab19b679f3ba service nova] [instance: 4ea2b70b-0cfe-4fbf-817a-baa028b8372a] Updated VIF entry in instance network info cache for port a97348ff-93e0-442b-8a41-b56fa04e8cb6. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 903.737622] env[65726]: DEBUG nova.network.neutron [req-951802b2-3216-407b-b531-3d7188ed02a1 req-8cfe1f61-93a9-4f86-b842-ab19b679f3ba service nova] [instance: 4ea2b70b-0cfe-4fbf-817a-baa028b8372a] Updating instance_info_cache with network_info: [{"id": "a97348ff-93e0-442b-8a41-b56fa04e8cb6", "address": "fa:16:3e:f5:96:db", "network": {"id": "dd4bbaef-0a1e-4e00-b65a-f19e7a837cc1", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-255385230-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "55f93cdf51e84cf683879668c2acecf4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "04ccbc7a-cf8d-4ea2-8411-291a1e27df7b", "external-id": "nsx-vlan-transportzone-998", "segmentation_id": 998, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa97348ff-93", "ovs_interfaceid": "a97348ff-93e0-442b-8a41-b56fa04e8cb6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 903.791553] env[65726]: DEBUG oslo_concurrency.lockutils [req-344d1f3c-8710-4831-9fe1-b8d63b07998e req-15bc8f76-833a-4878-958b-5d5cc85edb88 service nova] Releasing lock "refresh_cache-29823bc7-3909-4ab7-8119-91ee59b289c5" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 903.792578] env[65726]: DEBUG nova.compute.manager [req-344d1f3c-8710-4831-9fe1-b8d63b07998e req-15bc8f76-833a-4878-958b-5d5cc85edb88 service nova] [instance: 70375746-06dd-4710-9136-95e9e5759a0e] Received event network-vif-plugged-7818914d-352b-408a-9d0f-4c2bb1a25f2e {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 903.792578] env[65726]: DEBUG oslo_concurrency.lockutils [req-344d1f3c-8710-4831-9fe1-b8d63b07998e req-15bc8f76-833a-4878-958b-5d5cc85edb88 service nova] Acquiring lock "70375746-06dd-4710-9136-95e9e5759a0e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 903.792578] env[65726]: DEBUG oslo_concurrency.lockutils [req-344d1f3c-8710-4831-9fe1-b8d63b07998e req-15bc8f76-833a-4878-958b-5d5cc85edb88 service nova] Lock "70375746-06dd-4710-9136-95e9e5759a0e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 903.792858] env[65726]: DEBUG oslo_concurrency.lockutils [req-344d1f3c-8710-4831-9fe1-b8d63b07998e req-15bc8f76-833a-4878-958b-5d5cc85edb88 service nova] Lock "70375746-06dd-4710-9136-95e9e5759a0e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 903.792894] env[65726]: DEBUG nova.compute.manager [req-344d1f3c-8710-4831-9fe1-b8d63b07998e req-15bc8f76-833a-4878-958b-5d5cc85edb88 service nova] [instance: 70375746-06dd-4710-9136-95e9e5759a0e] No waiting events found dispatching network-vif-plugged-7818914d-352b-408a-9d0f-4c2bb1a25f2e {{(pid=65726) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 903.793161] env[65726]: WARNING nova.compute.manager [req-344d1f3c-8710-4831-9fe1-b8d63b07998e req-15bc8f76-833a-4878-958b-5d5cc85edb88 service nova] [instance: 70375746-06dd-4710-9136-95e9e5759a0e] Received unexpected event network-vif-plugged-7818914d-352b-408a-9d0f-4c2bb1a25f2e for instance with vm_state building and task_state spawning. [ 903.793370] env[65726]: DEBUG nova.compute.manager [req-344d1f3c-8710-4831-9fe1-b8d63b07998e req-15bc8f76-833a-4878-958b-5d5cc85edb88 service nova] [instance: 70375746-06dd-4710-9136-95e9e5759a0e] Received event network-changed-7818914d-352b-408a-9d0f-4c2bb1a25f2e {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 903.793528] env[65726]: DEBUG nova.compute.manager [req-344d1f3c-8710-4831-9fe1-b8d63b07998e req-15bc8f76-833a-4878-958b-5d5cc85edb88 service nova] [instance: 70375746-06dd-4710-9136-95e9e5759a0e] Refreshing instance network info cache due to event network-changed-7818914d-352b-408a-9d0f-4c2bb1a25f2e. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 903.793743] env[65726]: DEBUG oslo_concurrency.lockutils [req-344d1f3c-8710-4831-9fe1-b8d63b07998e req-15bc8f76-833a-4878-958b-5d5cc85edb88 service nova] Acquiring lock "refresh_cache-70375746-06dd-4710-9136-95e9e5759a0e" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 903.793857] env[65726]: DEBUG oslo_concurrency.lockutils [req-344d1f3c-8710-4831-9fe1-b8d63b07998e req-15bc8f76-833a-4878-958b-5d5cc85edb88 service nova] Acquired lock "refresh_cache-70375746-06dd-4710-9136-95e9e5759a0e" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 903.794030] env[65726]: DEBUG nova.network.neutron [req-344d1f3c-8710-4831-9fe1-b8d63b07998e req-15bc8f76-833a-4878-958b-5d5cc85edb88 service nova] [instance: 70375746-06dd-4710-9136-95e9e5759a0e] Refreshing network info cache for port 7818914d-352b-408a-9d0f-4c2bb1a25f2e {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 903.984561] env[65726]: DEBUG nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] [instance: de5cad99-1dbc-4435-98b4-987e24ba8c3c] Instance has had 0 of 5 cleanup attempts {{(pid=65726) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 903.996125] env[65726]: DEBUG oslo_vmware.api [None req-4826faf0-9908-4f3c-a509-20fe2a708723 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Task: {'id': task-5116175, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.056418] env[65726]: DEBUG oslo_vmware.api [None req-9c76632b-07ea-476e-92d3-ffb457aef721 tempest-ServerAddressesTestJSON-569100133 tempest-ServerAddressesTestJSON-569100133-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52d0e126-c901-f96f-add0-6bc573b172e5, 'name': SearchDatastore_Task, 'duration_secs': 0.021944} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.056779] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9c76632b-07ea-476e-92d3-ffb457aef721 tempest-ServerAddressesTestJSON-569100133 tempest-ServerAddressesTestJSON-569100133-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 904.057170] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-9c76632b-07ea-476e-92d3-ffb457aef721 tempest-ServerAddressesTestJSON-569100133 tempest-ServerAddressesTestJSON-569100133-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore2] 70375746-06dd-4710-9136-95e9e5759a0e/70375746-06dd-4710-9136-95e9e5759a0e.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 904.057472] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-db84fd80-c029-4745-8cd6-44f3c4f98662 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.066223] env[65726]: DEBUG oslo_vmware.api [None req-9c76632b-07ea-476e-92d3-ffb457aef721 tempest-ServerAddressesTestJSON-569100133 tempest-ServerAddressesTestJSON-569100133-project-member] Waiting for the task: (returnval){ [ 904.066223] env[65726]: value = "task-5116176" [ 904.066223] env[65726]: _type = "Task" [ 904.066223] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.079182] env[65726]: DEBUG oslo_vmware.api [None req-9c76632b-07ea-476e-92d3-ffb457aef721 tempest-ServerAddressesTestJSON-569100133 tempest-ServerAddressesTestJSON-569100133-project-member] Task: {'id': task-5116176, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.123868] env[65726]: DEBUG oslo_concurrency.lockutils [None req-253a2ea4-406a-415c-8211-e66a1f71d581 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Releasing lock "refresh_cache-76249623-6f83-46a3-b8c5-c001111aa698" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 904.124593] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-253a2ea4-406a-415c-8211-e66a1f71d581 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] [instance: 76249623-6f83-46a3-b8c5-c001111aa698] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 904.124999] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-50d2db43-bcdb-4a42-b8de-6c0dd1155888 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.137421] env[65726]: DEBUG oslo_vmware.api [None req-253a2ea4-406a-415c-8211-e66a1f71d581 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Waiting for the task: (returnval){ [ 904.137421] env[65726]: value = "task-5116177" [ 904.137421] env[65726]: _type = "Task" [ 904.137421] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.144328] env[65726]: DEBUG nova.compute.utils [None req-e5e88d5e-aa68-4ff5-9280-b2cfe1d14b27 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Using /dev/sd instead of None {{(pid=65726) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 904.150422] env[65726]: DEBUG nova.compute.manager [None req-e5e88d5e-aa68-4ff5-9280-b2cfe1d14b27 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: 5a252ef1-93c9-4bff-842b-b64df2bc5d75] Allocating IP information in the background. {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 904.150422] env[65726]: DEBUG nova.network.neutron [None req-e5e88d5e-aa68-4ff5-9280-b2cfe1d14b27 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: 5a252ef1-93c9-4bff-842b-b64df2bc5d75] allocate_for_instance() {{(pid=65726) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 904.151754] env[65726]: WARNING neutronclient.v2_0.client [None req-e5e88d5e-aa68-4ff5-9280-b2cfe1d14b27 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 904.151754] env[65726]: WARNING neutronclient.v2_0.client [None req-e5e88d5e-aa68-4ff5-9280-b2cfe1d14b27 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 904.152428] env[65726]: WARNING openstack [None req-e5e88d5e-aa68-4ff5-9280-b2cfe1d14b27 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 904.154349] env[65726]: WARNING openstack [None req-e5e88d5e-aa68-4ff5-9280-b2cfe1d14b27 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 904.176091] env[65726]: DEBUG oslo_vmware.api [None req-253a2ea4-406a-415c-8211-e66a1f71d581 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Task: {'id': task-5116177, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.215152] env[65726]: DEBUG oslo_vmware.api [None req-a1a4c202-18b9-4cfc-8067-d677c8415fff tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Task: {'id': task-5116174, 'name': PowerOnVM_Task, 'duration_secs': 0.751845} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.215704] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-a1a4c202-18b9-4cfc-8067-d677c8415fff tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] [instance: 815ee20b-cb07-4514-bd43-1b74972508a5] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 904.215790] env[65726]: INFO nova.compute.manager [None req-a1a4c202-18b9-4cfc-8067-d677c8415fff tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] [instance: 815ee20b-cb07-4514-bd43-1b74972508a5] Took 10.25 seconds to spawn the instance on the hypervisor. [ 904.215921] env[65726]: DEBUG nova.compute.manager [None req-a1a4c202-18b9-4cfc-8067-d677c8415fff tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] [instance: 815ee20b-cb07-4514-bd43-1b74972508a5] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 904.216864] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b62be2fb-e279-44b3-8dbc-3664f754c9fc {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.242136] env[65726]: DEBUG oslo_concurrency.lockutils [req-951802b2-3216-407b-b531-3d7188ed02a1 req-8cfe1f61-93a9-4f86-b842-ab19b679f3ba service nova] Releasing lock "refresh_cache-4ea2b70b-0cfe-4fbf-817a-baa028b8372a" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 904.242689] env[65726]: DEBUG oslo_concurrency.lockutils [None req-140eec77-ac75-4bcd-8eca-c7b64f5e10aa tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Acquired lock "refresh_cache-4ea2b70b-0cfe-4fbf-817a-baa028b8372a" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 904.242876] env[65726]: DEBUG nova.network.neutron [None req-140eec77-ac75-4bcd-8eca-c7b64f5e10aa tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] [instance: 4ea2b70b-0cfe-4fbf-817a-baa028b8372a] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 904.272395] env[65726]: DEBUG nova.policy [None req-e5e88d5e-aa68-4ff5-9280-b2cfe1d14b27 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4cd1ed63db174f21be78f86554fada0a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'cbb17a09b35c4c22ade5c4082b10bda9', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65726) authorize /opt/stack/nova/nova/policy.py:192}} [ 904.290857] env[65726]: DEBUG nova.compute.manager [None req-31642568-e231-4328-a969-4d06d89d9576 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] [instance: c4cc4f85-cb35-4edc-a58b-adfee0ce1265] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 904.296026] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9ea09d8-f01c-431c-be2d-5ea456184211 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.302150] env[65726]: WARNING openstack [req-344d1f3c-8710-4831-9fe1-b8d63b07998e req-15bc8f76-833a-4878-958b-5d5cc85edb88 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 904.302150] env[65726]: WARNING openstack [req-344d1f3c-8710-4831-9fe1-b8d63b07998e req-15bc8f76-833a-4878-958b-5d5cc85edb88 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 904.492037] env[65726]: DEBUG nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] [instance: aa2716a9-2cf8-4e51-800d-1c8f98b2cf05] Instance has had 0 of 5 cleanup attempts {{(pid=65726) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 904.494122] env[65726]: DEBUG oslo_vmware.api [None req-4826faf0-9908-4f3c-a509-20fe2a708723 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Task: {'id': task-5116175, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.756569} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.494558] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-4826faf0-9908-4f3c-a509-20fe2a708723 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore2] 29823bc7-3909-4ab7-8119-91ee59b289c5/29823bc7-3909-4ab7-8119-91ee59b289c5.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 904.494644] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-4826faf0-9908-4f3c-a509-20fe2a708723 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] [instance: 29823bc7-3909-4ab7-8119-91ee59b289c5] Extending root virtual disk to 1048576 {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 904.495260] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-43cb3c65-5648-499d-b963-985daf927eae {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.506504] env[65726]: DEBUG oslo_vmware.api [None req-4826faf0-9908-4f3c-a509-20fe2a708723 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Waiting for the task: (returnval){ [ 904.506504] env[65726]: value = "task-5116178" [ 904.506504] env[65726]: _type = "Task" [ 904.506504] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.518833] env[65726]: DEBUG oslo_vmware.api [None req-4826faf0-9908-4f3c-a509-20fe2a708723 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Task: {'id': task-5116178, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.581106] env[65726]: DEBUG oslo_vmware.api [None req-9c76632b-07ea-476e-92d3-ffb457aef721 tempest-ServerAddressesTestJSON-569100133 tempest-ServerAddressesTestJSON-569100133-project-member] Task: {'id': task-5116176, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.613334] env[65726]: DEBUG nova.network.neutron [None req-e5e88d5e-aa68-4ff5-9280-b2cfe1d14b27 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: 5a252ef1-93c9-4bff-842b-b64df2bc5d75] Successfully created port: 8cd2f231-f537-4509-92da-dc52e57441ba {{(pid=65726) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 904.647919] env[65726]: DEBUG oslo_vmware.api [None req-253a2ea4-406a-415c-8211-e66a1f71d581 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Task: {'id': task-5116177, 'name': PowerOffVM_Task, 'duration_secs': 0.240857} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.648188] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-253a2ea4-406a-415c-8211-e66a1f71d581 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] [instance: 76249623-6f83-46a3-b8c5-c001111aa698] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 904.648925] env[65726]: DEBUG nova.virt.hardware [None req-253a2ea4-406a-415c-8211-e66a1f71d581 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:32:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='d5a398f3-3467-4467-9a69-ab6d1bafa43b',id=37,is_public=True,memory_mb=192,name='tempest-test_resize_flavor_-1568765588',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 904.649171] env[65726]: DEBUG nova.virt.hardware [None req-253a2ea4-406a-415c-8211-e66a1f71d581 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 904.649328] env[65726]: DEBUG nova.virt.hardware [None req-253a2ea4-406a-415c-8211-e66a1f71d581 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 904.649503] env[65726]: DEBUG nova.virt.hardware [None req-253a2ea4-406a-415c-8211-e66a1f71d581 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 904.649639] env[65726]: DEBUG nova.virt.hardware [None req-253a2ea4-406a-415c-8211-e66a1f71d581 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 904.650263] env[65726]: DEBUG nova.virt.hardware [None req-253a2ea4-406a-415c-8211-e66a1f71d581 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 904.650263] env[65726]: DEBUG nova.virt.hardware [None req-253a2ea4-406a-415c-8211-e66a1f71d581 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 904.650263] env[65726]: DEBUG nova.virt.hardware [None req-253a2ea4-406a-415c-8211-e66a1f71d581 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 904.650405] env[65726]: DEBUG nova.virt.hardware [None req-253a2ea4-406a-415c-8211-e66a1f71d581 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 904.650463] env[65726]: DEBUG nova.virt.hardware [None req-253a2ea4-406a-415c-8211-e66a1f71d581 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 904.650596] env[65726]: DEBUG nova.virt.hardware [None req-253a2ea4-406a-415c-8211-e66a1f71d581 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 904.661833] env[65726]: DEBUG nova.compute.manager [None req-e5e88d5e-aa68-4ff5-9280-b2cfe1d14b27 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: 5a252ef1-93c9-4bff-842b-b64df2bc5d75] Start building block device mappings for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 904.664840] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-05015bb4-9638-4fd9-8186-2bb269ed5fb8 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.685815] env[65726]: DEBUG oslo_vmware.api [None req-253a2ea4-406a-415c-8211-e66a1f71d581 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Waiting for the task: (returnval){ [ 904.685815] env[65726]: value = "task-5116179" [ 904.685815] env[65726]: _type = "Task" [ 904.685815] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.695716] env[65726]: DEBUG oslo_vmware.api [None req-253a2ea4-406a-415c-8211-e66a1f71d581 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Task: {'id': task-5116179, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.747067] env[65726]: INFO nova.compute.manager [None req-a1a4c202-18b9-4cfc-8067-d677c8415fff tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] [instance: 815ee20b-cb07-4514-bd43-1b74972508a5] Took 62.67 seconds to build instance. [ 904.750686] env[65726]: WARNING openstack [None req-140eec77-ac75-4bcd-8eca-c7b64f5e10aa tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 904.750686] env[65726]: WARNING openstack [None req-140eec77-ac75-4bcd-8eca-c7b64f5e10aa tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 904.827026] env[65726]: INFO nova.compute.manager [None req-31642568-e231-4328-a969-4d06d89d9576 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] [instance: c4cc4f85-cb35-4edc-a58b-adfee0ce1265] instance snapshotting [ 904.830261] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc0b36be-21e7-41bf-8f8e-0d5275e97034 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.856237] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-944d0eb8-c578-4752-a397-ae4e9e99d609 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.991561] env[65726]: WARNING openstack [req-344d1f3c-8710-4831-9fe1-b8d63b07998e req-15bc8f76-833a-4878-958b-5d5cc85edb88 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 904.991965] env[65726]: WARNING openstack [req-344d1f3c-8710-4831-9fe1-b8d63b07998e req-15bc8f76-833a-4878-958b-5d5cc85edb88 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 904.999690] env[65726]: DEBUG nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] [instance: cd539d8b-921b-4947-954c-8573380dbdc8] Instance has had 0 of 5 cleanup attempts {{(pid=65726) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 905.022403] env[65726]: DEBUG oslo_vmware.api [None req-4826faf0-9908-4f3c-a509-20fe2a708723 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Task: {'id': task-5116178, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.099094} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.023769] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-4826faf0-9908-4f3c-a509-20fe2a708723 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] [instance: 29823bc7-3909-4ab7-8119-91ee59b289c5] Extended root virtual disk {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 905.025234] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6106488f-264c-43c5-be85-5b437f7db6b3 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.053437] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-4826faf0-9908-4f3c-a509-20fe2a708723 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] [instance: 29823bc7-3909-4ab7-8119-91ee59b289c5] Reconfiguring VM instance instance-00000042 to attach disk [datastore2] 29823bc7-3909-4ab7-8119-91ee59b289c5/29823bc7-3909-4ab7-8119-91ee59b289c5.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 905.062269] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e1cb8f99-02f6-4d79-842b-c249bbbcc582 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.092286] env[65726]: DEBUG oslo_vmware.api [None req-4826faf0-9908-4f3c-a509-20fe2a708723 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Waiting for the task: (returnval){ [ 905.092286] env[65726]: value = "task-5116180" [ 905.092286] env[65726]: _type = "Task" [ 905.092286] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.102991] env[65726]: DEBUG oslo_vmware.api [None req-9c76632b-07ea-476e-92d3-ffb457aef721 tempest-ServerAddressesTestJSON-569100133 tempest-ServerAddressesTestJSON-569100133-project-member] Task: {'id': task-5116176, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.617613} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.107295] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-9c76632b-07ea-476e-92d3-ffb457aef721 tempest-ServerAddressesTestJSON-569100133 tempest-ServerAddressesTestJSON-569100133-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore2] 70375746-06dd-4710-9136-95e9e5759a0e/70375746-06dd-4710-9136-95e9e5759a0e.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 905.107295] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-9c76632b-07ea-476e-92d3-ffb457aef721 tempest-ServerAddressesTestJSON-569100133 tempest-ServerAddressesTestJSON-569100133-project-member] [instance: 70375746-06dd-4710-9136-95e9e5759a0e] Extending root virtual disk to 1048576 {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 905.107295] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5936f5a6-aec1-40ac-8eca-786b6f7c9e80 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.113253] env[65726]: DEBUG oslo_vmware.api [None req-4826faf0-9908-4f3c-a509-20fe2a708723 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Task: {'id': task-5116180, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.122425] env[65726]: DEBUG oslo_vmware.api [None req-9c76632b-07ea-476e-92d3-ffb457aef721 tempest-ServerAddressesTestJSON-569100133 tempest-ServerAddressesTestJSON-569100133-project-member] Waiting for the task: (returnval){ [ 905.122425] env[65726]: value = "task-5116181" [ 905.122425] env[65726]: _type = "Task" [ 905.122425] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.134966] env[65726]: DEBUG oslo_vmware.api [None req-9c76632b-07ea-476e-92d3-ffb457aef721 tempest-ServerAddressesTestJSON-569100133 tempest-ServerAddressesTestJSON-569100133-project-member] Task: {'id': task-5116181, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.194069] env[65726]: WARNING openstack [req-344d1f3c-8710-4831-9fe1-b8d63b07998e req-15bc8f76-833a-4878-958b-5d5cc85edb88 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 905.194678] env[65726]: WARNING openstack [req-344d1f3c-8710-4831-9fe1-b8d63b07998e req-15bc8f76-833a-4878-958b-5d5cc85edb88 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 905.216204] env[65726]: DEBUG oslo_vmware.api [None req-253a2ea4-406a-415c-8211-e66a1f71d581 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Task: {'id': task-5116179, 'name': ReconfigVM_Task, 'duration_secs': 0.171198} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.218140] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c8092e7-aca0-4418-a08a-1f303345a65e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.255429] env[65726]: DEBUG nova.virt.hardware [None req-253a2ea4-406a-415c-8211-e66a1f71d581 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:32:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='d5a398f3-3467-4467-9a69-ab6d1bafa43b',id=37,is_public=True,memory_mb=192,name='tempest-test_resize_flavor_-1568765588',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 905.255871] env[65726]: DEBUG nova.virt.hardware [None req-253a2ea4-406a-415c-8211-e66a1f71d581 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 905.256150] env[65726]: DEBUG nova.virt.hardware [None req-253a2ea4-406a-415c-8211-e66a1f71d581 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 905.256457] env[65726]: DEBUG nova.virt.hardware [None req-253a2ea4-406a-415c-8211-e66a1f71d581 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 905.256700] env[65726]: DEBUG nova.virt.hardware [None req-253a2ea4-406a-415c-8211-e66a1f71d581 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 905.256967] env[65726]: DEBUG nova.virt.hardware [None req-253a2ea4-406a-415c-8211-e66a1f71d581 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 905.257342] env[65726]: DEBUG nova.virt.hardware [None req-253a2ea4-406a-415c-8211-e66a1f71d581 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 905.257609] env[65726]: DEBUG nova.virt.hardware [None req-253a2ea4-406a-415c-8211-e66a1f71d581 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 905.257886] env[65726]: DEBUG nova.virt.hardware [None req-253a2ea4-406a-415c-8211-e66a1f71d581 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 905.258188] env[65726]: DEBUG nova.virt.hardware [None req-253a2ea4-406a-415c-8211-e66a1f71d581 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 905.258488] env[65726]: DEBUG nova.virt.hardware [None req-253a2ea4-406a-415c-8211-e66a1f71d581 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 905.264798] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a1a4c202-18b9-4cfc-8067-d677c8415fff tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Lock "815ee20b-cb07-4514-bd43-1b74972508a5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 71.011s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 905.265228] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-70fb34a0-372b-4ab1-9846-33ef454f4dd5 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.271135] env[65726]: WARNING openstack [None req-140eec77-ac75-4bcd-8eca-c7b64f5e10aa tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 905.271135] env[65726]: WARNING openstack [None req-140eec77-ac75-4bcd-8eca-c7b64f5e10aa tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 905.282402] env[65726]: DEBUG oslo_vmware.api [None req-253a2ea4-406a-415c-8211-e66a1f71d581 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Waiting for the task: (returnval){ [ 905.282402] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]524b27f6-aff5-f9b4-a51b-191d09e71f0f" [ 905.282402] env[65726]: _type = "Task" [ 905.282402] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.295078] env[65726]: DEBUG oslo_vmware.api [None req-253a2ea4-406a-415c-8211-e66a1f71d581 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]524b27f6-aff5-f9b4-a51b-191d09e71f0f, 'name': SearchDatastore_Task, 'duration_secs': 0.008463} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.302337] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-253a2ea4-406a-415c-8211-e66a1f71d581 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] [instance: 76249623-6f83-46a3-b8c5-c001111aa698] Reconfiguring VM instance instance-0000002e to detach disk 2000 {{(pid=65726) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 905.303255] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-324db885-5b29-4aa3-b99a-a115ee964fab {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.327227] env[65726]: DEBUG oslo_vmware.api [None req-253a2ea4-406a-415c-8211-e66a1f71d581 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Waiting for the task: (returnval){ [ 905.327227] env[65726]: value = "task-5116182" [ 905.327227] env[65726]: _type = "Task" [ 905.327227] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.344396] env[65726]: DEBUG oslo_vmware.api [None req-253a2ea4-406a-415c-8211-e66a1f71d581 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Task: {'id': task-5116182, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.369133] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-31642568-e231-4328-a969-4d06d89d9576 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] [instance: c4cc4f85-cb35-4edc-a58b-adfee0ce1265] Creating Snapshot of the VM instance {{(pid=65726) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 905.369819] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-aadcca39-c3f7-4f58-ad5f-db6e2c0f07a9 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.381216] env[65726]: DEBUG oslo_vmware.api [None req-31642568-e231-4328-a969-4d06d89d9576 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Waiting for the task: (returnval){ [ 905.381216] env[65726]: value = "task-5116183" [ 905.381216] env[65726]: _type = "Task" [ 905.381216] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.382984] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de49ee8a-cdf5-4279-b552-58bc537e6ed4 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.397952] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-605b9255-4742-44a8-97d2-dbffd7541e11 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.401536] env[65726]: DEBUG oslo_vmware.api [None req-31642568-e231-4328-a969-4d06d89d9576 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Task: {'id': task-5116183, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.435762] env[65726]: DEBUG nova.network.neutron [req-344d1f3c-8710-4831-9fe1-b8d63b07998e req-15bc8f76-833a-4878-958b-5d5cc85edb88 service nova] [instance: 70375746-06dd-4710-9136-95e9e5759a0e] Updated VIF entry in instance network info cache for port 7818914d-352b-408a-9d0f-4c2bb1a25f2e. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 905.436301] env[65726]: DEBUG nova.network.neutron [req-344d1f3c-8710-4831-9fe1-b8d63b07998e req-15bc8f76-833a-4878-958b-5d5cc85edb88 service nova] [instance: 70375746-06dd-4710-9136-95e9e5759a0e] Updating instance_info_cache with network_info: [{"id": "7818914d-352b-408a-9d0f-4c2bb1a25f2e", "address": "fa:16:3e:8a:3a:b3", "network": {"id": "56a899e2-6ddb-4f9f-bcf6-0f1905f60e6f", "bridge": "br-int", "label": "tempest-ServerAddressesTestJSON-445774747-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "04fbc0c66f094d73b87bccb72aa6f37a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8c979f78-8597-41f8-b1de-995014032689", "external-id": "nsx-vlan-transportzone-477", "segmentation_id": 477, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7818914d-35", "ovs_interfaceid": "7818914d-352b-408a-9d0f-4c2bb1a25f2e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 905.438053] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d45ec54a-2e79-4b31-bd35-816319d38a60 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.448656] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-400201a0-ed3b-48d1-85aa-cd7fe4e34326 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.464243] env[65726]: DEBUG nova.compute.provider_tree [None req-89fee262-eea7-4ed4-83cb-eb9998cc3059 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 905.473362] env[65726]: WARNING openstack [None req-140eec77-ac75-4bcd-8eca-c7b64f5e10aa tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 905.474832] env[65726]: WARNING openstack [None req-140eec77-ac75-4bcd-8eca-c7b64f5e10aa tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 905.506849] env[65726]: DEBUG nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] [instance: 4d4dc88b-0d6c-430c-ab4f-c81a56e4657b] Instance has had 0 of 5 cleanup attempts {{(pid=65726) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 905.583928] env[65726]: DEBUG nova.network.neutron [None req-140eec77-ac75-4bcd-8eca-c7b64f5e10aa tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] [instance: 4ea2b70b-0cfe-4fbf-817a-baa028b8372a] Updating instance_info_cache with network_info: [{"id": "a97348ff-93e0-442b-8a41-b56fa04e8cb6", "address": "fa:16:3e:f5:96:db", "network": {"id": "dd4bbaef-0a1e-4e00-b65a-f19e7a837cc1", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-255385230-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "55f93cdf51e84cf683879668c2acecf4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "04ccbc7a-cf8d-4ea2-8411-291a1e27df7b", "external-id": "nsx-vlan-transportzone-998", "segmentation_id": 998, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa97348ff-93", "ovs_interfaceid": "a97348ff-93e0-442b-8a41-b56fa04e8cb6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 905.604411] env[65726]: DEBUG oslo_vmware.api [None req-4826faf0-9908-4f3c-a509-20fe2a708723 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Task: {'id': task-5116180, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.630366] env[65726]: DEBUG oslo_vmware.api [None req-9c76632b-07ea-476e-92d3-ffb457aef721 tempest-ServerAddressesTestJSON-569100133 tempest-ServerAddressesTestJSON-569100133-project-member] Task: {'id': task-5116181, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.136537} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.630675] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-9c76632b-07ea-476e-92d3-ffb457aef721 tempest-ServerAddressesTestJSON-569100133 tempest-ServerAddressesTestJSON-569100133-project-member] [instance: 70375746-06dd-4710-9136-95e9e5759a0e] Extended root virtual disk {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 905.631573] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26c82591-ddfc-46bd-bb93-9fcd31f94227 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.656930] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-9c76632b-07ea-476e-92d3-ffb457aef721 tempest-ServerAddressesTestJSON-569100133 tempest-ServerAddressesTestJSON-569100133-project-member] [instance: 70375746-06dd-4710-9136-95e9e5759a0e] Reconfiguring VM instance instance-00000043 to attach disk [datastore2] 70375746-06dd-4710-9136-95e9e5759a0e/70375746-06dd-4710-9136-95e9e5759a0e.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 905.657170] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2e9a1ae3-30b1-40ce-9d1e-a77e2a27e078 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.681747] env[65726]: DEBUG nova.compute.manager [None req-e5e88d5e-aa68-4ff5-9280-b2cfe1d14b27 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: 5a252ef1-93c9-4bff-842b-b64df2bc5d75] Start spawning the instance on the hypervisor. {{(pid=65726) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 905.684353] env[65726]: DEBUG oslo_vmware.api [None req-9c76632b-07ea-476e-92d3-ffb457aef721 tempest-ServerAddressesTestJSON-569100133 tempest-ServerAddressesTestJSON-569100133-project-member] Waiting for the task: (returnval){ [ 905.684353] env[65726]: value = "task-5116184" [ 905.684353] env[65726]: _type = "Task" [ 905.684353] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.695148] env[65726]: DEBUG oslo_vmware.api [None req-9c76632b-07ea-476e-92d3-ffb457aef721 tempest-ServerAddressesTestJSON-569100133 tempest-ServerAddressesTestJSON-569100133-project-member] Task: {'id': task-5116184, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.712710] env[65726]: DEBUG nova.virt.hardware [None req-e5e88d5e-aa68-4ff5-9280-b2cfe1d14b27 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T19:28:34Z,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1539f06fa3534e90acbc3a60c4b8bd3f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T19:28:35Z,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 905.713032] env[65726]: DEBUG nova.virt.hardware [None req-e5e88d5e-aa68-4ff5-9280-b2cfe1d14b27 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 905.713239] env[65726]: DEBUG nova.virt.hardware [None req-e5e88d5e-aa68-4ff5-9280-b2cfe1d14b27 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 905.713441] env[65726]: DEBUG nova.virt.hardware [None req-e5e88d5e-aa68-4ff5-9280-b2cfe1d14b27 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 905.713601] env[65726]: DEBUG nova.virt.hardware [None req-e5e88d5e-aa68-4ff5-9280-b2cfe1d14b27 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 905.713752] env[65726]: DEBUG nova.virt.hardware [None req-e5e88d5e-aa68-4ff5-9280-b2cfe1d14b27 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 905.713970] env[65726]: DEBUG nova.virt.hardware [None req-e5e88d5e-aa68-4ff5-9280-b2cfe1d14b27 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 905.714224] env[65726]: DEBUG nova.virt.hardware [None req-e5e88d5e-aa68-4ff5-9280-b2cfe1d14b27 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 905.714412] env[65726]: DEBUG nova.virt.hardware [None req-e5e88d5e-aa68-4ff5-9280-b2cfe1d14b27 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 905.714632] env[65726]: DEBUG nova.virt.hardware [None req-e5e88d5e-aa68-4ff5-9280-b2cfe1d14b27 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 905.714871] env[65726]: DEBUG nova.virt.hardware [None req-e5e88d5e-aa68-4ff5-9280-b2cfe1d14b27 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 905.716032] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f0284b3-d1d2-4294-83ce-fe78cc8c7cff {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.726059] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9270ac6d-9b8c-45ce-ab7d-fc398aa5a46f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.844841] env[65726]: DEBUG oslo_vmware.api [None req-253a2ea4-406a-415c-8211-e66a1f71d581 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Task: {'id': task-5116182, 'name': ReconfigVM_Task, 'duration_secs': 0.511162} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.845165] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-253a2ea4-406a-415c-8211-e66a1f71d581 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] [instance: 76249623-6f83-46a3-b8c5-c001111aa698] Reconfigured VM instance instance-0000002e to detach disk 2000 {{(pid=65726) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 905.845965] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c87e56ed-1e68-48d2-9773-de67ba068e98 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.876202] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-253a2ea4-406a-415c-8211-e66a1f71d581 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] [instance: 76249623-6f83-46a3-b8c5-c001111aa698] Reconfiguring VM instance instance-0000002e to attach disk [datastore1] 76249623-6f83-46a3-b8c5-c001111aa698/76249623-6f83-46a3-b8c5-c001111aa698.vmdk or device None with type thin {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 905.876993] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d446c884-e018-4f5f-bdb3-80864f9d68e8 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.901379] env[65726]: DEBUG oslo_vmware.api [None req-31642568-e231-4328-a969-4d06d89d9576 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Task: {'id': task-5116183, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.902157] env[65726]: DEBUG oslo_vmware.api [None req-253a2ea4-406a-415c-8211-e66a1f71d581 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Waiting for the task: (returnval){ [ 905.902157] env[65726]: value = "task-5116185" [ 905.902157] env[65726]: _type = "Task" [ 905.902157] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.913465] env[65726]: DEBUG oslo_vmware.api [None req-253a2ea4-406a-415c-8211-e66a1f71d581 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Task: {'id': task-5116185, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.942388] env[65726]: DEBUG oslo_concurrency.lockutils [req-344d1f3c-8710-4831-9fe1-b8d63b07998e req-15bc8f76-833a-4878-958b-5d5cc85edb88 service nova] Releasing lock "refresh_cache-70375746-06dd-4710-9136-95e9e5759a0e" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 905.967098] env[65726]: DEBUG nova.scheduler.client.report [None req-89fee262-eea7-4ed4-83cb-eb9998cc3059 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 906.008372] env[65726]: DEBUG nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] [instance: 21730277-cef6-4bfe-9f67-de71f6f615ff] Instance has had 0 of 5 cleanup attempts {{(pid=65726) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 906.090175] env[65726]: DEBUG oslo_concurrency.lockutils [None req-140eec77-ac75-4bcd-8eca-c7b64f5e10aa tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Releasing lock "refresh_cache-4ea2b70b-0cfe-4fbf-817a-baa028b8372a" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 906.104146] env[65726]: DEBUG oslo_vmware.api [None req-4826faf0-9908-4f3c-a509-20fe2a708723 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Task: {'id': task-5116180, 'name': ReconfigVM_Task, 'duration_secs': 0.814833} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.104146] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-4826faf0-9908-4f3c-a509-20fe2a708723 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] [instance: 29823bc7-3909-4ab7-8119-91ee59b289c5] Reconfigured VM instance instance-00000042 to attach disk [datastore2] 29823bc7-3909-4ab7-8119-91ee59b289c5/29823bc7-3909-4ab7-8119-91ee59b289c5.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 906.104572] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d7c9b777-48ed-467c-af4f-fad58a41c498 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.112314] env[65726]: DEBUG oslo_vmware.api [None req-4826faf0-9908-4f3c-a509-20fe2a708723 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Waiting for the task: (returnval){ [ 906.112314] env[65726]: value = "task-5116186" [ 906.112314] env[65726]: _type = "Task" [ 906.112314] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.122579] env[65726]: DEBUG oslo_vmware.api [None req-4826faf0-9908-4f3c-a509-20fe2a708723 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Task: {'id': task-5116186, 'name': Rename_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.207810] env[65726]: DEBUG oslo_vmware.api [None req-9c76632b-07ea-476e-92d3-ffb457aef721 tempest-ServerAddressesTestJSON-569100133 tempest-ServerAddressesTestJSON-569100133-project-member] Task: {'id': task-5116184, 'name': ReconfigVM_Task, 'duration_secs': 0.307181} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.208681] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-9c76632b-07ea-476e-92d3-ffb457aef721 tempest-ServerAddressesTestJSON-569100133 tempest-ServerAddressesTestJSON-569100133-project-member] [instance: 70375746-06dd-4710-9136-95e9e5759a0e] Reconfigured VM instance instance-00000043 to attach disk [datastore2] 70375746-06dd-4710-9136-95e9e5759a0e/70375746-06dd-4710-9136-95e9e5759a0e.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 906.209518] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3771060e-ff9f-4d5c-8205-357fc82ab3f9 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.220910] env[65726]: DEBUG oslo_vmware.api [None req-9c76632b-07ea-476e-92d3-ffb457aef721 tempest-ServerAddressesTestJSON-569100133 tempest-ServerAddressesTestJSON-569100133-project-member] Waiting for the task: (returnval){ [ 906.220910] env[65726]: value = "task-5116187" [ 906.220910] env[65726]: _type = "Task" [ 906.220910] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.230751] env[65726]: DEBUG oslo_vmware.api [None req-9c76632b-07ea-476e-92d3-ffb457aef721 tempest-ServerAddressesTestJSON-569100133 tempest-ServerAddressesTestJSON-569100133-project-member] Task: {'id': task-5116187, 'name': Rename_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.339143] env[65726]: DEBUG nova.network.neutron [None req-e5e88d5e-aa68-4ff5-9280-b2cfe1d14b27 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: 5a252ef1-93c9-4bff-842b-b64df2bc5d75] Successfully updated port: 8cd2f231-f537-4509-92da-dc52e57441ba {{(pid=65726) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 906.403030] env[65726]: DEBUG oslo_vmware.api [None req-31642568-e231-4328-a969-4d06d89d9576 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Task: {'id': task-5116183, 'name': CreateSnapshot_Task, 'duration_secs': 0.906958} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.407050] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-31642568-e231-4328-a969-4d06d89d9576 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] [instance: c4cc4f85-cb35-4edc-a58b-adfee0ce1265] Created Snapshot of the VM instance {{(pid=65726) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 906.410546] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afbbc1be-f8fc-4a5e-9453-330bbe00a3fb {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.427914] env[65726]: DEBUG oslo_vmware.api [None req-253a2ea4-406a-415c-8211-e66a1f71d581 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Task: {'id': task-5116185, 'name': ReconfigVM_Task, 'duration_secs': 0.43893} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.428533] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-253a2ea4-406a-415c-8211-e66a1f71d581 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] [instance: 76249623-6f83-46a3-b8c5-c001111aa698] Reconfigured VM instance instance-0000002e to attach disk [datastore1] 76249623-6f83-46a3-b8c5-c001111aa698/76249623-6f83-46a3-b8c5-c001111aa698.vmdk or device None with type thin {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 906.429700] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0808dba1-4354-4ff5-b34b-b2866e02f6e0 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.452136] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b98c7109-429c-4e34-82a9-98757ddfb957 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.475194] env[65726]: DEBUG oslo_concurrency.lockutils [None req-89fee262-eea7-4ed4-83cb-eb9998cc3059 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.842s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 906.475629] env[65726]: DEBUG nova.compute.manager [None req-89fee262-eea7-4ed4-83cb-eb9998cc3059 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 6f91b053-772a-4497-b29d-349b960c55eb] Start building networks asynchronously for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 906.483136] env[65726]: DEBUG oslo_concurrency.lockutils [None req-7734f530-a2df-4d41-b25a-6f45fb216a4e tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 35.720s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 906.483430] env[65726]: DEBUG oslo_concurrency.lockutils [None req-7734f530-a2df-4d41-b25a-6f45fb216a4e tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.004s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 906.485590] env[65726]: DEBUG oslo_concurrency.lockutils [None req-17b5a1c7-4bc8-41d1-988f-c933bb6421ad tempest-ServersTestManualDisk-1753452859 tempest-ServersTestManualDisk-1753452859-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 35.206s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 906.487314] env[65726]: INFO nova.compute.claims [None req-17b5a1c7-4bc8-41d1-988f-c933bb6421ad tempest-ServersTestManualDisk-1753452859 tempest-ServersTestManualDisk-1753452859-project-member] [instance: 3088ee5e-0d27-4058-a94b-2e04c3b52add] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 906.491252] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af9f31bb-27d7-4f9a-9775-8abe7360be2a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.518630] env[65726]: DEBUG nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] [instance: 8493ba52-b950-4727-970a-19b1797b2ca1] Instance has had 0 of 5 cleanup attempts {{(pid=65726) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 906.524183] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dcc0712b-cb9c-4809-8044-3ba9fd203a46 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.535579] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-253a2ea4-406a-415c-8211-e66a1f71d581 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] [instance: 76249623-6f83-46a3-b8c5-c001111aa698] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 906.537079] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-fdeb4260-5014-4520-bbd9-179b426af166 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.540442] env[65726]: INFO nova.scheduler.client.report [None req-7734f530-a2df-4d41-b25a-6f45fb216a4e tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Deleted allocations for instance 44d73b2d-2b6a-4501-9944-432da3c9330d [ 906.549765] env[65726]: DEBUG oslo_vmware.api [None req-253a2ea4-406a-415c-8211-e66a1f71d581 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Waiting for the task: (returnval){ [ 906.549765] env[65726]: value = "task-5116188" [ 906.549765] env[65726]: _type = "Task" [ 906.549765] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.562183] env[65726]: DEBUG oslo_vmware.api [None req-253a2ea4-406a-415c-8211-e66a1f71d581 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Task: {'id': task-5116188, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.593500] env[65726]: DEBUG nova.compute.manager [None req-140eec77-ac75-4bcd-8eca-c7b64f5e10aa tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] [instance: 4ea2b70b-0cfe-4fbf-817a-baa028b8372a] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 906.595626] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70c4f3ba-0f02-4555-94b7-a787c0fc49fb {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.624425] env[65726]: DEBUG oslo_vmware.api [None req-4826faf0-9908-4f3c-a509-20fe2a708723 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Task: {'id': task-5116186, 'name': Rename_Task, 'duration_secs': 0.225551} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.624425] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-4826faf0-9908-4f3c-a509-20fe2a708723 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] [instance: 29823bc7-3909-4ab7-8119-91ee59b289c5] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 906.624425] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-35d61c26-114a-48e8-b340-2f3b58623d07 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.633042] env[65726]: DEBUG oslo_vmware.api [None req-4826faf0-9908-4f3c-a509-20fe2a708723 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Waiting for the task: (returnval){ [ 906.633042] env[65726]: value = "task-5116189" [ 906.633042] env[65726]: _type = "Task" [ 906.633042] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.642662] env[65726]: DEBUG oslo_vmware.api [None req-4826faf0-9908-4f3c-a509-20fe2a708723 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Task: {'id': task-5116189, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.733603] env[65726]: DEBUG oslo_vmware.api [None req-9c76632b-07ea-476e-92d3-ffb457aef721 tempest-ServerAddressesTestJSON-569100133 tempest-ServerAddressesTestJSON-569100133-project-member] Task: {'id': task-5116187, 'name': Rename_Task, 'duration_secs': 0.194795} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.733892] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-9c76632b-07ea-476e-92d3-ffb457aef721 tempest-ServerAddressesTestJSON-569100133 tempest-ServerAddressesTestJSON-569100133-project-member] [instance: 70375746-06dd-4710-9136-95e9e5759a0e] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 906.734190] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5e1b92e3-4466-4a70-9e7f-2215b7c01995 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.742563] env[65726]: DEBUG oslo_vmware.api [None req-9c76632b-07ea-476e-92d3-ffb457aef721 tempest-ServerAddressesTestJSON-569100133 tempest-ServerAddressesTestJSON-569100133-project-member] Waiting for the task: (returnval){ [ 906.742563] env[65726]: value = "task-5116190" [ 906.742563] env[65726]: _type = "Task" [ 906.742563] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.753171] env[65726]: DEBUG oslo_vmware.api [None req-9c76632b-07ea-476e-92d3-ffb457aef721 tempest-ServerAddressesTestJSON-569100133 tempest-ServerAddressesTestJSON-569100133-project-member] Task: {'id': task-5116190, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.847315] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e5e88d5e-aa68-4ff5-9280-b2cfe1d14b27 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Acquiring lock "refresh_cache-5a252ef1-93c9-4bff-842b-b64df2bc5d75" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 906.847315] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e5e88d5e-aa68-4ff5-9280-b2cfe1d14b27 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Acquired lock "refresh_cache-5a252ef1-93c9-4bff-842b-b64df2bc5d75" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 906.847315] env[65726]: DEBUG nova.network.neutron [None req-e5e88d5e-aa68-4ff5-9280-b2cfe1d14b27 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: 5a252ef1-93c9-4bff-842b-b64df2bc5d75] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 906.938327] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-31642568-e231-4328-a969-4d06d89d9576 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] [instance: c4cc4f85-cb35-4edc-a58b-adfee0ce1265] Creating linked-clone VM from snapshot {{(pid=65726) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 906.938744] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-d041b429-6d37-4cda-ba8f-2a50e82701a5 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.946894] env[65726]: DEBUG nova.compute.manager [req-19564907-8491-4023-bb40-deac48f23b01 req-ef4cc451-0aef-4e9e-868d-4bc305643424 service nova] [instance: 5a252ef1-93c9-4bff-842b-b64df2bc5d75] Received event network-vif-plugged-8cd2f231-f537-4509-92da-dc52e57441ba {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 906.947208] env[65726]: DEBUG oslo_concurrency.lockutils [req-19564907-8491-4023-bb40-deac48f23b01 req-ef4cc451-0aef-4e9e-868d-4bc305643424 service nova] Acquiring lock "5a252ef1-93c9-4bff-842b-b64df2bc5d75-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 906.947870] env[65726]: DEBUG oslo_concurrency.lockutils [req-19564907-8491-4023-bb40-deac48f23b01 req-ef4cc451-0aef-4e9e-868d-4bc305643424 service nova] Lock "5a252ef1-93c9-4bff-842b-b64df2bc5d75-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 906.949132] env[65726]: DEBUG oslo_concurrency.lockutils [req-19564907-8491-4023-bb40-deac48f23b01 req-ef4cc451-0aef-4e9e-868d-4bc305643424 service nova] Lock "5a252ef1-93c9-4bff-842b-b64df2bc5d75-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.002s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 906.949391] env[65726]: DEBUG nova.compute.manager [req-19564907-8491-4023-bb40-deac48f23b01 req-ef4cc451-0aef-4e9e-868d-4bc305643424 service nova] [instance: 5a252ef1-93c9-4bff-842b-b64df2bc5d75] No waiting events found dispatching network-vif-plugged-8cd2f231-f537-4509-92da-dc52e57441ba {{(pid=65726) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 906.949598] env[65726]: WARNING nova.compute.manager [req-19564907-8491-4023-bb40-deac48f23b01 req-ef4cc451-0aef-4e9e-868d-4bc305643424 service nova] [instance: 5a252ef1-93c9-4bff-842b-b64df2bc5d75] Received unexpected event network-vif-plugged-8cd2f231-f537-4509-92da-dc52e57441ba for instance with vm_state building and task_state spawning. [ 906.957168] env[65726]: DEBUG oslo_vmware.api [None req-31642568-e231-4328-a969-4d06d89d9576 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Waiting for the task: (returnval){ [ 906.957168] env[65726]: value = "task-5116191" [ 906.957168] env[65726]: _type = "Task" [ 906.957168] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.974416] env[65726]: DEBUG oslo_vmware.api [None req-31642568-e231-4328-a969-4d06d89d9576 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Task: {'id': task-5116191, 'name': CloneVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.999427] env[65726]: DEBUG nova.compute.utils [None req-89fee262-eea7-4ed4-83cb-eb9998cc3059 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Using /dev/sd instead of None {{(pid=65726) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 907.005790] env[65726]: DEBUG nova.compute.manager [None req-89fee262-eea7-4ed4-83cb-eb9998cc3059 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 6f91b053-772a-4497-b29d-349b960c55eb] Allocating IP information in the background. {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 907.005790] env[65726]: DEBUG nova.network.neutron [None req-89fee262-eea7-4ed4-83cb-eb9998cc3059 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 6f91b053-772a-4497-b29d-349b960c55eb] allocate_for_instance() {{(pid=65726) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 907.005978] env[65726]: WARNING neutronclient.v2_0.client [None req-89fee262-eea7-4ed4-83cb-eb9998cc3059 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 907.006393] env[65726]: WARNING neutronclient.v2_0.client [None req-89fee262-eea7-4ed4-83cb-eb9998cc3059 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 907.007108] env[65726]: WARNING openstack [None req-89fee262-eea7-4ed4-83cb-eb9998cc3059 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 907.007529] env[65726]: WARNING openstack [None req-89fee262-eea7-4ed4-83cb-eb9998cc3059 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 907.035730] env[65726]: DEBUG nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] [instance: ee676e3b-3326-46a3-940d-ebbb8b108991] Instance has had 0 of 5 cleanup attempts {{(pid=65726) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 907.058249] env[65726]: DEBUG oslo_concurrency.lockutils [None req-7734f530-a2df-4d41-b25a-6f45fb216a4e tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Lock "44d73b2d-2b6a-4501-9944-432da3c9330d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 40.614s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 907.074928] env[65726]: DEBUG oslo_vmware.api [None req-253a2ea4-406a-415c-8211-e66a1f71d581 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Task: {'id': task-5116188, 'name': PowerOnVM_Task, 'duration_secs': 0.456896} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.074928] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-253a2ea4-406a-415c-8211-e66a1f71d581 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] [instance: 76249623-6f83-46a3-b8c5-c001111aa698] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 907.128214] env[65726]: DEBUG nova.policy [None req-89fee262-eea7-4ed4-83cb-eb9998cc3059 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9b6087cfa55d41a4bad9f6cee89865d5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '39067912d091464192516c08eb576696', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65726) authorize /opt/stack/nova/nova/policy.py:192}} [ 907.149930] env[65726]: DEBUG oslo_vmware.api [None req-4826faf0-9908-4f3c-a509-20fe2a708723 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Task: {'id': task-5116189, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.258198] env[65726]: DEBUG oslo_vmware.api [None req-9c76632b-07ea-476e-92d3-ffb457aef721 tempest-ServerAddressesTestJSON-569100133 tempest-ServerAddressesTestJSON-569100133-project-member] Task: {'id': task-5116190, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.353610] env[65726]: WARNING openstack [None req-e5e88d5e-aa68-4ff5-9280-b2cfe1d14b27 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 907.353971] env[65726]: WARNING openstack [None req-e5e88d5e-aa68-4ff5-9280-b2cfe1d14b27 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 907.359322] env[65726]: DEBUG nova.network.neutron [None req-e5e88d5e-aa68-4ff5-9280-b2cfe1d14b27 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: 5a252ef1-93c9-4bff-842b-b64df2bc5d75] Instance cache missing network info. {{(pid=65726) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 907.382346] env[65726]: DEBUG oslo_concurrency.lockutils [None req-231bdff4-2d2e-4577-9552-cfa1b1d4e74f tempest-ServersTestFqdnHostnames-1699296007 tempest-ServersTestFqdnHostnames-1699296007-project-member] Acquiring lock "7aa4c0d9-f7ca-42ba-9fd0-4075cd76752b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 907.382650] env[65726]: DEBUG oslo_concurrency.lockutils [None req-231bdff4-2d2e-4577-9552-cfa1b1d4e74f tempest-ServersTestFqdnHostnames-1699296007 tempest-ServersTestFqdnHostnames-1699296007-project-member] Lock "7aa4c0d9-f7ca-42ba-9fd0-4075cd76752b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 907.447813] env[65726]: WARNING openstack [None req-e5e88d5e-aa68-4ff5-9280-b2cfe1d14b27 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 907.448012] env[65726]: WARNING openstack [None req-e5e88d5e-aa68-4ff5-9280-b2cfe1d14b27 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 907.472245] env[65726]: DEBUG oslo_vmware.api [None req-31642568-e231-4328-a969-4d06d89d9576 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Task: {'id': task-5116191, 'name': CloneVM_Task} progress is 94%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.478312] env[65726]: DEBUG nova.network.neutron [None req-89fee262-eea7-4ed4-83cb-eb9998cc3059 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 6f91b053-772a-4497-b29d-349b960c55eb] Successfully created port: fbccb731-d98d-40ad-a144-4b4f1e48a1ce {{(pid=65726) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 907.508408] env[65726]: DEBUG nova.compute.manager [None req-89fee262-eea7-4ed4-83cb-eb9998cc3059 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 6f91b053-772a-4497-b29d-349b960c55eb] Start building block device mappings for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 907.540763] env[65726]: DEBUG nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] [instance: 703fff2d-5aec-4ed1-b987-9e1cc3a25f67] Instance has had 0 of 5 cleanup attempts {{(pid=65726) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 907.589106] env[65726]: WARNING openstack [None req-e5e88d5e-aa68-4ff5-9280-b2cfe1d14b27 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 907.589106] env[65726]: WARNING openstack [None req-e5e88d5e-aa68-4ff5-9280-b2cfe1d14b27 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 907.620529] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b195971-f822-4377-8292-61ee0a887115 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.635978] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-140eec77-ac75-4bcd-8eca-c7b64f5e10aa tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] [instance: 4ea2b70b-0cfe-4fbf-817a-baa028b8372a] Doing hard reboot of VM {{(pid=65726) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1063}} [ 907.637838] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ResetVM_Task with opID=oslo.vmware-67510548-f5a7-4ae2-b222-1bd1e2abd9c3 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.663330] env[65726]: DEBUG oslo_vmware.api [None req-140eec77-ac75-4bcd-8eca-c7b64f5e10aa tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Waiting for the task: (returnval){ [ 907.663330] env[65726]: value = "task-5116192" [ 907.663330] env[65726]: _type = "Task" [ 907.663330] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.669540] env[65726]: DEBUG oslo_vmware.api [None req-4826faf0-9908-4f3c-a509-20fe2a708723 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Task: {'id': task-5116189, 'name': PowerOnVM_Task, 'duration_secs': 0.756259} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.673947] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-4826faf0-9908-4f3c-a509-20fe2a708723 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] [instance: 29823bc7-3909-4ab7-8119-91ee59b289c5] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 907.674217] env[65726]: INFO nova.compute.manager [None req-4826faf0-9908-4f3c-a509-20fe2a708723 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] [instance: 29823bc7-3909-4ab7-8119-91ee59b289c5] Took 11.37 seconds to spawn the instance on the hypervisor. [ 907.674452] env[65726]: DEBUG nova.compute.manager [None req-4826faf0-9908-4f3c-a509-20fe2a708723 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] [instance: 29823bc7-3909-4ab7-8119-91ee59b289c5] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 907.678636] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99ac038e-4d08-4f50-a860-d17d43aa745b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.690318] env[65726]: DEBUG oslo_vmware.api [None req-140eec77-ac75-4bcd-8eca-c7b64f5e10aa tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Task: {'id': task-5116192, 'name': ResetVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.719887] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e712963c-3c79-4770-9204-f04bbcef8a2e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.729363] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a4c0b40-94a9-42ef-bbdd-3552d73658e7 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.738073] env[65726]: DEBUG nova.network.neutron [None req-e5e88d5e-aa68-4ff5-9280-b2cfe1d14b27 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: 5a252ef1-93c9-4bff-842b-b64df2bc5d75] Updating instance_info_cache with network_info: [{"id": "8cd2f231-f537-4509-92da-dc52e57441ba", "address": "fa:16:3e:f7:41:63", "network": {"id": "4c972d9d-db3b-4779-ac39-39d9a2d141dd", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1418913861-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cbb17a09b35c4c22ade5c4082b10bda9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "50cf0a70-948d-4611-af05-94c1483064ed", "external-id": "nsx-vlan-transportzone-536", "segmentation_id": 536, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8cd2f231-f5", "ovs_interfaceid": "8cd2f231-f537-4509-92da-dc52e57441ba", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 907.777576] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef5b1081-0d11-457b-b8fe-149624444540 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.785113] env[65726]: DEBUG oslo_vmware.api [None req-9c76632b-07ea-476e-92d3-ffb457aef721 tempest-ServerAddressesTestJSON-569100133 tempest-ServerAddressesTestJSON-569100133-project-member] Task: {'id': task-5116190, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.793624] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef4b2141-89fc-4cf6-95bb-25e11d48d7df {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.808163] env[65726]: DEBUG nova.compute.provider_tree [None req-17b5a1c7-4bc8-41d1-988f-c933bb6421ad tempest-ServersTestManualDisk-1753452859 tempest-ServersTestManualDisk-1753452859-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 907.885409] env[65726]: DEBUG nova.compute.manager [None req-231bdff4-2d2e-4577-9552-cfa1b1d4e74f tempest-ServersTestFqdnHostnames-1699296007 tempest-ServersTestFqdnHostnames-1699296007-project-member] [instance: 7aa4c0d9-f7ca-42ba-9fd0-4075cd76752b] Starting instance... {{(pid=65726) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 907.971905] env[65726]: DEBUG oslo_vmware.api [None req-31642568-e231-4328-a969-4d06d89d9576 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Task: {'id': task-5116191, 'name': CloneVM_Task} progress is 94%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.048604] env[65726]: DEBUG nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] [instance: 88fb7235-cd0e-49dc-9d78-3583d5d3f528] Instance has had 0 of 5 cleanup attempts {{(pid=65726) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 908.103604] env[65726]: INFO nova.compute.manager [None req-253a2ea4-406a-415c-8211-e66a1f71d581 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] [instance: 76249623-6f83-46a3-b8c5-c001111aa698] Updating instance to original state: 'active' [ 908.155628] env[65726]: DEBUG oslo_concurrency.lockutils [None req-fa9b16af-8474-4ab3-a554-b5fbaf4d16fd tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Acquiring lock "1cfb1bba-1e21-47eb-b50f-a86575846a65" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 908.155993] env[65726]: DEBUG oslo_concurrency.lockutils [None req-fa9b16af-8474-4ab3-a554-b5fbaf4d16fd tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Lock "1cfb1bba-1e21-47eb-b50f-a86575846a65" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 908.156248] env[65726]: DEBUG oslo_concurrency.lockutils [None req-fa9b16af-8474-4ab3-a554-b5fbaf4d16fd tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Acquiring lock "1cfb1bba-1e21-47eb-b50f-a86575846a65-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 908.156461] env[65726]: DEBUG oslo_concurrency.lockutils [None req-fa9b16af-8474-4ab3-a554-b5fbaf4d16fd tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Lock "1cfb1bba-1e21-47eb-b50f-a86575846a65-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 908.156664] env[65726]: DEBUG oslo_concurrency.lockutils [None req-fa9b16af-8474-4ab3-a554-b5fbaf4d16fd tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Lock "1cfb1bba-1e21-47eb-b50f-a86575846a65-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 908.158865] env[65726]: INFO nova.compute.manager [None req-fa9b16af-8474-4ab3-a554-b5fbaf4d16fd tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] [instance: 1cfb1bba-1e21-47eb-b50f-a86575846a65] Terminating instance [ 908.184394] env[65726]: DEBUG oslo_vmware.api [None req-140eec77-ac75-4bcd-8eca-c7b64f5e10aa tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Task: {'id': task-5116192, 'name': ResetVM_Task, 'duration_secs': 0.125076} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.184749] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-140eec77-ac75-4bcd-8eca-c7b64f5e10aa tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] [instance: 4ea2b70b-0cfe-4fbf-817a-baa028b8372a] Did hard reboot of VM {{(pid=65726) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1067}} [ 908.184923] env[65726]: DEBUG nova.compute.manager [None req-140eec77-ac75-4bcd-8eca-c7b64f5e10aa tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] [instance: 4ea2b70b-0cfe-4fbf-817a-baa028b8372a] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 908.185780] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87d488f7-5797-47f5-aa1c-d0c6fa1af7cc {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.206439] env[65726]: INFO nova.compute.manager [None req-4826faf0-9908-4f3c-a509-20fe2a708723 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] [instance: 29823bc7-3909-4ab7-8119-91ee59b289c5] Took 54.80 seconds to build instance. [ 908.243519] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e5e88d5e-aa68-4ff5-9280-b2cfe1d14b27 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Releasing lock "refresh_cache-5a252ef1-93c9-4bff-842b-b64df2bc5d75" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 908.243519] env[65726]: DEBUG nova.compute.manager [None req-e5e88d5e-aa68-4ff5-9280-b2cfe1d14b27 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: 5a252ef1-93c9-4bff-842b-b64df2bc5d75] Instance network_info: |[{"id": "8cd2f231-f537-4509-92da-dc52e57441ba", "address": "fa:16:3e:f7:41:63", "network": {"id": "4c972d9d-db3b-4779-ac39-39d9a2d141dd", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1418913861-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cbb17a09b35c4c22ade5c4082b10bda9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "50cf0a70-948d-4611-af05-94c1483064ed", "external-id": "nsx-vlan-transportzone-536", "segmentation_id": 536, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8cd2f231-f5", "ovs_interfaceid": "8cd2f231-f537-4509-92da-dc52e57441ba", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 908.244274] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-e5e88d5e-aa68-4ff5-9280-b2cfe1d14b27 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: 5a252ef1-93c9-4bff-842b-b64df2bc5d75] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f7:41:63', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '50cf0a70-948d-4611-af05-94c1483064ed', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '8cd2f231-f537-4509-92da-dc52e57441ba', 'vif_model': 'vmxnet3'}] {{(pid=65726) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 908.253236] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-e5e88d5e-aa68-4ff5-9280-b2cfe1d14b27 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 908.254603] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5a252ef1-93c9-4bff-842b-b64df2bc5d75] Creating VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 908.258633] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9b3143dc-099b-47b3-ab8e-8067786b0519 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.284051] env[65726]: DEBUG oslo_vmware.api [None req-9c76632b-07ea-476e-92d3-ffb457aef721 tempest-ServerAddressesTestJSON-569100133 tempest-ServerAddressesTestJSON-569100133-project-member] Task: {'id': task-5116190, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.286224] env[65726]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 908.286224] env[65726]: value = "task-5116193" [ 908.286224] env[65726]: _type = "Task" [ 908.286224] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.297448] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116193, 'name': CreateVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.312673] env[65726]: DEBUG nova.scheduler.client.report [None req-17b5a1c7-4bc8-41d1-988f-c933bb6421ad tempest-ServersTestManualDisk-1753452859 tempest-ServersTestManualDisk-1753452859-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 908.414296] env[65726]: DEBUG oslo_concurrency.lockutils [None req-231bdff4-2d2e-4577-9552-cfa1b1d4e74f tempest-ServersTestFqdnHostnames-1699296007 tempest-ServersTestFqdnHostnames-1699296007-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 908.475104] env[65726]: DEBUG oslo_vmware.api [None req-31642568-e231-4328-a969-4d06d89d9576 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Task: {'id': task-5116191, 'name': CloneVM_Task} progress is 94%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.523037] env[65726]: DEBUG nova.compute.manager [None req-89fee262-eea7-4ed4-83cb-eb9998cc3059 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 6f91b053-772a-4497-b29d-349b960c55eb] Start spawning the instance on the hypervisor. {{(pid=65726) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 908.555276] env[65726]: DEBUG nova.virt.hardware [None req-89fee262-eea7-4ed4-83cb-eb9998cc3059 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='9bd3b439cff07ea969b0f5c7def7a486',container_format='bare',created_at=2025-12-12T19:35:00Z,direct_url=,disk_format='vmdk',id=8ecee2b1-8703-4e91-8a32-dc17932b237e,min_disk=1,min_ram=0,name='tempest-test-snap-418227437',owner='39067912d091464192516c08eb576696',properties=ImageMetaProps,protected=,size=21334016,status='active',tags=,updated_at=2025-12-12T19:35:16Z,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 908.555276] env[65726]: DEBUG nova.virt.hardware [None req-89fee262-eea7-4ed4-83cb-eb9998cc3059 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 908.555276] env[65726]: DEBUG nova.virt.hardware [None req-89fee262-eea7-4ed4-83cb-eb9998cc3059 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 908.555276] env[65726]: DEBUG nova.virt.hardware [None req-89fee262-eea7-4ed4-83cb-eb9998cc3059 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 908.555781] env[65726]: DEBUG nova.virt.hardware [None req-89fee262-eea7-4ed4-83cb-eb9998cc3059 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 908.556185] env[65726]: DEBUG nova.virt.hardware [None req-89fee262-eea7-4ed4-83cb-eb9998cc3059 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 908.556566] env[65726]: DEBUG nova.virt.hardware [None req-89fee262-eea7-4ed4-83cb-eb9998cc3059 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 908.556865] env[65726]: DEBUG nova.virt.hardware [None req-89fee262-eea7-4ed4-83cb-eb9998cc3059 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 908.557198] env[65726]: DEBUG nova.virt.hardware [None req-89fee262-eea7-4ed4-83cb-eb9998cc3059 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 908.557525] env[65726]: DEBUG nova.virt.hardware [None req-89fee262-eea7-4ed4-83cb-eb9998cc3059 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 908.557849] env[65726]: DEBUG nova.virt.hardware [None req-89fee262-eea7-4ed4-83cb-eb9998cc3059 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 908.558467] env[65726]: DEBUG nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] [instance: 8e9c27f6-e9fa-4ed0-b6ae-3b5df52d1c44] Instance has had 0 of 5 cleanup attempts {{(pid=65726) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 908.561338] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07ec8caf-1673-4674-b141-1e78c119d452 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.573590] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da5b2319-f1c2-4746-8cbb-e98d27524e4c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.666033] env[65726]: DEBUG oslo_concurrency.lockutils [None req-fa9b16af-8474-4ab3-a554-b5fbaf4d16fd tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Acquiring lock "refresh_cache-1cfb1bba-1e21-47eb-b50f-a86575846a65" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 908.666033] env[65726]: DEBUG oslo_concurrency.lockutils [None req-fa9b16af-8474-4ab3-a554-b5fbaf4d16fd tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Acquired lock "refresh_cache-1cfb1bba-1e21-47eb-b50f-a86575846a65" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 908.666033] env[65726]: DEBUG nova.network.neutron [None req-fa9b16af-8474-4ab3-a554-b5fbaf4d16fd tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] [instance: 1cfb1bba-1e21-47eb-b50f-a86575846a65] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 908.702198] env[65726]: DEBUG oslo_concurrency.lockutils [None req-140eec77-ac75-4bcd-8eca-c7b64f5e10aa tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Lock "4ea2b70b-0cfe-4fbf-817a-baa028b8372a" "released" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: held 6.555s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 908.709258] env[65726]: DEBUG oslo_concurrency.lockutils [None req-4826faf0-9908-4f3c-a509-20fe2a708723 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Lock "29823bc7-3909-4ab7-8119-91ee59b289c5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 73.323s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 908.766857] env[65726]: DEBUG oslo_vmware.api [None req-9c76632b-07ea-476e-92d3-ffb457aef721 tempest-ServerAddressesTestJSON-569100133 tempest-ServerAddressesTestJSON-569100133-project-member] Task: {'id': task-5116190, 'name': PowerOnVM_Task, 'duration_secs': 1.588088} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.767286] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-9c76632b-07ea-476e-92d3-ffb457aef721 tempest-ServerAddressesTestJSON-569100133 tempest-ServerAddressesTestJSON-569100133-project-member] [instance: 70375746-06dd-4710-9136-95e9e5759a0e] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 908.767568] env[65726]: INFO nova.compute.manager [None req-9c76632b-07ea-476e-92d3-ffb457aef721 tempest-ServerAddressesTestJSON-569100133 tempest-ServerAddressesTestJSON-569100133-project-member] [instance: 70375746-06dd-4710-9136-95e9e5759a0e] Took 9.14 seconds to spawn the instance on the hypervisor. [ 908.767746] env[65726]: DEBUG nova.compute.manager [None req-9c76632b-07ea-476e-92d3-ffb457aef721 tempest-ServerAddressesTestJSON-569100133 tempest-ServerAddressesTestJSON-569100133-project-member] [instance: 70375746-06dd-4710-9136-95e9e5759a0e] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 908.768775] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97fe79d3-7526-4b4a-a6b8-d3c63d451358 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.800775] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116193, 'name': CreateVM_Task, 'duration_secs': 0.484214} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.800775] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5a252ef1-93c9-4bff-842b-b64df2bc5d75] Created VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 908.801813] env[65726]: WARNING openstack [None req-e5e88d5e-aa68-4ff5-9280-b2cfe1d14b27 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 908.802432] env[65726]: WARNING openstack [None req-e5e88d5e-aa68-4ff5-9280-b2cfe1d14b27 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 908.810422] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e5e88d5e-aa68-4ff5-9280-b2cfe1d14b27 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 908.810687] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e5e88d5e-aa68-4ff5-9280-b2cfe1d14b27 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 908.811238] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e5e88d5e-aa68-4ff5-9280-b2cfe1d14b27 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 908.811598] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-216ac068-7723-4a46-82e7-035da91d8855 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.819285] env[65726]: DEBUG oslo_concurrency.lockutils [None req-17b5a1c7-4bc8-41d1-988f-c933bb6421ad tempest-ServersTestManualDisk-1753452859 tempest-ServersTestManualDisk-1753452859-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.333s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 908.820034] env[65726]: DEBUG nova.compute.manager [None req-17b5a1c7-4bc8-41d1-988f-c933bb6421ad tempest-ServersTestManualDisk-1753452859 tempest-ServersTestManualDisk-1753452859-project-member] [instance: 3088ee5e-0d27-4058-a94b-2e04c3b52add] Start building networks asynchronously for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 908.824840] env[65726]: DEBUG oslo_vmware.api [None req-e5e88d5e-aa68-4ff5-9280-b2cfe1d14b27 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Waiting for the task: (returnval){ [ 908.824840] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52641dff-251c-17aa-e950-96cd329c8cb2" [ 908.824840] env[65726]: _type = "Task" [ 908.824840] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.826344] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a4b79740-ccaa-4794-bd78-55e7c8646033 tempest-ImagesNegativeTestJSON-1015404931 tempest-ImagesNegativeTestJSON-1015404931-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 36.432s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 908.826686] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a4b79740-ccaa-4794-bd78-55e7c8646033 tempest-ImagesNegativeTestJSON-1015404931 tempest-ImagesNegativeTestJSON-1015404931-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 908.829431] env[65726]: DEBUG oslo_concurrency.lockutils [None req-8333c704-ca27-426e-acee-8cfd63213196 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 32.573s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 908.829652] env[65726]: DEBUG nova.objects.instance [None req-8333c704-ca27-426e-acee-8cfd63213196 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: f2d9090c-988f-43f4-9c81-7aa718a3438a] Trying to apply a migration context that does not seem to be set for this instance {{(pid=65726) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 908.845039] env[65726]: DEBUG oslo_vmware.api [None req-e5e88d5e-aa68-4ff5-9280-b2cfe1d14b27 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52641dff-251c-17aa-e950-96cd329c8cb2, 'name': SearchDatastore_Task, 'duration_secs': 0.01843} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.845039] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e5e88d5e-aa68-4ff5-9280-b2cfe1d14b27 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 908.845195] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-e5e88d5e-aa68-4ff5-9280-b2cfe1d14b27 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: 5a252ef1-93c9-4bff-842b-b64df2bc5d75] Processing image b52362a3-ee8a-4cbf-b06f-513b0cc8f95c {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 908.845468] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e5e88d5e-aa68-4ff5-9280-b2cfe1d14b27 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 908.845616] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e5e88d5e-aa68-4ff5-9280-b2cfe1d14b27 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 908.845793] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-e5e88d5e-aa68-4ff5-9280-b2cfe1d14b27 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 908.846395] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0a039409-3a82-4a35-99ce-0e6468b3db67 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.858745] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-e5e88d5e-aa68-4ff5-9280-b2cfe1d14b27 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 908.859249] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-e5e88d5e-aa68-4ff5-9280-b2cfe1d14b27 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65726) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 908.861054] env[65726]: INFO nova.scheduler.client.report [None req-a4b79740-ccaa-4794-bd78-55e7c8646033 tempest-ImagesNegativeTestJSON-1015404931 tempest-ImagesNegativeTestJSON-1015404931-project-member] Deleted allocations for instance c4177e20-b1bd-4b54-a275-c93582359a07 [ 908.862437] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e2f051e8-4d01-4e80-8963-fcb07f556043 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.874759] env[65726]: DEBUG oslo_vmware.api [None req-e5e88d5e-aa68-4ff5-9280-b2cfe1d14b27 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Waiting for the task: (returnval){ [ 908.874759] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]521babf2-5070-fbeb-1fd4-87011519db41" [ 908.874759] env[65726]: _type = "Task" [ 908.874759] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.889018] env[65726]: DEBUG oslo_vmware.api [None req-e5e88d5e-aa68-4ff5-9280-b2cfe1d14b27 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]521babf2-5070-fbeb-1fd4-87011519db41, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.973295] env[65726]: DEBUG oslo_vmware.api [None req-31642568-e231-4328-a969-4d06d89d9576 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Task: {'id': task-5116191, 'name': CloneVM_Task, 'duration_secs': 1.916167} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.973563] env[65726]: INFO nova.virt.vmwareapi.vmops [None req-31642568-e231-4328-a969-4d06d89d9576 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] [instance: c4cc4f85-cb35-4edc-a58b-adfee0ce1265] Created linked-clone VM from snapshot [ 908.974780] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13141abb-45c9-402b-820e-0261c1115d46 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.982973] env[65726]: DEBUG nova.virt.vmwareapi.images [None req-31642568-e231-4328-a969-4d06d89d9576 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] [instance: c4cc4f85-cb35-4edc-a58b-adfee0ce1265] Uploading image fa1d1620-a4ff-4711-acf9-5d56f6b9b63a {{(pid=65726) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 909.009372] env[65726]: DEBUG oslo_vmware.rw_handles [None req-31642568-e231-4328-a969-4d06d89d9576 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 909.009372] env[65726]: value = "vm-995198" [ 909.009372] env[65726]: _type = "VirtualMachine" [ 909.009372] env[65726]: }. {{(pid=65726) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 909.009957] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-cfe7a766-bd0b-4264-8a93-67a83a66a15b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.019335] env[65726]: DEBUG oslo_vmware.rw_handles [None req-31642568-e231-4328-a969-4d06d89d9576 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Lease: (returnval){ [ 909.019335] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52ef282c-7de5-5f44-1e1d-e9ce835da96a" [ 909.019335] env[65726]: _type = "HttpNfcLease" [ 909.019335] env[65726]: } obtained for exporting VM: (result){ [ 909.019335] env[65726]: value = "vm-995198" [ 909.019335] env[65726]: _type = "VirtualMachine" [ 909.019335] env[65726]: }. {{(pid=65726) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 909.019730] env[65726]: DEBUG oslo_vmware.api [None req-31642568-e231-4328-a969-4d06d89d9576 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Waiting for the lease: (returnval){ [ 909.019730] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52ef282c-7de5-5f44-1e1d-e9ce835da96a" [ 909.019730] env[65726]: _type = "HttpNfcLease" [ 909.019730] env[65726]: } to be ready. {{(pid=65726) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 909.027133] env[65726]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 909.027133] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52ef282c-7de5-5f44-1e1d-e9ce835da96a" [ 909.027133] env[65726]: _type = "HttpNfcLease" [ 909.027133] env[65726]: } is initializing. {{(pid=65726) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 909.067702] env[65726]: DEBUG nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] [instance: bb64fc41-19d7-49dc-a17a-9dd74730130f] Instance has had 0 of 5 cleanup attempts {{(pid=65726) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 909.131278] env[65726]: DEBUG nova.network.neutron [None req-89fee262-eea7-4ed4-83cb-eb9998cc3059 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 6f91b053-772a-4497-b29d-349b960c55eb] Successfully updated port: fbccb731-d98d-40ad-a144-4b4f1e48a1ce {{(pid=65726) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 909.169153] env[65726]: WARNING openstack [None req-fa9b16af-8474-4ab3-a554-b5fbaf4d16fd tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 909.169153] env[65726]: WARNING openstack [None req-fa9b16af-8474-4ab3-a554-b5fbaf4d16fd tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 909.177726] env[65726]: DEBUG nova.network.neutron [None req-fa9b16af-8474-4ab3-a554-b5fbaf4d16fd tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] [instance: 1cfb1bba-1e21-47eb-b50f-a86575846a65] Instance cache missing network info. {{(pid=65726) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 909.232338] env[65726]: DEBUG nova.compute.manager [req-9d33368f-00d1-4c23-b602-24ad159961cd req-90ffd4c2-f693-435e-996e-3a5f1b73fa56 service nova] [instance: 5a252ef1-93c9-4bff-842b-b64df2bc5d75] Received event network-changed-8cd2f231-f537-4509-92da-dc52e57441ba {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 909.232338] env[65726]: DEBUG nova.compute.manager [req-9d33368f-00d1-4c23-b602-24ad159961cd req-90ffd4c2-f693-435e-996e-3a5f1b73fa56 service nova] [instance: 5a252ef1-93c9-4bff-842b-b64df2bc5d75] Refreshing instance network info cache due to event network-changed-8cd2f231-f537-4509-92da-dc52e57441ba. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 909.232574] env[65726]: DEBUG oslo_concurrency.lockutils [req-9d33368f-00d1-4c23-b602-24ad159961cd req-90ffd4c2-f693-435e-996e-3a5f1b73fa56 service nova] Acquiring lock "refresh_cache-5a252ef1-93c9-4bff-842b-b64df2bc5d75" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 909.232733] env[65726]: DEBUG oslo_concurrency.lockutils [req-9d33368f-00d1-4c23-b602-24ad159961cd req-90ffd4c2-f693-435e-996e-3a5f1b73fa56 service nova] Acquired lock "refresh_cache-5a252ef1-93c9-4bff-842b-b64df2bc5d75" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 909.233040] env[65726]: DEBUG nova.network.neutron [req-9d33368f-00d1-4c23-b602-24ad159961cd req-90ffd4c2-f693-435e-996e-3a5f1b73fa56 service nova] [instance: 5a252ef1-93c9-4bff-842b-b64df2bc5d75] Refreshing network info cache for port 8cd2f231-f537-4509-92da-dc52e57441ba {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 909.276516] env[65726]: DEBUG nova.network.neutron [None req-fa9b16af-8474-4ab3-a554-b5fbaf4d16fd tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] [instance: 1cfb1bba-1e21-47eb-b50f-a86575846a65] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 909.291030] env[65726]: INFO nova.compute.manager [None req-9c76632b-07ea-476e-92d3-ffb457aef721 tempest-ServerAddressesTestJSON-569100133 tempest-ServerAddressesTestJSON-569100133-project-member] [instance: 70375746-06dd-4710-9136-95e9e5759a0e] Took 44.49 seconds to build instance. [ 909.328347] env[65726]: DEBUG nova.compute.utils [None req-17b5a1c7-4bc8-41d1-988f-c933bb6421ad tempest-ServersTestManualDisk-1753452859 tempest-ServersTestManualDisk-1753452859-project-member] Using /dev/sd instead of None {{(pid=65726) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 909.330585] env[65726]: DEBUG nova.compute.manager [None req-17b5a1c7-4bc8-41d1-988f-c933bb6421ad tempest-ServersTestManualDisk-1753452859 tempest-ServersTestManualDisk-1753452859-project-member] [instance: 3088ee5e-0d27-4058-a94b-2e04c3b52add] Allocating IP information in the background. {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 909.331036] env[65726]: DEBUG nova.network.neutron [None req-17b5a1c7-4bc8-41d1-988f-c933bb6421ad tempest-ServersTestManualDisk-1753452859 tempest-ServersTestManualDisk-1753452859-project-member] [instance: 3088ee5e-0d27-4058-a94b-2e04c3b52add] allocate_for_instance() {{(pid=65726) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 909.331516] env[65726]: WARNING neutronclient.v2_0.client [None req-17b5a1c7-4bc8-41d1-988f-c933bb6421ad tempest-ServersTestManualDisk-1753452859 tempest-ServersTestManualDisk-1753452859-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 909.331980] env[65726]: WARNING neutronclient.v2_0.client [None req-17b5a1c7-4bc8-41d1-988f-c933bb6421ad tempest-ServersTestManualDisk-1753452859 tempest-ServersTestManualDisk-1753452859-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 909.332593] env[65726]: WARNING openstack [None req-17b5a1c7-4bc8-41d1-988f-c933bb6421ad tempest-ServersTestManualDisk-1753452859 tempest-ServersTestManualDisk-1753452859-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 909.333127] env[65726]: WARNING openstack [None req-17b5a1c7-4bc8-41d1-988f-c933bb6421ad tempest-ServersTestManualDisk-1753452859 tempest-ServersTestManualDisk-1753452859-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 909.356625] env[65726]: DEBUG nova.compute.manager [req-faada6ad-da73-42c9-8a27-de5f8fc8fb30 req-a837b40f-9ce7-4eaa-82dc-85274f92866e service nova] [instance: 6f91b053-772a-4497-b29d-349b960c55eb] Received event network-vif-plugged-fbccb731-d98d-40ad-a144-4b4f1e48a1ce {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 909.357212] env[65726]: DEBUG oslo_concurrency.lockutils [req-faada6ad-da73-42c9-8a27-de5f8fc8fb30 req-a837b40f-9ce7-4eaa-82dc-85274f92866e service nova] Acquiring lock "6f91b053-772a-4497-b29d-349b960c55eb-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 909.357212] env[65726]: DEBUG oslo_concurrency.lockutils [req-faada6ad-da73-42c9-8a27-de5f8fc8fb30 req-a837b40f-9ce7-4eaa-82dc-85274f92866e service nova] Lock "6f91b053-772a-4497-b29d-349b960c55eb-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 909.357372] env[65726]: DEBUG oslo_concurrency.lockutils [req-faada6ad-da73-42c9-8a27-de5f8fc8fb30 req-a837b40f-9ce7-4eaa-82dc-85274f92866e service nova] Lock "6f91b053-772a-4497-b29d-349b960c55eb-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 909.357535] env[65726]: DEBUG nova.compute.manager [req-faada6ad-da73-42c9-8a27-de5f8fc8fb30 req-a837b40f-9ce7-4eaa-82dc-85274f92866e service nova] [instance: 6f91b053-772a-4497-b29d-349b960c55eb] No waiting events found dispatching network-vif-plugged-fbccb731-d98d-40ad-a144-4b4f1e48a1ce {{(pid=65726) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 909.357723] env[65726]: WARNING nova.compute.manager [req-faada6ad-da73-42c9-8a27-de5f8fc8fb30 req-a837b40f-9ce7-4eaa-82dc-85274f92866e service nova] [instance: 6f91b053-772a-4497-b29d-349b960c55eb] Received unexpected event network-vif-plugged-fbccb731-d98d-40ad-a144-4b4f1e48a1ce for instance with vm_state building and task_state spawning. [ 909.375914] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a4b79740-ccaa-4794-bd78-55e7c8646033 tempest-ImagesNegativeTestJSON-1015404931 tempest-ImagesNegativeTestJSON-1015404931-project-member] Lock "c4177e20-b1bd-4b54-a275-c93582359a07" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 40.765s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 909.388700] env[65726]: DEBUG oslo_vmware.api [None req-e5e88d5e-aa68-4ff5-9280-b2cfe1d14b27 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]521babf2-5070-fbeb-1fd4-87011519db41, 'name': SearchDatastore_Task, 'duration_secs': 0.022951} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.392693] env[65726]: DEBUG nova.policy [None req-17b5a1c7-4bc8-41d1-988f-c933bb6421ad tempest-ServersTestManualDisk-1753452859 tempest-ServersTestManualDisk-1753452859-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'fa3d109114b749239c1a4685653deb25', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '308ec5a4cf844d2782db77b3edcb4127', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65726) authorize /opt/stack/nova/nova/policy.py:192}} [ 909.396757] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-53a5ad61-4874-4973-b344-9f3ea43d6fe5 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.404530] env[65726]: DEBUG oslo_vmware.api [None req-e5e88d5e-aa68-4ff5-9280-b2cfe1d14b27 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Waiting for the task: (returnval){ [ 909.404530] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52955c64-a317-644b-ced4-529f84382b2f" [ 909.404530] env[65726]: _type = "Task" [ 909.404530] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.414826] env[65726]: DEBUG oslo_vmware.api [None req-e5e88d5e-aa68-4ff5-9280-b2cfe1d14b27 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52955c64-a317-644b-ced4-529f84382b2f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.531081] env[65726]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 909.531081] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52ef282c-7de5-5f44-1e1d-e9ce835da96a" [ 909.531081] env[65726]: _type = "HttpNfcLease" [ 909.531081] env[65726]: } is ready. {{(pid=65726) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 909.531462] env[65726]: DEBUG oslo_vmware.rw_handles [None req-31642568-e231-4328-a969-4d06d89d9576 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 909.531462] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52ef282c-7de5-5f44-1e1d-e9ce835da96a" [ 909.531462] env[65726]: _type = "HttpNfcLease" [ 909.531462] env[65726]: }. {{(pid=65726) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 909.532472] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc57fdbc-6f28-422a-ac2c-335a898ac6c5 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.543695] env[65726]: DEBUG oslo_vmware.rw_handles [None req-31642568-e231-4328-a969-4d06d89d9576 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/523e089c-0ef7-904e-9b60-5ea26927a04d/disk-0.vmdk from lease info. {{(pid=65726) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 909.544015] env[65726]: DEBUG oslo_vmware.rw_handles [None req-31642568-e231-4328-a969-4d06d89d9576 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/523e089c-0ef7-904e-9b60-5ea26927a04d/disk-0.vmdk for reading. {{(pid=65726) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 909.605342] env[65726]: DEBUG oslo_service.periodic_task [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=65726) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 909.605529] env[65726]: DEBUG nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Cleaning up deleted instances with incomplete migration {{(pid=65726) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11970}} [ 909.634564] env[65726]: DEBUG oslo_concurrency.lockutils [None req-89fee262-eea7-4ed4-83cb-eb9998cc3059 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Acquiring lock "refresh_cache-6f91b053-772a-4497-b29d-349b960c55eb" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 909.635017] env[65726]: DEBUG oslo_concurrency.lockutils [None req-89fee262-eea7-4ed4-83cb-eb9998cc3059 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Acquired lock "refresh_cache-6f91b053-772a-4497-b29d-349b960c55eb" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 909.635017] env[65726]: DEBUG nova.network.neutron [None req-89fee262-eea7-4ed4-83cb-eb9998cc3059 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 6f91b053-772a-4497-b29d-349b960c55eb] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 909.650936] env[65726]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-b0bf53ee-b727-4ff0-a4ed-454718bc4307 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.686917] env[65726]: DEBUG nova.network.neutron [None req-17b5a1c7-4bc8-41d1-988f-c933bb6421ad tempest-ServersTestManualDisk-1753452859 tempest-ServersTestManualDisk-1753452859-project-member] [instance: 3088ee5e-0d27-4058-a94b-2e04c3b52add] Successfully created port: 2ffc16a0-351c-4f5b-8a4a-993ae7adb4f3 {{(pid=65726) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 909.737224] env[65726]: WARNING openstack [req-9d33368f-00d1-4c23-b602-24ad159961cd req-90ffd4c2-f693-435e-996e-3a5f1b73fa56 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 909.737224] env[65726]: WARNING openstack [req-9d33368f-00d1-4c23-b602-24ad159961cd req-90ffd4c2-f693-435e-996e-3a5f1b73fa56 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 909.782754] env[65726]: DEBUG oslo_concurrency.lockutils [None req-fa9b16af-8474-4ab3-a554-b5fbaf4d16fd tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Releasing lock "refresh_cache-1cfb1bba-1e21-47eb-b50f-a86575846a65" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 909.782754] env[65726]: DEBUG nova.compute.manager [None req-fa9b16af-8474-4ab3-a554-b5fbaf4d16fd tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] [instance: 1cfb1bba-1e21-47eb-b50f-a86575846a65] Start destroying the instance on the hypervisor. {{(pid=65726) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 909.782754] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-fa9b16af-8474-4ab3-a554-b5fbaf4d16fd tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] [instance: 1cfb1bba-1e21-47eb-b50f-a86575846a65] Destroying instance {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 909.784212] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a33f2de-29cf-40f3-b9d5-1f95353591b9 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.794094] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9c76632b-07ea-476e-92d3-ffb457aef721 tempest-ServerAddressesTestJSON-569100133 tempest-ServerAddressesTestJSON-569100133-project-member] Lock "70375746-06dd-4710-9136-95e9e5759a0e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 61.660s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 909.794525] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-fa9b16af-8474-4ab3-a554-b5fbaf4d16fd tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] [instance: 1cfb1bba-1e21-47eb-b50f-a86575846a65] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 909.794748] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b9788f0c-ee89-477f-8875-63888af1677d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.803840] env[65726]: DEBUG oslo_vmware.api [None req-fa9b16af-8474-4ab3-a554-b5fbaf4d16fd tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Waiting for the task: (returnval){ [ 909.803840] env[65726]: value = "task-5116195" [ 909.803840] env[65726]: _type = "Task" [ 909.803840] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.813951] env[65726]: DEBUG oslo_vmware.api [None req-fa9b16af-8474-4ab3-a554-b5fbaf4d16fd tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Task: {'id': task-5116195, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.842289] env[65726]: DEBUG nova.compute.manager [None req-17b5a1c7-4bc8-41d1-988f-c933bb6421ad tempest-ServersTestManualDisk-1753452859 tempest-ServersTestManualDisk-1753452859-project-member] [instance: 3088ee5e-0d27-4058-a94b-2e04c3b52add] Start building block device mappings for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 909.848128] env[65726]: DEBUG oslo_concurrency.lockutils [None req-8333c704-ca27-426e-acee-8cfd63213196 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.019s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 909.849567] env[65726]: DEBUG oslo_concurrency.lockutils [None req-037b42a1-7b2a-4493-8518-be35c6d71062 tempest-ServerMetadataNegativeTestJSON-936475568 tempest-ServerMetadataNegativeTestJSON-936475568-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 32.125s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 909.849764] env[65726]: DEBUG oslo_concurrency.lockutils [None req-037b42a1-7b2a-4493-8518-be35c6d71062 tempest-ServerMetadataNegativeTestJSON-936475568 tempest-ServerMetadataNegativeTestJSON-936475568-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 909.852585] env[65726]: DEBUG oslo_concurrency.lockutils [None req-2cc737a8-ad7d-449c-b270-9db9e18d971d tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 28.429s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 909.854381] env[65726]: INFO nova.compute.claims [None req-2cc737a8-ad7d-449c-b270-9db9e18d971d tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: dcde25df-87a5-47fb-94ce-334f68894e04] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 909.887636] env[65726]: WARNING openstack [req-9d33368f-00d1-4c23-b602-24ad159961cd req-90ffd4c2-f693-435e-996e-3a5f1b73fa56 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 909.887781] env[65726]: WARNING openstack [req-9d33368f-00d1-4c23-b602-24ad159961cd req-90ffd4c2-f693-435e-996e-3a5f1b73fa56 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 909.901083] env[65726]: INFO nova.scheduler.client.report [None req-037b42a1-7b2a-4493-8518-be35c6d71062 tempest-ServerMetadataNegativeTestJSON-936475568 tempest-ServerMetadataNegativeTestJSON-936475568-project-member] Deleted allocations for instance d179030a-e5ba-45b1-ad11-4d2f71ed7bd4 [ 909.924270] env[65726]: DEBUG oslo_vmware.api [None req-e5e88d5e-aa68-4ff5-9280-b2cfe1d14b27 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52955c64-a317-644b-ced4-529f84382b2f, 'name': SearchDatastore_Task, 'duration_secs': 0.013519} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.924270] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e5e88d5e-aa68-4ff5-9280-b2cfe1d14b27 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 909.924658] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-e5e88d5e-aa68-4ff5-9280-b2cfe1d14b27 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] 5a252ef1-93c9-4bff-842b-b64df2bc5d75/5a252ef1-93c9-4bff-842b-b64df2bc5d75.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 909.925458] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-122fd6ad-7195-40c3-8f04-76740a260970 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.936083] env[65726]: DEBUG oslo_vmware.api [None req-e5e88d5e-aa68-4ff5-9280-b2cfe1d14b27 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Waiting for the task: (returnval){ [ 909.936083] env[65726]: value = "task-5116196" [ 909.936083] env[65726]: _type = "Task" [ 909.936083] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.955063] env[65726]: DEBUG oslo_vmware.api [None req-e5e88d5e-aa68-4ff5-9280-b2cfe1d14b27 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': task-5116196, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.021892] env[65726]: WARNING openstack [req-9d33368f-00d1-4c23-b602-24ad159961cd req-90ffd4c2-f693-435e-996e-3a5f1b73fa56 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 910.022539] env[65726]: WARNING openstack [req-9d33368f-00d1-4c23-b602-24ad159961cd req-90ffd4c2-f693-435e-996e-3a5f1b73fa56 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 910.113482] env[65726]: DEBUG oslo_service.periodic_task [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=65726) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 910.124541] env[65726]: DEBUG nova.network.neutron [req-9d33368f-00d1-4c23-b602-24ad159961cd req-90ffd4c2-f693-435e-996e-3a5f1b73fa56 service nova] [instance: 5a252ef1-93c9-4bff-842b-b64df2bc5d75] Updated VIF entry in instance network info cache for port 8cd2f231-f537-4509-92da-dc52e57441ba. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 910.125344] env[65726]: DEBUG nova.network.neutron [req-9d33368f-00d1-4c23-b602-24ad159961cd req-90ffd4c2-f693-435e-996e-3a5f1b73fa56 service nova] [instance: 5a252ef1-93c9-4bff-842b-b64df2bc5d75] Updating instance_info_cache with network_info: [{"id": "8cd2f231-f537-4509-92da-dc52e57441ba", "address": "fa:16:3e:f7:41:63", "network": {"id": "4c972d9d-db3b-4779-ac39-39d9a2d141dd", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1418913861-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cbb17a09b35c4c22ade5c4082b10bda9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "50cf0a70-948d-4611-af05-94c1483064ed", "external-id": "nsx-vlan-transportzone-536", "segmentation_id": 536, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8cd2f231-f5", "ovs_interfaceid": "8cd2f231-f537-4509-92da-dc52e57441ba", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 910.139569] env[65726]: WARNING openstack [None req-89fee262-eea7-4ed4-83cb-eb9998cc3059 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 910.142526] env[65726]: WARNING openstack [None req-89fee262-eea7-4ed4-83cb-eb9998cc3059 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 910.148683] env[65726]: DEBUG nova.network.neutron [None req-89fee262-eea7-4ed4-83cb-eb9998cc3059 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 6f91b053-772a-4497-b29d-349b960c55eb] Instance cache missing network info. {{(pid=65726) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 910.228121] env[65726]: WARNING openstack [None req-89fee262-eea7-4ed4-83cb-eb9998cc3059 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 910.228787] env[65726]: WARNING openstack [None req-89fee262-eea7-4ed4-83cb-eb9998cc3059 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 910.322063] env[65726]: DEBUG oslo_vmware.api [None req-fa9b16af-8474-4ab3-a554-b5fbaf4d16fd tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Task: {'id': task-5116195, 'name': PowerOffVM_Task, 'duration_secs': 0.154628} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.322539] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-fa9b16af-8474-4ab3-a554-b5fbaf4d16fd tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] [instance: 1cfb1bba-1e21-47eb-b50f-a86575846a65] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 910.322813] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-fa9b16af-8474-4ab3-a554-b5fbaf4d16fd tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] [instance: 1cfb1bba-1e21-47eb-b50f-a86575846a65] Unregistering the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 910.324030] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7197749e-6ba3-42a8-8d33-19ac71bb4a31 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.363935] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-fa9b16af-8474-4ab3-a554-b5fbaf4d16fd tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] [instance: 1cfb1bba-1e21-47eb-b50f-a86575846a65] Unregistered the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 910.364950] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-fa9b16af-8474-4ab3-a554-b5fbaf4d16fd tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] [instance: 1cfb1bba-1e21-47eb-b50f-a86575846a65] Deleting contents of the VM from datastore datastore2 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 910.365087] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-fa9b16af-8474-4ab3-a554-b5fbaf4d16fd tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Deleting the datastore file [datastore2] 1cfb1bba-1e21-47eb-b50f-a86575846a65 {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 910.365657] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ee145c17-63eb-465a-816d-08055df27088 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.388766] env[65726]: WARNING openstack [None req-89fee262-eea7-4ed4-83cb-eb9998cc3059 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 910.389370] env[65726]: WARNING openstack [None req-89fee262-eea7-4ed4-83cb-eb9998cc3059 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 910.407279] env[65726]: DEBUG oslo_vmware.api [None req-fa9b16af-8474-4ab3-a554-b5fbaf4d16fd tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Waiting for the task: (returnval){ [ 910.407279] env[65726]: value = "task-5116198" [ 910.407279] env[65726]: _type = "Task" [ 910.407279] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.419589] env[65726]: DEBUG oslo_concurrency.lockutils [None req-037b42a1-7b2a-4493-8518-be35c6d71062 tempest-ServerMetadataNegativeTestJSON-936475568 tempest-ServerMetadataNegativeTestJSON-936475568-project-member] Lock "d179030a-e5ba-45b1-ad11-4d2f71ed7bd4" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 36.931s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 910.430024] env[65726]: DEBUG oslo_vmware.api [None req-fa9b16af-8474-4ab3-a554-b5fbaf4d16fd tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Task: {'id': task-5116198, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.452931] env[65726]: DEBUG oslo_vmware.api [None req-e5e88d5e-aa68-4ff5-9280-b2cfe1d14b27 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': task-5116196, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.599338] env[65726]: DEBUG oslo_concurrency.lockutils [None req-7acc7b96-c5e1-4976-bfde-d3cb706afaa3 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Acquiring lock "76249623-6f83-46a3-b8c5-c001111aa698" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 910.600320] env[65726]: DEBUG oslo_concurrency.lockutils [None req-7acc7b96-c5e1-4976-bfde-d3cb706afaa3 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Lock "76249623-6f83-46a3-b8c5-c001111aa698" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 910.600320] env[65726]: DEBUG oslo_concurrency.lockutils [None req-7acc7b96-c5e1-4976-bfde-d3cb706afaa3 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Acquiring lock "76249623-6f83-46a3-b8c5-c001111aa698-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 910.601101] env[65726]: DEBUG oslo_concurrency.lockutils [None req-7acc7b96-c5e1-4976-bfde-d3cb706afaa3 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Lock "76249623-6f83-46a3-b8c5-c001111aa698-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 910.601101] env[65726]: DEBUG oslo_concurrency.lockutils [None req-7acc7b96-c5e1-4976-bfde-d3cb706afaa3 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Lock "76249623-6f83-46a3-b8c5-c001111aa698-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 910.605857] env[65726]: INFO nova.compute.manager [None req-7acc7b96-c5e1-4976-bfde-d3cb706afaa3 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] [instance: 76249623-6f83-46a3-b8c5-c001111aa698] Terminating instance [ 910.630340] env[65726]: DEBUG oslo_concurrency.lockutils [req-9d33368f-00d1-4c23-b602-24ad159961cd req-90ffd4c2-f693-435e-996e-3a5f1b73fa56 service nova] Releasing lock "refresh_cache-5a252ef1-93c9-4bff-842b-b64df2bc5d75" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 910.859943] env[65726]: DEBUG nova.compute.manager [None req-17b5a1c7-4bc8-41d1-988f-c933bb6421ad tempest-ServersTestManualDisk-1753452859 tempest-ServersTestManualDisk-1753452859-project-member] [instance: 3088ee5e-0d27-4058-a94b-2e04c3b52add] Start spawning the instance on the hypervisor. {{(pid=65726) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 910.875370] env[65726]: DEBUG nova.network.neutron [None req-89fee262-eea7-4ed4-83cb-eb9998cc3059 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 6f91b053-772a-4497-b29d-349b960c55eb] Updating instance_info_cache with network_info: [{"id": "fbccb731-d98d-40ad-a144-4b4f1e48a1ce", "address": "fa:16:3e:06:3e:65", "network": {"id": "e206b5b9-da36-4e6f-926f-914dec38fac2", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1053201655-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "39067912d091464192516c08eb576696", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7c80243e-93a7-4a95-bc8d-e9534bacd66e", "external-id": "nsx-vlan-transportzone-306", "segmentation_id": 306, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfbccb731-d9", "ovs_interfaceid": "fbccb731-d98d-40ad-a144-4b4f1e48a1ce", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 910.910658] env[65726]: DEBUG nova.virt.hardware [None req-17b5a1c7-4bc8-41d1-988f-c933bb6421ad tempest-ServersTestManualDisk-1753452859 tempest-ServersTestManualDisk-1753452859-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T19:28:34Z,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1539f06fa3534e90acbc3a60c4b8bd3f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T19:28:35Z,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 910.910978] env[65726]: DEBUG nova.virt.hardware [None req-17b5a1c7-4bc8-41d1-988f-c933bb6421ad tempest-ServersTestManualDisk-1753452859 tempest-ServersTestManualDisk-1753452859-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 910.911268] env[65726]: DEBUG nova.virt.hardware [None req-17b5a1c7-4bc8-41d1-988f-c933bb6421ad tempest-ServersTestManualDisk-1753452859 tempest-ServersTestManualDisk-1753452859-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 910.911624] env[65726]: DEBUG nova.virt.hardware [None req-17b5a1c7-4bc8-41d1-988f-c933bb6421ad tempest-ServersTestManualDisk-1753452859 tempest-ServersTestManualDisk-1753452859-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 910.911979] env[65726]: DEBUG nova.virt.hardware [None req-17b5a1c7-4bc8-41d1-988f-c933bb6421ad tempest-ServersTestManualDisk-1753452859 tempest-ServersTestManualDisk-1753452859-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 910.912144] env[65726]: DEBUG nova.virt.hardware [None req-17b5a1c7-4bc8-41d1-988f-c933bb6421ad tempest-ServersTestManualDisk-1753452859 tempest-ServersTestManualDisk-1753452859-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 910.912370] env[65726]: DEBUG nova.virt.hardware [None req-17b5a1c7-4bc8-41d1-988f-c933bb6421ad tempest-ServersTestManualDisk-1753452859 tempest-ServersTestManualDisk-1753452859-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 910.912541] env[65726]: DEBUG nova.virt.hardware [None req-17b5a1c7-4bc8-41d1-988f-c933bb6421ad tempest-ServersTestManualDisk-1753452859 tempest-ServersTestManualDisk-1753452859-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 910.912712] env[65726]: DEBUG nova.virt.hardware [None req-17b5a1c7-4bc8-41d1-988f-c933bb6421ad tempest-ServersTestManualDisk-1753452859 tempest-ServersTestManualDisk-1753452859-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 910.912878] env[65726]: DEBUG nova.virt.hardware [None req-17b5a1c7-4bc8-41d1-988f-c933bb6421ad tempest-ServersTestManualDisk-1753452859 tempest-ServersTestManualDisk-1753452859-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 910.913452] env[65726]: DEBUG nova.virt.hardware [None req-17b5a1c7-4bc8-41d1-988f-c933bb6421ad tempest-ServersTestManualDisk-1753452859 tempest-ServersTestManualDisk-1753452859-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 910.914677] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64e27d03-7a4b-4a57-ac32-a71af9e3816c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.938924] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0eb2314-533a-4de3-a392-b8ddd75076ee {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.943689] env[65726]: DEBUG oslo_vmware.api [None req-fa9b16af-8474-4ab3-a554-b5fbaf4d16fd tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Task: {'id': task-5116198, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.256532} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.952500] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-fa9b16af-8474-4ab3-a554-b5fbaf4d16fd tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Deleted the datastore file {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 910.952500] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-fa9b16af-8474-4ab3-a554-b5fbaf4d16fd tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] [instance: 1cfb1bba-1e21-47eb-b50f-a86575846a65] Deleted contents of the VM from datastore datastore2 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 910.952500] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-fa9b16af-8474-4ab3-a554-b5fbaf4d16fd tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] [instance: 1cfb1bba-1e21-47eb-b50f-a86575846a65] Instance destroyed {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 910.952500] env[65726]: INFO nova.compute.manager [None req-fa9b16af-8474-4ab3-a554-b5fbaf4d16fd tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] [instance: 1cfb1bba-1e21-47eb-b50f-a86575846a65] Took 1.17 seconds to destroy the instance on the hypervisor. [ 910.952500] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-fa9b16af-8474-4ab3-a554-b5fbaf4d16fd tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 910.953172] env[65726]: DEBUG nova.compute.manager [-] [instance: 1cfb1bba-1e21-47eb-b50f-a86575846a65] Deallocating network for instance {{(pid=65726) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 910.953304] env[65726]: DEBUG nova.network.neutron [-] [instance: 1cfb1bba-1e21-47eb-b50f-a86575846a65] deallocate_for_instance() {{(pid=65726) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 910.953986] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 910.954404] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 910.978282] env[65726]: DEBUG oslo_vmware.api [None req-e5e88d5e-aa68-4ff5-9280-b2cfe1d14b27 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': task-5116196, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.602343} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.978363] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-e5e88d5e-aa68-4ff5-9280-b2cfe1d14b27 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] 5a252ef1-93c9-4bff-842b-b64df2bc5d75/5a252ef1-93c9-4bff-842b-b64df2bc5d75.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 910.978604] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-e5e88d5e-aa68-4ff5-9280-b2cfe1d14b27 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: 5a252ef1-93c9-4bff-842b-b64df2bc5d75] Extending root virtual disk to 1048576 {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 910.978987] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-27094df3-b3c0-47a5-9aa8-45551c14cd30 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.989198] env[65726]: DEBUG oslo_vmware.api [None req-e5e88d5e-aa68-4ff5-9280-b2cfe1d14b27 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Waiting for the task: (returnval){ [ 910.989198] env[65726]: value = "task-5116199" [ 910.989198] env[65726]: _type = "Task" [ 910.989198] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.000040] env[65726]: DEBUG oslo_vmware.api [None req-e5e88d5e-aa68-4ff5-9280-b2cfe1d14b27 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': task-5116199, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.112197] env[65726]: DEBUG nova.compute.manager [None req-7acc7b96-c5e1-4976-bfde-d3cb706afaa3 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] [instance: 76249623-6f83-46a3-b8c5-c001111aa698] Start destroying the instance on the hypervisor. {{(pid=65726) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 911.112627] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-7acc7b96-c5e1-4976-bfde-d3cb706afaa3 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] [instance: 76249623-6f83-46a3-b8c5-c001111aa698] Destroying instance {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 911.113629] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acf391b4-597d-4e34-a8c7-275202206f32 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.127490] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-7acc7b96-c5e1-4976-bfde-d3cb706afaa3 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] [instance: 76249623-6f83-46a3-b8c5-c001111aa698] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 911.127490] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b1452c39-e202-4f17-ba24-9485d7e188f5 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.135083] env[65726]: DEBUG oslo_vmware.api [None req-7acc7b96-c5e1-4976-bfde-d3cb706afaa3 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Waiting for the task: (returnval){ [ 911.135083] env[65726]: value = "task-5116200" [ 911.135083] env[65726]: _type = "Task" [ 911.135083] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.153729] env[65726]: DEBUG oslo_vmware.api [None req-7acc7b96-c5e1-4976-bfde-d3cb706afaa3 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Task: {'id': task-5116200, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.189486] env[65726]: DEBUG nova.compute.manager [None req-e4fa020c-c0fd-49d8-8a8a-f7bebcb49337 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] [instance: 815ee20b-cb07-4514-bd43-1b74972508a5] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 911.190948] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f398cb13-2cbd-4bb3-996b-c71529e134d2 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.381955] env[65726]: DEBUG oslo_concurrency.lockutils [None req-89fee262-eea7-4ed4-83cb-eb9998cc3059 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Releasing lock "refresh_cache-6f91b053-772a-4497-b29d-349b960c55eb" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 911.382611] env[65726]: DEBUG nova.compute.manager [None req-89fee262-eea7-4ed4-83cb-eb9998cc3059 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 6f91b053-772a-4497-b29d-349b960c55eb] Instance network_info: |[{"id": "fbccb731-d98d-40ad-a144-4b4f1e48a1ce", "address": "fa:16:3e:06:3e:65", "network": {"id": "e206b5b9-da36-4e6f-926f-914dec38fac2", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1053201655-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "39067912d091464192516c08eb576696", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7c80243e-93a7-4a95-bc8d-e9534bacd66e", "external-id": "nsx-vlan-transportzone-306", "segmentation_id": 306, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfbccb731-d9", "ovs_interfaceid": "fbccb731-d98d-40ad-a144-4b4f1e48a1ce", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 911.383377] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-89fee262-eea7-4ed4-83cb-eb9998cc3059 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 6f91b053-772a-4497-b29d-349b960c55eb] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:06:3e:65', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '7c80243e-93a7-4a95-bc8d-e9534bacd66e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'fbccb731-d98d-40ad-a144-4b4f1e48a1ce', 'vif_model': 'vmxnet3'}] {{(pid=65726) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 911.395970] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-89fee262-eea7-4ed4-83cb-eb9998cc3059 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 911.403577] env[65726]: DEBUG nova.network.neutron [None req-17b5a1c7-4bc8-41d1-988f-c933bb6421ad tempest-ServersTestManualDisk-1753452859 tempest-ServersTestManualDisk-1753452859-project-member] [instance: 3088ee5e-0d27-4058-a94b-2e04c3b52add] Successfully updated port: 2ffc16a0-351c-4f5b-8a4a-993ae7adb4f3 {{(pid=65726) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 911.405457] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6f91b053-772a-4497-b29d-349b960c55eb] Creating VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 911.407355] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-96ac5edb-a235-4e3c-abc6-da21a6f87ecb {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.443640] env[65726]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 911.443640] env[65726]: value = "task-5116201" [ 911.443640] env[65726]: _type = "Task" [ 911.443640] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.455528] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116201, 'name': CreateVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.459470] env[65726]: DEBUG nova.network.neutron [-] [instance: 1cfb1bba-1e21-47eb-b50f-a86575846a65] Instance cache missing network info. {{(pid=65726) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 911.460255] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 911.460549] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 911.500452] env[65726]: DEBUG oslo_vmware.api [None req-e5e88d5e-aa68-4ff5-9280-b2cfe1d14b27 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': task-5116199, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.210684} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.502125] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-e5e88d5e-aa68-4ff5-9280-b2cfe1d14b27 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: 5a252ef1-93c9-4bff-842b-b64df2bc5d75] Extended root virtual disk {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 911.503102] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26bf3c28-5b08-492c-a1a2-272deee1ceb2 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.506656] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae569ecd-359a-4b0e-b638-b57080a0678e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.535964] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-e5e88d5e-aa68-4ff5-9280-b2cfe1d14b27 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: 5a252ef1-93c9-4bff-842b-b64df2bc5d75] Reconfiguring VM instance instance-00000044 to attach disk [datastore1] 5a252ef1-93c9-4bff-842b-b64df2bc5d75/5a252ef1-93c9-4bff-842b-b64df2bc5d75.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 911.537360] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-674ae300-ded1-48f1-a2ea-a92b0cfdc7ec {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.541035] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-66462860-32d5-4f09-9cfb-030940abf01c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.594317] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc4badef-0cbb-4518-95cb-f62336c6689c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.599213] env[65726]: DEBUG oslo_vmware.api [None req-e5e88d5e-aa68-4ff5-9280-b2cfe1d14b27 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Waiting for the task: (returnval){ [ 911.599213] env[65726]: value = "task-5116202" [ 911.599213] env[65726]: _type = "Task" [ 911.599213] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.607822] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9777c2b9-44de-476e-9b65-22d2b866040f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.618769] env[65726]: DEBUG oslo_vmware.api [None req-e5e88d5e-aa68-4ff5-9280-b2cfe1d14b27 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': task-5116202, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.636424] env[65726]: DEBUG nova.compute.provider_tree [None req-2cc737a8-ad7d-449c-b270-9db9e18d971d tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 911.651459] env[65726]: DEBUG oslo_vmware.api [None req-7acc7b96-c5e1-4976-bfde-d3cb706afaa3 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Task: {'id': task-5116200, 'name': PowerOffVM_Task, 'duration_secs': 0.347756} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.652924] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-7acc7b96-c5e1-4976-bfde-d3cb706afaa3 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] [instance: 76249623-6f83-46a3-b8c5-c001111aa698] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 911.653232] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-7acc7b96-c5e1-4976-bfde-d3cb706afaa3 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] [instance: 76249623-6f83-46a3-b8c5-c001111aa698] Unregistering the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 911.653612] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-927a4f8b-1f57-4f1f-b317-d2cd373326b1 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.707501] env[65726]: INFO nova.compute.manager [None req-e4fa020c-c0fd-49d8-8a8a-f7bebcb49337 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] [instance: 815ee20b-cb07-4514-bd43-1b74972508a5] instance snapshotting [ 911.710983] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b270fcb1-acb3-41c9-a843-5a03dda571f6 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.734496] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc06f31c-3dc0-48ac-9193-a7d04355dc0a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.741210] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-7acc7b96-c5e1-4976-bfde-d3cb706afaa3 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] [instance: 76249623-6f83-46a3-b8c5-c001111aa698] Unregistered the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 911.741505] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-7acc7b96-c5e1-4976-bfde-d3cb706afaa3 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] [instance: 76249623-6f83-46a3-b8c5-c001111aa698] Deleting contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 911.741797] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-7acc7b96-c5e1-4976-bfde-d3cb706afaa3 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Deleting the datastore file [datastore1] 76249623-6f83-46a3-b8c5-c001111aa698 {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 911.742825] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-13ea766b-8f24-4d37-b5fb-eaf25f9e0771 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.763404] env[65726]: DEBUG oslo_vmware.api [None req-7acc7b96-c5e1-4976-bfde-d3cb706afaa3 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Waiting for the task: (returnval){ [ 911.763404] env[65726]: value = "task-5116204" [ 911.763404] env[65726]: _type = "Task" [ 911.763404] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.779612] env[65726]: DEBUG oslo_vmware.api [None req-7acc7b96-c5e1-4976-bfde-d3cb706afaa3 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Task: {'id': task-5116204, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.908076] env[65726]: DEBUG oslo_concurrency.lockutils [None req-17b5a1c7-4bc8-41d1-988f-c933bb6421ad tempest-ServersTestManualDisk-1753452859 tempest-ServersTestManualDisk-1753452859-project-member] Acquiring lock "refresh_cache-3088ee5e-0d27-4058-a94b-2e04c3b52add" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 911.908361] env[65726]: DEBUG oslo_concurrency.lockutils [None req-17b5a1c7-4bc8-41d1-988f-c933bb6421ad tempest-ServersTestManualDisk-1753452859 tempest-ServersTestManualDisk-1753452859-project-member] Acquired lock "refresh_cache-3088ee5e-0d27-4058-a94b-2e04c3b52add" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 911.908632] env[65726]: DEBUG nova.network.neutron [None req-17b5a1c7-4bc8-41d1-988f-c933bb6421ad tempest-ServersTestManualDisk-1753452859 tempest-ServersTestManualDisk-1753452859-project-member] [instance: 3088ee5e-0d27-4058-a94b-2e04c3b52add] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 911.959427] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116201, 'name': CreateVM_Task} progress is 99%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.970492] env[65726]: DEBUG nova.network.neutron [-] [instance: 1cfb1bba-1e21-47eb-b50f-a86575846a65] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 912.112467] env[65726]: DEBUG oslo_vmware.api [None req-e5e88d5e-aa68-4ff5-9280-b2cfe1d14b27 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': task-5116202, 'name': ReconfigVM_Task, 'duration_secs': 0.493553} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.112791] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-e5e88d5e-aa68-4ff5-9280-b2cfe1d14b27 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: 5a252ef1-93c9-4bff-842b-b64df2bc5d75] Reconfigured VM instance instance-00000044 to attach disk [datastore1] 5a252ef1-93c9-4bff-842b-b64df2bc5d75/5a252ef1-93c9-4bff-842b-b64df2bc5d75.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 912.113639] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-edaf977d-595c-4dbf-9af0-ba13f1935a17 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.126736] env[65726]: DEBUG oslo_vmware.api [None req-e5e88d5e-aa68-4ff5-9280-b2cfe1d14b27 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Waiting for the task: (returnval){ [ 912.126736] env[65726]: value = "task-5116205" [ 912.126736] env[65726]: _type = "Task" [ 912.126736] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.127650] env[65726]: DEBUG nova.compute.manager [req-3ab9b4ea-1548-4202-a06d-9d0833802ee2 req-79e605b8-473a-412e-8aff-b2bc04c493fc service nova] [instance: 4ea2b70b-0cfe-4fbf-817a-baa028b8372a] Received event network-changed-a97348ff-93e0-442b-8a41-b56fa04e8cb6 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 912.127773] env[65726]: DEBUG nova.compute.manager [req-3ab9b4ea-1548-4202-a06d-9d0833802ee2 req-79e605b8-473a-412e-8aff-b2bc04c493fc service nova] [instance: 4ea2b70b-0cfe-4fbf-817a-baa028b8372a] Refreshing instance network info cache due to event network-changed-a97348ff-93e0-442b-8a41-b56fa04e8cb6. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 912.128173] env[65726]: DEBUG oslo_concurrency.lockutils [req-3ab9b4ea-1548-4202-a06d-9d0833802ee2 req-79e605b8-473a-412e-8aff-b2bc04c493fc service nova] Acquiring lock "refresh_cache-4ea2b70b-0cfe-4fbf-817a-baa028b8372a" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 912.128354] env[65726]: DEBUG oslo_concurrency.lockutils [req-3ab9b4ea-1548-4202-a06d-9d0833802ee2 req-79e605b8-473a-412e-8aff-b2bc04c493fc service nova] Acquired lock "refresh_cache-4ea2b70b-0cfe-4fbf-817a-baa028b8372a" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 912.128966] env[65726]: DEBUG nova.network.neutron [req-3ab9b4ea-1548-4202-a06d-9d0833802ee2 req-79e605b8-473a-412e-8aff-b2bc04c493fc service nova] [instance: 4ea2b70b-0cfe-4fbf-817a-baa028b8372a] Refreshing network info cache for port a97348ff-93e0-442b-8a41-b56fa04e8cb6 {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 912.148289] env[65726]: DEBUG nova.scheduler.client.report [None req-2cc737a8-ad7d-449c-b270-9db9e18d971d tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 912.152759] env[65726]: DEBUG oslo_vmware.api [None req-e5e88d5e-aa68-4ff5-9280-b2cfe1d14b27 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': task-5116205, 'name': Rename_Task} progress is 10%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.157436] env[65726]: DEBUG nova.compute.manager [req-3d3d8609-65c4-497b-bab5-096199dfa99c req-b299532c-4637-401d-b584-8cb96169958e service nova] [instance: 6f91b053-772a-4497-b29d-349b960c55eb] Received event network-changed-fbccb731-d98d-40ad-a144-4b4f1e48a1ce {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 912.157436] env[65726]: DEBUG nova.compute.manager [req-3d3d8609-65c4-497b-bab5-096199dfa99c req-b299532c-4637-401d-b584-8cb96169958e service nova] [instance: 6f91b053-772a-4497-b29d-349b960c55eb] Refreshing instance network info cache due to event network-changed-fbccb731-d98d-40ad-a144-4b4f1e48a1ce. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 912.157833] env[65726]: DEBUG oslo_concurrency.lockutils [req-3d3d8609-65c4-497b-bab5-096199dfa99c req-b299532c-4637-401d-b584-8cb96169958e service nova] Acquiring lock "refresh_cache-6f91b053-772a-4497-b29d-349b960c55eb" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 912.158454] env[65726]: DEBUG oslo_concurrency.lockutils [req-3d3d8609-65c4-497b-bab5-096199dfa99c req-b299532c-4637-401d-b584-8cb96169958e service nova] Acquired lock "refresh_cache-6f91b053-772a-4497-b29d-349b960c55eb" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 912.158454] env[65726]: DEBUG nova.network.neutron [req-3d3d8609-65c4-497b-bab5-096199dfa99c req-b299532c-4637-401d-b584-8cb96169958e service nova] [instance: 6f91b053-772a-4497-b29d-349b960c55eb] Refreshing network info cache for port fbccb731-d98d-40ad-a144-4b4f1e48a1ce {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 912.255700] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-e4fa020c-c0fd-49d8-8a8a-f7bebcb49337 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] [instance: 815ee20b-cb07-4514-bd43-1b74972508a5] Creating Snapshot of the VM instance {{(pid=65726) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 912.255700] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-1ac9aafb-1efd-41a2-acaa-14f1e8d18b1b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.269661] env[65726]: DEBUG oslo_vmware.api [None req-e4fa020c-c0fd-49d8-8a8a-f7bebcb49337 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Waiting for the task: (returnval){ [ 912.269661] env[65726]: value = "task-5116206" [ 912.269661] env[65726]: _type = "Task" [ 912.269661] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.282071] env[65726]: DEBUG oslo_vmware.api [None req-7acc7b96-c5e1-4976-bfde-d3cb706afaa3 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Task: {'id': task-5116204, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.236096} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.283153] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-7acc7b96-c5e1-4976-bfde-d3cb706afaa3 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Deleted the datastore file {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 912.285299] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-7acc7b96-c5e1-4976-bfde-d3cb706afaa3 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] [instance: 76249623-6f83-46a3-b8c5-c001111aa698] Deleted contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 912.286903] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-7acc7b96-c5e1-4976-bfde-d3cb706afaa3 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] [instance: 76249623-6f83-46a3-b8c5-c001111aa698] Instance destroyed {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 912.286903] env[65726]: INFO nova.compute.manager [None req-7acc7b96-c5e1-4976-bfde-d3cb706afaa3 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] [instance: 76249623-6f83-46a3-b8c5-c001111aa698] Took 1.17 seconds to destroy the instance on the hypervisor. [ 912.286903] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-7acc7b96-c5e1-4976-bfde-d3cb706afaa3 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 912.292234] env[65726]: DEBUG nova.compute.manager [-] [instance: 76249623-6f83-46a3-b8c5-c001111aa698] Deallocating network for instance {{(pid=65726) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 912.292234] env[65726]: DEBUG nova.network.neutron [-] [instance: 76249623-6f83-46a3-b8c5-c001111aa698] deallocate_for_instance() {{(pid=65726) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 912.292234] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 912.292234] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 912.302050] env[65726]: DEBUG oslo_vmware.api [None req-e4fa020c-c0fd-49d8-8a8a-f7bebcb49337 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Task: {'id': task-5116206, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.385943] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 912.385943] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 912.413036] env[65726]: WARNING openstack [None req-17b5a1c7-4bc8-41d1-988f-c933bb6421ad tempest-ServersTestManualDisk-1753452859 tempest-ServersTestManualDisk-1753452859-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 912.414651] env[65726]: WARNING openstack [None req-17b5a1c7-4bc8-41d1-988f-c933bb6421ad tempest-ServersTestManualDisk-1753452859 tempest-ServersTestManualDisk-1753452859-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 912.419861] env[65726]: DEBUG nova.network.neutron [None req-17b5a1c7-4bc8-41d1-988f-c933bb6421ad tempest-ServersTestManualDisk-1753452859 tempest-ServersTestManualDisk-1753452859-project-member] [instance: 3088ee5e-0d27-4058-a94b-2e04c3b52add] Instance cache missing network info. {{(pid=65726) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 912.454981] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116201, 'name': CreateVM_Task, 'duration_secs': 0.586185} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.455248] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6f91b053-772a-4497-b29d-349b960c55eb] Created VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 912.456197] env[65726]: WARNING openstack [None req-89fee262-eea7-4ed4-83cb-eb9998cc3059 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 912.456555] env[65726]: WARNING openstack [None req-89fee262-eea7-4ed4-83cb-eb9998cc3059 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 912.464271] env[65726]: DEBUG oslo_concurrency.lockutils [None req-89fee262-eea7-4ed4-83cb-eb9998cc3059 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8ecee2b1-8703-4e91-8a32-dc17932b237e" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 912.464509] env[65726]: DEBUG oslo_concurrency.lockutils [None req-89fee262-eea7-4ed4-83cb-eb9998cc3059 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8ecee2b1-8703-4e91-8a32-dc17932b237e" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 912.464932] env[65726]: DEBUG oslo_concurrency.lockutils [None req-89fee262-eea7-4ed4-83cb-eb9998cc3059 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8ecee2b1-8703-4e91-8a32-dc17932b237e" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 912.465702] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-532199af-ef7f-41e0-95e0-8e9cfa18599b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.472326] env[65726]: DEBUG oslo_vmware.api [None req-89fee262-eea7-4ed4-83cb-eb9998cc3059 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Waiting for the task: (returnval){ [ 912.472326] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5275a855-39ff-2951-9dad-7e51819fffbc" [ 912.472326] env[65726]: _type = "Task" [ 912.472326] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.476093] env[65726]: INFO nova.compute.manager [-] [instance: 1cfb1bba-1e21-47eb-b50f-a86575846a65] Took 1.52 seconds to deallocate network for instance. [ 912.489914] env[65726]: DEBUG oslo_concurrency.lockutils [None req-89fee262-eea7-4ed4-83cb-eb9998cc3059 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8ecee2b1-8703-4e91-8a32-dc17932b237e" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 912.490313] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-89fee262-eea7-4ed4-83cb-eb9998cc3059 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 6f91b053-772a-4497-b29d-349b960c55eb] Processing image 8ecee2b1-8703-4e91-8a32-dc17932b237e {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 912.490497] env[65726]: DEBUG oslo_concurrency.lockutils [None req-89fee262-eea7-4ed4-83cb-eb9998cc3059 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8ecee2b1-8703-4e91-8a32-dc17932b237e/8ecee2b1-8703-4e91-8a32-dc17932b237e.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 912.490574] env[65726]: DEBUG oslo_concurrency.lockutils [None req-89fee262-eea7-4ed4-83cb-eb9998cc3059 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8ecee2b1-8703-4e91-8a32-dc17932b237e/8ecee2b1-8703-4e91-8a32-dc17932b237e.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 912.490747] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-89fee262-eea7-4ed4-83cb-eb9998cc3059 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 912.491080] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0bda4aba-803b-4416-825c-3655bf790a73 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.506857] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-89fee262-eea7-4ed4-83cb-eb9998cc3059 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 912.507490] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-89fee262-eea7-4ed4-83cb-eb9998cc3059 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65726) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 912.508206] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-159777c3-a5ea-42ec-b2ea-30c58a20859b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.515823] env[65726]: DEBUG oslo_vmware.api [None req-89fee262-eea7-4ed4-83cb-eb9998cc3059 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Waiting for the task: (returnval){ [ 912.515823] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52d3a848-7782-cf06-12d0-0f73176ddb25" [ 912.515823] env[65726]: _type = "Task" [ 912.515823] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.528033] env[65726]: DEBUG oslo_vmware.api [None req-89fee262-eea7-4ed4-83cb-eb9998cc3059 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52d3a848-7782-cf06-12d0-0f73176ddb25, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.636518] env[65726]: WARNING openstack [req-3ab9b4ea-1548-4202-a06d-9d0833802ee2 req-79e605b8-473a-412e-8aff-b2bc04c493fc service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 912.636936] env[65726]: WARNING openstack [req-3ab9b4ea-1548-4202-a06d-9d0833802ee2 req-79e605b8-473a-412e-8aff-b2bc04c493fc service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 912.659072] env[65726]: DEBUG oslo_concurrency.lockutils [None req-2cc737a8-ad7d-449c-b270-9db9e18d971d tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.806s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 912.659683] env[65726]: DEBUG nova.compute.manager [None req-2cc737a8-ad7d-449c-b270-9db9e18d971d tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: dcde25df-87a5-47fb-94ce-334f68894e04] Start building networks asynchronously for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 912.666258] env[65726]: DEBUG oslo_vmware.api [None req-e5e88d5e-aa68-4ff5-9280-b2cfe1d14b27 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': task-5116205, 'name': Rename_Task, 'duration_secs': 0.181929} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.666592] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c85b0a62-807d-4793-b1be-86b28d9a6745 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 29.013s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 912.666777] env[65726]: DEBUG nova.objects.instance [None req-c85b0a62-807d-4793-b1be-86b28d9a6745 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] [instance: 2cbee359-a1fc-45c0-bcc5-a2c24ded9c69] Trying to apply a migration context that does not seem to be set for this instance {{(pid=65726) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 912.675552] env[65726]: WARNING openstack [req-3d3d8609-65c4-497b-bab5-096199dfa99c req-b299532c-4637-401d-b584-8cb96169958e service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 912.676454] env[65726]: WARNING openstack [req-3d3d8609-65c4-497b-bab5-096199dfa99c req-b299532c-4637-401d-b584-8cb96169958e service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 912.684044] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-e5e88d5e-aa68-4ff5-9280-b2cfe1d14b27 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: 5a252ef1-93c9-4bff-842b-b64df2bc5d75] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 912.684937] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-854d3a70-e047-494b-abdd-4525e064c7bf {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.693949] env[65726]: DEBUG oslo_vmware.api [None req-e5e88d5e-aa68-4ff5-9280-b2cfe1d14b27 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Waiting for the task: (returnval){ [ 912.693949] env[65726]: value = "task-5116207" [ 912.693949] env[65726]: _type = "Task" [ 912.693949] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.704844] env[65726]: DEBUG oslo_vmware.api [None req-e5e88d5e-aa68-4ff5-9280-b2cfe1d14b27 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': task-5116207, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.782254] env[65726]: DEBUG oslo_vmware.api [None req-e4fa020c-c0fd-49d8-8a8a-f7bebcb49337 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Task: {'id': task-5116206, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.863204] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c4dc675f-0696-4840-ac89-d68c37ca6fa4 tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Acquiring lock "4ea2b70b-0cfe-4fbf-817a-baa028b8372a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 912.863503] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c4dc675f-0696-4840-ac89-d68c37ca6fa4 tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Lock "4ea2b70b-0cfe-4fbf-817a-baa028b8372a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 912.863782] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c4dc675f-0696-4840-ac89-d68c37ca6fa4 tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Acquiring lock "4ea2b70b-0cfe-4fbf-817a-baa028b8372a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 912.863892] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c4dc675f-0696-4840-ac89-d68c37ca6fa4 tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Lock "4ea2b70b-0cfe-4fbf-817a-baa028b8372a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 912.864095] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c4dc675f-0696-4840-ac89-d68c37ca6fa4 tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Lock "4ea2b70b-0cfe-4fbf-817a-baa028b8372a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 912.866655] env[65726]: INFO nova.compute.manager [None req-c4dc675f-0696-4840-ac89-d68c37ca6fa4 tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] [instance: 4ea2b70b-0cfe-4fbf-817a-baa028b8372a] Terminating instance [ 912.986146] env[65726]: DEBUG oslo_concurrency.lockutils [None req-fa9b16af-8474-4ab3-a554-b5fbaf4d16fd tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 913.012219] env[65726]: DEBUG oslo_concurrency.lockutils [None req-2eeb1a84-0a41-40be-a094-58835ef7befd tempest-ServerAddressesTestJSON-569100133 tempest-ServerAddressesTestJSON-569100133-project-member] Acquiring lock "70375746-06dd-4710-9136-95e9e5759a0e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 913.012638] env[65726]: DEBUG oslo_concurrency.lockutils [None req-2eeb1a84-0a41-40be-a094-58835ef7befd tempest-ServerAddressesTestJSON-569100133 tempest-ServerAddressesTestJSON-569100133-project-member] Lock "70375746-06dd-4710-9136-95e9e5759a0e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 913.012930] env[65726]: DEBUG oslo_concurrency.lockutils [None req-2eeb1a84-0a41-40be-a094-58835ef7befd tempest-ServerAddressesTestJSON-569100133 tempest-ServerAddressesTestJSON-569100133-project-member] Acquiring lock "70375746-06dd-4710-9136-95e9e5759a0e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 913.013645] env[65726]: DEBUG oslo_concurrency.lockutils [None req-2eeb1a84-0a41-40be-a094-58835ef7befd tempest-ServerAddressesTestJSON-569100133 tempest-ServerAddressesTestJSON-569100133-project-member] Lock "70375746-06dd-4710-9136-95e9e5759a0e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 913.013645] env[65726]: DEBUG oslo_concurrency.lockutils [None req-2eeb1a84-0a41-40be-a094-58835ef7befd tempest-ServerAddressesTestJSON-569100133 tempest-ServerAddressesTestJSON-569100133-project-member] Lock "70375746-06dd-4710-9136-95e9e5759a0e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 913.021141] env[65726]: INFO nova.compute.manager [None req-2eeb1a84-0a41-40be-a094-58835ef7befd tempest-ServerAddressesTestJSON-569100133 tempest-ServerAddressesTestJSON-569100133-project-member] [instance: 70375746-06dd-4710-9136-95e9e5759a0e] Terminating instance [ 913.025527] env[65726]: WARNING openstack [None req-17b5a1c7-4bc8-41d1-988f-c933bb6421ad tempest-ServersTestManualDisk-1753452859 tempest-ServersTestManualDisk-1753452859-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 913.025943] env[65726]: WARNING openstack [None req-17b5a1c7-4bc8-41d1-988f-c933bb6421ad tempest-ServersTestManualDisk-1753452859 tempest-ServersTestManualDisk-1753452859-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 913.040717] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-89fee262-eea7-4ed4-83cb-eb9998cc3059 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 6f91b053-772a-4497-b29d-349b960c55eb] Preparing fetch location {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 913.041165] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-89fee262-eea7-4ed4-83cb-eb9998cc3059 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 6f91b053-772a-4497-b29d-349b960c55eb] Fetch image to [datastore1] OSTACK_IMG_1f8b9f24-a5fc-420d-a36b-40a6efa61bbf/OSTACK_IMG_1f8b9f24-a5fc-420d-a36b-40a6efa61bbf.vmdk {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 913.041362] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-89fee262-eea7-4ed4-83cb-eb9998cc3059 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 6f91b053-772a-4497-b29d-349b960c55eb] Downloading stream optimized image 8ecee2b1-8703-4e91-8a32-dc17932b237e to [datastore1] OSTACK_IMG_1f8b9f24-a5fc-420d-a36b-40a6efa61bbf/OSTACK_IMG_1f8b9f24-a5fc-420d-a36b-40a6efa61bbf.vmdk on the data store datastore1 as vApp {{(pid=65726) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 913.041479] env[65726]: DEBUG nova.virt.vmwareapi.images [None req-89fee262-eea7-4ed4-83cb-eb9998cc3059 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 6f91b053-772a-4497-b29d-349b960c55eb] Downloading image file data 8ecee2b1-8703-4e91-8a32-dc17932b237e to the ESX as VM named 'OSTACK_IMG_1f8b9f24-a5fc-420d-a36b-40a6efa61bbf' {{(pid=65726) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 913.118723] env[65726]: DEBUG nova.network.neutron [-] [instance: 76249623-6f83-46a3-b8c5-c001111aa698] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 913.156301] env[65726]: DEBUG oslo_vmware.rw_handles [None req-89fee262-eea7-4ed4-83cb-eb9998cc3059 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 913.156301] env[65726]: value = "resgroup-9" [ 913.156301] env[65726]: _type = "ResourcePool" [ 913.156301] env[65726]: }. {{(pid=65726) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 913.156445] env[65726]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-7eb2dada-4724-4705-8a2a-7a92ab870ecf {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.183521] env[65726]: DEBUG nova.compute.utils [None req-2cc737a8-ad7d-449c-b270-9db9e18d971d tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Using /dev/sd instead of None {{(pid=65726) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 913.188369] env[65726]: DEBUG nova.compute.manager [None req-2cc737a8-ad7d-449c-b270-9db9e18d971d tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: dcde25df-87a5-47fb-94ce-334f68894e04] Allocating IP information in the background. {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 913.188589] env[65726]: DEBUG nova.network.neutron [None req-2cc737a8-ad7d-449c-b270-9db9e18d971d tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: dcde25df-87a5-47fb-94ce-334f68894e04] allocate_for_instance() {{(pid=65726) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 913.188944] env[65726]: WARNING neutronclient.v2_0.client [None req-2cc737a8-ad7d-449c-b270-9db9e18d971d tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 913.189505] env[65726]: WARNING neutronclient.v2_0.client [None req-2cc737a8-ad7d-449c-b270-9db9e18d971d tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 913.189968] env[65726]: WARNING openstack [None req-2cc737a8-ad7d-449c-b270-9db9e18d971d tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 913.190316] env[65726]: WARNING openstack [None req-2cc737a8-ad7d-449c-b270-9db9e18d971d tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 913.214044] env[65726]: DEBUG oslo_vmware.rw_handles [None req-89fee262-eea7-4ed4-83cb-eb9998cc3059 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Lease: (returnval){ [ 913.214044] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52c23fd6-d39a-65ad-1e34-5d6f64a43f1e" [ 913.214044] env[65726]: _type = "HttpNfcLease" [ 913.214044] env[65726]: } obtained for vApp import into resource pool (val){ [ 913.214044] env[65726]: value = "resgroup-9" [ 913.214044] env[65726]: _type = "ResourcePool" [ 913.214044] env[65726]: }. {{(pid=65726) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 913.214044] env[65726]: DEBUG oslo_vmware.api [None req-89fee262-eea7-4ed4-83cb-eb9998cc3059 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Waiting for the lease: (returnval){ [ 913.214044] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52c23fd6-d39a-65ad-1e34-5d6f64a43f1e" [ 913.214044] env[65726]: _type = "HttpNfcLease" [ 913.214044] env[65726]: } to be ready. {{(pid=65726) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 913.222105] env[65726]: DEBUG oslo_vmware.api [None req-e5e88d5e-aa68-4ff5-9280-b2cfe1d14b27 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': task-5116207, 'name': PowerOnVM_Task} progress is 96%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.227330] env[65726]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 913.227330] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52c23fd6-d39a-65ad-1e34-5d6f64a43f1e" [ 913.227330] env[65726]: _type = "HttpNfcLease" [ 913.227330] env[65726]: } is initializing. {{(pid=65726) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 913.282716] env[65726]: DEBUG oslo_vmware.api [None req-e4fa020c-c0fd-49d8-8a8a-f7bebcb49337 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Task: {'id': task-5116206, 'name': CreateSnapshot_Task, 'duration_secs': 0.606285} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.283826] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-e4fa020c-c0fd-49d8-8a8a-f7bebcb49337 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] [instance: 815ee20b-cb07-4514-bd43-1b74972508a5] Created Snapshot of the VM instance {{(pid=65726) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 913.285189] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f251874-9450-4f57-92de-b2dc1f6f3510 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.326051] env[65726]: WARNING openstack [req-3ab9b4ea-1548-4202-a06d-9d0833802ee2 req-79e605b8-473a-412e-8aff-b2bc04c493fc service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 913.326535] env[65726]: WARNING openstack [req-3ab9b4ea-1548-4202-a06d-9d0833802ee2 req-79e605b8-473a-412e-8aff-b2bc04c493fc service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 913.343694] env[65726]: WARNING openstack [None req-17b5a1c7-4bc8-41d1-988f-c933bb6421ad tempest-ServersTestManualDisk-1753452859 tempest-ServersTestManualDisk-1753452859-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 913.344113] env[65726]: WARNING openstack [None req-17b5a1c7-4bc8-41d1-988f-c933bb6421ad tempest-ServersTestManualDisk-1753452859 tempest-ServersTestManualDisk-1753452859-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 913.371342] env[65726]: DEBUG nova.compute.manager [None req-c4dc675f-0696-4840-ac89-d68c37ca6fa4 tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] [instance: 4ea2b70b-0cfe-4fbf-817a-baa028b8372a] Start destroying the instance on the hypervisor. {{(pid=65726) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 913.371678] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-c4dc675f-0696-4840-ac89-d68c37ca6fa4 tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] [instance: 4ea2b70b-0cfe-4fbf-817a-baa028b8372a] Destroying instance {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 913.372594] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa931f63-65cd-4bc2-80a6-cfd897fb1da2 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.383244] env[65726]: DEBUG nova.policy [None req-2cc737a8-ad7d-449c-b270-9db9e18d971d tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9b65b9194933469c908c7cc04478444b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd46ccb00794f458b85da4a93879139ee', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65726) authorize /opt/stack/nova/nova/policy.py:192}} [ 913.388779] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-c4dc675f-0696-4840-ac89-d68c37ca6fa4 tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] [instance: 4ea2b70b-0cfe-4fbf-817a-baa028b8372a] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 913.388779] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-dab37add-5534-47af-a0b5-9ee048b5f86b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.396550] env[65726]: DEBUG oslo_vmware.api [None req-c4dc675f-0696-4840-ac89-d68c37ca6fa4 tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Waiting for the task: (returnval){ [ 913.396550] env[65726]: value = "task-5116209" [ 913.396550] env[65726]: _type = "Task" [ 913.396550] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.402309] env[65726]: WARNING openstack [req-3d3d8609-65c4-497b-bab5-096199dfa99c req-b299532c-4637-401d-b584-8cb96169958e service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 913.402837] env[65726]: WARNING openstack [req-3d3d8609-65c4-497b-bab5-096199dfa99c req-b299532c-4637-401d-b584-8cb96169958e service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 913.416924] env[65726]: DEBUG oslo_vmware.api [None req-c4dc675f-0696-4840-ac89-d68c37ca6fa4 tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Task: {'id': task-5116209, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.536875] env[65726]: DEBUG nova.compute.manager [None req-2eeb1a84-0a41-40be-a094-58835ef7befd tempest-ServerAddressesTestJSON-569100133 tempest-ServerAddressesTestJSON-569100133-project-member] [instance: 70375746-06dd-4710-9136-95e9e5759a0e] Start destroying the instance on the hypervisor. {{(pid=65726) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 913.537286] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-2eeb1a84-0a41-40be-a094-58835ef7befd tempest-ServerAddressesTestJSON-569100133 tempest-ServerAddressesTestJSON-569100133-project-member] [instance: 70375746-06dd-4710-9136-95e9e5759a0e] Destroying instance {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 913.538316] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98a9a317-5d9f-41ce-992c-0af0276733c4 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.557059] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-2eeb1a84-0a41-40be-a094-58835ef7befd tempest-ServerAddressesTestJSON-569100133 tempest-ServerAddressesTestJSON-569100133-project-member] [instance: 70375746-06dd-4710-9136-95e9e5759a0e] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 913.557059] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-dc7af236-85bd-499b-adbb-3352ab7ad37f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.565476] env[65726]: DEBUG oslo_vmware.api [None req-2eeb1a84-0a41-40be-a094-58835ef7befd tempest-ServerAddressesTestJSON-569100133 tempest-ServerAddressesTestJSON-569100133-project-member] Waiting for the task: (returnval){ [ 913.565476] env[65726]: value = "task-5116210" [ 913.565476] env[65726]: _type = "Task" [ 913.565476] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.583464] env[65726]: DEBUG oslo_vmware.api [None req-2eeb1a84-0a41-40be-a094-58835ef7befd tempest-ServerAddressesTestJSON-569100133 tempest-ServerAddressesTestJSON-569100133-project-member] Task: {'id': task-5116210, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.624272] env[65726]: INFO nova.compute.manager [-] [instance: 76249623-6f83-46a3-b8c5-c001111aa698] Took 1.33 seconds to deallocate network for instance. [ 913.649450] env[65726]: WARNING openstack [req-3ab9b4ea-1548-4202-a06d-9d0833802ee2 req-79e605b8-473a-412e-8aff-b2bc04c493fc service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 913.649450] env[65726]: WARNING openstack [req-3ab9b4ea-1548-4202-a06d-9d0833802ee2 req-79e605b8-473a-412e-8aff-b2bc04c493fc service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 913.689214] env[65726]: DEBUG nova.compute.manager [None req-2cc737a8-ad7d-449c-b270-9db9e18d971d tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: dcde25df-87a5-47fb-94ce-334f68894e04] Start building block device mappings for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 913.699212] env[65726]: DEBUG nova.network.neutron [None req-17b5a1c7-4bc8-41d1-988f-c933bb6421ad tempest-ServersTestManualDisk-1753452859 tempest-ServersTestManualDisk-1753452859-project-member] [instance: 3088ee5e-0d27-4058-a94b-2e04c3b52add] Updating instance_info_cache with network_info: [{"id": "2ffc16a0-351c-4f5b-8a4a-993ae7adb4f3", "address": "fa:16:3e:c8:5a:78", "network": {"id": "6e131d80-bd85-46b6-9b8d-c4d34af67f1d", "bridge": "br-int", "label": "tempest-ServersTestManualDisk-1706238551-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "308ec5a4cf844d2782db77b3edcb4127", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fd0eb882-ab95-4373-aa20-ee565a9919e3", "external-id": "nsx-vlan-transportzone-510", "segmentation_id": 510, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2ffc16a0-35", "ovs_interfaceid": "2ffc16a0-351c-4f5b-8a4a-993ae7adb4f3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 913.704089] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c85b0a62-807d-4793-b1be-86b28d9a6745 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.037s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 913.706709] env[65726]: DEBUG oslo_concurrency.lockutils [None req-801b9b62-b5bd-4b5f-80fa-32388dd4cb88 tempest-ServerTagsTestJSON-59142761 tempest-ServerTagsTestJSON-59142761-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 28.037s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 913.711971] env[65726]: INFO nova.compute.claims [None req-801b9b62-b5bd-4b5f-80fa-32388dd4cb88 tempest-ServerTagsTestJSON-59142761 tempest-ServerTagsTestJSON-59142761-project-member] [instance: 9029549c-1914-4bae-91e2-8812b79051ec] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 913.734490] env[65726]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 913.734490] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52c23fd6-d39a-65ad-1e34-5d6f64a43f1e" [ 913.734490] env[65726]: _type = "HttpNfcLease" [ 913.734490] env[65726]: } is initializing. {{(pid=65726) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 913.740656] env[65726]: DEBUG oslo_vmware.api [None req-e5e88d5e-aa68-4ff5-9280-b2cfe1d14b27 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': task-5116207, 'name': PowerOnVM_Task, 'duration_secs': 0.983568} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.740656] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-e5e88d5e-aa68-4ff5-9280-b2cfe1d14b27 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: 5a252ef1-93c9-4bff-842b-b64df2bc5d75] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 913.740779] env[65726]: INFO nova.compute.manager [None req-e5e88d5e-aa68-4ff5-9280-b2cfe1d14b27 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: 5a252ef1-93c9-4bff-842b-b64df2bc5d75] Took 8.06 seconds to spawn the instance on the hypervisor. [ 913.741651] env[65726]: DEBUG nova.compute.manager [None req-e5e88d5e-aa68-4ff5-9280-b2cfe1d14b27 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: 5a252ef1-93c9-4bff-842b-b64df2bc5d75] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 913.742145] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c0188fa-6e83-4086-9e08-861cfcbc212c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.808502] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-e4fa020c-c0fd-49d8-8a8a-f7bebcb49337 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] [instance: 815ee20b-cb07-4514-bd43-1b74972508a5] Creating linked-clone VM from snapshot {{(pid=65726) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 913.809643] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-a9868ee8-40cf-433a-8299-c4054c8d2148 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.821041] env[65726]: DEBUG oslo_vmware.api [None req-e4fa020c-c0fd-49d8-8a8a-f7bebcb49337 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Waiting for the task: (returnval){ [ 913.821041] env[65726]: value = "task-5116211" [ 913.821041] env[65726]: _type = "Task" [ 913.821041] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.843530] env[65726]: DEBUG oslo_vmware.api [None req-e4fa020c-c0fd-49d8-8a8a-f7bebcb49337 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Task: {'id': task-5116211, 'name': CloneVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.912013] env[65726]: DEBUG oslo_vmware.api [None req-c4dc675f-0696-4840-ac89-d68c37ca6fa4 tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Task: {'id': task-5116209, 'name': PowerOffVM_Task, 'duration_secs': 0.495712} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.912751] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-c4dc675f-0696-4840-ac89-d68c37ca6fa4 tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] [instance: 4ea2b70b-0cfe-4fbf-817a-baa028b8372a] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 913.913131] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-c4dc675f-0696-4840-ac89-d68c37ca6fa4 tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] [instance: 4ea2b70b-0cfe-4fbf-817a-baa028b8372a] Unregistering the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 913.917987] env[65726]: WARNING openstack [req-3d3d8609-65c4-497b-bab5-096199dfa99c req-b299532c-4637-401d-b584-8cb96169958e service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 913.918556] env[65726]: WARNING openstack [req-3d3d8609-65c4-497b-bab5-096199dfa99c req-b299532c-4637-401d-b584-8cb96169958e service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 913.929200] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9f664ae1-ee08-48b5-9d50-a5e7c627261a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.015216] env[65726]: DEBUG nova.network.neutron [None req-2cc737a8-ad7d-449c-b270-9db9e18d971d tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: dcde25df-87a5-47fb-94ce-334f68894e04] Successfully created port: bf51c96c-63c5-46de-a40f-58e6769c5fcf {{(pid=65726) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 914.047876] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-c4dc675f-0696-4840-ac89-d68c37ca6fa4 tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] [instance: 4ea2b70b-0cfe-4fbf-817a-baa028b8372a] Unregistered the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 914.048210] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-c4dc675f-0696-4840-ac89-d68c37ca6fa4 tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] [instance: 4ea2b70b-0cfe-4fbf-817a-baa028b8372a] Deleting contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 914.049978] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-c4dc675f-0696-4840-ac89-d68c37ca6fa4 tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Deleting the datastore file [datastore1] 4ea2b70b-0cfe-4fbf-817a-baa028b8372a {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 914.051033] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8085055c-a942-4407-8cfe-6238a0d1046f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.065066] env[65726]: DEBUG oslo_vmware.api [None req-c4dc675f-0696-4840-ac89-d68c37ca6fa4 tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Waiting for the task: (returnval){ [ 914.065066] env[65726]: value = "task-5116213" [ 914.065066] env[65726]: _type = "Task" [ 914.065066] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.087036] env[65726]: DEBUG oslo_vmware.api [None req-c4dc675f-0696-4840-ac89-d68c37ca6fa4 tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Task: {'id': task-5116213, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.087279] env[65726]: DEBUG oslo_vmware.api [None req-2eeb1a84-0a41-40be-a094-58835ef7befd tempest-ServerAddressesTestJSON-569100133 tempest-ServerAddressesTestJSON-569100133-project-member] Task: {'id': task-5116210, 'name': PowerOffVM_Task, 'duration_secs': 0.335305} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.089351] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-2eeb1a84-0a41-40be-a094-58835ef7befd tempest-ServerAddressesTestJSON-569100133 tempest-ServerAddressesTestJSON-569100133-project-member] [instance: 70375746-06dd-4710-9136-95e9e5759a0e] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 914.089351] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-2eeb1a84-0a41-40be-a094-58835ef7befd tempest-ServerAddressesTestJSON-569100133 tempest-ServerAddressesTestJSON-569100133-project-member] [instance: 70375746-06dd-4710-9136-95e9e5759a0e] Unregistering the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 914.089351] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2e3fb148-88d9-4685-8311-edbbca17a9f6 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.099430] env[65726]: DEBUG nova.network.neutron [req-3ab9b4ea-1548-4202-a06d-9d0833802ee2 req-79e605b8-473a-412e-8aff-b2bc04c493fc service nova] [instance: 4ea2b70b-0cfe-4fbf-817a-baa028b8372a] Updated VIF entry in instance network info cache for port a97348ff-93e0-442b-8a41-b56fa04e8cb6. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 914.099802] env[65726]: DEBUG nova.network.neutron [req-3ab9b4ea-1548-4202-a06d-9d0833802ee2 req-79e605b8-473a-412e-8aff-b2bc04c493fc service nova] [instance: 4ea2b70b-0cfe-4fbf-817a-baa028b8372a] Updating instance_info_cache with network_info: [{"id": "a97348ff-93e0-442b-8a41-b56fa04e8cb6", "address": "fa:16:3e:f5:96:db", "network": {"id": "dd4bbaef-0a1e-4e00-b65a-f19e7a837cc1", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-255385230-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "55f93cdf51e84cf683879668c2acecf4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "04ccbc7a-cf8d-4ea2-8411-291a1e27df7b", "external-id": "nsx-vlan-transportzone-998", "segmentation_id": 998, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa97348ff-93", "ovs_interfaceid": "a97348ff-93e0-442b-8a41-b56fa04e8cb6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 914.137311] env[65726]: DEBUG oslo_concurrency.lockutils [None req-7acc7b96-c5e1-4976-bfde-d3cb706afaa3 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 914.169842] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-2eeb1a84-0a41-40be-a094-58835ef7befd tempest-ServerAddressesTestJSON-569100133 tempest-ServerAddressesTestJSON-569100133-project-member] [instance: 70375746-06dd-4710-9136-95e9e5759a0e] Unregistered the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 914.170400] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-2eeb1a84-0a41-40be-a094-58835ef7befd tempest-ServerAddressesTestJSON-569100133 tempest-ServerAddressesTestJSON-569100133-project-member] [instance: 70375746-06dd-4710-9136-95e9e5759a0e] Deleting contents of the VM from datastore datastore2 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 914.170472] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-2eeb1a84-0a41-40be-a094-58835ef7befd tempest-ServerAddressesTestJSON-569100133 tempest-ServerAddressesTestJSON-569100133-project-member] Deleting the datastore file [datastore2] 70375746-06dd-4710-9136-95e9e5759a0e {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 914.170952] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f2e3234d-8188-4c68-966d-edbe5ae0504f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.183895] env[65726]: DEBUG oslo_vmware.api [None req-2eeb1a84-0a41-40be-a094-58835ef7befd tempest-ServerAddressesTestJSON-569100133 tempest-ServerAddressesTestJSON-569100133-project-member] Waiting for the task: (returnval){ [ 914.183895] env[65726]: value = "task-5116215" [ 914.183895] env[65726]: _type = "Task" [ 914.183895] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.199752] env[65726]: DEBUG oslo_vmware.api [None req-2eeb1a84-0a41-40be-a094-58835ef7befd tempest-ServerAddressesTestJSON-569100133 tempest-ServerAddressesTestJSON-569100133-project-member] Task: {'id': task-5116215, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.202843] env[65726]: DEBUG oslo_concurrency.lockutils [None req-17b5a1c7-4bc8-41d1-988f-c933bb6421ad tempest-ServersTestManualDisk-1753452859 tempest-ServersTestManualDisk-1753452859-project-member] Releasing lock "refresh_cache-3088ee5e-0d27-4058-a94b-2e04c3b52add" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 914.203373] env[65726]: DEBUG nova.compute.manager [None req-17b5a1c7-4bc8-41d1-988f-c933bb6421ad tempest-ServersTestManualDisk-1753452859 tempest-ServersTestManualDisk-1753452859-project-member] [instance: 3088ee5e-0d27-4058-a94b-2e04c3b52add] Instance network_info: |[{"id": "2ffc16a0-351c-4f5b-8a4a-993ae7adb4f3", "address": "fa:16:3e:c8:5a:78", "network": {"id": "6e131d80-bd85-46b6-9b8d-c4d34af67f1d", "bridge": "br-int", "label": "tempest-ServersTestManualDisk-1706238551-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "308ec5a4cf844d2782db77b3edcb4127", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fd0eb882-ab95-4373-aa20-ee565a9919e3", "external-id": "nsx-vlan-transportzone-510", "segmentation_id": 510, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2ffc16a0-35", "ovs_interfaceid": "2ffc16a0-351c-4f5b-8a4a-993ae7adb4f3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 914.203760] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-17b5a1c7-4bc8-41d1-988f-c933bb6421ad tempest-ServersTestManualDisk-1753452859 tempest-ServersTestManualDisk-1753452859-project-member] [instance: 3088ee5e-0d27-4058-a94b-2e04c3b52add] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c8:5a:78', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'fd0eb882-ab95-4373-aa20-ee565a9919e3', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2ffc16a0-351c-4f5b-8a4a-993ae7adb4f3', 'vif_model': 'vmxnet3'}] {{(pid=65726) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 914.213911] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-17b5a1c7-4bc8-41d1-988f-c933bb6421ad tempest-ServersTestManualDisk-1753452859 tempest-ServersTestManualDisk-1753452859-project-member] Creating folder: Project (308ec5a4cf844d2782db77b3edcb4127). Parent ref: group-v995008. {{(pid=65726) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 914.214134] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d4892b1f-dbe3-4ab7-90f4-885a6f8343a1 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.232321] env[65726]: INFO nova.virt.vmwareapi.vm_util [None req-17b5a1c7-4bc8-41d1-988f-c933bb6421ad tempest-ServersTestManualDisk-1753452859 tempest-ServersTestManualDisk-1753452859-project-member] Created folder: Project (308ec5a4cf844d2782db77b3edcb4127) in parent group-v995008. [ 914.232321] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-17b5a1c7-4bc8-41d1-988f-c933bb6421ad tempest-ServersTestManualDisk-1753452859 tempest-ServersTestManualDisk-1753452859-project-member] Creating folder: Instances. Parent ref: group-v995204. {{(pid=65726) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 914.232669] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7106c6a0-3681-4da0-a3bd-335923d0c559 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.238321] env[65726]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 914.238321] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52c23fd6-d39a-65ad-1e34-5d6f64a43f1e" [ 914.238321] env[65726]: _type = "HttpNfcLease" [ 914.238321] env[65726]: } is ready. {{(pid=65726) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 914.239333] env[65726]: DEBUG oslo_vmware.rw_handles [None req-89fee262-eea7-4ed4-83cb-eb9998cc3059 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 914.239333] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52c23fd6-d39a-65ad-1e34-5d6f64a43f1e" [ 914.239333] env[65726]: _type = "HttpNfcLease" [ 914.239333] env[65726]: }. {{(pid=65726) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 914.240714] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f28bd28-c2b0-471b-babc-0458c273bd38 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.253044] env[65726]: DEBUG oslo_vmware.rw_handles [None req-89fee262-eea7-4ed4-83cb-eb9998cc3059 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52ae69b9-d36e-5ceb-eed2-a1b20831ecd5/disk-0.vmdk from lease info. {{(pid=65726) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 914.253044] env[65726]: DEBUG oslo_vmware.rw_handles [None req-89fee262-eea7-4ed4-83cb-eb9998cc3059 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Creating HTTP connection to write to file with size = 21334016 and URL = https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52ae69b9-d36e-5ceb-eed2-a1b20831ecd5/disk-0.vmdk. {{(pid=65726) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 914.255479] env[65726]: INFO nova.virt.vmwareapi.vm_util [None req-17b5a1c7-4bc8-41d1-988f-c933bb6421ad tempest-ServersTestManualDisk-1753452859 tempest-ServersTestManualDisk-1753452859-project-member] Created folder: Instances in parent group-v995204. [ 914.255800] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-17b5a1c7-4bc8-41d1-988f-c933bb6421ad tempest-ServersTestManualDisk-1753452859 tempest-ServersTestManualDisk-1753452859-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 914.256972] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3088ee5e-0d27-4058-a94b-2e04c3b52add] Creating VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 914.319616] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b5b86dd0-572f-46fa-ac57-adc4f71c8677 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.345712] env[65726]: DEBUG nova.network.neutron [req-3d3d8609-65c4-497b-bab5-096199dfa99c req-b299532c-4637-401d-b584-8cb96169958e service nova] [instance: 6f91b053-772a-4497-b29d-349b960c55eb] Updated VIF entry in instance network info cache for port fbccb731-d98d-40ad-a144-4b4f1e48a1ce. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 914.346268] env[65726]: DEBUG nova.network.neutron [req-3d3d8609-65c4-497b-bab5-096199dfa99c req-b299532c-4637-401d-b584-8cb96169958e service nova] [instance: 6f91b053-772a-4497-b29d-349b960c55eb] Updating instance_info_cache with network_info: [{"id": "fbccb731-d98d-40ad-a144-4b4f1e48a1ce", "address": "fa:16:3e:06:3e:65", "network": {"id": "e206b5b9-da36-4e6f-926f-914dec38fac2", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1053201655-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "39067912d091464192516c08eb576696", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7c80243e-93a7-4a95-bc8d-e9534bacd66e", "external-id": "nsx-vlan-transportzone-306", "segmentation_id": 306, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfbccb731-d9", "ovs_interfaceid": "fbccb731-d98d-40ad-a144-4b4f1e48a1ce", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 914.357788] env[65726]: INFO nova.compute.manager [None req-e5e88d5e-aa68-4ff5-9280-b2cfe1d14b27 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: 5a252ef1-93c9-4bff-842b-b64df2bc5d75] Took 45.25 seconds to build instance. [ 914.368307] env[65726]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-99e0fa56-4291-420f-a809-61ab1f3e045d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.372744] env[65726]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 914.372744] env[65726]: value = "task-5116218" [ 914.372744] env[65726]: _type = "Task" [ 914.372744] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.377052] env[65726]: DEBUG oslo_vmware.api [None req-e4fa020c-c0fd-49d8-8a8a-f7bebcb49337 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Task: {'id': task-5116211, 'name': CloneVM_Task} progress is 94%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.387719] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116218, 'name': CreateVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.575892] env[65726]: DEBUG oslo_vmware.api [None req-c4dc675f-0696-4840-ac89-d68c37ca6fa4 tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Task: {'id': task-5116213, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.242595} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.576165] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-c4dc675f-0696-4840-ac89-d68c37ca6fa4 tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Deleted the datastore file {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 914.576386] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-c4dc675f-0696-4840-ac89-d68c37ca6fa4 tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] [instance: 4ea2b70b-0cfe-4fbf-817a-baa028b8372a] Deleted contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 914.576584] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-c4dc675f-0696-4840-ac89-d68c37ca6fa4 tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] [instance: 4ea2b70b-0cfe-4fbf-817a-baa028b8372a] Instance destroyed {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 914.576753] env[65726]: INFO nova.compute.manager [None req-c4dc675f-0696-4840-ac89-d68c37ca6fa4 tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] [instance: 4ea2b70b-0cfe-4fbf-817a-baa028b8372a] Took 1.21 seconds to destroy the instance on the hypervisor. [ 914.577075] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-c4dc675f-0696-4840-ac89-d68c37ca6fa4 tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 914.577355] env[65726]: DEBUG nova.compute.manager [-] [instance: 4ea2b70b-0cfe-4fbf-817a-baa028b8372a] Deallocating network for instance {{(pid=65726) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 914.577517] env[65726]: DEBUG nova.network.neutron [-] [instance: 4ea2b70b-0cfe-4fbf-817a-baa028b8372a] deallocate_for_instance() {{(pid=65726) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 914.578191] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 914.578504] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 914.605411] env[65726]: DEBUG oslo_concurrency.lockutils [req-3ab9b4ea-1548-4202-a06d-9d0833802ee2 req-79e605b8-473a-412e-8aff-b2bc04c493fc service nova] Releasing lock "refresh_cache-4ea2b70b-0cfe-4fbf-817a-baa028b8372a" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 914.676551] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 914.677451] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 914.702754] env[65726]: DEBUG nova.compute.manager [None req-2cc737a8-ad7d-449c-b270-9db9e18d971d tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: dcde25df-87a5-47fb-94ce-334f68894e04] Start spawning the instance on the hypervisor. {{(pid=65726) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 914.713605] env[65726]: DEBUG oslo_vmware.api [None req-2eeb1a84-0a41-40be-a094-58835ef7befd tempest-ServerAddressesTestJSON-569100133 tempest-ServerAddressesTestJSON-569100133-project-member] Task: {'id': task-5116215, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.190144} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.716213] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-2eeb1a84-0a41-40be-a094-58835ef7befd tempest-ServerAddressesTestJSON-569100133 tempest-ServerAddressesTestJSON-569100133-project-member] Deleted the datastore file {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 914.716540] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-2eeb1a84-0a41-40be-a094-58835ef7befd tempest-ServerAddressesTestJSON-569100133 tempest-ServerAddressesTestJSON-569100133-project-member] [instance: 70375746-06dd-4710-9136-95e9e5759a0e] Deleted contents of the VM from datastore datastore2 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 914.717291] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-2eeb1a84-0a41-40be-a094-58835ef7befd tempest-ServerAddressesTestJSON-569100133 tempest-ServerAddressesTestJSON-569100133-project-member] [instance: 70375746-06dd-4710-9136-95e9e5759a0e] Instance destroyed {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 914.717291] env[65726]: INFO nova.compute.manager [None req-2eeb1a84-0a41-40be-a094-58835ef7befd tempest-ServerAddressesTestJSON-569100133 tempest-ServerAddressesTestJSON-569100133-project-member] [instance: 70375746-06dd-4710-9136-95e9e5759a0e] Took 1.18 seconds to destroy the instance on the hypervisor. [ 914.717291] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-2eeb1a84-0a41-40be-a094-58835ef7befd tempest-ServerAddressesTestJSON-569100133 tempest-ServerAddressesTestJSON-569100133-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 914.717975] env[65726]: DEBUG nova.compute.manager [-] [instance: 70375746-06dd-4710-9136-95e9e5759a0e] Deallocating network for instance {{(pid=65726) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 914.717975] env[65726]: DEBUG nova.network.neutron [-] [instance: 70375746-06dd-4710-9136-95e9e5759a0e] deallocate_for_instance() {{(pid=65726) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 914.718310] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 914.718575] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 914.786956] env[65726]: DEBUG nova.virt.hardware [None req-2cc737a8-ad7d-449c-b270-9db9e18d971d tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T19:28:34Z,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1539f06fa3534e90acbc3a60c4b8bd3f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T19:28:35Z,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 914.787329] env[65726]: DEBUG nova.virt.hardware [None req-2cc737a8-ad7d-449c-b270-9db9e18d971d tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 914.787506] env[65726]: DEBUG nova.virt.hardware [None req-2cc737a8-ad7d-449c-b270-9db9e18d971d tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 914.787705] env[65726]: DEBUG nova.virt.hardware [None req-2cc737a8-ad7d-449c-b270-9db9e18d971d tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 914.787899] env[65726]: DEBUG nova.virt.hardware [None req-2cc737a8-ad7d-449c-b270-9db9e18d971d tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 914.788138] env[65726]: DEBUG nova.virt.hardware [None req-2cc737a8-ad7d-449c-b270-9db9e18d971d tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 914.788373] env[65726]: DEBUG nova.virt.hardware [None req-2cc737a8-ad7d-449c-b270-9db9e18d971d tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 914.788538] env[65726]: DEBUG nova.virt.hardware [None req-2cc737a8-ad7d-449c-b270-9db9e18d971d tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 914.788812] env[65726]: DEBUG nova.virt.hardware [None req-2cc737a8-ad7d-449c-b270-9db9e18d971d tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 914.789051] env[65726]: DEBUG nova.virt.hardware [None req-2cc737a8-ad7d-449c-b270-9db9e18d971d tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 914.789271] env[65726]: DEBUG nova.virt.hardware [None req-2cc737a8-ad7d-449c-b270-9db9e18d971d tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 914.790214] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04c3c8fd-411a-4636-a8e8-aeae12fb8551 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.813217] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45ae1826-9311-44ec-b131-d322b0a32986 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.857310] env[65726]: DEBUG oslo_concurrency.lockutils [req-3d3d8609-65c4-497b-bab5-096199dfa99c req-b299532c-4637-401d-b584-8cb96169958e service nova] Releasing lock "refresh_cache-6f91b053-772a-4497-b29d-349b960c55eb" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 914.857741] env[65726]: DEBUG nova.compute.manager [req-3d3d8609-65c4-497b-bab5-096199dfa99c req-b299532c-4637-401d-b584-8cb96169958e service nova] [instance: 3088ee5e-0d27-4058-a94b-2e04c3b52add] Received event network-vif-plugged-2ffc16a0-351c-4f5b-8a4a-993ae7adb4f3 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 914.858164] env[65726]: DEBUG oslo_concurrency.lockutils [req-3d3d8609-65c4-497b-bab5-096199dfa99c req-b299532c-4637-401d-b584-8cb96169958e service nova] Acquiring lock "3088ee5e-0d27-4058-a94b-2e04c3b52add-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 914.858440] env[65726]: DEBUG oslo_concurrency.lockutils [req-3d3d8609-65c4-497b-bab5-096199dfa99c req-b299532c-4637-401d-b584-8cb96169958e service nova] Lock "3088ee5e-0d27-4058-a94b-2e04c3b52add-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 914.858720] env[65726]: DEBUG oslo_concurrency.lockutils [req-3d3d8609-65c4-497b-bab5-096199dfa99c req-b299532c-4637-401d-b584-8cb96169958e service nova] Lock "3088ee5e-0d27-4058-a94b-2e04c3b52add-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 914.859227] env[65726]: DEBUG nova.compute.manager [req-3d3d8609-65c4-497b-bab5-096199dfa99c req-b299532c-4637-401d-b584-8cb96169958e service nova] [instance: 3088ee5e-0d27-4058-a94b-2e04c3b52add] No waiting events found dispatching network-vif-plugged-2ffc16a0-351c-4f5b-8a4a-993ae7adb4f3 {{(pid=65726) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 914.859332] env[65726]: WARNING nova.compute.manager [req-3d3d8609-65c4-497b-bab5-096199dfa99c req-b299532c-4637-401d-b584-8cb96169958e service nova] [instance: 3088ee5e-0d27-4058-a94b-2e04c3b52add] Received unexpected event network-vif-plugged-2ffc16a0-351c-4f5b-8a4a-993ae7adb4f3 for instance with vm_state building and task_state spawning. [ 914.867463] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e5e88d5e-aa68-4ff5-9280-b2cfe1d14b27 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Lock "5a252ef1-93c9-4bff-842b-b64df2bc5d75" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 59.296s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 914.873775] env[65726]: DEBUG oslo_vmware.api [None req-e4fa020c-c0fd-49d8-8a8a-f7bebcb49337 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Task: {'id': task-5116211, 'name': CloneVM_Task} progress is 94%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.888144] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116218, 'name': CreateVM_Task, 'duration_secs': 0.374684} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.888604] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3088ee5e-0d27-4058-a94b-2e04c3b52add] Created VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 914.889593] env[65726]: WARNING openstack [None req-17b5a1c7-4bc8-41d1-988f-c933bb6421ad tempest-ServersTestManualDisk-1753452859 tempest-ServersTestManualDisk-1753452859-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 914.890052] env[65726]: WARNING openstack [None req-17b5a1c7-4bc8-41d1-988f-c933bb6421ad tempest-ServersTestManualDisk-1753452859 tempest-ServersTestManualDisk-1753452859-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 914.896446] env[65726]: DEBUG oslo_concurrency.lockutils [None req-17b5a1c7-4bc8-41d1-988f-c933bb6421ad tempest-ServersTestManualDisk-1753452859 tempest-ServersTestManualDisk-1753452859-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 914.896446] env[65726]: DEBUG oslo_concurrency.lockutils [None req-17b5a1c7-4bc8-41d1-988f-c933bb6421ad tempest-ServersTestManualDisk-1753452859 tempest-ServersTestManualDisk-1753452859-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 914.896959] env[65726]: DEBUG oslo_concurrency.lockutils [None req-17b5a1c7-4bc8-41d1-988f-c933bb6421ad tempest-ServersTestManualDisk-1753452859 tempest-ServersTestManualDisk-1753452859-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 914.899333] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dfca0b28-5e97-402b-9d6c-d1de28fa4fbd {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.908771] env[65726]: DEBUG oslo_vmware.api [None req-17b5a1c7-4bc8-41d1-988f-c933bb6421ad tempest-ServersTestManualDisk-1753452859 tempest-ServersTestManualDisk-1753452859-project-member] Waiting for the task: (returnval){ [ 914.908771] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]528abfb5-701d-d54e-6142-ff533b6e2c8b" [ 914.908771] env[65726]: _type = "Task" [ 914.908771] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.922634] env[65726]: DEBUG oslo_vmware.api [None req-17b5a1c7-4bc8-41d1-988f-c933bb6421ad tempest-ServersTestManualDisk-1753452859 tempest-ServersTestManualDisk-1753452859-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]528abfb5-701d-d54e-6142-ff533b6e2c8b, 'name': SearchDatastore_Task, 'duration_secs': 0.010738} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.924526] env[65726]: DEBUG oslo_concurrency.lockutils [None req-17b5a1c7-4bc8-41d1-988f-c933bb6421ad tempest-ServersTestManualDisk-1753452859 tempest-ServersTestManualDisk-1753452859-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 914.924786] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-17b5a1c7-4bc8-41d1-988f-c933bb6421ad tempest-ServersTestManualDisk-1753452859 tempest-ServersTestManualDisk-1753452859-project-member] [instance: 3088ee5e-0d27-4058-a94b-2e04c3b52add] Processing image b52362a3-ee8a-4cbf-b06f-513b0cc8f95c {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 914.925077] env[65726]: DEBUG oslo_concurrency.lockutils [None req-17b5a1c7-4bc8-41d1-988f-c933bb6421ad tempest-ServersTestManualDisk-1753452859 tempest-ServersTestManualDisk-1753452859-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 914.925222] env[65726]: DEBUG oslo_concurrency.lockutils [None req-17b5a1c7-4bc8-41d1-988f-c933bb6421ad tempest-ServersTestManualDisk-1753452859 tempest-ServersTestManualDisk-1753452859-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 914.925495] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-17b5a1c7-4bc8-41d1-988f-c933bb6421ad tempest-ServersTestManualDisk-1753452859 tempest-ServersTestManualDisk-1753452859-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 914.928879] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0d924a58-3c6f-418b-9b70-105fef4cd01c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.944474] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-17b5a1c7-4bc8-41d1-988f-c933bb6421ad tempest-ServersTestManualDisk-1753452859 tempest-ServersTestManualDisk-1753452859-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 914.944474] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-17b5a1c7-4bc8-41d1-988f-c933bb6421ad tempest-ServersTestManualDisk-1753452859 tempest-ServersTestManualDisk-1753452859-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65726) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 914.945974] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-aad93459-77a5-4365-8747-974413e1ab43 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.956317] env[65726]: DEBUG oslo_vmware.api [None req-17b5a1c7-4bc8-41d1-988f-c933bb6421ad tempest-ServersTestManualDisk-1753452859 tempest-ServersTestManualDisk-1753452859-project-member] Waiting for the task: (returnval){ [ 914.956317] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52ea9604-dab7-8cbf-7b8e-72b4cd776cbb" [ 914.956317] env[65726]: _type = "Task" [ 914.956317] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.968621] env[65726]: DEBUG oslo_vmware.api [None req-17b5a1c7-4bc8-41d1-988f-c933bb6421ad tempest-ServersTestManualDisk-1753452859 tempest-ServersTestManualDisk-1753452859-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52ea9604-dab7-8cbf-7b8e-72b4cd776cbb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.016276] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 915.016328] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 915.363986] env[65726]: DEBUG oslo_vmware.api [None req-e4fa020c-c0fd-49d8-8a8a-f7bebcb49337 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Task: {'id': task-5116211, 'name': CloneVM_Task, 'duration_secs': 1.515191} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.368884] env[65726]: INFO nova.virt.vmwareapi.vmops [None req-e4fa020c-c0fd-49d8-8a8a-f7bebcb49337 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] [instance: 815ee20b-cb07-4514-bd43-1b74972508a5] Created linked-clone VM from snapshot [ 915.369982] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96d4ef51-b6b9-4557-bc00-3b961845076f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.381552] env[65726]: DEBUG nova.virt.vmwareapi.images [None req-e4fa020c-c0fd-49d8-8a8a-f7bebcb49337 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] [instance: 815ee20b-cb07-4514-bd43-1b74972508a5] Uploading image 952406eb-2f8d-4d9d-9b5d-e189174387a1 {{(pid=65726) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 915.390012] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-110ada68-b785-46f9-9e39-46f0e570447c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.398822] env[65726]: DEBUG oslo_vmware.rw_handles [None req-89fee262-eea7-4ed4-83cb-eb9998cc3059 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Completed reading data from the image iterator. {{(pid=65726) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 915.399258] env[65726]: DEBUG oslo_vmware.rw_handles [None req-89fee262-eea7-4ed4-83cb-eb9998cc3059 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52ae69b9-d36e-5ceb-eed2-a1b20831ecd5/disk-0.vmdk. {{(pid=65726) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 915.400710] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f36a907-5a25-48ff-998f-c751381cad09 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.412212] env[65726]: DEBUG oslo_vmware.rw_handles [None req-89fee262-eea7-4ed4-83cb-eb9998cc3059 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52ae69b9-d36e-5ceb-eed2-a1b20831ecd5/disk-0.vmdk is in state: ready. {{(pid=65726) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 915.412435] env[65726]: DEBUG oslo_vmware.rw_handles [None req-89fee262-eea7-4ed4-83cb-eb9998cc3059 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Releasing lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52ae69b9-d36e-5ceb-eed2-a1b20831ecd5/disk-0.vmdk. {{(pid=65726) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 915.416580] env[65726]: DEBUG oslo_vmware.rw_handles [None req-e4fa020c-c0fd-49d8-8a8a-f7bebcb49337 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 915.416580] env[65726]: value = "vm-995203" [ 915.416580] env[65726]: _type = "VirtualMachine" [ 915.416580] env[65726]: }. {{(pid=65726) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 915.416912] env[65726]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-b6f88d1d-80b2-4e6e-bf6a-f8537df0d0bf {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.419101] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-ddbf446b-604b-47b9-9fd2-fd33c5a019ca {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.421951] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0c7785d-3686-4510-b1d4-58b85efd5aaf {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.463128] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ac3b18a-5d92-4041-9e48-f9dd6fcafaf7 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.465753] env[65726]: DEBUG oslo_vmware.rw_handles [None req-e4fa020c-c0fd-49d8-8a8a-f7bebcb49337 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Lease: (returnval){ [ 915.465753] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52860f3a-6578-28f0-b984-d21421976786" [ 915.465753] env[65726]: _type = "HttpNfcLease" [ 915.465753] env[65726]: } obtained for exporting VM: (result){ [ 915.465753] env[65726]: value = "vm-995203" [ 915.465753] env[65726]: _type = "VirtualMachine" [ 915.465753] env[65726]: }. {{(pid=65726) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 915.466065] env[65726]: DEBUG oslo_vmware.api [None req-e4fa020c-c0fd-49d8-8a8a-f7bebcb49337 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Waiting for the lease: (returnval){ [ 915.466065] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52860f3a-6578-28f0-b984-d21421976786" [ 915.466065] env[65726]: _type = "HttpNfcLease" [ 915.466065] env[65726]: } to be ready. {{(pid=65726) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 915.475714] env[65726]: DEBUG oslo_vmware.api [None req-17b5a1c7-4bc8-41d1-988f-c933bb6421ad tempest-ServersTestManualDisk-1753452859 tempest-ServersTestManualDisk-1753452859-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52ea9604-dab7-8cbf-7b8e-72b4cd776cbb, 'name': SearchDatastore_Task, 'duration_secs': 0.010971} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.477438] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-265e3cf3-3c14-42a8-84b5-3d6fc8852565 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.483775] env[65726]: DEBUG nova.network.neutron [-] [instance: 4ea2b70b-0cfe-4fbf-817a-baa028b8372a] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 915.485134] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5959ca65-b5a8-4025-aeda-74487daf9caa {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.487454] env[65726]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 915.487454] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52860f3a-6578-28f0-b984-d21421976786" [ 915.487454] env[65726]: _type = "HttpNfcLease" [ 915.487454] env[65726]: } is ready. {{(pid=65726) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 915.488379] env[65726]: DEBUG oslo_vmware.rw_handles [None req-e4fa020c-c0fd-49d8-8a8a-f7bebcb49337 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 915.488379] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52860f3a-6578-28f0-b984-d21421976786" [ 915.488379] env[65726]: _type = "HttpNfcLease" [ 915.488379] env[65726]: }. {{(pid=65726) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 915.489851] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-baebfb60-970b-49e7-8579-bb679cfc68ad {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.503839] env[65726]: DEBUG nova.compute.provider_tree [None req-801b9b62-b5bd-4b5f-80fa-32388dd4cb88 tempest-ServerTagsTestJSON-59142761 tempest-ServerTagsTestJSON-59142761-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 915.505335] env[65726]: DEBUG oslo_vmware.api [None req-17b5a1c7-4bc8-41d1-988f-c933bb6421ad tempest-ServersTestManualDisk-1753452859 tempest-ServersTestManualDisk-1753452859-project-member] Waiting for the task: (returnval){ [ 915.505335] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]522a041e-c3a2-81a8-7b3b-a7c18e185760" [ 915.505335] env[65726]: _type = "Task" [ 915.505335] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.513947] env[65726]: DEBUG oslo_vmware.rw_handles [None req-e4fa020c-c0fd-49d8-8a8a-f7bebcb49337 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5265049a-4106-25b3-49fd-ca4b56af6a22/disk-0.vmdk from lease info. {{(pid=65726) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 915.514171] env[65726]: DEBUG oslo_vmware.rw_handles [None req-e4fa020c-c0fd-49d8-8a8a-f7bebcb49337 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5265049a-4106-25b3-49fd-ca4b56af6a22/disk-0.vmdk for reading. {{(pid=65726) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 915.573870] env[65726]: DEBUG oslo_vmware.api [None req-17b5a1c7-4bc8-41d1-988f-c933bb6421ad tempest-ServersTestManualDisk-1753452859 tempest-ServersTestManualDisk-1753452859-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]522a041e-c3a2-81a8-7b3b-a7c18e185760, 'name': SearchDatastore_Task, 'duration_secs': 0.011886} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.575964] env[65726]: DEBUG oslo_concurrency.lockutils [None req-17b5a1c7-4bc8-41d1-988f-c933bb6421ad tempest-ServersTestManualDisk-1753452859 tempest-ServersTestManualDisk-1753452859-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 915.576291] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-17b5a1c7-4bc8-41d1-988f-c933bb6421ad tempest-ServersTestManualDisk-1753452859 tempest-ServersTestManualDisk-1753452859-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore2] 3088ee5e-0d27-4058-a94b-2e04c3b52add/3088ee5e-0d27-4058-a94b-2e04c3b52add.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 915.577349] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b72b6fda-3675-4da0-853f-2f552acd3c1e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.586505] env[65726]: DEBUG oslo_vmware.api [None req-17b5a1c7-4bc8-41d1-988f-c933bb6421ad tempest-ServersTestManualDisk-1753452859 tempest-ServersTestManualDisk-1753452859-project-member] Waiting for the task: (returnval){ [ 915.586505] env[65726]: value = "task-5116220" [ 915.586505] env[65726]: _type = "Task" [ 915.586505] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.596134] env[65726]: DEBUG oslo_vmware.api [None req-17b5a1c7-4bc8-41d1-988f-c933bb6421ad tempest-ServersTestManualDisk-1753452859 tempest-ServersTestManualDisk-1753452859-project-member] Task: {'id': task-5116220, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.636251] env[65726]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-78273f04-3f17-4862-8942-4f5444cec6fa {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.737687] env[65726]: DEBUG oslo_vmware.rw_handles [None req-89fee262-eea7-4ed4-83cb-eb9998cc3059 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Closed VMDK write handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52ae69b9-d36e-5ceb-eed2-a1b20831ecd5/disk-0.vmdk. {{(pid=65726) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 915.738046] env[65726]: INFO nova.virt.vmwareapi.images [None req-89fee262-eea7-4ed4-83cb-eb9998cc3059 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 6f91b053-772a-4497-b29d-349b960c55eb] Downloaded image file data 8ecee2b1-8703-4e91-8a32-dc17932b237e [ 915.739063] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8eb6777f-a83b-4c40-a241-b1f72ff91724 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.758398] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3818bf6e-d915-483f-bad9-ec36dca97df3 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.799169] env[65726]: INFO nova.virt.vmwareapi.images [None req-89fee262-eea7-4ed4-83cb-eb9998cc3059 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 6f91b053-772a-4497-b29d-349b960c55eb] The imported VM was unregistered [ 915.803908] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-89fee262-eea7-4ed4-83cb-eb9998cc3059 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 6f91b053-772a-4497-b29d-349b960c55eb] Caching image {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 915.804492] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-89fee262-eea7-4ed4-83cb-eb9998cc3059 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Creating directory with path [datastore1] devstack-image-cache_base/8ecee2b1-8703-4e91-8a32-dc17932b237e {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 915.804825] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-811a7458-de65-4724-b105-d7a329ecbfb5 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.825453] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-89fee262-eea7-4ed4-83cb-eb9998cc3059 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Created directory with path [datastore1] devstack-image-cache_base/8ecee2b1-8703-4e91-8a32-dc17932b237e {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 915.825711] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-89fee262-eea7-4ed4-83cb-eb9998cc3059 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Moving virtual disk from [datastore1] OSTACK_IMG_1f8b9f24-a5fc-420d-a36b-40a6efa61bbf/OSTACK_IMG_1f8b9f24-a5fc-420d-a36b-40a6efa61bbf.vmdk to [datastore1] devstack-image-cache_base/8ecee2b1-8703-4e91-8a32-dc17932b237e/8ecee2b1-8703-4e91-8a32-dc17932b237e.vmdk. {{(pid=65726) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 915.826060] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-80dcd263-674d-429e-8744-af2721f0d1ba {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.838564] env[65726]: DEBUG oslo_vmware.api [None req-89fee262-eea7-4ed4-83cb-eb9998cc3059 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Waiting for the task: (returnval){ [ 915.838564] env[65726]: value = "task-5116222" [ 915.838564] env[65726]: _type = "Task" [ 915.838564] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.851102] env[65726]: DEBUG oslo_vmware.api [None req-89fee262-eea7-4ed4-83cb-eb9998cc3059 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': task-5116222, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.899820] env[65726]: DEBUG nova.network.neutron [None req-2cc737a8-ad7d-449c-b270-9db9e18d971d tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: dcde25df-87a5-47fb-94ce-334f68894e04] Successfully updated port: bf51c96c-63c5-46de-a40f-58e6769c5fcf {{(pid=65726) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 915.904432] env[65726]: DEBUG nova.network.neutron [-] [instance: 70375746-06dd-4710-9136-95e9e5759a0e] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 915.990278] env[65726]: INFO nova.compute.manager [-] [instance: 4ea2b70b-0cfe-4fbf-817a-baa028b8372a] Took 1.41 seconds to deallocate network for instance. [ 916.008164] env[65726]: DEBUG nova.scheduler.client.report [None req-801b9b62-b5bd-4b5f-80fa-32388dd4cb88 tempest-ServerTagsTestJSON-59142761 tempest-ServerTagsTestJSON-59142761-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 916.077666] env[65726]: DEBUG nova.compute.manager [req-79681db2-6f01-440b-96ca-ce88a4aab1c5 req-5d23e393-cd62-49fb-b86c-65f48f6f774e service nova] [instance: 3088ee5e-0d27-4058-a94b-2e04c3b52add] Received event network-changed-2ffc16a0-351c-4f5b-8a4a-993ae7adb4f3 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 916.078144] env[65726]: DEBUG nova.compute.manager [req-79681db2-6f01-440b-96ca-ce88a4aab1c5 req-5d23e393-cd62-49fb-b86c-65f48f6f774e service nova] [instance: 3088ee5e-0d27-4058-a94b-2e04c3b52add] Refreshing instance network info cache due to event network-changed-2ffc16a0-351c-4f5b-8a4a-993ae7adb4f3. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 916.078795] env[65726]: DEBUG oslo_concurrency.lockutils [req-79681db2-6f01-440b-96ca-ce88a4aab1c5 req-5d23e393-cd62-49fb-b86c-65f48f6f774e service nova] Acquiring lock "refresh_cache-3088ee5e-0d27-4058-a94b-2e04c3b52add" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 916.079113] env[65726]: DEBUG oslo_concurrency.lockutils [req-79681db2-6f01-440b-96ca-ce88a4aab1c5 req-5d23e393-cd62-49fb-b86c-65f48f6f774e service nova] Acquired lock "refresh_cache-3088ee5e-0d27-4058-a94b-2e04c3b52add" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 916.079814] env[65726]: DEBUG nova.network.neutron [req-79681db2-6f01-440b-96ca-ce88a4aab1c5 req-5d23e393-cd62-49fb-b86c-65f48f6f774e service nova] [instance: 3088ee5e-0d27-4058-a94b-2e04c3b52add] Refreshing network info cache for port 2ffc16a0-351c-4f5b-8a4a-993ae7adb4f3 {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 916.107344] env[65726]: DEBUG oslo_vmware.api [None req-17b5a1c7-4bc8-41d1-988f-c933bb6421ad tempest-ServersTestManualDisk-1753452859 tempest-ServersTestManualDisk-1753452859-project-member] Task: {'id': task-5116220, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.284213] env[65726]: DEBUG nova.compute.manager [req-be2b8aae-b722-4c41-8e33-f250ddeaf354 req-4d32bb67-aca0-4d3b-b4f8-22919df00736 service nova] [instance: dcde25df-87a5-47fb-94ce-334f68894e04] Received event network-vif-plugged-bf51c96c-63c5-46de-a40f-58e6769c5fcf {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 916.284571] env[65726]: DEBUG oslo_concurrency.lockutils [req-be2b8aae-b722-4c41-8e33-f250ddeaf354 req-4d32bb67-aca0-4d3b-b4f8-22919df00736 service nova] Acquiring lock "dcde25df-87a5-47fb-94ce-334f68894e04-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 916.284811] env[65726]: DEBUG oslo_concurrency.lockutils [req-be2b8aae-b722-4c41-8e33-f250ddeaf354 req-4d32bb67-aca0-4d3b-b4f8-22919df00736 service nova] Lock "dcde25df-87a5-47fb-94ce-334f68894e04-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 916.285567] env[65726]: DEBUG oslo_concurrency.lockutils [req-be2b8aae-b722-4c41-8e33-f250ddeaf354 req-4d32bb67-aca0-4d3b-b4f8-22919df00736 service nova] Lock "dcde25df-87a5-47fb-94ce-334f68894e04-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 916.286104] env[65726]: DEBUG nova.compute.manager [req-be2b8aae-b722-4c41-8e33-f250ddeaf354 req-4d32bb67-aca0-4d3b-b4f8-22919df00736 service nova] [instance: dcde25df-87a5-47fb-94ce-334f68894e04] No waiting events found dispatching network-vif-plugged-bf51c96c-63c5-46de-a40f-58e6769c5fcf {{(pid=65726) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 916.286104] env[65726]: WARNING nova.compute.manager [req-be2b8aae-b722-4c41-8e33-f250ddeaf354 req-4d32bb67-aca0-4d3b-b4f8-22919df00736 service nova] [instance: dcde25df-87a5-47fb-94ce-334f68894e04] Received unexpected event network-vif-plugged-bf51c96c-63c5-46de-a40f-58e6769c5fcf for instance with vm_state building and task_state spawning. [ 916.355038] env[65726]: DEBUG oslo_vmware.api [None req-89fee262-eea7-4ed4-83cb-eb9998cc3059 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': task-5116222, 'name': MoveVirtualDisk_Task} progress is 12%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.373850] env[65726]: DEBUG oslo_concurrency.lockutils [None req-edb92956-8e7a-4023-9c72-6f4afa47cb43 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Acquiring lock "5a252ef1-93c9-4bff-842b-b64df2bc5d75" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 916.374219] env[65726]: DEBUG oslo_concurrency.lockutils [None req-edb92956-8e7a-4023-9c72-6f4afa47cb43 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Lock "5a252ef1-93c9-4bff-842b-b64df2bc5d75" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 916.407898] env[65726]: DEBUG oslo_concurrency.lockutils [None req-2cc737a8-ad7d-449c-b270-9db9e18d971d tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Acquiring lock "refresh_cache-dcde25df-87a5-47fb-94ce-334f68894e04" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 916.408775] env[65726]: DEBUG oslo_concurrency.lockutils [None req-2cc737a8-ad7d-449c-b270-9db9e18d971d tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Acquired lock "refresh_cache-dcde25df-87a5-47fb-94ce-334f68894e04" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 916.409216] env[65726]: DEBUG nova.network.neutron [None req-2cc737a8-ad7d-449c-b270-9db9e18d971d tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: dcde25df-87a5-47fb-94ce-334f68894e04] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 916.411920] env[65726]: INFO nova.compute.manager [-] [instance: 70375746-06dd-4710-9136-95e9e5759a0e] Took 1.69 seconds to deallocate network for instance. [ 916.503038] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c4dc675f-0696-4840-ac89-d68c37ca6fa4 tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 916.513894] env[65726]: DEBUG oslo_concurrency.lockutils [None req-801b9b62-b5bd-4b5f-80fa-32388dd4cb88 tempest-ServerTagsTestJSON-59142761 tempest-ServerTagsTestJSON-59142761-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.807s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 916.514625] env[65726]: DEBUG nova.compute.manager [None req-801b9b62-b5bd-4b5f-80fa-32388dd4cb88 tempest-ServerTagsTestJSON-59142761 tempest-ServerTagsTestJSON-59142761-project-member] [instance: 9029549c-1914-4bae-91e2-8812b79051ec] Start building networks asynchronously for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 916.517752] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5782ba22-8056-4347-9a97-4b1e5a35ed3b tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 27.498s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 916.517973] env[65726]: DEBUG nova.objects.instance [None req-5782ba22-8056-4347-9a97-4b1e5a35ed3b tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: f2d9090c-988f-43f4-9c81-7aa718a3438a] Trying to apply a migration context that does not seem to be set for this instance {{(pid=65726) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 916.584832] env[65726]: WARNING openstack [req-79681db2-6f01-440b-96ca-ce88a4aab1c5 req-5d23e393-cd62-49fb-b86c-65f48f6f774e service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 916.585496] env[65726]: WARNING openstack [req-79681db2-6f01-440b-96ca-ce88a4aab1c5 req-5d23e393-cd62-49fb-b86c-65f48f6f774e service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 916.613990] env[65726]: DEBUG oslo_vmware.api [None req-17b5a1c7-4bc8-41d1-988f-c933bb6421ad tempest-ServersTestManualDisk-1753452859 tempest-ServersTestManualDisk-1753452859-project-member] Task: {'id': task-5116220, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.55615} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.614575] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-17b5a1c7-4bc8-41d1-988f-c933bb6421ad tempest-ServersTestManualDisk-1753452859 tempest-ServersTestManualDisk-1753452859-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore2] 3088ee5e-0d27-4058-a94b-2e04c3b52add/3088ee5e-0d27-4058-a94b-2e04c3b52add.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 916.614929] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-17b5a1c7-4bc8-41d1-988f-c933bb6421ad tempest-ServersTestManualDisk-1753452859 tempest-ServersTestManualDisk-1753452859-project-member] [instance: 3088ee5e-0d27-4058-a94b-2e04c3b52add] Extending root virtual disk to 1048576 {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 916.615478] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5022326e-0d82-4c4a-8d93-d4b328200d8e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.631729] env[65726]: DEBUG oslo_vmware.api [None req-17b5a1c7-4bc8-41d1-988f-c933bb6421ad tempest-ServersTestManualDisk-1753452859 tempest-ServersTestManualDisk-1753452859-project-member] Waiting for the task: (returnval){ [ 916.631729] env[65726]: value = "task-5116223" [ 916.631729] env[65726]: _type = "Task" [ 916.631729] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.652942] env[65726]: DEBUG oslo_vmware.api [None req-17b5a1c7-4bc8-41d1-988f-c933bb6421ad tempest-ServersTestManualDisk-1753452859 tempest-ServersTestManualDisk-1753452859-project-member] Task: {'id': task-5116223, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.720888] env[65726]: WARNING openstack [req-79681db2-6f01-440b-96ca-ce88a4aab1c5 req-5d23e393-cd62-49fb-b86c-65f48f6f774e service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 916.723797] env[65726]: WARNING openstack [req-79681db2-6f01-440b-96ca-ce88a4aab1c5 req-5d23e393-cd62-49fb-b86c-65f48f6f774e service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 916.847025] env[65726]: WARNING openstack [req-79681db2-6f01-440b-96ca-ce88a4aab1c5 req-5d23e393-cd62-49fb-b86c-65f48f6f774e service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 916.847025] env[65726]: WARNING openstack [req-79681db2-6f01-440b-96ca-ce88a4aab1c5 req-5d23e393-cd62-49fb-b86c-65f48f6f774e service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 916.869907] env[65726]: DEBUG oslo_vmware.api [None req-89fee262-eea7-4ed4-83cb-eb9998cc3059 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': task-5116222, 'name': MoveVirtualDisk_Task} progress is 32%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.878706] env[65726]: DEBUG nova.compute.utils [None req-edb92956-8e7a-4023-9c72-6f4afa47cb43 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Using /dev/sd instead of None {{(pid=65726) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 916.913961] env[65726]: WARNING openstack [None req-2cc737a8-ad7d-449c-b270-9db9e18d971d tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 916.914406] env[65726]: WARNING openstack [None req-2cc737a8-ad7d-449c-b270-9db9e18d971d tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 916.922670] env[65726]: DEBUG nova.network.neutron [None req-2cc737a8-ad7d-449c-b270-9db9e18d971d tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: dcde25df-87a5-47fb-94ce-334f68894e04] Instance cache missing network info. {{(pid=65726) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 916.926807] env[65726]: DEBUG oslo_concurrency.lockutils [None req-2eeb1a84-0a41-40be-a094-58835ef7befd tempest-ServerAddressesTestJSON-569100133 tempest-ServerAddressesTestJSON-569100133-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 916.964315] env[65726]: DEBUG nova.network.neutron [req-79681db2-6f01-440b-96ca-ce88a4aab1c5 req-5d23e393-cd62-49fb-b86c-65f48f6f774e service nova] [instance: 3088ee5e-0d27-4058-a94b-2e04c3b52add] Updated VIF entry in instance network info cache for port 2ffc16a0-351c-4f5b-8a4a-993ae7adb4f3. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 916.966260] env[65726]: DEBUG nova.network.neutron [req-79681db2-6f01-440b-96ca-ce88a4aab1c5 req-5d23e393-cd62-49fb-b86c-65f48f6f774e service nova] [instance: 3088ee5e-0d27-4058-a94b-2e04c3b52add] Updating instance_info_cache with network_info: [{"id": "2ffc16a0-351c-4f5b-8a4a-993ae7adb4f3", "address": "fa:16:3e:c8:5a:78", "network": {"id": "6e131d80-bd85-46b6-9b8d-c4d34af67f1d", "bridge": "br-int", "label": "tempest-ServersTestManualDisk-1706238551-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "308ec5a4cf844d2782db77b3edcb4127", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fd0eb882-ab95-4373-aa20-ee565a9919e3", "external-id": "nsx-vlan-transportzone-510", "segmentation_id": 510, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2ffc16a0-35", "ovs_interfaceid": "2ffc16a0-351c-4f5b-8a4a-993ae7adb4f3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 917.019265] env[65726]: WARNING openstack [None req-2cc737a8-ad7d-449c-b270-9db9e18d971d tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 917.019783] env[65726]: WARNING openstack [None req-2cc737a8-ad7d-449c-b270-9db9e18d971d tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 917.030737] env[65726]: DEBUG nova.compute.utils [None req-801b9b62-b5bd-4b5f-80fa-32388dd4cb88 tempest-ServerTagsTestJSON-59142761 tempest-ServerTagsTestJSON-59142761-project-member] Using /dev/sd instead of None {{(pid=65726) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 917.037238] env[65726]: DEBUG nova.compute.manager [None req-801b9b62-b5bd-4b5f-80fa-32388dd4cb88 tempest-ServerTagsTestJSON-59142761 tempest-ServerTagsTestJSON-59142761-project-member] [instance: 9029549c-1914-4bae-91e2-8812b79051ec] Allocating IP information in the background. {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 917.037574] env[65726]: DEBUG nova.network.neutron [None req-801b9b62-b5bd-4b5f-80fa-32388dd4cb88 tempest-ServerTagsTestJSON-59142761 tempest-ServerTagsTestJSON-59142761-project-member] [instance: 9029549c-1914-4bae-91e2-8812b79051ec] allocate_for_instance() {{(pid=65726) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 917.038151] env[65726]: WARNING neutronclient.v2_0.client [None req-801b9b62-b5bd-4b5f-80fa-32388dd4cb88 tempest-ServerTagsTestJSON-59142761 tempest-ServerTagsTestJSON-59142761-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 917.038590] env[65726]: WARNING neutronclient.v2_0.client [None req-801b9b62-b5bd-4b5f-80fa-32388dd4cb88 tempest-ServerTagsTestJSON-59142761 tempest-ServerTagsTestJSON-59142761-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 917.039282] env[65726]: WARNING openstack [None req-801b9b62-b5bd-4b5f-80fa-32388dd4cb88 tempest-ServerTagsTestJSON-59142761 tempest-ServerTagsTestJSON-59142761-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 917.039692] env[65726]: WARNING openstack [None req-801b9b62-b5bd-4b5f-80fa-32388dd4cb88 tempest-ServerTagsTestJSON-59142761 tempest-ServerTagsTestJSON-59142761-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 917.049486] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5782ba22-8056-4347-9a97-4b1e5a35ed3b tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 0.532s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 917.052342] env[65726]: DEBUG oslo_concurrency.lockutils [None req-1cae8219-3b3b-4d11-bdce-863de7bcd2d6 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 26.843s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 917.052873] env[65726]: DEBUG nova.objects.instance [None req-1cae8219-3b3b-4d11-bdce-863de7bcd2d6 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] Lazy-loading 'resources' on Instance uuid 2cbee359-a1fc-45c0-bcc5-a2c24ded9c69 {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 917.114852] env[65726]: DEBUG nova.policy [None req-801b9b62-b5bd-4b5f-80fa-32388dd4cb88 tempest-ServerTagsTestJSON-59142761 tempest-ServerTagsTestJSON-59142761-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2c69b45a7d7c4dd79a7748b0035bb54e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b635b43ad6824e99a609092196631b58', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65726) authorize /opt/stack/nova/nova/policy.py:192}} [ 917.146634] env[65726]: DEBUG oslo_vmware.api [None req-17b5a1c7-4bc8-41d1-988f-c933bb6421ad tempest-ServersTestManualDisk-1753452859 tempest-ServersTestManualDisk-1753452859-project-member] Task: {'id': task-5116223, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.113678} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.150590] env[65726]: WARNING openstack [None req-2cc737a8-ad7d-449c-b270-9db9e18d971d tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 917.151210] env[65726]: WARNING openstack [None req-2cc737a8-ad7d-449c-b270-9db9e18d971d tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 917.160133] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-17b5a1c7-4bc8-41d1-988f-c933bb6421ad tempest-ServersTestManualDisk-1753452859 tempest-ServersTestManualDisk-1753452859-project-member] [instance: 3088ee5e-0d27-4058-a94b-2e04c3b52add] Extended root virtual disk {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 917.162218] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b87e42ac-df79-46aa-976b-e47fa963ac36 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.199918] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-17b5a1c7-4bc8-41d1-988f-c933bb6421ad tempest-ServersTestManualDisk-1753452859 tempest-ServersTestManualDisk-1753452859-project-member] [instance: 3088ee5e-0d27-4058-a94b-2e04c3b52add] Reconfiguring VM instance instance-00000046 to attach disk [datastore2] 3088ee5e-0d27-4058-a94b-2e04c3b52add/3088ee5e-0d27-4058-a94b-2e04c3b52add.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 917.201568] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4595121b-d46c-4bf0-ac09-a152f5ad0c03 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.233157] env[65726]: DEBUG oslo_vmware.api [None req-17b5a1c7-4bc8-41d1-988f-c933bb6421ad tempest-ServersTestManualDisk-1753452859 tempest-ServersTestManualDisk-1753452859-project-member] Waiting for the task: (returnval){ [ 917.233157] env[65726]: value = "task-5116224" [ 917.233157] env[65726]: _type = "Task" [ 917.233157] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.248809] env[65726]: DEBUG oslo_vmware.api [None req-17b5a1c7-4bc8-41d1-988f-c933bb6421ad tempest-ServersTestManualDisk-1753452859 tempest-ServersTestManualDisk-1753452859-project-member] Task: {'id': task-5116224, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.370258] env[65726]: DEBUG oslo_vmware.api [None req-89fee262-eea7-4ed4-83cb-eb9998cc3059 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': task-5116222, 'name': MoveVirtualDisk_Task} progress is 52%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.384417] env[65726]: DEBUG oslo_concurrency.lockutils [None req-edb92956-8e7a-4023-9c72-6f4afa47cb43 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Lock "5a252ef1-93c9-4bff-842b-b64df2bc5d75" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.010s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 917.469381] env[65726]: DEBUG oslo_concurrency.lockutils [req-79681db2-6f01-440b-96ca-ce88a4aab1c5 req-5d23e393-cd62-49fb-b86c-65f48f6f774e service nova] Releasing lock "refresh_cache-3088ee5e-0d27-4058-a94b-2e04c3b52add" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 917.469680] env[65726]: DEBUG nova.compute.manager [req-79681db2-6f01-440b-96ca-ce88a4aab1c5 req-5d23e393-cd62-49fb-b86c-65f48f6f774e service nova] [instance: 76249623-6f83-46a3-b8c5-c001111aa698] Received event network-vif-deleted-e97a5da4-5540-4c0c-a33d-be1182e82aa7 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 917.481819] env[65726]: DEBUG nova.network.neutron [None req-801b9b62-b5bd-4b5f-80fa-32388dd4cb88 tempest-ServerTagsTestJSON-59142761 tempest-ServerTagsTestJSON-59142761-project-member] [instance: 9029549c-1914-4bae-91e2-8812b79051ec] Successfully created port: 3104d757-f32f-4bcb-a4f3-fb6b80f1ad5c {{(pid=65726) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 917.538291] env[65726]: DEBUG nova.compute.manager [None req-801b9b62-b5bd-4b5f-80fa-32388dd4cb88 tempest-ServerTagsTestJSON-59142761 tempest-ServerTagsTestJSON-59142761-project-member] [instance: 9029549c-1914-4bae-91e2-8812b79051ec] Start building block device mappings for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 917.673138] env[65726]: DEBUG nova.network.neutron [None req-2cc737a8-ad7d-449c-b270-9db9e18d971d tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: dcde25df-87a5-47fb-94ce-334f68894e04] Updating instance_info_cache with network_info: [{"id": "bf51c96c-63c5-46de-a40f-58e6769c5fcf", "address": "fa:16:3e:4c:88:87", "network": {"id": "51a66448-0f72-4266-8868-856b48be839f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-155658961-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d46ccb00794f458b85da4a93879139ee", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b1f3e6c3-5584-4852-9017-476ab8ac4946", "external-id": "nsx-vlan-transportzone-304", "segmentation_id": 304, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbf51c96c-63", "ovs_interfaceid": "bf51c96c-63c5-46de-a40f-58e6769c5fcf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 917.763475] env[65726]: DEBUG oslo_vmware.api [None req-17b5a1c7-4bc8-41d1-988f-c933bb6421ad tempest-ServersTestManualDisk-1753452859 tempest-ServersTestManualDisk-1753452859-project-member] Task: {'id': task-5116224, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.873999] env[65726]: DEBUG oslo_vmware.api [None req-89fee262-eea7-4ed4-83cb-eb9998cc3059 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': task-5116222, 'name': MoveVirtualDisk_Task} progress is 74%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.178537] env[65726]: DEBUG oslo_concurrency.lockutils [None req-2cc737a8-ad7d-449c-b270-9db9e18d971d tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Releasing lock "refresh_cache-dcde25df-87a5-47fb-94ce-334f68894e04" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 918.179014] env[65726]: DEBUG nova.compute.manager [None req-2cc737a8-ad7d-449c-b270-9db9e18d971d tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: dcde25df-87a5-47fb-94ce-334f68894e04] Instance network_info: |[{"id": "bf51c96c-63c5-46de-a40f-58e6769c5fcf", "address": "fa:16:3e:4c:88:87", "network": {"id": "51a66448-0f72-4266-8868-856b48be839f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-155658961-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d46ccb00794f458b85da4a93879139ee", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b1f3e6c3-5584-4852-9017-476ab8ac4946", "external-id": "nsx-vlan-transportzone-304", "segmentation_id": 304, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbf51c96c-63", "ovs_interfaceid": "bf51c96c-63c5-46de-a40f-58e6769c5fcf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 918.180288] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-2cc737a8-ad7d-449c-b270-9db9e18d971d tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: dcde25df-87a5-47fb-94ce-334f68894e04] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:4c:88:87', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'b1f3e6c3-5584-4852-9017-476ab8ac4946', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'bf51c96c-63c5-46de-a40f-58e6769c5fcf', 'vif_model': 'vmxnet3'}] {{(pid=65726) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 918.189065] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-2cc737a8-ad7d-449c-b270-9db9e18d971d tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 918.191620] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: dcde25df-87a5-47fb-94ce-334f68894e04] Creating VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 918.191620] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f038fd37-c323-4d19-adcf-5254cef4cb6e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.194335] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a08babce-8e86-4700-93eb-6b92dcf6b470 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.217702] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e45e8680-f783-4235-bd42-190d331c10f5 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.223652] env[65726]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 918.223652] env[65726]: value = "task-5116225" [ 918.223652] env[65726]: _type = "Task" [ 918.223652] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.260195] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f135a06-a8dd-4e16-872d-4b3ae2f0f658 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.266900] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116225, 'name': CreateVM_Task} progress is 10%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.281284] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2430abd3-7259-4d58-b837-4667537e4ef4 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.288592] env[65726]: DEBUG oslo_vmware.api [None req-17b5a1c7-4bc8-41d1-988f-c933bb6421ad tempest-ServersTestManualDisk-1753452859 tempest-ServersTestManualDisk-1753452859-project-member] Task: {'id': task-5116224, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.309273] env[65726]: DEBUG nova.compute.provider_tree [None req-1cae8219-3b3b-4d11-bdce-863de7bcd2d6 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 918.374607] env[65726]: DEBUG oslo_vmware.api [None req-89fee262-eea7-4ed4-83cb-eb9998cc3059 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': task-5116222, 'name': MoveVirtualDisk_Task} progress is 94%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.488890] env[65726]: DEBUG oslo_concurrency.lockutils [None req-edb92956-8e7a-4023-9c72-6f4afa47cb43 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Acquiring lock "5a252ef1-93c9-4bff-842b-b64df2bc5d75" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 918.488890] env[65726]: DEBUG oslo_concurrency.lockutils [None req-edb92956-8e7a-4023-9c72-6f4afa47cb43 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Lock "5a252ef1-93c9-4bff-842b-b64df2bc5d75" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 918.489298] env[65726]: INFO nova.compute.manager [None req-edb92956-8e7a-4023-9c72-6f4afa47cb43 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: 5a252ef1-93c9-4bff-842b-b64df2bc5d75] Attaching volume 2e0dafa1-dc04-4dd8-b30f-16c07c062b44 to /dev/sdb [ 918.534453] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33d96502-803c-4f54-9596-c517abbac0c0 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.544885] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5db8af1c-7edd-49dd-af9b-a5b7d5e2e283 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.549881] env[65726]: DEBUG nova.compute.manager [None req-801b9b62-b5bd-4b5f-80fa-32388dd4cb88 tempest-ServerTagsTestJSON-59142761 tempest-ServerTagsTestJSON-59142761-project-member] [instance: 9029549c-1914-4bae-91e2-8812b79051ec] Start spawning the instance on the hypervisor. {{(pid=65726) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 918.563193] env[65726]: DEBUG nova.virt.block_device [None req-edb92956-8e7a-4023-9c72-6f4afa47cb43 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: 5a252ef1-93c9-4bff-842b-b64df2bc5d75] Updating existing volume attachment record: 7526aa56-eb1e-479f-89f9-636c759395f9 {{(pid=65726) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 918.738369] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116225, 'name': CreateVM_Task} progress is 99%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.770037] env[65726]: DEBUG oslo_vmware.api [None req-17b5a1c7-4bc8-41d1-988f-c933bb6421ad tempest-ServersTestManualDisk-1753452859 tempest-ServersTestManualDisk-1753452859-project-member] Task: {'id': task-5116224, 'name': ReconfigVM_Task, 'duration_secs': 1.12242} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.770037] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-17b5a1c7-4bc8-41d1-988f-c933bb6421ad tempest-ServersTestManualDisk-1753452859 tempest-ServersTestManualDisk-1753452859-project-member] [instance: 3088ee5e-0d27-4058-a94b-2e04c3b52add] Reconfigured VM instance instance-00000046 to attach disk [datastore2] 3088ee5e-0d27-4058-a94b-2e04c3b52add/3088ee5e-0d27-4058-a94b-2e04c3b52add.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 918.770037] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ea570596-2150-4c3e-953f-0698a2c80031 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.777185] env[65726]: DEBUG oslo_vmware.api [None req-17b5a1c7-4bc8-41d1-988f-c933bb6421ad tempest-ServersTestManualDisk-1753452859 tempest-ServersTestManualDisk-1753452859-project-member] Waiting for the task: (returnval){ [ 918.777185] env[65726]: value = "task-5116227" [ 918.777185] env[65726]: _type = "Task" [ 918.777185] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.787457] env[65726]: DEBUG oslo_vmware.api [None req-17b5a1c7-4bc8-41d1-988f-c933bb6421ad tempest-ServersTestManualDisk-1753452859 tempest-ServersTestManualDisk-1753452859-project-member] Task: {'id': task-5116227, 'name': Rename_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.813855] env[65726]: DEBUG nova.scheduler.client.report [None req-1cae8219-3b3b-4d11-bdce-863de7bcd2d6 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 918.854807] env[65726]: DEBUG nova.virt.hardware [None req-801b9b62-b5bd-4b5f-80fa-32388dd4cb88 tempest-ServerTagsTestJSON-59142761 tempest-ServerTagsTestJSON-59142761-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T19:28:34Z,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1539f06fa3534e90acbc3a60c4b8bd3f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T19:28:35Z,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 918.857236] env[65726]: DEBUG nova.virt.hardware [None req-801b9b62-b5bd-4b5f-80fa-32388dd4cb88 tempest-ServerTagsTestJSON-59142761 tempest-ServerTagsTestJSON-59142761-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 918.857906] env[65726]: DEBUG nova.virt.hardware [None req-801b9b62-b5bd-4b5f-80fa-32388dd4cb88 tempest-ServerTagsTestJSON-59142761 tempest-ServerTagsTestJSON-59142761-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 918.859523] env[65726]: DEBUG nova.virt.hardware [None req-801b9b62-b5bd-4b5f-80fa-32388dd4cb88 tempest-ServerTagsTestJSON-59142761 tempest-ServerTagsTestJSON-59142761-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 918.859736] env[65726]: DEBUG nova.virt.hardware [None req-801b9b62-b5bd-4b5f-80fa-32388dd4cb88 tempest-ServerTagsTestJSON-59142761 tempest-ServerTagsTestJSON-59142761-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 918.859893] env[65726]: DEBUG nova.virt.hardware [None req-801b9b62-b5bd-4b5f-80fa-32388dd4cb88 tempest-ServerTagsTestJSON-59142761 tempest-ServerTagsTestJSON-59142761-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 918.860155] env[65726]: DEBUG nova.virt.hardware [None req-801b9b62-b5bd-4b5f-80fa-32388dd4cb88 tempest-ServerTagsTestJSON-59142761 tempest-ServerTagsTestJSON-59142761-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 918.860323] env[65726]: DEBUG nova.virt.hardware [None req-801b9b62-b5bd-4b5f-80fa-32388dd4cb88 tempest-ServerTagsTestJSON-59142761 tempest-ServerTagsTestJSON-59142761-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 918.860494] env[65726]: DEBUG nova.virt.hardware [None req-801b9b62-b5bd-4b5f-80fa-32388dd4cb88 tempest-ServerTagsTestJSON-59142761 tempest-ServerTagsTestJSON-59142761-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 918.860656] env[65726]: DEBUG nova.virt.hardware [None req-801b9b62-b5bd-4b5f-80fa-32388dd4cb88 tempest-ServerTagsTestJSON-59142761 tempest-ServerTagsTestJSON-59142761-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 918.860826] env[65726]: DEBUG nova.virt.hardware [None req-801b9b62-b5bd-4b5f-80fa-32388dd4cb88 tempest-ServerTagsTestJSON-59142761 tempest-ServerTagsTestJSON-59142761-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 918.865055] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ffb498d-bfca-40a5-95b8-9c84fce3ca51 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.887168] env[65726]: DEBUG oslo_vmware.api [None req-89fee262-eea7-4ed4-83cb-eb9998cc3059 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': task-5116222, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.730162} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.888843] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8039e5d7-b551-47e5-93ef-36dd01ff8d9e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.894947] env[65726]: INFO nova.virt.vmwareapi.ds_util [None req-89fee262-eea7-4ed4-83cb-eb9998cc3059 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Moved virtual disk from [datastore1] OSTACK_IMG_1f8b9f24-a5fc-420d-a36b-40a6efa61bbf/OSTACK_IMG_1f8b9f24-a5fc-420d-a36b-40a6efa61bbf.vmdk to [datastore1] devstack-image-cache_base/8ecee2b1-8703-4e91-8a32-dc17932b237e/8ecee2b1-8703-4e91-8a32-dc17932b237e.vmdk. [ 918.894947] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-89fee262-eea7-4ed4-83cb-eb9998cc3059 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 6f91b053-772a-4497-b29d-349b960c55eb] Cleaning up location [datastore1] OSTACK_IMG_1f8b9f24-a5fc-420d-a36b-40a6efa61bbf {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 918.894947] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-89fee262-eea7-4ed4-83cb-eb9998cc3059 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Deleting the datastore file [datastore1] OSTACK_IMG_1f8b9f24-a5fc-420d-a36b-40a6efa61bbf {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 918.897550] env[65726]: DEBUG oslo_vmware.rw_handles [None req-31642568-e231-4328-a969-4d06d89d9576 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/523e089c-0ef7-904e-9b60-5ea26927a04d/disk-0.vmdk. {{(pid=65726) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 918.898840] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6dbd5cb5-a32a-44c8-acae-04af93b3ccd2 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.900854] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-037df06f-05be-4ab9-ae10-5f324eb8cdb7 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.914324] env[65726]: DEBUG oslo_vmware.rw_handles [None req-31642568-e231-4328-a969-4d06d89d9576 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/523e089c-0ef7-904e-9b60-5ea26927a04d/disk-0.vmdk is in state: ready. {{(pid=65726) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 918.914526] env[65726]: ERROR oslo_vmware.rw_handles [None req-31642568-e231-4328-a969-4d06d89d9576 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/523e089c-0ef7-904e-9b60-5ea26927a04d/disk-0.vmdk due to incomplete transfer. [ 918.924493] env[65726]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-2393f814-9d50-4912-8a18-c47f04aeef9d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.927116] env[65726]: DEBUG oslo_vmware.api [None req-89fee262-eea7-4ed4-83cb-eb9998cc3059 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Waiting for the task: (returnval){ [ 918.927116] env[65726]: value = "task-5116230" [ 918.927116] env[65726]: _type = "Task" [ 918.927116] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.940378] env[65726]: DEBUG oslo_vmware.api [None req-89fee262-eea7-4ed4-83cb-eb9998cc3059 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': task-5116230, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.944556] env[65726]: DEBUG oslo_vmware.rw_handles [None req-31642568-e231-4328-a969-4d06d89d9576 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/523e089c-0ef7-904e-9b60-5ea26927a04d/disk-0.vmdk. {{(pid=65726) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 918.944556] env[65726]: DEBUG nova.virt.vmwareapi.images [None req-31642568-e231-4328-a969-4d06d89d9576 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] [instance: c4cc4f85-cb35-4edc-a58b-adfee0ce1265] Uploaded image fa1d1620-a4ff-4711-acf9-5d56f6b9b63a to the Glance image server {{(pid=65726) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 918.946302] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-31642568-e231-4328-a969-4d06d89d9576 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] [instance: c4cc4f85-cb35-4edc-a58b-adfee0ce1265] Destroying the VM {{(pid=65726) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 918.947250] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-7acaea1e-cbfc-4e17-ac6f-a0a443581b62 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.954019] env[65726]: DEBUG oslo_vmware.api [None req-31642568-e231-4328-a969-4d06d89d9576 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Waiting for the task: (returnval){ [ 918.954019] env[65726]: value = "task-5116231" [ 918.954019] env[65726]: _type = "Task" [ 918.954019] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.964131] env[65726]: DEBUG oslo_vmware.api [None req-31642568-e231-4328-a969-4d06d89d9576 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Task: {'id': task-5116231, 'name': Destroy_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.236599] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116225, 'name': CreateVM_Task, 'duration_secs': 0.578055} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.236818] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: dcde25df-87a5-47fb-94ce-334f68894e04] Created VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 919.239467] env[65726]: WARNING openstack [None req-2cc737a8-ad7d-449c-b270-9db9e18d971d tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 919.239467] env[65726]: WARNING openstack [None req-2cc737a8-ad7d-449c-b270-9db9e18d971d tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 919.246481] env[65726]: DEBUG oslo_concurrency.lockutils [None req-2cc737a8-ad7d-449c-b270-9db9e18d971d tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 919.246738] env[65726]: DEBUG oslo_concurrency.lockutils [None req-2cc737a8-ad7d-449c-b270-9db9e18d971d tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 919.247185] env[65726]: DEBUG oslo_concurrency.lockutils [None req-2cc737a8-ad7d-449c-b270-9db9e18d971d tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 919.247569] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-167d4784-2230-498f-a7da-db9432c358f2 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.255729] env[65726]: DEBUG oslo_vmware.api [None req-2cc737a8-ad7d-449c-b270-9db9e18d971d tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Waiting for the task: (returnval){ [ 919.255729] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]528f9dc2-3d44-b0c8-54a4-ecd539836346" [ 919.255729] env[65726]: _type = "Task" [ 919.255729] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.268684] env[65726]: DEBUG oslo_vmware.api [None req-2cc737a8-ad7d-449c-b270-9db9e18d971d tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]528f9dc2-3d44-b0c8-54a4-ecd539836346, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.270851] env[65726]: DEBUG nova.network.neutron [None req-801b9b62-b5bd-4b5f-80fa-32388dd4cb88 tempest-ServerTagsTestJSON-59142761 tempest-ServerTagsTestJSON-59142761-project-member] [instance: 9029549c-1914-4bae-91e2-8812b79051ec] Successfully updated port: 3104d757-f32f-4bcb-a4f3-fb6b80f1ad5c {{(pid=65726) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 919.289189] env[65726]: DEBUG oslo_vmware.api [None req-17b5a1c7-4bc8-41d1-988f-c933bb6421ad tempest-ServersTestManualDisk-1753452859 tempest-ServersTestManualDisk-1753452859-project-member] Task: {'id': task-5116227, 'name': Rename_Task, 'duration_secs': 0.173661} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.290584] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-17b5a1c7-4bc8-41d1-988f-c933bb6421ad tempest-ServersTestManualDisk-1753452859 tempest-ServersTestManualDisk-1753452859-project-member] [instance: 3088ee5e-0d27-4058-a94b-2e04c3b52add] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 919.290882] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e5c2d367-d8ca-44a1-bc45-e5b0be454688 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.299512] env[65726]: DEBUG oslo_vmware.api [None req-17b5a1c7-4bc8-41d1-988f-c933bb6421ad tempest-ServersTestManualDisk-1753452859 tempest-ServersTestManualDisk-1753452859-project-member] Waiting for the task: (returnval){ [ 919.299512] env[65726]: value = "task-5116232" [ 919.299512] env[65726]: _type = "Task" [ 919.299512] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.309945] env[65726]: DEBUG oslo_vmware.api [None req-17b5a1c7-4bc8-41d1-988f-c933bb6421ad tempest-ServersTestManualDisk-1753452859 tempest-ServersTestManualDisk-1753452859-project-member] Task: {'id': task-5116232, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.321410] env[65726]: DEBUG oslo_concurrency.lockutils [None req-1cae8219-3b3b-4d11-bdce-863de7bcd2d6 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.269s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 919.324380] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9a0984e5-cc6d-419e-a851-17965236c39e tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 27.273s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 919.326300] env[65726]: INFO nova.compute.claims [None req-9a0984e5-cc6d-419e-a851-17965236c39e tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: 060646cc-f0fb-4e6c-9f67-7b2daa6b6fd6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 919.354641] env[65726]: INFO nova.scheduler.client.report [None req-1cae8219-3b3b-4d11-bdce-863de7bcd2d6 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] Deleted allocations for instance 2cbee359-a1fc-45c0-bcc5-a2c24ded9c69 [ 919.439579] env[65726]: DEBUG oslo_vmware.api [None req-89fee262-eea7-4ed4-83cb-eb9998cc3059 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': task-5116230, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.069727} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.439830] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-89fee262-eea7-4ed4-83cb-eb9998cc3059 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Deleted the datastore file {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 919.439992] env[65726]: DEBUG oslo_concurrency.lockutils [None req-89fee262-eea7-4ed4-83cb-eb9998cc3059 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8ecee2b1-8703-4e91-8a32-dc17932b237e/8ecee2b1-8703-4e91-8a32-dc17932b237e.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 919.440255] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-89fee262-eea7-4ed4-83cb-eb9998cc3059 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8ecee2b1-8703-4e91-8a32-dc17932b237e/8ecee2b1-8703-4e91-8a32-dc17932b237e.vmdk to [datastore1] 6f91b053-772a-4497-b29d-349b960c55eb/6f91b053-772a-4497-b29d-349b960c55eb.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 919.440514] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-93860458-ef32-4f24-9c32-a81aef1a0ee9 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.451091] env[65726]: DEBUG oslo_vmware.api [None req-89fee262-eea7-4ed4-83cb-eb9998cc3059 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Waiting for the task: (returnval){ [ 919.451091] env[65726]: value = "task-5116233" [ 919.451091] env[65726]: _type = "Task" [ 919.451091] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.464832] env[65726]: DEBUG oslo_vmware.api [None req-89fee262-eea7-4ed4-83cb-eb9998cc3059 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': task-5116233, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.468609] env[65726]: DEBUG oslo_vmware.api [None req-31642568-e231-4328-a969-4d06d89d9576 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Task: {'id': task-5116231, 'name': Destroy_Task} progress is 100%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.624294] env[65726]: DEBUG nova.compute.manager [req-61cef4b2-f6e3-411e-8762-2c6b540da4cb req-90bb357a-04b1-4256-813c-a03a64a80917 service nova] [instance: dcde25df-87a5-47fb-94ce-334f68894e04] Received event network-changed-bf51c96c-63c5-46de-a40f-58e6769c5fcf {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 919.624529] env[65726]: DEBUG nova.compute.manager [req-61cef4b2-f6e3-411e-8762-2c6b540da4cb req-90bb357a-04b1-4256-813c-a03a64a80917 service nova] [instance: dcde25df-87a5-47fb-94ce-334f68894e04] Refreshing instance network info cache due to event network-changed-bf51c96c-63c5-46de-a40f-58e6769c5fcf. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 919.624688] env[65726]: DEBUG oslo_concurrency.lockutils [req-61cef4b2-f6e3-411e-8762-2c6b540da4cb req-90bb357a-04b1-4256-813c-a03a64a80917 service nova] Acquiring lock "refresh_cache-dcde25df-87a5-47fb-94ce-334f68894e04" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 919.624823] env[65726]: DEBUG oslo_concurrency.lockutils [req-61cef4b2-f6e3-411e-8762-2c6b540da4cb req-90bb357a-04b1-4256-813c-a03a64a80917 service nova] Acquired lock "refresh_cache-dcde25df-87a5-47fb-94ce-334f68894e04" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 919.624973] env[65726]: DEBUG nova.network.neutron [req-61cef4b2-f6e3-411e-8762-2c6b540da4cb req-90bb357a-04b1-4256-813c-a03a64a80917 service nova] [instance: dcde25df-87a5-47fb-94ce-334f68894e04] Refreshing network info cache for port bf51c96c-63c5-46de-a40f-58e6769c5fcf {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 919.630526] env[65726]: DEBUG nova.compute.manager [req-9b45281a-d1e0-4fd6-9572-2b19f4337dd8 req-4f710651-78f8-4c3f-a954-9043f8bcd153 service nova] [instance: 4ea2b70b-0cfe-4fbf-817a-baa028b8372a] Received event network-vif-deleted-a97348ff-93e0-442b-8a41-b56fa04e8cb6 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 919.630526] env[65726]: DEBUG nova.compute.manager [req-9b45281a-d1e0-4fd6-9572-2b19f4337dd8 req-4f710651-78f8-4c3f-a954-9043f8bcd153 service nova] [instance: 70375746-06dd-4710-9136-95e9e5759a0e] Received event network-vif-deleted-7818914d-352b-408a-9d0f-4c2bb1a25f2e {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 919.767561] env[65726]: DEBUG oslo_vmware.api [None req-2cc737a8-ad7d-449c-b270-9db9e18d971d tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]528f9dc2-3d44-b0c8-54a4-ecd539836346, 'name': SearchDatastore_Task, 'duration_secs': 0.013893} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.767871] env[65726]: DEBUG oslo_concurrency.lockutils [None req-2cc737a8-ad7d-449c-b270-9db9e18d971d tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 919.768138] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-2cc737a8-ad7d-449c-b270-9db9e18d971d tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: dcde25df-87a5-47fb-94ce-334f68894e04] Processing image b52362a3-ee8a-4cbf-b06f-513b0cc8f95c {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 919.768371] env[65726]: DEBUG oslo_concurrency.lockutils [None req-2cc737a8-ad7d-449c-b270-9db9e18d971d tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 919.768505] env[65726]: DEBUG oslo_concurrency.lockutils [None req-2cc737a8-ad7d-449c-b270-9db9e18d971d tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 919.768677] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-2cc737a8-ad7d-449c-b270-9db9e18d971d tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 919.771041] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-cda6fd6e-c165-4c1e-9003-60a030e4472b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.778284] env[65726]: DEBUG oslo_concurrency.lockutils [None req-801b9b62-b5bd-4b5f-80fa-32388dd4cb88 tempest-ServerTagsTestJSON-59142761 tempest-ServerTagsTestJSON-59142761-project-member] Acquiring lock "refresh_cache-9029549c-1914-4bae-91e2-8812b79051ec" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 919.778284] env[65726]: DEBUG oslo_concurrency.lockutils [None req-801b9b62-b5bd-4b5f-80fa-32388dd4cb88 tempest-ServerTagsTestJSON-59142761 tempest-ServerTagsTestJSON-59142761-project-member] Acquired lock "refresh_cache-9029549c-1914-4bae-91e2-8812b79051ec" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 919.778284] env[65726]: DEBUG nova.network.neutron [None req-801b9b62-b5bd-4b5f-80fa-32388dd4cb88 tempest-ServerTagsTestJSON-59142761 tempest-ServerTagsTestJSON-59142761-project-member] [instance: 9029549c-1914-4bae-91e2-8812b79051ec] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 919.781977] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-2cc737a8-ad7d-449c-b270-9db9e18d971d tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 919.782324] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-2cc737a8-ad7d-449c-b270-9db9e18d971d tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65726) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 919.784031] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-14ec023f-6432-4f33-b030-f981c192ef57 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.798917] env[65726]: DEBUG oslo_vmware.api [None req-2cc737a8-ad7d-449c-b270-9db9e18d971d tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Waiting for the task: (returnval){ [ 919.798917] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52df8ccc-42c0-db11-6294-ee2ca6a78db4" [ 919.798917] env[65726]: _type = "Task" [ 919.798917] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.814682] env[65726]: DEBUG oslo_vmware.api [None req-17b5a1c7-4bc8-41d1-988f-c933bb6421ad tempest-ServersTestManualDisk-1753452859 tempest-ServersTestManualDisk-1753452859-project-member] Task: {'id': task-5116232, 'name': PowerOnVM_Task} progress is 88%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.819828] env[65726]: DEBUG oslo_vmware.api [None req-2cc737a8-ad7d-449c-b270-9db9e18d971d tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52df8ccc-42c0-db11-6294-ee2ca6a78db4, 'name': SearchDatastore_Task, 'duration_secs': 0.013869} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.820064] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-16ab2894-dcec-4c08-9704-d2690ac03420 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.836210] env[65726]: DEBUG oslo_vmware.api [None req-2cc737a8-ad7d-449c-b270-9db9e18d971d tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Waiting for the task: (returnval){ [ 919.836210] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52613c1d-f498-e223-c686-06f7a8fab9a0" [ 919.836210] env[65726]: _type = "Task" [ 919.836210] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.853973] env[65726]: DEBUG oslo_vmware.api [None req-2cc737a8-ad7d-449c-b270-9db9e18d971d tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52613c1d-f498-e223-c686-06f7a8fab9a0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.867606] env[65726]: DEBUG oslo_concurrency.lockutils [None req-1cae8219-3b3b-4d11-bdce-863de7bcd2d6 tempest-ServerShowV254Test-1467304924 tempest-ServerShowV254Test-1467304924-project-member] Lock "2cbee359-a1fc-45c0-bcc5-a2c24ded9c69" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 34.007s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 919.972394] env[65726]: DEBUG oslo_vmware.api [None req-31642568-e231-4328-a969-4d06d89d9576 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Task: {'id': task-5116231, 'name': Destroy_Task} progress is 100%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.977211] env[65726]: DEBUG oslo_vmware.api [None req-89fee262-eea7-4ed4-83cb-eb9998cc3059 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': task-5116233, 'name': CopyVirtualDisk_Task} progress is 12%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.129891] env[65726]: WARNING openstack [req-61cef4b2-f6e3-411e-8762-2c6b540da4cb req-90bb357a-04b1-4256-813c-a03a64a80917 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 920.130493] env[65726]: WARNING openstack [req-61cef4b2-f6e3-411e-8762-2c6b540da4cb req-90bb357a-04b1-4256-813c-a03a64a80917 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 920.287025] env[65726]: WARNING openstack [None req-801b9b62-b5bd-4b5f-80fa-32388dd4cb88 tempest-ServerTagsTestJSON-59142761 tempest-ServerTagsTestJSON-59142761-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 920.287025] env[65726]: WARNING openstack [None req-801b9b62-b5bd-4b5f-80fa-32388dd4cb88 tempest-ServerTagsTestJSON-59142761 tempest-ServerTagsTestJSON-59142761-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 920.294177] env[65726]: DEBUG nova.network.neutron [None req-801b9b62-b5bd-4b5f-80fa-32388dd4cb88 tempest-ServerTagsTestJSON-59142761 tempest-ServerTagsTestJSON-59142761-project-member] [instance: 9029549c-1914-4bae-91e2-8812b79051ec] Instance cache missing network info. {{(pid=65726) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 920.326652] env[65726]: DEBUG oslo_vmware.api [None req-17b5a1c7-4bc8-41d1-988f-c933bb6421ad tempest-ServersTestManualDisk-1753452859 tempest-ServersTestManualDisk-1753452859-project-member] Task: {'id': task-5116232, 'name': PowerOnVM_Task, 'duration_secs': 0.876164} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.326964] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-17b5a1c7-4bc8-41d1-988f-c933bb6421ad tempest-ServersTestManualDisk-1753452859 tempest-ServersTestManualDisk-1753452859-project-member] [instance: 3088ee5e-0d27-4058-a94b-2e04c3b52add] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 920.327256] env[65726]: INFO nova.compute.manager [None req-17b5a1c7-4bc8-41d1-988f-c933bb6421ad tempest-ServersTestManualDisk-1753452859 tempest-ServersTestManualDisk-1753452859-project-member] [instance: 3088ee5e-0d27-4058-a94b-2e04c3b52add] Took 9.47 seconds to spawn the instance on the hypervisor. [ 920.327515] env[65726]: DEBUG nova.compute.manager [None req-17b5a1c7-4bc8-41d1-988f-c933bb6421ad tempest-ServersTestManualDisk-1753452859 tempest-ServersTestManualDisk-1753452859-project-member] [instance: 3088ee5e-0d27-4058-a94b-2e04c3b52add] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 920.328571] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc1fbf58-a6c0-4611-a638-0127c45d5f1a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.372781] env[65726]: DEBUG oslo_vmware.api [None req-2cc737a8-ad7d-449c-b270-9db9e18d971d tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52613c1d-f498-e223-c686-06f7a8fab9a0, 'name': SearchDatastore_Task, 'duration_secs': 0.087669} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.373261] env[65726]: DEBUG oslo_concurrency.lockutils [None req-2cc737a8-ad7d-449c-b270-9db9e18d971d tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 920.373683] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-2cc737a8-ad7d-449c-b270-9db9e18d971d tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] dcde25df-87a5-47fb-94ce-334f68894e04/dcde25df-87a5-47fb-94ce-334f68894e04.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 920.374264] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8971ba2c-17db-47f8-9f9d-21f834a8509a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.401052] env[65726]: DEBUG oslo_vmware.api [None req-2cc737a8-ad7d-449c-b270-9db9e18d971d tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Waiting for the task: (returnval){ [ 920.401052] env[65726]: value = "task-5116234" [ 920.401052] env[65726]: _type = "Task" [ 920.401052] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.421731] env[65726]: DEBUG oslo_vmware.api [None req-2cc737a8-ad7d-449c-b270-9db9e18d971d tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': task-5116234, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.473931] env[65726]: DEBUG oslo_vmware.api [None req-31642568-e231-4328-a969-4d06d89d9576 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Task: {'id': task-5116231, 'name': Destroy_Task} progress is 100%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.478317] env[65726]: DEBUG oslo_vmware.api [None req-89fee262-eea7-4ed4-83cb-eb9998cc3059 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': task-5116233, 'name': CopyVirtualDisk_Task} progress is 32%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.485178] env[65726]: WARNING openstack [None req-801b9b62-b5bd-4b5f-80fa-32388dd4cb88 tempest-ServerTagsTestJSON-59142761 tempest-ServerTagsTestJSON-59142761-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 920.485590] env[65726]: WARNING openstack [None req-801b9b62-b5bd-4b5f-80fa-32388dd4cb88 tempest-ServerTagsTestJSON-59142761 tempest-ServerTagsTestJSON-59142761-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 920.576193] env[65726]: WARNING openstack [req-61cef4b2-f6e3-411e-8762-2c6b540da4cb req-90bb357a-04b1-4256-813c-a03a64a80917 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 920.576337] env[65726]: WARNING openstack [req-61cef4b2-f6e3-411e-8762-2c6b540da4cb req-90bb357a-04b1-4256-813c-a03a64a80917 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 920.707557] env[65726]: WARNING openstack [None req-801b9b62-b5bd-4b5f-80fa-32388dd4cb88 tempest-ServerTagsTestJSON-59142761 tempest-ServerTagsTestJSON-59142761-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 920.708216] env[65726]: WARNING openstack [None req-801b9b62-b5bd-4b5f-80fa-32388dd4cb88 tempest-ServerTagsTestJSON-59142761 tempest-ServerTagsTestJSON-59142761-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 920.851342] env[65726]: WARNING openstack [req-61cef4b2-f6e3-411e-8762-2c6b540da4cb req-90bb357a-04b1-4256-813c-a03a64a80917 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 920.851536] env[65726]: WARNING openstack [req-61cef4b2-f6e3-411e-8762-2c6b540da4cb req-90bb357a-04b1-4256-813c-a03a64a80917 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 920.875578] env[65726]: INFO nova.compute.manager [None req-17b5a1c7-4bc8-41d1-988f-c933bb6421ad tempest-ServersTestManualDisk-1753452859 tempest-ServersTestManualDisk-1753452859-project-member] [instance: 3088ee5e-0d27-4058-a94b-2e04c3b52add] Took 49.62 seconds to build instance. [ 920.919381] env[65726]: DEBUG oslo_vmware.api [None req-2cc737a8-ad7d-449c-b270-9db9e18d971d tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': task-5116234, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.988293] env[65726]: DEBUG oslo_vmware.api [None req-31642568-e231-4328-a969-4d06d89d9576 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Task: {'id': task-5116231, 'name': Destroy_Task, 'duration_secs': 1.546506} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.991986] env[65726]: INFO nova.virt.vmwareapi.vm_util [None req-31642568-e231-4328-a969-4d06d89d9576 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] [instance: c4cc4f85-cb35-4edc-a58b-adfee0ce1265] Destroyed the VM [ 920.992164] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-31642568-e231-4328-a969-4d06d89d9576 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] [instance: c4cc4f85-cb35-4edc-a58b-adfee0ce1265] Deleting Snapshot of the VM instance {{(pid=65726) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 920.992411] env[65726]: DEBUG oslo_vmware.api [None req-89fee262-eea7-4ed4-83cb-eb9998cc3059 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': task-5116233, 'name': CopyVirtualDisk_Task} progress is 54%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.996178] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-0abfd311-aa53-45bc-87df-d7e58adafb2e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.010537] env[65726]: DEBUG oslo_vmware.api [None req-31642568-e231-4328-a969-4d06d89d9576 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Waiting for the task: (returnval){ [ 921.010537] env[65726]: value = "task-5116236" [ 921.010537] env[65726]: _type = "Task" [ 921.010537] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.013041] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc05c63a-7f76-448c-9de5-cfe9e5a83a3d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.029927] env[65726]: DEBUG oslo_vmware.api [None req-31642568-e231-4328-a969-4d06d89d9576 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Task: {'id': task-5116236, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.035260] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e50e531-4b59-4e1e-8f44-b453306ca881 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.042865] env[65726]: DEBUG nova.network.neutron [None req-801b9b62-b5bd-4b5f-80fa-32388dd4cb88 tempest-ServerTagsTestJSON-59142761 tempest-ServerTagsTestJSON-59142761-project-member] [instance: 9029549c-1914-4bae-91e2-8812b79051ec] Updating instance_info_cache with network_info: [{"id": "3104d757-f32f-4bcb-a4f3-fb6b80f1ad5c", "address": "fa:16:3e:d7:5b:f6", "network": {"id": "23c16258-986d-4240-96bc-f43c98a10035", "bridge": "br-int", "label": "tempest-ServerTagsTestJSON-451731892-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b635b43ad6824e99a609092196631b58", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8cbc9b8f-ce19-4262-bf4d-88cd4f259a1c", "external-id": "nsx-vlan-transportzone-630", "segmentation_id": 630, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3104d757-f3", "ovs_interfaceid": "3104d757-f32f-4bcb-a4f3-fb6b80f1ad5c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 921.082829] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5558a209-6378-49c6-93c7-035f52d1d604 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.096198] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19947d19-3e7e-419a-816c-98cb29615ba5 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.118681] env[65726]: DEBUG nova.compute.provider_tree [None req-9a0984e5-cc6d-419e-a851-17965236c39e tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 921.151841] env[65726]: DEBUG nova.network.neutron [req-61cef4b2-f6e3-411e-8762-2c6b540da4cb req-90bb357a-04b1-4256-813c-a03a64a80917 service nova] [instance: dcde25df-87a5-47fb-94ce-334f68894e04] Updated VIF entry in instance network info cache for port bf51c96c-63c5-46de-a40f-58e6769c5fcf. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 921.152964] env[65726]: DEBUG nova.network.neutron [req-61cef4b2-f6e3-411e-8762-2c6b540da4cb req-90bb357a-04b1-4256-813c-a03a64a80917 service nova] [instance: dcde25df-87a5-47fb-94ce-334f68894e04] Updating instance_info_cache with network_info: [{"id": "bf51c96c-63c5-46de-a40f-58e6769c5fcf", "address": "fa:16:3e:4c:88:87", "network": {"id": "51a66448-0f72-4266-8868-856b48be839f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-155658961-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d46ccb00794f458b85da4a93879139ee", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b1f3e6c3-5584-4852-9017-476ab8ac4946", "external-id": "nsx-vlan-transportzone-304", "segmentation_id": 304, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbf51c96c-63", "ovs_interfaceid": "bf51c96c-63c5-46de-a40f-58e6769c5fcf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 921.380165] env[65726]: DEBUG oslo_concurrency.lockutils [None req-17b5a1c7-4bc8-41d1-988f-c933bb6421ad tempest-ServersTestManualDisk-1753452859 tempest-ServersTestManualDisk-1753452859-project-member] Lock "3088ee5e-0d27-4058-a94b-2e04c3b52add" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 63.697s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 921.417979] env[65726]: DEBUG oslo_vmware.api [None req-2cc737a8-ad7d-449c-b270-9db9e18d971d tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': task-5116234, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.477606] env[65726]: DEBUG oslo_vmware.api [None req-89fee262-eea7-4ed4-83cb-eb9998cc3059 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': task-5116233, 'name': CopyVirtualDisk_Task} progress is 74%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.528740] env[65726]: DEBUG oslo_vmware.api [None req-31642568-e231-4328-a969-4d06d89d9576 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Task: {'id': task-5116236, 'name': RemoveSnapshot_Task} progress is 66%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.545996] env[65726]: DEBUG oslo_concurrency.lockutils [None req-801b9b62-b5bd-4b5f-80fa-32388dd4cb88 tempest-ServerTagsTestJSON-59142761 tempest-ServerTagsTestJSON-59142761-project-member] Releasing lock "refresh_cache-9029549c-1914-4bae-91e2-8812b79051ec" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 921.546228] env[65726]: DEBUG nova.compute.manager [None req-801b9b62-b5bd-4b5f-80fa-32388dd4cb88 tempest-ServerTagsTestJSON-59142761 tempest-ServerTagsTestJSON-59142761-project-member] [instance: 9029549c-1914-4bae-91e2-8812b79051ec] Instance network_info: |[{"id": "3104d757-f32f-4bcb-a4f3-fb6b80f1ad5c", "address": "fa:16:3e:d7:5b:f6", "network": {"id": "23c16258-986d-4240-96bc-f43c98a10035", "bridge": "br-int", "label": "tempest-ServerTagsTestJSON-451731892-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b635b43ad6824e99a609092196631b58", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8cbc9b8f-ce19-4262-bf4d-88cd4f259a1c", "external-id": "nsx-vlan-transportzone-630", "segmentation_id": 630, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3104d757-f3", "ovs_interfaceid": "3104d757-f32f-4bcb-a4f3-fb6b80f1ad5c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 921.546731] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-801b9b62-b5bd-4b5f-80fa-32388dd4cb88 tempest-ServerTagsTestJSON-59142761 tempest-ServerTagsTestJSON-59142761-project-member] [instance: 9029549c-1914-4bae-91e2-8812b79051ec] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d7:5b:f6', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '8cbc9b8f-ce19-4262-bf4d-88cd4f259a1c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3104d757-f32f-4bcb-a4f3-fb6b80f1ad5c', 'vif_model': 'vmxnet3'}] {{(pid=65726) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 921.558513] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-801b9b62-b5bd-4b5f-80fa-32388dd4cb88 tempest-ServerTagsTestJSON-59142761 tempest-ServerTagsTestJSON-59142761-project-member] Creating folder: Project (b635b43ad6824e99a609092196631b58). Parent ref: group-v995008. {{(pid=65726) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 921.559396] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-dd4ebe8b-3f63-4594-b021-cd28987b6ef5 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.579876] env[65726]: INFO nova.virt.vmwareapi.vm_util [None req-801b9b62-b5bd-4b5f-80fa-32388dd4cb88 tempest-ServerTagsTestJSON-59142761 tempest-ServerTagsTestJSON-59142761-project-member] Created folder: Project (b635b43ad6824e99a609092196631b58) in parent group-v995008. [ 921.580042] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-801b9b62-b5bd-4b5f-80fa-32388dd4cb88 tempest-ServerTagsTestJSON-59142761 tempest-ServerTagsTestJSON-59142761-project-member] Creating folder: Instances. Parent ref: group-v995210. {{(pid=65726) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 921.580390] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-8f94e093-69cc-4350-9895-570f08a156c0 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.599731] env[65726]: INFO nova.virt.vmwareapi.vm_util [None req-801b9b62-b5bd-4b5f-80fa-32388dd4cb88 tempest-ServerTagsTestJSON-59142761 tempest-ServerTagsTestJSON-59142761-project-member] Created folder: Instances in parent group-v995210. [ 921.600315] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-801b9b62-b5bd-4b5f-80fa-32388dd4cb88 tempest-ServerTagsTestJSON-59142761 tempest-ServerTagsTestJSON-59142761-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 921.600692] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9029549c-1914-4bae-91e2-8812b79051ec] Creating VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 921.601064] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-298b13ba-7cdf-4430-b2dc-a2494982d840 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.622395] env[65726]: DEBUG nova.scheduler.client.report [None req-9a0984e5-cc6d-419e-a851-17965236c39e tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 921.631547] env[65726]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 921.631547] env[65726]: value = "task-5116239" [ 921.631547] env[65726]: _type = "Task" [ 921.631547] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.649912] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116239, 'name': CreateVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.656927] env[65726]: DEBUG oslo_concurrency.lockutils [None req-371355d3-22ed-4e9f-8bfc-a310d98eb1b4 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Acquiring lock "df46141c-b545-4e03-b3a3-fd9f5feda0d2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 921.657532] env[65726]: DEBUG oslo_concurrency.lockutils [None req-371355d3-22ed-4e9f-8bfc-a310d98eb1b4 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Lock "df46141c-b545-4e03-b3a3-fd9f5feda0d2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 921.659160] env[65726]: DEBUG oslo_concurrency.lockutils [req-61cef4b2-f6e3-411e-8762-2c6b540da4cb req-90bb357a-04b1-4256-813c-a03a64a80917 service nova] Releasing lock "refresh_cache-dcde25df-87a5-47fb-94ce-334f68894e04" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 921.921780] env[65726]: DEBUG oslo_vmware.api [None req-2cc737a8-ad7d-449c-b270-9db9e18d971d tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': task-5116234, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.979132] env[65726]: DEBUG oslo_vmware.api [None req-89fee262-eea7-4ed4-83cb-eb9998cc3059 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': task-5116233, 'name': CopyVirtualDisk_Task} progress is 94%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.991065] env[65726]: DEBUG nova.compute.manager [req-28a050aa-a772-454f-800b-80505084b022 req-38a54af4-c92f-4c16-aa52-870050abcc93 service nova] [instance: 9029549c-1914-4bae-91e2-8812b79051ec] Received event network-vif-plugged-3104d757-f32f-4bcb-a4f3-fb6b80f1ad5c {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 921.991345] env[65726]: DEBUG oslo_concurrency.lockutils [req-28a050aa-a772-454f-800b-80505084b022 req-38a54af4-c92f-4c16-aa52-870050abcc93 service nova] Acquiring lock "9029549c-1914-4bae-91e2-8812b79051ec-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 921.992826] env[65726]: DEBUG oslo_concurrency.lockutils [req-28a050aa-a772-454f-800b-80505084b022 req-38a54af4-c92f-4c16-aa52-870050abcc93 service nova] Lock "9029549c-1914-4bae-91e2-8812b79051ec-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 921.993026] env[65726]: DEBUG oslo_concurrency.lockutils [req-28a050aa-a772-454f-800b-80505084b022 req-38a54af4-c92f-4c16-aa52-870050abcc93 service nova] Lock "9029549c-1914-4bae-91e2-8812b79051ec-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 921.993296] env[65726]: DEBUG nova.compute.manager [req-28a050aa-a772-454f-800b-80505084b022 req-38a54af4-c92f-4c16-aa52-870050abcc93 service nova] [instance: 9029549c-1914-4bae-91e2-8812b79051ec] No waiting events found dispatching network-vif-plugged-3104d757-f32f-4bcb-a4f3-fb6b80f1ad5c {{(pid=65726) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 921.993478] env[65726]: WARNING nova.compute.manager [req-28a050aa-a772-454f-800b-80505084b022 req-38a54af4-c92f-4c16-aa52-870050abcc93 service nova] [instance: 9029549c-1914-4bae-91e2-8812b79051ec] Received unexpected event network-vif-plugged-3104d757-f32f-4bcb-a4f3-fb6b80f1ad5c for instance with vm_state building and task_state spawning. [ 921.993637] env[65726]: DEBUG nova.compute.manager [req-28a050aa-a772-454f-800b-80505084b022 req-38a54af4-c92f-4c16-aa52-870050abcc93 service nova] [instance: 9029549c-1914-4bae-91e2-8812b79051ec] Received event network-changed-3104d757-f32f-4bcb-a4f3-fb6b80f1ad5c {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 921.993817] env[65726]: DEBUG nova.compute.manager [req-28a050aa-a772-454f-800b-80505084b022 req-38a54af4-c92f-4c16-aa52-870050abcc93 service nova] [instance: 9029549c-1914-4bae-91e2-8812b79051ec] Refreshing instance network info cache due to event network-changed-3104d757-f32f-4bcb-a4f3-fb6b80f1ad5c. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 921.993967] env[65726]: DEBUG oslo_concurrency.lockutils [req-28a050aa-a772-454f-800b-80505084b022 req-38a54af4-c92f-4c16-aa52-870050abcc93 service nova] Acquiring lock "refresh_cache-9029549c-1914-4bae-91e2-8812b79051ec" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 921.994155] env[65726]: DEBUG oslo_concurrency.lockutils [req-28a050aa-a772-454f-800b-80505084b022 req-38a54af4-c92f-4c16-aa52-870050abcc93 service nova] Acquired lock "refresh_cache-9029549c-1914-4bae-91e2-8812b79051ec" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 921.994351] env[65726]: DEBUG nova.network.neutron [req-28a050aa-a772-454f-800b-80505084b022 req-38a54af4-c92f-4c16-aa52-870050abcc93 service nova] [instance: 9029549c-1914-4bae-91e2-8812b79051ec] Refreshing network info cache for port 3104d757-f32f-4bcb-a4f3-fb6b80f1ad5c {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 922.032322] env[65726]: DEBUG oslo_vmware.api [None req-31642568-e231-4328-a969-4d06d89d9576 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Task: {'id': task-5116236, 'name': RemoveSnapshot_Task, 'duration_secs': 0.792844} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.034551] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-31642568-e231-4328-a969-4d06d89d9576 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] [instance: c4cc4f85-cb35-4edc-a58b-adfee0ce1265] Deleted Snapshot of the VM instance {{(pid=65726) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 922.034551] env[65726]: INFO nova.compute.manager [None req-31642568-e231-4328-a969-4d06d89d9576 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] [instance: c4cc4f85-cb35-4edc-a58b-adfee0ce1265] Took 17.20 seconds to snapshot the instance on the hypervisor. [ 922.132763] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9a0984e5-cc6d-419e-a851-17965236c39e tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.806s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 922.132763] env[65726]: DEBUG nova.compute.manager [None req-9a0984e5-cc6d-419e-a851-17965236c39e tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: 060646cc-f0fb-4e6c-9f67-7b2daa6b6fd6] Start building networks asynchronously for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 922.136062] env[65726]: DEBUG oslo_concurrency.lockutils [None req-294000af-a13c-4aa5-b362-e66461f19340 tempest-ServersV294TestFqdnHostnames-819811677 tempest-ServersV294TestFqdnHostnames-819811677-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 28.322s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 922.136545] env[65726]: DEBUG nova.objects.instance [None req-294000af-a13c-4aa5-b362-e66461f19340 tempest-ServersV294TestFqdnHostnames-819811677 tempest-ServersV294TestFqdnHostnames-819811677-project-member] Lazy-loading 'resources' on Instance uuid c8be64a3-1bfa-41a3-b9be-68a7e6e052e0 {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 922.152478] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116239, 'name': CreateVM_Task} progress is 25%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.163768] env[65726]: DEBUG nova.compute.manager [None req-371355d3-22ed-4e9f-8bfc-a310d98eb1b4 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] [instance: df46141c-b545-4e03-b3a3-fd9f5feda0d2] Starting instance... {{(pid=65726) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 922.430365] env[65726]: DEBUG oslo_vmware.api [None req-2cc737a8-ad7d-449c-b270-9db9e18d971d tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': task-5116234, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.471399] env[65726]: DEBUG oslo_vmware.api [None req-89fee262-eea7-4ed4-83cb-eb9998cc3059 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': task-5116233, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.688896} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.471758] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-89fee262-eea7-4ed4-83cb-eb9998cc3059 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8ecee2b1-8703-4e91-8a32-dc17932b237e/8ecee2b1-8703-4e91-8a32-dc17932b237e.vmdk to [datastore1] 6f91b053-772a-4497-b29d-349b960c55eb/6f91b053-772a-4497-b29d-349b960c55eb.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 922.473174] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afa810cd-ebc4-4858-b0d9-54d4999e9ba0 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.505255] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-89fee262-eea7-4ed4-83cb-eb9998cc3059 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 6f91b053-772a-4497-b29d-349b960c55eb] Reconfiguring VM instance instance-00000045 to attach disk [datastore1] 6f91b053-772a-4497-b29d-349b960c55eb/6f91b053-772a-4497-b29d-349b960c55eb.vmdk or device None with type streamOptimized {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 922.506191] env[65726]: WARNING openstack [req-28a050aa-a772-454f-800b-80505084b022 req-38a54af4-c92f-4c16-aa52-870050abcc93 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 922.506568] env[65726]: WARNING openstack [req-28a050aa-a772-454f-800b-80505084b022 req-38a54af4-c92f-4c16-aa52-870050abcc93 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 922.514275] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-49521930-9932-41c4-8c01-0c85e187e5ec {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.542629] env[65726]: DEBUG oslo_vmware.api [None req-89fee262-eea7-4ed4-83cb-eb9998cc3059 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Waiting for the task: (returnval){ [ 922.542629] env[65726]: value = "task-5116240" [ 922.542629] env[65726]: _type = "Task" [ 922.542629] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.558040] env[65726]: DEBUG oslo_vmware.api [None req-89fee262-eea7-4ed4-83cb-eb9998cc3059 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': task-5116240, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.642612] env[65726]: DEBUG nova.compute.utils [None req-9a0984e5-cc6d-419e-a851-17965236c39e tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Using /dev/sd instead of None {{(pid=65726) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 922.648188] env[65726]: DEBUG nova.compute.manager [None req-9a0984e5-cc6d-419e-a851-17965236c39e tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: 060646cc-f0fb-4e6c-9f67-7b2daa6b6fd6] Allocating IP information in the background. {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 922.648506] env[65726]: DEBUG nova.network.neutron [None req-9a0984e5-cc6d-419e-a851-17965236c39e tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: 060646cc-f0fb-4e6c-9f67-7b2daa6b6fd6] allocate_for_instance() {{(pid=65726) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 922.648956] env[65726]: WARNING neutronclient.v2_0.client [None req-9a0984e5-cc6d-419e-a851-17965236c39e tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 922.649392] env[65726]: WARNING neutronclient.v2_0.client [None req-9a0984e5-cc6d-419e-a851-17965236c39e tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 922.649987] env[65726]: WARNING openstack [None req-9a0984e5-cc6d-419e-a851-17965236c39e tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 922.650522] env[65726]: WARNING openstack [None req-9a0984e5-cc6d-419e-a851-17965236c39e tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 922.682677] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116239, 'name': CreateVM_Task} progress is 99%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.704031] env[65726]: DEBUG oslo_concurrency.lockutils [None req-371355d3-22ed-4e9f-8bfc-a310d98eb1b4 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 922.772871] env[65726]: DEBUG nova.policy [None req-9a0984e5-cc6d-419e-a851-17965236c39e tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f3050294b8f34a0f97d10b038b048779', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a090d553766847e58e3231b966c92565', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65726) authorize /opt/stack/nova/nova/policy.py:192}} [ 922.926012] env[65726]: DEBUG oslo_vmware.api [None req-2cc737a8-ad7d-449c-b270-9db9e18d971d tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': task-5116234, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.149955} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.929915] env[65726]: WARNING openstack [req-28a050aa-a772-454f-800b-80505084b022 req-38a54af4-c92f-4c16-aa52-870050abcc93 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 922.930382] env[65726]: WARNING openstack [req-28a050aa-a772-454f-800b-80505084b022 req-38a54af4-c92f-4c16-aa52-870050abcc93 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 922.942171] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-2cc737a8-ad7d-449c-b270-9db9e18d971d tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] dcde25df-87a5-47fb-94ce-334f68894e04/dcde25df-87a5-47fb-94ce-334f68894e04.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 922.942171] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-2cc737a8-ad7d-449c-b270-9db9e18d971d tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: dcde25df-87a5-47fb-94ce-334f68894e04] Extending root virtual disk to 1048576 {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 922.942171] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-34c1db9d-38b3-4a18-bf7d-b6f0b5abf7aa {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.953156] env[65726]: DEBUG oslo_vmware.api [None req-2cc737a8-ad7d-449c-b270-9db9e18d971d tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Waiting for the task: (returnval){ [ 922.953156] env[65726]: value = "task-5116241" [ 922.953156] env[65726]: _type = "Task" [ 922.953156] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.967362] env[65726]: DEBUG oslo_vmware.api [None req-2cc737a8-ad7d-449c-b270-9db9e18d971d tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': task-5116241, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.061908] env[65726]: DEBUG oslo_vmware.api [None req-89fee262-eea7-4ed4-83cb-eb9998cc3059 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': task-5116240, 'name': ReconfigVM_Task, 'duration_secs': 0.335554} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.065687] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-89fee262-eea7-4ed4-83cb-eb9998cc3059 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 6f91b053-772a-4497-b29d-349b960c55eb] Reconfigured VM instance instance-00000045 to attach disk [datastore1] 6f91b053-772a-4497-b29d-349b960c55eb/6f91b053-772a-4497-b29d-349b960c55eb.vmdk or device None with type streamOptimized {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 923.070821] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-783c82f0-3cc6-452c-852e-f363b2d52704 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.076594] env[65726]: DEBUG oslo_vmware.api [None req-89fee262-eea7-4ed4-83cb-eb9998cc3059 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Waiting for the task: (returnval){ [ 923.076594] env[65726]: value = "task-5116242" [ 923.076594] env[65726]: _type = "Task" [ 923.076594] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.089248] env[65726]: DEBUG oslo_vmware.api [None req-89fee262-eea7-4ed4-83cb-eb9998cc3059 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': task-5116242, 'name': Rename_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.122506] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-edb92956-8e7a-4023-9c72-6f4afa47cb43 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: 5a252ef1-93c9-4bff-842b-b64df2bc5d75] Volume attach. Driver type: vmdk {{(pid=65726) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 923.122782] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-edb92956-8e7a-4023-9c72-6f4afa47cb43 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: 5a252ef1-93c9-4bff-842b-b64df2bc5d75] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-995209', 'volume_id': '2e0dafa1-dc04-4dd8-b30f-16c07c062b44', 'name': 'volume-2e0dafa1-dc04-4dd8-b30f-16c07c062b44', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '5a252ef1-93c9-4bff-842b-b64df2bc5d75', 'attached_at': '', 'detached_at': '', 'volume_id': '2e0dafa1-dc04-4dd8-b30f-16c07c062b44', 'serial': '2e0dafa1-dc04-4dd8-b30f-16c07c062b44'} {{(pid=65726) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 923.123896] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58c9c1e0-2db8-41f2-be8f-fc5107c26499 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.161136] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12df75d2-5b6f-4108-a353-a934b559d72f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.165710] env[65726]: DEBUG nova.compute.manager [None req-9a0984e5-cc6d-419e-a851-17965236c39e tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: 060646cc-f0fb-4e6c-9f67-7b2daa6b6fd6] Start building block device mappings for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 923.172342] env[65726]: DEBUG nova.network.neutron [None req-9a0984e5-cc6d-419e-a851-17965236c39e tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: 060646cc-f0fb-4e6c-9f67-7b2daa6b6fd6] Successfully created port: f923178f-ee20-4582-a2e8-6bfd87020d4b {{(pid=65726) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 923.179582] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116239, 'name': CreateVM_Task, 'duration_secs': 1.14189} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.195696] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9029549c-1914-4bae-91e2-8812b79051ec] Created VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 923.202656] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-edb92956-8e7a-4023-9c72-6f4afa47cb43 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: 5a252ef1-93c9-4bff-842b-b64df2bc5d75] Reconfiguring VM instance instance-00000044 to attach disk [datastore1] volume-2e0dafa1-dc04-4dd8-b30f-16c07c062b44/volume-2e0dafa1-dc04-4dd8-b30f-16c07c062b44.vmdk or device None with type thin {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 923.206399] env[65726]: WARNING openstack [None req-801b9b62-b5bd-4b5f-80fa-32388dd4cb88 tempest-ServerTagsTestJSON-59142761 tempest-ServerTagsTestJSON-59142761-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 923.206777] env[65726]: WARNING openstack [None req-801b9b62-b5bd-4b5f-80fa-32388dd4cb88 tempest-ServerTagsTestJSON-59142761 tempest-ServerTagsTestJSON-59142761-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 923.215116] env[65726]: DEBUG oslo_concurrency.lockutils [None req-801b9b62-b5bd-4b5f-80fa-32388dd4cb88 tempest-ServerTagsTestJSON-59142761 tempest-ServerTagsTestJSON-59142761-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 923.215497] env[65726]: DEBUG oslo_concurrency.lockutils [None req-801b9b62-b5bd-4b5f-80fa-32388dd4cb88 tempest-ServerTagsTestJSON-59142761 tempest-ServerTagsTestJSON-59142761-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 923.215982] env[65726]: DEBUG oslo_concurrency.lockutils [None req-801b9b62-b5bd-4b5f-80fa-32388dd4cb88 tempest-ServerTagsTestJSON-59142761 tempest-ServerTagsTestJSON-59142761-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 923.218088] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4a8ae524-765b-4311-b670-2b42cfbaa86a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.244733] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-928dbe58-4592-4d66-9baa-ab676b971196 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.252505] env[65726]: DEBUG oslo_vmware.api [None req-801b9b62-b5bd-4b5f-80fa-32388dd4cb88 tempest-ServerTagsTestJSON-59142761 tempest-ServerTagsTestJSON-59142761-project-member] Waiting for the task: (returnval){ [ 923.252505] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]527d6394-1d81-68a8-67bc-ef8b73905370" [ 923.252505] env[65726]: _type = "Task" [ 923.252505] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.258231] env[65726]: DEBUG oslo_vmware.api [None req-edb92956-8e7a-4023-9c72-6f4afa47cb43 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Waiting for the task: (returnval){ [ 923.258231] env[65726]: value = "task-5116243" [ 923.258231] env[65726]: _type = "Task" [ 923.258231] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.270470] env[65726]: DEBUG oslo_vmware.api [None req-801b9b62-b5bd-4b5f-80fa-32388dd4cb88 tempest-ServerTagsTestJSON-59142761 tempest-ServerTagsTestJSON-59142761-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]527d6394-1d81-68a8-67bc-ef8b73905370, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.282791] env[65726]: DEBUG oslo_vmware.api [None req-edb92956-8e7a-4023-9c72-6f4afa47cb43 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': task-5116243, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.396527] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e12bf6c-ea3a-4f90-9c42-03a8bd982570 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.410190] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2df0f085-75d5-4a4e-8daf-6eec2f40d24a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.444720] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eaaaa5d9-3c12-4ca3-a01b-3e0ace97d9e2 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.453758] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2177ddae-50b0-4cbf-b65a-ee3f913c77f1 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.468959] env[65726]: DEBUG oslo_vmware.api [None req-2cc737a8-ad7d-449c-b270-9db9e18d971d tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': task-5116241, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.095744} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.479521] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-2cc737a8-ad7d-449c-b270-9db9e18d971d tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: dcde25df-87a5-47fb-94ce-334f68894e04] Extended root virtual disk {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 923.480126] env[65726]: DEBUG nova.compute.provider_tree [None req-294000af-a13c-4aa5-b362-e66461f19340 tempest-ServersV294TestFqdnHostnames-819811677 tempest-ServersV294TestFqdnHostnames-819811677-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 923.485206] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63e2042f-189f-4db2-a65c-cf7d93281751 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.506087] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-2cc737a8-ad7d-449c-b270-9db9e18d971d tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: dcde25df-87a5-47fb-94ce-334f68894e04] Reconfiguring VM instance instance-00000047 to attach disk [datastore1] dcde25df-87a5-47fb-94ce-334f68894e04/dcde25df-87a5-47fb-94ce-334f68894e04.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 923.507141] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a9cbb54e-d59b-449f-82f7-99168d45f9f5 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.528459] env[65726]: DEBUG oslo_vmware.api [None req-2cc737a8-ad7d-449c-b270-9db9e18d971d tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Waiting for the task: (returnval){ [ 923.528459] env[65726]: value = "task-5116244" [ 923.528459] env[65726]: _type = "Task" [ 923.528459] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.538082] env[65726]: DEBUG oslo_vmware.api [None req-2cc737a8-ad7d-449c-b270-9db9e18d971d tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': task-5116244, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.589711] env[65726]: DEBUG oslo_vmware.api [None req-89fee262-eea7-4ed4-83cb-eb9998cc3059 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': task-5116242, 'name': Rename_Task, 'duration_secs': 0.194919} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.590084] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-89fee262-eea7-4ed4-83cb-eb9998cc3059 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 6f91b053-772a-4497-b29d-349b960c55eb] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 923.590265] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a19a8d38-22b8-4286-8aa6-92c2bfedfe2c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.599188] env[65726]: DEBUG oslo_vmware.api [None req-89fee262-eea7-4ed4-83cb-eb9998cc3059 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Waiting for the task: (returnval){ [ 923.599188] env[65726]: value = "task-5116245" [ 923.599188] env[65726]: _type = "Task" [ 923.599188] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.608581] env[65726]: DEBUG oslo_vmware.api [None req-89fee262-eea7-4ed4-83cb-eb9998cc3059 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': task-5116245, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.767659] env[65726]: DEBUG oslo_vmware.api [None req-801b9b62-b5bd-4b5f-80fa-32388dd4cb88 tempest-ServerTagsTestJSON-59142761 tempest-ServerTagsTestJSON-59142761-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]527d6394-1d81-68a8-67bc-ef8b73905370, 'name': SearchDatastore_Task, 'duration_secs': 0.020055} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.768426] env[65726]: DEBUG oslo_concurrency.lockutils [None req-801b9b62-b5bd-4b5f-80fa-32388dd4cb88 tempest-ServerTagsTestJSON-59142761 tempest-ServerTagsTestJSON-59142761-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 923.768649] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-801b9b62-b5bd-4b5f-80fa-32388dd4cb88 tempest-ServerTagsTestJSON-59142761 tempest-ServerTagsTestJSON-59142761-project-member] [instance: 9029549c-1914-4bae-91e2-8812b79051ec] Processing image b52362a3-ee8a-4cbf-b06f-513b0cc8f95c {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 923.768877] env[65726]: DEBUG oslo_concurrency.lockutils [None req-801b9b62-b5bd-4b5f-80fa-32388dd4cb88 tempest-ServerTagsTestJSON-59142761 tempest-ServerTagsTestJSON-59142761-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 923.769088] env[65726]: DEBUG oslo_concurrency.lockutils [None req-801b9b62-b5bd-4b5f-80fa-32388dd4cb88 tempest-ServerTagsTestJSON-59142761 tempest-ServerTagsTestJSON-59142761-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 923.769257] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-801b9b62-b5bd-4b5f-80fa-32388dd4cb88 tempest-ServerTagsTestJSON-59142761 tempest-ServerTagsTestJSON-59142761-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 923.769528] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c982ae2b-1921-45b6-99a0-4b35605b0a81 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.774897] env[65726]: DEBUG oslo_vmware.api [None req-edb92956-8e7a-4023-9c72-6f4afa47cb43 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': task-5116243, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.784453] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-801b9b62-b5bd-4b5f-80fa-32388dd4cb88 tempest-ServerTagsTestJSON-59142761 tempest-ServerTagsTestJSON-59142761-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 923.784725] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-801b9b62-b5bd-4b5f-80fa-32388dd4cb88 tempest-ServerTagsTestJSON-59142761 tempest-ServerTagsTestJSON-59142761-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65726) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 923.785626] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a9b9d7f5-e340-470f-a0a5-2b610b79e822 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.794360] env[65726]: WARNING openstack [req-28a050aa-a772-454f-800b-80505084b022 req-38a54af4-c92f-4c16-aa52-870050abcc93 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 923.794875] env[65726]: WARNING openstack [req-28a050aa-a772-454f-800b-80505084b022 req-38a54af4-c92f-4c16-aa52-870050abcc93 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 923.806267] env[65726]: DEBUG oslo_vmware.api [None req-801b9b62-b5bd-4b5f-80fa-32388dd4cb88 tempest-ServerTagsTestJSON-59142761 tempest-ServerTagsTestJSON-59142761-project-member] Waiting for the task: (returnval){ [ 923.806267] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]524feb1d-792c-d14e-75ca-6e4d483c4e49" [ 923.806267] env[65726]: _type = "Task" [ 923.806267] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.817452] env[65726]: DEBUG oslo_vmware.api [None req-801b9b62-b5bd-4b5f-80fa-32388dd4cb88 tempest-ServerTagsTestJSON-59142761 tempest-ServerTagsTestJSON-59142761-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]524feb1d-792c-d14e-75ca-6e4d483c4e49, 'name': SearchDatastore_Task, 'duration_secs': 0.016147} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.818349] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-871853ad-2884-41ce-b999-ae32e1d6ddbd {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.825943] env[65726]: DEBUG oslo_vmware.api [None req-801b9b62-b5bd-4b5f-80fa-32388dd4cb88 tempest-ServerTagsTestJSON-59142761 tempest-ServerTagsTestJSON-59142761-project-member] Waiting for the task: (returnval){ [ 923.825943] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52eb3296-90f3-93ea-6502-d4864d33eb73" [ 923.825943] env[65726]: _type = "Task" [ 923.825943] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.835943] env[65726]: DEBUG oslo_vmware.api [None req-801b9b62-b5bd-4b5f-80fa-32388dd4cb88 tempest-ServerTagsTestJSON-59142761 tempest-ServerTagsTestJSON-59142761-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52eb3296-90f3-93ea-6502-d4864d33eb73, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.986749] env[65726]: DEBUG nova.scheduler.client.report [None req-294000af-a13c-4aa5-b362-e66461f19340 tempest-ServersV294TestFqdnHostnames-819811677 tempest-ServersV294TestFqdnHostnames-819811677-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 924.040586] env[65726]: DEBUG oslo_vmware.api [None req-2cc737a8-ad7d-449c-b270-9db9e18d971d tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': task-5116244, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.110698] env[65726]: DEBUG oslo_vmware.api [None req-89fee262-eea7-4ed4-83cb-eb9998cc3059 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': task-5116245, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.184644] env[65726]: DEBUG nova.compute.manager [None req-9a0984e5-cc6d-419e-a851-17965236c39e tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: 060646cc-f0fb-4e6c-9f67-7b2daa6b6fd6] Start spawning the instance on the hypervisor. {{(pid=65726) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 924.232144] env[65726]: DEBUG nova.virt.hardware [None req-9a0984e5-cc6d-419e-a851-17965236c39e tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T19:28:34Z,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1539f06fa3534e90acbc3a60c4b8bd3f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T19:28:35Z,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 924.232505] env[65726]: DEBUG nova.virt.hardware [None req-9a0984e5-cc6d-419e-a851-17965236c39e tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 924.232677] env[65726]: DEBUG nova.virt.hardware [None req-9a0984e5-cc6d-419e-a851-17965236c39e tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 924.232858] env[65726]: DEBUG nova.virt.hardware [None req-9a0984e5-cc6d-419e-a851-17965236c39e tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 924.233038] env[65726]: DEBUG nova.virt.hardware [None req-9a0984e5-cc6d-419e-a851-17965236c39e tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 924.233779] env[65726]: DEBUG nova.virt.hardware [None req-9a0984e5-cc6d-419e-a851-17965236c39e tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 924.234388] env[65726]: DEBUG nova.virt.hardware [None req-9a0984e5-cc6d-419e-a851-17965236c39e tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 924.234884] env[65726]: DEBUG nova.virt.hardware [None req-9a0984e5-cc6d-419e-a851-17965236c39e tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 924.235209] env[65726]: DEBUG nova.virt.hardware [None req-9a0984e5-cc6d-419e-a851-17965236c39e tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 924.235498] env[65726]: DEBUG nova.virt.hardware [None req-9a0984e5-cc6d-419e-a851-17965236c39e tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 924.235795] env[65726]: DEBUG nova.virt.hardware [None req-9a0984e5-cc6d-419e-a851-17965236c39e tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 924.237485] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48363f33-9fdd-4073-bfa6-6e3a41cbab15 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.248131] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd99324c-ad50-47b2-b94f-1e29646605ee {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.279665] env[65726]: DEBUG oslo_vmware.api [None req-edb92956-8e7a-4023-9c72-6f4afa47cb43 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': task-5116243, 'name': ReconfigVM_Task, 'duration_secs': 0.712331} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.280795] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-edb92956-8e7a-4023-9c72-6f4afa47cb43 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: 5a252ef1-93c9-4bff-842b-b64df2bc5d75] Reconfigured VM instance instance-00000044 to attach disk [datastore1] volume-2e0dafa1-dc04-4dd8-b30f-16c07c062b44/volume-2e0dafa1-dc04-4dd8-b30f-16c07c062b44.vmdk or device None with type thin {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 924.285598] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d6836c14-832d-4f9d-98b3-a021f35fd2e2 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.304845] env[65726]: DEBUG oslo_vmware.api [None req-edb92956-8e7a-4023-9c72-6f4afa47cb43 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Waiting for the task: (returnval){ [ 924.304845] env[65726]: value = "task-5116246" [ 924.304845] env[65726]: _type = "Task" [ 924.304845] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.315858] env[65726]: DEBUG oslo_vmware.api [None req-edb92956-8e7a-4023-9c72-6f4afa47cb43 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': task-5116246, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.341270] env[65726]: DEBUG oslo_vmware.api [None req-801b9b62-b5bd-4b5f-80fa-32388dd4cb88 tempest-ServerTagsTestJSON-59142761 tempest-ServerTagsTestJSON-59142761-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52eb3296-90f3-93ea-6502-d4864d33eb73, 'name': SearchDatastore_Task, 'duration_secs': 0.011406} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.341270] env[65726]: DEBUG oslo_concurrency.lockutils [None req-801b9b62-b5bd-4b5f-80fa-32388dd4cb88 tempest-ServerTagsTestJSON-59142761 tempest-ServerTagsTestJSON-59142761-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 924.341270] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-801b9b62-b5bd-4b5f-80fa-32388dd4cb88 tempest-ServerTagsTestJSON-59142761 tempest-ServerTagsTestJSON-59142761-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] 9029549c-1914-4bae-91e2-8812b79051ec/9029549c-1914-4bae-91e2-8812b79051ec.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 924.341270] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8c4286ef-3892-4513-a66f-5d5cf6411276 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.350128] env[65726]: DEBUG oslo_vmware.api [None req-801b9b62-b5bd-4b5f-80fa-32388dd4cb88 tempest-ServerTagsTestJSON-59142761 tempest-ServerTagsTestJSON-59142761-project-member] Waiting for the task: (returnval){ [ 924.350128] env[65726]: value = "task-5116247" [ 924.350128] env[65726]: _type = "Task" [ 924.350128] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.360022] env[65726]: DEBUG oslo_vmware.api [None req-801b9b62-b5bd-4b5f-80fa-32388dd4cb88 tempest-ServerTagsTestJSON-59142761 tempest-ServerTagsTestJSON-59142761-project-member] Task: {'id': task-5116247, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.494825] env[65726]: DEBUG oslo_concurrency.lockutils [None req-294000af-a13c-4aa5-b362-e66461f19340 tempest-ServersV294TestFqdnHostnames-819811677 tempest-ServersV294TestFqdnHostnames-819811677-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.359s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 924.498547] env[65726]: DEBUG oslo_concurrency.lockutils [None req-73ce0e73-1afb-4566-995f-f51b4749346e tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 27.301s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 924.498870] env[65726]: DEBUG nova.objects.instance [None req-73ce0e73-1afb-4566-995f-f51b4749346e tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Lazy-loading 'resources' on Instance uuid cb8855b6-589d-4863-b86f-7fb4e30fb29e {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 924.524833] env[65726]: INFO nova.scheduler.client.report [None req-294000af-a13c-4aa5-b362-e66461f19340 tempest-ServersV294TestFqdnHostnames-819811677 tempest-ServersV294TestFqdnHostnames-819811677-project-member] Deleted allocations for instance c8be64a3-1bfa-41a3-b9be-68a7e6e052e0 [ 924.538728] env[65726]: DEBUG oslo_vmware.api [None req-2cc737a8-ad7d-449c-b270-9db9e18d971d tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': task-5116244, 'name': ReconfigVM_Task, 'duration_secs': 0.517354} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.539018] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-2cc737a8-ad7d-449c-b270-9db9e18d971d tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: dcde25df-87a5-47fb-94ce-334f68894e04] Reconfigured VM instance instance-00000047 to attach disk [datastore1] dcde25df-87a5-47fb-94ce-334f68894e04/dcde25df-87a5-47fb-94ce-334f68894e04.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 924.539709] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1fbe4b43-5f23-4d28-a33d-517e92cf02e6 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.547121] env[65726]: DEBUG oslo_vmware.api [None req-2cc737a8-ad7d-449c-b270-9db9e18d971d tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Waiting for the task: (returnval){ [ 924.547121] env[65726]: value = "task-5116248" [ 924.547121] env[65726]: _type = "Task" [ 924.547121] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.556816] env[65726]: DEBUG oslo_vmware.api [None req-2cc737a8-ad7d-449c-b270-9db9e18d971d tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': task-5116248, 'name': Rename_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.614597] env[65726]: DEBUG oslo_vmware.api [None req-89fee262-eea7-4ed4-83cb-eb9998cc3059 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': task-5116245, 'name': PowerOnVM_Task, 'duration_secs': 0.664655} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.615061] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-89fee262-eea7-4ed4-83cb-eb9998cc3059 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 6f91b053-772a-4497-b29d-349b960c55eb] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 924.615382] env[65726]: INFO nova.compute.manager [None req-89fee262-eea7-4ed4-83cb-eb9998cc3059 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 6f91b053-772a-4497-b29d-349b960c55eb] Took 16.09 seconds to spawn the instance on the hypervisor. [ 924.615685] env[65726]: DEBUG nova.compute.manager [None req-89fee262-eea7-4ed4-83cb-eb9998cc3059 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 6f91b053-772a-4497-b29d-349b960c55eb] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 924.616972] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2627311e-0ca7-4ca6-b371-47de133a5dcf {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.824219] env[65726]: DEBUG oslo_vmware.api [None req-edb92956-8e7a-4023-9c72-6f4afa47cb43 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': task-5116246, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.873319] env[65726]: DEBUG oslo_vmware.api [None req-801b9b62-b5bd-4b5f-80fa-32388dd4cb88 tempest-ServerTagsTestJSON-59142761 tempest-ServerTagsTestJSON-59142761-project-member] Task: {'id': task-5116247, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.919225] env[65726]: DEBUG nova.network.neutron [None req-9a0984e5-cc6d-419e-a851-17965236c39e tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: 060646cc-f0fb-4e6c-9f67-7b2daa6b6fd6] Successfully updated port: f923178f-ee20-4582-a2e8-6bfd87020d4b {{(pid=65726) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 924.932641] env[65726]: DEBUG nova.network.neutron [req-28a050aa-a772-454f-800b-80505084b022 req-38a54af4-c92f-4c16-aa52-870050abcc93 service nova] [instance: 9029549c-1914-4bae-91e2-8812b79051ec] Updated VIF entry in instance network info cache for port 3104d757-f32f-4bcb-a4f3-fb6b80f1ad5c. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 924.933021] env[65726]: DEBUG nova.network.neutron [req-28a050aa-a772-454f-800b-80505084b022 req-38a54af4-c92f-4c16-aa52-870050abcc93 service nova] [instance: 9029549c-1914-4bae-91e2-8812b79051ec] Updating instance_info_cache with network_info: [{"id": "3104d757-f32f-4bcb-a4f3-fb6b80f1ad5c", "address": "fa:16:3e:d7:5b:f6", "network": {"id": "23c16258-986d-4240-96bc-f43c98a10035", "bridge": "br-int", "label": "tempest-ServerTagsTestJSON-451731892-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b635b43ad6824e99a609092196631b58", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8cbc9b8f-ce19-4262-bf4d-88cd4f259a1c", "external-id": "nsx-vlan-transportzone-630", "segmentation_id": 630, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3104d757-f3", "ovs_interfaceid": "3104d757-f32f-4bcb-a4f3-fb6b80f1ad5c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 925.046471] env[65726]: DEBUG oslo_concurrency.lockutils [None req-294000af-a13c-4aa5-b362-e66461f19340 tempest-ServersV294TestFqdnHostnames-819811677 tempest-ServersV294TestFqdnHostnames-819811677-project-member] Lock "c8be64a3-1bfa-41a3-b9be-68a7e6e052e0" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 34.986s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 925.059259] env[65726]: DEBUG oslo_vmware.api [None req-2cc737a8-ad7d-449c-b270-9db9e18d971d tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': task-5116248, 'name': Rename_Task, 'duration_secs': 0.503892} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.059583] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-2cc737a8-ad7d-449c-b270-9db9e18d971d tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: dcde25df-87a5-47fb-94ce-334f68894e04] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 925.059851] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-51f7163b-f47d-471c-9547-c7f69c7f0b36 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.071162] env[65726]: DEBUG oslo_vmware.api [None req-2cc737a8-ad7d-449c-b270-9db9e18d971d tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Waiting for the task: (returnval){ [ 925.071162] env[65726]: value = "task-5116249" [ 925.071162] env[65726]: _type = "Task" [ 925.071162] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.083377] env[65726]: DEBUG oslo_vmware.api [None req-2cc737a8-ad7d-449c-b270-9db9e18d971d tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': task-5116249, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.139406] env[65726]: INFO nova.compute.manager [None req-89fee262-eea7-4ed4-83cb-eb9998cc3059 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 6f91b053-772a-4497-b29d-349b960c55eb] Took 54.65 seconds to build instance. [ 925.323473] env[65726]: DEBUG oslo_vmware.api [None req-edb92956-8e7a-4023-9c72-6f4afa47cb43 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': task-5116246, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.368021] env[65726]: DEBUG oslo_vmware.api [None req-801b9b62-b5bd-4b5f-80fa-32388dd4cb88 tempest-ServerTagsTestJSON-59142761 tempest-ServerTagsTestJSON-59142761-project-member] Task: {'id': task-5116247, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.681914} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.371089] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-801b9b62-b5bd-4b5f-80fa-32388dd4cb88 tempest-ServerTagsTestJSON-59142761 tempest-ServerTagsTestJSON-59142761-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] 9029549c-1914-4bae-91e2-8812b79051ec/9029549c-1914-4bae-91e2-8812b79051ec.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 925.371353] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-801b9b62-b5bd-4b5f-80fa-32388dd4cb88 tempest-ServerTagsTestJSON-59142761 tempest-ServerTagsTestJSON-59142761-project-member] [instance: 9029549c-1914-4bae-91e2-8812b79051ec] Extending root virtual disk to 1048576 {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 925.371911] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b0bc1288-dd9b-4224-b3c0-7f48048f914b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.380606] env[65726]: DEBUG oslo_vmware.api [None req-801b9b62-b5bd-4b5f-80fa-32388dd4cb88 tempest-ServerTagsTestJSON-59142761 tempest-ServerTagsTestJSON-59142761-project-member] Waiting for the task: (returnval){ [ 925.380606] env[65726]: value = "task-5116250" [ 925.380606] env[65726]: _type = "Task" [ 925.380606] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.393586] env[65726]: DEBUG oslo_vmware.api [None req-801b9b62-b5bd-4b5f-80fa-32388dd4cb88 tempest-ServerTagsTestJSON-59142761 tempest-ServerTagsTestJSON-59142761-project-member] Task: {'id': task-5116250, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.424892] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9a0984e5-cc6d-419e-a851-17965236c39e tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Acquiring lock "refresh_cache-060646cc-f0fb-4e6c-9f67-7b2daa6b6fd6" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 925.426157] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9a0984e5-cc6d-419e-a851-17965236c39e tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Acquired lock "refresh_cache-060646cc-f0fb-4e6c-9f67-7b2daa6b6fd6" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 925.426157] env[65726]: DEBUG nova.network.neutron [None req-9a0984e5-cc6d-419e-a851-17965236c39e tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: 060646cc-f0fb-4e6c-9f67-7b2daa6b6fd6] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 925.435613] env[65726]: DEBUG oslo_concurrency.lockutils [req-28a050aa-a772-454f-800b-80505084b022 req-38a54af4-c92f-4c16-aa52-870050abcc93 service nova] Releasing lock "refresh_cache-9029549c-1914-4bae-91e2-8812b79051ec" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 925.554052] env[65726]: DEBUG nova.compute.manager [req-77ad4d82-3aef-4dfa-9940-ecf112cbe33d req-5dbb92c0-14f2-44f7-a825-5d359a95ed4e service nova] [instance: 3088ee5e-0d27-4058-a94b-2e04c3b52add] Received event network-changed-2ffc16a0-351c-4f5b-8a4a-993ae7adb4f3 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 925.554262] env[65726]: DEBUG nova.compute.manager [req-77ad4d82-3aef-4dfa-9940-ecf112cbe33d req-5dbb92c0-14f2-44f7-a825-5d359a95ed4e service nova] [instance: 3088ee5e-0d27-4058-a94b-2e04c3b52add] Refreshing instance network info cache due to event network-changed-2ffc16a0-351c-4f5b-8a4a-993ae7adb4f3. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 925.554469] env[65726]: DEBUG oslo_concurrency.lockutils [req-77ad4d82-3aef-4dfa-9940-ecf112cbe33d req-5dbb92c0-14f2-44f7-a825-5d359a95ed4e service nova] Acquiring lock "refresh_cache-3088ee5e-0d27-4058-a94b-2e04c3b52add" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 925.554602] env[65726]: DEBUG oslo_concurrency.lockutils [req-77ad4d82-3aef-4dfa-9940-ecf112cbe33d req-5dbb92c0-14f2-44f7-a825-5d359a95ed4e service nova] Acquired lock "refresh_cache-3088ee5e-0d27-4058-a94b-2e04c3b52add" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 925.554757] env[65726]: DEBUG nova.network.neutron [req-77ad4d82-3aef-4dfa-9940-ecf112cbe33d req-5dbb92c0-14f2-44f7-a825-5d359a95ed4e service nova] [instance: 3088ee5e-0d27-4058-a94b-2e04c3b52add] Refreshing network info cache for port 2ffc16a0-351c-4f5b-8a4a-993ae7adb4f3 {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 925.588751] env[65726]: DEBUG oslo_vmware.api [None req-2cc737a8-ad7d-449c-b270-9db9e18d971d tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': task-5116249, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.614187] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0ae65d3-58d1-48f6-93c1-031b9d9ba14a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.624483] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8aae824-f6b2-4f0a-b62b-0ebe8417499b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.660403] env[65726]: DEBUG oslo_concurrency.lockutils [None req-89fee262-eea7-4ed4-83cb-eb9998cc3059 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Lock "6f91b053-772a-4497-b29d-349b960c55eb" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 69.877s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 925.662584] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17f61185-038e-4d37-9b21-f82668248014 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.672584] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60d52cb7-ac63-4c1f-b5a9-3fbdbd7717d9 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.690459] env[65726]: DEBUG nova.compute.provider_tree [None req-73ce0e73-1afb-4566-995f-f51b4749346e tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 925.823175] env[65726]: DEBUG oslo_vmware.api [None req-edb92956-8e7a-4023-9c72-6f4afa47cb43 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': task-5116246, 'name': ReconfigVM_Task, 'duration_secs': 1.194437} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.823623] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-edb92956-8e7a-4023-9c72-6f4afa47cb43 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: 5a252ef1-93c9-4bff-842b-b64df2bc5d75] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-995209', 'volume_id': '2e0dafa1-dc04-4dd8-b30f-16c07c062b44', 'name': 'volume-2e0dafa1-dc04-4dd8-b30f-16c07c062b44', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '5a252ef1-93c9-4bff-842b-b64df2bc5d75', 'attached_at': '', 'detached_at': '', 'volume_id': '2e0dafa1-dc04-4dd8-b30f-16c07c062b44', 'serial': '2e0dafa1-dc04-4dd8-b30f-16c07c062b44'} {{(pid=65726) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 925.892793] env[65726]: DEBUG oslo_vmware.api [None req-801b9b62-b5bd-4b5f-80fa-32388dd4cb88 tempest-ServerTagsTestJSON-59142761 tempest-ServerTagsTestJSON-59142761-project-member] Task: {'id': task-5116250, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.073185} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.893325] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-801b9b62-b5bd-4b5f-80fa-32388dd4cb88 tempest-ServerTagsTestJSON-59142761 tempest-ServerTagsTestJSON-59142761-project-member] [instance: 9029549c-1914-4bae-91e2-8812b79051ec] Extended root virtual disk {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 925.894267] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b18ff6b3-b432-4155-a116-1b4cca5a77c2 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.928972] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-801b9b62-b5bd-4b5f-80fa-32388dd4cb88 tempest-ServerTagsTestJSON-59142761 tempest-ServerTagsTestJSON-59142761-project-member] [instance: 9029549c-1914-4bae-91e2-8812b79051ec] Reconfiguring VM instance instance-00000048 to attach disk [datastore1] 9029549c-1914-4bae-91e2-8812b79051ec/9029549c-1914-4bae-91e2-8812b79051ec.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 925.930667] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-60fefe12-cbdf-4d1f-b22a-3c5ee6bc7434 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.947480] env[65726]: WARNING openstack [None req-9a0984e5-cc6d-419e-a851-17965236c39e tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 925.947934] env[65726]: WARNING openstack [None req-9a0984e5-cc6d-419e-a851-17965236c39e tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 925.954986] env[65726]: DEBUG nova.network.neutron [None req-9a0984e5-cc6d-419e-a851-17965236c39e tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: 060646cc-f0fb-4e6c-9f67-7b2daa6b6fd6] Instance cache missing network info. {{(pid=65726) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 925.963797] env[65726]: DEBUG oslo_vmware.api [None req-801b9b62-b5bd-4b5f-80fa-32388dd4cb88 tempest-ServerTagsTestJSON-59142761 tempest-ServerTagsTestJSON-59142761-project-member] Waiting for the task: (returnval){ [ 925.963797] env[65726]: value = "task-5116251" [ 925.963797] env[65726]: _type = "Task" [ 925.963797] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.973554] env[65726]: DEBUG oslo_vmware.api [None req-801b9b62-b5bd-4b5f-80fa-32388dd4cb88 tempest-ServerTagsTestJSON-59142761 tempest-ServerTagsTestJSON-59142761-project-member] Task: {'id': task-5116251, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.058197] env[65726]: WARNING openstack [req-77ad4d82-3aef-4dfa-9940-ecf112cbe33d req-5dbb92c0-14f2-44f7-a825-5d359a95ed4e service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 926.058681] env[65726]: WARNING openstack [req-77ad4d82-3aef-4dfa-9940-ecf112cbe33d req-5dbb92c0-14f2-44f7-a825-5d359a95ed4e service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 926.074230] env[65726]: DEBUG oslo_vmware.rw_handles [None req-e4fa020c-c0fd-49d8-8a8a-f7bebcb49337 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5265049a-4106-25b3-49fd-ca4b56af6a22/disk-0.vmdk. {{(pid=65726) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 926.075189] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91d75175-2bdb-459a-924c-eb22e89413dd {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.079120] env[65726]: WARNING openstack [None req-9a0984e5-cc6d-419e-a851-17965236c39e tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 926.079500] env[65726]: WARNING openstack [None req-9a0984e5-cc6d-419e-a851-17965236c39e tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 926.093590] env[65726]: DEBUG oslo_vmware.rw_handles [None req-e4fa020c-c0fd-49d8-8a8a-f7bebcb49337 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5265049a-4106-25b3-49fd-ca4b56af6a22/disk-0.vmdk is in state: ready. {{(pid=65726) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 926.093761] env[65726]: ERROR oslo_vmware.rw_handles [None req-e4fa020c-c0fd-49d8-8a8a-f7bebcb49337 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5265049a-4106-25b3-49fd-ca4b56af6a22/disk-0.vmdk due to incomplete transfer. [ 926.097361] env[65726]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-54557d5c-8518-4594-84d1-b68c3b12d07c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.098413] env[65726]: DEBUG oslo_vmware.api [None req-2cc737a8-ad7d-449c-b270-9db9e18d971d tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': task-5116249, 'name': PowerOnVM_Task, 'duration_secs': 0.733497} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.098657] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-2cc737a8-ad7d-449c-b270-9db9e18d971d tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: dcde25df-87a5-47fb-94ce-334f68894e04] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 926.098892] env[65726]: INFO nova.compute.manager [None req-2cc737a8-ad7d-449c-b270-9db9e18d971d tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: dcde25df-87a5-47fb-94ce-334f68894e04] Took 11.40 seconds to spawn the instance on the hypervisor. [ 926.099102] env[65726]: DEBUG nova.compute.manager [None req-2cc737a8-ad7d-449c-b270-9db9e18d971d tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: dcde25df-87a5-47fb-94ce-334f68894e04] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 926.100171] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-775f0568-cbfa-49a7-ab92-1ac5ccc68a01 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.110130] env[65726]: DEBUG oslo_vmware.rw_handles [None req-e4fa020c-c0fd-49d8-8a8a-f7bebcb49337 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5265049a-4106-25b3-49fd-ca4b56af6a22/disk-0.vmdk. {{(pid=65726) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 926.110408] env[65726]: DEBUG nova.virt.vmwareapi.images [None req-e4fa020c-c0fd-49d8-8a8a-f7bebcb49337 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] [instance: 815ee20b-cb07-4514-bd43-1b74972508a5] Uploaded image 952406eb-2f8d-4d9d-9b5d-e189174387a1 to the Glance image server {{(pid=65726) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 926.112135] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-e4fa020c-c0fd-49d8-8a8a-f7bebcb49337 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] [instance: 815ee20b-cb07-4514-bd43-1b74972508a5] Destroying the VM {{(pid=65726) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 926.112518] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-c686bdfe-4c9f-4457-b485-1bd1c37dc2c0 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.120896] env[65726]: DEBUG oslo_vmware.api [None req-e4fa020c-c0fd-49d8-8a8a-f7bebcb49337 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Waiting for the task: (returnval){ [ 926.120896] env[65726]: value = "task-5116252" [ 926.120896] env[65726]: _type = "Task" [ 926.120896] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.129835] env[65726]: DEBUG oslo_vmware.api [None req-e4fa020c-c0fd-49d8-8a8a-f7bebcb49337 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Task: {'id': task-5116252, 'name': Destroy_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.194423] env[65726]: DEBUG nova.scheduler.client.report [None req-73ce0e73-1afb-4566-995f-f51b4749346e tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 926.287406] env[65726]: WARNING openstack [None req-9a0984e5-cc6d-419e-a851-17965236c39e tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 926.287943] env[65726]: WARNING openstack [None req-9a0984e5-cc6d-419e-a851-17965236c39e tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 926.337038] env[65726]: WARNING openstack [req-77ad4d82-3aef-4dfa-9940-ecf112cbe33d req-5dbb92c0-14f2-44f7-a825-5d359a95ed4e service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 926.337038] env[65726]: WARNING openstack [req-77ad4d82-3aef-4dfa-9940-ecf112cbe33d req-5dbb92c0-14f2-44f7-a825-5d359a95ed4e service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 926.414487] env[65726]: DEBUG oslo_concurrency.lockutils [None req-7cd9e703-73b1-4767-92f9-46012e80c63d tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Acquiring lock "6f91b053-772a-4497-b29d-349b960c55eb" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 926.414487] env[65726]: DEBUG oslo_concurrency.lockutils [None req-7cd9e703-73b1-4767-92f9-46012e80c63d tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Lock "6f91b053-772a-4497-b29d-349b960c55eb" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 926.414487] env[65726]: DEBUG oslo_concurrency.lockutils [None req-7cd9e703-73b1-4767-92f9-46012e80c63d tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Acquiring lock "6f91b053-772a-4497-b29d-349b960c55eb-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 926.414487] env[65726]: DEBUG oslo_concurrency.lockutils [None req-7cd9e703-73b1-4767-92f9-46012e80c63d tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Lock "6f91b053-772a-4497-b29d-349b960c55eb-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 926.414487] env[65726]: DEBUG oslo_concurrency.lockutils [None req-7cd9e703-73b1-4767-92f9-46012e80c63d tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Lock "6f91b053-772a-4497-b29d-349b960c55eb-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 926.423207] env[65726]: INFO nova.compute.manager [None req-7cd9e703-73b1-4767-92f9-46012e80c63d tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 6f91b053-772a-4497-b29d-349b960c55eb] Terminating instance [ 926.449981] env[65726]: DEBUG nova.network.neutron [None req-9a0984e5-cc6d-419e-a851-17965236c39e tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: 060646cc-f0fb-4e6c-9f67-7b2daa6b6fd6] Updating instance_info_cache with network_info: [{"id": "f923178f-ee20-4582-a2e8-6bfd87020d4b", "address": "fa:16:3e:1e:3b:76", "network": {"id": "ad1373b5-e0ed-4ee6-a97c-af931afbab35", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-909418417-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a090d553766847e58e3231b966c92565", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5f60c972-a72d-4c5f-a250-faadfd6eafbe", "external-id": "nsx-vlan-transportzone-932", "segmentation_id": 932, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf923178f-ee", "ovs_interfaceid": "f923178f-ee20-4582-a2e8-6bfd87020d4b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 926.476250] env[65726]: DEBUG oslo_vmware.api [None req-801b9b62-b5bd-4b5f-80fa-32388dd4cb88 tempest-ServerTagsTestJSON-59142761 tempest-ServerTagsTestJSON-59142761-project-member] Task: {'id': task-5116251, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.480516] env[65726]: WARNING openstack [req-77ad4d82-3aef-4dfa-9940-ecf112cbe33d req-5dbb92c0-14f2-44f7-a825-5d359a95ed4e service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 926.480516] env[65726]: WARNING openstack [req-77ad4d82-3aef-4dfa-9940-ecf112cbe33d req-5dbb92c0-14f2-44f7-a825-5d359a95ed4e service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 926.627441] env[65726]: INFO nova.compute.manager [None req-2cc737a8-ad7d-449c-b270-9db9e18d971d tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: dcde25df-87a5-47fb-94ce-334f68894e04] Took 45.23 seconds to build instance. [ 926.634961] env[65726]: DEBUG oslo_vmware.api [None req-e4fa020c-c0fd-49d8-8a8a-f7bebcb49337 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Task: {'id': task-5116252, 'name': Destroy_Task, 'duration_secs': 0.366577} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.636859] env[65726]: DEBUG nova.network.neutron [req-77ad4d82-3aef-4dfa-9940-ecf112cbe33d req-5dbb92c0-14f2-44f7-a825-5d359a95ed4e service nova] [instance: 3088ee5e-0d27-4058-a94b-2e04c3b52add] Updated VIF entry in instance network info cache for port 2ffc16a0-351c-4f5b-8a4a-993ae7adb4f3. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 926.637571] env[65726]: DEBUG nova.network.neutron [req-77ad4d82-3aef-4dfa-9940-ecf112cbe33d req-5dbb92c0-14f2-44f7-a825-5d359a95ed4e service nova] [instance: 3088ee5e-0d27-4058-a94b-2e04c3b52add] Updating instance_info_cache with network_info: [{"id": "2ffc16a0-351c-4f5b-8a4a-993ae7adb4f3", "address": "fa:16:3e:c8:5a:78", "network": {"id": "6e131d80-bd85-46b6-9b8d-c4d34af67f1d", "bridge": "br-int", "label": "tempest-ServersTestManualDisk-1706238551-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.200", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "308ec5a4cf844d2782db77b3edcb4127", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fd0eb882-ab95-4373-aa20-ee565a9919e3", "external-id": "nsx-vlan-transportzone-510", "segmentation_id": 510, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2ffc16a0-35", "ovs_interfaceid": "2ffc16a0-351c-4f5b-8a4a-993ae7adb4f3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 926.639156] env[65726]: INFO nova.virt.vmwareapi.vm_util [None req-e4fa020c-c0fd-49d8-8a8a-f7bebcb49337 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] [instance: 815ee20b-cb07-4514-bd43-1b74972508a5] Destroyed the VM [ 926.639156] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-e4fa020c-c0fd-49d8-8a8a-f7bebcb49337 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] [instance: 815ee20b-cb07-4514-bd43-1b74972508a5] Deleting Snapshot of the VM instance {{(pid=65726) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 926.639522] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-99ad87ac-15f1-41ee-8412-76a7c6e4ecf3 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.648928] env[65726]: DEBUG oslo_vmware.api [None req-e4fa020c-c0fd-49d8-8a8a-f7bebcb49337 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Waiting for the task: (returnval){ [ 926.648928] env[65726]: value = "task-5116253" [ 926.648928] env[65726]: _type = "Task" [ 926.648928] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.661928] env[65726]: DEBUG oslo_vmware.api [None req-e4fa020c-c0fd-49d8-8a8a-f7bebcb49337 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Task: {'id': task-5116253, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.704201] env[65726]: DEBUG oslo_concurrency.lockutils [None req-73ce0e73-1afb-4566-995f-f51b4749346e tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.206s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 926.707709] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e9f6f3ba-cd10-4f45-bec2-5e0b6cfdfe85 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 27.611s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 926.708087] env[65726]: DEBUG nova.objects.instance [None req-e9f6f3ba-cd10-4f45-bec2-5e0b6cfdfe85 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Lazy-loading 'resources' on Instance uuid 6c8db442-daaa-4eca-ae24-b9d4afaf3a71 {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 926.719945] env[65726]: DEBUG oslo_concurrency.lockutils [None req-1fa711c5-810b-46ee-a4d3-beb8e1bea121 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Acquiring lock "c4cc4f85-cb35-4edc-a58b-adfee0ce1265" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 926.719945] env[65726]: DEBUG oslo_concurrency.lockutils [None req-1fa711c5-810b-46ee-a4d3-beb8e1bea121 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Lock "c4cc4f85-cb35-4edc-a58b-adfee0ce1265" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 926.719945] env[65726]: DEBUG oslo_concurrency.lockutils [None req-1fa711c5-810b-46ee-a4d3-beb8e1bea121 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Acquiring lock "c4cc4f85-cb35-4edc-a58b-adfee0ce1265-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 926.719945] env[65726]: DEBUG oslo_concurrency.lockutils [None req-1fa711c5-810b-46ee-a4d3-beb8e1bea121 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Lock "c4cc4f85-cb35-4edc-a58b-adfee0ce1265-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 926.719945] env[65726]: DEBUG oslo_concurrency.lockutils [None req-1fa711c5-810b-46ee-a4d3-beb8e1bea121 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Lock "c4cc4f85-cb35-4edc-a58b-adfee0ce1265-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 926.721915] env[65726]: INFO nova.compute.manager [None req-1fa711c5-810b-46ee-a4d3-beb8e1bea121 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] [instance: c4cc4f85-cb35-4edc-a58b-adfee0ce1265] Terminating instance [ 926.752388] env[65726]: INFO nova.scheduler.client.report [None req-73ce0e73-1afb-4566-995f-f51b4749346e tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Deleted allocations for instance cb8855b6-589d-4863-b86f-7fb4e30fb29e [ 926.880268] env[65726]: DEBUG nova.objects.instance [None req-edb92956-8e7a-4023-9c72-6f4afa47cb43 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Lazy-loading 'flavor' on Instance uuid 5a252ef1-93c9-4bff-842b-b64df2bc5d75 {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 926.928289] env[65726]: DEBUG nova.compute.manager [None req-7cd9e703-73b1-4767-92f9-46012e80c63d tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 6f91b053-772a-4497-b29d-349b960c55eb] Start destroying the instance on the hypervisor. {{(pid=65726) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 926.928289] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-7cd9e703-73b1-4767-92f9-46012e80c63d tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 6f91b053-772a-4497-b29d-349b960c55eb] Destroying instance {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 926.928903] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-030da3c7-66ed-45f9-a3ac-8aa054453ff4 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.938544] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-7cd9e703-73b1-4767-92f9-46012e80c63d tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 6f91b053-772a-4497-b29d-349b960c55eb] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 926.938850] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-db7696d6-1917-45aa-80e9-9c8b13e0ddc8 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.946826] env[65726]: DEBUG oslo_vmware.api [None req-7cd9e703-73b1-4767-92f9-46012e80c63d tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Waiting for the task: (returnval){ [ 926.946826] env[65726]: value = "task-5116254" [ 926.946826] env[65726]: _type = "Task" [ 926.946826] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.955511] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9a0984e5-cc6d-419e-a851-17965236c39e tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Releasing lock "refresh_cache-060646cc-f0fb-4e6c-9f67-7b2daa6b6fd6" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 926.957371] env[65726]: DEBUG nova.compute.manager [None req-9a0984e5-cc6d-419e-a851-17965236c39e tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: 060646cc-f0fb-4e6c-9f67-7b2daa6b6fd6] Instance network_info: |[{"id": "f923178f-ee20-4582-a2e8-6bfd87020d4b", "address": "fa:16:3e:1e:3b:76", "network": {"id": "ad1373b5-e0ed-4ee6-a97c-af931afbab35", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-909418417-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a090d553766847e58e3231b966c92565", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5f60c972-a72d-4c5f-a250-faadfd6eafbe", "external-id": "nsx-vlan-transportzone-932", "segmentation_id": 932, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf923178f-ee", "ovs_interfaceid": "f923178f-ee20-4582-a2e8-6bfd87020d4b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 926.957995] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-9a0984e5-cc6d-419e-a851-17965236c39e tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: 060646cc-f0fb-4e6c-9f67-7b2daa6b6fd6] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:1e:3b:76', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '5f60c972-a72d-4c5f-a250-faadfd6eafbe', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f923178f-ee20-4582-a2e8-6bfd87020d4b', 'vif_model': 'vmxnet3'}] {{(pid=65726) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 926.975847] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-9a0984e5-cc6d-419e-a851-17965236c39e tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 926.983516] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 060646cc-f0fb-4e6c-9f67-7b2daa6b6fd6] Creating VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 926.999472] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b39aa128-aeb0-43dc-99d4-3135c2bcc635 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.035934] env[65726]: DEBUG oslo_vmware.api [None req-7cd9e703-73b1-4767-92f9-46012e80c63d tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': task-5116254, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.045032] env[65726]: DEBUG oslo_vmware.api [None req-801b9b62-b5bd-4b5f-80fa-32388dd4cb88 tempest-ServerTagsTestJSON-59142761 tempest-ServerTagsTestJSON-59142761-project-member] Task: {'id': task-5116251, 'name': ReconfigVM_Task, 'duration_secs': 0.908876} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.045032] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-801b9b62-b5bd-4b5f-80fa-32388dd4cb88 tempest-ServerTagsTestJSON-59142761 tempest-ServerTagsTestJSON-59142761-project-member] [instance: 9029549c-1914-4bae-91e2-8812b79051ec] Reconfigured VM instance instance-00000048 to attach disk [datastore1] 9029549c-1914-4bae-91e2-8812b79051ec/9029549c-1914-4bae-91e2-8812b79051ec.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 927.045032] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-89fdf800-2445-4761-9abf-153bdaba832c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.047327] env[65726]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 927.047327] env[65726]: value = "task-5116255" [ 927.047327] env[65726]: _type = "Task" [ 927.047327] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.054917] env[65726]: DEBUG oslo_vmware.api [None req-801b9b62-b5bd-4b5f-80fa-32388dd4cb88 tempest-ServerTagsTestJSON-59142761 tempest-ServerTagsTestJSON-59142761-project-member] Waiting for the task: (returnval){ [ 927.054917] env[65726]: value = "task-5116256" [ 927.054917] env[65726]: _type = "Task" [ 927.054917] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.069785] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116255, 'name': CreateVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.069785] env[65726]: DEBUG oslo_vmware.api [None req-801b9b62-b5bd-4b5f-80fa-32388dd4cb88 tempest-ServerTagsTestJSON-59142761 tempest-ServerTagsTestJSON-59142761-project-member] Task: {'id': task-5116256, 'name': Rename_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.131438] env[65726]: DEBUG oslo_concurrency.lockutils [None req-2cc737a8-ad7d-449c-b270-9db9e18d971d tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Lock "dcde25df-87a5-47fb-94ce-334f68894e04" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 55.698s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 927.140468] env[65726]: DEBUG oslo_concurrency.lockutils [req-77ad4d82-3aef-4dfa-9940-ecf112cbe33d req-5dbb92c0-14f2-44f7-a825-5d359a95ed4e service nova] Releasing lock "refresh_cache-3088ee5e-0d27-4058-a94b-2e04c3b52add" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 927.160533] env[65726]: DEBUG oslo_vmware.api [None req-e4fa020c-c0fd-49d8-8a8a-f7bebcb49337 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Task: {'id': task-5116253, 'name': RemoveSnapshot_Task} progress is 80%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.227218] env[65726]: DEBUG nova.compute.manager [None req-1fa711c5-810b-46ee-a4d3-beb8e1bea121 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] [instance: c4cc4f85-cb35-4edc-a58b-adfee0ce1265] Start destroying the instance on the hypervisor. {{(pid=65726) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 927.227943] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-1fa711c5-810b-46ee-a4d3-beb8e1bea121 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] [instance: c4cc4f85-cb35-4edc-a58b-adfee0ce1265] Destroying instance {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 927.228643] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d7e4c6a-dd17-42de-9d23-190ec60d9a7f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.245739] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-1fa711c5-810b-46ee-a4d3-beb8e1bea121 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] [instance: c4cc4f85-cb35-4edc-a58b-adfee0ce1265] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 927.246516] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1ef6eb24-80bd-4bda-91d9-a3bf2ed91772 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.256461] env[65726]: DEBUG oslo_vmware.api [None req-1fa711c5-810b-46ee-a4d3-beb8e1bea121 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Waiting for the task: (returnval){ [ 927.256461] env[65726]: value = "task-5116257" [ 927.256461] env[65726]: _type = "Task" [ 927.256461] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.264081] env[65726]: DEBUG oslo_concurrency.lockutils [None req-73ce0e73-1afb-4566-995f-f51b4749346e tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Lock "cb8855b6-589d-4863-b86f-7fb4e30fb29e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 35.989s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 927.284315] env[65726]: DEBUG oslo_vmware.api [None req-1fa711c5-810b-46ee-a4d3-beb8e1bea121 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Task: {'id': task-5116257, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.390027] env[65726]: DEBUG oslo_concurrency.lockutils [None req-edb92956-8e7a-4023-9c72-6f4afa47cb43 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Lock "5a252ef1-93c9-4bff-842b-b64df2bc5d75" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 8.898s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 927.462383] env[65726]: DEBUG oslo_vmware.api [None req-7cd9e703-73b1-4767-92f9-46012e80c63d tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': task-5116254, 'name': PowerOffVM_Task, 'duration_secs': 0.238289} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.462676] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-7cd9e703-73b1-4767-92f9-46012e80c63d tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 6f91b053-772a-4497-b29d-349b960c55eb] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 927.462839] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-7cd9e703-73b1-4767-92f9-46012e80c63d tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 6f91b053-772a-4497-b29d-349b960c55eb] Unregistering the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 927.463117] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8e1036f1-411b-4585-b7f0-66af609eb41d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.502086] env[65726]: DEBUG oslo_concurrency.lockutils [None req-57ebeb7d-b6c8-4c85-80c2-d441f18e658e tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Acquiring lock "6fd70b4a-63e4-4258-9cff-f3b582500b9f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 927.502379] env[65726]: DEBUG oslo_concurrency.lockutils [None req-57ebeb7d-b6c8-4c85-80c2-d441f18e658e tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Lock "6fd70b4a-63e4-4258-9cff-f3b582500b9f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 927.530392] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-7cd9e703-73b1-4767-92f9-46012e80c63d tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 6f91b053-772a-4497-b29d-349b960c55eb] Unregistered the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 927.531501] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-7cd9e703-73b1-4767-92f9-46012e80c63d tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 6f91b053-772a-4497-b29d-349b960c55eb] Deleting contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 927.531501] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-7cd9e703-73b1-4767-92f9-46012e80c63d tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Deleting the datastore file [datastore1] 6f91b053-772a-4497-b29d-349b960c55eb {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 927.531501] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-408c1b67-b3f5-4765-90ad-7c47d6b380a6 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.539385] env[65726]: DEBUG oslo_vmware.api [None req-7cd9e703-73b1-4767-92f9-46012e80c63d tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Waiting for the task: (returnval){ [ 927.539385] env[65726]: value = "task-5116259" [ 927.539385] env[65726]: _type = "Task" [ 927.539385] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.552211] env[65726]: DEBUG oslo_vmware.api [None req-7cd9e703-73b1-4767-92f9-46012e80c63d tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': task-5116259, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.566867] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116255, 'name': CreateVM_Task, 'duration_secs': 0.443067} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.567069] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 060646cc-f0fb-4e6c-9f67-7b2daa6b6fd6] Created VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 927.568254] env[65726]: WARNING openstack [None req-9a0984e5-cc6d-419e-a851-17965236c39e tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 927.568608] env[65726]: WARNING openstack [None req-9a0984e5-cc6d-419e-a851-17965236c39e tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 927.574338] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9a0984e5-cc6d-419e-a851-17965236c39e tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 927.574495] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9a0984e5-cc6d-419e-a851-17965236c39e tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 927.574813] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9a0984e5-cc6d-419e-a851-17965236c39e tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 927.578656] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-be5d25db-366b-47ea-b5e2-08c1337fb3a7 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.580624] env[65726]: DEBUG oslo_vmware.api [None req-801b9b62-b5bd-4b5f-80fa-32388dd4cb88 tempest-ServerTagsTestJSON-59142761 tempest-ServerTagsTestJSON-59142761-project-member] Task: {'id': task-5116256, 'name': Rename_Task, 'duration_secs': 0.206684} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.583102] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-801b9b62-b5bd-4b5f-80fa-32388dd4cb88 tempest-ServerTagsTestJSON-59142761 tempest-ServerTagsTestJSON-59142761-project-member] [instance: 9029549c-1914-4bae-91e2-8812b79051ec] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 927.583933] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-037197b2-69fd-41b2-9852-043baed83002 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.587280] env[65726]: DEBUG oslo_vmware.api [None req-9a0984e5-cc6d-419e-a851-17965236c39e tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Waiting for the task: (returnval){ [ 927.587280] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52d14413-e5fb-de4c-505f-91e1255b23e3" [ 927.587280] env[65726]: _type = "Task" [ 927.587280] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.593195] env[65726]: DEBUG oslo_vmware.api [None req-801b9b62-b5bd-4b5f-80fa-32388dd4cb88 tempest-ServerTagsTestJSON-59142761 tempest-ServerTagsTestJSON-59142761-project-member] Waiting for the task: (returnval){ [ 927.593195] env[65726]: value = "task-5116260" [ 927.593195] env[65726]: _type = "Task" [ 927.593195] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.601014] env[65726]: DEBUG oslo_vmware.api [None req-9a0984e5-cc6d-419e-a851-17965236c39e tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52d14413-e5fb-de4c-505f-91e1255b23e3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.613540] env[65726]: DEBUG oslo_vmware.api [None req-801b9b62-b5bd-4b5f-80fa-32388dd4cb88 tempest-ServerTagsTestJSON-59142761 tempest-ServerTagsTestJSON-59142761-project-member] Task: {'id': task-5116260, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.664321] env[65726]: DEBUG oslo_vmware.api [None req-e4fa020c-c0fd-49d8-8a8a-f7bebcb49337 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Task: {'id': task-5116253, 'name': RemoveSnapshot_Task, 'duration_secs': 0.615967} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.667499] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-e4fa020c-c0fd-49d8-8a8a-f7bebcb49337 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] [instance: 815ee20b-cb07-4514-bd43-1b74972508a5] Deleted Snapshot of the VM instance {{(pid=65726) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 927.667742] env[65726]: INFO nova.compute.manager [None req-e4fa020c-c0fd-49d8-8a8a-f7bebcb49337 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] [instance: 815ee20b-cb07-4514-bd43-1b74972508a5] Took 15.96 seconds to snapshot the instance on the hypervisor. [ 927.770996] env[65726]: DEBUG oslo_vmware.api [None req-1fa711c5-810b-46ee-a4d3-beb8e1bea121 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Task: {'id': task-5116257, 'name': PowerOffVM_Task, 'duration_secs': 0.203959} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.774998] env[65726]: DEBUG oslo_concurrency.lockutils [None req-816608ea-3ce8-4b57-8b12-cbe3b89f21e9 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Acquiring lock "5a252ef1-93c9-4bff-842b-b64df2bc5d75" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 927.775291] env[65726]: DEBUG oslo_concurrency.lockutils [None req-816608ea-3ce8-4b57-8b12-cbe3b89f21e9 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Lock "5a252ef1-93c9-4bff-842b-b64df2bc5d75" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 927.776812] env[65726]: DEBUG oslo_concurrency.lockutils [None req-816608ea-3ce8-4b57-8b12-cbe3b89f21e9 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Acquiring lock "5a252ef1-93c9-4bff-842b-b64df2bc5d75-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 927.776928] env[65726]: DEBUG oslo_concurrency.lockutils [None req-816608ea-3ce8-4b57-8b12-cbe3b89f21e9 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Lock "5a252ef1-93c9-4bff-842b-b64df2bc5d75-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 927.777096] env[65726]: DEBUG oslo_concurrency.lockutils [None req-816608ea-3ce8-4b57-8b12-cbe3b89f21e9 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Lock "5a252ef1-93c9-4bff-842b-b64df2bc5d75-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 927.779029] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-1fa711c5-810b-46ee-a4d3-beb8e1bea121 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] [instance: c4cc4f85-cb35-4edc-a58b-adfee0ce1265] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 927.779109] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-1fa711c5-810b-46ee-a4d3-beb8e1bea121 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] [instance: c4cc4f85-cb35-4edc-a58b-adfee0ce1265] Unregistering the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 927.779888] env[65726]: INFO nova.compute.manager [None req-816608ea-3ce8-4b57-8b12-cbe3b89f21e9 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: 5a252ef1-93c9-4bff-842b-b64df2bc5d75] Terminating instance [ 927.781586] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4707c332-b8e9-4742-855c-59e4e5220811 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.850567] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-1fa711c5-810b-46ee-a4d3-beb8e1bea121 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] [instance: c4cc4f85-cb35-4edc-a58b-adfee0ce1265] Unregistered the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 927.850830] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-1fa711c5-810b-46ee-a4d3-beb8e1bea121 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] [instance: c4cc4f85-cb35-4edc-a58b-adfee0ce1265] Deleting contents of the VM from datastore datastore2 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 927.851046] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-1fa711c5-810b-46ee-a4d3-beb8e1bea121 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Deleting the datastore file [datastore2] c4cc4f85-cb35-4edc-a58b-adfee0ce1265 {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 927.851344] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e2526df8-2e86-435a-8f6a-2762d3613812 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.862851] env[65726]: DEBUG oslo_vmware.api [None req-1fa711c5-810b-46ee-a4d3-beb8e1bea121 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Waiting for the task: (returnval){ [ 927.862851] env[65726]: value = "task-5116262" [ 927.862851] env[65726]: _type = "Task" [ 927.862851] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.879434] env[65726]: DEBUG oslo_vmware.api [None req-1fa711c5-810b-46ee-a4d3-beb8e1bea121 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Task: {'id': task-5116262, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.882293] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26377a06-1c9f-4458-88ed-ee2cf8fd5911 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.890841] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3875ecab-9ed2-4ab9-9802-969b64eacc60 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.936624] env[65726]: DEBUG oslo_concurrency.lockutils [None req-2eacb245-5155-4571-979a-d9aaa1a355b7 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Acquiring lock "c108f2a5-031b-47ee-9a5b-d62c2f42c26b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 927.937466] env[65726]: DEBUG oslo_concurrency.lockutils [None req-2eacb245-5155-4571-979a-d9aaa1a355b7 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Lock "c108f2a5-031b-47ee-9a5b-d62c2f42c26b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 927.937466] env[65726]: DEBUG oslo_concurrency.lockutils [None req-2eacb245-5155-4571-979a-d9aaa1a355b7 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Acquiring lock "c108f2a5-031b-47ee-9a5b-d62c2f42c26b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 927.937466] env[65726]: DEBUG oslo_concurrency.lockutils [None req-2eacb245-5155-4571-979a-d9aaa1a355b7 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Lock "c108f2a5-031b-47ee-9a5b-d62c2f42c26b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 927.937466] env[65726]: DEBUG oslo_concurrency.lockutils [None req-2eacb245-5155-4571-979a-d9aaa1a355b7 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Lock "c108f2a5-031b-47ee-9a5b-d62c2f42c26b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 927.939588] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f8b044b-abd6-44a0-9530-a0c30269ef16 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.943227] env[65726]: INFO nova.compute.manager [None req-2eacb245-5155-4571-979a-d9aaa1a355b7 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: c108f2a5-031b-47ee-9a5b-d62c2f42c26b] Terminating instance [ 927.952142] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da64941a-e4e8-48da-88a4-4d60414084c5 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.968875] env[65726]: DEBUG nova.compute.provider_tree [None req-e9f6f3ba-cd10-4f45-bec2-5e0b6cfdfe85 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 928.007514] env[65726]: DEBUG nova.compute.manager [None req-57ebeb7d-b6c8-4c85-80c2-d441f18e658e tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] [instance: 6fd70b4a-63e4-4258-9cff-f3b582500b9f] Starting instance... {{(pid=65726) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 928.054243] env[65726]: DEBUG oslo_vmware.api [None req-7cd9e703-73b1-4767-92f9-46012e80c63d tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': task-5116259, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.387147} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.054243] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-7cd9e703-73b1-4767-92f9-46012e80c63d tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Deleted the datastore file {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 928.054243] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-7cd9e703-73b1-4767-92f9-46012e80c63d tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 6f91b053-772a-4497-b29d-349b960c55eb] Deleted contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 928.054604] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-7cd9e703-73b1-4767-92f9-46012e80c63d tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 6f91b053-772a-4497-b29d-349b960c55eb] Instance destroyed {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 928.054604] env[65726]: INFO nova.compute.manager [None req-7cd9e703-73b1-4767-92f9-46012e80c63d tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 6f91b053-772a-4497-b29d-349b960c55eb] Took 1.13 seconds to destroy the instance on the hypervisor. [ 928.055320] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-7cd9e703-73b1-4767-92f9-46012e80c63d tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 928.055320] env[65726]: DEBUG nova.compute.manager [-] [instance: 6f91b053-772a-4497-b29d-349b960c55eb] Deallocating network for instance {{(pid=65726) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 928.055320] env[65726]: DEBUG nova.network.neutron [-] [instance: 6f91b053-772a-4497-b29d-349b960c55eb] deallocate_for_instance() {{(pid=65726) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 928.055856] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 928.056263] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 928.103044] env[65726]: DEBUG oslo_vmware.api [None req-9a0984e5-cc6d-419e-a851-17965236c39e tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52d14413-e5fb-de4c-505f-91e1255b23e3, 'name': SearchDatastore_Task, 'duration_secs': 0.023784} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.103533] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9a0984e5-cc6d-419e-a851-17965236c39e tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 928.103756] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-9a0984e5-cc6d-419e-a851-17965236c39e tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: 060646cc-f0fb-4e6c-9f67-7b2daa6b6fd6] Processing image b52362a3-ee8a-4cbf-b06f-513b0cc8f95c {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 928.104107] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9a0984e5-cc6d-419e-a851-17965236c39e tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 928.104276] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9a0984e5-cc6d-419e-a851-17965236c39e tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 928.104463] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-9a0984e5-cc6d-419e-a851-17965236c39e tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 928.105639] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f53b3b31-9d46-4cd4-bdd0-981bf7f2029c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.112846] env[65726]: DEBUG oslo_vmware.api [None req-801b9b62-b5bd-4b5f-80fa-32388dd4cb88 tempest-ServerTagsTestJSON-59142761 tempest-ServerTagsTestJSON-59142761-project-member] Task: {'id': task-5116260, 'name': PowerOnVM_Task} progress is 87%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.127683] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-9a0984e5-cc6d-419e-a851-17965236c39e tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 928.127932] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-9a0984e5-cc6d-419e-a851-17965236c39e tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65726) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 928.128875] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-30f65338-3094-4713-befd-a804c9c19bce {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.135171] env[65726]: DEBUG oslo_vmware.api [None req-9a0984e5-cc6d-419e-a851-17965236c39e tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Waiting for the task: (returnval){ [ 928.135171] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]528d5ca4-0079-7bd7-db05-2c445c490f07" [ 928.135171] env[65726]: _type = "Task" [ 928.135171] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.147580] env[65726]: DEBUG oslo_vmware.api [None req-9a0984e5-cc6d-419e-a851-17965236c39e tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]528d5ca4-0079-7bd7-db05-2c445c490f07, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.289569] env[65726]: DEBUG nova.compute.manager [None req-816608ea-3ce8-4b57-8b12-cbe3b89f21e9 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: 5a252ef1-93c9-4bff-842b-b64df2bc5d75] Start destroying the instance on the hypervisor. {{(pid=65726) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 928.289569] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-816608ea-3ce8-4b57-8b12-cbe3b89f21e9 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: 5a252ef1-93c9-4bff-842b-b64df2bc5d75] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 928.289943] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fdbc664e-b131-43fe-b5bf-c775e7b81ae4 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.303503] env[65726]: DEBUG oslo_vmware.api [None req-816608ea-3ce8-4b57-8b12-cbe3b89f21e9 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Waiting for the task: (returnval){ [ 928.303503] env[65726]: value = "task-5116263" [ 928.303503] env[65726]: _type = "Task" [ 928.303503] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.314932] env[65726]: DEBUG oslo_vmware.api [None req-816608ea-3ce8-4b57-8b12-cbe3b89f21e9 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': task-5116263, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.378398] env[65726]: DEBUG oslo_vmware.api [None req-1fa711c5-810b-46ee-a4d3-beb8e1bea121 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Task: {'id': task-5116262, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.283254} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.378721] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-1fa711c5-810b-46ee-a4d3-beb8e1bea121 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Deleted the datastore file {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 928.379089] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-1fa711c5-810b-46ee-a4d3-beb8e1bea121 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] [instance: c4cc4f85-cb35-4edc-a58b-adfee0ce1265] Deleted contents of the VM from datastore datastore2 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 928.379252] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-1fa711c5-810b-46ee-a4d3-beb8e1bea121 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] [instance: c4cc4f85-cb35-4edc-a58b-adfee0ce1265] Instance destroyed {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 928.381415] env[65726]: INFO nova.compute.manager [None req-1fa711c5-810b-46ee-a4d3-beb8e1bea121 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] [instance: c4cc4f85-cb35-4edc-a58b-adfee0ce1265] Took 1.15 seconds to destroy the instance on the hypervisor. [ 928.381415] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-1fa711c5-810b-46ee-a4d3-beb8e1bea121 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 928.381415] env[65726]: DEBUG nova.compute.manager [-] [instance: c4cc4f85-cb35-4edc-a58b-adfee0ce1265] Deallocating network for instance {{(pid=65726) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 928.381415] env[65726]: DEBUG nova.network.neutron [-] [instance: c4cc4f85-cb35-4edc-a58b-adfee0ce1265] deallocate_for_instance() {{(pid=65726) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 928.381415] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 928.385737] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 928.404975] env[65726]: DEBUG nova.compute.manager [req-00c6ba26-63af-4954-ba22-fecbe93f2e8f req-d3bbc8b8-26ac-4f68-921a-cbb9181e437e service nova] [instance: 060646cc-f0fb-4e6c-9f67-7b2daa6b6fd6] Received event network-vif-plugged-f923178f-ee20-4582-a2e8-6bfd87020d4b {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 928.405266] env[65726]: DEBUG oslo_concurrency.lockutils [req-00c6ba26-63af-4954-ba22-fecbe93f2e8f req-d3bbc8b8-26ac-4f68-921a-cbb9181e437e service nova] Acquiring lock "060646cc-f0fb-4e6c-9f67-7b2daa6b6fd6-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 928.405834] env[65726]: DEBUG oslo_concurrency.lockutils [req-00c6ba26-63af-4954-ba22-fecbe93f2e8f req-d3bbc8b8-26ac-4f68-921a-cbb9181e437e service nova] Lock "060646cc-f0fb-4e6c-9f67-7b2daa6b6fd6-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 928.406083] env[65726]: DEBUG oslo_concurrency.lockutils [req-00c6ba26-63af-4954-ba22-fecbe93f2e8f req-d3bbc8b8-26ac-4f68-921a-cbb9181e437e service nova] Lock "060646cc-f0fb-4e6c-9f67-7b2daa6b6fd6-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 928.406716] env[65726]: DEBUG nova.compute.manager [req-00c6ba26-63af-4954-ba22-fecbe93f2e8f req-d3bbc8b8-26ac-4f68-921a-cbb9181e437e service nova] [instance: 060646cc-f0fb-4e6c-9f67-7b2daa6b6fd6] No waiting events found dispatching network-vif-plugged-f923178f-ee20-4582-a2e8-6bfd87020d4b {{(pid=65726) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 928.406716] env[65726]: WARNING nova.compute.manager [req-00c6ba26-63af-4954-ba22-fecbe93f2e8f req-d3bbc8b8-26ac-4f68-921a-cbb9181e437e service nova] [instance: 060646cc-f0fb-4e6c-9f67-7b2daa6b6fd6] Received unexpected event network-vif-plugged-f923178f-ee20-4582-a2e8-6bfd87020d4b for instance with vm_state building and task_state spawning. [ 928.406716] env[65726]: DEBUG nova.compute.manager [req-00c6ba26-63af-4954-ba22-fecbe93f2e8f req-d3bbc8b8-26ac-4f68-921a-cbb9181e437e service nova] [instance: 060646cc-f0fb-4e6c-9f67-7b2daa6b6fd6] Received event network-changed-f923178f-ee20-4582-a2e8-6bfd87020d4b {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 928.408262] env[65726]: DEBUG nova.compute.manager [req-00c6ba26-63af-4954-ba22-fecbe93f2e8f req-d3bbc8b8-26ac-4f68-921a-cbb9181e437e service nova] [instance: 060646cc-f0fb-4e6c-9f67-7b2daa6b6fd6] Refreshing instance network info cache due to event network-changed-f923178f-ee20-4582-a2e8-6bfd87020d4b. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 928.408262] env[65726]: DEBUG oslo_concurrency.lockutils [req-00c6ba26-63af-4954-ba22-fecbe93f2e8f req-d3bbc8b8-26ac-4f68-921a-cbb9181e437e service nova] Acquiring lock "refresh_cache-060646cc-f0fb-4e6c-9f67-7b2daa6b6fd6" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 928.408262] env[65726]: DEBUG oslo_concurrency.lockutils [req-00c6ba26-63af-4954-ba22-fecbe93f2e8f req-d3bbc8b8-26ac-4f68-921a-cbb9181e437e service nova] Acquired lock "refresh_cache-060646cc-f0fb-4e6c-9f67-7b2daa6b6fd6" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 928.408262] env[65726]: DEBUG nova.network.neutron [req-00c6ba26-63af-4954-ba22-fecbe93f2e8f req-d3bbc8b8-26ac-4f68-921a-cbb9181e437e service nova] [instance: 060646cc-f0fb-4e6c-9f67-7b2daa6b6fd6] Refreshing network info cache for port f923178f-ee20-4582-a2e8-6bfd87020d4b {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 928.413048] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 928.415196] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 928.450315] env[65726]: DEBUG nova.compute.manager [None req-2eacb245-5155-4571-979a-d9aaa1a355b7 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: c108f2a5-031b-47ee-9a5b-d62c2f42c26b] Start destroying the instance on the hypervisor. {{(pid=65726) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 928.451323] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-2eacb245-5155-4571-979a-d9aaa1a355b7 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: c108f2a5-031b-47ee-9a5b-d62c2f42c26b] Destroying instance {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 928.452671] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acf11fc9-1c82-4839-a2f8-bd441786da2e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.462513] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-2eacb245-5155-4571-979a-d9aaa1a355b7 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: c108f2a5-031b-47ee-9a5b-d62c2f42c26b] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 928.462942] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-955d6e82-f77b-47db-9d43-a35e8d5c7009 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.473275] env[65726]: DEBUG nova.scheduler.client.report [None req-e9f6f3ba-cd10-4f45-bec2-5e0b6cfdfe85 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 928.477907] env[65726]: DEBUG oslo_vmware.api [None req-2eacb245-5155-4571-979a-d9aaa1a355b7 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Waiting for the task: (returnval){ [ 928.477907] env[65726]: value = "task-5116264" [ 928.477907] env[65726]: _type = "Task" [ 928.477907] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.493907] env[65726]: DEBUG oslo_vmware.api [None req-2eacb245-5155-4571-979a-d9aaa1a355b7 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Task: {'id': task-5116264, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.543388] env[65726]: DEBUG oslo_concurrency.lockutils [None req-57ebeb7d-b6c8-4c85-80c2-d441f18e658e tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 928.574506] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 928.575016] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 928.608972] env[65726]: DEBUG oslo_vmware.api [None req-801b9b62-b5bd-4b5f-80fa-32388dd4cb88 tempest-ServerTagsTestJSON-59142761 tempest-ServerTagsTestJSON-59142761-project-member] Task: {'id': task-5116260, 'name': PowerOnVM_Task, 'duration_secs': 0.791968} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.609431] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-801b9b62-b5bd-4b5f-80fa-32388dd4cb88 tempest-ServerTagsTestJSON-59142761 tempest-ServerTagsTestJSON-59142761-project-member] [instance: 9029549c-1914-4bae-91e2-8812b79051ec] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 928.610322] env[65726]: INFO nova.compute.manager [None req-801b9b62-b5bd-4b5f-80fa-32388dd4cb88 tempest-ServerTagsTestJSON-59142761 tempest-ServerTagsTestJSON-59142761-project-member] [instance: 9029549c-1914-4bae-91e2-8812b79051ec] Took 10.06 seconds to spawn the instance on the hypervisor. [ 928.610984] env[65726]: DEBUG nova.compute.manager [None req-801b9b62-b5bd-4b5f-80fa-32388dd4cb88 tempest-ServerTagsTestJSON-59142761 tempest-ServerTagsTestJSON-59142761-project-member] [instance: 9029549c-1914-4bae-91e2-8812b79051ec] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 928.611548] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a107b4fe-375d-4e10-9d47-b05c144519b8 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.646646] env[65726]: DEBUG oslo_vmware.api [None req-9a0984e5-cc6d-419e-a851-17965236c39e tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]528d5ca4-0079-7bd7-db05-2c445c490f07, 'name': SearchDatastore_Task, 'duration_secs': 0.012003} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.648411] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d17538db-8138-4079-9f9f-5febfa4c0fe8 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.655375] env[65726]: DEBUG oslo_vmware.api [None req-9a0984e5-cc6d-419e-a851-17965236c39e tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Waiting for the task: (returnval){ [ 928.655375] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52636646-d776-1cc2-fd6d-008d8ab68e24" [ 928.655375] env[65726]: _type = "Task" [ 928.655375] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.670706] env[65726]: DEBUG oslo_vmware.api [None req-9a0984e5-cc6d-419e-a851-17965236c39e tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52636646-d776-1cc2-fd6d-008d8ab68e24, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.816751] env[65726]: DEBUG oslo_vmware.api [None req-816608ea-3ce8-4b57-8b12-cbe3b89f21e9 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': task-5116263, 'name': PowerOffVM_Task, 'duration_secs': 0.224315} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.816751] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-816608ea-3ce8-4b57-8b12-cbe3b89f21e9 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: 5a252ef1-93c9-4bff-842b-b64df2bc5d75] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 928.816751] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-816608ea-3ce8-4b57-8b12-cbe3b89f21e9 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: 5a252ef1-93c9-4bff-842b-b64df2bc5d75] Volume detach. Driver type: vmdk {{(pid=65726) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 928.816751] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-816608ea-3ce8-4b57-8b12-cbe3b89f21e9 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: 5a252ef1-93c9-4bff-842b-b64df2bc5d75] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-995209', 'volume_id': '2e0dafa1-dc04-4dd8-b30f-16c07c062b44', 'name': 'volume-2e0dafa1-dc04-4dd8-b30f-16c07c062b44', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '5a252ef1-93c9-4bff-842b-b64df2bc5d75', 'attached_at': '', 'detached_at': '', 'volume_id': '2e0dafa1-dc04-4dd8-b30f-16c07c062b44', 'serial': '2e0dafa1-dc04-4dd8-b30f-16c07c062b44'} {{(pid=65726) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 928.816751] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d709ccfc-8935-4ed2-a7ee-7e4c18d2e2ac {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.846224] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07f318e1-efca-44df-96e0-a89adaa80fd5 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.855019] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09afbdaf-c3eb-40d7-b24e-3c0716cec0f2 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.882139] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbdcf082-5901-4f62-93a4-3c5fbb92063c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.899879] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-816608ea-3ce8-4b57-8b12-cbe3b89f21e9 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] The volume has not been displaced from its original location: [datastore1] volume-2e0dafa1-dc04-4dd8-b30f-16c07c062b44/volume-2e0dafa1-dc04-4dd8-b30f-16c07c062b44.vmdk. No consolidation needed. {{(pid=65726) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 928.905717] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-816608ea-3ce8-4b57-8b12-cbe3b89f21e9 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: 5a252ef1-93c9-4bff-842b-b64df2bc5d75] Reconfiguring VM instance instance-00000044 to detach disk 2001 {{(pid=65726) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 928.906728] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bb9e317a-bcd4-4aca-89bb-d7bc94f52d42 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.924246] env[65726]: WARNING openstack [req-00c6ba26-63af-4954-ba22-fecbe93f2e8f req-d3bbc8b8-26ac-4f68-921a-cbb9181e437e service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 928.924246] env[65726]: WARNING openstack [req-00c6ba26-63af-4954-ba22-fecbe93f2e8f req-d3bbc8b8-26ac-4f68-921a-cbb9181e437e service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 928.935641] env[65726]: DEBUG oslo_vmware.api [None req-816608ea-3ce8-4b57-8b12-cbe3b89f21e9 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Waiting for the task: (returnval){ [ 928.935641] env[65726]: value = "task-5116265" [ 928.935641] env[65726]: _type = "Task" [ 928.935641] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.945657] env[65726]: DEBUG oslo_vmware.api [None req-816608ea-3ce8-4b57-8b12-cbe3b89f21e9 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': task-5116265, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.979413] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e9f6f3ba-cd10-4f45-bec2-5e0b6cfdfe85 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.272s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 928.982744] env[65726]: DEBUG oslo_concurrency.lockutils [None req-21960d76-f281-41f6-8236-c885b049da01 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 26.342s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 928.996141] env[65726]: DEBUG oslo_vmware.api [None req-2eacb245-5155-4571-979a-d9aaa1a355b7 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Task: {'id': task-5116264, 'name': PowerOffVM_Task, 'duration_secs': 0.265807} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.996511] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-2eacb245-5155-4571-979a-d9aaa1a355b7 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: c108f2a5-031b-47ee-9a5b-d62c2f42c26b] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 928.996674] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-2eacb245-5155-4571-979a-d9aaa1a355b7 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: c108f2a5-031b-47ee-9a5b-d62c2f42c26b] Unregistering the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 928.997339] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c15364f4-b493-4ba3-88d8-4b6841fe1936 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.023430] env[65726]: INFO nova.scheduler.client.report [None req-e9f6f3ba-cd10-4f45-bec2-5e0b6cfdfe85 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Deleted allocations for instance 6c8db442-daaa-4eca-ae24-b9d4afaf3a71 [ 929.076054] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-2eacb245-5155-4571-979a-d9aaa1a355b7 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: c108f2a5-031b-47ee-9a5b-d62c2f42c26b] Unregistered the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 929.076054] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-2eacb245-5155-4571-979a-d9aaa1a355b7 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: c108f2a5-031b-47ee-9a5b-d62c2f42c26b] Deleting contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 929.076054] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-2eacb245-5155-4571-979a-d9aaa1a355b7 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Deleting the datastore file [datastore1] c108f2a5-031b-47ee-9a5b-d62c2f42c26b {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 929.076328] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-19a24ea3-1d2e-42a8-84fb-269bb53496e8 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.085388] env[65726]: DEBUG oslo_vmware.api [None req-2eacb245-5155-4571-979a-d9aaa1a355b7 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Waiting for the task: (returnval){ [ 929.085388] env[65726]: value = "task-5116267" [ 929.085388] env[65726]: _type = "Task" [ 929.085388] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.086845] env[65726]: WARNING openstack [req-00c6ba26-63af-4954-ba22-fecbe93f2e8f req-d3bbc8b8-26ac-4f68-921a-cbb9181e437e service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 929.087279] env[65726]: WARNING openstack [req-00c6ba26-63af-4954-ba22-fecbe93f2e8f req-d3bbc8b8-26ac-4f68-921a-cbb9181e437e service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 929.106217] env[65726]: DEBUG oslo_vmware.api [None req-2eacb245-5155-4571-979a-d9aaa1a355b7 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Task: {'id': task-5116267, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.141522] env[65726]: INFO nova.compute.manager [None req-801b9b62-b5bd-4b5f-80fa-32388dd4cb88 tempest-ServerTagsTestJSON-59142761 tempest-ServerTagsTestJSON-59142761-project-member] [instance: 9029549c-1914-4bae-91e2-8812b79051ec] Took 43.50 seconds to build instance. [ 929.179910] env[65726]: DEBUG oslo_vmware.api [None req-9a0984e5-cc6d-419e-a851-17965236c39e tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52636646-d776-1cc2-fd6d-008d8ab68e24, 'name': SearchDatastore_Task, 'duration_secs': 0.01679} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.180478] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9a0984e5-cc6d-419e-a851-17965236c39e tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 929.180940] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-9a0984e5-cc6d-419e-a851-17965236c39e tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore2] 060646cc-f0fb-4e6c-9f67-7b2daa6b6fd6/060646cc-f0fb-4e6c-9f67-7b2daa6b6fd6.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 929.181661] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-eb7e6c8d-4e80-4ebf-94a9-4c559f8a40f8 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.195683] env[65726]: DEBUG oslo_vmware.api [None req-9a0984e5-cc6d-419e-a851-17965236c39e tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Waiting for the task: (returnval){ [ 929.195683] env[65726]: value = "task-5116268" [ 929.195683] env[65726]: _type = "Task" [ 929.195683] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.210223] env[65726]: DEBUG oslo_vmware.api [None req-9a0984e5-cc6d-419e-a851-17965236c39e tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Task: {'id': task-5116268, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.284203] env[65726]: WARNING openstack [req-00c6ba26-63af-4954-ba22-fecbe93f2e8f req-d3bbc8b8-26ac-4f68-921a-cbb9181e437e service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 929.284203] env[65726]: WARNING openstack [req-00c6ba26-63af-4954-ba22-fecbe93f2e8f req-d3bbc8b8-26ac-4f68-921a-cbb9181e437e service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 929.389367] env[65726]: DEBUG nova.network.neutron [-] [instance: 6f91b053-772a-4497-b29d-349b960c55eb] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 929.390788] env[65726]: DEBUG nova.network.neutron [-] [instance: c4cc4f85-cb35-4edc-a58b-adfee0ce1265] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 929.445964] env[65726]: DEBUG oslo_vmware.api [None req-816608ea-3ce8-4b57-8b12-cbe3b89f21e9 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': task-5116265, 'name': ReconfigVM_Task, 'duration_secs': 0.385693} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.450316] env[65726]: DEBUG nova.network.neutron [req-00c6ba26-63af-4954-ba22-fecbe93f2e8f req-d3bbc8b8-26ac-4f68-921a-cbb9181e437e service nova] [instance: 060646cc-f0fb-4e6c-9f67-7b2daa6b6fd6] Updated VIF entry in instance network info cache for port f923178f-ee20-4582-a2e8-6bfd87020d4b. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 929.450316] env[65726]: DEBUG nova.network.neutron [req-00c6ba26-63af-4954-ba22-fecbe93f2e8f req-d3bbc8b8-26ac-4f68-921a-cbb9181e437e service nova] [instance: 060646cc-f0fb-4e6c-9f67-7b2daa6b6fd6] Updating instance_info_cache with network_info: [{"id": "f923178f-ee20-4582-a2e8-6bfd87020d4b", "address": "fa:16:3e:1e:3b:76", "network": {"id": "ad1373b5-e0ed-4ee6-a97c-af931afbab35", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-909418417-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a090d553766847e58e3231b966c92565", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5f60c972-a72d-4c5f-a250-faadfd6eafbe", "external-id": "nsx-vlan-transportzone-932", "segmentation_id": 932, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf923178f-ee", "ovs_interfaceid": "f923178f-ee20-4582-a2e8-6bfd87020d4b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 929.450607] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-816608ea-3ce8-4b57-8b12-cbe3b89f21e9 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: 5a252ef1-93c9-4bff-842b-b64df2bc5d75] Reconfigured VM instance instance-00000044 to detach disk 2001 {{(pid=65726) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 929.455860] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bf6bc120-7e90-4873-a9ac-cf94255c3766 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.474496] env[65726]: DEBUG oslo_vmware.api [None req-816608ea-3ce8-4b57-8b12-cbe3b89f21e9 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Waiting for the task: (returnval){ [ 929.474496] env[65726]: value = "task-5116269" [ 929.474496] env[65726]: _type = "Task" [ 929.474496] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.484377] env[65726]: DEBUG oslo_vmware.api [None req-816608ea-3ce8-4b57-8b12-cbe3b89f21e9 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': task-5116269, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.493986] env[65726]: DEBUG nova.compute.manager [req-3f037a5c-b0be-4b32-8989-c02bd1ca2a9a req-c9626198-5781-4087-a0e3-eac59c84c3a8 service nova] [instance: dcde25df-87a5-47fb-94ce-334f68894e04] Received event network-changed-bf51c96c-63c5-46de-a40f-58e6769c5fcf {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 929.494214] env[65726]: DEBUG nova.compute.manager [req-3f037a5c-b0be-4b32-8989-c02bd1ca2a9a req-c9626198-5781-4087-a0e3-eac59c84c3a8 service nova] [instance: dcde25df-87a5-47fb-94ce-334f68894e04] Refreshing instance network info cache due to event network-changed-bf51c96c-63c5-46de-a40f-58e6769c5fcf. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 929.494437] env[65726]: DEBUG oslo_concurrency.lockutils [req-3f037a5c-b0be-4b32-8989-c02bd1ca2a9a req-c9626198-5781-4087-a0e3-eac59c84c3a8 service nova] Acquiring lock "refresh_cache-dcde25df-87a5-47fb-94ce-334f68894e04" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 929.494605] env[65726]: DEBUG oslo_concurrency.lockutils [req-3f037a5c-b0be-4b32-8989-c02bd1ca2a9a req-c9626198-5781-4087-a0e3-eac59c84c3a8 service nova] Acquired lock "refresh_cache-dcde25df-87a5-47fb-94ce-334f68894e04" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 929.494719] env[65726]: DEBUG nova.network.neutron [req-3f037a5c-b0be-4b32-8989-c02bd1ca2a9a req-c9626198-5781-4087-a0e3-eac59c84c3a8 service nova] [instance: dcde25df-87a5-47fb-94ce-334f68894e04] Refreshing network info cache for port bf51c96c-63c5-46de-a40f-58e6769c5fcf {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 929.533053] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e9f6f3ba-cd10-4f45-bec2-5e0b6cfdfe85 tempest-ServersWithSpecificFlavorTestJSON-1255420205 tempest-ServersWithSpecificFlavorTestJSON-1255420205-project-member] Lock "6c8db442-daaa-4eca-ae24-b9d4afaf3a71" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 34.706s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 929.619378] env[65726]: DEBUG oslo_vmware.api [None req-2eacb245-5155-4571-979a-d9aaa1a355b7 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Task: {'id': task-5116267, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.220443} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.619681] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-2eacb245-5155-4571-979a-d9aaa1a355b7 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Deleted the datastore file {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 929.619872] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-2eacb245-5155-4571-979a-d9aaa1a355b7 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: c108f2a5-031b-47ee-9a5b-d62c2f42c26b] Deleted contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 929.620098] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-2eacb245-5155-4571-979a-d9aaa1a355b7 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: c108f2a5-031b-47ee-9a5b-d62c2f42c26b] Instance destroyed {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 929.621047] env[65726]: INFO nova.compute.manager [None req-2eacb245-5155-4571-979a-d9aaa1a355b7 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: c108f2a5-031b-47ee-9a5b-d62c2f42c26b] Took 1.17 seconds to destroy the instance on the hypervisor. [ 929.621047] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-2eacb245-5155-4571-979a-d9aaa1a355b7 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 929.621047] env[65726]: DEBUG nova.compute.manager [-] [instance: c108f2a5-031b-47ee-9a5b-d62c2f42c26b] Deallocating network for instance {{(pid=65726) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 929.621047] env[65726]: DEBUG nova.network.neutron [-] [instance: c108f2a5-031b-47ee-9a5b-d62c2f42c26b] deallocate_for_instance() {{(pid=65726) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 929.621453] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 929.621736] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 929.647761] env[65726]: DEBUG oslo_concurrency.lockutils [None req-801b9b62-b5bd-4b5f-80fa-32388dd4cb88 tempest-ServerTagsTestJSON-59142761 tempest-ServerTagsTestJSON-59142761-project-member] Lock "9029549c-1914-4bae-91e2-8812b79051ec" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 58.095s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 929.696396] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 929.696890] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 929.729063] env[65726]: DEBUG oslo_vmware.api [None req-9a0984e5-cc6d-419e-a851-17965236c39e tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Task: {'id': task-5116268, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.893321] env[65726]: INFO nova.compute.manager [-] [instance: 6f91b053-772a-4497-b29d-349b960c55eb] Took 1.84 seconds to deallocate network for instance. [ 929.893813] env[65726]: INFO nova.compute.manager [-] [instance: c4cc4f85-cb35-4edc-a58b-adfee0ce1265] Took 1.51 seconds to deallocate network for instance. [ 929.958912] env[65726]: DEBUG oslo_concurrency.lockutils [req-00c6ba26-63af-4954-ba22-fecbe93f2e8f req-d3bbc8b8-26ac-4f68-921a-cbb9181e437e service nova] Releasing lock "refresh_cache-060646cc-f0fb-4e6c-9f67-7b2daa6b6fd6" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 929.988027] env[65726]: DEBUG oslo_vmware.api [None req-816608ea-3ce8-4b57-8b12-cbe3b89f21e9 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': task-5116269, 'name': ReconfigVM_Task, 'duration_secs': 0.232677} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.988027] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-816608ea-3ce8-4b57-8b12-cbe3b89f21e9 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: 5a252ef1-93c9-4bff-842b-b64df2bc5d75] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-995209', 'volume_id': '2e0dafa1-dc04-4dd8-b30f-16c07c062b44', 'name': 'volume-2e0dafa1-dc04-4dd8-b30f-16c07c062b44', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '5a252ef1-93c9-4bff-842b-b64df2bc5d75', 'attached_at': '', 'detached_at': '', 'volume_id': '2e0dafa1-dc04-4dd8-b30f-16c07c062b44', 'serial': '2e0dafa1-dc04-4dd8-b30f-16c07c062b44'} {{(pid=65726) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 929.988598] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-816608ea-3ce8-4b57-8b12-cbe3b89f21e9 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: 5a252ef1-93c9-4bff-842b-b64df2bc5d75] Destroying instance {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 929.989165] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73564053-0302-4c3a-9063-063136de12b4 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.002023] env[65726]: WARNING openstack [req-3f037a5c-b0be-4b32-8989-c02bd1ca2a9a req-c9626198-5781-4087-a0e3-eac59c84c3a8 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 930.002331] env[65726]: WARNING openstack [req-3f037a5c-b0be-4b32-8989-c02bd1ca2a9a req-c9626198-5781-4087-a0e3-eac59c84c3a8 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 930.008896] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-816608ea-3ce8-4b57-8b12-cbe3b89f21e9 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: 5a252ef1-93c9-4bff-842b-b64df2bc5d75] Unregistering the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 930.010430] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6ccfda74-31df-4dbd-98cc-4a6b16a5cb5a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.077639] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-caf4672d-d21f-4da3-a556-e8263f0a4c1b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.087504] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b6074bf-7423-4c10-8858-23e91f06470d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.131468] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07867483-1bea-430f-aaf8-ce3c94b07370 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.134211] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-816608ea-3ce8-4b57-8b12-cbe3b89f21e9 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: 5a252ef1-93c9-4bff-842b-b64df2bc5d75] Unregistered the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 930.134410] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-816608ea-3ce8-4b57-8b12-cbe3b89f21e9 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: 5a252ef1-93c9-4bff-842b-b64df2bc5d75] Deleting contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 930.134583] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-816608ea-3ce8-4b57-8b12-cbe3b89f21e9 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Deleting the datastore file [datastore1] 5a252ef1-93c9-4bff-842b-b64df2bc5d75 {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 930.135073] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c2a16230-f46b-401f-a41f-0d0b21e25625 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.143808] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2be6df3a-17eb-43f6-a88c-b507409bca9b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.147999] env[65726]: DEBUG oslo_vmware.api [None req-816608ea-3ce8-4b57-8b12-cbe3b89f21e9 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Waiting for the task: (returnval){ [ 930.147999] env[65726]: value = "task-5116271" [ 930.147999] env[65726]: _type = "Task" [ 930.147999] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.159414] env[65726]: DEBUG nova.compute.provider_tree [None req-21960d76-f281-41f6-8236-c885b049da01 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 930.166871] env[65726]: DEBUG oslo_vmware.api [None req-816608ea-3ce8-4b57-8b12-cbe3b89f21e9 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': task-5116271, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.173229] env[65726]: WARNING openstack [req-3f037a5c-b0be-4b32-8989-c02bd1ca2a9a req-c9626198-5781-4087-a0e3-eac59c84c3a8 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 930.173607] env[65726]: WARNING openstack [req-3f037a5c-b0be-4b32-8989-c02bd1ca2a9a req-c9626198-5781-4087-a0e3-eac59c84c3a8 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 930.219596] env[65726]: DEBUG oslo_vmware.api [None req-9a0984e5-cc6d-419e-a851-17965236c39e tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Task: {'id': task-5116268, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.731497} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.219596] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-9a0984e5-cc6d-419e-a851-17965236c39e tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore2] 060646cc-f0fb-4e6c-9f67-7b2daa6b6fd6/060646cc-f0fb-4e6c-9f67-7b2daa6b6fd6.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 930.219881] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-9a0984e5-cc6d-419e-a851-17965236c39e tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: 060646cc-f0fb-4e6c-9f67-7b2daa6b6fd6] Extending root virtual disk to 1048576 {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 930.220197] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d7380294-8866-4a64-a2d9-9beaef733f01 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.229400] env[65726]: DEBUG oslo_vmware.api [None req-9a0984e5-cc6d-419e-a851-17965236c39e tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Waiting for the task: (returnval){ [ 930.229400] env[65726]: value = "task-5116272" [ 930.229400] env[65726]: _type = "Task" [ 930.229400] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.240777] env[65726]: DEBUG oslo_vmware.api [None req-9a0984e5-cc6d-419e-a851-17965236c39e tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Task: {'id': task-5116272, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.265726] env[65726]: WARNING openstack [req-3f037a5c-b0be-4b32-8989-c02bd1ca2a9a req-c9626198-5781-4087-a0e3-eac59c84c3a8 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 930.266962] env[65726]: WARNING openstack [req-3f037a5c-b0be-4b32-8989-c02bd1ca2a9a req-c9626198-5781-4087-a0e3-eac59c84c3a8 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 930.383926] env[65726]: DEBUG nova.network.neutron [req-3f037a5c-b0be-4b32-8989-c02bd1ca2a9a req-c9626198-5781-4087-a0e3-eac59c84c3a8 service nova] [instance: dcde25df-87a5-47fb-94ce-334f68894e04] Updated VIF entry in instance network info cache for port bf51c96c-63c5-46de-a40f-58e6769c5fcf. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 930.384433] env[65726]: DEBUG nova.network.neutron [req-3f037a5c-b0be-4b32-8989-c02bd1ca2a9a req-c9626198-5781-4087-a0e3-eac59c84c3a8 service nova] [instance: dcde25df-87a5-47fb-94ce-334f68894e04] Updating instance_info_cache with network_info: [{"id": "bf51c96c-63c5-46de-a40f-58e6769c5fcf", "address": "fa:16:3e:4c:88:87", "network": {"id": "51a66448-0f72-4266-8868-856b48be839f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-155658961-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.146", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d46ccb00794f458b85da4a93879139ee", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b1f3e6c3-5584-4852-9017-476ab8ac4946", "external-id": "nsx-vlan-transportzone-304", "segmentation_id": 304, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbf51c96c-63", "ovs_interfaceid": "bf51c96c-63c5-46de-a40f-58e6769c5fcf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 930.406415] env[65726]: DEBUG oslo_concurrency.lockutils [None req-1fa711c5-810b-46ee-a4d3-beb8e1bea121 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 930.407564] env[65726]: DEBUG oslo_concurrency.lockutils [None req-7cd9e703-73b1-4767-92f9-46012e80c63d tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 930.516456] env[65726]: DEBUG nova.network.neutron [-] [instance: c108f2a5-031b-47ee-9a5b-d62c2f42c26b] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 930.639783] env[65726]: DEBUG nova.compute.manager [req-1c9b3045-5a59-45d0-8f5f-0bc83507f772 req-1175009b-58ee-401b-b5bc-c1571948cb1e service nova] [instance: 6f91b053-772a-4497-b29d-349b960c55eb] Received event network-vif-deleted-fbccb731-d98d-40ad-a144-4b4f1e48a1ce {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 930.661955] env[65726]: DEBUG oslo_vmware.api [None req-816608ea-3ce8-4b57-8b12-cbe3b89f21e9 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': task-5116271, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.339918} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.661955] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-816608ea-3ce8-4b57-8b12-cbe3b89f21e9 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Deleted the datastore file {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 930.661955] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-816608ea-3ce8-4b57-8b12-cbe3b89f21e9 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: 5a252ef1-93c9-4bff-842b-b64df2bc5d75] Deleted contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 930.661955] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-816608ea-3ce8-4b57-8b12-cbe3b89f21e9 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: 5a252ef1-93c9-4bff-842b-b64df2bc5d75] Instance destroyed {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 930.661955] env[65726]: INFO nova.compute.manager [None req-816608ea-3ce8-4b57-8b12-cbe3b89f21e9 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: 5a252ef1-93c9-4bff-842b-b64df2bc5d75] Took 2.37 seconds to destroy the instance on the hypervisor. [ 930.661955] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-816608ea-3ce8-4b57-8b12-cbe3b89f21e9 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 930.665864] env[65726]: DEBUG nova.scheduler.client.report [None req-21960d76-f281-41f6-8236-c885b049da01 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 930.669229] env[65726]: DEBUG nova.compute.manager [-] [instance: 5a252ef1-93c9-4bff-842b-b64df2bc5d75] Deallocating network for instance {{(pid=65726) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 930.669338] env[65726]: DEBUG nova.network.neutron [-] [instance: 5a252ef1-93c9-4bff-842b-b64df2bc5d75] deallocate_for_instance() {{(pid=65726) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 930.670058] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 930.672295] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 930.716495] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 930.716789] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 930.740283] env[65726]: DEBUG oslo_vmware.api [None req-9a0984e5-cc6d-419e-a851-17965236c39e tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Task: {'id': task-5116272, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.078039} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.740901] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-9a0984e5-cc6d-419e-a851-17965236c39e tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: 060646cc-f0fb-4e6c-9f67-7b2daa6b6fd6] Extended root virtual disk {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 930.741857] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4e9fdf3-509c-4313-94f5-ccfbfef8639f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.769353] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-9a0984e5-cc6d-419e-a851-17965236c39e tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: 060646cc-f0fb-4e6c-9f67-7b2daa6b6fd6] Reconfiguring VM instance instance-00000049 to attach disk [datastore2] 060646cc-f0fb-4e6c-9f67-7b2daa6b6fd6/060646cc-f0fb-4e6c-9f67-7b2daa6b6fd6.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 930.769605] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-06e692c2-bae9-4d8e-9ba0-7b2db2621e52 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.791896] env[65726]: DEBUG oslo_vmware.api [None req-9a0984e5-cc6d-419e-a851-17965236c39e tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Waiting for the task: (returnval){ [ 930.791896] env[65726]: value = "task-5116273" [ 930.791896] env[65726]: _type = "Task" [ 930.791896] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.800681] env[65726]: DEBUG oslo_vmware.api [None req-9a0984e5-cc6d-419e-a851-17965236c39e tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Task: {'id': task-5116273, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.888027] env[65726]: DEBUG oslo_concurrency.lockutils [req-3f037a5c-b0be-4b32-8989-c02bd1ca2a9a req-c9626198-5781-4087-a0e3-eac59c84c3a8 service nova] Releasing lock "refresh_cache-dcde25df-87a5-47fb-94ce-334f68894e04" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 931.019162] env[65726]: INFO nova.compute.manager [-] [instance: c108f2a5-031b-47ee-9a5b-d62c2f42c26b] Took 1.40 seconds to deallocate network for instance. [ 931.179655] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a58f60ab-ca30-4ed0-bbe9-1d319a9c18f8 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] Acquiring lock "73f5b23e-4463-4068-8994-dd2752a2abc9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 931.179938] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a58f60ab-ca30-4ed0-bbe9-1d319a9c18f8 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] Lock "73f5b23e-4463-4068-8994-dd2752a2abc9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 931.305012] env[65726]: DEBUG oslo_vmware.api [None req-9a0984e5-cc6d-419e-a851-17965236c39e tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Task: {'id': task-5116273, 'name': ReconfigVM_Task, 'duration_secs': 0.461203} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.306160] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-9a0984e5-cc6d-419e-a851-17965236c39e tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: 060646cc-f0fb-4e6c-9f67-7b2daa6b6fd6] Reconfigured VM instance instance-00000049 to attach disk [datastore2] 060646cc-f0fb-4e6c-9f67-7b2daa6b6fd6/060646cc-f0fb-4e6c-9f67-7b2daa6b6fd6.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 931.306314] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6586a006-4902-4c2a-82b7-f9c402141e59 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.314697] env[65726]: DEBUG oslo_vmware.api [None req-9a0984e5-cc6d-419e-a851-17965236c39e tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Waiting for the task: (returnval){ [ 931.314697] env[65726]: value = "task-5116274" [ 931.314697] env[65726]: _type = "Task" [ 931.314697] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.324930] env[65726]: DEBUG oslo_vmware.api [None req-9a0984e5-cc6d-419e-a851-17965236c39e tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Task: {'id': task-5116274, 'name': Rename_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.333292] env[65726]: DEBUG nova.compute.manager [None req-2a7383f5-244c-480c-bf85-0a55eabf2c00 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] [instance: 29823bc7-3909-4ab7-8119-91ee59b289c5] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 931.334635] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03b5b164-3ad6-496d-b02f-7935d276ee2d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.468961] env[65726]: DEBUG nova.network.neutron [-] [instance: 5a252ef1-93c9-4bff-842b-b64df2bc5d75] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 931.527617] env[65726]: DEBUG oslo_concurrency.lockutils [None req-2eacb245-5155-4571-979a-d9aaa1a355b7 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 931.552821] env[65726]: DEBUG nova.compute.manager [req-8fd4531f-ab34-4cd3-9aae-a3b365bbd63f req-8fd977ac-f43c-4d61-b7a5-5c56ae612a99 service nova] [instance: c4cc4f85-cb35-4edc-a58b-adfee0ce1265] Received event network-vif-deleted-01aa3d90-3b27-4f98-a3ef-eb316f36f0cc {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 931.553063] env[65726]: DEBUG nova.compute.manager [req-8fd4531f-ab34-4cd3-9aae-a3b365bbd63f req-8fd977ac-f43c-4d61-b7a5-5c56ae612a99 service nova] [instance: c108f2a5-031b-47ee-9a5b-d62c2f42c26b] Received event network-vif-deleted-981741c6-69f0-42d4-87be-e1c8f62d0e17 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 931.682874] env[65726]: DEBUG nova.compute.manager [None req-a58f60ab-ca30-4ed0-bbe9-1d319a9c18f8 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] [instance: 73f5b23e-4463-4068-8994-dd2752a2abc9] Starting instance... {{(pid=65726) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 931.688648] env[65726]: DEBUG oslo_concurrency.lockutils [None req-21960d76-f281-41f6-8236-c885b049da01 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.706s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 931.688991] env[65726]: DEBUG nova.compute.manager [None req-21960d76-f281-41f6-8236-c885b049da01 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 608cc0ea-3f6b-4b4f-83c1-01aa50999d63] Resized/migrated instance is powered off. Setting vm_state to 'stopped'. {{(pid=65726) _confirm_resize /opt/stack/nova/nova/compute/manager.py:5433}} [ 931.695869] env[65726]: DEBUG oslo_concurrency.lockutils [None req-231bdff4-2d2e-4577-9552-cfa1b1d4e74f tempest-ServersTestFqdnHostnames-1699296007 tempest-ServersTestFqdnHostnames-1699296007-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.282s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 931.698347] env[65726]: INFO nova.compute.claims [None req-231bdff4-2d2e-4577-9552-cfa1b1d4e74f tempest-ServersTestFqdnHostnames-1699296007 tempest-ServersTestFqdnHostnames-1699296007-project-member] [instance: 7aa4c0d9-f7ca-42ba-9fd0-4075cd76752b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 931.828315] env[65726]: DEBUG oslo_vmware.api [None req-9a0984e5-cc6d-419e-a851-17965236c39e tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Task: {'id': task-5116274, 'name': Rename_Task, 'duration_secs': 0.347015} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.828618] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-9a0984e5-cc6d-419e-a851-17965236c39e tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: 060646cc-f0fb-4e6c-9f67-7b2daa6b6fd6] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 931.828897] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b43f5696-fda7-4e76-b6c9-c567280e1845 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.836345] env[65726]: DEBUG oslo_vmware.api [None req-9a0984e5-cc6d-419e-a851-17965236c39e tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Waiting for the task: (returnval){ [ 931.836345] env[65726]: value = "task-5116275" [ 931.836345] env[65726]: _type = "Task" [ 931.836345] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.845583] env[65726]: DEBUG oslo_vmware.api [None req-9a0984e5-cc6d-419e-a851-17965236c39e tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Task: {'id': task-5116275, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.847861] env[65726]: INFO nova.compute.manager [None req-2a7383f5-244c-480c-bf85-0a55eabf2c00 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] [instance: 29823bc7-3909-4ab7-8119-91ee59b289c5] instance snapshotting [ 931.850722] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90c85f05-8376-42b1-87a5-5d878c0b36f8 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.872190] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33f2407e-1359-4ff5-ba4c-4a818781eed9 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.972394] env[65726]: INFO nova.compute.manager [-] [instance: 5a252ef1-93c9-4bff-842b-b64df2bc5d75] Took 1.30 seconds to deallocate network for instance. [ 932.231192] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a58f60ab-ca30-4ed0-bbe9-1d319a9c18f8 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 932.282430] env[65726]: INFO nova.scheduler.client.report [None req-21960d76-f281-41f6-8236-c885b049da01 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Deleted allocation for migration 3a215dd8-f815-44fc-8620-fb9bc12739f7 [ 932.347819] env[65726]: DEBUG oslo_vmware.api [None req-9a0984e5-cc6d-419e-a851-17965236c39e tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Task: {'id': task-5116275, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.387144] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-2a7383f5-244c-480c-bf85-0a55eabf2c00 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] [instance: 29823bc7-3909-4ab7-8119-91ee59b289c5] Creating Snapshot of the VM instance {{(pid=65726) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 932.387629] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-0d306200-bf02-4694-9f26-9789c0e1bfda {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.396168] env[65726]: DEBUG oslo_vmware.api [None req-2a7383f5-244c-480c-bf85-0a55eabf2c00 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Waiting for the task: (returnval){ [ 932.396168] env[65726]: value = "task-5116276" [ 932.396168] env[65726]: _type = "Task" [ 932.396168] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 932.405465] env[65726]: DEBUG oslo_vmware.api [None req-2a7383f5-244c-480c-bf85-0a55eabf2c00 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Task: {'id': task-5116276, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.523505] env[65726]: DEBUG oslo_concurrency.lockutils [None req-eb162c35-1e60-4d30-91d4-e0662af06e95 tempest-ServerTagsTestJSON-59142761 tempest-ServerTagsTestJSON-59142761-project-member] Acquiring lock "9029549c-1914-4bae-91e2-8812b79051ec" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 932.523791] env[65726]: DEBUG oslo_concurrency.lockutils [None req-eb162c35-1e60-4d30-91d4-e0662af06e95 tempest-ServerTagsTestJSON-59142761 tempest-ServerTagsTestJSON-59142761-project-member] Lock "9029549c-1914-4bae-91e2-8812b79051ec" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 932.523993] env[65726]: DEBUG oslo_concurrency.lockutils [None req-eb162c35-1e60-4d30-91d4-e0662af06e95 tempest-ServerTagsTestJSON-59142761 tempest-ServerTagsTestJSON-59142761-project-member] Acquiring lock "9029549c-1914-4bae-91e2-8812b79051ec-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 932.524216] env[65726]: DEBUG oslo_concurrency.lockutils [None req-eb162c35-1e60-4d30-91d4-e0662af06e95 tempest-ServerTagsTestJSON-59142761 tempest-ServerTagsTestJSON-59142761-project-member] Lock "9029549c-1914-4bae-91e2-8812b79051ec-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 932.524355] env[65726]: DEBUG oslo_concurrency.lockutils [None req-eb162c35-1e60-4d30-91d4-e0662af06e95 tempest-ServerTagsTestJSON-59142761 tempest-ServerTagsTestJSON-59142761-project-member] Lock "9029549c-1914-4bae-91e2-8812b79051ec-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 932.526858] env[65726]: INFO nova.compute.manager [None req-eb162c35-1e60-4d30-91d4-e0662af06e95 tempest-ServerTagsTestJSON-59142761 tempest-ServerTagsTestJSON-59142761-project-member] [instance: 9029549c-1914-4bae-91e2-8812b79051ec] Terminating instance [ 932.534114] env[65726]: INFO nova.compute.manager [None req-816608ea-3ce8-4b57-8b12-cbe3b89f21e9 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: 5a252ef1-93c9-4bff-842b-b64df2bc5d75] Took 0.56 seconds to detach 1 volumes for instance. [ 932.687947] env[65726]: DEBUG nova.compute.manager [req-01b35fb3-644d-4c3d-bc5a-dc15062d3db6 req-62259cf3-a663-423d-ba2d-164037bbed3a service nova] [instance: 5a252ef1-93c9-4bff-842b-b64df2bc5d75] Received event network-vif-deleted-8cd2f231-f537-4509-92da-dc52e57441ba {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 932.791668] env[65726]: DEBUG oslo_concurrency.lockutils [None req-21960d76-f281-41f6-8236-c885b049da01 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Lock "608cc0ea-3f6b-4b4f-83c1-01aa50999d63" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 33.831s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 932.850296] env[65726]: DEBUG oslo_vmware.api [None req-9a0984e5-cc6d-419e-a851-17965236c39e tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Task: {'id': task-5116275, 'name': PowerOnVM_Task, 'duration_secs': 0.827132} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.850597] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-9a0984e5-cc6d-419e-a851-17965236c39e tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: 060646cc-f0fb-4e6c-9f67-7b2daa6b6fd6] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 932.850820] env[65726]: INFO nova.compute.manager [None req-9a0984e5-cc6d-419e-a851-17965236c39e tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: 060646cc-f0fb-4e6c-9f67-7b2daa6b6fd6] Took 8.67 seconds to spawn the instance on the hypervisor. [ 932.851042] env[65726]: DEBUG nova.compute.manager [None req-9a0984e5-cc6d-419e-a851-17965236c39e tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: 060646cc-f0fb-4e6c-9f67-7b2daa6b6fd6] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 932.852038] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-662aaaf1-d7f7-41a6-8c0b-7d84c027e502 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.907426] env[65726]: DEBUG oslo_vmware.api [None req-2a7383f5-244c-480c-bf85-0a55eabf2c00 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Task: {'id': task-5116276, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.033035] env[65726]: DEBUG nova.compute.manager [None req-eb162c35-1e60-4d30-91d4-e0662af06e95 tempest-ServerTagsTestJSON-59142761 tempest-ServerTagsTestJSON-59142761-project-member] [instance: 9029549c-1914-4bae-91e2-8812b79051ec] Start destroying the instance on the hypervisor. {{(pid=65726) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 933.033035] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-eb162c35-1e60-4d30-91d4-e0662af06e95 tempest-ServerTagsTestJSON-59142761 tempest-ServerTagsTestJSON-59142761-project-member] [instance: 9029549c-1914-4bae-91e2-8812b79051ec] Destroying instance {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 933.033035] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4cb5fc9-688e-4a95-b02c-c381c4c32a2b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.040093] env[65726]: DEBUG oslo_concurrency.lockutils [None req-816608ea-3ce8-4b57-8b12-cbe3b89f21e9 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 933.044563] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-eb162c35-1e60-4d30-91d4-e0662af06e95 tempest-ServerTagsTestJSON-59142761 tempest-ServerTagsTestJSON-59142761-project-member] [instance: 9029549c-1914-4bae-91e2-8812b79051ec] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 933.044822] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f2b2a839-a4d5-4437-ab6a-3956f6e6ca81 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.052633] env[65726]: DEBUG oslo_vmware.api [None req-eb162c35-1e60-4d30-91d4-e0662af06e95 tempest-ServerTagsTestJSON-59142761 tempest-ServerTagsTestJSON-59142761-project-member] Waiting for the task: (returnval){ [ 933.052633] env[65726]: value = "task-5116277" [ 933.052633] env[65726]: _type = "Task" [ 933.052633] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.061953] env[65726]: DEBUG oslo_vmware.api [None req-eb162c35-1e60-4d30-91d4-e0662af06e95 tempest-ServerTagsTestJSON-59142761 tempest-ServerTagsTestJSON-59142761-project-member] Task: {'id': task-5116277, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.146523] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3f7f8bc-a54a-4027-8ea7-b63b1076bcb8 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.155848] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd0551b4-453e-43bf-8808-881bf5e92fe4 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.188754] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca359311-1072-45ef-bc1e-af1ea9ec8360 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.197123] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba9c56e1-adb7-4011-9762-ddc2cbd7290e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.215083] env[65726]: DEBUG nova.compute.provider_tree [None req-231bdff4-2d2e-4577-9552-cfa1b1d4e74f tempest-ServersTestFqdnHostnames-1699296007 tempest-ServersTestFqdnHostnames-1699296007-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 933.375452] env[65726]: INFO nova.compute.manager [None req-9a0984e5-cc6d-419e-a851-17965236c39e tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: 060646cc-f0fb-4e6c-9f67-7b2daa6b6fd6] Took 41.35 seconds to build instance. [ 933.407956] env[65726]: DEBUG oslo_vmware.api [None req-2a7383f5-244c-480c-bf85-0a55eabf2c00 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Task: {'id': task-5116276, 'name': CreateSnapshot_Task, 'duration_secs': 0.565207} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.408305] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-2a7383f5-244c-480c-bf85-0a55eabf2c00 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] [instance: 29823bc7-3909-4ab7-8119-91ee59b289c5] Created Snapshot of the VM instance {{(pid=65726) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 933.409037] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18a8d2aa-3b4a-40c0-a038-481ee0259535 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.563781] env[65726]: DEBUG oslo_vmware.api [None req-eb162c35-1e60-4d30-91d4-e0662af06e95 tempest-ServerTagsTestJSON-59142761 tempest-ServerTagsTestJSON-59142761-project-member] Task: {'id': task-5116277, 'name': PowerOffVM_Task, 'duration_secs': 0.312352} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.565320] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-eb162c35-1e60-4d30-91d4-e0662af06e95 tempest-ServerTagsTestJSON-59142761 tempest-ServerTagsTestJSON-59142761-project-member] [instance: 9029549c-1914-4bae-91e2-8812b79051ec] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 933.565320] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-eb162c35-1e60-4d30-91d4-e0662af06e95 tempest-ServerTagsTestJSON-59142761 tempest-ServerTagsTestJSON-59142761-project-member] [instance: 9029549c-1914-4bae-91e2-8812b79051ec] Unregistering the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 933.565320] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7592b309-4131-4656-aa40-04e26dffcd1b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.634347] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-eb162c35-1e60-4d30-91d4-e0662af06e95 tempest-ServerTagsTestJSON-59142761 tempest-ServerTagsTestJSON-59142761-project-member] [instance: 9029549c-1914-4bae-91e2-8812b79051ec] Unregistered the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 933.634961] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-eb162c35-1e60-4d30-91d4-e0662af06e95 tempest-ServerTagsTestJSON-59142761 tempest-ServerTagsTestJSON-59142761-project-member] [instance: 9029549c-1914-4bae-91e2-8812b79051ec] Deleting contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 933.634961] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-eb162c35-1e60-4d30-91d4-e0662af06e95 tempest-ServerTagsTestJSON-59142761 tempest-ServerTagsTestJSON-59142761-project-member] Deleting the datastore file [datastore1] 9029549c-1914-4bae-91e2-8812b79051ec {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 933.635162] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0beec3ce-0838-4a25-9aaf-c76e9c2c9e2a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.643944] env[65726]: DEBUG oslo_vmware.api [None req-eb162c35-1e60-4d30-91d4-e0662af06e95 tempest-ServerTagsTestJSON-59142761 tempest-ServerTagsTestJSON-59142761-project-member] Waiting for the task: (returnval){ [ 933.643944] env[65726]: value = "task-5116279" [ 933.643944] env[65726]: _type = "Task" [ 933.643944] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.652738] env[65726]: DEBUG oslo_vmware.api [None req-eb162c35-1e60-4d30-91d4-e0662af06e95 tempest-ServerTagsTestJSON-59142761 tempest-ServerTagsTestJSON-59142761-project-member] Task: {'id': task-5116279, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.717512] env[65726]: DEBUG nova.scheduler.client.report [None req-231bdff4-2d2e-4577-9552-cfa1b1d4e74f tempest-ServersTestFqdnHostnames-1699296007 tempest-ServersTestFqdnHostnames-1699296007-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 933.757685] env[65726]: DEBUG nova.objects.instance [None req-2f669eab-49fc-4ee0-87eb-a72d9b593723 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Lazy-loading 'flavor' on Instance uuid 608cc0ea-3f6b-4b4f-83c1-01aa50999d63 {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 933.879718] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9a0984e5-cc6d-419e-a851-17965236c39e tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Lock "060646cc-f0fb-4e6c-9f67-7b2daa6b6fd6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 42.869s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 933.930161] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-2a7383f5-244c-480c-bf85-0a55eabf2c00 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] [instance: 29823bc7-3909-4ab7-8119-91ee59b289c5] Creating linked-clone VM from snapshot {{(pid=65726) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 933.930911] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-1cff5f6f-7006-44b1-b221-e2f26b1dcd1a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.944157] env[65726]: DEBUG oslo_vmware.api [None req-2a7383f5-244c-480c-bf85-0a55eabf2c00 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Waiting for the task: (returnval){ [ 933.944157] env[65726]: value = "task-5116280" [ 933.944157] env[65726]: _type = "Task" [ 933.944157] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.956707] env[65726]: DEBUG oslo_vmware.api [None req-2a7383f5-244c-480c-bf85-0a55eabf2c00 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Task: {'id': task-5116280, 'name': CloneVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.160700] env[65726]: DEBUG oslo_vmware.api [None req-eb162c35-1e60-4d30-91d4-e0662af06e95 tempest-ServerTagsTestJSON-59142761 tempest-ServerTagsTestJSON-59142761-project-member] Task: {'id': task-5116279, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.348286} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.161527] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-eb162c35-1e60-4d30-91d4-e0662af06e95 tempest-ServerTagsTestJSON-59142761 tempest-ServerTagsTestJSON-59142761-project-member] Deleted the datastore file {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 934.161792] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-eb162c35-1e60-4d30-91d4-e0662af06e95 tempest-ServerTagsTestJSON-59142761 tempest-ServerTagsTestJSON-59142761-project-member] [instance: 9029549c-1914-4bae-91e2-8812b79051ec] Deleted contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 934.161997] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-eb162c35-1e60-4d30-91d4-e0662af06e95 tempest-ServerTagsTestJSON-59142761 tempest-ServerTagsTestJSON-59142761-project-member] [instance: 9029549c-1914-4bae-91e2-8812b79051ec] Instance destroyed {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 934.162391] env[65726]: INFO nova.compute.manager [None req-eb162c35-1e60-4d30-91d4-e0662af06e95 tempest-ServerTagsTestJSON-59142761 tempest-ServerTagsTestJSON-59142761-project-member] [instance: 9029549c-1914-4bae-91e2-8812b79051ec] Took 1.13 seconds to destroy the instance on the hypervisor. [ 934.162480] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-eb162c35-1e60-4d30-91d4-e0662af06e95 tempest-ServerTagsTestJSON-59142761 tempest-ServerTagsTestJSON-59142761-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 934.162679] env[65726]: DEBUG nova.compute.manager [-] [instance: 9029549c-1914-4bae-91e2-8812b79051ec] Deallocating network for instance {{(pid=65726) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 934.162777] env[65726]: DEBUG nova.network.neutron [-] [instance: 9029549c-1914-4bae-91e2-8812b79051ec] deallocate_for_instance() {{(pid=65726) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 934.163333] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 934.163544] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 934.227447] env[65726]: DEBUG oslo_concurrency.lockutils [None req-231bdff4-2d2e-4577-9552-cfa1b1d4e74f tempest-ServersTestFqdnHostnames-1699296007 tempest-ServersTestFqdnHostnames-1699296007-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.531s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 934.228228] env[65726]: DEBUG nova.compute.manager [None req-231bdff4-2d2e-4577-9552-cfa1b1d4e74f tempest-ServersTestFqdnHostnames-1699296007 tempest-ServersTestFqdnHostnames-1699296007-project-member] [instance: 7aa4c0d9-f7ca-42ba-9fd0-4075cd76752b] Start building networks asynchronously for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 934.231784] env[65726]: DEBUG oslo_concurrency.lockutils [None req-fa9b16af-8474-4ab3-a554-b5fbaf4d16fd tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 21.246s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 934.232095] env[65726]: DEBUG nova.objects.instance [None req-fa9b16af-8474-4ab3-a554-b5fbaf4d16fd tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Lazy-loading 'resources' on Instance uuid 1cfb1bba-1e21-47eb-b50f-a86575846a65 {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 934.255295] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 934.255612] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 934.267421] env[65726]: DEBUG oslo_concurrency.lockutils [None req-2f669eab-49fc-4ee0-87eb-a72d9b593723 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Acquiring lock "refresh_cache-608cc0ea-3f6b-4b4f-83c1-01aa50999d63" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 934.269023] env[65726]: DEBUG oslo_concurrency.lockutils [None req-2f669eab-49fc-4ee0-87eb-a72d9b593723 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Acquired lock "refresh_cache-608cc0ea-3f6b-4b4f-83c1-01aa50999d63" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 934.269023] env[65726]: DEBUG nova.network.neutron [None req-2f669eab-49fc-4ee0-87eb-a72d9b593723 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 608cc0ea-3f6b-4b4f-83c1-01aa50999d63] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 934.269023] env[65726]: DEBUG nova.objects.instance [None req-2f669eab-49fc-4ee0-87eb-a72d9b593723 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Lazy-loading 'info_cache' on Instance uuid 608cc0ea-3f6b-4b4f-83c1-01aa50999d63 {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 934.456421] env[65726]: DEBUG oslo_vmware.api [None req-2a7383f5-244c-480c-bf85-0a55eabf2c00 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Task: {'id': task-5116280, 'name': CloneVM_Task} progress is 94%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.715887] env[65726]: DEBUG nova.compute.manager [req-77c25f0e-9fac-4754-a067-04e421eacb44 req-7c072834-1e0e-47c1-b1ba-4ae70df07e49 service nova] [instance: 060646cc-f0fb-4e6c-9f67-7b2daa6b6fd6] Received event network-changed-f923178f-ee20-4582-a2e8-6bfd87020d4b {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 934.716723] env[65726]: DEBUG nova.compute.manager [req-77c25f0e-9fac-4754-a067-04e421eacb44 req-7c072834-1e0e-47c1-b1ba-4ae70df07e49 service nova] [instance: 060646cc-f0fb-4e6c-9f67-7b2daa6b6fd6] Refreshing instance network info cache due to event network-changed-f923178f-ee20-4582-a2e8-6bfd87020d4b. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 934.717011] env[65726]: DEBUG oslo_concurrency.lockutils [req-77c25f0e-9fac-4754-a067-04e421eacb44 req-7c072834-1e0e-47c1-b1ba-4ae70df07e49 service nova] Acquiring lock "refresh_cache-060646cc-f0fb-4e6c-9f67-7b2daa6b6fd6" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 934.717180] env[65726]: DEBUG oslo_concurrency.lockutils [req-77c25f0e-9fac-4754-a067-04e421eacb44 req-7c072834-1e0e-47c1-b1ba-4ae70df07e49 service nova] Acquired lock "refresh_cache-060646cc-f0fb-4e6c-9f67-7b2daa6b6fd6" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 934.717348] env[65726]: DEBUG nova.network.neutron [req-77c25f0e-9fac-4754-a067-04e421eacb44 req-7c072834-1e0e-47c1-b1ba-4ae70df07e49 service nova] [instance: 060646cc-f0fb-4e6c-9f67-7b2daa6b6fd6] Refreshing network info cache for port f923178f-ee20-4582-a2e8-6bfd87020d4b {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 934.739719] env[65726]: DEBUG nova.compute.utils [None req-231bdff4-2d2e-4577-9552-cfa1b1d4e74f tempest-ServersTestFqdnHostnames-1699296007 tempest-ServersTestFqdnHostnames-1699296007-project-member] Using /dev/sd instead of None {{(pid=65726) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 934.744154] env[65726]: DEBUG nova.compute.manager [None req-231bdff4-2d2e-4577-9552-cfa1b1d4e74f tempest-ServersTestFqdnHostnames-1699296007 tempest-ServersTestFqdnHostnames-1699296007-project-member] [instance: 7aa4c0d9-f7ca-42ba-9fd0-4075cd76752b] Allocating IP information in the background. {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 934.744154] env[65726]: DEBUG nova.network.neutron [None req-231bdff4-2d2e-4577-9552-cfa1b1d4e74f tempest-ServersTestFqdnHostnames-1699296007 tempest-ServersTestFqdnHostnames-1699296007-project-member] [instance: 7aa4c0d9-f7ca-42ba-9fd0-4075cd76752b] allocate_for_instance() {{(pid=65726) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 934.744154] env[65726]: WARNING neutronclient.v2_0.client [None req-231bdff4-2d2e-4577-9552-cfa1b1d4e74f tempest-ServersTestFqdnHostnames-1699296007 tempest-ServersTestFqdnHostnames-1699296007-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 934.744384] env[65726]: WARNING neutronclient.v2_0.client [None req-231bdff4-2d2e-4577-9552-cfa1b1d4e74f tempest-ServersTestFqdnHostnames-1699296007 tempest-ServersTestFqdnHostnames-1699296007-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 934.745096] env[65726]: WARNING openstack [None req-231bdff4-2d2e-4577-9552-cfa1b1d4e74f tempest-ServersTestFqdnHostnames-1699296007 tempest-ServersTestFqdnHostnames-1699296007-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 934.745416] env[65726]: WARNING openstack [None req-231bdff4-2d2e-4577-9552-cfa1b1d4e74f tempest-ServersTestFqdnHostnames-1699296007 tempest-ServersTestFqdnHostnames-1699296007-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 934.773137] env[65726]: DEBUG nova.objects.base [None req-2f669eab-49fc-4ee0-87eb-a72d9b593723 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Object Instance<608cc0ea-3f6b-4b4f-83c1-01aa50999d63> lazy-loaded attributes: flavor,info_cache {{(pid=65726) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 934.810416] env[65726]: DEBUG nova.policy [None req-231bdff4-2d2e-4577-9552-cfa1b1d4e74f tempest-ServersTestFqdnHostnames-1699296007 tempest-ServersTestFqdnHostnames-1699296007-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b5fa4de46c2541d1859828c02aa25576', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '534461fc8c62483d875b8e165ef0d5c0', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65726) authorize /opt/stack/nova/nova/policy.py:192}} [ 934.957884] env[65726]: DEBUG oslo_vmware.api [None req-2a7383f5-244c-480c-bf85-0a55eabf2c00 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Task: {'id': task-5116280, 'name': CloneVM_Task} progress is 94%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.025884] env[65726]: DEBUG nova.network.neutron [-] [instance: 9029549c-1914-4bae-91e2-8812b79051ec] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 935.214323] env[65726]: DEBUG nova.network.neutron [None req-231bdff4-2d2e-4577-9552-cfa1b1d4e74f tempest-ServersTestFqdnHostnames-1699296007 tempest-ServersTestFqdnHostnames-1699296007-project-member] [instance: 7aa4c0d9-f7ca-42ba-9fd0-4075cd76752b] Successfully created port: 47831805-14e7-41d0-91a6-54a9da3ff27c {{(pid=65726) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 935.220835] env[65726]: WARNING openstack [req-77c25f0e-9fac-4754-a067-04e421eacb44 req-7c072834-1e0e-47c1-b1ba-4ae70df07e49 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 935.221741] env[65726]: WARNING openstack [req-77c25f0e-9fac-4754-a067-04e421eacb44 req-7c072834-1e0e-47c1-b1ba-4ae70df07e49 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 935.246167] env[65726]: DEBUG nova.compute.manager [None req-231bdff4-2d2e-4577-9552-cfa1b1d4e74f tempest-ServersTestFqdnHostnames-1699296007 tempest-ServersTestFqdnHostnames-1699296007-project-member] [instance: 7aa4c0d9-f7ca-42ba-9fd0-4075cd76752b] Start building block device mappings for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 935.279929] env[65726]: WARNING openstack [None req-2f669eab-49fc-4ee0-87eb-a72d9b593723 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 935.279929] env[65726]: WARNING openstack [None req-2f669eab-49fc-4ee0-87eb-a72d9b593723 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 935.307953] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-309f6832-6e55-4c09-a0a3-20ab3ff16c0d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.318023] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-243de4f8-aa1e-4f4b-b0f8-4b8633e233f2 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.357788] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12965a12-8b09-48e5-90ef-dab53f7b7625 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.372041] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8169f49-e9d0-4527-8288-25afba81caf5 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.388514] env[65726]: DEBUG nova.compute.provider_tree [None req-fa9b16af-8474-4ab3-a554-b5fbaf4d16fd tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 935.457708] env[65726]: DEBUG oslo_vmware.api [None req-2a7383f5-244c-480c-bf85-0a55eabf2c00 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Task: {'id': task-5116280, 'name': CloneVM_Task, 'duration_secs': 1.511447} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.458082] env[65726]: INFO nova.virt.vmwareapi.vmops [None req-2a7383f5-244c-480c-bf85-0a55eabf2c00 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] [instance: 29823bc7-3909-4ab7-8119-91ee59b289c5] Created linked-clone VM from snapshot [ 935.458666] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09b26667-0798-4727-8a27-608df3697de8 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.468207] env[65726]: DEBUG nova.virt.vmwareapi.images [None req-2a7383f5-244c-480c-bf85-0a55eabf2c00 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] [instance: 29823bc7-3909-4ab7-8119-91ee59b289c5] Uploading image 49be9d2d-40dd-437d-91c3-541d9773bbee {{(pid=65726) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 935.497922] env[65726]: DEBUG oslo_vmware.rw_handles [None req-2a7383f5-244c-480c-bf85-0a55eabf2c00 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 935.497922] env[65726]: value = "vm-995215" [ 935.497922] env[65726]: _type = "VirtualMachine" [ 935.497922] env[65726]: }. {{(pid=65726) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 935.498259] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-95ce79ed-ff9b-4eb2-8d2d-a938e7a74952 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.506144] env[65726]: DEBUG oslo_vmware.rw_handles [None req-2a7383f5-244c-480c-bf85-0a55eabf2c00 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Lease: (returnval){ [ 935.506144] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]520355c3-de88-0f90-10ae-b06f447ee2f2" [ 935.506144] env[65726]: _type = "HttpNfcLease" [ 935.506144] env[65726]: } obtained for exporting VM: (result){ [ 935.506144] env[65726]: value = "vm-995215" [ 935.506144] env[65726]: _type = "VirtualMachine" [ 935.506144] env[65726]: }. {{(pid=65726) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 935.506434] env[65726]: DEBUG oslo_vmware.api [None req-2a7383f5-244c-480c-bf85-0a55eabf2c00 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Waiting for the lease: (returnval){ [ 935.506434] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]520355c3-de88-0f90-10ae-b06f447ee2f2" [ 935.506434] env[65726]: _type = "HttpNfcLease" [ 935.506434] env[65726]: } to be ready. {{(pid=65726) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 935.515144] env[65726]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 935.515144] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]520355c3-de88-0f90-10ae-b06f447ee2f2" [ 935.515144] env[65726]: _type = "HttpNfcLease" [ 935.515144] env[65726]: } is initializing. {{(pid=65726) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 935.535919] env[65726]: WARNING openstack [req-77c25f0e-9fac-4754-a067-04e421eacb44 req-7c072834-1e0e-47c1-b1ba-4ae70df07e49 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 935.536547] env[65726]: WARNING openstack [req-77c25f0e-9fac-4754-a067-04e421eacb44 req-7c072834-1e0e-47c1-b1ba-4ae70df07e49 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 935.543612] env[65726]: INFO nova.compute.manager [-] [instance: 9029549c-1914-4bae-91e2-8812b79051ec] Took 1.38 seconds to deallocate network for instance. [ 935.614182] env[65726]: WARNING openstack [None req-2f669eab-49fc-4ee0-87eb-a72d9b593723 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 935.614182] env[65726]: WARNING openstack [None req-2f669eab-49fc-4ee0-87eb-a72d9b593723 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 935.821207] env[65726]: WARNING openstack [req-77c25f0e-9fac-4754-a067-04e421eacb44 req-7c072834-1e0e-47c1-b1ba-4ae70df07e49 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 935.821763] env[65726]: WARNING openstack [req-77c25f0e-9fac-4754-a067-04e421eacb44 req-7c072834-1e0e-47c1-b1ba-4ae70df07e49 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 935.861426] env[65726]: WARNING openstack [None req-2f669eab-49fc-4ee0-87eb-a72d9b593723 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 935.861834] env[65726]: WARNING openstack [None req-2f669eab-49fc-4ee0-87eb-a72d9b593723 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 935.895011] env[65726]: DEBUG nova.scheduler.client.report [None req-fa9b16af-8474-4ab3-a554-b5fbaf4d16fd tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 936.013622] env[65726]: DEBUG nova.network.neutron [req-77c25f0e-9fac-4754-a067-04e421eacb44 req-7c072834-1e0e-47c1-b1ba-4ae70df07e49 service nova] [instance: 060646cc-f0fb-4e6c-9f67-7b2daa6b6fd6] Updated VIF entry in instance network info cache for port f923178f-ee20-4582-a2e8-6bfd87020d4b. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 936.013978] env[65726]: DEBUG nova.network.neutron [req-77c25f0e-9fac-4754-a067-04e421eacb44 req-7c072834-1e0e-47c1-b1ba-4ae70df07e49 service nova] [instance: 060646cc-f0fb-4e6c-9f67-7b2daa6b6fd6] Updating instance_info_cache with network_info: [{"id": "f923178f-ee20-4582-a2e8-6bfd87020d4b", "address": "fa:16:3e:1e:3b:76", "network": {"id": "ad1373b5-e0ed-4ee6-a97c-af931afbab35", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-909418417-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.252", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a090d553766847e58e3231b966c92565", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5f60c972-a72d-4c5f-a250-faadfd6eafbe", "external-id": "nsx-vlan-transportzone-932", "segmentation_id": 932, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf923178f-ee", "ovs_interfaceid": "f923178f-ee20-4582-a2e8-6bfd87020d4b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 936.019192] env[65726]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 936.019192] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]520355c3-de88-0f90-10ae-b06f447ee2f2" [ 936.019192] env[65726]: _type = "HttpNfcLease" [ 936.019192] env[65726]: } is ready. {{(pid=65726) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 936.019463] env[65726]: DEBUG oslo_vmware.rw_handles [None req-2a7383f5-244c-480c-bf85-0a55eabf2c00 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 936.019463] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]520355c3-de88-0f90-10ae-b06f447ee2f2" [ 936.019463] env[65726]: _type = "HttpNfcLease" [ 936.019463] env[65726]: }. {{(pid=65726) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 936.020215] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5684316-026b-49f3-93c8-ca8c9d7de9de {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.028298] env[65726]: DEBUG oslo_vmware.rw_handles [None req-2a7383f5-244c-480c-bf85-0a55eabf2c00 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52c522f2-60fc-0fa2-d6d8-5fc55a919213/disk-0.vmdk from lease info. {{(pid=65726) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 936.028478] env[65726]: DEBUG oslo_vmware.rw_handles [None req-2a7383f5-244c-480c-bf85-0a55eabf2c00 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52c522f2-60fc-0fa2-d6d8-5fc55a919213/disk-0.vmdk for reading. {{(pid=65726) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 936.086518] env[65726]: DEBUG oslo_concurrency.lockutils [None req-eb162c35-1e60-4d30-91d4-e0662af06e95 tempest-ServerTagsTestJSON-59142761 tempest-ServerTagsTestJSON-59142761-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 936.116604] env[65726]: DEBUG nova.network.neutron [None req-2f669eab-49fc-4ee0-87eb-a72d9b593723 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 608cc0ea-3f6b-4b4f-83c1-01aa50999d63] Updating instance_info_cache with network_info: [{"id": "e9f7132a-8d5a-47fe-afc6-d45bf3687ea4", "address": "fa:16:3e:fe:4d:df", "network": {"id": "0d0f561d-ac41-4ef5-9926-c707409799f7", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1134321217-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.160", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4eece77569624f90bf64e5c51974173f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2ce62383-8e84-4e26-955b-74c11392f4c9", "external-id": "nsx-vlan-transportzone-215", "segmentation_id": 215, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape9f7132a-8d", "ovs_interfaceid": "e9f7132a-8d5a-47fe-afc6-d45bf3687ea4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 936.127272] env[65726]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-78548c31-e20e-45f6-a7e8-8077bb5309cd {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.258822] env[65726]: DEBUG nova.compute.manager [None req-231bdff4-2d2e-4577-9552-cfa1b1d4e74f tempest-ServersTestFqdnHostnames-1699296007 tempest-ServersTestFqdnHostnames-1699296007-project-member] [instance: 7aa4c0d9-f7ca-42ba-9fd0-4075cd76752b] Start spawning the instance on the hypervisor. {{(pid=65726) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 936.304805] env[65726]: DEBUG nova.virt.hardware [None req-231bdff4-2d2e-4577-9552-cfa1b1d4e74f tempest-ServersTestFqdnHostnames-1699296007 tempest-ServersTestFqdnHostnames-1699296007-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T19:28:34Z,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1539f06fa3534e90acbc3a60c4b8bd3f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T19:28:35Z,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 936.305078] env[65726]: DEBUG nova.virt.hardware [None req-231bdff4-2d2e-4577-9552-cfa1b1d4e74f tempest-ServersTestFqdnHostnames-1699296007 tempest-ServersTestFqdnHostnames-1699296007-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 936.305232] env[65726]: DEBUG nova.virt.hardware [None req-231bdff4-2d2e-4577-9552-cfa1b1d4e74f tempest-ServersTestFqdnHostnames-1699296007 tempest-ServersTestFqdnHostnames-1699296007-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 936.305409] env[65726]: DEBUG nova.virt.hardware [None req-231bdff4-2d2e-4577-9552-cfa1b1d4e74f tempest-ServersTestFqdnHostnames-1699296007 tempest-ServersTestFqdnHostnames-1699296007-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 936.305546] env[65726]: DEBUG nova.virt.hardware [None req-231bdff4-2d2e-4577-9552-cfa1b1d4e74f tempest-ServersTestFqdnHostnames-1699296007 tempest-ServersTestFqdnHostnames-1699296007-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 936.305687] env[65726]: DEBUG nova.virt.hardware [None req-231bdff4-2d2e-4577-9552-cfa1b1d4e74f tempest-ServersTestFqdnHostnames-1699296007 tempest-ServersTestFqdnHostnames-1699296007-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 936.305895] env[65726]: DEBUG nova.virt.hardware [None req-231bdff4-2d2e-4577-9552-cfa1b1d4e74f tempest-ServersTestFqdnHostnames-1699296007 tempest-ServersTestFqdnHostnames-1699296007-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 936.306057] env[65726]: DEBUG nova.virt.hardware [None req-231bdff4-2d2e-4577-9552-cfa1b1d4e74f tempest-ServersTestFqdnHostnames-1699296007 tempest-ServersTestFqdnHostnames-1699296007-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 936.306233] env[65726]: DEBUG nova.virt.hardware [None req-231bdff4-2d2e-4577-9552-cfa1b1d4e74f tempest-ServersTestFqdnHostnames-1699296007 tempest-ServersTestFqdnHostnames-1699296007-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 936.306392] env[65726]: DEBUG nova.virt.hardware [None req-231bdff4-2d2e-4577-9552-cfa1b1d4e74f tempest-ServersTestFqdnHostnames-1699296007 tempest-ServersTestFqdnHostnames-1699296007-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 936.306565] env[65726]: DEBUG nova.virt.hardware [None req-231bdff4-2d2e-4577-9552-cfa1b1d4e74f tempest-ServersTestFqdnHostnames-1699296007 tempest-ServersTestFqdnHostnames-1699296007-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 936.307509] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e460ed0-a7c5-4cc8-99b0-533fcc068894 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.400612] env[65726]: DEBUG oslo_concurrency.lockutils [None req-fa9b16af-8474-4ab3-a554-b5fbaf4d16fd tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.169s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 936.408133] env[65726]: DEBUG oslo_concurrency.lockutils [None req-7acc7b96-c5e1-4976-bfde-d3cb706afaa3 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 22.271s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 936.408375] env[65726]: DEBUG oslo_concurrency.lockutils [None req-7acc7b96-c5e1-4976-bfde-d3cb706afaa3 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 936.410419] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c4dc675f-0696-4840-ac89-d68c37ca6fa4 tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 19.908s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 936.410692] env[65726]: DEBUG nova.objects.instance [None req-c4dc675f-0696-4840-ac89-d68c37ca6fa4 tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Lazy-loading 'resources' on Instance uuid 4ea2b70b-0cfe-4fbf-817a-baa028b8372a {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 936.432824] env[65726]: INFO nova.scheduler.client.report [None req-fa9b16af-8474-4ab3-a554-b5fbaf4d16fd tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Deleted allocations for instance 1cfb1bba-1e21-47eb-b50f-a86575846a65 [ 936.440022] env[65726]: INFO nova.scheduler.client.report [None req-7acc7b96-c5e1-4976-bfde-d3cb706afaa3 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Deleted allocations for instance 76249623-6f83-46a3-b8c5-c001111aa698 [ 936.517137] env[65726]: DEBUG oslo_concurrency.lockutils [req-77c25f0e-9fac-4754-a067-04e421eacb44 req-7c072834-1e0e-47c1-b1ba-4ae70df07e49 service nova] Releasing lock "refresh_cache-060646cc-f0fb-4e6c-9f67-7b2daa6b6fd6" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 936.573062] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d09cd1c-184f-4d4b-8a57-5b0fbf4aace9 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.619114] env[65726]: DEBUG oslo_concurrency.lockutils [None req-2f669eab-49fc-4ee0-87eb-a72d9b593723 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Releasing lock "refresh_cache-608cc0ea-3f6b-4b4f-83c1-01aa50999d63" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 936.767741] env[65726]: DEBUG nova.compute.manager [req-f863fbf8-f235-4cfc-a079-ff67e94ea2f4 req-e3956e71-85b5-4438-8e0f-79c6ab87d035 service nova] [instance: 9029549c-1914-4bae-91e2-8812b79051ec] Received event network-vif-deleted-3104d757-f32f-4bcb-a4f3-fb6b80f1ad5c {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 936.943329] env[65726]: DEBUG oslo_concurrency.lockutils [None req-fa9b16af-8474-4ab3-a554-b5fbaf4d16fd tempest-ServerShowV247Test-1840010593 tempest-ServerShowV247Test-1840010593-project-member] Lock "1cfb1bba-1e21-47eb-b50f-a86575846a65" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 28.787s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 936.960067] env[65726]: DEBUG nova.network.neutron [None req-231bdff4-2d2e-4577-9552-cfa1b1d4e74f tempest-ServersTestFqdnHostnames-1699296007 tempest-ServersTestFqdnHostnames-1699296007-project-member] [instance: 7aa4c0d9-f7ca-42ba-9fd0-4075cd76752b] Successfully updated port: 47831805-14e7-41d0-91a6-54a9da3ff27c {{(pid=65726) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 936.963408] env[65726]: DEBUG oslo_concurrency.lockutils [None req-7acc7b96-c5e1-4976-bfde-d3cb706afaa3 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Lock "76249623-6f83-46a3-b8c5-c001111aa698" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 26.364s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 937.447992] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5209953d-1dcd-45bf-9041-1463f6a7dc20 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.458631] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57104d08-4988-430c-987f-4bf3005eb97d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.496920] env[65726]: DEBUG oslo_concurrency.lockutils [None req-231bdff4-2d2e-4577-9552-cfa1b1d4e74f tempest-ServersTestFqdnHostnames-1699296007 tempest-ServersTestFqdnHostnames-1699296007-project-member] Acquiring lock "refresh_cache-7aa4c0d9-f7ca-42ba-9fd0-4075cd76752b" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 937.497133] env[65726]: DEBUG oslo_concurrency.lockutils [None req-231bdff4-2d2e-4577-9552-cfa1b1d4e74f tempest-ServersTestFqdnHostnames-1699296007 tempest-ServersTestFqdnHostnames-1699296007-project-member] Acquired lock "refresh_cache-7aa4c0d9-f7ca-42ba-9fd0-4075cd76752b" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 937.497309] env[65726]: DEBUG nova.network.neutron [None req-231bdff4-2d2e-4577-9552-cfa1b1d4e74f tempest-ServersTestFqdnHostnames-1699296007 tempest-ServersTestFqdnHostnames-1699296007-project-member] [instance: 7aa4c0d9-f7ca-42ba-9fd0-4075cd76752b] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 937.500210] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8479caa-604a-47e1-8e1d-82bdccbeeffd {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.508957] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88341026-b18f-43f8-a99c-edc100a7469a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.525695] env[65726]: DEBUG nova.compute.provider_tree [None req-c4dc675f-0696-4840-ac89-d68c37ca6fa4 tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 937.625541] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-2f669eab-49fc-4ee0-87eb-a72d9b593723 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 608cc0ea-3f6b-4b4f-83c1-01aa50999d63] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 937.625921] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-93508814-8cd0-4fac-954d-6e23d5c5cd3d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.635025] env[65726]: DEBUG oslo_vmware.api [None req-2f669eab-49fc-4ee0-87eb-a72d9b593723 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Waiting for the task: (returnval){ [ 937.635025] env[65726]: value = "task-5116282" [ 937.635025] env[65726]: _type = "Task" [ 937.635025] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.646038] env[65726]: DEBUG oslo_vmware.api [None req-2f669eab-49fc-4ee0-87eb-a72d9b593723 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5116282, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.674715] env[65726]: DEBUG oslo_concurrency.lockutils [None req-f5777e4a-de17-4332-8f51-ad1e8cbe3f19 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Acquiring lock "0e064341-4e4a-407b-8c26-3eb04b409029" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 937.675045] env[65726]: DEBUG oslo_concurrency.lockutils [None req-f5777e4a-de17-4332-8f51-ad1e8cbe3f19 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Lock "0e064341-4e4a-407b-8c26-3eb04b409029" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 937.675329] env[65726]: DEBUG oslo_concurrency.lockutils [None req-f5777e4a-de17-4332-8f51-ad1e8cbe3f19 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Acquiring lock "0e064341-4e4a-407b-8c26-3eb04b409029-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 937.675545] env[65726]: DEBUG oslo_concurrency.lockutils [None req-f5777e4a-de17-4332-8f51-ad1e8cbe3f19 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Lock "0e064341-4e4a-407b-8c26-3eb04b409029-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 937.675807] env[65726]: DEBUG oslo_concurrency.lockutils [None req-f5777e4a-de17-4332-8f51-ad1e8cbe3f19 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Lock "0e064341-4e4a-407b-8c26-3eb04b409029-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 937.678990] env[65726]: INFO nova.compute.manager [None req-f5777e4a-de17-4332-8f51-ad1e8cbe3f19 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] [instance: 0e064341-4e4a-407b-8c26-3eb04b409029] Terminating instance [ 938.000673] env[65726]: WARNING openstack [None req-231bdff4-2d2e-4577-9552-cfa1b1d4e74f tempest-ServersTestFqdnHostnames-1699296007 tempest-ServersTestFqdnHostnames-1699296007-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 938.001620] env[65726]: WARNING openstack [None req-231bdff4-2d2e-4577-9552-cfa1b1d4e74f tempest-ServersTestFqdnHostnames-1699296007 tempest-ServersTestFqdnHostnames-1699296007-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 938.009191] env[65726]: DEBUG nova.network.neutron [None req-231bdff4-2d2e-4577-9552-cfa1b1d4e74f tempest-ServersTestFqdnHostnames-1699296007 tempest-ServersTestFqdnHostnames-1699296007-project-member] [instance: 7aa4c0d9-f7ca-42ba-9fd0-4075cd76752b] Instance cache missing network info. {{(pid=65726) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 938.029198] env[65726]: DEBUG nova.scheduler.client.report [None req-c4dc675f-0696-4840-ac89-d68c37ca6fa4 tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 938.067815] env[65726]: WARNING openstack [None req-231bdff4-2d2e-4577-9552-cfa1b1d4e74f tempest-ServersTestFqdnHostnames-1699296007 tempest-ServersTestFqdnHostnames-1699296007-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 938.068353] env[65726]: WARNING openstack [None req-231bdff4-2d2e-4577-9552-cfa1b1d4e74f tempest-ServersTestFqdnHostnames-1699296007 tempest-ServersTestFqdnHostnames-1699296007-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 938.147570] env[65726]: DEBUG oslo_vmware.api [None req-2f669eab-49fc-4ee0-87eb-a72d9b593723 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5116282, 'name': PowerOnVM_Task} progress is 90%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.154346] env[65726]: WARNING openstack [None req-231bdff4-2d2e-4577-9552-cfa1b1d4e74f tempest-ServersTestFqdnHostnames-1699296007 tempest-ServersTestFqdnHostnames-1699296007-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 938.155313] env[65726]: WARNING openstack [None req-231bdff4-2d2e-4577-9552-cfa1b1d4e74f tempest-ServersTestFqdnHostnames-1699296007 tempest-ServersTestFqdnHostnames-1699296007-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 938.184360] env[65726]: DEBUG nova.compute.manager [None req-f5777e4a-de17-4332-8f51-ad1e8cbe3f19 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] [instance: 0e064341-4e4a-407b-8c26-3eb04b409029] Start destroying the instance on the hypervisor. {{(pid=65726) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 938.185313] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-f5777e4a-de17-4332-8f51-ad1e8cbe3f19 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] [instance: 0e064341-4e4a-407b-8c26-3eb04b409029] Destroying instance {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 938.186012] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a3108e2-50da-4947-91ef-7425b28e136a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.196487] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-f5777e4a-de17-4332-8f51-ad1e8cbe3f19 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] [instance: 0e064341-4e4a-407b-8c26-3eb04b409029] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 938.196902] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ae849835-84b0-4e30-a0bd-125b68b390e6 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.206530] env[65726]: DEBUG oslo_vmware.api [None req-f5777e4a-de17-4332-8f51-ad1e8cbe3f19 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Waiting for the task: (returnval){ [ 938.206530] env[65726]: value = "task-5116283" [ 938.206530] env[65726]: _type = "Task" [ 938.206530] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.218491] env[65726]: DEBUG oslo_vmware.api [None req-f5777e4a-de17-4332-8f51-ad1e8cbe3f19 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Task: {'id': task-5116283, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.307659] env[65726]: DEBUG nova.network.neutron [None req-231bdff4-2d2e-4577-9552-cfa1b1d4e74f tempest-ServersTestFqdnHostnames-1699296007 tempest-ServersTestFqdnHostnames-1699296007-project-member] [instance: 7aa4c0d9-f7ca-42ba-9fd0-4075cd76752b] Updating instance_info_cache with network_info: [{"id": "47831805-14e7-41d0-91a6-54a9da3ff27c", "address": "fa:16:3e:b0:42:a2", "network": {"id": "e7826d02-16e9-49e4-8794-b5b7870aa663", "bridge": "br-int", "label": "tempest-ServersTestFqdnHostnames-212080893-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "534461fc8c62483d875b8e165ef0d5c0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f625f389-b7cf-49b9-998a-87f3a9e3f234", "external-id": "nsx-vlan-transportzone-280", "segmentation_id": 280, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap47831805-14", "ovs_interfaceid": "47831805-14e7-41d0-91a6-54a9da3ff27c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 938.538969] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c4dc675f-0696-4840-ac89-d68c37ca6fa4 tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.128s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 938.542241] env[65726]: DEBUG oslo_concurrency.lockutils [None req-2eeb1a84-0a41-40be-a094-58835ef7befd tempest-ServerAddressesTestJSON-569100133 tempest-ServerAddressesTestJSON-569100133-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 21.615s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 938.542619] env[65726]: DEBUG nova.objects.instance [None req-2eeb1a84-0a41-40be-a094-58835ef7befd tempest-ServerAddressesTestJSON-569100133 tempest-ServerAddressesTestJSON-569100133-project-member] Lazy-loading 'resources' on Instance uuid 70375746-06dd-4710-9136-95e9e5759a0e {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 938.569102] env[65726]: INFO nova.scheduler.client.report [None req-c4dc675f-0696-4840-ac89-d68c37ca6fa4 tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Deleted allocations for instance 4ea2b70b-0cfe-4fbf-817a-baa028b8372a [ 938.647894] env[65726]: DEBUG oslo_vmware.api [None req-2f669eab-49fc-4ee0-87eb-a72d9b593723 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5116282, 'name': PowerOnVM_Task, 'duration_secs': 0.624894} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.648522] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-2f669eab-49fc-4ee0-87eb-a72d9b593723 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 608cc0ea-3f6b-4b4f-83c1-01aa50999d63] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 938.648783] env[65726]: DEBUG nova.compute.manager [None req-2f669eab-49fc-4ee0-87eb-a72d9b593723 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 608cc0ea-3f6b-4b4f-83c1-01aa50999d63] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 938.649654] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07743761-73f1-4f3b-92a6-6f3372208886 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.719462] env[65726]: DEBUG oslo_vmware.api [None req-f5777e4a-de17-4332-8f51-ad1e8cbe3f19 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Task: {'id': task-5116283, 'name': PowerOffVM_Task, 'duration_secs': 0.274857} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.719855] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-f5777e4a-de17-4332-8f51-ad1e8cbe3f19 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] [instance: 0e064341-4e4a-407b-8c26-3eb04b409029] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 938.720051] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-f5777e4a-de17-4332-8f51-ad1e8cbe3f19 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] [instance: 0e064341-4e4a-407b-8c26-3eb04b409029] Unregistering the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 938.720388] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-60d34056-5fb0-4270-8c0e-6035df5c58a5 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.802991] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-f5777e4a-de17-4332-8f51-ad1e8cbe3f19 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] [instance: 0e064341-4e4a-407b-8c26-3eb04b409029] Unregistered the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 938.803334] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-f5777e4a-de17-4332-8f51-ad1e8cbe3f19 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] [instance: 0e064341-4e4a-407b-8c26-3eb04b409029] Deleting contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 938.803580] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-f5777e4a-de17-4332-8f51-ad1e8cbe3f19 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Deleting the datastore file [datastore1] 0e064341-4e4a-407b-8c26-3eb04b409029 {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 938.804016] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-df06c24f-3ab5-4dda-ae19-451ceb624fbe {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.812748] env[65726]: DEBUG oslo_concurrency.lockutils [None req-231bdff4-2d2e-4577-9552-cfa1b1d4e74f tempest-ServersTestFqdnHostnames-1699296007 tempest-ServersTestFqdnHostnames-1699296007-project-member] Releasing lock "refresh_cache-7aa4c0d9-f7ca-42ba-9fd0-4075cd76752b" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 938.813170] env[65726]: DEBUG nova.compute.manager [None req-231bdff4-2d2e-4577-9552-cfa1b1d4e74f tempest-ServersTestFqdnHostnames-1699296007 tempest-ServersTestFqdnHostnames-1699296007-project-member] [instance: 7aa4c0d9-f7ca-42ba-9fd0-4075cd76752b] Instance network_info: |[{"id": "47831805-14e7-41d0-91a6-54a9da3ff27c", "address": "fa:16:3e:b0:42:a2", "network": {"id": "e7826d02-16e9-49e4-8794-b5b7870aa663", "bridge": "br-int", "label": "tempest-ServersTestFqdnHostnames-212080893-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "534461fc8c62483d875b8e165ef0d5c0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f625f389-b7cf-49b9-998a-87f3a9e3f234", "external-id": "nsx-vlan-transportzone-280", "segmentation_id": 280, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap47831805-14", "ovs_interfaceid": "47831805-14e7-41d0-91a6-54a9da3ff27c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 938.813770] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-231bdff4-2d2e-4577-9552-cfa1b1d4e74f tempest-ServersTestFqdnHostnames-1699296007 tempest-ServersTestFqdnHostnames-1699296007-project-member] [instance: 7aa4c0d9-f7ca-42ba-9fd0-4075cd76752b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b0:42:a2', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f625f389-b7cf-49b9-998a-87f3a9e3f234', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '47831805-14e7-41d0-91a6-54a9da3ff27c', 'vif_model': 'vmxnet3'}] {{(pid=65726) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 938.825296] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-231bdff4-2d2e-4577-9552-cfa1b1d4e74f tempest-ServersTestFqdnHostnames-1699296007 tempest-ServersTestFqdnHostnames-1699296007-project-member] Creating folder: Project (534461fc8c62483d875b8e165ef0d5c0). Parent ref: group-v995008. {{(pid=65726) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 938.827163] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-df6d0565-4a95-4fed-9668-9f644a618a09 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.829598] env[65726]: DEBUG oslo_vmware.api [None req-f5777e4a-de17-4332-8f51-ad1e8cbe3f19 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Waiting for the task: (returnval){ [ 938.829598] env[65726]: value = "task-5116285" [ 938.829598] env[65726]: _type = "Task" [ 938.829598] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.842905] env[65726]: DEBUG oslo_vmware.api [None req-f5777e4a-de17-4332-8f51-ad1e8cbe3f19 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Task: {'id': task-5116285, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.844807] env[65726]: INFO nova.virt.vmwareapi.vm_util [None req-231bdff4-2d2e-4577-9552-cfa1b1d4e74f tempest-ServersTestFqdnHostnames-1699296007 tempest-ServersTestFqdnHostnames-1699296007-project-member] Created folder: Project (534461fc8c62483d875b8e165ef0d5c0) in parent group-v995008. [ 938.845009] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-231bdff4-2d2e-4577-9552-cfa1b1d4e74f tempest-ServersTestFqdnHostnames-1699296007 tempest-ServersTestFqdnHostnames-1699296007-project-member] Creating folder: Instances. Parent ref: group-v995216. {{(pid=65726) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 938.845302] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-6bf36012-35b0-420f-bd43-50ef10ec5d18 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.860112] env[65726]: INFO nova.virt.vmwareapi.vm_util [None req-231bdff4-2d2e-4577-9552-cfa1b1d4e74f tempest-ServersTestFqdnHostnames-1699296007 tempest-ServersTestFqdnHostnames-1699296007-project-member] Created folder: Instances in parent group-v995216. [ 938.860112] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-231bdff4-2d2e-4577-9552-cfa1b1d4e74f tempest-ServersTestFqdnHostnames-1699296007 tempest-ServersTestFqdnHostnames-1699296007-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 938.860332] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7aa4c0d9-f7ca-42ba-9fd0-4075cd76752b] Creating VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 938.860398] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-bfdea6b4-8682-49c1-98d5-722adf114c41 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.887449] env[65726]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 938.887449] env[65726]: value = "task-5116288" [ 938.887449] env[65726]: _type = "Task" [ 938.887449] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.897436] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116288, 'name': CreateVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.034418] env[65726]: DEBUG nova.compute.manager [req-128ee42c-f918-43e6-96e0-6b5511cc7ea4 req-d64e1b52-5b22-4859-8bfa-4ac4f36a3a1c service nova] [instance: 7aa4c0d9-f7ca-42ba-9fd0-4075cd76752b] Received event network-vif-plugged-47831805-14e7-41d0-91a6-54a9da3ff27c {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 939.034782] env[65726]: DEBUG oslo_concurrency.lockutils [req-128ee42c-f918-43e6-96e0-6b5511cc7ea4 req-d64e1b52-5b22-4859-8bfa-4ac4f36a3a1c service nova] Acquiring lock "7aa4c0d9-f7ca-42ba-9fd0-4075cd76752b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 939.035050] env[65726]: DEBUG oslo_concurrency.lockutils [req-128ee42c-f918-43e6-96e0-6b5511cc7ea4 req-d64e1b52-5b22-4859-8bfa-4ac4f36a3a1c service nova] Lock "7aa4c0d9-f7ca-42ba-9fd0-4075cd76752b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 939.035342] env[65726]: DEBUG oslo_concurrency.lockutils [req-128ee42c-f918-43e6-96e0-6b5511cc7ea4 req-d64e1b52-5b22-4859-8bfa-4ac4f36a3a1c service nova] Lock "7aa4c0d9-f7ca-42ba-9fd0-4075cd76752b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 939.035631] env[65726]: DEBUG nova.compute.manager [req-128ee42c-f918-43e6-96e0-6b5511cc7ea4 req-d64e1b52-5b22-4859-8bfa-4ac4f36a3a1c service nova] [instance: 7aa4c0d9-f7ca-42ba-9fd0-4075cd76752b] No waiting events found dispatching network-vif-plugged-47831805-14e7-41d0-91a6-54a9da3ff27c {{(pid=65726) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 939.035912] env[65726]: WARNING nova.compute.manager [req-128ee42c-f918-43e6-96e0-6b5511cc7ea4 req-d64e1b52-5b22-4859-8bfa-4ac4f36a3a1c service nova] [instance: 7aa4c0d9-f7ca-42ba-9fd0-4075cd76752b] Received unexpected event network-vif-plugged-47831805-14e7-41d0-91a6-54a9da3ff27c for instance with vm_state building and task_state spawning. [ 939.036195] env[65726]: DEBUG nova.compute.manager [req-128ee42c-f918-43e6-96e0-6b5511cc7ea4 req-d64e1b52-5b22-4859-8bfa-4ac4f36a3a1c service nova] [instance: 7aa4c0d9-f7ca-42ba-9fd0-4075cd76752b] Received event network-changed-47831805-14e7-41d0-91a6-54a9da3ff27c {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 939.036404] env[65726]: DEBUG nova.compute.manager [req-128ee42c-f918-43e6-96e0-6b5511cc7ea4 req-d64e1b52-5b22-4859-8bfa-4ac4f36a3a1c service nova] [instance: 7aa4c0d9-f7ca-42ba-9fd0-4075cd76752b] Refreshing instance network info cache due to event network-changed-47831805-14e7-41d0-91a6-54a9da3ff27c. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 939.036738] env[65726]: DEBUG oslo_concurrency.lockutils [req-128ee42c-f918-43e6-96e0-6b5511cc7ea4 req-d64e1b52-5b22-4859-8bfa-4ac4f36a3a1c service nova] Acquiring lock "refresh_cache-7aa4c0d9-f7ca-42ba-9fd0-4075cd76752b" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 939.036932] env[65726]: DEBUG oslo_concurrency.lockutils [req-128ee42c-f918-43e6-96e0-6b5511cc7ea4 req-d64e1b52-5b22-4859-8bfa-4ac4f36a3a1c service nova] Acquired lock "refresh_cache-7aa4c0d9-f7ca-42ba-9fd0-4075cd76752b" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 939.037159] env[65726]: DEBUG nova.network.neutron [req-128ee42c-f918-43e6-96e0-6b5511cc7ea4 req-d64e1b52-5b22-4859-8bfa-4ac4f36a3a1c service nova] [instance: 7aa4c0d9-f7ca-42ba-9fd0-4075cd76752b] Refreshing network info cache for port 47831805-14e7-41d0-91a6-54a9da3ff27c {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 939.080368] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c4dc675f-0696-4840-ac89-d68c37ca6fa4 tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Lock "4ea2b70b-0cfe-4fbf-817a-baa028b8372a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 26.217s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 939.347999] env[65726]: DEBUG oslo_vmware.api [None req-f5777e4a-de17-4332-8f51-ad1e8cbe3f19 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Task: {'id': task-5116285, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.266004} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.347999] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-f5777e4a-de17-4332-8f51-ad1e8cbe3f19 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Deleted the datastore file {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 939.347999] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-f5777e4a-de17-4332-8f51-ad1e8cbe3f19 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] [instance: 0e064341-4e4a-407b-8c26-3eb04b409029] Deleted contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 939.347999] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-f5777e4a-de17-4332-8f51-ad1e8cbe3f19 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] [instance: 0e064341-4e4a-407b-8c26-3eb04b409029] Instance destroyed {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 939.348421] env[65726]: INFO nova.compute.manager [None req-f5777e4a-de17-4332-8f51-ad1e8cbe3f19 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] [instance: 0e064341-4e4a-407b-8c26-3eb04b409029] Took 1.16 seconds to destroy the instance on the hypervisor. [ 939.348516] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-f5777e4a-de17-4332-8f51-ad1e8cbe3f19 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 939.348804] env[65726]: DEBUG nova.compute.manager [-] [instance: 0e064341-4e4a-407b-8c26-3eb04b409029] Deallocating network for instance {{(pid=65726) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 939.348923] env[65726]: DEBUG nova.network.neutron [-] [instance: 0e064341-4e4a-407b-8c26-3eb04b409029] deallocate_for_instance() {{(pid=65726) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 939.351473] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 939.353866] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 939.401024] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116288, 'name': CreateVM_Task} progress is 99%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.413114] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 939.413506] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 939.517853] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b1f1b0b-9dfe-43f7-949a-5d6de37c5f69 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.530053] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed9b0dca-8a04-4314-bbf0-7d90d324dc97 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.567699] env[65726]: WARNING openstack [req-128ee42c-f918-43e6-96e0-6b5511cc7ea4 req-d64e1b52-5b22-4859-8bfa-4ac4f36a3a1c service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 939.568260] env[65726]: WARNING openstack [req-128ee42c-f918-43e6-96e0-6b5511cc7ea4 req-d64e1b52-5b22-4859-8bfa-4ac4f36a3a1c service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 939.577633] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b2b1da6-b279-47c0-837d-ec827b9bae79 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.587417] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-222702b7-71b5-46ee-885c-d823e5ddf216 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.614174] env[65726]: DEBUG nova.compute.provider_tree [None req-2eeb1a84-0a41-40be-a094-58835ef7befd tempest-ServerAddressesTestJSON-569100133 tempest-ServerAddressesTestJSON-569100133-project-member] Updating inventory in ProviderTree for provider 07c4692f-bdb4-4058-9173-ff9664830295 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 939.715461] env[65726]: WARNING openstack [req-128ee42c-f918-43e6-96e0-6b5511cc7ea4 req-d64e1b52-5b22-4859-8bfa-4ac4f36a3a1c service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 939.716049] env[65726]: WARNING openstack [req-128ee42c-f918-43e6-96e0-6b5511cc7ea4 req-d64e1b52-5b22-4859-8bfa-4ac4f36a3a1c service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 939.807605] env[65726]: WARNING openstack [req-128ee42c-f918-43e6-96e0-6b5511cc7ea4 req-d64e1b52-5b22-4859-8bfa-4ac4f36a3a1c service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 939.807605] env[65726]: WARNING openstack [req-128ee42c-f918-43e6-96e0-6b5511cc7ea4 req-d64e1b52-5b22-4859-8bfa-4ac4f36a3a1c service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 939.898906] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116288, 'name': CreateVM_Task, 'duration_secs': 0.512947} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.899170] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7aa4c0d9-f7ca-42ba-9fd0-4075cd76752b] Created VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 939.899990] env[65726]: WARNING openstack [None req-231bdff4-2d2e-4577-9552-cfa1b1d4e74f tempest-ServersTestFqdnHostnames-1699296007 tempest-ServersTestFqdnHostnames-1699296007-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 939.900399] env[65726]: WARNING openstack [None req-231bdff4-2d2e-4577-9552-cfa1b1d4e74f tempest-ServersTestFqdnHostnames-1699296007 tempest-ServersTestFqdnHostnames-1699296007-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 939.908050] env[65726]: DEBUG oslo_concurrency.lockutils [None req-231bdff4-2d2e-4577-9552-cfa1b1d4e74f tempest-ServersTestFqdnHostnames-1699296007 tempest-ServersTestFqdnHostnames-1699296007-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 939.908423] env[65726]: DEBUG oslo_concurrency.lockutils [None req-231bdff4-2d2e-4577-9552-cfa1b1d4e74f tempest-ServersTestFqdnHostnames-1699296007 tempest-ServersTestFqdnHostnames-1699296007-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 939.908790] env[65726]: DEBUG oslo_concurrency.lockutils [None req-231bdff4-2d2e-4577-9552-cfa1b1d4e74f tempest-ServersTestFqdnHostnames-1699296007 tempest-ServersTestFqdnHostnames-1699296007-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 939.912192] env[65726]: DEBUG nova.network.neutron [req-128ee42c-f918-43e6-96e0-6b5511cc7ea4 req-d64e1b52-5b22-4859-8bfa-4ac4f36a3a1c service nova] [instance: 7aa4c0d9-f7ca-42ba-9fd0-4075cd76752b] Updated VIF entry in instance network info cache for port 47831805-14e7-41d0-91a6-54a9da3ff27c. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 939.912295] env[65726]: DEBUG nova.network.neutron [req-128ee42c-f918-43e6-96e0-6b5511cc7ea4 req-d64e1b52-5b22-4859-8bfa-4ac4f36a3a1c service nova] [instance: 7aa4c0d9-f7ca-42ba-9fd0-4075cd76752b] Updating instance_info_cache with network_info: [{"id": "47831805-14e7-41d0-91a6-54a9da3ff27c", "address": "fa:16:3e:b0:42:a2", "network": {"id": "e7826d02-16e9-49e4-8794-b5b7870aa663", "bridge": "br-int", "label": "tempest-ServersTestFqdnHostnames-212080893-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "534461fc8c62483d875b8e165ef0d5c0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f625f389-b7cf-49b9-998a-87f3a9e3f234", "external-id": "nsx-vlan-transportzone-280", "segmentation_id": 280, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap47831805-14", "ovs_interfaceid": "47831805-14e7-41d0-91a6-54a9da3ff27c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 939.913626] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-80260e95-9d44-4f80-a6a5-9f1af945333f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.921464] env[65726]: DEBUG oslo_vmware.api [None req-231bdff4-2d2e-4577-9552-cfa1b1d4e74f tempest-ServersTestFqdnHostnames-1699296007 tempest-ServersTestFqdnHostnames-1699296007-project-member] Waiting for the task: (returnval){ [ 939.921464] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52908bbd-d371-9a3f-5872-0bae35887a6b" [ 939.921464] env[65726]: _type = "Task" [ 939.921464] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.932019] env[65726]: DEBUG oslo_vmware.api [None req-231bdff4-2d2e-4577-9552-cfa1b1d4e74f tempest-ServersTestFqdnHostnames-1699296007 tempest-ServersTestFqdnHostnames-1699296007-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52908bbd-d371-9a3f-5872-0bae35887a6b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.104040] env[65726]: DEBUG oslo_concurrency.lockutils [None req-98994c54-47ac-4003-a1d5-85175217e75d tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Acquiring lock "608cc0ea-3f6b-4b4f-83c1-01aa50999d63" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 940.104339] env[65726]: DEBUG oslo_concurrency.lockutils [None req-98994c54-47ac-4003-a1d5-85175217e75d tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Lock "608cc0ea-3f6b-4b4f-83c1-01aa50999d63" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 940.104548] env[65726]: DEBUG oslo_concurrency.lockutils [None req-98994c54-47ac-4003-a1d5-85175217e75d tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Acquiring lock "608cc0ea-3f6b-4b4f-83c1-01aa50999d63-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 940.104725] env[65726]: DEBUG oslo_concurrency.lockutils [None req-98994c54-47ac-4003-a1d5-85175217e75d tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Lock "608cc0ea-3f6b-4b4f-83c1-01aa50999d63-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 940.104895] env[65726]: DEBUG oslo_concurrency.lockutils [None req-98994c54-47ac-4003-a1d5-85175217e75d tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Lock "608cc0ea-3f6b-4b4f-83c1-01aa50999d63-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 940.109137] env[65726]: INFO nova.compute.manager [None req-98994c54-47ac-4003-a1d5-85175217e75d tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 608cc0ea-3f6b-4b4f-83c1-01aa50999d63] Terminating instance [ 940.139912] env[65726]: ERROR nova.scheduler.client.report [None req-2eeb1a84-0a41-40be-a094-58835ef7befd tempest-ServerAddressesTestJSON-569100133 tempest-ServerAddressesTestJSON-569100133-project-member] [req-2ac3cf13-7a1e-42c4-90f6-b783fb126862] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 07c4692f-bdb4-4058-9173-ff9664830295. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-2ac3cf13-7a1e-42c4-90f6-b783fb126862"}]} [ 940.165899] env[65726]: DEBUG nova.scheduler.client.report [None req-2eeb1a84-0a41-40be-a094-58835ef7befd tempest-ServerAddressesTestJSON-569100133 tempest-ServerAddressesTestJSON-569100133-project-member] Refreshing inventories for resource provider 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 940.184256] env[65726]: DEBUG nova.scheduler.client.report [None req-2eeb1a84-0a41-40be-a094-58835ef7befd tempest-ServerAddressesTestJSON-569100133 tempest-ServerAddressesTestJSON-569100133-project-member] Updating ProviderTree inventory for provider 07c4692f-bdb4-4058-9173-ff9664830295 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 940.184575] env[65726]: DEBUG nova.compute.provider_tree [None req-2eeb1a84-0a41-40be-a094-58835ef7befd tempest-ServerAddressesTestJSON-569100133 tempest-ServerAddressesTestJSON-569100133-project-member] Updating inventory in ProviderTree for provider 07c4692f-bdb4-4058-9173-ff9664830295 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 940.197416] env[65726]: DEBUG nova.network.neutron [-] [instance: 0e064341-4e4a-407b-8c26-3eb04b409029] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 940.209517] env[65726]: DEBUG nova.scheduler.client.report [None req-2eeb1a84-0a41-40be-a094-58835ef7befd tempest-ServerAddressesTestJSON-569100133 tempest-ServerAddressesTestJSON-569100133-project-member] Refreshing aggregate associations for resource provider 07c4692f-bdb4-4058-9173-ff9664830295, aggregates: None {{(pid=65726) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 940.249120] env[65726]: DEBUG nova.scheduler.client.report [None req-2eeb1a84-0a41-40be-a094-58835ef7befd tempest-ServerAddressesTestJSON-569100133 tempest-ServerAddressesTestJSON-569100133-project-member] Refreshing trait associations for resource provider 07c4692f-bdb4-4058-9173-ff9664830295, traits: COMPUTE_IMAGE_TYPE_ISO,HW_ARCH_X86_64,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=65726) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 940.419712] env[65726]: DEBUG oslo_concurrency.lockutils [req-128ee42c-f918-43e6-96e0-6b5511cc7ea4 req-d64e1b52-5b22-4859-8bfa-4ac4f36a3a1c service nova] Releasing lock "refresh_cache-7aa4c0d9-f7ca-42ba-9fd0-4075cd76752b" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 940.441258] env[65726]: DEBUG oslo_vmware.api [None req-231bdff4-2d2e-4577-9552-cfa1b1d4e74f tempest-ServersTestFqdnHostnames-1699296007 tempest-ServersTestFqdnHostnames-1699296007-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52908bbd-d371-9a3f-5872-0bae35887a6b, 'name': SearchDatastore_Task, 'duration_secs': 0.016461} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.441636] env[65726]: DEBUG oslo_concurrency.lockutils [None req-231bdff4-2d2e-4577-9552-cfa1b1d4e74f tempest-ServersTestFqdnHostnames-1699296007 tempest-ServersTestFqdnHostnames-1699296007-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 940.441913] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-231bdff4-2d2e-4577-9552-cfa1b1d4e74f tempest-ServersTestFqdnHostnames-1699296007 tempest-ServersTestFqdnHostnames-1699296007-project-member] [instance: 7aa4c0d9-f7ca-42ba-9fd0-4075cd76752b] Processing image b52362a3-ee8a-4cbf-b06f-513b0cc8f95c {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 940.442241] env[65726]: DEBUG oslo_concurrency.lockutils [None req-231bdff4-2d2e-4577-9552-cfa1b1d4e74f tempest-ServersTestFqdnHostnames-1699296007 tempest-ServersTestFqdnHostnames-1699296007-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 940.442362] env[65726]: DEBUG oslo_concurrency.lockutils [None req-231bdff4-2d2e-4577-9552-cfa1b1d4e74f tempest-ServersTestFqdnHostnames-1699296007 tempest-ServersTestFqdnHostnames-1699296007-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 940.442547] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-231bdff4-2d2e-4577-9552-cfa1b1d4e74f tempest-ServersTestFqdnHostnames-1699296007 tempest-ServersTestFqdnHostnames-1699296007-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 940.446046] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c3d3d2ff-8964-44cc-8be0-b9283988294b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.456906] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-231bdff4-2d2e-4577-9552-cfa1b1d4e74f tempest-ServersTestFqdnHostnames-1699296007 tempest-ServersTestFqdnHostnames-1699296007-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 940.457912] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-231bdff4-2d2e-4577-9552-cfa1b1d4e74f tempest-ServersTestFqdnHostnames-1699296007 tempest-ServersTestFqdnHostnames-1699296007-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65726) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 940.462598] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5e3db9ff-6ddd-414f-ba91-f750cf00c3a3 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.471276] env[65726]: DEBUG oslo_vmware.api [None req-231bdff4-2d2e-4577-9552-cfa1b1d4e74f tempest-ServersTestFqdnHostnames-1699296007 tempest-ServersTestFqdnHostnames-1699296007-project-member] Waiting for the task: (returnval){ [ 940.471276] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52b19774-cb84-0c30-9d9c-e31e653ed538" [ 940.471276] env[65726]: _type = "Task" [ 940.471276] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.485015] env[65726]: DEBUG oslo_vmware.api [None req-231bdff4-2d2e-4577-9552-cfa1b1d4e74f tempest-ServersTestFqdnHostnames-1699296007 tempest-ServersTestFqdnHostnames-1699296007-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52b19774-cb84-0c30-9d9c-e31e653ed538, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.615579] env[65726]: DEBUG nova.compute.manager [None req-98994c54-47ac-4003-a1d5-85175217e75d tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 608cc0ea-3f6b-4b4f-83c1-01aa50999d63] Start destroying the instance on the hypervisor. {{(pid=65726) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 940.615579] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-98994c54-47ac-4003-a1d5-85175217e75d tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 608cc0ea-3f6b-4b4f-83c1-01aa50999d63] Destroying instance {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 940.615579] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1658bf86-7f79-4915-9568-fbe078ca5855 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.626060] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-98994c54-47ac-4003-a1d5-85175217e75d tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 608cc0ea-3f6b-4b4f-83c1-01aa50999d63] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 940.629560] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-aae0c3e9-d21f-483e-a531-edb59dfc2947 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.640277] env[65726]: DEBUG oslo_vmware.api [None req-98994c54-47ac-4003-a1d5-85175217e75d tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Waiting for the task: (returnval){ [ 940.640277] env[65726]: value = "task-5116289" [ 940.640277] env[65726]: _type = "Task" [ 940.640277] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.652412] env[65726]: DEBUG oslo_vmware.api [None req-98994c54-47ac-4003-a1d5-85175217e75d tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5116289, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.703272] env[65726]: INFO nova.compute.manager [-] [instance: 0e064341-4e4a-407b-8c26-3eb04b409029] Took 1.35 seconds to deallocate network for instance. [ 940.754024] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a278716-b14f-492e-bb1d-6d58de150761 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.766513] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af0dfb06-3421-4416-b6bf-8df06207a2c9 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.815083] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35f17516-c595-4c0b-bcb7-f22bb4151d4d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.825825] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-717c1545-bdbb-4e3b-b3d5-3cb3c4e84272 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.844855] env[65726]: DEBUG nova.compute.provider_tree [None req-2eeb1a84-0a41-40be-a094-58835ef7befd tempest-ServerAddressesTestJSON-569100133 tempest-ServerAddressesTestJSON-569100133-project-member] Updating inventory in ProviderTree for provider 07c4692f-bdb4-4058-9173-ff9664830295 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 940.983505] env[65726]: DEBUG oslo_vmware.api [None req-231bdff4-2d2e-4577-9552-cfa1b1d4e74f tempest-ServersTestFqdnHostnames-1699296007 tempest-ServersTestFqdnHostnames-1699296007-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52b19774-cb84-0c30-9d9c-e31e653ed538, 'name': SearchDatastore_Task, 'duration_secs': 0.015222} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.984817] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b179430c-7254-4cad-b7ba-529fdc657219 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.992389] env[65726]: DEBUG oslo_vmware.api [None req-231bdff4-2d2e-4577-9552-cfa1b1d4e74f tempest-ServersTestFqdnHostnames-1699296007 tempest-ServersTestFqdnHostnames-1699296007-project-member] Waiting for the task: (returnval){ [ 940.992389] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]525c21bd-3e9c-7b27-4d7d-984734e44dc2" [ 940.992389] env[65726]: _type = "Task" [ 940.992389] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.005472] env[65726]: DEBUG oslo_vmware.api [None req-231bdff4-2d2e-4577-9552-cfa1b1d4e74f tempest-ServersTestFqdnHostnames-1699296007 tempest-ServersTestFqdnHostnames-1699296007-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]525c21bd-3e9c-7b27-4d7d-984734e44dc2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.082478] env[65726]: DEBUG nova.compute.manager [req-b379db3b-39d8-428f-a86b-c8b748448781 req-983f2a88-46e9-4389-9f3d-341df787d06e service nova] [instance: 0e064341-4e4a-407b-8c26-3eb04b409029] Received event network-vif-deleted-9d3c5bc9-aa25-4013-b8dc-f565887d2736 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 941.153476] env[65726]: DEBUG oslo_vmware.api [None req-98994c54-47ac-4003-a1d5-85175217e75d tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5116289, 'name': PowerOffVM_Task, 'duration_secs': 0.35763} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.153867] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-98994c54-47ac-4003-a1d5-85175217e75d tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 608cc0ea-3f6b-4b4f-83c1-01aa50999d63] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 941.154016] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-98994c54-47ac-4003-a1d5-85175217e75d tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 608cc0ea-3f6b-4b4f-83c1-01aa50999d63] Unregistering the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 941.154294] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f22ad5df-77dd-4065-b504-a6e21599675f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.182272] env[65726]: DEBUG oslo_concurrency.lockutils [None req-2f755edc-fef9-4d79-8fd3-503b461c6f4b tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Acquiring lock "cf950ec3-9914-4eb9-99db-048b3969bb21" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 941.182470] env[65726]: DEBUG oslo_concurrency.lockutils [None req-2f755edc-fef9-4d79-8fd3-503b461c6f4b tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Lock "cf950ec3-9914-4eb9-99db-048b3969bb21" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 941.214733] env[65726]: DEBUG oslo_concurrency.lockutils [None req-f5777e4a-de17-4332-8f51-ad1e8cbe3f19 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 941.238103] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-98994c54-47ac-4003-a1d5-85175217e75d tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 608cc0ea-3f6b-4b4f-83c1-01aa50999d63] Unregistered the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 941.238103] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-98994c54-47ac-4003-a1d5-85175217e75d tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 608cc0ea-3f6b-4b4f-83c1-01aa50999d63] Deleting contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 941.238478] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-98994c54-47ac-4003-a1d5-85175217e75d tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Deleting the datastore file [datastore1] 608cc0ea-3f6b-4b4f-83c1-01aa50999d63 {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 941.238984] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b7574491-ba96-4221-aac5-ee155cf34a1e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.248276] env[65726]: DEBUG oslo_vmware.api [None req-98994c54-47ac-4003-a1d5-85175217e75d tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Waiting for the task: (returnval){ [ 941.248276] env[65726]: value = "task-5116291" [ 941.248276] env[65726]: _type = "Task" [ 941.248276] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.260285] env[65726]: DEBUG oslo_vmware.api [None req-98994c54-47ac-4003-a1d5-85175217e75d tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5116291, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.388023] env[65726]: DEBUG nova.scheduler.client.report [None req-2eeb1a84-0a41-40be-a094-58835ef7befd tempest-ServerAddressesTestJSON-569100133 tempest-ServerAddressesTestJSON-569100133-project-member] Updated inventory for provider 07c4692f-bdb4-4058-9173-ff9664830295 with generation 106 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 941.388399] env[65726]: DEBUG nova.compute.provider_tree [None req-2eeb1a84-0a41-40be-a094-58835ef7befd tempest-ServerAddressesTestJSON-569100133 tempest-ServerAddressesTestJSON-569100133-project-member] Updating resource provider 07c4692f-bdb4-4058-9173-ff9664830295 generation from 106 to 107 during operation: update_inventory {{(pid=65726) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 941.388594] env[65726]: DEBUG nova.compute.provider_tree [None req-2eeb1a84-0a41-40be-a094-58835ef7befd tempest-ServerAddressesTestJSON-569100133 tempest-ServerAddressesTestJSON-569100133-project-member] Updating inventory in ProviderTree for provider 07c4692f-bdb4-4058-9173-ff9664830295 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 941.503875] env[65726]: DEBUG oslo_vmware.api [None req-231bdff4-2d2e-4577-9552-cfa1b1d4e74f tempest-ServersTestFqdnHostnames-1699296007 tempest-ServersTestFqdnHostnames-1699296007-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]525c21bd-3e9c-7b27-4d7d-984734e44dc2, 'name': SearchDatastore_Task, 'duration_secs': 0.014014} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.504196] env[65726]: DEBUG oslo_concurrency.lockutils [None req-231bdff4-2d2e-4577-9552-cfa1b1d4e74f tempest-ServersTestFqdnHostnames-1699296007 tempest-ServersTestFqdnHostnames-1699296007-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 941.504486] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-231bdff4-2d2e-4577-9552-cfa1b1d4e74f tempest-ServersTestFqdnHostnames-1699296007 tempest-ServersTestFqdnHostnames-1699296007-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] 7aa4c0d9-f7ca-42ba-9fd0-4075cd76752b/7aa4c0d9-f7ca-42ba-9fd0-4075cd76752b.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 941.504735] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-10446587-ac24-4dfa-b054-543121da1766 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.513075] env[65726]: DEBUG oslo_vmware.api [None req-231bdff4-2d2e-4577-9552-cfa1b1d4e74f tempest-ServersTestFqdnHostnames-1699296007 tempest-ServersTestFqdnHostnames-1699296007-project-member] Waiting for the task: (returnval){ [ 941.513075] env[65726]: value = "task-5116292" [ 941.513075] env[65726]: _type = "Task" [ 941.513075] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.522841] env[65726]: DEBUG oslo_vmware.api [None req-231bdff4-2d2e-4577-9552-cfa1b1d4e74f tempest-ServersTestFqdnHostnames-1699296007 tempest-ServersTestFqdnHostnames-1699296007-project-member] Task: {'id': task-5116292, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.688888] env[65726]: DEBUG nova.compute.utils [None req-2f755edc-fef9-4d79-8fd3-503b461c6f4b tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Using /dev/sd instead of None {{(pid=65726) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 941.760562] env[65726]: DEBUG oslo_vmware.api [None req-98994c54-47ac-4003-a1d5-85175217e75d tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5116291, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.366906} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.760939] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-98994c54-47ac-4003-a1d5-85175217e75d tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Deleted the datastore file {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 941.761224] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-98994c54-47ac-4003-a1d5-85175217e75d tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 608cc0ea-3f6b-4b4f-83c1-01aa50999d63] Deleted contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 941.761464] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-98994c54-47ac-4003-a1d5-85175217e75d tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 608cc0ea-3f6b-4b4f-83c1-01aa50999d63] Instance destroyed {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 941.761697] env[65726]: INFO nova.compute.manager [None req-98994c54-47ac-4003-a1d5-85175217e75d tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 608cc0ea-3f6b-4b4f-83c1-01aa50999d63] Took 1.15 seconds to destroy the instance on the hypervisor. [ 941.762017] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-98994c54-47ac-4003-a1d5-85175217e75d tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 941.762318] env[65726]: DEBUG nova.compute.manager [-] [instance: 608cc0ea-3f6b-4b4f-83c1-01aa50999d63] Deallocating network for instance {{(pid=65726) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 941.762448] env[65726]: DEBUG nova.network.neutron [-] [instance: 608cc0ea-3f6b-4b4f-83c1-01aa50999d63] deallocate_for_instance() {{(pid=65726) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 941.763094] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 941.763438] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 941.807263] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 941.807815] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 941.896291] env[65726]: DEBUG oslo_concurrency.lockutils [None req-2eeb1a84-0a41-40be-a094-58835ef7befd tempest-ServerAddressesTestJSON-569100133 tempest-ServerAddressesTestJSON-569100133-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 3.354s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 941.899172] env[65726]: DEBUG oslo_concurrency.lockutils [None req-371355d3-22ed-4e9f-8bfc-a310d98eb1b4 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.195s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 941.901407] env[65726]: INFO nova.compute.claims [None req-371355d3-22ed-4e9f-8bfc-a310d98eb1b4 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] [instance: df46141c-b545-4e03-b3a3-fd9f5feda0d2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 941.926676] env[65726]: INFO nova.scheduler.client.report [None req-2eeb1a84-0a41-40be-a094-58835ef7befd tempest-ServerAddressesTestJSON-569100133 tempest-ServerAddressesTestJSON-569100133-project-member] Deleted allocations for instance 70375746-06dd-4710-9136-95e9e5759a0e [ 942.035349] env[65726]: DEBUG oslo_vmware.api [None req-231bdff4-2d2e-4577-9552-cfa1b1d4e74f tempest-ServersTestFqdnHostnames-1699296007 tempest-ServersTestFqdnHostnames-1699296007-project-member] Task: {'id': task-5116292, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.172320] env[65726]: DEBUG oslo_concurrency.lockutils [None req-242f797f-8330-4ee8-bd46-f537d017011f tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Acquiring lock "bc2c12e0-0d06-432f-b42f-be468e3b6ee4" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 942.172692] env[65726]: DEBUG oslo_concurrency.lockutils [None req-242f797f-8330-4ee8-bd46-f537d017011f tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Lock "bc2c12e0-0d06-432f-b42f-be468e3b6ee4" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 942.173148] env[65726]: DEBUG oslo_concurrency.lockutils [None req-242f797f-8330-4ee8-bd46-f537d017011f tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Acquiring lock "bc2c12e0-0d06-432f-b42f-be468e3b6ee4-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 942.173148] env[65726]: DEBUG oslo_concurrency.lockutils [None req-242f797f-8330-4ee8-bd46-f537d017011f tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Lock "bc2c12e0-0d06-432f-b42f-be468e3b6ee4-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 942.173604] env[65726]: DEBUG oslo_concurrency.lockutils [None req-242f797f-8330-4ee8-bd46-f537d017011f tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Lock "bc2c12e0-0d06-432f-b42f-be468e3b6ee4-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 942.175881] env[65726]: INFO nova.compute.manager [None req-242f797f-8330-4ee8-bd46-f537d017011f tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] [instance: bc2c12e0-0d06-432f-b42f-be468e3b6ee4] Terminating instance [ 942.192783] env[65726]: DEBUG oslo_concurrency.lockutils [None req-2f755edc-fef9-4d79-8fd3-503b461c6f4b tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Lock "cf950ec3-9914-4eb9-99db-048b3969bb21" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.010s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 942.436489] env[65726]: DEBUG oslo_concurrency.lockutils [None req-2eeb1a84-0a41-40be-a094-58835ef7befd tempest-ServerAddressesTestJSON-569100133 tempest-ServerAddressesTestJSON-569100133-project-member] Lock "70375746-06dd-4710-9136-95e9e5759a0e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 29.424s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 942.525439] env[65726]: DEBUG oslo_vmware.api [None req-231bdff4-2d2e-4577-9552-cfa1b1d4e74f tempest-ServersTestFqdnHostnames-1699296007 tempest-ServersTestFqdnHostnames-1699296007-project-member] Task: {'id': task-5116292, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.641616} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 942.525776] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-231bdff4-2d2e-4577-9552-cfa1b1d4e74f tempest-ServersTestFqdnHostnames-1699296007 tempest-ServersTestFqdnHostnames-1699296007-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] 7aa4c0d9-f7ca-42ba-9fd0-4075cd76752b/7aa4c0d9-f7ca-42ba-9fd0-4075cd76752b.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 942.525996] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-231bdff4-2d2e-4577-9552-cfa1b1d4e74f tempest-ServersTestFqdnHostnames-1699296007 tempest-ServersTestFqdnHostnames-1699296007-project-member] [instance: 7aa4c0d9-f7ca-42ba-9fd0-4075cd76752b] Extending root virtual disk to 1048576 {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 942.526210] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c370228b-7aeb-4272-a04d-ea20668271ac {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.533855] env[65726]: DEBUG oslo_vmware.api [None req-231bdff4-2d2e-4577-9552-cfa1b1d4e74f tempest-ServersTestFqdnHostnames-1699296007 tempest-ServersTestFqdnHostnames-1699296007-project-member] Waiting for the task: (returnval){ [ 942.533855] env[65726]: value = "task-5116293" [ 942.533855] env[65726]: _type = "Task" [ 942.533855] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.545331] env[65726]: DEBUG oslo_vmware.api [None req-231bdff4-2d2e-4577-9552-cfa1b1d4e74f tempest-ServersTestFqdnHostnames-1699296007 tempest-ServersTestFqdnHostnames-1699296007-project-member] Task: {'id': task-5116293, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.680147] env[65726]: DEBUG nova.compute.manager [None req-242f797f-8330-4ee8-bd46-f537d017011f tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] [instance: bc2c12e0-0d06-432f-b42f-be468e3b6ee4] Start destroying the instance on the hypervisor. {{(pid=65726) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 942.680522] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-242f797f-8330-4ee8-bd46-f537d017011f tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] [instance: bc2c12e0-0d06-432f-b42f-be468e3b6ee4] Destroying instance {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 942.681631] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afd333da-c5e6-46f7-818b-8398d205bcc5 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.691961] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-242f797f-8330-4ee8-bd46-f537d017011f tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] [instance: bc2c12e0-0d06-432f-b42f-be468e3b6ee4] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 942.692420] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8c99fd79-6291-4d26-8e9a-a634d8bcaf39 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.703924] env[65726]: DEBUG oslo_vmware.api [None req-242f797f-8330-4ee8-bd46-f537d017011f tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Waiting for the task: (returnval){ [ 942.703924] env[65726]: value = "task-5116294" [ 942.703924] env[65726]: _type = "Task" [ 942.703924] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.722038] env[65726]: DEBUG oslo_vmware.api [None req-242f797f-8330-4ee8-bd46-f537d017011f tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Task: {'id': task-5116294, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.846631] env[65726]: DEBUG nova.network.neutron [-] [instance: 608cc0ea-3f6b-4b4f-83c1-01aa50999d63] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 943.046481] env[65726]: DEBUG oslo_vmware.api [None req-231bdff4-2d2e-4577-9552-cfa1b1d4e74f tempest-ServersTestFqdnHostnames-1699296007 tempest-ServersTestFqdnHostnames-1699296007-project-member] Task: {'id': task-5116293, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.072958} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.049021] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-231bdff4-2d2e-4577-9552-cfa1b1d4e74f tempest-ServersTestFqdnHostnames-1699296007 tempest-ServersTestFqdnHostnames-1699296007-project-member] [instance: 7aa4c0d9-f7ca-42ba-9fd0-4075cd76752b] Extended root virtual disk {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 943.049021] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-319f9fc8-90cc-4a39-be09-04225fe72a6f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.071279] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-231bdff4-2d2e-4577-9552-cfa1b1d4e74f tempest-ServersTestFqdnHostnames-1699296007 tempest-ServersTestFqdnHostnames-1699296007-project-member] [instance: 7aa4c0d9-f7ca-42ba-9fd0-4075cd76752b] Reconfiguring VM instance instance-0000004a to attach disk [datastore1] 7aa4c0d9-f7ca-42ba-9fd0-4075cd76752b/7aa4c0d9-f7ca-42ba-9fd0-4075cd76752b.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 943.074555] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-59bfbc11-3452-4cc7-a5b6-520021934e8f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.096840] env[65726]: DEBUG oslo_vmware.api [None req-231bdff4-2d2e-4577-9552-cfa1b1d4e74f tempest-ServersTestFqdnHostnames-1699296007 tempest-ServersTestFqdnHostnames-1699296007-project-member] Waiting for the task: (returnval){ [ 943.096840] env[65726]: value = "task-5116295" [ 943.096840] env[65726]: _type = "Task" [ 943.096840] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.108657] env[65726]: DEBUG oslo_vmware.api [None req-231bdff4-2d2e-4577-9552-cfa1b1d4e74f tempest-ServersTestFqdnHostnames-1699296007 tempest-ServersTestFqdnHostnames-1699296007-project-member] Task: {'id': task-5116295, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.219935] env[65726]: DEBUG oslo_vmware.api [None req-242f797f-8330-4ee8-bd46-f537d017011f tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Task: {'id': task-5116294, 'name': PowerOffVM_Task, 'duration_secs': 0.231315} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.220199] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-242f797f-8330-4ee8-bd46-f537d017011f tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] [instance: bc2c12e0-0d06-432f-b42f-be468e3b6ee4] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 943.220377] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-242f797f-8330-4ee8-bd46-f537d017011f tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] [instance: bc2c12e0-0d06-432f-b42f-be468e3b6ee4] Unregistering the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 943.220659] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9a7901a3-49a6-453b-9f43-99e9b94f6fd5 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.280109] env[65726]: DEBUG oslo_concurrency.lockutils [None req-2f755edc-fef9-4d79-8fd3-503b461c6f4b tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Acquiring lock "cf950ec3-9914-4eb9-99db-048b3969bb21" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 943.282030] env[65726]: DEBUG oslo_concurrency.lockutils [None req-2f755edc-fef9-4d79-8fd3-503b461c6f4b tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Lock "cf950ec3-9914-4eb9-99db-048b3969bb21" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 943.282030] env[65726]: INFO nova.compute.manager [None req-2f755edc-fef9-4d79-8fd3-503b461c6f4b tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: cf950ec3-9914-4eb9-99db-048b3969bb21] Attaching volume dbb0735d-409f-439d-a32a-4de84d10f7a9 to /dev/sdb [ 943.304364] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-242f797f-8330-4ee8-bd46-f537d017011f tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] [instance: bc2c12e0-0d06-432f-b42f-be468e3b6ee4] Unregistered the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 943.304591] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-242f797f-8330-4ee8-bd46-f537d017011f tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] [instance: bc2c12e0-0d06-432f-b42f-be468e3b6ee4] Deleting contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 943.306734] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-242f797f-8330-4ee8-bd46-f537d017011f tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Deleting the datastore file [datastore1] bc2c12e0-0d06-432f-b42f-be468e3b6ee4 {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 943.306734] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d85084b1-80d3-4ae0-b3a7-936ca0c7a993 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.315180] env[65726]: DEBUG oslo_vmware.api [None req-242f797f-8330-4ee8-bd46-f537d017011f tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Waiting for the task: (returnval){ [ 943.315180] env[65726]: value = "task-5116297" [ 943.315180] env[65726]: _type = "Task" [ 943.315180] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.326127] env[65726]: DEBUG oslo_vmware.api [None req-242f797f-8330-4ee8-bd46-f537d017011f tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Task: {'id': task-5116297, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.339742] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a996bca-584c-453a-8042-250972255a68 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.351580] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80ed31f0-8102-4677-9ce0-fa29064b82fa {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.355422] env[65726]: INFO nova.compute.manager [-] [instance: 608cc0ea-3f6b-4b4f-83c1-01aa50999d63] Took 1.59 seconds to deallocate network for instance. [ 943.371212] env[65726]: DEBUG nova.virt.block_device [None req-2f755edc-fef9-4d79-8fd3-503b461c6f4b tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: cf950ec3-9914-4eb9-99db-048b3969bb21] Updating existing volume attachment record: 939eccd8-7ce7-4518-b3b4-c7f97be3f1f3 {{(pid=65726) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 943.390232] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93832e74-52a2-4010-8f87-49bdaa0f3730 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.401122] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c97548c5-3862-4ddb-989b-4de90fba6cb8 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.437110] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d6dfadf-bef7-456c-b9fe-871a06ca2994 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.446617] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4d6a8d3-a852-4729-bc61-d0772eb9d9d3 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.463693] env[65726]: DEBUG nova.compute.provider_tree [None req-371355d3-22ed-4e9f-8bfc-a310d98eb1b4 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 943.609408] env[65726]: DEBUG oslo_vmware.api [None req-231bdff4-2d2e-4577-9552-cfa1b1d4e74f tempest-ServersTestFqdnHostnames-1699296007 tempest-ServersTestFqdnHostnames-1699296007-project-member] Task: {'id': task-5116295, 'name': ReconfigVM_Task, 'duration_secs': 0.427157} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.609712] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-231bdff4-2d2e-4577-9552-cfa1b1d4e74f tempest-ServersTestFqdnHostnames-1699296007 tempest-ServersTestFqdnHostnames-1699296007-project-member] [instance: 7aa4c0d9-f7ca-42ba-9fd0-4075cd76752b] Reconfigured VM instance instance-0000004a to attach disk [datastore1] 7aa4c0d9-f7ca-42ba-9fd0-4075cd76752b/7aa4c0d9-f7ca-42ba-9fd0-4075cd76752b.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 943.612071] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a5b4b4a6-7378-4b14-870b-c5ae286e935d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.619238] env[65726]: DEBUG oslo_vmware.api [None req-231bdff4-2d2e-4577-9552-cfa1b1d4e74f tempest-ServersTestFqdnHostnames-1699296007 tempest-ServersTestFqdnHostnames-1699296007-project-member] Waiting for the task: (returnval){ [ 943.619238] env[65726]: value = "task-5116299" [ 943.619238] env[65726]: _type = "Task" [ 943.619238] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.632803] env[65726]: DEBUG oslo_vmware.api [None req-231bdff4-2d2e-4577-9552-cfa1b1d4e74f tempest-ServersTestFqdnHostnames-1699296007 tempest-ServersTestFqdnHostnames-1699296007-project-member] Task: {'id': task-5116299, 'name': Rename_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.779269] env[65726]: DEBUG nova.compute.manager [req-64743bda-c65c-4e21-9181-90748e1178e1 req-bc9eed45-ebf9-44e1-be86-625fe7cda9f7 service nova] [instance: 608cc0ea-3f6b-4b4f-83c1-01aa50999d63] Received event network-vif-deleted-e9f7132a-8d5a-47fe-afc6-d45bf3687ea4 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 943.829394] env[65726]: DEBUG oslo_vmware.api [None req-242f797f-8330-4ee8-bd46-f537d017011f tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Task: {'id': task-5116297, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.146757} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.829952] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-242f797f-8330-4ee8-bd46-f537d017011f tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Deleted the datastore file {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 943.830194] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-242f797f-8330-4ee8-bd46-f537d017011f tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] [instance: bc2c12e0-0d06-432f-b42f-be468e3b6ee4] Deleted contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 943.830389] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-242f797f-8330-4ee8-bd46-f537d017011f tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] [instance: bc2c12e0-0d06-432f-b42f-be468e3b6ee4] Instance destroyed {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 943.830567] env[65726]: INFO nova.compute.manager [None req-242f797f-8330-4ee8-bd46-f537d017011f tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] [instance: bc2c12e0-0d06-432f-b42f-be468e3b6ee4] Took 1.15 seconds to destroy the instance on the hypervisor. [ 943.830883] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-242f797f-8330-4ee8-bd46-f537d017011f tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 943.831208] env[65726]: DEBUG nova.compute.manager [-] [instance: bc2c12e0-0d06-432f-b42f-be468e3b6ee4] Deallocating network for instance {{(pid=65726) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 943.831340] env[65726]: DEBUG nova.network.neutron [-] [instance: bc2c12e0-0d06-432f-b42f-be468e3b6ee4] deallocate_for_instance() {{(pid=65726) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 943.831872] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 943.833169] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 943.873082] env[65726]: DEBUG oslo_concurrency.lockutils [None req-98994c54-47ac-4003-a1d5-85175217e75d tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 943.974032] env[65726]: DEBUG nova.scheduler.client.report [None req-371355d3-22ed-4e9f-8bfc-a310d98eb1b4 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 944.136027] env[65726]: DEBUG oslo_vmware.api [None req-231bdff4-2d2e-4577-9552-cfa1b1d4e74f tempest-ServersTestFqdnHostnames-1699296007 tempest-ServersTestFqdnHostnames-1699296007-project-member] Task: {'id': task-5116299, 'name': Rename_Task, 'duration_secs': 0.308584} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.136027] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-231bdff4-2d2e-4577-9552-cfa1b1d4e74f tempest-ServersTestFqdnHostnames-1699296007 tempest-ServersTestFqdnHostnames-1699296007-project-member] [instance: 7aa4c0d9-f7ca-42ba-9fd0-4075cd76752b] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 944.136027] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-316a0938-2469-49a1-a1e0-77a15ed0cf12 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.143430] env[65726]: DEBUG oslo_vmware.api [None req-231bdff4-2d2e-4577-9552-cfa1b1d4e74f tempest-ServersTestFqdnHostnames-1699296007 tempest-ServersTestFqdnHostnames-1699296007-project-member] Waiting for the task: (returnval){ [ 944.143430] env[65726]: value = "task-5116302" [ 944.143430] env[65726]: _type = "Task" [ 944.143430] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.153446] env[65726]: DEBUG oslo_vmware.api [None req-231bdff4-2d2e-4577-9552-cfa1b1d4e74f tempest-ServersTestFqdnHostnames-1699296007 tempest-ServersTestFqdnHostnames-1699296007-project-member] Task: {'id': task-5116302, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.256952] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 944.258406] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 944.479106] env[65726]: DEBUG oslo_concurrency.lockutils [None req-371355d3-22ed-4e9f-8bfc-a310d98eb1b4 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.580s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 944.479852] env[65726]: DEBUG nova.compute.manager [None req-371355d3-22ed-4e9f-8bfc-a310d98eb1b4 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] [instance: df46141c-b545-4e03-b3a3-fd9f5feda0d2] Start building networks asynchronously for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 944.483357] env[65726]: DEBUG oslo_concurrency.lockutils [None req-57ebeb7d-b6c8-4c85-80c2-d441f18e658e tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.940s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 944.485015] env[65726]: INFO nova.compute.claims [None req-57ebeb7d-b6c8-4c85-80c2-d441f18e658e tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] [instance: 6fd70b4a-63e4-4258-9cff-f3b582500b9f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 944.656240] env[65726]: DEBUG oslo_vmware.api [None req-231bdff4-2d2e-4577-9552-cfa1b1d4e74f tempest-ServersTestFqdnHostnames-1699296007 tempest-ServersTestFqdnHostnames-1699296007-project-member] Task: {'id': task-5116302, 'name': PowerOnVM_Task} progress is 88%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.960638] env[65726]: DEBUG nova.compute.manager [req-7ef8a2ad-6d74-4917-acfe-1f454eb1f929 req-9078d245-1fa3-481c-8caf-4d05df191320 service nova] [instance: bc2c12e0-0d06-432f-b42f-be468e3b6ee4] Received event network-vif-deleted-a1c6ed0b-f1f6-4da3-aeaf-9f815ef90b64 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 944.962740] env[65726]: INFO nova.compute.manager [req-7ef8a2ad-6d74-4917-acfe-1f454eb1f929 req-9078d245-1fa3-481c-8caf-4d05df191320 service nova] [instance: bc2c12e0-0d06-432f-b42f-be468e3b6ee4] Neutron deleted interface a1c6ed0b-f1f6-4da3-aeaf-9f815ef90b64; detaching it from the instance and deleting it from the info cache [ 944.963031] env[65726]: DEBUG nova.network.neutron [req-7ef8a2ad-6d74-4917-acfe-1f454eb1f929 req-9078d245-1fa3-481c-8caf-4d05df191320 service nova] [instance: bc2c12e0-0d06-432f-b42f-be468e3b6ee4] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 944.992076] env[65726]: DEBUG nova.compute.utils [None req-371355d3-22ed-4e9f-8bfc-a310d98eb1b4 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Using /dev/sd instead of None {{(pid=65726) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 944.996322] env[65726]: DEBUG nova.compute.manager [None req-371355d3-22ed-4e9f-8bfc-a310d98eb1b4 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] [instance: df46141c-b545-4e03-b3a3-fd9f5feda0d2] Allocating IP information in the background. {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 944.996322] env[65726]: DEBUG nova.network.neutron [None req-371355d3-22ed-4e9f-8bfc-a310d98eb1b4 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] [instance: df46141c-b545-4e03-b3a3-fd9f5feda0d2] allocate_for_instance() {{(pid=65726) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 944.996322] env[65726]: WARNING neutronclient.v2_0.client [None req-371355d3-22ed-4e9f-8bfc-a310d98eb1b4 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 944.996675] env[65726]: WARNING neutronclient.v2_0.client [None req-371355d3-22ed-4e9f-8bfc-a310d98eb1b4 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 944.997063] env[65726]: WARNING openstack [None req-371355d3-22ed-4e9f-8bfc-a310d98eb1b4 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 944.999266] env[65726]: WARNING openstack [None req-371355d3-22ed-4e9f-8bfc-a310d98eb1b4 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 945.073936] env[65726]: DEBUG nova.policy [None req-371355d3-22ed-4e9f-8bfc-a310d98eb1b4 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2614400ef47b4d20ae684c79641ec2a6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1d5cb47905144da49aaf07ddb7d263af', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65726) authorize /opt/stack/nova/nova/policy.py:192}} [ 945.085735] env[65726]: DEBUG nova.network.neutron [-] [instance: bc2c12e0-0d06-432f-b42f-be468e3b6ee4] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 945.155928] env[65726]: DEBUG oslo_vmware.api [None req-231bdff4-2d2e-4577-9552-cfa1b1d4e74f tempest-ServersTestFqdnHostnames-1699296007 tempest-ServersTestFqdnHostnames-1699296007-project-member] Task: {'id': task-5116302, 'name': PowerOnVM_Task, 'duration_secs': 0.80556} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.155928] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-231bdff4-2d2e-4577-9552-cfa1b1d4e74f tempest-ServersTestFqdnHostnames-1699296007 tempest-ServersTestFqdnHostnames-1699296007-project-member] [instance: 7aa4c0d9-f7ca-42ba-9fd0-4075cd76752b] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 945.159381] env[65726]: INFO nova.compute.manager [None req-231bdff4-2d2e-4577-9552-cfa1b1d4e74f tempest-ServersTestFqdnHostnames-1699296007 tempest-ServersTestFqdnHostnames-1699296007-project-member] [instance: 7aa4c0d9-f7ca-42ba-9fd0-4075cd76752b] Took 8.90 seconds to spawn the instance on the hypervisor. [ 945.159381] env[65726]: DEBUG nova.compute.manager [None req-231bdff4-2d2e-4577-9552-cfa1b1d4e74f tempest-ServersTestFqdnHostnames-1699296007 tempest-ServersTestFqdnHostnames-1699296007-project-member] [instance: 7aa4c0d9-f7ca-42ba-9fd0-4075cd76752b] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 945.159381] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7194032-fc53-43a0-a210-7e081d2a5a5e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.398022] env[65726]: DEBUG nova.network.neutron [None req-371355d3-22ed-4e9f-8bfc-a310d98eb1b4 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] [instance: df46141c-b545-4e03-b3a3-fd9f5feda0d2] Successfully created port: cddd4b3f-5986-4368-8626-1bcb77f31c8e {{(pid=65726) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 945.466583] env[65726]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-21c42018-5190-4907-a9cb-7fd5dee83ad5 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.485656] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ede21a55-3bd7-4ecc-bf2f-63dc9d2fc250 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.516075] env[65726]: DEBUG nova.compute.manager [None req-371355d3-22ed-4e9f-8bfc-a310d98eb1b4 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] [instance: df46141c-b545-4e03-b3a3-fd9f5feda0d2] Start building block device mappings for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 945.535289] env[65726]: DEBUG nova.compute.manager [req-7ef8a2ad-6d74-4917-acfe-1f454eb1f929 req-9078d245-1fa3-481c-8caf-4d05df191320 service nova] [instance: bc2c12e0-0d06-432f-b42f-be468e3b6ee4] Detach interface failed, port_id=a1c6ed0b-f1f6-4da3-aeaf-9f815ef90b64, reason: Instance bc2c12e0-0d06-432f-b42f-be468e3b6ee4 could not be found. {{(pid=65726) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11669}} [ 945.593661] env[65726]: INFO nova.compute.manager [-] [instance: bc2c12e0-0d06-432f-b42f-be468e3b6ee4] Took 1.76 seconds to deallocate network for instance. [ 945.686208] env[65726]: INFO nova.compute.manager [None req-231bdff4-2d2e-4577-9552-cfa1b1d4e74f tempest-ServersTestFqdnHostnames-1699296007 tempest-ServersTestFqdnHostnames-1699296007-project-member] [instance: 7aa4c0d9-f7ca-42ba-9fd0-4075cd76752b] Took 37.30 seconds to build instance. [ 945.980108] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6b20b38-ca9a-4c1c-a732-537693a20b00 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.990595] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-982c1fd8-14cf-4bc9-93a3-a1d26a0ec2be {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.029289] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7199023-4982-451f-a6b7-3fbc612ea270 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.045158] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-817278fa-efb5-42eb-bf36-33922e26ffd7 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.066227] env[65726]: DEBUG nova.compute.provider_tree [None req-57ebeb7d-b6c8-4c85-80c2-d441f18e658e tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 946.070103] env[65726]: DEBUG oslo_vmware.rw_handles [None req-2a7383f5-244c-480c-bf85-0a55eabf2c00 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52c522f2-60fc-0fa2-d6d8-5fc55a919213/disk-0.vmdk. {{(pid=65726) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 946.071363] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e957a968-a944-4a77-a3f2-806ebb6702cd {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.079635] env[65726]: DEBUG oslo_vmware.rw_handles [None req-2a7383f5-244c-480c-bf85-0a55eabf2c00 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52c522f2-60fc-0fa2-d6d8-5fc55a919213/disk-0.vmdk is in state: ready. {{(pid=65726) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 946.079809] env[65726]: ERROR oslo_vmware.rw_handles [None req-2a7383f5-244c-480c-bf85-0a55eabf2c00 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52c522f2-60fc-0fa2-d6d8-5fc55a919213/disk-0.vmdk due to incomplete transfer. [ 946.080712] env[65726]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-978904e0-cedd-4d48-9448-ab4f8ecb3e06 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.089466] env[65726]: DEBUG oslo_vmware.rw_handles [None req-2a7383f5-244c-480c-bf85-0a55eabf2c00 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52c522f2-60fc-0fa2-d6d8-5fc55a919213/disk-0.vmdk. {{(pid=65726) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 946.089978] env[65726]: DEBUG nova.virt.vmwareapi.images [None req-2a7383f5-244c-480c-bf85-0a55eabf2c00 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] [instance: 29823bc7-3909-4ab7-8119-91ee59b289c5] Uploaded image 49be9d2d-40dd-437d-91c3-541d9773bbee to the Glance image server {{(pid=65726) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 946.092434] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-2a7383f5-244c-480c-bf85-0a55eabf2c00 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] [instance: 29823bc7-3909-4ab7-8119-91ee59b289c5] Destroying the VM {{(pid=65726) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 946.093628] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-c6aafad9-1677-4fa0-8907-0815176f1401 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.101059] env[65726]: DEBUG oslo_vmware.api [None req-2a7383f5-244c-480c-bf85-0a55eabf2c00 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Waiting for the task: (returnval){ [ 946.101059] env[65726]: value = "task-5116304" [ 946.101059] env[65726]: _type = "Task" [ 946.101059] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.112881] env[65726]: DEBUG oslo_concurrency.lockutils [None req-242f797f-8330-4ee8-bd46-f537d017011f tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 946.113144] env[65726]: DEBUG oslo_vmware.api [None req-2a7383f5-244c-480c-bf85-0a55eabf2c00 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Task: {'id': task-5116304, 'name': Destroy_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.188027] env[65726]: DEBUG oslo_concurrency.lockutils [None req-231bdff4-2d2e-4577-9552-cfa1b1d4e74f tempest-ServersTestFqdnHostnames-1699296007 tempest-ServersTestFqdnHostnames-1699296007-project-member] Lock "7aa4c0d9-f7ca-42ba-9fd0-4075cd76752b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 38.805s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 946.543791] env[65726]: DEBUG nova.compute.manager [None req-371355d3-22ed-4e9f-8bfc-a310d98eb1b4 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] [instance: df46141c-b545-4e03-b3a3-fd9f5feda0d2] Start spawning the instance on the hypervisor. {{(pid=65726) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 946.577529] env[65726]: DEBUG nova.virt.hardware [None req-371355d3-22ed-4e9f-8bfc-a310d98eb1b4 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T19:28:34Z,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1539f06fa3534e90acbc3a60c4b8bd3f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T19:28:35Z,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 946.577529] env[65726]: DEBUG nova.virt.hardware [None req-371355d3-22ed-4e9f-8bfc-a310d98eb1b4 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 946.577708] env[65726]: DEBUG nova.virt.hardware [None req-371355d3-22ed-4e9f-8bfc-a310d98eb1b4 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 946.577750] env[65726]: DEBUG nova.virt.hardware [None req-371355d3-22ed-4e9f-8bfc-a310d98eb1b4 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 946.577887] env[65726]: DEBUG nova.virt.hardware [None req-371355d3-22ed-4e9f-8bfc-a310d98eb1b4 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 946.578050] env[65726]: DEBUG nova.virt.hardware [None req-371355d3-22ed-4e9f-8bfc-a310d98eb1b4 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 946.578243] env[65726]: DEBUG nova.virt.hardware [None req-371355d3-22ed-4e9f-8bfc-a310d98eb1b4 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 946.578399] env[65726]: DEBUG nova.virt.hardware [None req-371355d3-22ed-4e9f-8bfc-a310d98eb1b4 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 946.578665] env[65726]: DEBUG nova.virt.hardware [None req-371355d3-22ed-4e9f-8bfc-a310d98eb1b4 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 946.578834] env[65726]: DEBUG nova.virt.hardware [None req-371355d3-22ed-4e9f-8bfc-a310d98eb1b4 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 946.579013] env[65726]: DEBUG nova.virt.hardware [None req-371355d3-22ed-4e9f-8bfc-a310d98eb1b4 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 946.579907] env[65726]: DEBUG nova.scheduler.client.report [None req-57ebeb7d-b6c8-4c85-80c2-d441f18e658e tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 946.584027] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d219af6-f1d0-4759-9860-ca7b6b9bf5b4 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.594735] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fffe487e-00cf-42bd-8ad5-2b93bd82a7f4 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.621809] env[65726]: DEBUG oslo_vmware.api [None req-2a7383f5-244c-480c-bf85-0a55eabf2c00 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Task: {'id': task-5116304, 'name': Destroy_Task} progress is 33%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.972808] env[65726]: DEBUG nova.network.neutron [None req-371355d3-22ed-4e9f-8bfc-a310d98eb1b4 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] [instance: df46141c-b545-4e03-b3a3-fd9f5feda0d2] Successfully updated port: cddd4b3f-5986-4368-8626-1bcb77f31c8e {{(pid=65726) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 947.089311] env[65726]: DEBUG oslo_concurrency.lockutils [None req-57ebeb7d-b6c8-4c85-80c2-d441f18e658e tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.606s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 947.089653] env[65726]: DEBUG nova.compute.manager [None req-57ebeb7d-b6c8-4c85-80c2-d441f18e658e tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] [instance: 6fd70b4a-63e4-4258-9cff-f3b582500b9f] Start building networks asynchronously for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 947.092790] env[65726]: DEBUG oslo_concurrency.lockutils [None req-1fa711c5-810b-46ee-a4d3-beb8e1bea121 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 16.686s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 947.093072] env[65726]: DEBUG nova.objects.instance [None req-1fa711c5-810b-46ee-a4d3-beb8e1bea121 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Lazy-loading 'resources' on Instance uuid c4cc4f85-cb35-4edc-a58b-adfee0ce1265 {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 947.119929] env[65726]: DEBUG oslo_vmware.api [None req-2a7383f5-244c-480c-bf85-0a55eabf2c00 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Task: {'id': task-5116304, 'name': Destroy_Task, 'duration_secs': 0.811859} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.120397] env[65726]: INFO nova.virt.vmwareapi.vm_util [None req-2a7383f5-244c-480c-bf85-0a55eabf2c00 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] [instance: 29823bc7-3909-4ab7-8119-91ee59b289c5] Destroyed the VM [ 947.120841] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-2a7383f5-244c-480c-bf85-0a55eabf2c00 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] [instance: 29823bc7-3909-4ab7-8119-91ee59b289c5] Deleting Snapshot of the VM instance {{(pid=65726) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 947.121254] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-860f21b3-5b6a-4dad-a5f9-3b8964380837 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.132024] env[65726]: DEBUG oslo_vmware.api [None req-2a7383f5-244c-480c-bf85-0a55eabf2c00 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Waiting for the task: (returnval){ [ 947.132024] env[65726]: value = "task-5116305" [ 947.132024] env[65726]: _type = "Task" [ 947.132024] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.145543] env[65726]: DEBUG oslo_vmware.api [None req-2a7383f5-244c-480c-bf85-0a55eabf2c00 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Task: {'id': task-5116305, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.477251] env[65726]: DEBUG oslo_concurrency.lockutils [None req-371355d3-22ed-4e9f-8bfc-a310d98eb1b4 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Acquiring lock "refresh_cache-df46141c-b545-4e03-b3a3-fd9f5feda0d2" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 947.477439] env[65726]: DEBUG oslo_concurrency.lockutils [None req-371355d3-22ed-4e9f-8bfc-a310d98eb1b4 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Acquired lock "refresh_cache-df46141c-b545-4e03-b3a3-fd9f5feda0d2" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 947.477615] env[65726]: DEBUG nova.network.neutron [None req-371355d3-22ed-4e9f-8bfc-a310d98eb1b4 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] [instance: df46141c-b545-4e03-b3a3-fd9f5feda0d2] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 947.596712] env[65726]: DEBUG nova.compute.utils [None req-57ebeb7d-b6c8-4c85-80c2-d441f18e658e tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Using /dev/sd instead of None {{(pid=65726) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 947.603711] env[65726]: DEBUG nova.compute.manager [None req-57ebeb7d-b6c8-4c85-80c2-d441f18e658e tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] [instance: 6fd70b4a-63e4-4258-9cff-f3b582500b9f] Allocating IP information in the background. {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 947.603944] env[65726]: DEBUG nova.network.neutron [None req-57ebeb7d-b6c8-4c85-80c2-d441f18e658e tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] [instance: 6fd70b4a-63e4-4258-9cff-f3b582500b9f] allocate_for_instance() {{(pid=65726) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 947.604444] env[65726]: WARNING neutronclient.v2_0.client [None req-57ebeb7d-b6c8-4c85-80c2-d441f18e658e tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 947.604598] env[65726]: WARNING neutronclient.v2_0.client [None req-57ebeb7d-b6c8-4c85-80c2-d441f18e658e tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 947.605252] env[65726]: WARNING openstack [None req-57ebeb7d-b6c8-4c85-80c2-d441f18e658e tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 947.607062] env[65726]: WARNING openstack [None req-57ebeb7d-b6c8-4c85-80c2-d441f18e658e tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 947.646940] env[65726]: DEBUG oslo_vmware.api [None req-2a7383f5-244c-480c-bf85-0a55eabf2c00 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Task: {'id': task-5116305, 'name': RemoveSnapshot_Task} progress is 74%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.759037] env[65726]: DEBUG nova.policy [None req-57ebeb7d-b6c8-4c85-80c2-d441f18e658e tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8ebf00cfba6e4d4c8a72c463707b7acd', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '06e221dc693640929cac7bc5af45948f', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65726) authorize /opt/stack/nova/nova/policy.py:192}} [ 947.986370] env[65726]: WARNING openstack [None req-371355d3-22ed-4e9f-8bfc-a310d98eb1b4 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 947.988952] env[65726]: WARNING openstack [None req-371355d3-22ed-4e9f-8bfc-a310d98eb1b4 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 947.995013] env[65726]: DEBUG nova.network.neutron [None req-371355d3-22ed-4e9f-8bfc-a310d98eb1b4 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] [instance: df46141c-b545-4e03-b3a3-fd9f5feda0d2] Instance cache missing network info. {{(pid=65726) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 948.019275] env[65726]: DEBUG nova.network.neutron [None req-57ebeb7d-b6c8-4c85-80c2-d441f18e658e tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] [instance: 6fd70b4a-63e4-4258-9cff-f3b582500b9f] Successfully created port: 153b4c9d-d01d-4254-9aa6-040705be347a {{(pid=65726) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 948.091847] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a7b6d62-5423-470c-bbfc-9a852ba555ae {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.104197] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-326a94f4-be2e-4a5b-b285-a899a848e0cb {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.107838] env[65726]: DEBUG nova.compute.manager [None req-57ebeb7d-b6c8-4c85-80c2-d441f18e658e tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] [instance: 6fd70b4a-63e4-4258-9cff-f3b582500b9f] Start building block device mappings for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 948.141607] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-954d3033-125f-4456-b5db-df195f4c08a9 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.150455] env[65726]: DEBUG oslo_vmware.api [None req-2a7383f5-244c-480c-bf85-0a55eabf2c00 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Task: {'id': task-5116305, 'name': RemoveSnapshot_Task, 'duration_secs': 0.774554} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.152687] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-2a7383f5-244c-480c-bf85-0a55eabf2c00 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] [instance: 29823bc7-3909-4ab7-8119-91ee59b289c5] Deleted Snapshot of the VM instance {{(pid=65726) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 948.152920] env[65726]: INFO nova.compute.manager [None req-2a7383f5-244c-480c-bf85-0a55eabf2c00 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] [instance: 29823bc7-3909-4ab7-8119-91ee59b289c5] Took 16.30 seconds to snapshot the instance on the hypervisor. [ 948.159030] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c75bab8-814d-4126-a33a-8b5d859e448d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.172345] env[65726]: DEBUG nova.compute.provider_tree [None req-1fa711c5-810b-46ee-a4d3-beb8e1bea121 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 948.278876] env[65726]: WARNING openstack [None req-371355d3-22ed-4e9f-8bfc-a310d98eb1b4 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 948.280813] env[65726]: WARNING openstack [None req-371355d3-22ed-4e9f-8bfc-a310d98eb1b4 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 948.367615] env[65726]: DEBUG nova.compute.manager [req-dbc9aa1e-9eea-4107-bf31-6e261e728fd5 req-60b74adc-bd0e-41a9-b2a0-0ec3a1b7f75c service nova] [instance: df46141c-b545-4e03-b3a3-fd9f5feda0d2] Received event network-vif-plugged-cddd4b3f-5986-4368-8626-1bcb77f31c8e {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 948.368398] env[65726]: DEBUG oslo_concurrency.lockutils [req-dbc9aa1e-9eea-4107-bf31-6e261e728fd5 req-60b74adc-bd0e-41a9-b2a0-0ec3a1b7f75c service nova] Acquiring lock "df46141c-b545-4e03-b3a3-fd9f5feda0d2-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 948.368398] env[65726]: DEBUG oslo_concurrency.lockutils [req-dbc9aa1e-9eea-4107-bf31-6e261e728fd5 req-60b74adc-bd0e-41a9-b2a0-0ec3a1b7f75c service nova] Lock "df46141c-b545-4e03-b3a3-fd9f5feda0d2-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 948.368969] env[65726]: DEBUG oslo_concurrency.lockutils [req-dbc9aa1e-9eea-4107-bf31-6e261e728fd5 req-60b74adc-bd0e-41a9-b2a0-0ec3a1b7f75c service nova] Lock "df46141c-b545-4e03-b3a3-fd9f5feda0d2-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 948.369366] env[65726]: DEBUG nova.compute.manager [req-dbc9aa1e-9eea-4107-bf31-6e261e728fd5 req-60b74adc-bd0e-41a9-b2a0-0ec3a1b7f75c service nova] [instance: df46141c-b545-4e03-b3a3-fd9f5feda0d2] No waiting events found dispatching network-vif-plugged-cddd4b3f-5986-4368-8626-1bcb77f31c8e {{(pid=65726) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 948.369993] env[65726]: WARNING nova.compute.manager [req-dbc9aa1e-9eea-4107-bf31-6e261e728fd5 req-60b74adc-bd0e-41a9-b2a0-0ec3a1b7f75c service nova] [instance: df46141c-b545-4e03-b3a3-fd9f5feda0d2] Received unexpected event network-vif-plugged-cddd4b3f-5986-4368-8626-1bcb77f31c8e for instance with vm_state building and task_state spawning. [ 948.417945] env[65726]: WARNING openstack [None req-371355d3-22ed-4e9f-8bfc-a310d98eb1b4 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 948.421454] env[65726]: WARNING openstack [None req-371355d3-22ed-4e9f-8bfc-a310d98eb1b4 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 948.447825] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-2f755edc-fef9-4d79-8fd3-503b461c6f4b tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: cf950ec3-9914-4eb9-99db-048b3969bb21] Volume attach. Driver type: vmdk {{(pid=65726) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 948.448062] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-2f755edc-fef9-4d79-8fd3-503b461c6f4b tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: cf950ec3-9914-4eb9-99db-048b3969bb21] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-995220', 'volume_id': 'dbb0735d-409f-439d-a32a-4de84d10f7a9', 'name': 'volume-dbb0735d-409f-439d-a32a-4de84d10f7a9', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'cf950ec3-9914-4eb9-99db-048b3969bb21', 'attached_at': '', 'detached_at': '', 'volume_id': 'dbb0735d-409f-439d-a32a-4de84d10f7a9', 'serial': 'dbb0735d-409f-439d-a32a-4de84d10f7a9'} {{(pid=65726) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 948.449452] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-533b3219-6af2-44a4-9ce3-22bb0de0c018 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.476156] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0dc42a2f-e179-40f2-8539-a6c52e37495f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.504416] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-2f755edc-fef9-4d79-8fd3-503b461c6f4b tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: cf950ec3-9914-4eb9-99db-048b3969bb21] Reconfiguring VM instance instance-0000003e to attach disk [datastore1] volume-dbb0735d-409f-439d-a32a-4de84d10f7a9/volume-dbb0735d-409f-439d-a32a-4de84d10f7a9.vmdk or device None with type thin {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 948.504746] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-99cdb9cb-e19c-40f7-a71c-d07768e7e356 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.529622] env[65726]: DEBUG oslo_vmware.api [None req-2f755edc-fef9-4d79-8fd3-503b461c6f4b tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Waiting for the task: (returnval){ [ 948.529622] env[65726]: value = "task-5116306" [ 948.529622] env[65726]: _type = "Task" [ 948.529622] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 948.539795] env[65726]: DEBUG oslo_vmware.api [None req-2f755edc-fef9-4d79-8fd3-503b461c6f4b tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Task: {'id': task-5116306, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.675550] env[65726]: DEBUG nova.scheduler.client.report [None req-1fa711c5-810b-46ee-a4d3-beb8e1bea121 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 949.004781] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9f65b40f-615d-4bae-ac4c-c122fff748fd tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Acquiring lock "aa6ce489-c62f-4481-87b7-e74242aeb8ca" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 949.004781] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9f65b40f-615d-4bae-ac4c-c122fff748fd tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Lock "aa6ce489-c62f-4481-87b7-e74242aeb8ca" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 949.045023] env[65726]: DEBUG oslo_vmware.api [None req-2f755edc-fef9-4d79-8fd3-503b461c6f4b tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Task: {'id': task-5116306, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.119168] env[65726]: DEBUG nova.compute.manager [None req-57ebeb7d-b6c8-4c85-80c2-d441f18e658e tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] [instance: 6fd70b4a-63e4-4258-9cff-f3b582500b9f] Start spawning the instance on the hypervisor. {{(pid=65726) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 949.148984] env[65726]: DEBUG nova.virt.hardware [None req-57ebeb7d-b6c8-4c85-80c2-d441f18e658e tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T19:28:34Z,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1539f06fa3534e90acbc3a60c4b8bd3f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T19:28:35Z,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 949.150061] env[65726]: DEBUG nova.virt.hardware [None req-57ebeb7d-b6c8-4c85-80c2-d441f18e658e tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 949.150061] env[65726]: DEBUG nova.virt.hardware [None req-57ebeb7d-b6c8-4c85-80c2-d441f18e658e tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 949.150061] env[65726]: DEBUG nova.virt.hardware [None req-57ebeb7d-b6c8-4c85-80c2-d441f18e658e tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 949.150061] env[65726]: DEBUG nova.virt.hardware [None req-57ebeb7d-b6c8-4c85-80c2-d441f18e658e tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 949.150211] env[65726]: DEBUG nova.virt.hardware [None req-57ebeb7d-b6c8-4c85-80c2-d441f18e658e tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 949.151543] env[65726]: DEBUG nova.virt.hardware [None req-57ebeb7d-b6c8-4c85-80c2-d441f18e658e tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 949.151543] env[65726]: DEBUG nova.virt.hardware [None req-57ebeb7d-b6c8-4c85-80c2-d441f18e658e tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 949.151543] env[65726]: DEBUG nova.virt.hardware [None req-57ebeb7d-b6c8-4c85-80c2-d441f18e658e tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 949.151543] env[65726]: DEBUG nova.virt.hardware [None req-57ebeb7d-b6c8-4c85-80c2-d441f18e658e tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 949.151543] env[65726]: DEBUG nova.virt.hardware [None req-57ebeb7d-b6c8-4c85-80c2-d441f18e658e tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 949.151992] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aca1b55d-bca5-4680-82fd-754557483268 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.162709] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5dbe0845-e680-4989-8fed-38ee5ef17d37 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.185042] env[65726]: DEBUG nova.network.neutron [None req-371355d3-22ed-4e9f-8bfc-a310d98eb1b4 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] [instance: df46141c-b545-4e03-b3a3-fd9f5feda0d2] Updating instance_info_cache with network_info: [{"id": "cddd4b3f-5986-4368-8626-1bcb77f31c8e", "address": "fa:16:3e:55:63:02", "network": {"id": "a2c8f0c5-7fad-4b72-8e1a-376e03d3c0ff", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-2135463722-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1d5cb47905144da49aaf07ddb7d263af", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "98f447de-d71e-41ef-bc37-ed97b4a1f58f", "external-id": "nsx-vlan-transportzone-904", "segmentation_id": 904, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcddd4b3f-59", "ovs_interfaceid": "cddd4b3f-5986-4368-8626-1bcb77f31c8e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 949.186659] env[65726]: DEBUG oslo_concurrency.lockutils [None req-1fa711c5-810b-46ee-a4d3-beb8e1bea121 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.094s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 949.189835] env[65726]: DEBUG oslo_concurrency.lockutils [None req-7cd9e703-73b1-4767-92f9-46012e80c63d tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 18.782s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 949.189835] env[65726]: DEBUG nova.objects.instance [None req-7cd9e703-73b1-4767-92f9-46012e80c63d tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Lazy-loading 'resources' on Instance uuid 6f91b053-772a-4497-b29d-349b960c55eb {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 949.216752] env[65726]: INFO nova.scheduler.client.report [None req-1fa711c5-810b-46ee-a4d3-beb8e1bea121 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Deleted allocations for instance c4cc4f85-cb35-4edc-a58b-adfee0ce1265 [ 949.510263] env[65726]: DEBUG nova.compute.manager [None req-9f65b40f-615d-4bae-ac4c-c122fff748fd tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] [instance: aa6ce489-c62f-4481-87b7-e74242aeb8ca] Starting instance... {{(pid=65726) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 949.542182] env[65726]: DEBUG oslo_vmware.api [None req-2f755edc-fef9-4d79-8fd3-503b461c6f4b tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Task: {'id': task-5116306, 'name': ReconfigVM_Task, 'duration_secs': 0.705741} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 949.542182] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-2f755edc-fef9-4d79-8fd3-503b461c6f4b tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: cf950ec3-9914-4eb9-99db-048b3969bb21] Reconfigured VM instance instance-0000003e to attach disk [datastore1] volume-dbb0735d-409f-439d-a32a-4de84d10f7a9/volume-dbb0735d-409f-439d-a32a-4de84d10f7a9.vmdk or device None with type thin {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 949.547015] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-082dfa9f-b4cc-4191-b8fc-b544ca8fe410 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.564242] env[65726]: DEBUG oslo_vmware.api [None req-2f755edc-fef9-4d79-8fd3-503b461c6f4b tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Waiting for the task: (returnval){ [ 949.564242] env[65726]: value = "task-5116307" [ 949.564242] env[65726]: _type = "Task" [ 949.564242] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.577520] env[65726]: DEBUG oslo_vmware.api [None req-2f755edc-fef9-4d79-8fd3-503b461c6f4b tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Task: {'id': task-5116307, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.690416] env[65726]: DEBUG nova.compute.manager [req-5a443543-3560-4e57-b433-be7ad0ac436e req-dbf4c8eb-2444-4e34-9a61-eca82a8fea7a service nova] [instance: 7aa4c0d9-f7ca-42ba-9fd0-4075cd76752b] Received event network-changed-47831805-14e7-41d0-91a6-54a9da3ff27c {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 949.690789] env[65726]: DEBUG nova.compute.manager [req-5a443543-3560-4e57-b433-be7ad0ac436e req-dbf4c8eb-2444-4e34-9a61-eca82a8fea7a service nova] [instance: 7aa4c0d9-f7ca-42ba-9fd0-4075cd76752b] Refreshing instance network info cache due to event network-changed-47831805-14e7-41d0-91a6-54a9da3ff27c. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 949.690874] env[65726]: DEBUG oslo_concurrency.lockutils [req-5a443543-3560-4e57-b433-be7ad0ac436e req-dbf4c8eb-2444-4e34-9a61-eca82a8fea7a service nova] Acquiring lock "refresh_cache-7aa4c0d9-f7ca-42ba-9fd0-4075cd76752b" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 949.692156] env[65726]: DEBUG oslo_concurrency.lockutils [req-5a443543-3560-4e57-b433-be7ad0ac436e req-dbf4c8eb-2444-4e34-9a61-eca82a8fea7a service nova] Acquired lock "refresh_cache-7aa4c0d9-f7ca-42ba-9fd0-4075cd76752b" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 949.693979] env[65726]: DEBUG nova.network.neutron [req-5a443543-3560-4e57-b433-be7ad0ac436e req-dbf4c8eb-2444-4e34-9a61-eca82a8fea7a service nova] [instance: 7aa4c0d9-f7ca-42ba-9fd0-4075cd76752b] Refreshing network info cache for port 47831805-14e7-41d0-91a6-54a9da3ff27c {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 949.694073] env[65726]: DEBUG oslo_concurrency.lockutils [None req-371355d3-22ed-4e9f-8bfc-a310d98eb1b4 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Releasing lock "refresh_cache-df46141c-b545-4e03-b3a3-fd9f5feda0d2" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 949.694434] env[65726]: DEBUG nova.compute.manager [None req-371355d3-22ed-4e9f-8bfc-a310d98eb1b4 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] [instance: df46141c-b545-4e03-b3a3-fd9f5feda0d2] Instance network_info: |[{"id": "cddd4b3f-5986-4368-8626-1bcb77f31c8e", "address": "fa:16:3e:55:63:02", "network": {"id": "a2c8f0c5-7fad-4b72-8e1a-376e03d3c0ff", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-2135463722-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1d5cb47905144da49aaf07ddb7d263af", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "98f447de-d71e-41ef-bc37-ed97b4a1f58f", "external-id": "nsx-vlan-transportzone-904", "segmentation_id": 904, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcddd4b3f-59", "ovs_interfaceid": "cddd4b3f-5986-4368-8626-1bcb77f31c8e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 949.699072] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-371355d3-22ed-4e9f-8bfc-a310d98eb1b4 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] [instance: df46141c-b545-4e03-b3a3-fd9f5feda0d2] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:55:63:02', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '98f447de-d71e-41ef-bc37-ed97b4a1f58f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'cddd4b3f-5986-4368-8626-1bcb77f31c8e', 'vif_model': 'vmxnet3'}] {{(pid=65726) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 949.708898] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-371355d3-22ed-4e9f-8bfc-a310d98eb1b4 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Creating folder: Project (1d5cb47905144da49aaf07ddb7d263af). Parent ref: group-v995008. {{(pid=65726) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 949.710537] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a68d17f1-5c38-45fa-a01c-905cd1c909b1 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.725636] env[65726]: INFO nova.virt.vmwareapi.vm_util [None req-371355d3-22ed-4e9f-8bfc-a310d98eb1b4 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Created folder: Project (1d5cb47905144da49aaf07ddb7d263af) in parent group-v995008. [ 949.725854] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-371355d3-22ed-4e9f-8bfc-a310d98eb1b4 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Creating folder: Instances. Parent ref: group-v995221. {{(pid=65726) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 949.728962] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-bbee9823-12b8-4ce7-8f86-c8f796ce854b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.732519] env[65726]: DEBUG oslo_concurrency.lockutils [None req-1fa711c5-810b-46ee-a4d3-beb8e1bea121 tempest-ImagesOneServerTestJSON-994467277 tempest-ImagesOneServerTestJSON-994467277-project-member] Lock "c4cc4f85-cb35-4edc-a58b-adfee0ce1265" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 23.014s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 949.747670] env[65726]: INFO nova.virt.vmwareapi.vm_util [None req-371355d3-22ed-4e9f-8bfc-a310d98eb1b4 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Created folder: Instances in parent group-v995221. [ 949.748388] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-371355d3-22ed-4e9f-8bfc-a310d98eb1b4 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 949.751675] env[65726]: DEBUG nova.network.neutron [None req-57ebeb7d-b6c8-4c85-80c2-d441f18e658e tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] [instance: 6fd70b4a-63e4-4258-9cff-f3b582500b9f] Successfully updated port: 153b4c9d-d01d-4254-9aa6-040705be347a {{(pid=65726) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 949.753106] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: df46141c-b545-4e03-b3a3-fd9f5feda0d2] Creating VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 949.753754] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-552e5787-9a9c-426e-9c61-259cc9aa3a92 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.782298] env[65726]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 949.782298] env[65726]: value = "task-5116310" [ 949.782298] env[65726]: _type = "Task" [ 949.782298] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.791954] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116310, 'name': CreateVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.038999] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9f65b40f-615d-4bae-ac4c-c122fff748fd tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 950.078613] env[65726]: DEBUG oslo_vmware.api [None req-2f755edc-fef9-4d79-8fd3-503b461c6f4b tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Task: {'id': task-5116307, 'name': ReconfigVM_Task, 'duration_secs': 0.276647} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.079036] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-2f755edc-fef9-4d79-8fd3-503b461c6f4b tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: cf950ec3-9914-4eb9-99db-048b3969bb21] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-995220', 'volume_id': 'dbb0735d-409f-439d-a32a-4de84d10f7a9', 'name': 'volume-dbb0735d-409f-439d-a32a-4de84d10f7a9', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'cf950ec3-9914-4eb9-99db-048b3969bb21', 'attached_at': '', 'detached_at': '', 'volume_id': 'dbb0735d-409f-439d-a32a-4de84d10f7a9', 'serial': 'dbb0735d-409f-439d-a32a-4de84d10f7a9'} {{(pid=65726) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 950.149801] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0977dd7-c776-4f66-a2df-fad2c1fb5899 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.159091] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-299f7b75-9d36-4f92-a35f-db7aa12e4a38 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.197597] env[65726]: WARNING openstack [req-5a443543-3560-4e57-b433-be7ad0ac436e req-dbf4c8eb-2444-4e34-9a61-eca82a8fea7a service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 950.197597] env[65726]: WARNING openstack [req-5a443543-3560-4e57-b433-be7ad0ac436e req-dbf4c8eb-2444-4e34-9a61-eca82a8fea7a service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 950.204362] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b61a05b4-6055-43ac-9cc0-ebe00b9b69cc {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.214338] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81f19bd2-6cb8-4f95-a271-48897dbe99b3 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.233732] env[65726]: DEBUG nova.compute.provider_tree [None req-7cd9e703-73b1-4767-92f9-46012e80c63d tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 950.272069] env[65726]: DEBUG oslo_concurrency.lockutils [None req-57ebeb7d-b6c8-4c85-80c2-d441f18e658e tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Acquiring lock "refresh_cache-6fd70b4a-63e4-4258-9cff-f3b582500b9f" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 950.273124] env[65726]: DEBUG oslo_concurrency.lockutils [None req-57ebeb7d-b6c8-4c85-80c2-d441f18e658e tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Acquired lock "refresh_cache-6fd70b4a-63e4-4258-9cff-f3b582500b9f" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 950.273124] env[65726]: DEBUG nova.network.neutron [None req-57ebeb7d-b6c8-4c85-80c2-d441f18e658e tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] [instance: 6fd70b4a-63e4-4258-9cff-f3b582500b9f] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 950.297083] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116310, 'name': CreateVM_Task} progress is 99%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.737799] env[65726]: DEBUG nova.scheduler.client.report [None req-7cd9e703-73b1-4767-92f9-46012e80c63d tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 950.775460] env[65726]: WARNING openstack [None req-57ebeb7d-b6c8-4c85-80c2-d441f18e658e tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 950.775847] env[65726]: WARNING openstack [None req-57ebeb7d-b6c8-4c85-80c2-d441f18e658e tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 950.783469] env[65726]: DEBUG nova.network.neutron [None req-57ebeb7d-b6c8-4c85-80c2-d441f18e658e tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] [instance: 6fd70b4a-63e4-4258-9cff-f3b582500b9f] Instance cache missing network info. {{(pid=65726) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 950.795430] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116310, 'name': CreateVM_Task} progress is 99%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.139949] env[65726]: DEBUG nova.objects.instance [None req-2f755edc-fef9-4d79-8fd3-503b461c6f4b tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Lazy-loading 'flavor' on Instance uuid cf950ec3-9914-4eb9-99db-048b3969bb21 {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 951.247738] env[65726]: DEBUG oslo_concurrency.lockutils [None req-7cd9e703-73b1-4767-92f9-46012e80c63d tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.058s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 951.250946] env[65726]: DEBUG oslo_concurrency.lockutils [None req-2eacb245-5155-4571-979a-d9aaa1a355b7 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 19.723s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 951.251282] env[65726]: DEBUG nova.objects.instance [None req-2eacb245-5155-4571-979a-d9aaa1a355b7 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Lazy-loading 'resources' on Instance uuid c108f2a5-031b-47ee-9a5b-d62c2f42c26b {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 951.275773] env[65726]: DEBUG nova.compute.manager [None req-da525c49-2593-4b16-8428-bf88b8e252b1 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] [instance: 815ee20b-cb07-4514-bd43-1b74972508a5] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 951.276914] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffe59af5-db1c-4f5c-b7f2-1f36ed740587 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.293814] env[65726]: WARNING openstack [req-5a443543-3560-4e57-b433-be7ad0ac436e req-dbf4c8eb-2444-4e34-9a61-eca82a8fea7a service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 951.294223] env[65726]: WARNING openstack [req-5a443543-3560-4e57-b433-be7ad0ac436e req-dbf4c8eb-2444-4e34-9a61-eca82a8fea7a service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 951.307257] env[65726]: INFO nova.scheduler.client.report [None req-7cd9e703-73b1-4767-92f9-46012e80c63d tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Deleted allocations for instance 6f91b053-772a-4497-b29d-349b960c55eb [ 951.318338] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116310, 'name': CreateVM_Task} progress is 99%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.377494] env[65726]: WARNING openstack [None req-57ebeb7d-b6c8-4c85-80c2-d441f18e658e tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 951.378073] env[65726]: WARNING openstack [None req-57ebeb7d-b6c8-4c85-80c2-d441f18e658e tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 951.390724] env[65726]: DEBUG nova.compute.manager [req-002127c4-ff40-4b50-8f27-47f4bc24e364 req-c2134f96-4826-4614-9815-a3d04fd2dbe5 service nova] [instance: df46141c-b545-4e03-b3a3-fd9f5feda0d2] Received event network-changed-cddd4b3f-5986-4368-8626-1bcb77f31c8e {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 951.390724] env[65726]: DEBUG nova.compute.manager [req-002127c4-ff40-4b50-8f27-47f4bc24e364 req-c2134f96-4826-4614-9815-a3d04fd2dbe5 service nova] [instance: df46141c-b545-4e03-b3a3-fd9f5feda0d2] Refreshing instance network info cache due to event network-changed-cddd4b3f-5986-4368-8626-1bcb77f31c8e. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 951.390894] env[65726]: DEBUG oslo_concurrency.lockutils [req-002127c4-ff40-4b50-8f27-47f4bc24e364 req-c2134f96-4826-4614-9815-a3d04fd2dbe5 service nova] Acquiring lock "refresh_cache-df46141c-b545-4e03-b3a3-fd9f5feda0d2" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 951.390986] env[65726]: DEBUG oslo_concurrency.lockutils [req-002127c4-ff40-4b50-8f27-47f4bc24e364 req-c2134f96-4826-4614-9815-a3d04fd2dbe5 service nova] Acquired lock "refresh_cache-df46141c-b545-4e03-b3a3-fd9f5feda0d2" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 951.391324] env[65726]: DEBUG nova.network.neutron [req-002127c4-ff40-4b50-8f27-47f4bc24e364 req-c2134f96-4826-4614-9815-a3d04fd2dbe5 service nova] [instance: df46141c-b545-4e03-b3a3-fd9f5feda0d2] Refreshing network info cache for port cddd4b3f-5986-4368-8626-1bcb77f31c8e {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 951.655177] env[65726]: DEBUG oslo_concurrency.lockutils [None req-2f755edc-fef9-4d79-8fd3-503b461c6f4b tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Lock "cf950ec3-9914-4eb9-99db-048b3969bb21" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 8.375s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 951.721848] env[65726]: WARNING openstack [req-5a443543-3560-4e57-b433-be7ad0ac436e req-dbf4c8eb-2444-4e34-9a61-eca82a8fea7a service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 951.721848] env[65726]: WARNING openstack [req-5a443543-3560-4e57-b433-be7ad0ac436e req-dbf4c8eb-2444-4e34-9a61-eca82a8fea7a service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 951.757919] env[65726]: WARNING openstack [None req-57ebeb7d-b6c8-4c85-80c2-d441f18e658e tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 951.758719] env[65726]: WARNING openstack [None req-57ebeb7d-b6c8-4c85-80c2-d441f18e658e tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 951.803651] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116310, 'name': CreateVM_Task, 'duration_secs': 1.605879} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.803768] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: df46141c-b545-4e03-b3a3-fd9f5feda0d2] Created VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 951.804996] env[65726]: WARNING openstack [None req-371355d3-22ed-4e9f-8bfc-a310d98eb1b4 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 951.805410] env[65726]: WARNING openstack [None req-371355d3-22ed-4e9f-8bfc-a310d98eb1b4 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 951.811179] env[65726]: DEBUG oslo_concurrency.lockutils [None req-371355d3-22ed-4e9f-8bfc-a310d98eb1b4 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 951.811412] env[65726]: DEBUG oslo_concurrency.lockutils [None req-371355d3-22ed-4e9f-8bfc-a310d98eb1b4 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 951.811883] env[65726]: DEBUG oslo_concurrency.lockutils [None req-371355d3-22ed-4e9f-8bfc-a310d98eb1b4 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 951.816020] env[65726]: INFO nova.compute.manager [None req-da525c49-2593-4b16-8428-bf88b8e252b1 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] [instance: 815ee20b-cb07-4514-bd43-1b74972508a5] instance snapshotting [ 951.816020] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9c4140e0-3e8b-4721-90cf-a18b0160636b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.817776] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e326165b-aac0-4e35-8f60-317acb0b2a28 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.830065] env[65726]: DEBUG oslo_vmware.api [None req-371355d3-22ed-4e9f-8bfc-a310d98eb1b4 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Waiting for the task: (returnval){ [ 951.830065] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]528e2272-b68e-2536-435f-dd7e2650c9c8" [ 951.830065] env[65726]: _type = "Task" [ 951.830065] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.830792] env[65726]: DEBUG oslo_concurrency.lockutils [None req-7cd9e703-73b1-4767-92f9-46012e80c63d tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Lock "6f91b053-772a-4497-b29d-349b960c55eb" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 25.419s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 951.861270] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ce4a3c9-ddd0-4b6d-846c-70e6b5b71a20 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.880099] env[65726]: DEBUG oslo_vmware.api [None req-371355d3-22ed-4e9f-8bfc-a310d98eb1b4 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]528e2272-b68e-2536-435f-dd7e2650c9c8, 'name': SearchDatastore_Task, 'duration_secs': 0.012529} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.881303] env[65726]: DEBUG oslo_concurrency.lockutils [None req-371355d3-22ed-4e9f-8bfc-a310d98eb1b4 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 951.881607] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-371355d3-22ed-4e9f-8bfc-a310d98eb1b4 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] [instance: df46141c-b545-4e03-b3a3-fd9f5feda0d2] Processing image b52362a3-ee8a-4cbf-b06f-513b0cc8f95c {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 951.881791] env[65726]: DEBUG oslo_concurrency.lockutils [None req-371355d3-22ed-4e9f-8bfc-a310d98eb1b4 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 951.881925] env[65726]: DEBUG oslo_concurrency.lockutils [None req-371355d3-22ed-4e9f-8bfc-a310d98eb1b4 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 951.882232] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-371355d3-22ed-4e9f-8bfc-a310d98eb1b4 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 951.889042] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-03357c7b-0dc6-4515-b280-3d5cd9fa83e2 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.896852] env[65726]: WARNING openstack [req-002127c4-ff40-4b50-8f27-47f4bc24e364 req-c2134f96-4826-4614-9815-a3d04fd2dbe5 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 951.896852] env[65726]: WARNING openstack [req-002127c4-ff40-4b50-8f27-47f4bc24e364 req-c2134f96-4826-4614-9815-a3d04fd2dbe5 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 951.911774] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-371355d3-22ed-4e9f-8bfc-a310d98eb1b4 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 951.911991] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-371355d3-22ed-4e9f-8bfc-a310d98eb1b4 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65726) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 951.913223] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d90b4177-8e0f-4193-8e42-d6cffede8ad9 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.927408] env[65726]: DEBUG oslo_vmware.api [None req-371355d3-22ed-4e9f-8bfc-a310d98eb1b4 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Waiting for the task: (returnval){ [ 951.927408] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52267ec3-5f2a-57a3-fa6f-01f1df514491" [ 951.927408] env[65726]: _type = "Task" [ 951.927408] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.942604] env[65726]: DEBUG oslo_vmware.api [None req-371355d3-22ed-4e9f-8bfc-a310d98eb1b4 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52267ec3-5f2a-57a3-fa6f-01f1df514491, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.996673] env[65726]: DEBUG nova.network.neutron [req-5a443543-3560-4e57-b433-be7ad0ac436e req-dbf4c8eb-2444-4e34-9a61-eca82a8fea7a service nova] [instance: 7aa4c0d9-f7ca-42ba-9fd0-4075cd76752b] Updated VIF entry in instance network info cache for port 47831805-14e7-41d0-91a6-54a9da3ff27c. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 951.997067] env[65726]: DEBUG nova.network.neutron [req-5a443543-3560-4e57-b433-be7ad0ac436e req-dbf4c8eb-2444-4e34-9a61-eca82a8fea7a service nova] [instance: 7aa4c0d9-f7ca-42ba-9fd0-4075cd76752b] Updating instance_info_cache with network_info: [{"id": "47831805-14e7-41d0-91a6-54a9da3ff27c", "address": "fa:16:3e:b0:42:a2", "network": {"id": "e7826d02-16e9-49e4-8794-b5b7870aa663", "bridge": "br-int", "label": "tempest-ServersTestFqdnHostnames-212080893-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.187", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "534461fc8c62483d875b8e165ef0d5c0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f625f389-b7cf-49b9-998a-87f3a9e3f234", "external-id": "nsx-vlan-transportzone-280", "segmentation_id": 280, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap47831805-14", "ovs_interfaceid": "47831805-14e7-41d0-91a6-54a9da3ff27c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 952.083183] env[65726]: DEBUG nova.network.neutron [None req-57ebeb7d-b6c8-4c85-80c2-d441f18e658e tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] [instance: 6fd70b4a-63e4-4258-9cff-f3b582500b9f] Updating instance_info_cache with network_info: [{"id": "153b4c9d-d01d-4254-9aa6-040705be347a", "address": "fa:16:3e:17:18:d9", "network": {"id": "93e38d7e-b4d7-4d4a-8393-64ab2f6769de", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1445539922-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "06e221dc693640929cac7bc5af45948f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "399f3826-705c-45f7-9fe0-3a08a945151a", "external-id": "nsx-vlan-transportzone-936", "segmentation_id": 936, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap153b4c9d-d0", "ovs_interfaceid": "153b4c9d-d01d-4254-9aa6-040705be347a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 952.188077] env[65726]: WARNING openstack [req-002127c4-ff40-4b50-8f27-47f4bc24e364 req-c2134f96-4826-4614-9815-a3d04fd2dbe5 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 952.188077] env[65726]: WARNING openstack [req-002127c4-ff40-4b50-8f27-47f4bc24e364 req-c2134f96-4826-4614-9815-a3d04fd2dbe5 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 952.298452] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4fa7f1db-2b71-4e75-a451-340135178300 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.310427] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b003a431-e285-48f8-86f5-2083308d5f05 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.360450] env[65726]: WARNING openstack [req-002127c4-ff40-4b50-8f27-47f4bc24e364 req-c2134f96-4826-4614-9815-a3d04fd2dbe5 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 952.360943] env[65726]: WARNING openstack [req-002127c4-ff40-4b50-8f27-47f4bc24e364 req-c2134f96-4826-4614-9815-a3d04fd2dbe5 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 952.370700] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f288269e-293b-4975-8c8e-6cda897ef6d5 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.380822] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-087766fe-0e8b-42eb-8c28-d6cc7ff9888d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.386639] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-da525c49-2593-4b16-8428-bf88b8e252b1 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] [instance: 815ee20b-cb07-4514-bd43-1b74972508a5] Creating Snapshot of the VM instance {{(pid=65726) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 952.386639] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-09d50192-9b14-48b6-8871-b26e58ccf3c6 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.402480] env[65726]: DEBUG nova.compute.provider_tree [None req-2eacb245-5155-4571-979a-d9aaa1a355b7 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 952.406423] env[65726]: DEBUG oslo_vmware.api [None req-da525c49-2593-4b16-8428-bf88b8e252b1 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Waiting for the task: (returnval){ [ 952.406423] env[65726]: value = "task-5116311" [ 952.406423] env[65726]: _type = "Task" [ 952.406423] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.415762] env[65726]: DEBUG oslo_vmware.api [None req-da525c49-2593-4b16-8428-bf88b8e252b1 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Task: {'id': task-5116311, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.438970] env[65726]: DEBUG oslo_vmware.api [None req-371355d3-22ed-4e9f-8bfc-a310d98eb1b4 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52267ec3-5f2a-57a3-fa6f-01f1df514491, 'name': SearchDatastore_Task, 'duration_secs': 0.01141} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 952.440069] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7c4ed2b6-de3b-4465-8196-35e7e8864e48 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.446855] env[65726]: DEBUG oslo_vmware.api [None req-371355d3-22ed-4e9f-8bfc-a310d98eb1b4 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Waiting for the task: (returnval){ [ 952.446855] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]520d5f36-4595-ee22-d408-07308451f0af" [ 952.446855] env[65726]: _type = "Task" [ 952.446855] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.465148] env[65726]: DEBUG oslo_vmware.api [None req-371355d3-22ed-4e9f-8bfc-a310d98eb1b4 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]520d5f36-4595-ee22-d408-07308451f0af, 'name': SearchDatastore_Task, 'duration_secs': 0.010924} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 952.465148] env[65726]: DEBUG oslo_concurrency.lockutils [None req-371355d3-22ed-4e9f-8bfc-a310d98eb1b4 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 952.465148] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-371355d3-22ed-4e9f-8bfc-a310d98eb1b4 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] df46141c-b545-4e03-b3a3-fd9f5feda0d2/df46141c-b545-4e03-b3a3-fd9f5feda0d2.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 952.465148] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2536185d-8d08-43ed-9e33-69deb203978b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.471858] env[65726]: DEBUG oslo_vmware.api [None req-371355d3-22ed-4e9f-8bfc-a310d98eb1b4 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Waiting for the task: (returnval){ [ 952.471858] env[65726]: value = "task-5116312" [ 952.471858] env[65726]: _type = "Task" [ 952.471858] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.483846] env[65726]: DEBUG oslo_vmware.api [None req-371355d3-22ed-4e9f-8bfc-a310d98eb1b4 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Task: {'id': task-5116312, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.500556] env[65726]: DEBUG nova.network.neutron [req-002127c4-ff40-4b50-8f27-47f4bc24e364 req-c2134f96-4826-4614-9815-a3d04fd2dbe5 service nova] [instance: df46141c-b545-4e03-b3a3-fd9f5feda0d2] Updated VIF entry in instance network info cache for port cddd4b3f-5986-4368-8626-1bcb77f31c8e. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 952.500971] env[65726]: DEBUG nova.network.neutron [req-002127c4-ff40-4b50-8f27-47f4bc24e364 req-c2134f96-4826-4614-9815-a3d04fd2dbe5 service nova] [instance: df46141c-b545-4e03-b3a3-fd9f5feda0d2] Updating instance_info_cache with network_info: [{"id": "cddd4b3f-5986-4368-8626-1bcb77f31c8e", "address": "fa:16:3e:55:63:02", "network": {"id": "a2c8f0c5-7fad-4b72-8e1a-376e03d3c0ff", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-2135463722-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1d5cb47905144da49aaf07ddb7d263af", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "98f447de-d71e-41ef-bc37-ed97b4a1f58f", "external-id": "nsx-vlan-transportzone-904", "segmentation_id": 904, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcddd4b3f-59", "ovs_interfaceid": "cddd4b3f-5986-4368-8626-1bcb77f31c8e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 952.502703] env[65726]: DEBUG oslo_concurrency.lockutils [req-5a443543-3560-4e57-b433-be7ad0ac436e req-dbf4c8eb-2444-4e34-9a61-eca82a8fea7a service nova] Releasing lock "refresh_cache-7aa4c0d9-f7ca-42ba-9fd0-4075cd76752b" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 952.588360] env[65726]: DEBUG oslo_concurrency.lockutils [None req-57ebeb7d-b6c8-4c85-80c2-d441f18e658e tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Releasing lock "refresh_cache-6fd70b4a-63e4-4258-9cff-f3b582500b9f" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 952.588962] env[65726]: DEBUG nova.compute.manager [None req-57ebeb7d-b6c8-4c85-80c2-d441f18e658e tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] [instance: 6fd70b4a-63e4-4258-9cff-f3b582500b9f] Instance network_info: |[{"id": "153b4c9d-d01d-4254-9aa6-040705be347a", "address": "fa:16:3e:17:18:d9", "network": {"id": "93e38d7e-b4d7-4d4a-8393-64ab2f6769de", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1445539922-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "06e221dc693640929cac7bc5af45948f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "399f3826-705c-45f7-9fe0-3a08a945151a", "external-id": "nsx-vlan-transportzone-936", "segmentation_id": 936, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap153b4c9d-d0", "ovs_interfaceid": "153b4c9d-d01d-4254-9aa6-040705be347a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 952.589964] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-57ebeb7d-b6c8-4c85-80c2-d441f18e658e tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] [instance: 6fd70b4a-63e4-4258-9cff-f3b582500b9f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:17:18:d9', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '399f3826-705c-45f7-9fe0-3a08a945151a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '153b4c9d-d01d-4254-9aa6-040705be347a', 'vif_model': 'vmxnet3'}] {{(pid=65726) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 952.597753] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-57ebeb7d-b6c8-4c85-80c2-d441f18e658e tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Creating folder: Project (06e221dc693640929cac7bc5af45948f). Parent ref: group-v995008. {{(pid=65726) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 952.598095] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a1a575cf-d142-4252-aad7-75d677ea2584 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.611070] env[65726]: INFO nova.virt.vmwareapi.vm_util [None req-57ebeb7d-b6c8-4c85-80c2-d441f18e658e tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Created folder: Project (06e221dc693640929cac7bc5af45948f) in parent group-v995008. [ 952.611374] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-57ebeb7d-b6c8-4c85-80c2-d441f18e658e tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Creating folder: Instances. Parent ref: group-v995224. {{(pid=65726) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 952.611653] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-bc1916f5-18b0-477f-b45b-ab23cd196500 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.624839] env[65726]: INFO nova.virt.vmwareapi.vm_util [None req-57ebeb7d-b6c8-4c85-80c2-d441f18e658e tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Created folder: Instances in parent group-v995224. [ 952.625206] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-57ebeb7d-b6c8-4c85-80c2-d441f18e658e tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 952.625465] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6fd70b4a-63e4-4258-9cff-f3b582500b9f] Creating VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 952.627094] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-cf01547b-ca56-41ae-a60c-87ebc086f021 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.658652] env[65726]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 952.658652] env[65726]: value = "task-5116315" [ 952.658652] env[65726]: _type = "Task" [ 952.658652] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.672809] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116315, 'name': CreateVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.689751] env[65726]: INFO nova.compute.manager [None req-71fc3c6d-ac1b-4ee5-b5e9-31bd39e153cf tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: cf950ec3-9914-4eb9-99db-048b3969bb21] Rescuing [ 952.690026] env[65726]: DEBUG oslo_concurrency.lockutils [None req-71fc3c6d-ac1b-4ee5-b5e9-31bd39e153cf tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Acquiring lock "refresh_cache-cf950ec3-9914-4eb9-99db-048b3969bb21" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 952.690099] env[65726]: DEBUG oslo_concurrency.lockutils [None req-71fc3c6d-ac1b-4ee5-b5e9-31bd39e153cf tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Acquired lock "refresh_cache-cf950ec3-9914-4eb9-99db-048b3969bb21" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 952.690241] env[65726]: DEBUG nova.network.neutron [None req-71fc3c6d-ac1b-4ee5-b5e9-31bd39e153cf tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: cf950ec3-9914-4eb9-99db-048b3969bb21] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 952.908079] env[65726]: DEBUG nova.scheduler.client.report [None req-2eacb245-5155-4571-979a-d9aaa1a355b7 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 952.926907] env[65726]: DEBUG oslo_vmware.api [None req-da525c49-2593-4b16-8428-bf88b8e252b1 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Task: {'id': task-5116311, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.987037] env[65726]: DEBUG oslo_vmware.api [None req-371355d3-22ed-4e9f-8bfc-a310d98eb1b4 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Task: {'id': task-5116312, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.004615] env[65726]: DEBUG oslo_concurrency.lockutils [req-002127c4-ff40-4b50-8f27-47f4bc24e364 req-c2134f96-4826-4614-9815-a3d04fd2dbe5 service nova] Releasing lock "refresh_cache-df46141c-b545-4e03-b3a3-fd9f5feda0d2" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 953.004910] env[65726]: DEBUG nova.compute.manager [req-002127c4-ff40-4b50-8f27-47f4bc24e364 req-c2134f96-4826-4614-9815-a3d04fd2dbe5 service nova] [instance: 6fd70b4a-63e4-4258-9cff-f3b582500b9f] Received event network-vif-plugged-153b4c9d-d01d-4254-9aa6-040705be347a {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 953.005156] env[65726]: DEBUG oslo_concurrency.lockutils [req-002127c4-ff40-4b50-8f27-47f4bc24e364 req-c2134f96-4826-4614-9815-a3d04fd2dbe5 service nova] Acquiring lock "6fd70b4a-63e4-4258-9cff-f3b582500b9f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 953.005466] env[65726]: DEBUG oslo_concurrency.lockutils [req-002127c4-ff40-4b50-8f27-47f4bc24e364 req-c2134f96-4826-4614-9815-a3d04fd2dbe5 service nova] Lock "6fd70b4a-63e4-4258-9cff-f3b582500b9f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 953.005651] env[65726]: DEBUG oslo_concurrency.lockutils [req-002127c4-ff40-4b50-8f27-47f4bc24e364 req-c2134f96-4826-4614-9815-a3d04fd2dbe5 service nova] Lock "6fd70b4a-63e4-4258-9cff-f3b582500b9f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 953.005825] env[65726]: DEBUG nova.compute.manager [req-002127c4-ff40-4b50-8f27-47f4bc24e364 req-c2134f96-4826-4614-9815-a3d04fd2dbe5 service nova] [instance: 6fd70b4a-63e4-4258-9cff-f3b582500b9f] No waiting events found dispatching network-vif-plugged-153b4c9d-d01d-4254-9aa6-040705be347a {{(pid=65726) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 953.005994] env[65726]: WARNING nova.compute.manager [req-002127c4-ff40-4b50-8f27-47f4bc24e364 req-c2134f96-4826-4614-9815-a3d04fd2dbe5 service nova] [instance: 6fd70b4a-63e4-4258-9cff-f3b582500b9f] Received unexpected event network-vif-plugged-153b4c9d-d01d-4254-9aa6-040705be347a for instance with vm_state building and task_state spawning. [ 953.006194] env[65726]: DEBUG nova.compute.manager [req-002127c4-ff40-4b50-8f27-47f4bc24e364 req-c2134f96-4826-4614-9815-a3d04fd2dbe5 service nova] [instance: 6fd70b4a-63e4-4258-9cff-f3b582500b9f] Received event network-changed-153b4c9d-d01d-4254-9aa6-040705be347a {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 953.006871] env[65726]: DEBUG nova.compute.manager [req-002127c4-ff40-4b50-8f27-47f4bc24e364 req-c2134f96-4826-4614-9815-a3d04fd2dbe5 service nova] [instance: 6fd70b4a-63e4-4258-9cff-f3b582500b9f] Refreshing instance network info cache due to event network-changed-153b4c9d-d01d-4254-9aa6-040705be347a. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 953.006871] env[65726]: DEBUG oslo_concurrency.lockutils [req-002127c4-ff40-4b50-8f27-47f4bc24e364 req-c2134f96-4826-4614-9815-a3d04fd2dbe5 service nova] Acquiring lock "refresh_cache-6fd70b4a-63e4-4258-9cff-f3b582500b9f" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 953.006871] env[65726]: DEBUG oslo_concurrency.lockutils [req-002127c4-ff40-4b50-8f27-47f4bc24e364 req-c2134f96-4826-4614-9815-a3d04fd2dbe5 service nova] Acquired lock "refresh_cache-6fd70b4a-63e4-4258-9cff-f3b582500b9f" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 953.006871] env[65726]: DEBUG nova.network.neutron [req-002127c4-ff40-4b50-8f27-47f4bc24e364 req-c2134f96-4826-4614-9815-a3d04fd2dbe5 service nova] [instance: 6fd70b4a-63e4-4258-9cff-f3b582500b9f] Refreshing network info cache for port 153b4c9d-d01d-4254-9aa6-040705be347a {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 953.175608] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116315, 'name': CreateVM_Task} progress is 99%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.194293] env[65726]: WARNING openstack [None req-71fc3c6d-ac1b-4ee5-b5e9-31bd39e153cf tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 953.194795] env[65726]: WARNING openstack [None req-71fc3c6d-ac1b-4ee5-b5e9-31bd39e153cf tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 953.307943] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c4955188-25cc-47e9-912c-88328888d7f1 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Acquiring lock "01d86089-6b9b-4588-864e-ef91375a1eea" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 953.308268] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c4955188-25cc-47e9-912c-88328888d7f1 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Lock "01d86089-6b9b-4588-864e-ef91375a1eea" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 953.308613] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c4955188-25cc-47e9-912c-88328888d7f1 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Acquiring lock "01d86089-6b9b-4588-864e-ef91375a1eea-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 953.308893] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c4955188-25cc-47e9-912c-88328888d7f1 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Lock "01d86089-6b9b-4588-864e-ef91375a1eea-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 953.309171] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c4955188-25cc-47e9-912c-88328888d7f1 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Lock "01d86089-6b9b-4588-864e-ef91375a1eea-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 953.312275] env[65726]: INFO nova.compute.manager [None req-c4955188-25cc-47e9-912c-88328888d7f1 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 01d86089-6b9b-4588-864e-ef91375a1eea] Terminating instance [ 953.420682] env[65726]: DEBUG oslo_concurrency.lockutils [None req-2eacb245-5155-4571-979a-d9aaa1a355b7 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.170s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 953.426932] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a58f60ab-ca30-4ed0-bbe9-1d319a9c18f8 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.196s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 953.428612] env[65726]: INFO nova.compute.claims [None req-a58f60ab-ca30-4ed0-bbe9-1d319a9c18f8 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] [instance: 73f5b23e-4463-4068-8994-dd2752a2abc9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 953.432297] env[65726]: DEBUG oslo_vmware.api [None req-da525c49-2593-4b16-8428-bf88b8e252b1 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Task: {'id': task-5116311, 'name': CreateSnapshot_Task, 'duration_secs': 0.546028} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.432433] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-da525c49-2593-4b16-8428-bf88b8e252b1 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] [instance: 815ee20b-cb07-4514-bd43-1b74972508a5] Created Snapshot of the VM instance {{(pid=65726) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 953.433450] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd295eb6-7be8-4ab4-bd27-9a1a88302872 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.451827] env[65726]: INFO nova.scheduler.client.report [None req-2eacb245-5155-4571-979a-d9aaa1a355b7 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Deleted allocations for instance c108f2a5-031b-47ee-9a5b-d62c2f42c26b [ 953.483022] env[65726]: DEBUG oslo_vmware.api [None req-371355d3-22ed-4e9f-8bfc-a310d98eb1b4 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Task: {'id': task-5116312, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.534141} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.485031] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-371355d3-22ed-4e9f-8bfc-a310d98eb1b4 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] df46141c-b545-4e03-b3a3-fd9f5feda0d2/df46141c-b545-4e03-b3a3-fd9f5feda0d2.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 953.485031] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-371355d3-22ed-4e9f-8bfc-a310d98eb1b4 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] [instance: df46141c-b545-4e03-b3a3-fd9f5feda0d2] Extending root virtual disk to 1048576 {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 953.485031] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ee51eaf6-fe02-41ec-ab9c-3cbccff68d8d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.492842] env[65726]: DEBUG oslo_vmware.api [None req-371355d3-22ed-4e9f-8bfc-a310d98eb1b4 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Waiting for the task: (returnval){ [ 953.492842] env[65726]: value = "task-5116316" [ 953.492842] env[65726]: _type = "Task" [ 953.492842] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.504847] env[65726]: DEBUG oslo_vmware.api [None req-371355d3-22ed-4e9f-8bfc-a310d98eb1b4 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Task: {'id': task-5116316, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.512026] env[65726]: WARNING openstack [req-002127c4-ff40-4b50-8f27-47f4bc24e364 req-c2134f96-4826-4614-9815-a3d04fd2dbe5 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 953.512026] env[65726]: WARNING openstack [req-002127c4-ff40-4b50-8f27-47f4bc24e364 req-c2134f96-4826-4614-9815-a3d04fd2dbe5 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 953.656274] env[65726]: WARNING openstack [None req-71fc3c6d-ac1b-4ee5-b5e9-31bd39e153cf tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 953.656884] env[65726]: WARNING openstack [None req-71fc3c6d-ac1b-4ee5-b5e9-31bd39e153cf tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 953.676127] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116315, 'name': CreateVM_Task, 'duration_secs': 0.601782} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.676419] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6fd70b4a-63e4-4258-9cff-f3b582500b9f] Created VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 953.677169] env[65726]: WARNING openstack [None req-57ebeb7d-b6c8-4c85-80c2-d441f18e658e tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 953.677541] env[65726]: WARNING openstack [None req-57ebeb7d-b6c8-4c85-80c2-d441f18e658e tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 953.682784] env[65726]: DEBUG oslo_concurrency.lockutils [None req-57ebeb7d-b6c8-4c85-80c2-d441f18e658e tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 953.683025] env[65726]: DEBUG oslo_concurrency.lockutils [None req-57ebeb7d-b6c8-4c85-80c2-d441f18e658e tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 953.683322] env[65726]: DEBUG oslo_concurrency.lockutils [None req-57ebeb7d-b6c8-4c85-80c2-d441f18e658e tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 953.683652] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a8861134-6607-41cd-851f-43847e92e798 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.692189] env[65726]: DEBUG oslo_vmware.api [None req-57ebeb7d-b6c8-4c85-80c2-d441f18e658e tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Waiting for the task: (returnval){ [ 953.692189] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5260125a-e69d-8fee-c043-5dd1afb544e3" [ 953.692189] env[65726]: _type = "Task" [ 953.692189] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.701594] env[65726]: DEBUG oslo_vmware.api [None req-57ebeb7d-b6c8-4c85-80c2-d441f18e658e tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5260125a-e69d-8fee-c043-5dd1afb544e3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.785376] env[65726]: WARNING openstack [req-002127c4-ff40-4b50-8f27-47f4bc24e364 req-c2134f96-4826-4614-9815-a3d04fd2dbe5 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 953.785376] env[65726]: WARNING openstack [req-002127c4-ff40-4b50-8f27-47f4bc24e364 req-c2134f96-4826-4614-9815-a3d04fd2dbe5 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 953.817166] env[65726]: DEBUG nova.compute.manager [None req-c4955188-25cc-47e9-912c-88328888d7f1 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 01d86089-6b9b-4588-864e-ef91375a1eea] Start destroying the instance on the hypervisor. {{(pid=65726) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 953.817448] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-c4955188-25cc-47e9-912c-88328888d7f1 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 01d86089-6b9b-4588-864e-ef91375a1eea] Destroying instance {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 953.818377] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78e22657-54e3-4ee6-8797-2e22b29eb371 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.829740] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-c4955188-25cc-47e9-912c-88328888d7f1 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 01d86089-6b9b-4588-864e-ef91375a1eea] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 953.831228] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fa78e02f-3a6d-4741-af0e-dda3c8a6f99d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.838443] env[65726]: DEBUG oslo_vmware.api [None req-c4955188-25cc-47e9-912c-88328888d7f1 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Waiting for the task: (returnval){ [ 953.838443] env[65726]: value = "task-5116317" [ 953.838443] env[65726]: _type = "Task" [ 953.838443] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.850203] env[65726]: DEBUG oslo_vmware.api [None req-c4955188-25cc-47e9-912c-88328888d7f1 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': task-5116317, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.871671] env[65726]: WARNING openstack [None req-71fc3c6d-ac1b-4ee5-b5e9-31bd39e153cf tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 953.872183] env[65726]: WARNING openstack [None req-71fc3c6d-ac1b-4ee5-b5e9-31bd39e153cf tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 953.963437] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-da525c49-2593-4b16-8428-bf88b8e252b1 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] [instance: 815ee20b-cb07-4514-bd43-1b74972508a5] Creating linked-clone VM from snapshot {{(pid=65726) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 953.967820] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-ec38c5e1-1b2a-42a9-8d3e-1dd6adea5929 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.972098] env[65726]: DEBUG oslo_concurrency.lockutils [None req-2eacb245-5155-4571-979a-d9aaa1a355b7 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Lock "c108f2a5-031b-47ee-9a5b-d62c2f42c26b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 26.035s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 953.984670] env[65726]: DEBUG oslo_vmware.api [None req-da525c49-2593-4b16-8428-bf88b8e252b1 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Waiting for the task: (returnval){ [ 953.984670] env[65726]: value = "task-5116318" [ 953.984670] env[65726]: _type = "Task" [ 953.984670] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.998889] env[65726]: WARNING openstack [req-002127c4-ff40-4b50-8f27-47f4bc24e364 req-c2134f96-4826-4614-9815-a3d04fd2dbe5 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 953.998889] env[65726]: WARNING openstack [req-002127c4-ff40-4b50-8f27-47f4bc24e364 req-c2134f96-4826-4614-9815-a3d04fd2dbe5 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 954.015987] env[65726]: DEBUG oslo_vmware.api [None req-da525c49-2593-4b16-8428-bf88b8e252b1 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Task: {'id': task-5116318, 'name': CloneVM_Task} progress is 10%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.027815] env[65726]: DEBUG oslo_vmware.api [None req-371355d3-22ed-4e9f-8bfc-a310d98eb1b4 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Task: {'id': task-5116316, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.295952} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.028143] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-371355d3-22ed-4e9f-8bfc-a310d98eb1b4 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] [instance: df46141c-b545-4e03-b3a3-fd9f5feda0d2] Extended root virtual disk {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 954.030165] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab7a17d9-8431-4bf1-8fa1-ebc2ca1329f0 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.037913] env[65726]: DEBUG nova.network.neutron [None req-71fc3c6d-ac1b-4ee5-b5e9-31bd39e153cf tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: cf950ec3-9914-4eb9-99db-048b3969bb21] Updating instance_info_cache with network_info: [{"id": "58accf41-587f-4367-854e-fc571a6a3424", "address": "fa:16:3e:83:9d:0c", "network": {"id": "9d5ad167-99fb-4208-ab30-7053f35774e4", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-2082516762-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.132", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a8362295a62f4b51bae719a7ef5a4656", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dabbac20-1723-40ad-9da0-e53b28073651", "external-id": "nsx-vlan-transportzone-790", "segmentation_id": 790, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap58accf41-58", "ovs_interfaceid": "58accf41-587f-4367-854e-fc571a6a3424", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 954.066932] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-371355d3-22ed-4e9f-8bfc-a310d98eb1b4 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] [instance: df46141c-b545-4e03-b3a3-fd9f5feda0d2] Reconfiguring VM instance instance-0000004b to attach disk [datastore1] df46141c-b545-4e03-b3a3-fd9f5feda0d2/df46141c-b545-4e03-b3a3-fd9f5feda0d2.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 954.071307] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a7afde83-3b80-453f-81cb-751424186f9d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.096060] env[65726]: DEBUG oslo_vmware.api [None req-371355d3-22ed-4e9f-8bfc-a310d98eb1b4 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Waiting for the task: (returnval){ [ 954.096060] env[65726]: value = "task-5116319" [ 954.096060] env[65726]: _type = "Task" [ 954.096060] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.108055] env[65726]: DEBUG oslo_vmware.api [None req-371355d3-22ed-4e9f-8bfc-a310d98eb1b4 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Task: {'id': task-5116319, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.204965] env[65726]: DEBUG oslo_vmware.api [None req-57ebeb7d-b6c8-4c85-80c2-d441f18e658e tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5260125a-e69d-8fee-c043-5dd1afb544e3, 'name': SearchDatastore_Task, 'duration_secs': 0.046181} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.205490] env[65726]: DEBUG oslo_concurrency.lockutils [None req-57ebeb7d-b6c8-4c85-80c2-d441f18e658e tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 954.205861] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-57ebeb7d-b6c8-4c85-80c2-d441f18e658e tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] [instance: 6fd70b4a-63e4-4258-9cff-f3b582500b9f] Processing image b52362a3-ee8a-4cbf-b06f-513b0cc8f95c {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 954.206479] env[65726]: DEBUG oslo_concurrency.lockutils [None req-57ebeb7d-b6c8-4c85-80c2-d441f18e658e tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 954.208120] env[65726]: DEBUG oslo_concurrency.lockutils [None req-57ebeb7d-b6c8-4c85-80c2-d441f18e658e tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 954.208120] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-57ebeb7d-b6c8-4c85-80c2-d441f18e658e tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 954.208120] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9b50d70b-5415-4882-9bb6-a3161de7616c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.226058] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-57ebeb7d-b6c8-4c85-80c2-d441f18e658e tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 954.226287] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-57ebeb7d-b6c8-4c85-80c2-d441f18e658e tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65726) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 954.227013] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5861f3fd-aa78-45e6-8a4a-dc33f806e368 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.234552] env[65726]: DEBUG oslo_vmware.api [None req-57ebeb7d-b6c8-4c85-80c2-d441f18e658e tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Waiting for the task: (returnval){ [ 954.234552] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]525e847d-722b-5d4a-7fa9-b9d5761ab6ac" [ 954.234552] env[65726]: _type = "Task" [ 954.234552] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.248553] env[65726]: DEBUG oslo_vmware.api [None req-57ebeb7d-b6c8-4c85-80c2-d441f18e658e tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]525e847d-722b-5d4a-7fa9-b9d5761ab6ac, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.267894] env[65726]: DEBUG nova.network.neutron [req-002127c4-ff40-4b50-8f27-47f4bc24e364 req-c2134f96-4826-4614-9815-a3d04fd2dbe5 service nova] [instance: 6fd70b4a-63e4-4258-9cff-f3b582500b9f] Updated VIF entry in instance network info cache for port 153b4c9d-d01d-4254-9aa6-040705be347a. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 954.268188] env[65726]: DEBUG nova.network.neutron [req-002127c4-ff40-4b50-8f27-47f4bc24e364 req-c2134f96-4826-4614-9815-a3d04fd2dbe5 service nova] [instance: 6fd70b4a-63e4-4258-9cff-f3b582500b9f] Updating instance_info_cache with network_info: [{"id": "153b4c9d-d01d-4254-9aa6-040705be347a", "address": "fa:16:3e:17:18:d9", "network": {"id": "93e38d7e-b4d7-4d4a-8393-64ab2f6769de", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1445539922-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "06e221dc693640929cac7bc5af45948f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "399f3826-705c-45f7-9fe0-3a08a945151a", "external-id": "nsx-vlan-transportzone-936", "segmentation_id": 936, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap153b4c9d-d0", "ovs_interfaceid": "153b4c9d-d01d-4254-9aa6-040705be347a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 954.350750] env[65726]: DEBUG oslo_vmware.api [None req-c4955188-25cc-47e9-912c-88328888d7f1 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': task-5116317, 'name': PowerOffVM_Task} progress is 100%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.505512] env[65726]: DEBUG oslo_vmware.api [None req-da525c49-2593-4b16-8428-bf88b8e252b1 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Task: {'id': task-5116318, 'name': CloneVM_Task} progress is 94%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.541652] env[65726]: DEBUG oslo_concurrency.lockutils [None req-71fc3c6d-ac1b-4ee5-b5e9-31bd39e153cf tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Releasing lock "refresh_cache-cf950ec3-9914-4eb9-99db-048b3969bb21" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 954.597279] env[65726]: DEBUG oslo_concurrency.lockutils [None req-dd3cd0c9-6ca2-4870-ab72-9325931d98de tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Acquiring lock "274ab469-61a9-4b7e-852c-074c871e3abf" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 954.597678] env[65726]: DEBUG oslo_concurrency.lockutils [None req-dd3cd0c9-6ca2-4870-ab72-9325931d98de tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Lock "274ab469-61a9-4b7e-852c-074c871e3abf" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.002s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 954.597941] env[65726]: DEBUG oslo_concurrency.lockutils [None req-dd3cd0c9-6ca2-4870-ab72-9325931d98de tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Acquiring lock "274ab469-61a9-4b7e-852c-074c871e3abf-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 954.598245] env[65726]: DEBUG oslo_concurrency.lockutils [None req-dd3cd0c9-6ca2-4870-ab72-9325931d98de tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Lock "274ab469-61a9-4b7e-852c-074c871e3abf-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 954.598390] env[65726]: DEBUG oslo_concurrency.lockutils [None req-dd3cd0c9-6ca2-4870-ab72-9325931d98de tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Lock "274ab469-61a9-4b7e-852c-074c871e3abf-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 954.609369] env[65726]: INFO nova.compute.manager [None req-dd3cd0c9-6ca2-4870-ab72-9325931d98de tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: 274ab469-61a9-4b7e-852c-074c871e3abf] Terminating instance [ 954.621766] env[65726]: DEBUG oslo_vmware.api [None req-371355d3-22ed-4e9f-8bfc-a310d98eb1b4 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Task: {'id': task-5116319, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.636685] env[65726]: DEBUG oslo_concurrency.lockutils [None req-d903724e-08f6-4863-a079-5fadb490f6c5 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Acquiring lock "54e85161-5f63-405f-83f8-490c70645a3b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 954.636984] env[65726]: DEBUG oslo_concurrency.lockutils [None req-d903724e-08f6-4863-a079-5fadb490f6c5 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Lock "54e85161-5f63-405f-83f8-490c70645a3b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 954.750142] env[65726]: DEBUG oslo_vmware.api [None req-57ebeb7d-b6c8-4c85-80c2-d441f18e658e tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]525e847d-722b-5d4a-7fa9-b9d5761ab6ac, 'name': SearchDatastore_Task, 'duration_secs': 0.017081} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.751239] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e10c4288-3d55-4aa3-b32f-484096b29294 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.773024] env[65726]: DEBUG oslo_concurrency.lockutils [req-002127c4-ff40-4b50-8f27-47f4bc24e364 req-c2134f96-4826-4614-9815-a3d04fd2dbe5 service nova] Releasing lock "refresh_cache-6fd70b4a-63e4-4258-9cff-f3b582500b9f" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 954.773334] env[65726]: DEBUG oslo_vmware.api [None req-57ebeb7d-b6c8-4c85-80c2-d441f18e658e tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Waiting for the task: (returnval){ [ 954.773334] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52e32ef8-170e-2d96-1c2e-b9d5562f6581" [ 954.773334] env[65726]: _type = "Task" [ 954.773334] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.786736] env[65726]: DEBUG oslo_vmware.api [None req-57ebeb7d-b6c8-4c85-80c2-d441f18e658e tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52e32ef8-170e-2d96-1c2e-b9d5562f6581, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.860163] env[65726]: DEBUG oslo_vmware.api [None req-c4955188-25cc-47e9-912c-88328888d7f1 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': task-5116317, 'name': PowerOffVM_Task, 'duration_secs': 0.526568} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.860295] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-c4955188-25cc-47e9-912c-88328888d7f1 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 01d86089-6b9b-4588-864e-ef91375a1eea] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 954.860532] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-c4955188-25cc-47e9-912c-88328888d7f1 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 01d86089-6b9b-4588-864e-ef91375a1eea] Unregistering the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 954.861073] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-baf1ad90-0373-4ab9-a1e6-7daa84c41e8e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.932182] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5bc57a0e-e9ed-446e-b9b2-ff4d014c30c8 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.945211] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db150774-03e3-4d4e-9977-fee21c9e9b98 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.950461] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-c4955188-25cc-47e9-912c-88328888d7f1 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 01d86089-6b9b-4588-864e-ef91375a1eea] Unregistered the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 954.950676] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-c4955188-25cc-47e9-912c-88328888d7f1 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 01d86089-6b9b-4588-864e-ef91375a1eea] Deleting contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 954.950844] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-c4955188-25cc-47e9-912c-88328888d7f1 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Deleting the datastore file [datastore1] 01d86089-6b9b-4588-864e-ef91375a1eea {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 954.951675] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-cb7c6f46-401d-4932-b931-9c79b71b6802 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.984457] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc25b070-0d2a-4db4-afed-a3156632cb99 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.987721] env[65726]: DEBUG oslo_vmware.api [None req-c4955188-25cc-47e9-912c-88328888d7f1 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Waiting for the task: (returnval){ [ 954.987721] env[65726]: value = "task-5116321" [ 954.987721] env[65726]: _type = "Task" [ 954.987721] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.000453] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2de9be5e-423d-4cf8-a82c-980677b45734 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.013439] env[65726]: DEBUG oslo_vmware.api [None req-c4955188-25cc-47e9-912c-88328888d7f1 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': task-5116321, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.013800] env[65726]: DEBUG oslo_vmware.api [None req-da525c49-2593-4b16-8428-bf88b8e252b1 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Task: {'id': task-5116318, 'name': CloneVM_Task} progress is 94%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.025549] env[65726]: DEBUG nova.compute.provider_tree [None req-a58f60ab-ca30-4ed0-bbe9-1d319a9c18f8 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 955.115391] env[65726]: DEBUG oslo_vmware.api [None req-371355d3-22ed-4e9f-8bfc-a310d98eb1b4 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Task: {'id': task-5116319, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.122964] env[65726]: DEBUG nova.compute.manager [None req-dd3cd0c9-6ca2-4870-ab72-9325931d98de tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: 274ab469-61a9-4b7e-852c-074c871e3abf] Start destroying the instance on the hypervisor. {{(pid=65726) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 955.123237] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-dd3cd0c9-6ca2-4870-ab72-9325931d98de tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: 274ab469-61a9-4b7e-852c-074c871e3abf] Destroying instance {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 955.124482] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0eb455a8-126b-448b-bbd5-8a5b27d5715a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.137541] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-dd3cd0c9-6ca2-4870-ab72-9325931d98de tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: 274ab469-61a9-4b7e-852c-074c871e3abf] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 955.138519] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-eada0d57-50b6-42e6-8377-11b10aedc7fd {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.141555] env[65726]: DEBUG nova.compute.manager [None req-d903724e-08f6-4863-a079-5fadb490f6c5 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 54e85161-5f63-405f-83f8-490c70645a3b] Starting instance... {{(pid=65726) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 955.152483] env[65726]: DEBUG oslo_vmware.api [None req-dd3cd0c9-6ca2-4870-ab72-9325931d98de tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Waiting for the task: (returnval){ [ 955.152483] env[65726]: value = "task-5116322" [ 955.152483] env[65726]: _type = "Task" [ 955.152483] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.167282] env[65726]: DEBUG oslo_vmware.api [None req-dd3cd0c9-6ca2-4870-ab72-9325931d98de tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Task: {'id': task-5116322, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.290181] env[65726]: DEBUG oslo_vmware.api [None req-57ebeb7d-b6c8-4c85-80c2-d441f18e658e tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52e32ef8-170e-2d96-1c2e-b9d5562f6581, 'name': SearchDatastore_Task, 'duration_secs': 0.020726} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.290481] env[65726]: DEBUG oslo_concurrency.lockutils [None req-57ebeb7d-b6c8-4c85-80c2-d441f18e658e tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 955.290796] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-57ebeb7d-b6c8-4c85-80c2-d441f18e658e tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] 6fd70b4a-63e4-4258-9cff-f3b582500b9f/6fd70b4a-63e4-4258-9cff-f3b582500b9f.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 955.291128] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0fd6da34-1eb3-46e1-be1e-b8e83e5905ec {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.302345] env[65726]: DEBUG oslo_vmware.api [None req-57ebeb7d-b6c8-4c85-80c2-d441f18e658e tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Waiting for the task: (returnval){ [ 955.302345] env[65726]: value = "task-5116323" [ 955.302345] env[65726]: _type = "Task" [ 955.302345] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.310682] env[65726]: DEBUG oslo_vmware.api [None req-57ebeb7d-b6c8-4c85-80c2-d441f18e658e tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Task: {'id': task-5116323, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.502751] env[65726]: DEBUG oslo_vmware.api [None req-c4955188-25cc-47e9-912c-88328888d7f1 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': task-5116321, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.330595} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.514432] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-c4955188-25cc-47e9-912c-88328888d7f1 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Deleted the datastore file {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 955.514432] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-c4955188-25cc-47e9-912c-88328888d7f1 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 01d86089-6b9b-4588-864e-ef91375a1eea] Deleted contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 955.514432] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-c4955188-25cc-47e9-912c-88328888d7f1 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 01d86089-6b9b-4588-864e-ef91375a1eea] Instance destroyed {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 955.514432] env[65726]: INFO nova.compute.manager [None req-c4955188-25cc-47e9-912c-88328888d7f1 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 01d86089-6b9b-4588-864e-ef91375a1eea] Took 1.69 seconds to destroy the instance on the hypervisor. [ 955.514432] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-c4955188-25cc-47e9-912c-88328888d7f1 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 955.514432] env[65726]: DEBUG oslo_vmware.api [None req-da525c49-2593-4b16-8428-bf88b8e252b1 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Task: {'id': task-5116318, 'name': CloneVM_Task} progress is 94%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.514432] env[65726]: DEBUG nova.compute.manager [-] [instance: 01d86089-6b9b-4588-864e-ef91375a1eea] Deallocating network for instance {{(pid=65726) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 955.514432] env[65726]: DEBUG nova.network.neutron [-] [instance: 01d86089-6b9b-4588-864e-ef91375a1eea] deallocate_for_instance() {{(pid=65726) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 955.514432] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 955.514432] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 955.530547] env[65726]: DEBUG nova.scheduler.client.report [None req-a58f60ab-ca30-4ed0-bbe9-1d319a9c18f8 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 955.574029] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 955.574029] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 955.614245] env[65726]: DEBUG oslo_vmware.api [None req-371355d3-22ed-4e9f-8bfc-a310d98eb1b4 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Task: {'id': task-5116319, 'name': ReconfigVM_Task, 'duration_secs': 1.49692} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.617099] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-371355d3-22ed-4e9f-8bfc-a310d98eb1b4 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] [instance: df46141c-b545-4e03-b3a3-fd9f5feda0d2] Reconfigured VM instance instance-0000004b to attach disk [datastore1] df46141c-b545-4e03-b3a3-fd9f5feda0d2/df46141c-b545-4e03-b3a3-fd9f5feda0d2.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 955.617099] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3abbd634-4480-4f12-a2b7-b2b13b31f2ab {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.626389] env[65726]: DEBUG oslo_vmware.api [None req-371355d3-22ed-4e9f-8bfc-a310d98eb1b4 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Waiting for the task: (returnval){ [ 955.626389] env[65726]: value = "task-5116324" [ 955.626389] env[65726]: _type = "Task" [ 955.626389] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.636534] env[65726]: DEBUG oslo_vmware.api [None req-371355d3-22ed-4e9f-8bfc-a310d98eb1b4 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Task: {'id': task-5116324, 'name': Rename_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.668101] env[65726]: DEBUG oslo_vmware.api [None req-dd3cd0c9-6ca2-4870-ab72-9325931d98de tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Task: {'id': task-5116322, 'name': PowerOffVM_Task, 'duration_secs': 0.258623} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.668101] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-dd3cd0c9-6ca2-4870-ab72-9325931d98de tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: 274ab469-61a9-4b7e-852c-074c871e3abf] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 955.668232] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-dd3cd0c9-6ca2-4870-ab72-9325931d98de tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: 274ab469-61a9-4b7e-852c-074c871e3abf] Unregistering the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 955.668549] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b0c25da8-2112-449a-9200-c25392ad8495 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.675592] env[65726]: DEBUG oslo_concurrency.lockutils [None req-d903724e-08f6-4863-a079-5fadb490f6c5 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 955.754238] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-dd3cd0c9-6ca2-4870-ab72-9325931d98de tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: 274ab469-61a9-4b7e-852c-074c871e3abf] Unregistered the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 955.754573] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-dd3cd0c9-6ca2-4870-ab72-9325931d98de tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: 274ab469-61a9-4b7e-852c-074c871e3abf] Deleting contents of the VM from datastore datastore2 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 955.754814] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-dd3cd0c9-6ca2-4870-ab72-9325931d98de tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Deleting the datastore file [datastore2] 274ab469-61a9-4b7e-852c-074c871e3abf {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 955.755145] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f3215f0c-9a44-4283-bab0-f07e52a5b5b7 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.766420] env[65726]: DEBUG oslo_vmware.api [None req-dd3cd0c9-6ca2-4870-ab72-9325931d98de tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Waiting for the task: (returnval){ [ 955.766420] env[65726]: value = "task-5116326" [ 955.766420] env[65726]: _type = "Task" [ 955.766420] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.779063] env[65726]: DEBUG oslo_vmware.api [None req-dd3cd0c9-6ca2-4870-ab72-9325931d98de tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Task: {'id': task-5116326, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.813537] env[65726]: DEBUG oslo_vmware.api [None req-57ebeb7d-b6c8-4c85-80c2-d441f18e658e tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Task: {'id': task-5116323, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.999647] env[65726]: DEBUG oslo_vmware.api [None req-da525c49-2593-4b16-8428-bf88b8e252b1 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Task: {'id': task-5116318, 'name': CloneVM_Task, 'duration_secs': 1.91845} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.000076] env[65726]: INFO nova.virt.vmwareapi.vmops [None req-da525c49-2593-4b16-8428-bf88b8e252b1 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] [instance: 815ee20b-cb07-4514-bd43-1b74972508a5] Created linked-clone VM from snapshot [ 956.001022] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c48cbab0-b675-4a36-8dc7-4c51821d2721 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.009363] env[65726]: DEBUG nova.virt.vmwareapi.images [None req-da525c49-2593-4b16-8428-bf88b8e252b1 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] [instance: 815ee20b-cb07-4514-bd43-1b74972508a5] Uploading image eb5566e4-ef83-4dc3-89c5-bc440d53be6b {{(pid=65726) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 956.040748] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a58f60ab-ca30-4ed0-bbe9-1d319a9c18f8 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.614s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 956.041343] env[65726]: DEBUG nova.compute.manager [None req-a58f60ab-ca30-4ed0-bbe9-1d319a9c18f8 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] [instance: 73f5b23e-4463-4068-8994-dd2752a2abc9] Start building networks asynchronously for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 956.045793] env[65726]: DEBUG oslo_concurrency.lockutils [None req-816608ea-3ce8-4b57-8b12-cbe3b89f21e9 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 23.006s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 956.046180] env[65726]: DEBUG nova.objects.instance [None req-816608ea-3ce8-4b57-8b12-cbe3b89f21e9 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Lazy-loading 'resources' on Instance uuid 5a252ef1-93c9-4bff-842b-b64df2bc5d75 {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 956.049737] env[65726]: DEBUG oslo_vmware.rw_handles [None req-da525c49-2593-4b16-8428-bf88b8e252b1 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 956.049737] env[65726]: value = "vm-995228" [ 956.049737] env[65726]: _type = "VirtualMachine" [ 956.049737] env[65726]: }. {{(pid=65726) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 956.050654] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-550088df-2b7d-4429-b7f3-4bc97cb72553 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.064652] env[65726]: DEBUG oslo_vmware.rw_handles [None req-da525c49-2593-4b16-8428-bf88b8e252b1 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Lease: (returnval){ [ 956.064652] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52dcd249-8ea4-7c9f-b21b-61240de604db" [ 956.064652] env[65726]: _type = "HttpNfcLease" [ 956.064652] env[65726]: } obtained for exporting VM: (result){ [ 956.064652] env[65726]: value = "vm-995228" [ 956.064652] env[65726]: _type = "VirtualMachine" [ 956.064652] env[65726]: }. {{(pid=65726) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 956.065103] env[65726]: DEBUG oslo_vmware.api [None req-da525c49-2593-4b16-8428-bf88b8e252b1 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Waiting for the lease: (returnval){ [ 956.065103] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52dcd249-8ea4-7c9f-b21b-61240de604db" [ 956.065103] env[65726]: _type = "HttpNfcLease" [ 956.065103] env[65726]: } to be ready. {{(pid=65726) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 956.074077] env[65726]: DEBUG nova.compute.manager [req-aa84476f-6830-4989-95c4-9497c38a71c1 req-ab629779-4990-4e0c-8d86-185332336e14 service nova] [instance: 01d86089-6b9b-4588-864e-ef91375a1eea] Received event network-vif-deleted-beb674aa-cade-455d-a97a-4da9699c9a70 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 956.074077] env[65726]: INFO nova.compute.manager [req-aa84476f-6830-4989-95c4-9497c38a71c1 req-ab629779-4990-4e0c-8d86-185332336e14 service nova] [instance: 01d86089-6b9b-4588-864e-ef91375a1eea] Neutron deleted interface beb674aa-cade-455d-a97a-4da9699c9a70; detaching it from the instance and deleting it from the info cache [ 956.074283] env[65726]: DEBUG nova.network.neutron [req-aa84476f-6830-4989-95c4-9497c38a71c1 req-ab629779-4990-4e0c-8d86-185332336e14 service nova] [instance: 01d86089-6b9b-4588-864e-ef91375a1eea] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 956.081243] env[65726]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 956.081243] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52dcd249-8ea4-7c9f-b21b-61240de604db" [ 956.081243] env[65726]: _type = "HttpNfcLease" [ 956.081243] env[65726]: } is ready. {{(pid=65726) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 956.082071] env[65726]: DEBUG oslo_vmware.rw_handles [None req-da525c49-2593-4b16-8428-bf88b8e252b1 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 956.082071] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52dcd249-8ea4-7c9f-b21b-61240de604db" [ 956.082071] env[65726]: _type = "HttpNfcLease" [ 956.082071] env[65726]: }. {{(pid=65726) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 956.082917] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61fd7246-fd7a-47f6-8a65-1b4211df4a11 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.092143] env[65726]: DEBUG oslo_vmware.rw_handles [None req-da525c49-2593-4b16-8428-bf88b8e252b1 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/527908c0-5e5e-60f3-9262-0e56a59cf070/disk-0.vmdk from lease info. {{(pid=65726) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 956.092419] env[65726]: DEBUG oslo_vmware.rw_handles [None req-da525c49-2593-4b16-8428-bf88b8e252b1 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/527908c0-5e5e-60f3-9262-0e56a59cf070/disk-0.vmdk for reading. {{(pid=65726) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 956.151274] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-71fc3c6d-ac1b-4ee5-b5e9-31bd39e153cf tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: cf950ec3-9914-4eb9-99db-048b3969bb21] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 956.152364] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5ba736a3-64f8-401c-b015-f41fb0fbbe2a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.165381] env[65726]: DEBUG oslo_vmware.api [None req-371355d3-22ed-4e9f-8bfc-a310d98eb1b4 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Task: {'id': task-5116324, 'name': Rename_Task, 'duration_secs': 0.211016} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.167195] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-371355d3-22ed-4e9f-8bfc-a310d98eb1b4 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] [instance: df46141c-b545-4e03-b3a3-fd9f5feda0d2] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 956.167571] env[65726]: DEBUG oslo_vmware.api [None req-71fc3c6d-ac1b-4ee5-b5e9-31bd39e153cf tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Waiting for the task: (returnval){ [ 956.167571] env[65726]: value = "task-5116328" [ 956.167571] env[65726]: _type = "Task" [ 956.167571] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.167846] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e98bc973-4b14-48f9-a3f7-726e49296357 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.178469] env[65726]: DEBUG oslo_vmware.api [None req-71fc3c6d-ac1b-4ee5-b5e9-31bd39e153cf tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Task: {'id': task-5116328, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.179908] env[65726]: DEBUG oslo_vmware.api [None req-371355d3-22ed-4e9f-8bfc-a310d98eb1b4 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Waiting for the task: (returnval){ [ 956.179908] env[65726]: value = "task-5116329" [ 956.179908] env[65726]: _type = "Task" [ 956.179908] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.192376] env[65726]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-f1ded242-e436-43b4-85a7-0798a0505173 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.278201] env[65726]: DEBUG oslo_vmware.api [None req-dd3cd0c9-6ca2-4870-ab72-9325931d98de tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Task: {'id': task-5116326, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.316123} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.278453] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-dd3cd0c9-6ca2-4870-ab72-9325931d98de tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Deleted the datastore file {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 956.278634] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-dd3cd0c9-6ca2-4870-ab72-9325931d98de tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: 274ab469-61a9-4b7e-852c-074c871e3abf] Deleted contents of the VM from datastore datastore2 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 956.278810] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-dd3cd0c9-6ca2-4870-ab72-9325931d98de tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: 274ab469-61a9-4b7e-852c-074c871e3abf] Instance destroyed {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 956.278975] env[65726]: INFO nova.compute.manager [None req-dd3cd0c9-6ca2-4870-ab72-9325931d98de tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: 274ab469-61a9-4b7e-852c-074c871e3abf] Took 1.16 seconds to destroy the instance on the hypervisor. [ 956.279234] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-dd3cd0c9-6ca2-4870-ab72-9325931d98de tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 956.279661] env[65726]: DEBUG nova.compute.manager [-] [instance: 274ab469-61a9-4b7e-852c-074c871e3abf] Deallocating network for instance {{(pid=65726) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 956.279768] env[65726]: DEBUG nova.network.neutron [-] [instance: 274ab469-61a9-4b7e-852c-074c871e3abf] deallocate_for_instance() {{(pid=65726) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 956.280297] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 956.280545] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 956.311379] env[65726]: DEBUG oslo_vmware.api [None req-57ebeb7d-b6c8-4c85-80c2-d441f18e658e tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Task: {'id': task-5116323, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.690019} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.311805] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-57ebeb7d-b6c8-4c85-80c2-d441f18e658e tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] 6fd70b4a-63e4-4258-9cff-f3b582500b9f/6fd70b4a-63e4-4258-9cff-f3b582500b9f.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 956.311929] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-57ebeb7d-b6c8-4c85-80c2-d441f18e658e tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] [instance: 6fd70b4a-63e4-4258-9cff-f3b582500b9f] Extending root virtual disk to 1048576 {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 956.312414] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-74f89b23-4a80-4e52-b1d9-ff91233f8fd9 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.320602] env[65726]: DEBUG oslo_vmware.api [None req-57ebeb7d-b6c8-4c85-80c2-d441f18e658e tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Waiting for the task: (returnval){ [ 956.320602] env[65726]: value = "task-5116330" [ 956.320602] env[65726]: _type = "Task" [ 956.320602] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.325366] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 956.325667] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 956.339727] env[65726]: DEBUG oslo_vmware.api [None req-57ebeb7d-b6c8-4c85-80c2-d441f18e658e tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Task: {'id': task-5116330, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.409237] env[65726]: DEBUG nova.network.neutron [-] [instance: 01d86089-6b9b-4588-864e-ef91375a1eea] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 956.553227] env[65726]: DEBUG nova.compute.utils [None req-a58f60ab-ca30-4ed0-bbe9-1d319a9c18f8 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] Using /dev/sd instead of None {{(pid=65726) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 956.556729] env[65726]: DEBUG nova.compute.manager [None req-a58f60ab-ca30-4ed0-bbe9-1d319a9c18f8 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] [instance: 73f5b23e-4463-4068-8994-dd2752a2abc9] Not allocating networking since 'none' was specified. {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2015}} [ 956.589121] env[65726]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-3991238f-3f08-467d-9e1a-d3d5a5c328c1 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.604215] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ea41786-cacb-4ef7-9843-a9f46a6c9399 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.656691] env[65726]: DEBUG nova.compute.manager [req-aa84476f-6830-4989-95c4-9497c38a71c1 req-ab629779-4990-4e0c-8d86-185332336e14 service nova] [instance: 01d86089-6b9b-4588-864e-ef91375a1eea] Detach interface failed, port_id=beb674aa-cade-455d-a97a-4da9699c9a70, reason: Instance 01d86089-6b9b-4588-864e-ef91375a1eea could not be found. {{(pid=65726) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11669}} [ 956.680537] env[65726]: DEBUG oslo_vmware.api [None req-71fc3c6d-ac1b-4ee5-b5e9-31bd39e153cf tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Task: {'id': task-5116328, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.694695] env[65726]: DEBUG oslo_vmware.api [None req-371355d3-22ed-4e9f-8bfc-a310d98eb1b4 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Task: {'id': task-5116329, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.787977] env[65726]: DEBUG nova.compute.manager [req-27f84b54-98d6-46cd-a50e-50145b691553 req-81a7bfa1-9234-4564-b8b2-bd318bc3c1d5 service nova] [instance: 274ab469-61a9-4b7e-852c-074c871e3abf] Received event network-vif-deleted-757d86b3-4474-41c5-aaf8-14fc7216cf15 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 956.788948] env[65726]: INFO nova.compute.manager [req-27f84b54-98d6-46cd-a50e-50145b691553 req-81a7bfa1-9234-4564-b8b2-bd318bc3c1d5 service nova] [instance: 274ab469-61a9-4b7e-852c-074c871e3abf] Neutron deleted interface 757d86b3-4474-41c5-aaf8-14fc7216cf15; detaching it from the instance and deleting it from the info cache [ 956.789162] env[65726]: DEBUG nova.network.neutron [req-27f84b54-98d6-46cd-a50e-50145b691553 req-81a7bfa1-9234-4564-b8b2-bd318bc3c1d5 service nova] [instance: 274ab469-61a9-4b7e-852c-074c871e3abf] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 956.837155] env[65726]: DEBUG oslo_vmware.api [None req-57ebeb7d-b6c8-4c85-80c2-d441f18e658e tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Task: {'id': task-5116330, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.077192} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.841029] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-57ebeb7d-b6c8-4c85-80c2-d441f18e658e tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] [instance: 6fd70b4a-63e4-4258-9cff-f3b582500b9f] Extended root virtual disk {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 956.842483] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7aa9259a-f7df-4e02-af5d-cd4aa4eeb5a1 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.886280] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-57ebeb7d-b6c8-4c85-80c2-d441f18e658e tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] [instance: 6fd70b4a-63e4-4258-9cff-f3b582500b9f] Reconfiguring VM instance instance-0000004c to attach disk [datastore1] 6fd70b4a-63e4-4258-9cff-f3b582500b9f/6fd70b4a-63e4-4258-9cff-f3b582500b9f.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 956.891727] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3fcade5a-09ba-4031-9672-e589932aae7f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.916407] env[65726]: INFO nova.compute.manager [-] [instance: 01d86089-6b9b-4588-864e-ef91375a1eea] Took 1.40 seconds to deallocate network for instance. [ 956.924311] env[65726]: DEBUG oslo_vmware.api [None req-57ebeb7d-b6c8-4c85-80c2-d441f18e658e tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Waiting for the task: (returnval){ [ 956.924311] env[65726]: value = "task-5116331" [ 956.924311] env[65726]: _type = "Task" [ 956.924311] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.935468] env[65726]: DEBUG oslo_vmware.api [None req-57ebeb7d-b6c8-4c85-80c2-d441f18e658e tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Task: {'id': task-5116331, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.060457] env[65726]: DEBUG nova.compute.manager [None req-a58f60ab-ca30-4ed0-bbe9-1d319a9c18f8 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] [instance: 73f5b23e-4463-4068-8994-dd2752a2abc9] Start building block device mappings for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 957.123782] env[65726]: DEBUG nova.network.neutron [-] [instance: 274ab469-61a9-4b7e-852c-074c871e3abf] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 957.160817] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d31890b9-8ed8-4f75-a307-311943264a95 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.171239] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c936c51-0c85-4806-b607-48fcf9ebb21f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.220973] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d880fd8-ec57-4c6f-bf7b-bd4d112290ad {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.224089] env[65726]: DEBUG oslo_vmware.api [None req-71fc3c6d-ac1b-4ee5-b5e9-31bd39e153cf tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Task: {'id': task-5116328, 'name': PowerOffVM_Task, 'duration_secs': 0.944141} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.229277] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-71fc3c6d-ac1b-4ee5-b5e9-31bd39e153cf tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: cf950ec3-9914-4eb9-99db-048b3969bb21] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 957.229277] env[65726]: DEBUG oslo_vmware.api [None req-371355d3-22ed-4e9f-8bfc-a310d98eb1b4 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Task: {'id': task-5116329, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.229924] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d73ab68-7501-4c33-bfa9-ce05ffbc4047 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.239098] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0845cd01-adad-48ac-9365-69eccfe767e1 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.269097] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c063a471-8be2-4e8d-9e4d-a01c03b57aba {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.283662] env[65726]: DEBUG nova.compute.provider_tree [None req-816608ea-3ce8-4b57-8b12-cbe3b89f21e9 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 957.296868] env[65726]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5d8edbe0-d628-4e8f-8ff5-a5115c5c928d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.310325] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-224bd610-2dcb-43a9-9ea8-e03b01668538 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.339036] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-71fc3c6d-ac1b-4ee5-b5e9-31bd39e153cf tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: cf950ec3-9914-4eb9-99db-048b3969bb21] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 957.339911] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b7724088-acdb-42e7-8a60-fc8b35d272a5 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.361951] env[65726]: DEBUG nova.compute.manager [req-27f84b54-98d6-46cd-a50e-50145b691553 req-81a7bfa1-9234-4564-b8b2-bd318bc3c1d5 service nova] [instance: 274ab469-61a9-4b7e-852c-074c871e3abf] Detach interface failed, port_id=757d86b3-4474-41c5-aaf8-14fc7216cf15, reason: Instance 274ab469-61a9-4b7e-852c-074c871e3abf could not be found. {{(pid=65726) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11669}} [ 957.369655] env[65726]: DEBUG oslo_vmware.api [None req-71fc3c6d-ac1b-4ee5-b5e9-31bd39e153cf tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Waiting for the task: (returnval){ [ 957.369655] env[65726]: value = "task-5116332" [ 957.369655] env[65726]: _type = "Task" [ 957.369655] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.382326] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-71fc3c6d-ac1b-4ee5-b5e9-31bd39e153cf tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: cf950ec3-9914-4eb9-99db-048b3969bb21] VM already powered off {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 957.383387] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-71fc3c6d-ac1b-4ee5-b5e9-31bd39e153cf tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: cf950ec3-9914-4eb9-99db-048b3969bb21] Processing image b52362a3-ee8a-4cbf-b06f-513b0cc8f95c {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 957.383387] env[65726]: DEBUG oslo_concurrency.lockutils [None req-71fc3c6d-ac1b-4ee5-b5e9-31bd39e153cf tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 957.383387] env[65726]: DEBUG oslo_concurrency.lockutils [None req-71fc3c6d-ac1b-4ee5-b5e9-31bd39e153cf tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 957.383387] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-71fc3c6d-ac1b-4ee5-b5e9-31bd39e153cf tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 957.383671] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a94785ac-c294-48fe-b6e0-84d6ecc72398 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.394887] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-71fc3c6d-ac1b-4ee5-b5e9-31bd39e153cf tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 957.395039] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-71fc3c6d-ac1b-4ee5-b5e9-31bd39e153cf tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65726) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 957.396026] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a19260d2-a2d3-4267-9c6e-13272433f2a9 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.402321] env[65726]: DEBUG oslo_vmware.api [None req-71fc3c6d-ac1b-4ee5-b5e9-31bd39e153cf tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Waiting for the task: (returnval){ [ 957.402321] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52c9777f-28ce-a28d-5df7-3ca0d52401aa" [ 957.402321] env[65726]: _type = "Task" [ 957.402321] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.413183] env[65726]: DEBUG oslo_vmware.api [None req-71fc3c6d-ac1b-4ee5-b5e9-31bd39e153cf tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52c9777f-28ce-a28d-5df7-3ca0d52401aa, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.423866] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c4955188-25cc-47e9-912c-88328888d7f1 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 957.435718] env[65726]: DEBUG oslo_vmware.api [None req-57ebeb7d-b6c8-4c85-80c2-d441f18e658e tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Task: {'id': task-5116331, 'name': ReconfigVM_Task, 'duration_secs': 0.344013} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.436024] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-57ebeb7d-b6c8-4c85-80c2-d441f18e658e tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] [instance: 6fd70b4a-63e4-4258-9cff-f3b582500b9f] Reconfigured VM instance instance-0000004c to attach disk [datastore1] 6fd70b4a-63e4-4258-9cff-f3b582500b9f/6fd70b4a-63e4-4258-9cff-f3b582500b9f.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 957.436696] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e2345c93-d344-45a2-9a96-448f9f50518b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.444746] env[65726]: DEBUG oslo_vmware.api [None req-57ebeb7d-b6c8-4c85-80c2-d441f18e658e tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Waiting for the task: (returnval){ [ 957.444746] env[65726]: value = "task-5116333" [ 957.444746] env[65726]: _type = "Task" [ 957.444746] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.455277] env[65726]: DEBUG oslo_vmware.api [None req-57ebeb7d-b6c8-4c85-80c2-d441f18e658e tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Task: {'id': task-5116333, 'name': Rename_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.629511] env[65726]: INFO nova.compute.manager [-] [instance: 274ab469-61a9-4b7e-852c-074c871e3abf] Took 1.35 seconds to deallocate network for instance. [ 957.698716] env[65726]: DEBUG oslo_vmware.api [None req-371355d3-22ed-4e9f-8bfc-a310d98eb1b4 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Task: {'id': task-5116329, 'name': PowerOnVM_Task, 'duration_secs': 1.234039} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.699717] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-371355d3-22ed-4e9f-8bfc-a310d98eb1b4 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] [instance: df46141c-b545-4e03-b3a3-fd9f5feda0d2] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 957.699717] env[65726]: INFO nova.compute.manager [None req-371355d3-22ed-4e9f-8bfc-a310d98eb1b4 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] [instance: df46141c-b545-4e03-b3a3-fd9f5feda0d2] Took 11.16 seconds to spawn the instance on the hypervisor. [ 957.699867] env[65726]: DEBUG nova.compute.manager [None req-371355d3-22ed-4e9f-8bfc-a310d98eb1b4 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] [instance: df46141c-b545-4e03-b3a3-fd9f5feda0d2] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 957.701179] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a051521-3cb6-421b-be42-c8f21805004b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.795895] env[65726]: DEBUG nova.scheduler.client.report [None req-816608ea-3ce8-4b57-8b12-cbe3b89f21e9 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 957.914296] env[65726]: DEBUG oslo_vmware.api [None req-71fc3c6d-ac1b-4ee5-b5e9-31bd39e153cf tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52c9777f-28ce-a28d-5df7-3ca0d52401aa, 'name': SearchDatastore_Task, 'duration_secs': 0.011406} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.916591] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-85230643-1215-4544-885c-eb5c2543dd58 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.923227] env[65726]: DEBUG oslo_vmware.api [None req-71fc3c6d-ac1b-4ee5-b5e9-31bd39e153cf tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Waiting for the task: (returnval){ [ 957.923227] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52477533-2a4d-659c-66f7-7a0af7ecd7a5" [ 957.923227] env[65726]: _type = "Task" [ 957.923227] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.934906] env[65726]: DEBUG oslo_vmware.api [None req-71fc3c6d-ac1b-4ee5-b5e9-31bd39e153cf tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52477533-2a4d-659c-66f7-7a0af7ecd7a5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.956043] env[65726]: DEBUG oslo_vmware.api [None req-57ebeb7d-b6c8-4c85-80c2-d441f18e658e tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Task: {'id': task-5116333, 'name': Rename_Task, 'duration_secs': 0.16583} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.956043] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-57ebeb7d-b6c8-4c85-80c2-d441f18e658e tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] [instance: 6fd70b4a-63e4-4258-9cff-f3b582500b9f] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 957.956043] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a10a72c5-8549-49cf-a04a-65f0ecea69e9 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.963708] env[65726]: DEBUG oslo_vmware.api [None req-57ebeb7d-b6c8-4c85-80c2-d441f18e658e tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Waiting for the task: (returnval){ [ 957.963708] env[65726]: value = "task-5116334" [ 957.963708] env[65726]: _type = "Task" [ 957.963708] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.974310] env[65726]: DEBUG oslo_vmware.api [None req-57ebeb7d-b6c8-4c85-80c2-d441f18e658e tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Task: {'id': task-5116334, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.073898] env[65726]: DEBUG nova.compute.manager [None req-a58f60ab-ca30-4ed0-bbe9-1d319a9c18f8 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] [instance: 73f5b23e-4463-4068-8994-dd2752a2abc9] Start spawning the instance on the hypervisor. {{(pid=65726) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 958.103943] env[65726]: DEBUG nova.virt.hardware [None req-a58f60ab-ca30-4ed0-bbe9-1d319a9c18f8 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T19:28:34Z,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1539f06fa3534e90acbc3a60c4b8bd3f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T19:28:35Z,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 958.104588] env[65726]: DEBUG nova.virt.hardware [None req-a58f60ab-ca30-4ed0-bbe9-1d319a9c18f8 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 958.104990] env[65726]: DEBUG nova.virt.hardware [None req-a58f60ab-ca30-4ed0-bbe9-1d319a9c18f8 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 958.106263] env[65726]: DEBUG nova.virt.hardware [None req-a58f60ab-ca30-4ed0-bbe9-1d319a9c18f8 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 958.106263] env[65726]: DEBUG nova.virt.hardware [None req-a58f60ab-ca30-4ed0-bbe9-1d319a9c18f8 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 958.106263] env[65726]: DEBUG nova.virt.hardware [None req-a58f60ab-ca30-4ed0-bbe9-1d319a9c18f8 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 958.106263] env[65726]: DEBUG nova.virt.hardware [None req-a58f60ab-ca30-4ed0-bbe9-1d319a9c18f8 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 958.106716] env[65726]: DEBUG nova.virt.hardware [None req-a58f60ab-ca30-4ed0-bbe9-1d319a9c18f8 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 958.107048] env[65726]: DEBUG nova.virt.hardware [None req-a58f60ab-ca30-4ed0-bbe9-1d319a9c18f8 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 958.107413] env[65726]: DEBUG nova.virt.hardware [None req-a58f60ab-ca30-4ed0-bbe9-1d319a9c18f8 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 958.107714] env[65726]: DEBUG nova.virt.hardware [None req-a58f60ab-ca30-4ed0-bbe9-1d319a9c18f8 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 958.108856] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acadd50f-758f-4b37-ae52-171c350b1058 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.123036] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-184ade14-5035-4be5-ae7e-fe7cbcd7ff26 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.140209] env[65726]: DEBUG oslo_concurrency.lockutils [None req-dd3cd0c9-6ca2-4870-ab72-9325931d98de tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 958.140877] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-a58f60ab-ca30-4ed0-bbe9-1d319a9c18f8 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] [instance: 73f5b23e-4463-4068-8994-dd2752a2abc9] Instance VIF info [] {{(pid=65726) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 958.148179] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-a58f60ab-ca30-4ed0-bbe9-1d319a9c18f8 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] Creating folder: Project (6ceda9baf4a44a0db8539857f1005752). Parent ref: group-v995008. {{(pid=65726) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 958.152325] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-81fb4f38-02e8-456f-90c8-3e484f785eba {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.161573] env[65726]: INFO nova.virt.vmwareapi.vm_util [None req-a58f60ab-ca30-4ed0-bbe9-1d319a9c18f8 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] Created folder: Project (6ceda9baf4a44a0db8539857f1005752) in parent group-v995008. [ 958.161939] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-a58f60ab-ca30-4ed0-bbe9-1d319a9c18f8 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] Creating folder: Instances. Parent ref: group-v995229. {{(pid=65726) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 958.162337] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5a74c66a-37b1-413b-aa5c-3874119de20b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.174571] env[65726]: INFO nova.virt.vmwareapi.vm_util [None req-a58f60ab-ca30-4ed0-bbe9-1d319a9c18f8 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] Created folder: Instances in parent group-v995229. [ 958.175335] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-a58f60ab-ca30-4ed0-bbe9-1d319a9c18f8 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 958.175335] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 73f5b23e-4463-4068-8994-dd2752a2abc9] Creating VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 958.175466] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-038c997e-5169-4644-888f-a905ce1ddd5b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.194949] env[65726]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 958.194949] env[65726]: value = "task-5116337" [ 958.194949] env[65726]: _type = "Task" [ 958.194949] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 958.206113] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116337, 'name': CreateVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.229399] env[65726]: INFO nova.compute.manager [None req-371355d3-22ed-4e9f-8bfc-a310d98eb1b4 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] [instance: df46141c-b545-4e03-b3a3-fd9f5feda0d2] Took 35.56 seconds to build instance. [ 958.301722] env[65726]: DEBUG oslo_concurrency.lockutils [None req-816608ea-3ce8-4b57-8b12-cbe3b89f21e9 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.256s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 958.308505] env[65726]: DEBUG oslo_concurrency.lockutils [None req-eb162c35-1e60-4d30-91d4-e0662af06e95 tempest-ServerTagsTestJSON-59142761 tempest-ServerTagsTestJSON-59142761-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 22.222s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 958.308813] env[65726]: DEBUG nova.objects.instance [None req-eb162c35-1e60-4d30-91d4-e0662af06e95 tempest-ServerTagsTestJSON-59142761 tempest-ServerTagsTestJSON-59142761-project-member] Lazy-loading 'resources' on Instance uuid 9029549c-1914-4bae-91e2-8812b79051ec {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 958.335012] env[65726]: INFO nova.scheduler.client.report [None req-816608ea-3ce8-4b57-8b12-cbe3b89f21e9 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Deleted allocations for instance 5a252ef1-93c9-4bff-842b-b64df2bc5d75 [ 958.435787] env[65726]: DEBUG oslo_vmware.api [None req-71fc3c6d-ac1b-4ee5-b5e9-31bd39e153cf tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52477533-2a4d-659c-66f7-7a0af7ecd7a5, 'name': SearchDatastore_Task, 'duration_secs': 0.011895} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 958.436175] env[65726]: DEBUG oslo_concurrency.lockutils [None req-71fc3c6d-ac1b-4ee5-b5e9-31bd39e153cf tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 958.436830] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-71fc3c6d-ac1b-4ee5-b5e9-31bd39e153cf tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Copying virtual disk from [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] cf950ec3-9914-4eb9-99db-048b3969bb21/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c-rescue.vmdk. {{(pid=65726) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 958.437319] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6f64a8f2-633b-426c-93e9-36a929fea4b7 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.447052] env[65726]: DEBUG oslo_vmware.api [None req-71fc3c6d-ac1b-4ee5-b5e9-31bd39e153cf tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Waiting for the task: (returnval){ [ 958.447052] env[65726]: value = "task-5116338" [ 958.447052] env[65726]: _type = "Task" [ 958.447052] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 958.459581] env[65726]: DEBUG oslo_vmware.api [None req-71fc3c6d-ac1b-4ee5-b5e9-31bd39e153cf tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Task: {'id': task-5116338, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.476304] env[65726]: DEBUG oslo_vmware.api [None req-57ebeb7d-b6c8-4c85-80c2-d441f18e658e tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Task: {'id': task-5116334, 'name': PowerOnVM_Task} progress is 94%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.706055] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116337, 'name': CreateVM_Task, 'duration_secs': 0.379796} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 958.706351] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 73f5b23e-4463-4068-8994-dd2752a2abc9] Created VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 958.706858] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a58f60ab-ca30-4ed0-bbe9-1d319a9c18f8 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 958.707093] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a58f60ab-ca30-4ed0-bbe9-1d319a9c18f8 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 958.707453] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a58f60ab-ca30-4ed0-bbe9-1d319a9c18f8 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 958.707772] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-109784b7-8071-4efe-9848-209d62a3f553 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.715700] env[65726]: DEBUG oslo_vmware.api [None req-a58f60ab-ca30-4ed0-bbe9-1d319a9c18f8 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] Waiting for the task: (returnval){ [ 958.715700] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52c0d66c-726f-6e77-13ed-d51f27439f37" [ 958.715700] env[65726]: _type = "Task" [ 958.715700] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 958.728485] env[65726]: DEBUG oslo_vmware.api [None req-a58f60ab-ca30-4ed0-bbe9-1d319a9c18f8 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52c0d66c-726f-6e77-13ed-d51f27439f37, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.734255] env[65726]: DEBUG oslo_concurrency.lockutils [None req-371355d3-22ed-4e9f-8bfc-a310d98eb1b4 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Lock "df46141c-b545-4e03-b3a3-fd9f5feda0d2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 37.077s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 958.752699] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c75c9ed0-c9ca-4fd9-9695-87538dcf1f78 tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Acquiring lock "406c59e7-8eb8-4cfd-8e1e-e151ab908571" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 958.753013] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c75c9ed0-c9ca-4fd9-9695-87538dcf1f78 tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Lock "406c59e7-8eb8-4cfd-8e1e-e151ab908571" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 958.845581] env[65726]: DEBUG oslo_concurrency.lockutils [None req-816608ea-3ce8-4b57-8b12-cbe3b89f21e9 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Lock "5a252ef1-93c9-4bff-842b-b64df2bc5d75" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 31.070s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 958.967690] env[65726]: DEBUG oslo_vmware.api [None req-71fc3c6d-ac1b-4ee5-b5e9-31bd39e153cf tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Task: {'id': task-5116338, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.980744] env[65726]: DEBUG oslo_vmware.api [None req-57ebeb7d-b6c8-4c85-80c2-d441f18e658e tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Task: {'id': task-5116334, 'name': PowerOnVM_Task, 'duration_secs': 0.570166} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 958.981140] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-57ebeb7d-b6c8-4c85-80c2-d441f18e658e tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] [instance: 6fd70b4a-63e4-4258-9cff-f3b582500b9f] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 958.982080] env[65726]: INFO nova.compute.manager [None req-57ebeb7d-b6c8-4c85-80c2-d441f18e658e tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] [instance: 6fd70b4a-63e4-4258-9cff-f3b582500b9f] Took 9.86 seconds to spawn the instance on the hypervisor. [ 958.982080] env[65726]: DEBUG nova.compute.manager [None req-57ebeb7d-b6c8-4c85-80c2-d441f18e658e tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] [instance: 6fd70b4a-63e4-4258-9cff-f3b582500b9f] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 958.982638] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4e102b6-ef4b-4f99-ace5-c72f10204dd9 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.190599] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d407c43-700c-454a-9bfb-629b895ec1dd {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.198794] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c60e17a5-650c-401d-88ad-1f6516b46082 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.250396] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7c3ac9f-83df-4ffa-ac41-ef151cc60e4f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.267455] env[65726]: DEBUG nova.compute.manager [None req-c75c9ed0-c9ca-4fd9-9695-87538dcf1f78 tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] [instance: 406c59e7-8eb8-4cfd-8e1e-e151ab908571] Starting instance... {{(pid=65726) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 959.271293] env[65726]: DEBUG oslo_vmware.api [None req-a58f60ab-ca30-4ed0-bbe9-1d319a9c18f8 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52c0d66c-726f-6e77-13ed-d51f27439f37, 'name': SearchDatastore_Task, 'duration_secs': 0.082562} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.273085] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24816b23-2e49-42e2-a284-e25257e1f936 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.279895] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a58f60ab-ca30-4ed0-bbe9-1d319a9c18f8 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 959.280185] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-a58f60ab-ca30-4ed0-bbe9-1d319a9c18f8 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] [instance: 73f5b23e-4463-4068-8994-dd2752a2abc9] Processing image b52362a3-ee8a-4cbf-b06f-513b0cc8f95c {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 959.280489] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a58f60ab-ca30-4ed0-bbe9-1d319a9c18f8 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 959.280639] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a58f60ab-ca30-4ed0-bbe9-1d319a9c18f8 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 959.280829] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-a58f60ab-ca30-4ed0-bbe9-1d319a9c18f8 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 959.281633] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7c1c73ab-d7d5-4636-a871-401661564383 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.298861] env[65726]: DEBUG nova.compute.provider_tree [None req-eb162c35-1e60-4d30-91d4-e0662af06e95 tempest-ServerTagsTestJSON-59142761 tempest-ServerTagsTestJSON-59142761-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 959.302851] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-a58f60ab-ca30-4ed0-bbe9-1d319a9c18f8 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 959.302851] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-a58f60ab-ca30-4ed0-bbe9-1d319a9c18f8 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65726) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 959.304224] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0becdb32-708d-4a03-9c1e-975b2cc70235 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.316444] env[65726]: DEBUG oslo_vmware.api [None req-a58f60ab-ca30-4ed0-bbe9-1d319a9c18f8 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] Waiting for the task: (returnval){ [ 959.316444] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]525297d4-5a62-0a81-0b96-53edacef163c" [ 959.316444] env[65726]: _type = "Task" [ 959.316444] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.329346] env[65726]: DEBUG oslo_vmware.api [None req-a58f60ab-ca30-4ed0-bbe9-1d319a9c18f8 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]525297d4-5a62-0a81-0b96-53edacef163c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.460628] env[65726]: DEBUG oslo_vmware.api [None req-71fc3c6d-ac1b-4ee5-b5e9-31bd39e153cf tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Task: {'id': task-5116338, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.66293} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.461019] env[65726]: INFO nova.virt.vmwareapi.ds_util [None req-71fc3c6d-ac1b-4ee5-b5e9-31bd39e153cf tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Copied virtual disk from [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] cf950ec3-9914-4eb9-99db-048b3969bb21/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c-rescue.vmdk. [ 959.461973] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0dadc2e-ca55-459b-8c96-bd5263860bb1 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.493756] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-71fc3c6d-ac1b-4ee5-b5e9-31bd39e153cf tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: cf950ec3-9914-4eb9-99db-048b3969bb21] Reconfiguring VM instance instance-0000003e to attach disk [datastore1] cf950ec3-9914-4eb9-99db-048b3969bb21/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c-rescue.vmdk or device None with type thin {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 959.494195] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-63c9e0f3-0e9f-4ecc-9990-f40c1468fab3 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.516211] env[65726]: INFO nova.compute.manager [None req-57ebeb7d-b6c8-4c85-80c2-d441f18e658e tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] [instance: 6fd70b4a-63e4-4258-9cff-f3b582500b9f] Took 31.00 seconds to build instance. [ 959.519956] env[65726]: DEBUG oslo_vmware.api [None req-71fc3c6d-ac1b-4ee5-b5e9-31bd39e153cf tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Waiting for the task: (returnval){ [ 959.519956] env[65726]: value = "task-5116339" [ 959.519956] env[65726]: _type = "Task" [ 959.519956] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.529355] env[65726]: DEBUG oslo_vmware.api [None req-71fc3c6d-ac1b-4ee5-b5e9-31bd39e153cf tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Task: {'id': task-5116339, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.802742] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c75c9ed0-c9ca-4fd9-9695-87538dcf1f78 tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 959.806237] env[65726]: DEBUG nova.scheduler.client.report [None req-eb162c35-1e60-4d30-91d4-e0662af06e95 tempest-ServerTagsTestJSON-59142761 tempest-ServerTagsTestJSON-59142761-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 959.828341] env[65726]: DEBUG oslo_vmware.api [None req-a58f60ab-ca30-4ed0-bbe9-1d319a9c18f8 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]525297d4-5a62-0a81-0b96-53edacef163c, 'name': SearchDatastore_Task, 'duration_secs': 0.015835} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.829217] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-af046556-d1db-4f67-8263-a345dad688cd {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.835826] env[65726]: DEBUG oslo_vmware.api [None req-a58f60ab-ca30-4ed0-bbe9-1d319a9c18f8 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] Waiting for the task: (returnval){ [ 959.835826] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52b32044-7fa6-2b0a-4bfc-4ebf691e53c1" [ 959.835826] env[65726]: _type = "Task" [ 959.835826] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.845992] env[65726]: DEBUG oslo_vmware.api [None req-a58f60ab-ca30-4ed0-bbe9-1d319a9c18f8 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52b32044-7fa6-2b0a-4bfc-4ebf691e53c1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.018435] env[65726]: DEBUG oslo_concurrency.lockutils [None req-57ebeb7d-b6c8-4c85-80c2-d441f18e658e tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Lock "6fd70b4a-63e4-4258-9cff-f3b582500b9f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 32.516s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 960.031412] env[65726]: DEBUG oslo_vmware.api [None req-71fc3c6d-ac1b-4ee5-b5e9-31bd39e153cf tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Task: {'id': task-5116339, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.312014] env[65726]: DEBUG oslo_concurrency.lockutils [None req-eb162c35-1e60-4d30-91d4-e0662af06e95 tempest-ServerTagsTestJSON-59142761 tempest-ServerTagsTestJSON-59142761-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.004s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 960.315070] env[65726]: DEBUG oslo_concurrency.lockutils [None req-f5777e4a-de17-4332-8f51-ad1e8cbe3f19 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 19.100s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 960.315070] env[65726]: DEBUG nova.objects.instance [None req-f5777e4a-de17-4332-8f51-ad1e8cbe3f19 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Lazy-loading 'resources' on Instance uuid 0e064341-4e4a-407b-8c26-3eb04b409029 {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 960.350131] env[65726]: DEBUG oslo_vmware.api [None req-a58f60ab-ca30-4ed0-bbe9-1d319a9c18f8 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52b32044-7fa6-2b0a-4bfc-4ebf691e53c1, 'name': SearchDatastore_Task, 'duration_secs': 0.020595} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.351117] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a58f60ab-ca30-4ed0-bbe9-1d319a9c18f8 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 960.351117] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-a58f60ab-ca30-4ed0-bbe9-1d319a9c18f8 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] 73f5b23e-4463-4068-8994-dd2752a2abc9/73f5b23e-4463-4068-8994-dd2752a2abc9.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 960.351117] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7a0bb833-39ce-47a1-9b34-76c78ffe22b8 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.359930] env[65726]: DEBUG oslo_vmware.api [None req-a58f60ab-ca30-4ed0-bbe9-1d319a9c18f8 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] Waiting for the task: (returnval){ [ 960.359930] env[65726]: value = "task-5116340" [ 960.359930] env[65726]: _type = "Task" [ 960.359930] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 960.370026] env[65726]: DEBUG oslo_vmware.api [None req-a58f60ab-ca30-4ed0-bbe9-1d319a9c18f8 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] Task: {'id': task-5116340, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.439047] env[65726]: INFO nova.scheduler.client.report [None req-eb162c35-1e60-4d30-91d4-e0662af06e95 tempest-ServerTagsTestJSON-59142761 tempest-ServerTagsTestJSON-59142761-project-member] Deleted allocations for instance 9029549c-1914-4bae-91e2-8812b79051ec [ 960.533899] env[65726]: DEBUG oslo_vmware.api [None req-71fc3c6d-ac1b-4ee5-b5e9-31bd39e153cf tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Task: {'id': task-5116339, 'name': ReconfigVM_Task, 'duration_secs': 0.727719} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.534274] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-71fc3c6d-ac1b-4ee5-b5e9-31bd39e153cf tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: cf950ec3-9914-4eb9-99db-048b3969bb21] Reconfigured VM instance instance-0000003e to attach disk [datastore1] cf950ec3-9914-4eb9-99db-048b3969bb21/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c-rescue.vmdk or device None with type thin {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 960.536287] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9656d27d-1c50-4094-b2cc-6d5bf1f584a2 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.569964] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ade31399-ce50-4f55-89c3-31c21ea3f21b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.594111] env[65726]: DEBUG oslo_vmware.api [None req-71fc3c6d-ac1b-4ee5-b5e9-31bd39e153cf tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Waiting for the task: (returnval){ [ 960.594111] env[65726]: value = "task-5116341" [ 960.594111] env[65726]: _type = "Task" [ 960.594111] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 960.604973] env[65726]: DEBUG oslo_vmware.api [None req-71fc3c6d-ac1b-4ee5-b5e9-31bd39e153cf tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Task: {'id': task-5116341, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.872173] env[65726]: DEBUG oslo_vmware.api [None req-a58f60ab-ca30-4ed0-bbe9-1d319a9c18f8 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] Task: {'id': task-5116340, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.948027] env[65726]: DEBUG oslo_concurrency.lockutils [None req-eb162c35-1e60-4d30-91d4-e0662af06e95 tempest-ServerTagsTestJSON-59142761 tempest-ServerTagsTestJSON-59142761-project-member] Lock "9029549c-1914-4bae-91e2-8812b79051ec" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 28.424s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 961.090867] env[65726]: DEBUG nova.compute.manager [None req-ace31bce-a963-4a72-a87f-23a0d9efcfe0 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] [instance: df46141c-b545-4e03-b3a3-fd9f5feda0d2] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 961.098092] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb1245fc-97a1-4636-bb6f-a744471a96d8 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.114311] env[65726]: DEBUG oslo_vmware.api [None req-71fc3c6d-ac1b-4ee5-b5e9-31bd39e153cf tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Task: {'id': task-5116341, 'name': ReconfigVM_Task, 'duration_secs': 0.251358} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 961.114861] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-71fc3c6d-ac1b-4ee5-b5e9-31bd39e153cf tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: cf950ec3-9914-4eb9-99db-048b3969bb21] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 961.115195] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5ee09265-3419-4200-ba76-ba7801fcfac6 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.126417] env[65726]: DEBUG oslo_vmware.api [None req-71fc3c6d-ac1b-4ee5-b5e9-31bd39e153cf tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Waiting for the task: (returnval){ [ 961.126417] env[65726]: value = "task-5116342" [ 961.126417] env[65726]: _type = "Task" [ 961.126417] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 961.138873] env[65726]: DEBUG oslo_vmware.api [None req-71fc3c6d-ac1b-4ee5-b5e9-31bd39e153cf tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Task: {'id': task-5116342, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.287153] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30e1f3df-eac5-4323-ac8b-4a1bf17acd84 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.300883] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b8d8388-2894-45d6-bb1a-621cef0c4638 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.335946] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7818ee05-bd21-403c-8f3b-b0766c03a596 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.344919] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6e711cf-e1d1-4a44-9aae-19c773fe7d79 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.360547] env[65726]: DEBUG nova.compute.provider_tree [None req-f5777e4a-de17-4332-8f51-ad1e8cbe3f19 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Updating inventory in ProviderTree for provider 07c4692f-bdb4-4058-9173-ff9664830295 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 961.377293] env[65726]: DEBUG oslo_vmware.api [None req-a58f60ab-ca30-4ed0-bbe9-1d319a9c18f8 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] Task: {'id': task-5116340, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.68778} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 961.378304] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-a58f60ab-ca30-4ed0-bbe9-1d319a9c18f8 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] 73f5b23e-4463-4068-8994-dd2752a2abc9/73f5b23e-4463-4068-8994-dd2752a2abc9.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 961.378558] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-a58f60ab-ca30-4ed0-bbe9-1d319a9c18f8 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] [instance: 73f5b23e-4463-4068-8994-dd2752a2abc9] Extending root virtual disk to 1048576 {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 961.378884] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-6ef0df26-314e-48fe-9667-8c8f0d2c4946 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.388206] env[65726]: DEBUG oslo_vmware.api [None req-a58f60ab-ca30-4ed0-bbe9-1d319a9c18f8 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] Waiting for the task: (returnval){ [ 961.388206] env[65726]: value = "task-5116343" [ 961.388206] env[65726]: _type = "Task" [ 961.388206] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 961.397682] env[65726]: DEBUG oslo_vmware.api [None req-a58f60ab-ca30-4ed0-bbe9-1d319a9c18f8 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] Task: {'id': task-5116343, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.618628] env[65726]: INFO nova.compute.manager [None req-ace31bce-a963-4a72-a87f-23a0d9efcfe0 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] [instance: df46141c-b545-4e03-b3a3-fd9f5feda0d2] instance snapshotting [ 961.623613] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc4d7bb3-6baa-496f-a88a-82d0cfdb456f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.651047] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec17e63d-070b-45aa-b3d1-b6f25540ff72 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.658792] env[65726]: DEBUG oslo_vmware.api [None req-71fc3c6d-ac1b-4ee5-b5e9-31bd39e153cf tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Task: {'id': task-5116342, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.900279] env[65726]: ERROR nova.scheduler.client.report [None req-f5777e4a-de17-4332-8f51-ad1e8cbe3f19 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] [req-a92b3fd3-3bef-4d3e-b2c6-0f42775b7799] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 07c4692f-bdb4-4058-9173-ff9664830295. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-a92b3fd3-3bef-4d3e-b2c6-0f42775b7799"}]} [ 961.911028] env[65726]: DEBUG oslo_vmware.api [None req-a58f60ab-ca30-4ed0-bbe9-1d319a9c18f8 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] Task: {'id': task-5116343, 'name': ExtendVirtualDisk_Task} progress is 50%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.918617] env[65726]: DEBUG nova.scheduler.client.report [None req-f5777e4a-de17-4332-8f51-ad1e8cbe3f19 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Refreshing inventories for resource provider 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 961.934891] env[65726]: DEBUG nova.scheduler.client.report [None req-f5777e4a-de17-4332-8f51-ad1e8cbe3f19 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Updating ProviderTree inventory for provider 07c4692f-bdb4-4058-9173-ff9664830295 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 961.935079] env[65726]: DEBUG nova.compute.provider_tree [None req-f5777e4a-de17-4332-8f51-ad1e8cbe3f19 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Updating inventory in ProviderTree for provider 07c4692f-bdb4-4058-9173-ff9664830295 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 961.948673] env[65726]: DEBUG nova.scheduler.client.report [None req-f5777e4a-de17-4332-8f51-ad1e8cbe3f19 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Refreshing aggregate associations for resource provider 07c4692f-bdb4-4058-9173-ff9664830295, aggregates: None {{(pid=65726) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 961.970196] env[65726]: DEBUG nova.scheduler.client.report [None req-f5777e4a-de17-4332-8f51-ad1e8cbe3f19 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Refreshing trait associations for resource provider 07c4692f-bdb4-4058-9173-ff9664830295, traits: COMPUTE_IMAGE_TYPE_ISO,HW_ARCH_X86_64,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=65726) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 962.142250] env[65726]: DEBUG oslo_vmware.api [None req-71fc3c6d-ac1b-4ee5-b5e9-31bd39e153cf tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Task: {'id': task-5116342, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.174417] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-ace31bce-a963-4a72-a87f-23a0d9efcfe0 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] [instance: df46141c-b545-4e03-b3a3-fd9f5feda0d2] Creating Snapshot of the VM instance {{(pid=65726) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 962.175359] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-f9f8c770-6492-4337-8c4f-fe6d570027fb {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.192088] env[65726]: DEBUG oslo_vmware.api [None req-ace31bce-a963-4a72-a87f-23a0d9efcfe0 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Waiting for the task: (returnval){ [ 962.192088] env[65726]: value = "task-5116344" [ 962.192088] env[65726]: _type = "Task" [ 962.192088] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.204198] env[65726]: DEBUG oslo_vmware.api [None req-ace31bce-a963-4a72-a87f-23a0d9efcfe0 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Task: {'id': task-5116344, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.210637] env[65726]: DEBUG oslo_concurrency.lockutils [None req-83256413-d207-4cc8-b904-cfb2a5ea5628 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Acquiring lock "f9a40d36-dda9-454c-8774-ed011f3f50ae" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 962.210853] env[65726]: DEBUG oslo_concurrency.lockutils [None req-83256413-d207-4cc8-b904-cfb2a5ea5628 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Lock "f9a40d36-dda9-454c-8774-ed011f3f50ae" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 962.403204] env[65726]: DEBUG oslo_vmware.api [None req-a58f60ab-ca30-4ed0-bbe9-1d319a9c18f8 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] Task: {'id': task-5116343, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.574411} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 962.404801] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-a58f60ab-ca30-4ed0-bbe9-1d319a9c18f8 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] [instance: 73f5b23e-4463-4068-8994-dd2752a2abc9] Extended root virtual disk {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 962.405696] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd6b627e-3515-43d0-9d64-e17195720e28 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.409533] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d478bbb3-4b91-4a8c-93ed-1daa0f674a78 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.426368] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94e8e81a-afb2-40c4-a4ee-b5310bfee536 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.438036] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-a58f60ab-ca30-4ed0-bbe9-1d319a9c18f8 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] [instance: 73f5b23e-4463-4068-8994-dd2752a2abc9] Reconfiguring VM instance instance-0000004d to attach disk [datastore1] 73f5b23e-4463-4068-8994-dd2752a2abc9/73f5b23e-4463-4068-8994-dd2752a2abc9.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 962.438271] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-91763bef-8726-4fa0-86ef-85c8175ef875 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.487640] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e85aea91-07d1-4979-9d34-21e71c7ada8f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.491179] env[65726]: DEBUG oslo_vmware.api [None req-a58f60ab-ca30-4ed0-bbe9-1d319a9c18f8 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] Waiting for the task: (returnval){ [ 962.491179] env[65726]: value = "task-5116345" [ 962.491179] env[65726]: _type = "Task" [ 962.491179] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.501678] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebad8403-bbf5-4dab-8e9c-d4f7809b1979 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.509489] env[65726]: DEBUG oslo_vmware.api [None req-a58f60ab-ca30-4ed0-bbe9-1d319a9c18f8 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] Task: {'id': task-5116345, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.521025] env[65726]: DEBUG nova.compute.provider_tree [None req-f5777e4a-de17-4332-8f51-ad1e8cbe3f19 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Updating inventory in ProviderTree for provider 07c4692f-bdb4-4058-9173-ff9664830295 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 962.641884] env[65726]: DEBUG oslo_vmware.api [None req-71fc3c6d-ac1b-4ee5-b5e9-31bd39e153cf tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Task: {'id': task-5116342, 'name': PowerOnVM_Task, 'duration_secs': 1.283263} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 962.641884] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-71fc3c6d-ac1b-4ee5-b5e9-31bd39e153cf tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: cf950ec3-9914-4eb9-99db-048b3969bb21] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 962.646579] env[65726]: DEBUG nova.compute.manager [None req-71fc3c6d-ac1b-4ee5-b5e9-31bd39e153cf tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: cf950ec3-9914-4eb9-99db-048b3969bb21] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 962.650091] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07c9eb8c-33e8-475f-8959-d588f04b06b1 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.703258] env[65726]: DEBUG oslo_vmware.api [None req-ace31bce-a963-4a72-a87f-23a0d9efcfe0 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Task: {'id': task-5116344, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.714397] env[65726]: DEBUG nova.compute.manager [None req-83256413-d207-4cc8-b904-cfb2a5ea5628 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] [instance: f9a40d36-dda9-454c-8774-ed011f3f50ae] Starting instance... {{(pid=65726) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 962.734458] env[65726]: DEBUG oslo_concurrency.lockutils [None req-ed96d5ac-cdb1-4748-8154-6accb1ae5e65 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Acquiring lock "067c9332-fa2c-47fe-93ae-643de1fe77cf" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 962.734640] env[65726]: DEBUG oslo_concurrency.lockutils [None req-ed96d5ac-cdb1-4748-8154-6accb1ae5e65 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Lock "067c9332-fa2c-47fe-93ae-643de1fe77cf" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 963.002332] env[65726]: DEBUG oslo_vmware.api [None req-a58f60ab-ca30-4ed0-bbe9-1d319a9c18f8 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] Task: {'id': task-5116345, 'name': ReconfigVM_Task, 'duration_secs': 0.42757} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.002819] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-a58f60ab-ca30-4ed0-bbe9-1d319a9c18f8 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] [instance: 73f5b23e-4463-4068-8994-dd2752a2abc9] Reconfigured VM instance instance-0000004d to attach disk [datastore1] 73f5b23e-4463-4068-8994-dd2752a2abc9/73f5b23e-4463-4068-8994-dd2752a2abc9.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 963.003469] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1271dbfd-b112-4cec-8a7b-41f923bb223e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.012660] env[65726]: DEBUG oslo_vmware.api [None req-a58f60ab-ca30-4ed0-bbe9-1d319a9c18f8 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] Waiting for the task: (returnval){ [ 963.012660] env[65726]: value = "task-5116346" [ 963.012660] env[65726]: _type = "Task" [ 963.012660] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.027748] env[65726]: DEBUG oslo_vmware.api [None req-a58f60ab-ca30-4ed0-bbe9-1d319a9c18f8 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] Task: {'id': task-5116346, 'name': Rename_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.052264] env[65726]: ERROR nova.scheduler.client.report [None req-f5777e4a-de17-4332-8f51-ad1e8cbe3f19 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] [req-3d4c3575-f6ec-4b3f-b14b-2dc026e33732] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 07c4692f-bdb4-4058-9173-ff9664830295. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-3d4c3575-f6ec-4b3f-b14b-2dc026e33732"}]} [ 963.072549] env[65726]: DEBUG nova.scheduler.client.report [None req-f5777e4a-de17-4332-8f51-ad1e8cbe3f19 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Refreshing inventories for resource provider 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 963.091963] env[65726]: DEBUG nova.scheduler.client.report [None req-f5777e4a-de17-4332-8f51-ad1e8cbe3f19 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Updating ProviderTree inventory for provider 07c4692f-bdb4-4058-9173-ff9664830295 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 963.092228] env[65726]: DEBUG nova.compute.provider_tree [None req-f5777e4a-de17-4332-8f51-ad1e8cbe3f19 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Updating inventory in ProviderTree for provider 07c4692f-bdb4-4058-9173-ff9664830295 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 963.104434] env[65726]: DEBUG nova.scheduler.client.report [None req-f5777e4a-de17-4332-8f51-ad1e8cbe3f19 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Refreshing aggregate associations for resource provider 07c4692f-bdb4-4058-9173-ff9664830295, aggregates: None {{(pid=65726) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 963.124981] env[65726]: DEBUG nova.scheduler.client.report [None req-f5777e4a-de17-4332-8f51-ad1e8cbe3f19 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Refreshing trait associations for resource provider 07c4692f-bdb4-4058-9173-ff9664830295, traits: COMPUTE_IMAGE_TYPE_ISO,HW_ARCH_X86_64,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=65726) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 963.204131] env[65726]: DEBUG oslo_vmware.api [None req-ace31bce-a963-4a72-a87f-23a0d9efcfe0 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Task: {'id': task-5116344, 'name': CreateSnapshot_Task, 'duration_secs': 0.844236} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.204131] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-ace31bce-a963-4a72-a87f-23a0d9efcfe0 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] [instance: df46141c-b545-4e03-b3a3-fd9f5feda0d2] Created Snapshot of the VM instance {{(pid=65726) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 963.204710] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8cac363c-d593-42bd-b3b9-1c549fd247d4 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.245038] env[65726]: DEBUG nova.compute.manager [None req-ed96d5ac-cdb1-4748-8154-6accb1ae5e65 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: 067c9332-fa2c-47fe-93ae-643de1fe77cf] Starting instance... {{(pid=65726) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 963.261318] env[65726]: DEBUG oslo_concurrency.lockutils [None req-83256413-d207-4cc8-b904-cfb2a5ea5628 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 963.526174] env[65726]: DEBUG oslo_vmware.api [None req-a58f60ab-ca30-4ed0-bbe9-1d319a9c18f8 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] Task: {'id': task-5116346, 'name': Rename_Task, 'duration_secs': 0.21157} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.529254] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-a58f60ab-ca30-4ed0-bbe9-1d319a9c18f8 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] [instance: 73f5b23e-4463-4068-8994-dd2752a2abc9] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 963.531051] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-da92d9fb-0b9b-493e-8d5b-2e7da7af6a95 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.540611] env[65726]: DEBUG oslo_vmware.api [None req-a58f60ab-ca30-4ed0-bbe9-1d319a9c18f8 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] Waiting for the task: (returnval){ [ 963.540611] env[65726]: value = "task-5116347" [ 963.540611] env[65726]: _type = "Task" [ 963.540611] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.556173] env[65726]: DEBUG oslo_vmware.api [None req-a58f60ab-ca30-4ed0-bbe9-1d319a9c18f8 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] Task: {'id': task-5116347, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.614929] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db078705-daf6-4838-882c-4110b1041071 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.629759] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd0bc108-da68-4666-9f80-6e234f51837d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.662237] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eedb6b36-2de1-465c-85b2-17d89ae6c4cb {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.671162] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a17db456-3c86-4a52-a13f-6cb95826611d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.675869] env[65726]: INFO nova.compute.manager [None req-d7b5b063-4f6c-4203-a67b-66f369d80524 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: cf950ec3-9914-4eb9-99db-048b3969bb21] Unrescuing [ 963.676135] env[65726]: DEBUG oslo_concurrency.lockutils [None req-d7b5b063-4f6c-4203-a67b-66f369d80524 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Acquiring lock "refresh_cache-cf950ec3-9914-4eb9-99db-048b3969bb21" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 963.676319] env[65726]: DEBUG oslo_concurrency.lockutils [None req-d7b5b063-4f6c-4203-a67b-66f369d80524 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Acquired lock "refresh_cache-cf950ec3-9914-4eb9-99db-048b3969bb21" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 963.676501] env[65726]: DEBUG nova.network.neutron [None req-d7b5b063-4f6c-4203-a67b-66f369d80524 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: cf950ec3-9914-4eb9-99db-048b3969bb21] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 963.693026] env[65726]: DEBUG nova.compute.provider_tree [None req-f5777e4a-de17-4332-8f51-ad1e8cbe3f19 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Updating inventory in ProviderTree for provider 07c4692f-bdb4-4058-9173-ff9664830295 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 963.731491] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-ace31bce-a963-4a72-a87f-23a0d9efcfe0 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] [instance: df46141c-b545-4e03-b3a3-fd9f5feda0d2] Creating linked-clone VM from snapshot {{(pid=65726) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 963.731858] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-65a10991-c857-4f97-b598-c5493d5bdbdc {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.741832] env[65726]: DEBUG oslo_vmware.api [None req-ace31bce-a963-4a72-a87f-23a0d9efcfe0 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Waiting for the task: (returnval){ [ 963.741832] env[65726]: value = "task-5116348" [ 963.741832] env[65726]: _type = "Task" [ 963.741832] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.756242] env[65726]: DEBUG oslo_vmware.api [None req-ace31bce-a963-4a72-a87f-23a0d9efcfe0 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Task: {'id': task-5116348, 'name': CloneVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.778536] env[65726]: DEBUG oslo_concurrency.lockutils [None req-ed96d5ac-cdb1-4748-8154-6accb1ae5e65 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 964.055304] env[65726]: DEBUG oslo_vmware.api [None req-a58f60ab-ca30-4ed0-bbe9-1d319a9c18f8 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] Task: {'id': task-5116347, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.241873] env[65726]: WARNING openstack [None req-d7b5b063-4f6c-4203-a67b-66f369d80524 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 964.241873] env[65726]: WARNING openstack [None req-d7b5b063-4f6c-4203-a67b-66f369d80524 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 964.241873] env[65726]: DEBUG nova.scheduler.client.report [None req-f5777e4a-de17-4332-8f51-ad1e8cbe3f19 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Updated inventory for provider 07c4692f-bdb4-4058-9173-ff9664830295 with generation 113 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 964.241873] env[65726]: DEBUG nova.compute.provider_tree [None req-f5777e4a-de17-4332-8f51-ad1e8cbe3f19 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Updating resource provider 07c4692f-bdb4-4058-9173-ff9664830295 generation from 113 to 114 during operation: update_inventory {{(pid=65726) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 964.241873] env[65726]: DEBUG nova.compute.provider_tree [None req-f5777e4a-de17-4332-8f51-ad1e8cbe3f19 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Updating inventory in ProviderTree for provider 07c4692f-bdb4-4058-9173-ff9664830295 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 964.259034] env[65726]: DEBUG oslo_vmware.api [None req-ace31bce-a963-4a72-a87f-23a0d9efcfe0 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Task: {'id': task-5116348, 'name': CloneVM_Task} progress is 94%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.373856] env[65726]: WARNING openstack [None req-d7b5b063-4f6c-4203-a67b-66f369d80524 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 964.374608] env[65726]: WARNING openstack [None req-d7b5b063-4f6c-4203-a67b-66f369d80524 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 964.452516] env[65726]: WARNING openstack [None req-d7b5b063-4f6c-4203-a67b-66f369d80524 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 964.453029] env[65726]: WARNING openstack [None req-d7b5b063-4f6c-4203-a67b-66f369d80524 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 964.553686] env[65726]: DEBUG oslo_vmware.api [None req-a58f60ab-ca30-4ed0-bbe9-1d319a9c18f8 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] Task: {'id': task-5116347, 'name': PowerOnVM_Task, 'duration_secs': 0.800098} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 964.553964] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-a58f60ab-ca30-4ed0-bbe9-1d319a9c18f8 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] [instance: 73f5b23e-4463-4068-8994-dd2752a2abc9] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 964.554192] env[65726]: INFO nova.compute.manager [None req-a58f60ab-ca30-4ed0-bbe9-1d319a9c18f8 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] [instance: 73f5b23e-4463-4068-8994-dd2752a2abc9] Took 6.48 seconds to spawn the instance on the hypervisor. [ 964.554406] env[65726]: DEBUG nova.compute.manager [None req-a58f60ab-ca30-4ed0-bbe9-1d319a9c18f8 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] [instance: 73f5b23e-4463-4068-8994-dd2752a2abc9] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 964.555273] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90321dec-de3c-4ec7-9e2c-2353558d904a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.567389] env[65726]: DEBUG nova.network.neutron [None req-d7b5b063-4f6c-4203-a67b-66f369d80524 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: cf950ec3-9914-4eb9-99db-048b3969bb21] Updating instance_info_cache with network_info: [{"id": "58accf41-587f-4367-854e-fc571a6a3424", "address": "fa:16:3e:83:9d:0c", "network": {"id": "9d5ad167-99fb-4208-ab30-7053f35774e4", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-2082516762-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.132", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a8362295a62f4b51bae719a7ef5a4656", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dabbac20-1723-40ad-9da0-e53b28073651", "external-id": "nsx-vlan-transportzone-790", "segmentation_id": 790, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap58accf41-58", "ovs_interfaceid": "58accf41-587f-4367-854e-fc571a6a3424", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 964.600609] env[65726]: DEBUG oslo_vmware.rw_handles [None req-da525c49-2593-4b16-8428-bf88b8e252b1 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/527908c0-5e5e-60f3-9262-0e56a59cf070/disk-0.vmdk. {{(pid=65726) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 964.601678] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fbee4aa-0483-4bd4-af84-73318336c022 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.609466] env[65726]: DEBUG oslo_vmware.rw_handles [None req-da525c49-2593-4b16-8428-bf88b8e252b1 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/527908c0-5e5e-60f3-9262-0e56a59cf070/disk-0.vmdk is in state: ready. {{(pid=65726) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 964.609647] env[65726]: ERROR oslo_vmware.rw_handles [None req-da525c49-2593-4b16-8428-bf88b8e252b1 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/527908c0-5e5e-60f3-9262-0e56a59cf070/disk-0.vmdk due to incomplete transfer. [ 964.609890] env[65726]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-0f2ed11d-53af-45b1-82f8-bd1406c644dc {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.618905] env[65726]: DEBUG oslo_vmware.rw_handles [None req-da525c49-2593-4b16-8428-bf88b8e252b1 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/527908c0-5e5e-60f3-9262-0e56a59cf070/disk-0.vmdk. {{(pid=65726) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 964.619149] env[65726]: DEBUG nova.virt.vmwareapi.images [None req-da525c49-2593-4b16-8428-bf88b8e252b1 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] [instance: 815ee20b-cb07-4514-bd43-1b74972508a5] Uploaded image eb5566e4-ef83-4dc3-89c5-bc440d53be6b to the Glance image server {{(pid=65726) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 964.621904] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-da525c49-2593-4b16-8428-bf88b8e252b1 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] [instance: 815ee20b-cb07-4514-bd43-1b74972508a5] Destroying the VM {{(pid=65726) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 964.622789] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-66ef7a89-25b9-41f7-98ff-40ea1344ed74 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.633183] env[65726]: DEBUG oslo_vmware.api [None req-da525c49-2593-4b16-8428-bf88b8e252b1 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Waiting for the task: (returnval){ [ 964.633183] env[65726]: value = "task-5116349" [ 964.633183] env[65726]: _type = "Task" [ 964.633183] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 964.647190] env[65726]: DEBUG oslo_vmware.api [None req-da525c49-2593-4b16-8428-bf88b8e252b1 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Task: {'id': task-5116349, 'name': Destroy_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.744411] env[65726]: DEBUG oslo_concurrency.lockutils [None req-f5777e4a-de17-4332-8f51-ad1e8cbe3f19 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 4.430s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 964.748508] env[65726]: DEBUG oslo_concurrency.lockutils [None req-98994c54-47ac-4003-a1d5-85175217e75d tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 20.880s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 964.748620] env[65726]: DEBUG oslo_concurrency.lockutils [None req-98994c54-47ac-4003-a1d5-85175217e75d tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.001s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 964.750663] env[65726]: DEBUG oslo_concurrency.lockutils [None req-242f797f-8330-4ee8-bd46-f537d017011f tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 18.638s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 964.750991] env[65726]: DEBUG nova.objects.instance [None req-242f797f-8330-4ee8-bd46-f537d017011f tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Lazy-loading 'resources' on Instance uuid bc2c12e0-0d06-432f-b42f-be468e3b6ee4 {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 964.764024] env[65726]: DEBUG oslo_vmware.api [None req-ace31bce-a963-4a72-a87f-23a0d9efcfe0 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Task: {'id': task-5116348, 'name': CloneVM_Task} progress is 94%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.771266] env[65726]: INFO nova.scheduler.client.report [None req-f5777e4a-de17-4332-8f51-ad1e8cbe3f19 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Deleted allocations for instance 0e064341-4e4a-407b-8c26-3eb04b409029 [ 964.780365] env[65726]: INFO nova.scheduler.client.report [None req-98994c54-47ac-4003-a1d5-85175217e75d tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Deleted allocations for instance 608cc0ea-3f6b-4b4f-83c1-01aa50999d63 [ 965.076915] env[65726]: DEBUG oslo_concurrency.lockutils [None req-d7b5b063-4f6c-4203-a67b-66f369d80524 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Releasing lock "refresh_cache-cf950ec3-9914-4eb9-99db-048b3969bb21" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 965.077664] env[65726]: DEBUG nova.objects.instance [None req-d7b5b063-4f6c-4203-a67b-66f369d80524 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Lazy-loading 'flavor' on Instance uuid cf950ec3-9914-4eb9-99db-048b3969bb21 {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 965.083629] env[65726]: INFO nova.compute.manager [None req-a58f60ab-ca30-4ed0-bbe9-1d319a9c18f8 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] [instance: 73f5b23e-4463-4068-8994-dd2752a2abc9] Took 32.88 seconds to build instance. [ 965.151036] env[65726]: DEBUG oslo_vmware.api [None req-da525c49-2593-4b16-8428-bf88b8e252b1 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Task: {'id': task-5116349, 'name': Destroy_Task} progress is 100%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.266172] env[65726]: DEBUG oslo_vmware.api [None req-ace31bce-a963-4a72-a87f-23a0d9efcfe0 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Task: {'id': task-5116348, 'name': CloneVM_Task} progress is 100%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.281406] env[65726]: DEBUG oslo_concurrency.lockutils [None req-f5777e4a-de17-4332-8f51-ad1e8cbe3f19 tempest-MigrationsAdminTest-1257366526 tempest-MigrationsAdminTest-1257366526-project-member] Lock "0e064341-4e4a-407b-8c26-3eb04b409029" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 27.605s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 965.290453] env[65726]: DEBUG oslo_concurrency.lockutils [None req-98994c54-47ac-4003-a1d5-85175217e75d tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Lock "608cc0ea-3f6b-4b4f-83c1-01aa50999d63" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 25.186s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 965.584799] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a58f60ab-ca30-4ed0-bbe9-1d319a9c18f8 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] Lock "73f5b23e-4463-4068-8994-dd2752a2abc9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 34.405s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 965.587613] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-195a74b3-e39c-423a-a50a-6596d561ba69 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.642156] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-d7b5b063-4f6c-4203-a67b-66f369d80524 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: cf950ec3-9914-4eb9-99db-048b3969bb21] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 965.648847] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-25220896-8b8a-426e-8562-e34c5f480a80 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.658184] env[65726]: INFO nova.compute.manager [None req-de3a85d2-cc71-4c75-a1b0-fe6db6e57022 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] [instance: 73f5b23e-4463-4068-8994-dd2752a2abc9] Rebuilding instance [ 965.660589] env[65726]: DEBUG oslo_vmware.api [None req-da525c49-2593-4b16-8428-bf88b8e252b1 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Task: {'id': task-5116349, 'name': Destroy_Task} progress is 100%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.663574] env[65726]: DEBUG oslo_vmware.api [None req-d7b5b063-4f6c-4203-a67b-66f369d80524 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Waiting for the task: (returnval){ [ 965.663574] env[65726]: value = "task-5116350" [ 965.663574] env[65726]: _type = "Task" [ 965.663574] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.675414] env[65726]: DEBUG oslo_vmware.api [None req-d7b5b063-4f6c-4203-a67b-66f369d80524 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Task: {'id': task-5116350, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.715260] env[65726]: DEBUG nova.compute.manager [None req-de3a85d2-cc71-4c75-a1b0-fe6db6e57022 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] [instance: 73f5b23e-4463-4068-8994-dd2752a2abc9] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 965.716425] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-372e3e15-ac30-415b-adff-d38d4aa90a18 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.762039] env[65726]: DEBUG oslo_vmware.api [None req-ace31bce-a963-4a72-a87f-23a0d9efcfe0 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Task: {'id': task-5116348, 'name': CloneVM_Task, 'duration_secs': 1.642748} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 965.762437] env[65726]: INFO nova.virt.vmwareapi.vmops [None req-ace31bce-a963-4a72-a87f-23a0d9efcfe0 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] [instance: df46141c-b545-4e03-b3a3-fd9f5feda0d2] Created linked-clone VM from snapshot [ 965.764123] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1722bd1e-cef0-47d7-98b4-bd32e9bced63 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.780204] env[65726]: DEBUG nova.virt.vmwareapi.images [None req-ace31bce-a963-4a72-a87f-23a0d9efcfe0 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] [instance: df46141c-b545-4e03-b3a3-fd9f5feda0d2] Uploading image e8e6eb08-0499-4c5d-a1b2-1c9ed8f7ff91 {{(pid=65726) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 965.800738] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-ace31bce-a963-4a72-a87f-23a0d9efcfe0 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] [instance: df46141c-b545-4e03-b3a3-fd9f5feda0d2] Destroying the VM {{(pid=65726) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 965.801395] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-43bcfa46-ec4c-41b4-a627-14327ad79c94 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.811452] env[65726]: DEBUG oslo_vmware.api [None req-ace31bce-a963-4a72-a87f-23a0d9efcfe0 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Waiting for the task: (returnval){ [ 965.811452] env[65726]: value = "task-5116351" [ 965.811452] env[65726]: _type = "Task" [ 965.811452] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.827702] env[65726]: DEBUG oslo_vmware.api [None req-ace31bce-a963-4a72-a87f-23a0d9efcfe0 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Task: {'id': task-5116351, 'name': Destroy_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.850682] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dcbad227-8c3a-4d4d-8ca4-16618d82b068 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.864400] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76a3fe71-9946-4b98-82c2-5d8378c9df94 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.897057] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-888138b1-7779-4b5e-a108-57ba82163ece {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.908263] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16149221-21bf-4f42-b3a3-8e67460d3788 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.924540] env[65726]: DEBUG nova.compute.provider_tree [None req-242f797f-8330-4ee8-bd46-f537d017011f tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 966.153856] env[65726]: DEBUG oslo_vmware.api [None req-da525c49-2593-4b16-8428-bf88b8e252b1 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Task: {'id': task-5116349, 'name': Destroy_Task, 'duration_secs': 1.381074} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.154380] env[65726]: INFO nova.virt.vmwareapi.vm_util [None req-da525c49-2593-4b16-8428-bf88b8e252b1 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] [instance: 815ee20b-cb07-4514-bd43-1b74972508a5] Destroyed the VM [ 966.154686] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-da525c49-2593-4b16-8428-bf88b8e252b1 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] [instance: 815ee20b-cb07-4514-bd43-1b74972508a5] Deleting Snapshot of the VM instance {{(pid=65726) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 966.155036] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-a787539a-ab89-48b5-8dc8-32891661d69c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.167739] env[65726]: DEBUG oslo_vmware.api [None req-da525c49-2593-4b16-8428-bf88b8e252b1 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Waiting for the task: (returnval){ [ 966.167739] env[65726]: value = "task-5116352" [ 966.167739] env[65726]: _type = "Task" [ 966.167739] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.184618] env[65726]: DEBUG oslo_vmware.api [None req-d7b5b063-4f6c-4203-a67b-66f369d80524 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Task: {'id': task-5116350, 'name': PowerOffVM_Task, 'duration_secs': 0.377735} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.190605] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-d7b5b063-4f6c-4203-a67b-66f369d80524 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: cf950ec3-9914-4eb9-99db-048b3969bb21] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 966.197484] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-d7b5b063-4f6c-4203-a67b-66f369d80524 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: cf950ec3-9914-4eb9-99db-048b3969bb21] Reconfiguring VM instance instance-0000003e to detach disk 2002 {{(pid=65726) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 966.199084] env[65726]: DEBUG oslo_vmware.api [None req-da525c49-2593-4b16-8428-bf88b8e252b1 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Task: {'id': task-5116352, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.199829] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4c9bcddf-9230-44d9-9198-e4981ebe7c91 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.234050] env[65726]: DEBUG oslo_vmware.api [None req-d7b5b063-4f6c-4203-a67b-66f369d80524 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Waiting for the task: (returnval){ [ 966.234050] env[65726]: value = "task-5116353" [ 966.234050] env[65726]: _type = "Task" [ 966.234050] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.244835] env[65726]: DEBUG oslo_vmware.api [None req-d7b5b063-4f6c-4203-a67b-66f369d80524 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Task: {'id': task-5116353, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.325929] env[65726]: DEBUG oslo_vmware.api [None req-ace31bce-a963-4a72-a87f-23a0d9efcfe0 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Task: {'id': task-5116351, 'name': Destroy_Task} progress is 100%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.428121] env[65726]: DEBUG nova.scheduler.client.report [None req-242f797f-8330-4ee8-bd46-f537d017011f tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 966.687641] env[65726]: DEBUG oslo_vmware.api [None req-da525c49-2593-4b16-8428-bf88b8e252b1 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Task: {'id': task-5116352, 'name': RemoveSnapshot_Task} progress is 38%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.739242] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-de3a85d2-cc71-4c75-a1b0-fe6db6e57022 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] [instance: 73f5b23e-4463-4068-8994-dd2752a2abc9] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 966.743835] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-33f7fbd4-8e5f-49f5-93d4-84b6b32445fd {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.745844] env[65726]: DEBUG oslo_vmware.api [None req-d7b5b063-4f6c-4203-a67b-66f369d80524 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Task: {'id': task-5116353, 'name': ReconfigVM_Task, 'duration_secs': 0.349267} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.746193] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-d7b5b063-4f6c-4203-a67b-66f369d80524 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: cf950ec3-9914-4eb9-99db-048b3969bb21] Reconfigured VM instance instance-0000003e to detach disk 2002 {{(pid=65726) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 966.746414] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-d7b5b063-4f6c-4203-a67b-66f369d80524 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: cf950ec3-9914-4eb9-99db-048b3969bb21] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 966.747316] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9ab1b9e2-da2a-47a3-8d5d-b70c3ccdf9b5 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.755394] env[65726]: DEBUG oslo_vmware.api [None req-de3a85d2-cc71-4c75-a1b0-fe6db6e57022 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] Waiting for the task: (returnval){ [ 966.755394] env[65726]: value = "task-5116354" [ 966.755394] env[65726]: _type = "Task" [ 966.755394] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.758807] env[65726]: DEBUG oslo_vmware.api [None req-d7b5b063-4f6c-4203-a67b-66f369d80524 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Waiting for the task: (returnval){ [ 966.758807] env[65726]: value = "task-5116355" [ 966.758807] env[65726]: _type = "Task" [ 966.758807] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.775397] env[65726]: DEBUG oslo_vmware.api [None req-d7b5b063-4f6c-4203-a67b-66f369d80524 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Task: {'id': task-5116355, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.779862] env[65726]: DEBUG oslo_vmware.api [None req-de3a85d2-cc71-4c75-a1b0-fe6db6e57022 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] Task: {'id': task-5116354, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.826212] env[65726]: DEBUG oslo_vmware.api [None req-ace31bce-a963-4a72-a87f-23a0d9efcfe0 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Task: {'id': task-5116351, 'name': Destroy_Task, 'duration_secs': 0.518964} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.826212] env[65726]: INFO nova.virt.vmwareapi.vm_util [None req-ace31bce-a963-4a72-a87f-23a0d9efcfe0 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] [instance: df46141c-b545-4e03-b3a3-fd9f5feda0d2] Destroyed the VM [ 966.826212] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-ace31bce-a963-4a72-a87f-23a0d9efcfe0 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] [instance: df46141c-b545-4e03-b3a3-fd9f5feda0d2] Deleting Snapshot of the VM instance {{(pid=65726) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 966.826212] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-d837c917-5eba-4c6a-a62b-40e568bfe34c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.835749] env[65726]: DEBUG oslo_vmware.api [None req-ace31bce-a963-4a72-a87f-23a0d9efcfe0 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Waiting for the task: (returnval){ [ 966.835749] env[65726]: value = "task-5116356" [ 966.835749] env[65726]: _type = "Task" [ 966.835749] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.845618] env[65726]: DEBUG oslo_vmware.api [None req-ace31bce-a963-4a72-a87f-23a0d9efcfe0 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Task: {'id': task-5116356, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.934720] env[65726]: DEBUG oslo_concurrency.lockutils [None req-242f797f-8330-4ee8-bd46-f537d017011f tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.184s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 966.939020] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9f65b40f-615d-4bae-ac4c-c122fff748fd tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.900s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 966.940490] env[65726]: INFO nova.compute.claims [None req-9f65b40f-615d-4bae-ac4c-c122fff748fd tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] [instance: aa6ce489-c62f-4481-87b7-e74242aeb8ca] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 966.964507] env[65726]: INFO nova.scheduler.client.report [None req-242f797f-8330-4ee8-bd46-f537d017011f tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Deleted allocations for instance bc2c12e0-0d06-432f-b42f-be468e3b6ee4 [ 967.190197] env[65726]: DEBUG oslo_vmware.api [None req-da525c49-2593-4b16-8428-bf88b8e252b1 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Task: {'id': task-5116352, 'name': RemoveSnapshot_Task, 'duration_secs': 0.565788} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 967.190197] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-da525c49-2593-4b16-8428-bf88b8e252b1 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] [instance: 815ee20b-cb07-4514-bd43-1b74972508a5] Deleted Snapshot of the VM instance {{(pid=65726) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 967.190197] env[65726]: INFO nova.compute.manager [None req-da525c49-2593-4b16-8428-bf88b8e252b1 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] [instance: 815ee20b-cb07-4514-bd43-1b74972508a5] Took 15.37 seconds to snapshot the instance on the hypervisor. [ 967.274740] env[65726]: DEBUG oslo_vmware.api [None req-de3a85d2-cc71-4c75-a1b0-fe6db6e57022 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] Task: {'id': task-5116354, 'name': PowerOffVM_Task, 'duration_secs': 0.138772} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 967.275790] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-de3a85d2-cc71-4c75-a1b0-fe6db6e57022 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] [instance: 73f5b23e-4463-4068-8994-dd2752a2abc9] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 967.276083] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-de3a85d2-cc71-4c75-a1b0-fe6db6e57022 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] [instance: 73f5b23e-4463-4068-8994-dd2752a2abc9] Destroying instance {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 967.277115] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7656958-4324-4c2c-aa7d-dc48fcc68f61 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.289067] env[65726]: DEBUG oslo_vmware.api [None req-d7b5b063-4f6c-4203-a67b-66f369d80524 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Task: {'id': task-5116355, 'name': PowerOnVM_Task, 'duration_secs': 0.525347} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 967.289067] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-d7b5b063-4f6c-4203-a67b-66f369d80524 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: cf950ec3-9914-4eb9-99db-048b3969bb21] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 967.289067] env[65726]: DEBUG nova.compute.manager [None req-d7b5b063-4f6c-4203-a67b-66f369d80524 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: cf950ec3-9914-4eb9-99db-048b3969bb21] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 967.289067] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8d00680-d6d4-4e1e-9e0f-ca6d1ad4a623 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.292130] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-de3a85d2-cc71-4c75-a1b0-fe6db6e57022 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] [instance: 73f5b23e-4463-4068-8994-dd2752a2abc9] Unregistering the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 967.296168] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-96934985-3999-4151-a082-341c85f277f8 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.337821] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-de3a85d2-cc71-4c75-a1b0-fe6db6e57022 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] [instance: 73f5b23e-4463-4068-8994-dd2752a2abc9] Unregistered the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 967.338221] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-de3a85d2-cc71-4c75-a1b0-fe6db6e57022 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] [instance: 73f5b23e-4463-4068-8994-dd2752a2abc9] Deleting contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 967.338330] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-de3a85d2-cc71-4c75-a1b0-fe6db6e57022 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] Deleting the datastore file [datastore1] 73f5b23e-4463-4068-8994-dd2752a2abc9 {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 967.344718] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a4f43782-6586-45cc-85a8-1bf959baf6a0 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.354255] env[65726]: DEBUG oslo_vmware.api [None req-ace31bce-a963-4a72-a87f-23a0d9efcfe0 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Task: {'id': task-5116356, 'name': RemoveSnapshot_Task} progress is 80%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.356376] env[65726]: DEBUG oslo_vmware.api [None req-de3a85d2-cc71-4c75-a1b0-fe6db6e57022 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] Waiting for the task: (returnval){ [ 967.356376] env[65726]: value = "task-5116358" [ 967.356376] env[65726]: _type = "Task" [ 967.356376] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 967.365556] env[65726]: DEBUG oslo_vmware.api [None req-de3a85d2-cc71-4c75-a1b0-fe6db6e57022 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] Task: {'id': task-5116358, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.410190] env[65726]: DEBUG oslo_concurrency.lockutils [None req-775eaf93-375d-4ff3-9c57-60f5cc32d4ea tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Acquiring lock "a56be6fa-6027-46ee-9aa8-332fe4222d68" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 967.410395] env[65726]: DEBUG oslo_concurrency.lockutils [None req-775eaf93-375d-4ff3-9c57-60f5cc32d4ea tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Lock "a56be6fa-6027-46ee-9aa8-332fe4222d68" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 967.474142] env[65726]: DEBUG oslo_concurrency.lockutils [None req-242f797f-8330-4ee8-bd46-f537d017011f tempest-SecurityGroupsTestJSON-545899341 tempest-SecurityGroupsTestJSON-545899341-project-member] Lock "bc2c12e0-0d06-432f-b42f-be468e3b6ee4" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 25.301s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 967.850434] env[65726]: DEBUG oslo_vmware.api [None req-ace31bce-a963-4a72-a87f-23a0d9efcfe0 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Task: {'id': task-5116356, 'name': RemoveSnapshot_Task} progress is 80%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.869337] env[65726]: DEBUG oslo_vmware.api [None req-de3a85d2-cc71-4c75-a1b0-fe6db6e57022 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] Task: {'id': task-5116358, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.109628} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 967.869709] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-de3a85d2-cc71-4c75-a1b0-fe6db6e57022 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] Deleted the datastore file {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 967.869941] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-de3a85d2-cc71-4c75-a1b0-fe6db6e57022 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] [instance: 73f5b23e-4463-4068-8994-dd2752a2abc9] Deleted contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 967.870141] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-de3a85d2-cc71-4c75-a1b0-fe6db6e57022 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] [instance: 73f5b23e-4463-4068-8994-dd2752a2abc9] Instance destroyed {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 967.916186] env[65726]: DEBUG nova.compute.manager [None req-775eaf93-375d-4ff3-9c57-60f5cc32d4ea tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: a56be6fa-6027-46ee-9aa8-332fe4222d68] Starting instance... {{(pid=65726) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 968.340538] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-931aadd9-3ebb-4600-9550-6dc9f8771669 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.354803] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6319783c-c8a8-48a9-a830-814e8c596f1c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.359403] env[65726]: DEBUG oslo_vmware.api [None req-ace31bce-a963-4a72-a87f-23a0d9efcfe0 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Task: {'id': task-5116356, 'name': RemoveSnapshot_Task, 'duration_secs': 1.445019} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 968.359693] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-ace31bce-a963-4a72-a87f-23a0d9efcfe0 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] [instance: df46141c-b545-4e03-b3a3-fd9f5feda0d2] Deleted Snapshot of the VM instance {{(pid=65726) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 968.398137] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2f49dd7-8dad-48e2-a02d-ed92570831c3 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.406972] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4163c59a-838b-4bdc-9c01-2596cdce97fe {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.426462] env[65726]: DEBUG nova.compute.provider_tree [None req-9f65b40f-615d-4bae-ac4c-c122fff748fd tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 968.442808] env[65726]: DEBUG oslo_concurrency.lockutils [None req-775eaf93-375d-4ff3-9c57-60f5cc32d4ea tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 968.867839] env[65726]: WARNING nova.compute.manager [None req-ace31bce-a963-4a72-a87f-23a0d9efcfe0 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] [instance: df46141c-b545-4e03-b3a3-fd9f5feda0d2] Image not found during snapshot: nova.exception.ImageNotFound: Image e8e6eb08-0499-4c5d-a1b2-1c9ed8f7ff91 could not be found. [ 968.903628] env[65726]: DEBUG nova.compute.manager [req-10cea4aa-91dd-4704-815e-ccc59659ff50 req-097d62f3-1290-4ab3-8d20-8381ec43f400 service nova] [instance: cf950ec3-9914-4eb9-99db-048b3969bb21] Received event network-changed-58accf41-587f-4367-854e-fc571a6a3424 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 968.903875] env[65726]: DEBUG nova.compute.manager [req-10cea4aa-91dd-4704-815e-ccc59659ff50 req-097d62f3-1290-4ab3-8d20-8381ec43f400 service nova] [instance: cf950ec3-9914-4eb9-99db-048b3969bb21] Refreshing instance network info cache due to event network-changed-58accf41-587f-4367-854e-fc571a6a3424. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 968.904147] env[65726]: DEBUG oslo_concurrency.lockutils [req-10cea4aa-91dd-4704-815e-ccc59659ff50 req-097d62f3-1290-4ab3-8d20-8381ec43f400 service nova] Acquiring lock "refresh_cache-cf950ec3-9914-4eb9-99db-048b3969bb21" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 968.904326] env[65726]: DEBUG oslo_concurrency.lockutils [req-10cea4aa-91dd-4704-815e-ccc59659ff50 req-097d62f3-1290-4ab3-8d20-8381ec43f400 service nova] Acquired lock "refresh_cache-cf950ec3-9914-4eb9-99db-048b3969bb21" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 968.904518] env[65726]: DEBUG nova.network.neutron [req-10cea4aa-91dd-4704-815e-ccc59659ff50 req-097d62f3-1290-4ab3-8d20-8381ec43f400 service nova] [instance: cf950ec3-9914-4eb9-99db-048b3969bb21] Refreshing network info cache for port 58accf41-587f-4367-854e-fc571a6a3424 {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 968.929420] env[65726]: DEBUG nova.virt.hardware [None req-de3a85d2-cc71-4c75-a1b0-fe6db6e57022 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T19:28:34Z,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1539f06fa3534e90acbc3a60c4b8bd3f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T19:28:35Z,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 968.929664] env[65726]: DEBUG nova.virt.hardware [None req-de3a85d2-cc71-4c75-a1b0-fe6db6e57022 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 968.929814] env[65726]: DEBUG nova.virt.hardware [None req-de3a85d2-cc71-4c75-a1b0-fe6db6e57022 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 968.929990] env[65726]: DEBUG nova.virt.hardware [None req-de3a85d2-cc71-4c75-a1b0-fe6db6e57022 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 968.930309] env[65726]: DEBUG nova.virt.hardware [None req-de3a85d2-cc71-4c75-a1b0-fe6db6e57022 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 968.930506] env[65726]: DEBUG nova.virt.hardware [None req-de3a85d2-cc71-4c75-a1b0-fe6db6e57022 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 968.930767] env[65726]: DEBUG nova.virt.hardware [None req-de3a85d2-cc71-4c75-a1b0-fe6db6e57022 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 968.930938] env[65726]: DEBUG nova.virt.hardware [None req-de3a85d2-cc71-4c75-a1b0-fe6db6e57022 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 968.931132] env[65726]: DEBUG nova.virt.hardware [None req-de3a85d2-cc71-4c75-a1b0-fe6db6e57022 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 968.931366] env[65726]: DEBUG nova.virt.hardware [None req-de3a85d2-cc71-4c75-a1b0-fe6db6e57022 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 968.931567] env[65726]: DEBUG nova.virt.hardware [None req-de3a85d2-cc71-4c75-a1b0-fe6db6e57022 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 968.932678] env[65726]: DEBUG nova.scheduler.client.report [None req-9f65b40f-615d-4bae-ac4c-c122fff748fd tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 968.937281] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db717fc6-504e-47be-9e98-98c72ba4c0b9 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.952632] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-513dedcb-3895-4ba6-bdfd-fe2fd32467fa {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.967843] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-de3a85d2-cc71-4c75-a1b0-fe6db6e57022 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] [instance: 73f5b23e-4463-4068-8994-dd2752a2abc9] Instance VIF info [] {{(pid=65726) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 968.972952] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-de3a85d2-cc71-4c75-a1b0-fe6db6e57022 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 968.973461] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 73f5b23e-4463-4068-8994-dd2752a2abc9] Creating VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 968.973732] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8c1be725-558e-4a38-8b90-022dbbd3cc84 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.999864] env[65726]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 968.999864] env[65726]: value = "task-5116359" [ 968.999864] env[65726]: _type = "Task" [ 968.999864] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.010849] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116359, 'name': CreateVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.408568] env[65726]: WARNING openstack [req-10cea4aa-91dd-4704-815e-ccc59659ff50 req-097d62f3-1290-4ab3-8d20-8381ec43f400 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 969.409389] env[65726]: WARNING openstack [req-10cea4aa-91dd-4704-815e-ccc59659ff50 req-097d62f3-1290-4ab3-8d20-8381ec43f400 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 969.446870] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9f65b40f-615d-4bae-ac4c-c122fff748fd tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.508s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 969.447424] env[65726]: DEBUG nova.compute.manager [None req-9f65b40f-615d-4bae-ac4c-c122fff748fd tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] [instance: aa6ce489-c62f-4481-87b7-e74242aeb8ca] Start building networks asynchronously for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 969.450374] env[65726]: DEBUG oslo_concurrency.lockutils [None req-d903724e-08f6-4863-a079-5fadb490f6c5 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.775s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 969.456365] env[65726]: INFO nova.compute.claims [None req-d903724e-08f6-4863-a079-5fadb490f6c5 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 54e85161-5f63-405f-83f8-490c70645a3b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 969.513340] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116359, 'name': CreateVM_Task} progress is 99%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.776756] env[65726]: WARNING openstack [req-10cea4aa-91dd-4704-815e-ccc59659ff50 req-097d62f3-1290-4ab3-8d20-8381ec43f400 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 969.777252] env[65726]: WARNING openstack [req-10cea4aa-91dd-4704-815e-ccc59659ff50 req-097d62f3-1290-4ab3-8d20-8381ec43f400 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 969.926303] env[65726]: WARNING openstack [req-10cea4aa-91dd-4704-815e-ccc59659ff50 req-097d62f3-1290-4ab3-8d20-8381ec43f400 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 969.926765] env[65726]: WARNING openstack [req-10cea4aa-91dd-4704-815e-ccc59659ff50 req-097d62f3-1290-4ab3-8d20-8381ec43f400 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 969.946815] env[65726]: DEBUG oslo_concurrency.lockutils [None req-8940e544-9cc2-4a18-9dd0-e7f396520ac6 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Acquiring lock "2fe6fc5a-f5c7-4f8e-96df-4e621a252f04" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 969.947491] env[65726]: DEBUG oslo_concurrency.lockutils [None req-8940e544-9cc2-4a18-9dd0-e7f396520ac6 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Lock "2fe6fc5a-f5c7-4f8e-96df-4e621a252f04" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.001s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 969.947695] env[65726]: INFO nova.compute.manager [None req-8940e544-9cc2-4a18-9dd0-e7f396520ac6 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04] Shelving [ 969.961939] env[65726]: DEBUG nova.compute.utils [None req-9f65b40f-615d-4bae-ac4c-c122fff748fd tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Using /dev/sd instead of None {{(pid=65726) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 969.967036] env[65726]: DEBUG nova.compute.manager [None req-9f65b40f-615d-4bae-ac4c-c122fff748fd tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] [instance: aa6ce489-c62f-4481-87b7-e74242aeb8ca] Allocating IP information in the background. {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 969.967036] env[65726]: DEBUG nova.network.neutron [None req-9f65b40f-615d-4bae-ac4c-c122fff748fd tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] [instance: aa6ce489-c62f-4481-87b7-e74242aeb8ca] allocate_for_instance() {{(pid=65726) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 969.967036] env[65726]: WARNING neutronclient.v2_0.client [None req-9f65b40f-615d-4bae-ac4c-c122fff748fd tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 969.967036] env[65726]: WARNING neutronclient.v2_0.client [None req-9f65b40f-615d-4bae-ac4c-c122fff748fd tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 969.967036] env[65726]: WARNING openstack [None req-9f65b40f-615d-4bae-ac4c-c122fff748fd tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 969.967655] env[65726]: WARNING openstack [None req-9f65b40f-615d-4bae-ac4c-c122fff748fd tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 970.013673] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116359, 'name': CreateVM_Task} progress is 99%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.066936] env[65726]: DEBUG nova.policy [None req-9f65b40f-615d-4bae-ac4c-c122fff748fd tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'cc5f9964e96a460fadfec7a9fe8e861b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4ffd45f4a7a041199a4fc7f69f5e7e9b', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65726) authorize /opt/stack/nova/nova/policy.py:192}} [ 970.106924] env[65726]: DEBUG nova.network.neutron [req-10cea4aa-91dd-4704-815e-ccc59659ff50 req-097d62f3-1290-4ab3-8d20-8381ec43f400 service nova] [instance: cf950ec3-9914-4eb9-99db-048b3969bb21] Updated VIF entry in instance network info cache for port 58accf41-587f-4367-854e-fc571a6a3424. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 970.106924] env[65726]: DEBUG nova.network.neutron [req-10cea4aa-91dd-4704-815e-ccc59659ff50 req-097d62f3-1290-4ab3-8d20-8381ec43f400 service nova] [instance: cf950ec3-9914-4eb9-99db-048b3969bb21] Updating instance_info_cache with network_info: [{"id": "58accf41-587f-4367-854e-fc571a6a3424", "address": "fa:16:3e:83:9d:0c", "network": {"id": "9d5ad167-99fb-4208-ab30-7053f35774e4", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-2082516762-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.132", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a8362295a62f4b51bae719a7ef5a4656", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dabbac20-1723-40ad-9da0-e53b28073651", "external-id": "nsx-vlan-transportzone-790", "segmentation_id": 790, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap58accf41-58", "ovs_interfaceid": "58accf41-587f-4367-854e-fc571a6a3424", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 970.478752] env[65726]: DEBUG nova.compute.manager [None req-9f65b40f-615d-4bae-ac4c-c122fff748fd tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] [instance: aa6ce489-c62f-4481-87b7-e74242aeb8ca] Start building block device mappings for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 970.498173] env[65726]: DEBUG nova.network.neutron [None req-9f65b40f-615d-4bae-ac4c-c122fff748fd tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] [instance: aa6ce489-c62f-4481-87b7-e74242aeb8ca] Successfully created port: d5a805f5-b6a8-42ec-92a0-4a652b51c1e6 {{(pid=65726) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 970.500844] env[65726]: DEBUG oslo_concurrency.lockutils [None req-4d042142-5ccf-4c86-bcb5-fb6e60309d3b tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Acquiring lock "df46141c-b545-4e03-b3a3-fd9f5feda0d2" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 970.504014] env[65726]: DEBUG oslo_concurrency.lockutils [None req-4d042142-5ccf-4c86-bcb5-fb6e60309d3b tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Lock "df46141c-b545-4e03-b3a3-fd9f5feda0d2" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 970.504014] env[65726]: DEBUG oslo_concurrency.lockutils [None req-4d042142-5ccf-4c86-bcb5-fb6e60309d3b tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Acquiring lock "df46141c-b545-4e03-b3a3-fd9f5feda0d2-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 970.504014] env[65726]: DEBUG oslo_concurrency.lockutils [None req-4d042142-5ccf-4c86-bcb5-fb6e60309d3b tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Lock "df46141c-b545-4e03-b3a3-fd9f5feda0d2-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 970.504014] env[65726]: DEBUG oslo_concurrency.lockutils [None req-4d042142-5ccf-4c86-bcb5-fb6e60309d3b tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Lock "df46141c-b545-4e03-b3a3-fd9f5feda0d2-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 970.508199] env[65726]: INFO nova.compute.manager [None req-4d042142-5ccf-4c86-bcb5-fb6e60309d3b tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] [instance: df46141c-b545-4e03-b3a3-fd9f5feda0d2] Terminating instance [ 970.521314] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116359, 'name': CreateVM_Task} progress is 99%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.610118] env[65726]: DEBUG oslo_concurrency.lockutils [req-10cea4aa-91dd-4704-815e-ccc59659ff50 req-097d62f3-1290-4ab3-8d20-8381ec43f400 service nova] Releasing lock "refresh_cache-cf950ec3-9914-4eb9-99db-048b3969bb21" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 970.618094] env[65726]: DEBUG oslo_service.periodic_task [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=65726) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 970.620194] env[65726]: DEBUG oslo_service.periodic_task [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=65726) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 970.748986] env[65726]: DEBUG oslo_concurrency.lockutils [None req-618baf85-0896-4fc2-b1ef-24618b3b2fb7 tempest-FloatingIPsAssociationNegativeTestJSON-737087291 tempest-FloatingIPsAssociationNegativeTestJSON-737087291-project-member] Acquiring lock "d8468efa-c2d3-4dce-ab89-fc077011e3d8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 970.749218] env[65726]: DEBUG oslo_concurrency.lockutils [None req-618baf85-0896-4fc2-b1ef-24618b3b2fb7 tempest-FloatingIPsAssociationNegativeTestJSON-737087291 tempest-FloatingIPsAssociationNegativeTestJSON-737087291-project-member] Lock "d8468efa-c2d3-4dce-ab89-fc077011e3d8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 970.960664] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-8940e544-9cc2-4a18-9dd0-e7f396520ac6 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 970.961544] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-75d46bd8-1c64-45ad-bdd5-c1dd5236d3fb {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.966414] env[65726]: DEBUG nova.compute.manager [req-a4623df5-2a6b-4ab4-81c7-75a5bdd51804 req-449729b7-8acc-401d-a792-50b40ba68a74 service nova] [instance: cf950ec3-9914-4eb9-99db-048b3969bb21] Received event network-changed-58accf41-587f-4367-854e-fc571a6a3424 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 970.966622] env[65726]: DEBUG nova.compute.manager [req-a4623df5-2a6b-4ab4-81c7-75a5bdd51804 req-449729b7-8acc-401d-a792-50b40ba68a74 service nova] [instance: cf950ec3-9914-4eb9-99db-048b3969bb21] Refreshing instance network info cache due to event network-changed-58accf41-587f-4367-854e-fc571a6a3424. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 970.966833] env[65726]: DEBUG oslo_concurrency.lockutils [req-a4623df5-2a6b-4ab4-81c7-75a5bdd51804 req-449729b7-8acc-401d-a792-50b40ba68a74 service nova] Acquiring lock "refresh_cache-cf950ec3-9914-4eb9-99db-048b3969bb21" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 970.966967] env[65726]: DEBUG oslo_concurrency.lockutils [req-a4623df5-2a6b-4ab4-81c7-75a5bdd51804 req-449729b7-8acc-401d-a792-50b40ba68a74 service nova] Acquired lock "refresh_cache-cf950ec3-9914-4eb9-99db-048b3969bb21" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 970.967137] env[65726]: DEBUG nova.network.neutron [req-a4623df5-2a6b-4ab4-81c7-75a5bdd51804 req-449729b7-8acc-401d-a792-50b40ba68a74 service nova] [instance: cf950ec3-9914-4eb9-99db-048b3969bb21] Refreshing network info cache for port 58accf41-587f-4367-854e-fc571a6a3424 {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 970.976228] env[65726]: DEBUG oslo_vmware.api [None req-8940e544-9cc2-4a18-9dd0-e7f396520ac6 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Waiting for the task: (returnval){ [ 970.976228] env[65726]: value = "task-5116360" [ 970.976228] env[65726]: _type = "Task" [ 970.976228] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.978114] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae38ced6-2115-4a14-bf5c-a16f81693b79 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.000219] env[65726]: DEBUG oslo_vmware.api [None req-8940e544-9cc2-4a18-9dd0-e7f396520ac6 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Task: {'id': task-5116360, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.001357] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f5c430e-7a9b-4fbe-bae5-7dae3165b7f1 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.016115] env[65726]: DEBUG nova.compute.manager [None req-4d042142-5ccf-4c86-bcb5-fb6e60309d3b tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] [instance: df46141c-b545-4e03-b3a3-fd9f5feda0d2] Start destroying the instance on the hypervisor. {{(pid=65726) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 971.016339] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-4d042142-5ccf-4c86-bcb5-fb6e60309d3b tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] [instance: df46141c-b545-4e03-b3a3-fd9f5feda0d2] Destroying instance {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 971.017031] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116359, 'name': CreateVM_Task} progress is 99%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.044590] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a27e352-ae7e-400c-8743-b3c3a980eee4 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.049686] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fdf1413-305f-4158-8c10-e220898f1c89 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.062026] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52d4f529-4a83-400e-a883-8d132f5f25d7 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.066137] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-4d042142-5ccf-4c86-bcb5-fb6e60309d3b tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] [instance: df46141c-b545-4e03-b3a3-fd9f5feda0d2] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 971.066661] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b01fcfbc-837d-47c0-b522-c501ddd8e5a3 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.081342] env[65726]: DEBUG nova.compute.provider_tree [None req-d903724e-08f6-4863-a079-5fadb490f6c5 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 971.084700] env[65726]: DEBUG oslo_vmware.api [None req-4d042142-5ccf-4c86-bcb5-fb6e60309d3b tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Waiting for the task: (returnval){ [ 971.084700] env[65726]: value = "task-5116361" [ 971.084700] env[65726]: _type = "Task" [ 971.084700] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 971.094842] env[65726]: DEBUG oslo_vmware.api [None req-4d042142-5ccf-4c86-bcb5-fb6e60309d3b tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Task: {'id': task-5116361, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.128743] env[65726]: DEBUG oslo_service.periodic_task [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=65726) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 971.130955] env[65726]: DEBUG oslo_service.periodic_task [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=65726) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 971.130955] env[65726]: DEBUG oslo_service.periodic_task [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=65726) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 971.130955] env[65726]: DEBUG oslo_service.periodic_task [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=65726) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 971.130955] env[65726]: DEBUG oslo_service.periodic_task [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=65726) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 971.130955] env[65726]: DEBUG oslo_service.periodic_task [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=65726) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 971.130955] env[65726]: DEBUG nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=65726) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11251}} [ 971.130955] env[65726]: DEBUG oslo_service.periodic_task [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Running periodic task ComputeManager.update_available_resource {{(pid=65726) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 971.256654] env[65726]: DEBUG nova.compute.manager [None req-618baf85-0896-4fc2-b1ef-24618b3b2fb7 tempest-FloatingIPsAssociationNegativeTestJSON-737087291 tempest-FloatingIPsAssociationNegativeTestJSON-737087291-project-member] [instance: d8468efa-c2d3-4dce-ab89-fc077011e3d8] Starting instance... {{(pid=65726) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 971.375967] env[65726]: DEBUG oslo_concurrency.lockutils [None req-7ab7520f-2b6e-4be5-84c4-f04aa6f3af4f tempest-ServersTestManualDisk-1753452859 tempest-ServersTestManualDisk-1753452859-project-member] Acquiring lock "3088ee5e-0d27-4058-a94b-2e04c3b52add" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 971.376149] env[65726]: DEBUG oslo_concurrency.lockutils [None req-7ab7520f-2b6e-4be5-84c4-f04aa6f3af4f tempest-ServersTestManualDisk-1753452859 tempest-ServersTestManualDisk-1753452859-project-member] Lock "3088ee5e-0d27-4058-a94b-2e04c3b52add" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 971.376358] env[65726]: DEBUG oslo_concurrency.lockutils [None req-7ab7520f-2b6e-4be5-84c4-f04aa6f3af4f tempest-ServersTestManualDisk-1753452859 tempest-ServersTestManualDisk-1753452859-project-member] Acquiring lock "3088ee5e-0d27-4058-a94b-2e04c3b52add-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 971.376551] env[65726]: DEBUG oslo_concurrency.lockutils [None req-7ab7520f-2b6e-4be5-84c4-f04aa6f3af4f tempest-ServersTestManualDisk-1753452859 tempest-ServersTestManualDisk-1753452859-project-member] Lock "3088ee5e-0d27-4058-a94b-2e04c3b52add-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 971.376605] env[65726]: DEBUG oslo_concurrency.lockutils [None req-7ab7520f-2b6e-4be5-84c4-f04aa6f3af4f tempest-ServersTestManualDisk-1753452859 tempest-ServersTestManualDisk-1753452859-project-member] Lock "3088ee5e-0d27-4058-a94b-2e04c3b52add-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 971.380056] env[65726]: INFO nova.compute.manager [None req-7ab7520f-2b6e-4be5-84c4-f04aa6f3af4f tempest-ServersTestManualDisk-1753452859 tempest-ServersTestManualDisk-1753452859-project-member] [instance: 3088ee5e-0d27-4058-a94b-2e04c3b52add] Terminating instance [ 971.471033] env[65726]: WARNING openstack [req-a4623df5-2a6b-4ab4-81c7-75a5bdd51804 req-449729b7-8acc-401d-a792-50b40ba68a74 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 971.471277] env[65726]: WARNING openstack [req-a4623df5-2a6b-4ab4-81c7-75a5bdd51804 req-449729b7-8acc-401d-a792-50b40ba68a74 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 971.488821] env[65726]: DEBUG oslo_vmware.api [None req-8940e544-9cc2-4a18-9dd0-e7f396520ac6 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Task: {'id': task-5116360, 'name': PowerOffVM_Task} progress is 100%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.494358] env[65726]: DEBUG nova.compute.manager [None req-9f65b40f-615d-4bae-ac4c-c122fff748fd tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] [instance: aa6ce489-c62f-4481-87b7-e74242aeb8ca] Start spawning the instance on the hypervisor. {{(pid=65726) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 971.521188] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116359, 'name': CreateVM_Task, 'duration_secs': 2.334607} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 971.522256] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 73f5b23e-4463-4068-8994-dd2752a2abc9] Created VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 971.523272] env[65726]: DEBUG oslo_concurrency.lockutils [None req-de3a85d2-cc71-4c75-a1b0-fe6db6e57022 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 971.523435] env[65726]: DEBUG oslo_concurrency.lockutils [None req-de3a85d2-cc71-4c75-a1b0-fe6db6e57022 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 971.523769] env[65726]: DEBUG oslo_concurrency.lockutils [None req-de3a85d2-cc71-4c75-a1b0-fe6db6e57022 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 971.524052] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5c6881dd-cc63-4ff8-b62f-273ab701fcb3 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.530506] env[65726]: DEBUG nova.virt.hardware [None req-9f65b40f-615d-4bae-ac4c-c122fff748fd tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T19:28:34Z,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1539f06fa3534e90acbc3a60c4b8bd3f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T19:28:35Z,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 971.530787] env[65726]: DEBUG nova.virt.hardware [None req-9f65b40f-615d-4bae-ac4c-c122fff748fd tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 971.530952] env[65726]: DEBUG nova.virt.hardware [None req-9f65b40f-615d-4bae-ac4c-c122fff748fd tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 971.531144] env[65726]: DEBUG nova.virt.hardware [None req-9f65b40f-615d-4bae-ac4c-c122fff748fd tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 971.531347] env[65726]: DEBUG nova.virt.hardware [None req-9f65b40f-615d-4bae-ac4c-c122fff748fd tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 971.531449] env[65726]: DEBUG nova.virt.hardware [None req-9f65b40f-615d-4bae-ac4c-c122fff748fd tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 971.531662] env[65726]: DEBUG nova.virt.hardware [None req-9f65b40f-615d-4bae-ac4c-c122fff748fd tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 971.531818] env[65726]: DEBUG nova.virt.hardware [None req-9f65b40f-615d-4bae-ac4c-c122fff748fd tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 971.531975] env[65726]: DEBUG nova.virt.hardware [None req-9f65b40f-615d-4bae-ac4c-c122fff748fd tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 971.532149] env[65726]: DEBUG nova.virt.hardware [None req-9f65b40f-615d-4bae-ac4c-c122fff748fd tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 971.532446] env[65726]: DEBUG nova.virt.hardware [None req-9f65b40f-615d-4bae-ac4c-c122fff748fd tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 971.533804] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f87ea0b-0e63-4271-9bed-ea446cc296a9 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.542830] env[65726]: DEBUG oslo_vmware.api [None req-de3a85d2-cc71-4c75-a1b0-fe6db6e57022 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] Waiting for the task: (returnval){ [ 971.542830] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5270b060-173e-6c5e-26be-119982791fcc" [ 971.542830] env[65726]: _type = "Task" [ 971.542830] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 971.551682] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b2b69ec-4343-4695-9c25-639624774b0f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.560089] env[65726]: DEBUG oslo_vmware.api [None req-de3a85d2-cc71-4c75-a1b0-fe6db6e57022 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5270b060-173e-6c5e-26be-119982791fcc, 'name': SearchDatastore_Task, 'duration_secs': 0.010745} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 971.560789] env[65726]: DEBUG oslo_concurrency.lockutils [None req-de3a85d2-cc71-4c75-a1b0-fe6db6e57022 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 971.561107] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-de3a85d2-cc71-4c75-a1b0-fe6db6e57022 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] [instance: 73f5b23e-4463-4068-8994-dd2752a2abc9] Processing image b52362a3-ee8a-4cbf-b06f-513b0cc8f95c {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 971.561385] env[65726]: DEBUG oslo_concurrency.lockutils [None req-de3a85d2-cc71-4c75-a1b0-fe6db6e57022 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 971.561534] env[65726]: DEBUG oslo_concurrency.lockutils [None req-de3a85d2-cc71-4c75-a1b0-fe6db6e57022 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 971.561712] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-de3a85d2-cc71-4c75-a1b0-fe6db6e57022 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 971.561988] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5875daa7-186e-4ce6-bdb9-6921b9b5809e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.579051] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-de3a85d2-cc71-4c75-a1b0-fe6db6e57022 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 971.579248] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-de3a85d2-cc71-4c75-a1b0-fe6db6e57022 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65726) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 971.579980] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7e715eae-0495-472b-81df-e4e54c575720 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.586407] env[65726]: DEBUG nova.scheduler.client.report [None req-d903724e-08f6-4863-a079-5fadb490f6c5 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 971.589871] env[65726]: DEBUG oslo_vmware.api [None req-de3a85d2-cc71-4c75-a1b0-fe6db6e57022 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] Waiting for the task: (returnval){ [ 971.589871] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]527fb22c-0606-b445-32ee-e460d10dc882" [ 971.589871] env[65726]: _type = "Task" [ 971.589871] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 971.604766] env[65726]: DEBUG oslo_vmware.api [None req-4d042142-5ccf-4c86-bcb5-fb6e60309d3b tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Task: {'id': task-5116361, 'name': PowerOffVM_Task, 'duration_secs': 0.455526} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 971.608809] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-4d042142-5ccf-4c86-bcb5-fb6e60309d3b tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] [instance: df46141c-b545-4e03-b3a3-fd9f5feda0d2] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 971.609015] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-4d042142-5ccf-4c86-bcb5-fb6e60309d3b tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] [instance: df46141c-b545-4e03-b3a3-fd9f5feda0d2] Unregistering the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 971.609633] env[65726]: DEBUG oslo_vmware.api [None req-de3a85d2-cc71-4c75-a1b0-fe6db6e57022 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]527fb22c-0606-b445-32ee-e460d10dc882, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.609860] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-06d3bb7e-2542-4316-b7a9-c7fba49f20d7 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.633217] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 971.649541] env[65726]: WARNING openstack [req-a4623df5-2a6b-4ab4-81c7-75a5bdd51804 req-449729b7-8acc-401d-a792-50b40ba68a74 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 971.649922] env[65726]: WARNING openstack [req-a4623df5-2a6b-4ab4-81c7-75a5bdd51804 req-449729b7-8acc-401d-a792-50b40ba68a74 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 971.703813] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-4d042142-5ccf-4c86-bcb5-fb6e60309d3b tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] [instance: df46141c-b545-4e03-b3a3-fd9f5feda0d2] Unregistered the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 971.704271] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-4d042142-5ccf-4c86-bcb5-fb6e60309d3b tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] [instance: df46141c-b545-4e03-b3a3-fd9f5feda0d2] Deleting contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 971.704509] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-4d042142-5ccf-4c86-bcb5-fb6e60309d3b tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Deleting the datastore file [datastore1] df46141c-b545-4e03-b3a3-fd9f5feda0d2 {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 971.704807] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-920bf0d2-8f9e-402c-bfe2-9235bf5233f8 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.714674] env[65726]: DEBUG oslo_vmware.api [None req-4d042142-5ccf-4c86-bcb5-fb6e60309d3b tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Waiting for the task: (returnval){ [ 971.714674] env[65726]: value = "task-5116363" [ 971.714674] env[65726]: _type = "Task" [ 971.714674] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 971.726458] env[65726]: DEBUG oslo_vmware.api [None req-4d042142-5ccf-4c86-bcb5-fb6e60309d3b tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Task: {'id': task-5116363, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.755569] env[65726]: WARNING openstack [req-a4623df5-2a6b-4ab4-81c7-75a5bdd51804 req-449729b7-8acc-401d-a792-50b40ba68a74 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 971.756034] env[65726]: WARNING openstack [req-a4623df5-2a6b-4ab4-81c7-75a5bdd51804 req-449729b7-8acc-401d-a792-50b40ba68a74 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 971.794289] env[65726]: DEBUG oslo_concurrency.lockutils [None req-618baf85-0896-4fc2-b1ef-24618b3b2fb7 tempest-FloatingIPsAssociationNegativeTestJSON-737087291 tempest-FloatingIPsAssociationNegativeTestJSON-737087291-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 971.883136] env[65726]: DEBUG nova.network.neutron [req-a4623df5-2a6b-4ab4-81c7-75a5bdd51804 req-449729b7-8acc-401d-a792-50b40ba68a74 service nova] [instance: cf950ec3-9914-4eb9-99db-048b3969bb21] Updated VIF entry in instance network info cache for port 58accf41-587f-4367-854e-fc571a6a3424. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 971.883544] env[65726]: DEBUG nova.network.neutron [req-a4623df5-2a6b-4ab4-81c7-75a5bdd51804 req-449729b7-8acc-401d-a792-50b40ba68a74 service nova] [instance: cf950ec3-9914-4eb9-99db-048b3969bb21] Updating instance_info_cache with network_info: [{"id": "58accf41-587f-4367-854e-fc571a6a3424", "address": "fa:16:3e:83:9d:0c", "network": {"id": "9d5ad167-99fb-4208-ab30-7053f35774e4", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-2082516762-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.132", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a8362295a62f4b51bae719a7ef5a4656", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dabbac20-1723-40ad-9da0-e53b28073651", "external-id": "nsx-vlan-transportzone-790", "segmentation_id": 790, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap58accf41-58", "ovs_interfaceid": "58accf41-587f-4367-854e-fc571a6a3424", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 971.885107] env[65726]: DEBUG nova.compute.manager [None req-7ab7520f-2b6e-4be5-84c4-f04aa6f3af4f tempest-ServersTestManualDisk-1753452859 tempest-ServersTestManualDisk-1753452859-project-member] [instance: 3088ee5e-0d27-4058-a94b-2e04c3b52add] Start destroying the instance on the hypervisor. {{(pid=65726) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 971.885308] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-7ab7520f-2b6e-4be5-84c4-f04aa6f3af4f tempest-ServersTestManualDisk-1753452859 tempest-ServersTestManualDisk-1753452859-project-member] [instance: 3088ee5e-0d27-4058-a94b-2e04c3b52add] Destroying instance {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 971.886616] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43628b7f-e25b-439c-8a3b-94e63fc3d46f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.894526] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-7ab7520f-2b6e-4be5-84c4-f04aa6f3af4f tempest-ServersTestManualDisk-1753452859 tempest-ServersTestManualDisk-1753452859-project-member] [instance: 3088ee5e-0d27-4058-a94b-2e04c3b52add] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 971.894791] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-60739f6e-f4bb-4479-9765-00512d00e6b2 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.902296] env[65726]: DEBUG oslo_vmware.api [None req-7ab7520f-2b6e-4be5-84c4-f04aa6f3af4f tempest-ServersTestManualDisk-1753452859 tempest-ServersTestManualDisk-1753452859-project-member] Waiting for the task: (returnval){ [ 971.902296] env[65726]: value = "task-5116364" [ 971.902296] env[65726]: _type = "Task" [ 971.902296] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 971.912435] env[65726]: DEBUG oslo_vmware.api [None req-7ab7520f-2b6e-4be5-84c4-f04aa6f3af4f tempest-ServersTestManualDisk-1753452859 tempest-ServersTestManualDisk-1753452859-project-member] Task: {'id': task-5116364, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.989817] env[65726]: DEBUG oslo_vmware.api [None req-8940e544-9cc2-4a18-9dd0-e7f396520ac6 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Task: {'id': task-5116360, 'name': PowerOffVM_Task, 'duration_secs': 0.565074} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 971.990890] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-8940e544-9cc2-4a18-9dd0-e7f396520ac6 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 971.991107] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a800a73-cac1-477a-b88f-02b1e13d7d53 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.014978] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56e6c18a-b642-4c1e-bc3d-0b4c15baa366 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.092434] env[65726]: DEBUG oslo_concurrency.lockutils [None req-d903724e-08f6-4863-a079-5fadb490f6c5 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.642s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 972.093018] env[65726]: DEBUG nova.compute.manager [None req-d903724e-08f6-4863-a079-5fadb490f6c5 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 54e85161-5f63-405f-83f8-490c70645a3b] Start building networks asynchronously for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 972.095995] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c4955188-25cc-47e9-912c-88328888d7f1 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 14.672s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 972.096930] env[65726]: DEBUG nova.objects.instance [None req-c4955188-25cc-47e9-912c-88328888d7f1 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Lazy-loading 'resources' on Instance uuid 01d86089-6b9b-4588-864e-ef91375a1eea {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 972.109419] env[65726]: DEBUG oslo_vmware.api [None req-de3a85d2-cc71-4c75-a1b0-fe6db6e57022 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]527fb22c-0606-b445-32ee-e460d10dc882, 'name': SearchDatastore_Task, 'duration_secs': 0.021994} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.110523] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0be7e498-43eb-4307-b9fe-504dd915aa5b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.117657] env[65726]: DEBUG oslo_vmware.api [None req-de3a85d2-cc71-4c75-a1b0-fe6db6e57022 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] Waiting for the task: (returnval){ [ 972.117657] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]527ff19b-1899-20f9-fabd-c237ad67a898" [ 972.117657] env[65726]: _type = "Task" [ 972.117657] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.127117] env[65726]: DEBUG oslo_vmware.api [None req-de3a85d2-cc71-4c75-a1b0-fe6db6e57022 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]527ff19b-1899-20f9-fabd-c237ad67a898, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.170991] env[65726]: DEBUG nova.compute.manager [req-31c2aaeb-5897-4dd4-aff3-e8d158902f40 req-42cbaa1b-85cd-4bea-803a-35f438e39aea service nova] [instance: aa6ce489-c62f-4481-87b7-e74242aeb8ca] Received event network-vif-plugged-d5a805f5-b6a8-42ec-92a0-4a652b51c1e6 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 972.171319] env[65726]: DEBUG oslo_concurrency.lockutils [req-31c2aaeb-5897-4dd4-aff3-e8d158902f40 req-42cbaa1b-85cd-4bea-803a-35f438e39aea service nova] Acquiring lock "aa6ce489-c62f-4481-87b7-e74242aeb8ca-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 972.171508] env[65726]: DEBUG oslo_concurrency.lockutils [req-31c2aaeb-5897-4dd4-aff3-e8d158902f40 req-42cbaa1b-85cd-4bea-803a-35f438e39aea service nova] Lock "aa6ce489-c62f-4481-87b7-e74242aeb8ca-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 972.171681] env[65726]: DEBUG oslo_concurrency.lockutils [req-31c2aaeb-5897-4dd4-aff3-e8d158902f40 req-42cbaa1b-85cd-4bea-803a-35f438e39aea service nova] Lock "aa6ce489-c62f-4481-87b7-e74242aeb8ca-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 972.171840] env[65726]: DEBUG nova.compute.manager [req-31c2aaeb-5897-4dd4-aff3-e8d158902f40 req-42cbaa1b-85cd-4bea-803a-35f438e39aea service nova] [instance: aa6ce489-c62f-4481-87b7-e74242aeb8ca] No waiting events found dispatching network-vif-plugged-d5a805f5-b6a8-42ec-92a0-4a652b51c1e6 {{(pid=65726) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 972.172518] env[65726]: WARNING nova.compute.manager [req-31c2aaeb-5897-4dd4-aff3-e8d158902f40 req-42cbaa1b-85cd-4bea-803a-35f438e39aea service nova] [instance: aa6ce489-c62f-4481-87b7-e74242aeb8ca] Received unexpected event network-vif-plugged-d5a805f5-b6a8-42ec-92a0-4a652b51c1e6 for instance with vm_state building and task_state spawning. [ 972.227520] env[65726]: DEBUG oslo_vmware.api [None req-4d042142-5ccf-4c86-bcb5-fb6e60309d3b tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Task: {'id': task-5116363, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.141494} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.227943] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-4d042142-5ccf-4c86-bcb5-fb6e60309d3b tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Deleted the datastore file {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 972.228207] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-4d042142-5ccf-4c86-bcb5-fb6e60309d3b tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] [instance: df46141c-b545-4e03-b3a3-fd9f5feda0d2] Deleted contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 972.228388] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-4d042142-5ccf-4c86-bcb5-fb6e60309d3b tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] [instance: df46141c-b545-4e03-b3a3-fd9f5feda0d2] Instance destroyed {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 972.228555] env[65726]: INFO nova.compute.manager [None req-4d042142-5ccf-4c86-bcb5-fb6e60309d3b tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] [instance: df46141c-b545-4e03-b3a3-fd9f5feda0d2] Took 1.21 seconds to destroy the instance on the hypervisor. [ 972.228798] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-4d042142-5ccf-4c86-bcb5-fb6e60309d3b tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 972.228990] env[65726]: DEBUG nova.compute.manager [-] [instance: df46141c-b545-4e03-b3a3-fd9f5feda0d2] Deallocating network for instance {{(pid=65726) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 972.229177] env[65726]: DEBUG nova.network.neutron [-] [instance: df46141c-b545-4e03-b3a3-fd9f5feda0d2] deallocate_for_instance() {{(pid=65726) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 972.229754] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 972.230064] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 972.300595] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 972.300595] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 972.311658] env[65726]: DEBUG nova.network.neutron [None req-9f65b40f-615d-4bae-ac4c-c122fff748fd tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] [instance: aa6ce489-c62f-4481-87b7-e74242aeb8ca] Successfully updated port: d5a805f5-b6a8-42ec-92a0-4a652b51c1e6 {{(pid=65726) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 972.387424] env[65726]: DEBUG oslo_concurrency.lockutils [req-a4623df5-2a6b-4ab4-81c7-75a5bdd51804 req-449729b7-8acc-401d-a792-50b40ba68a74 service nova] Releasing lock "refresh_cache-cf950ec3-9914-4eb9-99db-048b3969bb21" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 972.414235] env[65726]: DEBUG oslo_vmware.api [None req-7ab7520f-2b6e-4be5-84c4-f04aa6f3af4f tempest-ServersTestManualDisk-1753452859 tempest-ServersTestManualDisk-1753452859-project-member] Task: {'id': task-5116364, 'name': PowerOffVM_Task, 'duration_secs': 0.263107} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.414549] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-7ab7520f-2b6e-4be5-84c4-f04aa6f3af4f tempest-ServersTestManualDisk-1753452859 tempest-ServersTestManualDisk-1753452859-project-member] [instance: 3088ee5e-0d27-4058-a94b-2e04c3b52add] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 972.414735] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-7ab7520f-2b6e-4be5-84c4-f04aa6f3af4f tempest-ServersTestManualDisk-1753452859 tempest-ServersTestManualDisk-1753452859-project-member] [instance: 3088ee5e-0d27-4058-a94b-2e04c3b52add] Unregistering the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 972.415084] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6223d83d-b358-4bdc-9b90-eca61d5bc292 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.482718] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-7ab7520f-2b6e-4be5-84c4-f04aa6f3af4f tempest-ServersTestManualDisk-1753452859 tempest-ServersTestManualDisk-1753452859-project-member] [instance: 3088ee5e-0d27-4058-a94b-2e04c3b52add] Unregistered the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 972.484459] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-7ab7520f-2b6e-4be5-84c4-f04aa6f3af4f tempest-ServersTestManualDisk-1753452859 tempest-ServersTestManualDisk-1753452859-project-member] [instance: 3088ee5e-0d27-4058-a94b-2e04c3b52add] Deleting contents of the VM from datastore datastore2 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 972.484459] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-7ab7520f-2b6e-4be5-84c4-f04aa6f3af4f tempest-ServersTestManualDisk-1753452859 tempest-ServersTestManualDisk-1753452859-project-member] Deleting the datastore file [datastore2] 3088ee5e-0d27-4058-a94b-2e04c3b52add {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 972.484459] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-39ce946d-e632-44fe-8444-1e6e40f17203 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.496031] env[65726]: DEBUG oslo_vmware.api [None req-7ab7520f-2b6e-4be5-84c4-f04aa6f3af4f tempest-ServersTestManualDisk-1753452859 tempest-ServersTestManualDisk-1753452859-project-member] Waiting for the task: (returnval){ [ 972.496031] env[65726]: value = "task-5116366" [ 972.496031] env[65726]: _type = "Task" [ 972.496031] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.505587] env[65726]: DEBUG oslo_vmware.api [None req-7ab7520f-2b6e-4be5-84c4-f04aa6f3af4f tempest-ServersTestManualDisk-1753452859 tempest-ServersTestManualDisk-1753452859-project-member] Task: {'id': task-5116366, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.528876] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-8940e544-9cc2-4a18-9dd0-e7f396520ac6 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04] Creating Snapshot of the VM instance {{(pid=65726) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 972.530475] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-43f64553-f14d-41ab-8305-fedae7bd8d77 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.540123] env[65726]: DEBUG oslo_vmware.api [None req-8940e544-9cc2-4a18-9dd0-e7f396520ac6 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Waiting for the task: (returnval){ [ 972.540123] env[65726]: value = "task-5116367" [ 972.540123] env[65726]: _type = "Task" [ 972.540123] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.552853] env[65726]: DEBUG oslo_vmware.api [None req-8940e544-9cc2-4a18-9dd0-e7f396520ac6 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Task: {'id': task-5116367, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.599879] env[65726]: DEBUG nova.compute.utils [None req-d903724e-08f6-4863-a079-5fadb490f6c5 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Using /dev/sd instead of None {{(pid=65726) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 972.602650] env[65726]: DEBUG nova.compute.manager [None req-d903724e-08f6-4863-a079-5fadb490f6c5 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 54e85161-5f63-405f-83f8-490c70645a3b] Allocating IP information in the background. {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 972.602650] env[65726]: DEBUG nova.network.neutron [None req-d903724e-08f6-4863-a079-5fadb490f6c5 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 54e85161-5f63-405f-83f8-490c70645a3b] allocate_for_instance() {{(pid=65726) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 972.602650] env[65726]: WARNING neutronclient.v2_0.client [None req-d903724e-08f6-4863-a079-5fadb490f6c5 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 972.602650] env[65726]: WARNING neutronclient.v2_0.client [None req-d903724e-08f6-4863-a079-5fadb490f6c5 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 972.603205] env[65726]: WARNING openstack [None req-d903724e-08f6-4863-a079-5fadb490f6c5 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 972.603915] env[65726]: WARNING openstack [None req-d903724e-08f6-4863-a079-5fadb490f6c5 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 972.628685] env[65726]: DEBUG oslo_vmware.api [None req-de3a85d2-cc71-4c75-a1b0-fe6db6e57022 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]527ff19b-1899-20f9-fabd-c237ad67a898, 'name': SearchDatastore_Task, 'duration_secs': 0.01204} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.628935] env[65726]: DEBUG oslo_concurrency.lockutils [None req-de3a85d2-cc71-4c75-a1b0-fe6db6e57022 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 972.629242] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-de3a85d2-cc71-4c75-a1b0-fe6db6e57022 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] 73f5b23e-4463-4068-8994-dd2752a2abc9/73f5b23e-4463-4068-8994-dd2752a2abc9.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 972.629507] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5e1d66dd-bbb6-4b21-8161-b57ebc790dc4 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.641742] env[65726]: DEBUG oslo_vmware.api [None req-de3a85d2-cc71-4c75-a1b0-fe6db6e57022 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] Waiting for the task: (returnval){ [ 972.641742] env[65726]: value = "task-5116368" [ 972.641742] env[65726]: _type = "Task" [ 972.641742] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.656451] env[65726]: DEBUG oslo_vmware.api [None req-de3a85d2-cc71-4c75-a1b0-fe6db6e57022 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] Task: {'id': task-5116368, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.668747] env[65726]: DEBUG nova.policy [None req-d903724e-08f6-4863-a079-5fadb490f6c5 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9b6087cfa55d41a4bad9f6cee89865d5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '39067912d091464192516c08eb576696', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65726) authorize /opt/stack/nova/nova/policy.py:192}} [ 972.818430] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9f65b40f-615d-4bae-ac4c-c122fff748fd tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Acquiring lock "refresh_cache-aa6ce489-c62f-4481-87b7-e74242aeb8ca" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 972.818643] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9f65b40f-615d-4bae-ac4c-c122fff748fd tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Acquired lock "refresh_cache-aa6ce489-c62f-4481-87b7-e74242aeb8ca" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 972.819229] env[65726]: DEBUG nova.network.neutron [None req-9f65b40f-615d-4bae-ac4c-c122fff748fd tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] [instance: aa6ce489-c62f-4481-87b7-e74242aeb8ca] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 972.964380] env[65726]: DEBUG nova.network.neutron [None req-d903724e-08f6-4863-a079-5fadb490f6c5 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 54e85161-5f63-405f-83f8-490c70645a3b] Successfully created port: 8adb72c3-8e2f-48d7-8197-301c8958dbfb {{(pid=65726) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 973.007148] env[65726]: DEBUG oslo_vmware.api [None req-7ab7520f-2b6e-4be5-84c4-f04aa6f3af4f tempest-ServersTestManualDisk-1753452859 tempest-ServersTestManualDisk-1753452859-project-member] Task: {'id': task-5116366, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.185289} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 973.010380] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-7ab7520f-2b6e-4be5-84c4-f04aa6f3af4f tempest-ServersTestManualDisk-1753452859 tempest-ServersTestManualDisk-1753452859-project-member] Deleted the datastore file {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 973.010665] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-7ab7520f-2b6e-4be5-84c4-f04aa6f3af4f tempest-ServersTestManualDisk-1753452859 tempest-ServersTestManualDisk-1753452859-project-member] [instance: 3088ee5e-0d27-4058-a94b-2e04c3b52add] Deleted contents of the VM from datastore datastore2 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 973.010882] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-7ab7520f-2b6e-4be5-84c4-f04aa6f3af4f tempest-ServersTestManualDisk-1753452859 tempest-ServersTestManualDisk-1753452859-project-member] [instance: 3088ee5e-0d27-4058-a94b-2e04c3b52add] Instance destroyed {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 973.011071] env[65726]: INFO nova.compute.manager [None req-7ab7520f-2b6e-4be5-84c4-f04aa6f3af4f tempest-ServersTestManualDisk-1753452859 tempest-ServersTestManualDisk-1753452859-project-member] [instance: 3088ee5e-0d27-4058-a94b-2e04c3b52add] Took 1.13 seconds to destroy the instance on the hypervisor. [ 973.011408] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-7ab7520f-2b6e-4be5-84c4-f04aa6f3af4f tempest-ServersTestManualDisk-1753452859 tempest-ServersTestManualDisk-1753452859-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 973.011874] env[65726]: DEBUG nova.compute.manager [-] [instance: 3088ee5e-0d27-4058-a94b-2e04c3b52add] Deallocating network for instance {{(pid=65726) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 973.012039] env[65726]: DEBUG nova.network.neutron [-] [instance: 3088ee5e-0d27-4058-a94b-2e04c3b52add] deallocate_for_instance() {{(pid=65726) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 973.012629] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 973.012776] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 973.026976] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42ddf461-48a1-443f-9b7c-8dff36b65472 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.030538] env[65726]: DEBUG oslo_concurrency.lockutils [None req-4d7e3d4f-40bd-4bd5-8d6c-4c6d705b54c9 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Acquiring lock "29823bc7-3909-4ab7-8119-91ee59b289c5" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 973.030782] env[65726]: DEBUG oslo_concurrency.lockutils [None req-4d7e3d4f-40bd-4bd5-8d6c-4c6d705b54c9 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Lock "29823bc7-3909-4ab7-8119-91ee59b289c5" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 973.030977] env[65726]: DEBUG oslo_concurrency.lockutils [None req-4d7e3d4f-40bd-4bd5-8d6c-4c6d705b54c9 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Acquiring lock "29823bc7-3909-4ab7-8119-91ee59b289c5-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 973.031175] env[65726]: DEBUG oslo_concurrency.lockutils [None req-4d7e3d4f-40bd-4bd5-8d6c-4c6d705b54c9 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Lock "29823bc7-3909-4ab7-8119-91ee59b289c5-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 973.031369] env[65726]: DEBUG oslo_concurrency.lockutils [None req-4d7e3d4f-40bd-4bd5-8d6c-4c6d705b54c9 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Lock "29823bc7-3909-4ab7-8119-91ee59b289c5-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 973.034290] env[65726]: INFO nova.compute.manager [None req-4d7e3d4f-40bd-4bd5-8d6c-4c6d705b54c9 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] [instance: 29823bc7-3909-4ab7-8119-91ee59b289c5] Terminating instance [ 973.045938] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf4d5499-8aa4-4b49-be68-7e89ddb0a590 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.082052] env[65726]: DEBUG nova.network.neutron [-] [instance: df46141c-b545-4e03-b3a3-fd9f5feda0d2] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 973.086982] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 973.087385] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 973.099951] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee00a7b8-8082-445f-aee1-4bc26ea61c10 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.103340] env[65726]: DEBUG oslo_vmware.api [None req-8940e544-9cc2-4a18-9dd0-e7f396520ac6 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Task: {'id': task-5116367, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.112752] env[65726]: DEBUG nova.compute.manager [None req-d903724e-08f6-4863-a079-5fadb490f6c5 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 54e85161-5f63-405f-83f8-490c70645a3b] Start building block device mappings for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 973.117347] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98bdf1b6-ce2a-4edf-9dc9-056af5033113 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.137967] env[65726]: DEBUG nova.compute.provider_tree [None req-c4955188-25cc-47e9-912c-88328888d7f1 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Updating inventory in ProviderTree for provider 07c4692f-bdb4-4058-9173-ff9664830295 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 973.155506] env[65726]: DEBUG oslo_vmware.api [None req-de3a85d2-cc71-4c75-a1b0-fe6db6e57022 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] Task: {'id': task-5116368, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.326533] env[65726]: WARNING openstack [None req-9f65b40f-615d-4bae-ac4c-c122fff748fd tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 973.326992] env[65726]: WARNING openstack [None req-9f65b40f-615d-4bae-ac4c-c122fff748fd tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 973.332930] env[65726]: DEBUG nova.network.neutron [None req-9f65b40f-615d-4bae-ac4c-c122fff748fd tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] [instance: aa6ce489-c62f-4481-87b7-e74242aeb8ca] Instance cache missing network info. {{(pid=65726) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 973.538795] env[65726]: DEBUG nova.compute.manager [None req-4d7e3d4f-40bd-4bd5-8d6c-4c6d705b54c9 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] [instance: 29823bc7-3909-4ab7-8119-91ee59b289c5] Start destroying the instance on the hypervisor. {{(pid=65726) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 973.539080] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-4d7e3d4f-40bd-4bd5-8d6c-4c6d705b54c9 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] [instance: 29823bc7-3909-4ab7-8119-91ee59b289c5] Destroying instance {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 973.540817] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4501da39-d47b-4f47-a0fb-88a218b8898a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.563868] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-4d7e3d4f-40bd-4bd5-8d6c-4c6d705b54c9 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] [instance: 29823bc7-3909-4ab7-8119-91ee59b289c5] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 973.564715] env[65726]: DEBUG oslo_vmware.api [None req-8940e544-9cc2-4a18-9dd0-e7f396520ac6 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Task: {'id': task-5116367, 'name': CreateSnapshot_Task, 'duration_secs': 0.609446} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 973.565342] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-adbfee93-ee12-4ce6-a282-0831b3c248f4 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.567042] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-8940e544-9cc2-4a18-9dd0-e7f396520ac6 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04] Created Snapshot of the VM instance {{(pid=65726) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 973.568032] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1923e3b8-b286-4f0c-868f-fdaf95108e4a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.582018] env[65726]: DEBUG oslo_vmware.api [None req-4d7e3d4f-40bd-4bd5-8d6c-4c6d705b54c9 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Waiting for the task: (returnval){ [ 973.582018] env[65726]: value = "task-5116369" [ 973.582018] env[65726]: _type = "Task" [ 973.582018] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 973.592663] env[65726]: DEBUG oslo_vmware.api [None req-4d7e3d4f-40bd-4bd5-8d6c-4c6d705b54c9 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Task: {'id': task-5116369, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.596343] env[65726]: INFO nova.compute.manager [-] [instance: df46141c-b545-4e03-b3a3-fd9f5feda0d2] Took 1.37 seconds to deallocate network for instance. [ 973.638280] env[65726]: WARNING openstack [None req-9f65b40f-615d-4bae-ac4c-c122fff748fd tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 973.638853] env[65726]: WARNING openstack [None req-9f65b40f-615d-4bae-ac4c-c122fff748fd tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 973.668563] env[65726]: DEBUG oslo_vmware.api [None req-de3a85d2-cc71-4c75-a1b0-fe6db6e57022 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] Task: {'id': task-5116368, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.531252} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 973.668660] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-de3a85d2-cc71-4c75-a1b0-fe6db6e57022 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] 73f5b23e-4463-4068-8994-dd2752a2abc9/73f5b23e-4463-4068-8994-dd2752a2abc9.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 973.668814] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-de3a85d2-cc71-4c75-a1b0-fe6db6e57022 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] [instance: 73f5b23e-4463-4068-8994-dd2752a2abc9] Extending root virtual disk to 1048576 {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 973.669208] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-cad0a078-5b83-4d02-b436-b2b2d6450c0a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.677748] env[65726]: ERROR nova.scheduler.client.report [None req-c4955188-25cc-47e9-912c-88328888d7f1 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [req-ac084434-43c3-4e76-80fd-7b37dd71ad16] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 07c4692f-bdb4-4058-9173-ff9664830295. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-ac084434-43c3-4e76-80fd-7b37dd71ad16"}]} [ 973.682131] env[65726]: DEBUG oslo_vmware.api [None req-de3a85d2-cc71-4c75-a1b0-fe6db6e57022 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] Waiting for the task: (returnval){ [ 973.682131] env[65726]: value = "task-5116370" [ 973.682131] env[65726]: _type = "Task" [ 973.682131] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 973.697102] env[65726]: DEBUG oslo_vmware.api [None req-de3a85d2-cc71-4c75-a1b0-fe6db6e57022 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] Task: {'id': task-5116370, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.705491] env[65726]: DEBUG nova.scheduler.client.report [None req-c4955188-25cc-47e9-912c-88328888d7f1 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Refreshing inventories for resource provider 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 973.721449] env[65726]: DEBUG nova.scheduler.client.report [None req-c4955188-25cc-47e9-912c-88328888d7f1 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Updating ProviderTree inventory for provider 07c4692f-bdb4-4058-9173-ff9664830295 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 973.721704] env[65726]: DEBUG nova.compute.provider_tree [None req-c4955188-25cc-47e9-912c-88328888d7f1 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Updating inventory in ProviderTree for provider 07c4692f-bdb4-4058-9173-ff9664830295 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 973.727113] env[65726]: WARNING openstack [None req-9f65b40f-615d-4bae-ac4c-c122fff748fd tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 973.727471] env[65726]: WARNING openstack [None req-9f65b40f-615d-4bae-ac4c-c122fff748fd tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 973.735691] env[65726]: DEBUG nova.scheduler.client.report [None req-c4955188-25cc-47e9-912c-88328888d7f1 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Refreshing aggregate associations for resource provider 07c4692f-bdb4-4058-9173-ff9664830295, aggregates: None {{(pid=65726) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 973.757363] env[65726]: DEBUG nova.scheduler.client.report [None req-c4955188-25cc-47e9-912c-88328888d7f1 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Refreshing trait associations for resource provider 07c4692f-bdb4-4058-9173-ff9664830295, traits: COMPUTE_IMAGE_TYPE_ISO,HW_ARCH_X86_64,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=65726) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 973.767326] env[65726]: DEBUG nova.compute.manager [req-21471198-dbcc-4485-846b-5493cfe5ac44 req-7ab8f941-dc97-4ccc-bb39-f5735d7e0963 service nova] [instance: 3088ee5e-0d27-4058-a94b-2e04c3b52add] Received event network-vif-deleted-2ffc16a0-351c-4f5b-8a4a-993ae7adb4f3 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 973.767509] env[65726]: INFO nova.compute.manager [req-21471198-dbcc-4485-846b-5493cfe5ac44 req-7ab8f941-dc97-4ccc-bb39-f5735d7e0963 service nova] [instance: 3088ee5e-0d27-4058-a94b-2e04c3b52add] Neutron deleted interface 2ffc16a0-351c-4f5b-8a4a-993ae7adb4f3; detaching it from the instance and deleting it from the info cache [ 973.767697] env[65726]: DEBUG nova.network.neutron [req-21471198-dbcc-4485-846b-5493cfe5ac44 req-7ab8f941-dc97-4ccc-bb39-f5735d7e0963 service nova] [instance: 3088ee5e-0d27-4058-a94b-2e04c3b52add] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 973.899825] env[65726]: DEBUG nova.network.neutron [None req-9f65b40f-615d-4bae-ac4c-c122fff748fd tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] [instance: aa6ce489-c62f-4481-87b7-e74242aeb8ca] Updating instance_info_cache with network_info: [{"id": "d5a805f5-b6a8-42ec-92a0-4a652b51c1e6", "address": "fa:16:3e:f5:d5:d7", "network": {"id": "547d7b16-4591-4d80-9e9b-c8a9a3a78dab", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-2077874536-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4ffd45f4a7a041199a4fc7f69f5e7e9b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2ed91b7b-b4ec-486d-ab34-af0afb7ec691", "external-id": "nsx-vlan-transportzone-75", "segmentation_id": 75, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd5a805f5-b6", "ovs_interfaceid": "d5a805f5-b6a8-42ec-92a0-4a652b51c1e6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 974.055292] env[65726]: DEBUG nova.network.neutron [-] [instance: 3088ee5e-0d27-4058-a94b-2e04c3b52add] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 974.094892] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-8940e544-9cc2-4a18-9dd0-e7f396520ac6 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04] Creating linked-clone VM from snapshot {{(pid=65726) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 974.097015] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-92d870c2-5e4b-4184-b560-8e3bba4279c1 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.105033] env[65726]: DEBUG oslo_concurrency.lockutils [None req-4d042142-5ccf-4c86-bcb5-fb6e60309d3b tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 974.120216] env[65726]: DEBUG oslo_vmware.api [None req-4d7e3d4f-40bd-4bd5-8d6c-4c6d705b54c9 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Task: {'id': task-5116369, 'name': PowerOffVM_Task, 'duration_secs': 0.40015} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 974.121693] env[65726]: DEBUG oslo_vmware.api [None req-8940e544-9cc2-4a18-9dd0-e7f396520ac6 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Waiting for the task: (returnval){ [ 974.121693] env[65726]: value = "task-5116371" [ 974.121693] env[65726]: _type = "Task" [ 974.121693] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.121693] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-4d7e3d4f-40bd-4bd5-8d6c-4c6d705b54c9 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] [instance: 29823bc7-3909-4ab7-8119-91ee59b289c5] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 974.121693] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-4d7e3d4f-40bd-4bd5-8d6c-4c6d705b54c9 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] [instance: 29823bc7-3909-4ab7-8119-91ee59b289c5] Unregistering the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 974.121693] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d63a3de5-7614-4a97-a58a-24a84a291735 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.129438] env[65726]: DEBUG nova.compute.manager [None req-d903724e-08f6-4863-a079-5fadb490f6c5 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 54e85161-5f63-405f-83f8-490c70645a3b] Start spawning the instance on the hypervisor. {{(pid=65726) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 974.136863] env[65726]: DEBUG oslo_vmware.api [None req-8940e544-9cc2-4a18-9dd0-e7f396520ac6 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Task: {'id': task-5116371, 'name': CloneVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.158948] env[65726]: DEBUG nova.virt.hardware [None req-d903724e-08f6-4863-a079-5fadb490f6c5 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T19:28:34Z,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1539f06fa3534e90acbc3a60c4b8bd3f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T19:28:35Z,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 974.159230] env[65726]: DEBUG nova.virt.hardware [None req-d903724e-08f6-4863-a079-5fadb490f6c5 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 974.159381] env[65726]: DEBUG nova.virt.hardware [None req-d903724e-08f6-4863-a079-5fadb490f6c5 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 974.159552] env[65726]: DEBUG nova.virt.hardware [None req-d903724e-08f6-4863-a079-5fadb490f6c5 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 974.159774] env[65726]: DEBUG nova.virt.hardware [None req-d903724e-08f6-4863-a079-5fadb490f6c5 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 974.159941] env[65726]: DEBUG nova.virt.hardware [None req-d903724e-08f6-4863-a079-5fadb490f6c5 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 974.160735] env[65726]: DEBUG nova.virt.hardware [None req-d903724e-08f6-4863-a079-5fadb490f6c5 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 974.160957] env[65726]: DEBUG nova.virt.hardware [None req-d903724e-08f6-4863-a079-5fadb490f6c5 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 974.161155] env[65726]: DEBUG nova.virt.hardware [None req-d903724e-08f6-4863-a079-5fadb490f6c5 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 974.161386] env[65726]: DEBUG nova.virt.hardware [None req-d903724e-08f6-4863-a079-5fadb490f6c5 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 974.161575] env[65726]: DEBUG nova.virt.hardware [None req-d903724e-08f6-4863-a079-5fadb490f6c5 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 974.165209] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-816a2a52-d4d6-4ea3-aa8c-84798fe26e2d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.178511] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65d8e2ec-37a5-4b43-8de9-907daa262fd5 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.186053] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d76bc97-e78e-427b-b013-a4a59e1635c3 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.207552] env[65726]: DEBUG oslo_vmware.api [None req-de3a85d2-cc71-4c75-a1b0-fe6db6e57022 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] Task: {'id': task-5116370, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.220118} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 974.208656] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a60bcaf-8f8d-4a4e-8856-80d0a69c9091 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.212470] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-de3a85d2-cc71-4c75-a1b0-fe6db6e57022 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] [instance: 73f5b23e-4463-4068-8994-dd2752a2abc9] Extended root virtual disk {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 974.213653] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c809f38f-e4f5-4aa0-8c83-a3e3f26181bc {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.238257] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-de3a85d2-cc71-4c75-a1b0-fe6db6e57022 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] [instance: 73f5b23e-4463-4068-8994-dd2752a2abc9] Reconfiguring VM instance instance-0000004d to attach disk [datastore1] 73f5b23e-4463-4068-8994-dd2752a2abc9/73f5b23e-4463-4068-8994-dd2752a2abc9.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 974.239868] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-921725af-407b-4da2-993e-311dc385a617 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.282718] env[65726]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-627beda0-d1f6-4d0d-8bf3-7d7ef11ec542 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.285754] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aefa58ec-3cd0-4553-9282-e8bb4b1e746d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.290835] env[65726]: DEBUG oslo_vmware.api [None req-de3a85d2-cc71-4c75-a1b0-fe6db6e57022 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] Waiting for the task: (returnval){ [ 974.290835] env[65726]: value = "task-5116373" [ 974.290835] env[65726]: _type = "Task" [ 974.290835] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.303435] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cac27fc5-eea9-47ed-b6fb-e412c4d0f5f3 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.325030] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31b9bff6-66e3-48f6-b37e-07258ce13864 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.327949] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-4d7e3d4f-40bd-4bd5-8d6c-4c6d705b54c9 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] [instance: 29823bc7-3909-4ab7-8119-91ee59b289c5] Unregistered the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 974.328218] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-4d7e3d4f-40bd-4bd5-8d6c-4c6d705b54c9 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] [instance: 29823bc7-3909-4ab7-8119-91ee59b289c5] Deleting contents of the VM from datastore datastore2 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 974.328337] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-4d7e3d4f-40bd-4bd5-8d6c-4c6d705b54c9 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Deleting the datastore file [datastore2] 29823bc7-3909-4ab7-8119-91ee59b289c5 {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 974.328652] env[65726]: DEBUG oslo_vmware.api [None req-de3a85d2-cc71-4c75-a1b0-fe6db6e57022 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] Task: {'id': task-5116373, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.329461] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-35ecebf5-6762-410a-b8ce-f36869d946f4 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.353646] env[65726]: DEBUG nova.compute.provider_tree [None req-c4955188-25cc-47e9-912c-88328888d7f1 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Updating inventory in ProviderTree for provider 07c4692f-bdb4-4058-9173-ff9664830295 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 974.370458] env[65726]: DEBUG oslo_vmware.api [None req-4d7e3d4f-40bd-4bd5-8d6c-4c6d705b54c9 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Waiting for the task: (returnval){ [ 974.370458] env[65726]: value = "task-5116374" [ 974.370458] env[65726]: _type = "Task" [ 974.370458] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.371532] env[65726]: DEBUG nova.compute.manager [req-21471198-dbcc-4485-846b-5493cfe5ac44 req-7ab8f941-dc97-4ccc-bb39-f5735d7e0963 service nova] [instance: 3088ee5e-0d27-4058-a94b-2e04c3b52add] Detach interface failed, port_id=2ffc16a0-351c-4f5b-8a4a-993ae7adb4f3, reason: Instance 3088ee5e-0d27-4058-a94b-2e04c3b52add could not be found. {{(pid=65726) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11669}} [ 974.383966] env[65726]: DEBUG oslo_vmware.api [None req-4d7e3d4f-40bd-4bd5-8d6c-4c6d705b54c9 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Task: {'id': task-5116374, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.402733] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9f65b40f-615d-4bae-ac4c-c122fff748fd tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Releasing lock "refresh_cache-aa6ce489-c62f-4481-87b7-e74242aeb8ca" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 974.402870] env[65726]: DEBUG nova.compute.manager [None req-9f65b40f-615d-4bae-ac4c-c122fff748fd tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] [instance: aa6ce489-c62f-4481-87b7-e74242aeb8ca] Instance network_info: |[{"id": "d5a805f5-b6a8-42ec-92a0-4a652b51c1e6", "address": "fa:16:3e:f5:d5:d7", "network": {"id": "547d7b16-4591-4d80-9e9b-c8a9a3a78dab", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-2077874536-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4ffd45f4a7a041199a4fc7f69f5e7e9b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2ed91b7b-b4ec-486d-ab34-af0afb7ec691", "external-id": "nsx-vlan-transportzone-75", "segmentation_id": 75, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd5a805f5-b6", "ovs_interfaceid": "d5a805f5-b6a8-42ec-92a0-4a652b51c1e6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 974.403556] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-9f65b40f-615d-4bae-ac4c-c122fff748fd tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] [instance: aa6ce489-c62f-4481-87b7-e74242aeb8ca] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f5:d5:d7', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '2ed91b7b-b4ec-486d-ab34-af0afb7ec691', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd5a805f5-b6a8-42ec-92a0-4a652b51c1e6', 'vif_model': 'vmxnet3'}] {{(pid=65726) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 974.412257] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-9f65b40f-615d-4bae-ac4c-c122fff748fd tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Creating folder: Project (4ffd45f4a7a041199a4fc7f69f5e7e9b). Parent ref: group-v995008. {{(pid=65726) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 974.413024] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-dd2dabe9-e206-41bd-b47d-cf60a43d29db {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.427369] env[65726]: INFO nova.virt.vmwareapi.vm_util [None req-9f65b40f-615d-4bae-ac4c-c122fff748fd tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Created folder: Project (4ffd45f4a7a041199a4fc7f69f5e7e9b) in parent group-v995008. [ 974.427369] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-9f65b40f-615d-4bae-ac4c-c122fff748fd tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Creating folder: Instances. Parent ref: group-v995237. {{(pid=65726) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 974.427369] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-3809af17-3aab-49ef-a6a7-a702f4344049 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.439796] env[65726]: INFO nova.virt.vmwareapi.vm_util [None req-9f65b40f-615d-4bae-ac4c-c122fff748fd tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Created folder: Instances in parent group-v995237. [ 974.440202] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-9f65b40f-615d-4bae-ac4c-c122fff748fd tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 974.441029] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: aa6ce489-c62f-4481-87b7-e74242aeb8ca] Creating VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 974.441029] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-0ec6f72b-0b00-4d4f-90a5-574e4ea9b0ad {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.463949] env[65726]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 974.463949] env[65726]: value = "task-5116377" [ 974.463949] env[65726]: _type = "Task" [ 974.463949] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.473809] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116377, 'name': CreateVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.543832] env[65726]: DEBUG nova.compute.manager [req-0172e8c1-1b40-4ad7-a064-b16dd49d8418 req-17239211-ffcc-474f-88fe-b08f65340fc4 service nova] [instance: aa6ce489-c62f-4481-87b7-e74242aeb8ca] Received event network-changed-d5a805f5-b6a8-42ec-92a0-4a652b51c1e6 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 974.544136] env[65726]: DEBUG nova.compute.manager [req-0172e8c1-1b40-4ad7-a064-b16dd49d8418 req-17239211-ffcc-474f-88fe-b08f65340fc4 service nova] [instance: aa6ce489-c62f-4481-87b7-e74242aeb8ca] Refreshing instance network info cache due to event network-changed-d5a805f5-b6a8-42ec-92a0-4a652b51c1e6. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 974.545012] env[65726]: DEBUG oslo_concurrency.lockutils [req-0172e8c1-1b40-4ad7-a064-b16dd49d8418 req-17239211-ffcc-474f-88fe-b08f65340fc4 service nova] Acquiring lock "refresh_cache-aa6ce489-c62f-4481-87b7-e74242aeb8ca" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 974.545012] env[65726]: DEBUG oslo_concurrency.lockutils [req-0172e8c1-1b40-4ad7-a064-b16dd49d8418 req-17239211-ffcc-474f-88fe-b08f65340fc4 service nova] Acquired lock "refresh_cache-aa6ce489-c62f-4481-87b7-e74242aeb8ca" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 974.545012] env[65726]: DEBUG nova.network.neutron [req-0172e8c1-1b40-4ad7-a064-b16dd49d8418 req-17239211-ffcc-474f-88fe-b08f65340fc4 service nova] [instance: aa6ce489-c62f-4481-87b7-e74242aeb8ca] Refreshing network info cache for port d5a805f5-b6a8-42ec-92a0-4a652b51c1e6 {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 974.559023] env[65726]: INFO nova.compute.manager [-] [instance: 3088ee5e-0d27-4058-a94b-2e04c3b52add] Took 1.55 seconds to deallocate network for instance. [ 974.634661] env[65726]: DEBUG oslo_vmware.api [None req-8940e544-9cc2-4a18-9dd0-e7f396520ac6 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Task: {'id': task-5116371, 'name': CloneVM_Task} progress is 94%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.671724] env[65726]: DEBUG oslo_concurrency.lockutils [None req-eb838c27-51fd-42ee-bf3e-5ee447005ae8 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Acquiring lock "interface-dcde25df-87a5-47fb-94ce-334f68894e04-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 974.673136] env[65726]: DEBUG oslo_concurrency.lockutils [None req-eb838c27-51fd-42ee-bf3e-5ee447005ae8 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Lock "interface-dcde25df-87a5-47fb-94ce-334f68894e04-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.001s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 974.673556] env[65726]: DEBUG nova.objects.instance [None req-eb838c27-51fd-42ee-bf3e-5ee447005ae8 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Lazy-loading 'flavor' on Instance uuid dcde25df-87a5-47fb-94ce-334f68894e04 {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 974.693716] env[65726]: DEBUG nova.network.neutron [None req-d903724e-08f6-4863-a079-5fadb490f6c5 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 54e85161-5f63-405f-83f8-490c70645a3b] Successfully updated port: 8adb72c3-8e2f-48d7-8197-301c8958dbfb {{(pid=65726) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 974.803024] env[65726]: DEBUG oslo_vmware.api [None req-de3a85d2-cc71-4c75-a1b0-fe6db6e57022 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] Task: {'id': task-5116373, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.884589] env[65726]: DEBUG oslo_vmware.api [None req-4d7e3d4f-40bd-4bd5-8d6c-4c6d705b54c9 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Task: {'id': task-5116374, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.163143} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 974.884780] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-4d7e3d4f-40bd-4bd5-8d6c-4c6d705b54c9 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Deleted the datastore file {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 974.884967] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-4d7e3d4f-40bd-4bd5-8d6c-4c6d705b54c9 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] [instance: 29823bc7-3909-4ab7-8119-91ee59b289c5] Deleted contents of the VM from datastore datastore2 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 974.885160] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-4d7e3d4f-40bd-4bd5-8d6c-4c6d705b54c9 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] [instance: 29823bc7-3909-4ab7-8119-91ee59b289c5] Instance destroyed {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 974.885331] env[65726]: INFO nova.compute.manager [None req-4d7e3d4f-40bd-4bd5-8d6c-4c6d705b54c9 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] [instance: 29823bc7-3909-4ab7-8119-91ee59b289c5] Took 1.35 seconds to destroy the instance on the hypervisor. [ 974.885582] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-4d7e3d4f-40bd-4bd5-8d6c-4c6d705b54c9 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 974.885784] env[65726]: DEBUG nova.compute.manager [-] [instance: 29823bc7-3909-4ab7-8119-91ee59b289c5] Deallocating network for instance {{(pid=65726) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 974.885873] env[65726]: DEBUG nova.network.neutron [-] [instance: 29823bc7-3909-4ab7-8119-91ee59b289c5] deallocate_for_instance() {{(pid=65726) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 974.886420] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 974.886687] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 974.909124] env[65726]: DEBUG nova.scheduler.client.report [None req-c4955188-25cc-47e9-912c-88328888d7f1 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Updated inventory for provider 07c4692f-bdb4-4058-9173-ff9664830295 with generation 116 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 974.909536] env[65726]: DEBUG nova.compute.provider_tree [None req-c4955188-25cc-47e9-912c-88328888d7f1 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Updating resource provider 07c4692f-bdb4-4058-9173-ff9664830295 generation from 116 to 117 during operation: update_inventory {{(pid=65726) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 974.909726] env[65726]: DEBUG nova.compute.provider_tree [None req-c4955188-25cc-47e9-912c-88328888d7f1 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Updating inventory in ProviderTree for provider 07c4692f-bdb4-4058-9173-ff9664830295 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 974.929231] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 974.929491] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 974.973835] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116377, 'name': CreateVM_Task} progress is 99%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.050721] env[65726]: WARNING openstack [req-0172e8c1-1b40-4ad7-a064-b16dd49d8418 req-17239211-ffcc-474f-88fe-b08f65340fc4 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 975.051132] env[65726]: WARNING openstack [req-0172e8c1-1b40-4ad7-a064-b16dd49d8418 req-17239211-ffcc-474f-88fe-b08f65340fc4 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 975.066519] env[65726]: DEBUG oslo_concurrency.lockutils [None req-7ab7520f-2b6e-4be5-84c4-f04aa6f3af4f tempest-ServersTestManualDisk-1753452859 tempest-ServersTestManualDisk-1753452859-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 975.136093] env[65726]: DEBUG oslo_vmware.api [None req-8940e544-9cc2-4a18-9dd0-e7f396520ac6 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Task: {'id': task-5116371, 'name': CloneVM_Task} progress is 94%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.176752] env[65726]: WARNING openstack [req-0172e8c1-1b40-4ad7-a064-b16dd49d8418 req-17239211-ffcc-474f-88fe-b08f65340fc4 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 975.177145] env[65726]: WARNING openstack [req-0172e8c1-1b40-4ad7-a064-b16dd49d8418 req-17239211-ffcc-474f-88fe-b08f65340fc4 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 975.186024] env[65726]: WARNING openstack [None req-eb838c27-51fd-42ee-bf3e-5ee447005ae8 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 975.186024] env[65726]: WARNING openstack [None req-eb838c27-51fd-42ee-bf3e-5ee447005ae8 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 975.190483] env[65726]: DEBUG nova.objects.instance [None req-eb838c27-51fd-42ee-bf3e-5ee447005ae8 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Lazy-loading 'pci_requests' on Instance uuid dcde25df-87a5-47fb-94ce-334f68894e04 {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 975.195260] env[65726]: DEBUG oslo_concurrency.lockutils [None req-d903724e-08f6-4863-a079-5fadb490f6c5 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Acquiring lock "refresh_cache-54e85161-5f63-405f-83f8-490c70645a3b" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 975.195260] env[65726]: DEBUG oslo_concurrency.lockutils [None req-d903724e-08f6-4863-a079-5fadb490f6c5 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Acquired lock "refresh_cache-54e85161-5f63-405f-83f8-490c70645a3b" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 975.195440] env[65726]: DEBUG nova.network.neutron [None req-d903724e-08f6-4863-a079-5fadb490f6c5 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 54e85161-5f63-405f-83f8-490c70645a3b] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 975.254069] env[65726]: WARNING openstack [req-0172e8c1-1b40-4ad7-a064-b16dd49d8418 req-17239211-ffcc-474f-88fe-b08f65340fc4 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 975.254564] env[65726]: WARNING openstack [req-0172e8c1-1b40-4ad7-a064-b16dd49d8418 req-17239211-ffcc-474f-88fe-b08f65340fc4 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 975.306914] env[65726]: DEBUG oslo_vmware.api [None req-de3a85d2-cc71-4c75-a1b0-fe6db6e57022 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] Task: {'id': task-5116373, 'name': ReconfigVM_Task, 'duration_secs': 0.659021} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 975.307640] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-de3a85d2-cc71-4c75-a1b0-fe6db6e57022 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] [instance: 73f5b23e-4463-4068-8994-dd2752a2abc9] Reconfigured VM instance instance-0000004d to attach disk [datastore1] 73f5b23e-4463-4068-8994-dd2752a2abc9/73f5b23e-4463-4068-8994-dd2752a2abc9.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 975.309008] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4125df1f-788c-41db-adbd-001062b2ac76 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.326912] env[65726]: DEBUG oslo_vmware.api [None req-de3a85d2-cc71-4c75-a1b0-fe6db6e57022 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] Waiting for the task: (returnval){ [ 975.326912] env[65726]: value = "task-5116378" [ 975.326912] env[65726]: _type = "Task" [ 975.326912] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 975.336449] env[65726]: DEBUG oslo_vmware.api [None req-de3a85d2-cc71-4c75-a1b0-fe6db6e57022 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] Task: {'id': task-5116378, 'name': Rename_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.359456] env[65726]: DEBUG nova.network.neutron [req-0172e8c1-1b40-4ad7-a064-b16dd49d8418 req-17239211-ffcc-474f-88fe-b08f65340fc4 service nova] [instance: aa6ce489-c62f-4481-87b7-e74242aeb8ca] Updated VIF entry in instance network info cache for port d5a805f5-b6a8-42ec-92a0-4a652b51c1e6. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 975.359837] env[65726]: DEBUG nova.network.neutron [req-0172e8c1-1b40-4ad7-a064-b16dd49d8418 req-17239211-ffcc-474f-88fe-b08f65340fc4 service nova] [instance: aa6ce489-c62f-4481-87b7-e74242aeb8ca] Updating instance_info_cache with network_info: [{"id": "d5a805f5-b6a8-42ec-92a0-4a652b51c1e6", "address": "fa:16:3e:f5:d5:d7", "network": {"id": "547d7b16-4591-4d80-9e9b-c8a9a3a78dab", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-2077874536-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4ffd45f4a7a041199a4fc7f69f5e7e9b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2ed91b7b-b4ec-486d-ab34-af0afb7ec691", "external-id": "nsx-vlan-transportzone-75", "segmentation_id": 75, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd5a805f5-b6", "ovs_interfaceid": "d5a805f5-b6a8-42ec-92a0-4a652b51c1e6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 975.414968] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c4955188-25cc-47e9-912c-88328888d7f1 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 3.319s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 975.417700] env[65726]: DEBUG oslo_concurrency.lockutils [None req-dd3cd0c9-6ca2-4870-ab72-9325931d98de tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 17.278s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 975.417931] env[65726]: DEBUG nova.objects.instance [None req-dd3cd0c9-6ca2-4870-ab72-9325931d98de tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Lazy-loading 'resources' on Instance uuid 274ab469-61a9-4b7e-852c-074c871e3abf {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 975.438251] env[65726]: INFO nova.scheduler.client.report [None req-c4955188-25cc-47e9-912c-88328888d7f1 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Deleted allocations for instance 01d86089-6b9b-4588-864e-ef91375a1eea [ 975.475618] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116377, 'name': CreateVM_Task, 'duration_secs': 0.535419} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 975.475825] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: aa6ce489-c62f-4481-87b7-e74242aeb8ca] Created VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 975.476593] env[65726]: WARNING openstack [None req-9f65b40f-615d-4bae-ac4c-c122fff748fd tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 975.476955] env[65726]: WARNING openstack [None req-9f65b40f-615d-4bae-ac4c-c122fff748fd tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 975.481982] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9f65b40f-615d-4bae-ac4c-c122fff748fd tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 975.482157] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9f65b40f-615d-4bae-ac4c-c122fff748fd tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 975.482503] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9f65b40f-615d-4bae-ac4c-c122fff748fd tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 975.482899] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ace5cdf5-f3a1-478b-9fcf-36269fa17790 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.488669] env[65726]: DEBUG oslo_vmware.api [None req-9f65b40f-615d-4bae-ac4c-c122fff748fd tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Waiting for the task: (returnval){ [ 975.488669] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5280a171-63e0-dc41-f5d4-3ad7c23b3a17" [ 975.488669] env[65726]: _type = "Task" [ 975.488669] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 975.499089] env[65726]: DEBUG oslo_vmware.api [None req-9f65b40f-615d-4bae-ac4c-c122fff748fd tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5280a171-63e0-dc41-f5d4-3ad7c23b3a17, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.621584] env[65726]: DEBUG nova.network.neutron [-] [instance: 29823bc7-3909-4ab7-8119-91ee59b289c5] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 975.634765] env[65726]: DEBUG oslo_vmware.api [None req-8940e544-9cc2-4a18-9dd0-e7f396520ac6 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Task: {'id': task-5116371, 'name': CloneVM_Task} progress is 95%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.694688] env[65726]: DEBUG nova.objects.base [None req-eb838c27-51fd-42ee-bf3e-5ee447005ae8 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Object Instance lazy-loaded attributes: flavor,pci_requests {{(pid=65726) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 975.694961] env[65726]: DEBUG nova.network.neutron [None req-eb838c27-51fd-42ee-bf3e-5ee447005ae8 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: dcde25df-87a5-47fb-94ce-334f68894e04] allocate_for_instance() {{(pid=65726) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 975.695324] env[65726]: WARNING neutronclient.v2_0.client [None req-eb838c27-51fd-42ee-bf3e-5ee447005ae8 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 975.695640] env[65726]: WARNING neutronclient.v2_0.client [None req-eb838c27-51fd-42ee-bf3e-5ee447005ae8 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 975.696263] env[65726]: WARNING openstack [None req-eb838c27-51fd-42ee-bf3e-5ee447005ae8 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 975.696632] env[65726]: WARNING openstack [None req-eb838c27-51fd-42ee-bf3e-5ee447005ae8 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 975.705242] env[65726]: WARNING openstack [None req-d903724e-08f6-4863-a079-5fadb490f6c5 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 975.705600] env[65726]: WARNING openstack [None req-d903724e-08f6-4863-a079-5fadb490f6c5 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 975.710196] env[65726]: DEBUG nova.network.neutron [None req-d903724e-08f6-4863-a079-5fadb490f6c5 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 54e85161-5f63-405f-83f8-490c70645a3b] Instance cache missing network info. {{(pid=65726) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 975.765447] env[65726]: WARNING openstack [None req-d903724e-08f6-4863-a079-5fadb490f6c5 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 975.765835] env[65726]: WARNING openstack [None req-d903724e-08f6-4863-a079-5fadb490f6c5 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 975.806275] env[65726]: DEBUG oslo_concurrency.lockutils [None req-eb838c27-51fd-42ee-bf3e-5ee447005ae8 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Lock "interface-dcde25df-87a5-47fb-94ce-334f68894e04-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 1.134s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 975.833028] env[65726]: WARNING openstack [None req-d903724e-08f6-4863-a079-5fadb490f6c5 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 975.833028] env[65726]: WARNING openstack [None req-d903724e-08f6-4863-a079-5fadb490f6c5 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 975.848599] env[65726]: DEBUG oslo_vmware.api [None req-de3a85d2-cc71-4c75-a1b0-fe6db6e57022 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] Task: {'id': task-5116378, 'name': Rename_Task, 'duration_secs': 0.264949} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 975.849822] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-de3a85d2-cc71-4c75-a1b0-fe6db6e57022 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] [instance: 73f5b23e-4463-4068-8994-dd2752a2abc9] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 975.850863] env[65726]: DEBUG nova.compute.manager [req-92d10cec-2c19-43ac-aca3-d24d910c977a req-3977ca4f-a7c5-4b80-951c-e19403c49a40 service nova] [instance: 29823bc7-3909-4ab7-8119-91ee59b289c5] Received event network-vif-deleted-ef57c09b-e512-44ad-adeb-8aee81e9dc7b {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 975.851166] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-061a160b-4b0d-4754-bbc2-047fd734613d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.858287] env[65726]: DEBUG oslo_vmware.api [None req-de3a85d2-cc71-4c75-a1b0-fe6db6e57022 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] Waiting for the task: (returnval){ [ 975.858287] env[65726]: value = "task-5116379" [ 975.858287] env[65726]: _type = "Task" [ 975.858287] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 975.862261] env[65726]: DEBUG oslo_concurrency.lockutils [req-0172e8c1-1b40-4ad7-a064-b16dd49d8418 req-17239211-ffcc-474f-88fe-b08f65340fc4 service nova] Releasing lock "refresh_cache-aa6ce489-c62f-4481-87b7-e74242aeb8ca" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 975.862748] env[65726]: DEBUG nova.compute.manager [req-0172e8c1-1b40-4ad7-a064-b16dd49d8418 req-17239211-ffcc-474f-88fe-b08f65340fc4 service nova] [instance: df46141c-b545-4e03-b3a3-fd9f5feda0d2] Received event network-vif-deleted-cddd4b3f-5986-4368-8626-1bcb77f31c8e {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 975.867952] env[65726]: DEBUG oslo_vmware.api [None req-de3a85d2-cc71-4c75-a1b0-fe6db6e57022 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] Task: {'id': task-5116379, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.929946] env[65726]: DEBUG nova.network.neutron [None req-d903724e-08f6-4863-a079-5fadb490f6c5 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 54e85161-5f63-405f-83f8-490c70645a3b] Updating instance_info_cache with network_info: [{"id": "8adb72c3-8e2f-48d7-8197-301c8958dbfb", "address": "fa:16:3e:6a:5f:09", "network": {"id": "e206b5b9-da36-4e6f-926f-914dec38fac2", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1053201655-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "39067912d091464192516c08eb576696", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7c80243e-93a7-4a95-bc8d-e9534bacd66e", "external-id": "nsx-vlan-transportzone-306", "segmentation_id": 306, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8adb72c3-8e", "ovs_interfaceid": "8adb72c3-8e2f-48d7-8197-301c8958dbfb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 975.949854] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c4955188-25cc-47e9-912c-88328888d7f1 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Lock "01d86089-6b9b-4588-864e-ef91375a1eea" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 22.642s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 976.002996] env[65726]: DEBUG oslo_vmware.api [None req-9f65b40f-615d-4bae-ac4c-c122fff748fd tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5280a171-63e0-dc41-f5d4-3ad7c23b3a17, 'name': SearchDatastore_Task, 'duration_secs': 0.011252} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 976.003304] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9f65b40f-615d-4bae-ac4c-c122fff748fd tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 976.003571] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-9f65b40f-615d-4bae-ac4c-c122fff748fd tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] [instance: aa6ce489-c62f-4481-87b7-e74242aeb8ca] Processing image b52362a3-ee8a-4cbf-b06f-513b0cc8f95c {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 976.003839] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9f65b40f-615d-4bae-ac4c-c122fff748fd tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 976.003985] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9f65b40f-615d-4bae-ac4c-c122fff748fd tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 976.004180] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-9f65b40f-615d-4bae-ac4c-c122fff748fd tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 976.004493] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e75a93c0-405b-43fc-b66a-7960d146ecb4 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.016860] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-9f65b40f-615d-4bae-ac4c-c122fff748fd tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 976.018371] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-9f65b40f-615d-4bae-ac4c-c122fff748fd tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65726) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 976.018371] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-12d5e4a6-a294-4bd6-8285-a567a8a134e1 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.027607] env[65726]: DEBUG oslo_vmware.api [None req-9f65b40f-615d-4bae-ac4c-c122fff748fd tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Waiting for the task: (returnval){ [ 976.027607] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52809680-34a5-1fdc-259a-c41bc32d18e7" [ 976.027607] env[65726]: _type = "Task" [ 976.027607] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 976.036612] env[65726]: DEBUG oslo_vmware.api [None req-9f65b40f-615d-4bae-ac4c-c122fff748fd tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52809680-34a5-1fdc-259a-c41bc32d18e7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.124793] env[65726]: INFO nova.compute.manager [-] [instance: 29823bc7-3909-4ab7-8119-91ee59b289c5] Took 1.24 seconds to deallocate network for instance. [ 976.139910] env[65726]: DEBUG oslo_vmware.api [None req-8940e544-9cc2-4a18-9dd0-e7f396520ac6 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Task: {'id': task-5116371, 'name': CloneVM_Task, 'duration_secs': 1.646699} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 976.140192] env[65726]: INFO nova.virt.vmwareapi.vmops [None req-8940e544-9cc2-4a18-9dd0-e7f396520ac6 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04] Created linked-clone VM from snapshot [ 976.140973] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ddb181ea-70b3-4e1c-9ef9-1b2f038860dd {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.149759] env[65726]: DEBUG nova.virt.vmwareapi.images [None req-8940e544-9cc2-4a18-9dd0-e7f396520ac6 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04] Uploading image 81f5ce39-0bc0-4d0e-ab15-42a80a4f614b {{(pid=65726) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 976.177567] env[65726]: DEBUG oslo_vmware.rw_handles [None req-8940e544-9cc2-4a18-9dd0-e7f396520ac6 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 976.177567] env[65726]: value = "vm-995236" [ 976.177567] env[65726]: _type = "VirtualMachine" [ 976.177567] env[65726]: }. {{(pid=65726) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 976.177943] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-4847b564-5bd8-4ff5-91f8-b000bae51846 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.188366] env[65726]: DEBUG oslo_vmware.rw_handles [None req-8940e544-9cc2-4a18-9dd0-e7f396520ac6 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Lease: (returnval){ [ 976.188366] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52da0b24-195e-d267-1a62-2d53525a5bc4" [ 976.188366] env[65726]: _type = "HttpNfcLease" [ 976.188366] env[65726]: } obtained for exporting VM: (result){ [ 976.188366] env[65726]: value = "vm-995236" [ 976.188366] env[65726]: _type = "VirtualMachine" [ 976.188366] env[65726]: }. {{(pid=65726) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 976.188723] env[65726]: DEBUG oslo_vmware.api [None req-8940e544-9cc2-4a18-9dd0-e7f396520ac6 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Waiting for the lease: (returnval){ [ 976.188723] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52da0b24-195e-d267-1a62-2d53525a5bc4" [ 976.188723] env[65726]: _type = "HttpNfcLease" [ 976.188723] env[65726]: } to be ready. {{(pid=65726) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 976.199108] env[65726]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 976.199108] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52da0b24-195e-d267-1a62-2d53525a5bc4" [ 976.199108] env[65726]: _type = "HttpNfcLease" [ 976.199108] env[65726]: } is initializing. {{(pid=65726) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 976.256263] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9914c0e0-cc52-465f-a094-4aab1d4f5568 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.268076] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e78e6bc9-60ff-44bf-ac91-e47f9e488d3b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.307266] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b9751a5-113b-466d-9317-ae3997ccc24f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.318166] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89eb7867-d3de-4392-936f-cdfc06cb5579 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.335199] env[65726]: DEBUG nova.compute.provider_tree [None req-dd3cd0c9-6ca2-4870-ab72-9325931d98de tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 976.369537] env[65726]: DEBUG oslo_vmware.api [None req-de3a85d2-cc71-4c75-a1b0-fe6db6e57022 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] Task: {'id': task-5116379, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.432246] env[65726]: DEBUG oslo_concurrency.lockutils [None req-d903724e-08f6-4863-a079-5fadb490f6c5 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Releasing lock "refresh_cache-54e85161-5f63-405f-83f8-490c70645a3b" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 976.432739] env[65726]: DEBUG nova.compute.manager [None req-d903724e-08f6-4863-a079-5fadb490f6c5 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 54e85161-5f63-405f-83f8-490c70645a3b] Instance network_info: |[{"id": "8adb72c3-8e2f-48d7-8197-301c8958dbfb", "address": "fa:16:3e:6a:5f:09", "network": {"id": "e206b5b9-da36-4e6f-926f-914dec38fac2", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1053201655-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "39067912d091464192516c08eb576696", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7c80243e-93a7-4a95-bc8d-e9534bacd66e", "external-id": "nsx-vlan-transportzone-306", "segmentation_id": 306, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8adb72c3-8e", "ovs_interfaceid": "8adb72c3-8e2f-48d7-8197-301c8958dbfb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 976.433277] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-d903724e-08f6-4863-a079-5fadb490f6c5 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 54e85161-5f63-405f-83f8-490c70645a3b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:6a:5f:09', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '7c80243e-93a7-4a95-bc8d-e9534bacd66e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '8adb72c3-8e2f-48d7-8197-301c8958dbfb', 'vif_model': 'vmxnet3'}] {{(pid=65726) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 976.441438] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-d903724e-08f6-4863-a079-5fadb490f6c5 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 976.441697] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 54e85161-5f63-405f-83f8-490c70645a3b] Creating VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 976.441963] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-138149f9-ce73-4a19-a7f6-6f24ac374abb {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.465366] env[65726]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 976.465366] env[65726]: value = "task-5116381" [ 976.465366] env[65726]: _type = "Task" [ 976.465366] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 976.475265] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116381, 'name': CreateVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.538773] env[65726]: DEBUG oslo_vmware.api [None req-9f65b40f-615d-4bae-ac4c-c122fff748fd tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52809680-34a5-1fdc-259a-c41bc32d18e7, 'name': SearchDatastore_Task, 'duration_secs': 0.010607} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 976.539721] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9a6e1edc-fa77-4c80-ad84-57d782b9a9bb {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.547275] env[65726]: DEBUG oslo_vmware.api [None req-9f65b40f-615d-4bae-ac4c-c122fff748fd tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Waiting for the task: (returnval){ [ 976.547275] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52b8d5ac-124f-edb7-4485-07ceaa243bcd" [ 976.547275] env[65726]: _type = "Task" [ 976.547275] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 976.556959] env[65726]: DEBUG oslo_vmware.api [None req-9f65b40f-615d-4bae-ac4c-c122fff748fd tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52b8d5ac-124f-edb7-4485-07ceaa243bcd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.605540] env[65726]: DEBUG nova.compute.manager [req-aca68845-ef75-45b0-83bb-9f4c583b08a0 req-d75ce563-aa53-4fa4-b955-6b57b967b1d9 service nova] [instance: 54e85161-5f63-405f-83f8-490c70645a3b] Received event network-vif-plugged-8adb72c3-8e2f-48d7-8197-301c8958dbfb {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 976.605779] env[65726]: DEBUG oslo_concurrency.lockutils [req-aca68845-ef75-45b0-83bb-9f4c583b08a0 req-d75ce563-aa53-4fa4-b955-6b57b967b1d9 service nova] Acquiring lock "54e85161-5f63-405f-83f8-490c70645a3b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 976.605989] env[65726]: DEBUG oslo_concurrency.lockutils [req-aca68845-ef75-45b0-83bb-9f4c583b08a0 req-d75ce563-aa53-4fa4-b955-6b57b967b1d9 service nova] Lock "54e85161-5f63-405f-83f8-490c70645a3b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 976.606248] env[65726]: DEBUG oslo_concurrency.lockutils [req-aca68845-ef75-45b0-83bb-9f4c583b08a0 req-d75ce563-aa53-4fa4-b955-6b57b967b1d9 service nova] Lock "54e85161-5f63-405f-83f8-490c70645a3b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 976.606432] env[65726]: DEBUG nova.compute.manager [req-aca68845-ef75-45b0-83bb-9f4c583b08a0 req-d75ce563-aa53-4fa4-b955-6b57b967b1d9 service nova] [instance: 54e85161-5f63-405f-83f8-490c70645a3b] No waiting events found dispatching network-vif-plugged-8adb72c3-8e2f-48d7-8197-301c8958dbfb {{(pid=65726) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 976.606615] env[65726]: WARNING nova.compute.manager [req-aca68845-ef75-45b0-83bb-9f4c583b08a0 req-d75ce563-aa53-4fa4-b955-6b57b967b1d9 service nova] [instance: 54e85161-5f63-405f-83f8-490c70645a3b] Received unexpected event network-vif-plugged-8adb72c3-8e2f-48d7-8197-301c8958dbfb for instance with vm_state building and task_state spawning. [ 976.606776] env[65726]: DEBUG nova.compute.manager [req-aca68845-ef75-45b0-83bb-9f4c583b08a0 req-d75ce563-aa53-4fa4-b955-6b57b967b1d9 service nova] [instance: 54e85161-5f63-405f-83f8-490c70645a3b] Received event network-changed-8adb72c3-8e2f-48d7-8197-301c8958dbfb {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 976.607240] env[65726]: DEBUG nova.compute.manager [req-aca68845-ef75-45b0-83bb-9f4c583b08a0 req-d75ce563-aa53-4fa4-b955-6b57b967b1d9 service nova] [instance: 54e85161-5f63-405f-83f8-490c70645a3b] Refreshing instance network info cache due to event network-changed-8adb72c3-8e2f-48d7-8197-301c8958dbfb. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 976.607240] env[65726]: DEBUG oslo_concurrency.lockutils [req-aca68845-ef75-45b0-83bb-9f4c583b08a0 req-d75ce563-aa53-4fa4-b955-6b57b967b1d9 service nova] Acquiring lock "refresh_cache-54e85161-5f63-405f-83f8-490c70645a3b" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 976.607408] env[65726]: DEBUG oslo_concurrency.lockutils [req-aca68845-ef75-45b0-83bb-9f4c583b08a0 req-d75ce563-aa53-4fa4-b955-6b57b967b1d9 service nova] Acquired lock "refresh_cache-54e85161-5f63-405f-83f8-490c70645a3b" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 976.607458] env[65726]: DEBUG nova.network.neutron [req-aca68845-ef75-45b0-83bb-9f4c583b08a0 req-d75ce563-aa53-4fa4-b955-6b57b967b1d9 service nova] [instance: 54e85161-5f63-405f-83f8-490c70645a3b] Refreshing network info cache for port 8adb72c3-8e2f-48d7-8197-301c8958dbfb {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 976.634076] env[65726]: DEBUG oslo_concurrency.lockutils [None req-4d7e3d4f-40bd-4bd5-8d6c-4c6d705b54c9 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 976.697436] env[65726]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 976.697436] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52da0b24-195e-d267-1a62-2d53525a5bc4" [ 976.697436] env[65726]: _type = "HttpNfcLease" [ 976.697436] env[65726]: } is ready. {{(pid=65726) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 976.697781] env[65726]: DEBUG oslo_vmware.rw_handles [None req-8940e544-9cc2-4a18-9dd0-e7f396520ac6 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 976.697781] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52da0b24-195e-d267-1a62-2d53525a5bc4" [ 976.697781] env[65726]: _type = "HttpNfcLease" [ 976.697781] env[65726]: }. {{(pid=65726) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 976.698554] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8da88257-679b-4e13-bab0-8f4291930e5c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.706389] env[65726]: DEBUG oslo_vmware.rw_handles [None req-8940e544-9cc2-4a18-9dd0-e7f396520ac6 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52e1447a-555d-c02a-cc2d-96e5d4ae2965/disk-0.vmdk from lease info. {{(pid=65726) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 976.706590] env[65726]: DEBUG oslo_vmware.rw_handles [None req-8940e544-9cc2-4a18-9dd0-e7f396520ac6 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52e1447a-555d-c02a-cc2d-96e5d4ae2965/disk-0.vmdk for reading. {{(pid=65726) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 976.801421] env[65726]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-cab9bdcb-a294-489f-86cc-800356a1eda8 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.841982] env[65726]: DEBUG nova.scheduler.client.report [None req-dd3cd0c9-6ca2-4870-ab72-9325931d98de tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 976.870459] env[65726]: DEBUG oslo_vmware.api [None req-de3a85d2-cc71-4c75-a1b0-fe6db6e57022 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] Task: {'id': task-5116379, 'name': PowerOnVM_Task, 'duration_secs': 0.790178} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 976.870732] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-de3a85d2-cc71-4c75-a1b0-fe6db6e57022 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] [instance: 73f5b23e-4463-4068-8994-dd2752a2abc9] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 976.870951] env[65726]: DEBUG nova.compute.manager [None req-de3a85d2-cc71-4c75-a1b0-fe6db6e57022 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] [instance: 73f5b23e-4463-4068-8994-dd2752a2abc9] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 976.871793] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6a804af-4120-48c7-9926-48092f397a0d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.982635] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116381, 'name': CreateVM_Task, 'duration_secs': 0.372291} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 976.982869] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 54e85161-5f63-405f-83f8-490c70645a3b] Created VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 976.983824] env[65726]: WARNING openstack [None req-d903724e-08f6-4863-a079-5fadb490f6c5 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 976.984995] env[65726]: WARNING openstack [None req-d903724e-08f6-4863-a079-5fadb490f6c5 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 976.989644] env[65726]: DEBUG oslo_concurrency.lockutils [None req-d903724e-08f6-4863-a079-5fadb490f6c5 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 976.989820] env[65726]: DEBUG oslo_concurrency.lockutils [None req-d903724e-08f6-4863-a079-5fadb490f6c5 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 976.990170] env[65726]: DEBUG oslo_concurrency.lockutils [None req-d903724e-08f6-4863-a079-5fadb490f6c5 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 976.991050] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-91127935-cb9a-4de1-8bcd-743851e3c776 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.998673] env[65726]: DEBUG oslo_vmware.api [None req-d903724e-08f6-4863-a079-5fadb490f6c5 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Waiting for the task: (returnval){ [ 976.998673] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5233b83b-37c9-7d17-8a92-d9ea34bd601a" [ 976.998673] env[65726]: _type = "Task" [ 976.998673] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 977.007580] env[65726]: DEBUG oslo_vmware.api [None req-d903724e-08f6-4863-a079-5fadb490f6c5 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5233b83b-37c9-7d17-8a92-d9ea34bd601a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.061934] env[65726]: DEBUG oslo_vmware.api [None req-9f65b40f-615d-4bae-ac4c-c122fff748fd tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52b8d5ac-124f-edb7-4485-07ceaa243bcd, 'name': SearchDatastore_Task, 'duration_secs': 0.01616} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 977.062445] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9f65b40f-615d-4bae-ac4c-c122fff748fd tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 977.062846] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-9f65b40f-615d-4bae-ac4c-c122fff748fd tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] aa6ce489-c62f-4481-87b7-e74242aeb8ca/aa6ce489-c62f-4481-87b7-e74242aeb8ca.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 977.063292] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-39ffcafb-c93f-4164-a276-5ba7ca64edcf {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.072082] env[65726]: DEBUG oslo_vmware.api [None req-9f65b40f-615d-4bae-ac4c-c122fff748fd tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Waiting for the task: (returnval){ [ 977.072082] env[65726]: value = "task-5116382" [ 977.072082] env[65726]: _type = "Task" [ 977.072082] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 977.081647] env[65726]: DEBUG oslo_vmware.api [None req-9f65b40f-615d-4bae-ac4c-c122fff748fd tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Task: {'id': task-5116382, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.112941] env[65726]: WARNING openstack [req-aca68845-ef75-45b0-83bb-9f4c583b08a0 req-d75ce563-aa53-4fa4-b955-6b57b967b1d9 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 977.113432] env[65726]: WARNING openstack [req-aca68845-ef75-45b0-83bb-9f4c583b08a0 req-d75ce563-aa53-4fa4-b955-6b57b967b1d9 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 977.254745] env[65726]: WARNING openstack [req-aca68845-ef75-45b0-83bb-9f4c583b08a0 req-d75ce563-aa53-4fa4-b955-6b57b967b1d9 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 977.255622] env[65726]: WARNING openstack [req-aca68845-ef75-45b0-83bb-9f4c583b08a0 req-d75ce563-aa53-4fa4-b955-6b57b967b1d9 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 977.347245] env[65726]: DEBUG oslo_concurrency.lockutils [None req-dd3cd0c9-6ca2-4870-ab72-9325931d98de tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.929s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 977.350768] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c75c9ed0-c9ca-4fd9-9695-87538dcf1f78 tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.548s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 977.353582] env[65726]: INFO nova.compute.claims [None req-c75c9ed0-c9ca-4fd9-9695-87538dcf1f78 tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] [instance: 406c59e7-8eb8-4cfd-8e1e-e151ab908571] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 977.378011] env[65726]: WARNING openstack [req-aca68845-ef75-45b0-83bb-9f4c583b08a0 req-d75ce563-aa53-4fa4-b955-6b57b967b1d9 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 977.378583] env[65726]: WARNING openstack [req-aca68845-ef75-45b0-83bb-9f4c583b08a0 req-d75ce563-aa53-4fa4-b955-6b57b967b1d9 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 977.393065] env[65726]: INFO nova.scheduler.client.report [None req-dd3cd0c9-6ca2-4870-ab72-9325931d98de tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Deleted allocations for instance 274ab469-61a9-4b7e-852c-074c871e3abf [ 977.402066] env[65726]: DEBUG oslo_concurrency.lockutils [None req-de3a85d2-cc71-4c75-a1b0-fe6db6e57022 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 977.516623] env[65726]: DEBUG nova.network.neutron [req-aca68845-ef75-45b0-83bb-9f4c583b08a0 req-d75ce563-aa53-4fa4-b955-6b57b967b1d9 service nova] [instance: 54e85161-5f63-405f-83f8-490c70645a3b] Updated VIF entry in instance network info cache for port 8adb72c3-8e2f-48d7-8197-301c8958dbfb. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 977.518367] env[65726]: DEBUG nova.network.neutron [req-aca68845-ef75-45b0-83bb-9f4c583b08a0 req-d75ce563-aa53-4fa4-b955-6b57b967b1d9 service nova] [instance: 54e85161-5f63-405f-83f8-490c70645a3b] Updating instance_info_cache with network_info: [{"id": "8adb72c3-8e2f-48d7-8197-301c8958dbfb", "address": "fa:16:3e:6a:5f:09", "network": {"id": "e206b5b9-da36-4e6f-926f-914dec38fac2", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1053201655-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "39067912d091464192516c08eb576696", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7c80243e-93a7-4a95-bc8d-e9534bacd66e", "external-id": "nsx-vlan-transportzone-306", "segmentation_id": 306, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8adb72c3-8e", "ovs_interfaceid": "8adb72c3-8e2f-48d7-8197-301c8958dbfb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 977.525904] env[65726]: DEBUG oslo_vmware.api [None req-d903724e-08f6-4863-a079-5fadb490f6c5 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5233b83b-37c9-7d17-8a92-d9ea34bd601a, 'name': SearchDatastore_Task, 'duration_secs': 0.010954} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 977.528541] env[65726]: DEBUG oslo_concurrency.lockutils [None req-d903724e-08f6-4863-a079-5fadb490f6c5 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 977.528836] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-d903724e-08f6-4863-a079-5fadb490f6c5 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 54e85161-5f63-405f-83f8-490c70645a3b] Processing image b52362a3-ee8a-4cbf-b06f-513b0cc8f95c {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 977.529133] env[65726]: DEBUG oslo_concurrency.lockutils [None req-d903724e-08f6-4863-a079-5fadb490f6c5 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 977.529257] env[65726]: DEBUG oslo_concurrency.lockutils [None req-d903724e-08f6-4863-a079-5fadb490f6c5 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 977.529463] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-d903724e-08f6-4863-a079-5fadb490f6c5 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 977.529779] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-219dcb88-4e1a-47ce-96a1-1238bf426906 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.552237] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-d903724e-08f6-4863-a079-5fadb490f6c5 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 977.552714] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-d903724e-08f6-4863-a079-5fadb490f6c5 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65726) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 977.553648] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-92d14bb3-a238-4cd4-956b-1f1f8702687b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.562806] env[65726]: DEBUG oslo_vmware.api [None req-d903724e-08f6-4863-a079-5fadb490f6c5 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Waiting for the task: (returnval){ [ 977.562806] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52fb2334-42fe-2df9-79d4-ff2cb3a8838b" [ 977.562806] env[65726]: _type = "Task" [ 977.562806] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 977.576459] env[65726]: DEBUG oslo_vmware.api [None req-d903724e-08f6-4863-a079-5fadb490f6c5 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52fb2334-42fe-2df9-79d4-ff2cb3a8838b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.587041] env[65726]: DEBUG oslo_vmware.api [None req-9f65b40f-615d-4bae-ac4c-c122fff748fd tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Task: {'id': task-5116382, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.707805] env[65726]: DEBUG oslo_concurrency.lockutils [None req-ce025067-bba1-472f-842e-5d842d2328d6 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Acquiring lock "interface-dcde25df-87a5-47fb-94ce-334f68894e04-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 977.707805] env[65726]: DEBUG oslo_concurrency.lockutils [None req-ce025067-bba1-472f-842e-5d842d2328d6 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Lock "interface-dcde25df-87a5-47fb-94ce-334f68894e04-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 977.707805] env[65726]: DEBUG nova.objects.instance [None req-ce025067-bba1-472f-842e-5d842d2328d6 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Lazy-loading 'flavor' on Instance uuid dcde25df-87a5-47fb-94ce-334f68894e04 {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 977.904777] env[65726]: DEBUG oslo_concurrency.lockutils [None req-dd3cd0c9-6ca2-4870-ab72-9325931d98de tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Lock "274ab469-61a9-4b7e-852c-074c871e3abf" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 23.307s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 978.028826] env[65726]: DEBUG oslo_concurrency.lockutils [req-aca68845-ef75-45b0-83bb-9f4c583b08a0 req-d75ce563-aa53-4fa4-b955-6b57b967b1d9 service nova] Releasing lock "refresh_cache-54e85161-5f63-405f-83f8-490c70645a3b" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 978.077329] env[65726]: DEBUG oslo_vmware.api [None req-d903724e-08f6-4863-a079-5fadb490f6c5 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52fb2334-42fe-2df9-79d4-ff2cb3a8838b, 'name': SearchDatastore_Task, 'duration_secs': 0.027762} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 978.081638] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6053cc07-97f1-4206-afe8-0b17414437b7 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.089032] env[65726]: DEBUG oslo_vmware.api [None req-d903724e-08f6-4863-a079-5fadb490f6c5 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Waiting for the task: (returnval){ [ 978.089032] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52a0d011-8527-bd33-aadb-048f4523ada9" [ 978.089032] env[65726]: _type = "Task" [ 978.089032] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 978.092495] env[65726]: DEBUG oslo_vmware.api [None req-9f65b40f-615d-4bae-ac4c-c122fff748fd tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Task: {'id': task-5116382, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.532192} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 978.096056] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-9f65b40f-615d-4bae-ac4c-c122fff748fd tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] aa6ce489-c62f-4481-87b7-e74242aeb8ca/aa6ce489-c62f-4481-87b7-e74242aeb8ca.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 978.096399] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-9f65b40f-615d-4bae-ac4c-c122fff748fd tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] [instance: aa6ce489-c62f-4481-87b7-e74242aeb8ca] Extending root virtual disk to 1048576 {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 978.096752] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e34f4690-4732-45d2-80d4-96ef296e111b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.104689] env[65726]: DEBUG oslo_vmware.api [None req-d903724e-08f6-4863-a079-5fadb490f6c5 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52a0d011-8527-bd33-aadb-048f4523ada9, 'name': SearchDatastore_Task, 'duration_secs': 0.010618} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 978.106220] env[65726]: DEBUG oslo_concurrency.lockutils [None req-d903724e-08f6-4863-a079-5fadb490f6c5 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 978.106612] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-d903724e-08f6-4863-a079-5fadb490f6c5 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] 54e85161-5f63-405f-83f8-490c70645a3b/54e85161-5f63-405f-83f8-490c70645a3b.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 978.107047] env[65726]: DEBUG oslo_vmware.api [None req-9f65b40f-615d-4bae-ac4c-c122fff748fd tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Waiting for the task: (returnval){ [ 978.107047] env[65726]: value = "task-5116383" [ 978.107047] env[65726]: _type = "Task" [ 978.107047] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 978.107393] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-97702ac1-81bc-4178-b6af-99fe21f5f6d2 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.122030] env[65726]: DEBUG oslo_vmware.api [None req-9f65b40f-615d-4bae-ac4c-c122fff748fd tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Task: {'id': task-5116383, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.122030] env[65726]: DEBUG oslo_vmware.api [None req-d903724e-08f6-4863-a079-5fadb490f6c5 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Waiting for the task: (returnval){ [ 978.122030] env[65726]: value = "task-5116384" [ 978.122030] env[65726]: _type = "Task" [ 978.122030] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 978.131639] env[65726]: DEBUG oslo_vmware.api [None req-d903724e-08f6-4863-a079-5fadb490f6c5 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': task-5116384, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.211735] env[65726]: WARNING openstack [None req-ce025067-bba1-472f-842e-5d842d2328d6 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 978.211988] env[65726]: WARNING openstack [None req-ce025067-bba1-472f-842e-5d842d2328d6 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 978.333139] env[65726]: DEBUG nova.objects.instance [None req-ce025067-bba1-472f-842e-5d842d2328d6 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Lazy-loading 'pci_requests' on Instance uuid dcde25df-87a5-47fb-94ce-334f68894e04 {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 978.433231] env[65726]: DEBUG oslo_concurrency.lockutils [None req-378e66fe-caea-4dff-a5bd-614d0f24e5ba tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] Acquiring lock "73f5b23e-4463-4068-8994-dd2752a2abc9" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 978.433781] env[65726]: DEBUG oslo_concurrency.lockutils [None req-378e66fe-caea-4dff-a5bd-614d0f24e5ba tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] Lock "73f5b23e-4463-4068-8994-dd2752a2abc9" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 978.434140] env[65726]: DEBUG oslo_concurrency.lockutils [None req-378e66fe-caea-4dff-a5bd-614d0f24e5ba tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] Acquiring lock "73f5b23e-4463-4068-8994-dd2752a2abc9-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 978.435289] env[65726]: DEBUG oslo_concurrency.lockutils [None req-378e66fe-caea-4dff-a5bd-614d0f24e5ba tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] Lock "73f5b23e-4463-4068-8994-dd2752a2abc9-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 978.435488] env[65726]: DEBUG oslo_concurrency.lockutils [None req-378e66fe-caea-4dff-a5bd-614d0f24e5ba tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] Lock "73f5b23e-4463-4068-8994-dd2752a2abc9-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 978.438818] env[65726]: INFO nova.compute.manager [None req-378e66fe-caea-4dff-a5bd-614d0f24e5ba tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] [instance: 73f5b23e-4463-4068-8994-dd2752a2abc9] Terminating instance [ 978.625023] env[65726]: DEBUG oslo_vmware.api [None req-9f65b40f-615d-4bae-ac4c-c122fff748fd tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Task: {'id': task-5116383, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.06765} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 978.633986] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-9f65b40f-615d-4bae-ac4c-c122fff748fd tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] [instance: aa6ce489-c62f-4481-87b7-e74242aeb8ca] Extended root virtual disk {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 978.635463] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd5054a9-ae9f-4429-88ce-d3b392175eeb {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.645340] env[65726]: DEBUG oslo_vmware.api [None req-d903724e-08f6-4863-a079-5fadb490f6c5 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': task-5116384, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.667643] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-9f65b40f-615d-4bae-ac4c-c122fff748fd tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] [instance: aa6ce489-c62f-4481-87b7-e74242aeb8ca] Reconfiguring VM instance instance-0000004e to attach disk [datastore1] aa6ce489-c62f-4481-87b7-e74242aeb8ca/aa6ce489-c62f-4481-87b7-e74242aeb8ca.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 978.668655] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9d7646d9-8842-43ac-a3f5-6e3917ddb1ba {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.692434] env[65726]: DEBUG oslo_vmware.api [None req-9f65b40f-615d-4bae-ac4c-c122fff748fd tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Waiting for the task: (returnval){ [ 978.692434] env[65726]: value = "task-5116385" [ 978.692434] env[65726]: _type = "Task" [ 978.692434] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 978.703384] env[65726]: DEBUG oslo_vmware.api [None req-9f65b40f-615d-4bae-ac4c-c122fff748fd tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Task: {'id': task-5116385, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.707933] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c5289ae6-5b56-421a-a0e1-7e74e3494692 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Acquiring lock "f2d9090c-988f-43f4-9c81-7aa718a3438a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 978.708221] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c5289ae6-5b56-421a-a0e1-7e74e3494692 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Lock "f2d9090c-988f-43f4-9c81-7aa718a3438a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 978.708381] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c5289ae6-5b56-421a-a0e1-7e74e3494692 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Acquiring lock "f2d9090c-988f-43f4-9c81-7aa718a3438a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 978.708556] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c5289ae6-5b56-421a-a0e1-7e74e3494692 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Lock "f2d9090c-988f-43f4-9c81-7aa718a3438a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 978.708716] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c5289ae6-5b56-421a-a0e1-7e74e3494692 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Lock "f2d9090c-988f-43f4-9c81-7aa718a3438a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 978.711426] env[65726]: INFO nova.compute.manager [None req-c5289ae6-5b56-421a-a0e1-7e74e3494692 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: f2d9090c-988f-43f4-9c81-7aa718a3438a] Terminating instance [ 978.742780] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1805da80-2195-4d32-aeac-1a3130d38c90 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.754534] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-422c11d7-28e4-4309-b198-8062cbde2dc6 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.619269] env[65726]: DEBUG nova.objects.base [None req-ce025067-bba1-472f-842e-5d842d2328d6 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Object Instance lazy-loaded attributes: flavor,pci_requests {{(pid=65726) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 979.619481] env[65726]: DEBUG nova.network.neutron [None req-ce025067-bba1-472f-842e-5d842d2328d6 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: dcde25df-87a5-47fb-94ce-334f68894e04] allocate_for_instance() {{(pid=65726) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 979.620863] env[65726]: WARNING neutronclient.v2_0.client [None req-ce025067-bba1-472f-842e-5d842d2328d6 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 979.620863] env[65726]: WARNING neutronclient.v2_0.client [None req-ce025067-bba1-472f-842e-5d842d2328d6 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 979.620863] env[65726]: WARNING openstack [None req-ce025067-bba1-472f-842e-5d842d2328d6 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 979.621287] env[65726]: WARNING openstack [None req-ce025067-bba1-472f-842e-5d842d2328d6 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 979.629094] env[65726]: DEBUG oslo_concurrency.lockutils [None req-378e66fe-caea-4dff-a5bd-614d0f24e5ba tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] Acquiring lock "refresh_cache-73f5b23e-4463-4068-8994-dd2752a2abc9" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 979.629094] env[65726]: DEBUG oslo_concurrency.lockutils [None req-378e66fe-caea-4dff-a5bd-614d0f24e5ba tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] Acquired lock "refresh_cache-73f5b23e-4463-4068-8994-dd2752a2abc9" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 979.629094] env[65726]: DEBUG nova.network.neutron [None req-378e66fe-caea-4dff-a5bd-614d0f24e5ba tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] [instance: 73f5b23e-4463-4068-8994-dd2752a2abc9] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 979.633018] env[65726]: DEBUG nova.compute.manager [None req-c5289ae6-5b56-421a-a0e1-7e74e3494692 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: f2d9090c-988f-43f4-9c81-7aa718a3438a] Start destroying the instance on the hypervisor. {{(pid=65726) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 979.633018] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-c5289ae6-5b56-421a-a0e1-7e74e3494692 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: f2d9090c-988f-43f4-9c81-7aa718a3438a] Destroying instance {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 979.638354] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d190f5fd-86c8-4586-a0e7-93372d5a9476 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.641288] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e333d54-7505-4e76-9b25-6a9f6e0fdb37 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.653629] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-c5289ae6-5b56-421a-a0e1-7e74e3494692 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: f2d9090c-988f-43f4-9c81-7aa718a3438a] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 979.658808] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7291273e-1215-4d6a-a608-a6af680199c3 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.661129] env[65726]: DEBUG oslo_vmware.api [None req-9f65b40f-615d-4bae-ac4c-c122fff748fd tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Task: {'id': task-5116385, 'name': ReconfigVM_Task, 'duration_secs': 0.630692} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.661328] env[65726]: DEBUG oslo_vmware.api [None req-d903724e-08f6-4863-a079-5fadb490f6c5 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': task-5116384, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.539095} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.661675] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-9f65b40f-615d-4bae-ac4c-c122fff748fd tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] [instance: aa6ce489-c62f-4481-87b7-e74242aeb8ca] Reconfigured VM instance instance-0000004e to attach disk [datastore1] aa6ce489-c62f-4481-87b7-e74242aeb8ca/aa6ce489-c62f-4481-87b7-e74242aeb8ca.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 979.663386] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae8fda88-c2bf-4025-8469-d724adaf5a77 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.668031] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-d903724e-08f6-4863-a079-5fadb490f6c5 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] 54e85161-5f63-405f-83f8-490c70645a3b/54e85161-5f63-405f-83f8-490c70645a3b.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 979.668031] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-d903724e-08f6-4863-a079-5fadb490f6c5 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 54e85161-5f63-405f-83f8-490c70645a3b] Extending root virtual disk to 1048576 {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 979.669760] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a184b403-86d5-4ffe-b8f7-d17b3d12a071 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.670666] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-9d642889-b3b4-4726-9e5a-808c2904f39f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.674999] env[65726]: DEBUG oslo_vmware.api [None req-c5289ae6-5b56-421a-a0e1-7e74e3494692 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Waiting for the task: (returnval){ [ 979.674999] env[65726]: value = "task-5116386" [ 979.674999] env[65726]: _type = "Task" [ 979.674999] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 979.690543] env[65726]: DEBUG nova.compute.provider_tree [None req-c75c9ed0-c9ca-4fd9-9695-87538dcf1f78 tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Updating inventory in ProviderTree for provider 07c4692f-bdb4-4058-9173-ff9664830295 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 979.694167] env[65726]: DEBUG oslo_vmware.api [None req-9f65b40f-615d-4bae-ac4c-c122fff748fd tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Waiting for the task: (returnval){ [ 979.694167] env[65726]: value = "task-5116388" [ 979.694167] env[65726]: _type = "Task" [ 979.694167] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 979.694167] env[65726]: DEBUG oslo_vmware.api [None req-d903724e-08f6-4863-a079-5fadb490f6c5 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Waiting for the task: (returnval){ [ 979.694167] env[65726]: value = "task-5116387" [ 979.694167] env[65726]: _type = "Task" [ 979.694167] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 979.708162] env[65726]: DEBUG nova.policy [None req-ce025067-bba1-472f-842e-5d842d2328d6 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9b65b9194933469c908c7cc04478444b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd46ccb00794f458b85da4a93879139ee', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65726) authorize /opt/stack/nova/nova/policy.py:192}} [ 979.711548] env[65726]: DEBUG oslo_vmware.api [None req-c5289ae6-5b56-421a-a0e1-7e74e3494692 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Task: {'id': task-5116386, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.719741] env[65726]: DEBUG oslo_vmware.api [None req-9f65b40f-615d-4bae-ac4c-c122fff748fd tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Task: {'id': task-5116388, 'name': Rename_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.723011] env[65726]: DEBUG oslo_vmware.api [None req-d903724e-08f6-4863-a079-5fadb490f6c5 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': task-5116387, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.945601] env[65726]: DEBUG nova.network.neutron [None req-ce025067-bba1-472f-842e-5d842d2328d6 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: dcde25df-87a5-47fb-94ce-334f68894e04] Successfully created port: b48313f9-a0d3-4fdd-88c6-d103aef0b4f2 {{(pid=65726) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 980.133560] env[65726]: WARNING openstack [None req-378e66fe-caea-4dff-a5bd-614d0f24e5ba tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 980.134106] env[65726]: WARNING openstack [None req-378e66fe-caea-4dff-a5bd-614d0f24e5ba tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 980.139050] env[65726]: DEBUG nova.network.neutron [None req-378e66fe-caea-4dff-a5bd-614d0f24e5ba tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] [instance: 73f5b23e-4463-4068-8994-dd2752a2abc9] Instance cache missing network info. {{(pid=65726) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 980.185839] env[65726]: DEBUG oslo_vmware.api [None req-c5289ae6-5b56-421a-a0e1-7e74e3494692 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Task: {'id': task-5116386, 'name': PowerOffVM_Task, 'duration_secs': 0.477391} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 980.186121] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-c5289ae6-5b56-421a-a0e1-7e74e3494692 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: f2d9090c-988f-43f4-9c81-7aa718a3438a] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 980.186284] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-c5289ae6-5b56-421a-a0e1-7e74e3494692 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: f2d9090c-988f-43f4-9c81-7aa718a3438a] Unregistering the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 980.186537] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3d0901f6-3a8c-4507-87ef-9ed22dc6c869 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.218094] env[65726]: DEBUG oslo_vmware.api [None req-9f65b40f-615d-4bae-ac4c-c122fff748fd tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Task: {'id': task-5116388, 'name': Rename_Task, 'duration_secs': 0.311184} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 980.218263] env[65726]: DEBUG oslo_vmware.api [None req-d903724e-08f6-4863-a079-5fadb490f6c5 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': task-5116387, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.131245} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 980.218551] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-9f65b40f-615d-4bae-ac4c-c122fff748fd tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] [instance: aa6ce489-c62f-4481-87b7-e74242aeb8ca] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 980.218810] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-d903724e-08f6-4863-a079-5fadb490f6c5 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 54e85161-5f63-405f-83f8-490c70645a3b] Extended root virtual disk {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 980.219115] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1aa9a720-5c2c-4e94-874d-c457d1b059db {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.222044] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1386afd8-5a7d-4fee-a3ee-a4ec9acfcb66 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.229394] env[65726]: DEBUG nova.scheduler.client.report [None req-c75c9ed0-c9ca-4fd9-9695-87538dcf1f78 tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Updated inventory for provider 07c4692f-bdb4-4058-9173-ff9664830295 with generation 117 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 980.229642] env[65726]: DEBUG nova.compute.provider_tree [None req-c75c9ed0-c9ca-4fd9-9695-87538dcf1f78 tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Updating resource provider 07c4692f-bdb4-4058-9173-ff9664830295 generation from 117 to 118 during operation: update_inventory {{(pid=65726) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 980.229815] env[65726]: DEBUG nova.compute.provider_tree [None req-c75c9ed0-c9ca-4fd9-9695-87538dcf1f78 tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Updating inventory in ProviderTree for provider 07c4692f-bdb4-4058-9173-ff9664830295 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 980.256725] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-d903724e-08f6-4863-a079-5fadb490f6c5 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 54e85161-5f63-405f-83f8-490c70645a3b] Reconfiguring VM instance instance-0000004f to attach disk [datastore1] 54e85161-5f63-405f-83f8-490c70645a3b/54e85161-5f63-405f-83f8-490c70645a3b.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 980.257503] env[65726]: DEBUG oslo_vmware.api [None req-9f65b40f-615d-4bae-ac4c-c122fff748fd tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Waiting for the task: (returnval){ [ 980.257503] env[65726]: value = "task-5116390" [ 980.257503] env[65726]: _type = "Task" [ 980.257503] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 980.259493] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0cac6d8e-ac94-469f-ad05-630333f1c52b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.283704] env[65726]: DEBUG nova.network.neutron [None req-378e66fe-caea-4dff-a5bd-614d0f24e5ba tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] [instance: 73f5b23e-4463-4068-8994-dd2752a2abc9] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 980.291000] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-c5289ae6-5b56-421a-a0e1-7e74e3494692 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: f2d9090c-988f-43f4-9c81-7aa718a3438a] Unregistered the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 980.291379] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-c5289ae6-5b56-421a-a0e1-7e74e3494692 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: f2d9090c-988f-43f4-9c81-7aa718a3438a] Deleting contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 980.291630] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-c5289ae6-5b56-421a-a0e1-7e74e3494692 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Deleting the datastore file [datastore1] f2d9090c-988f-43f4-9c81-7aa718a3438a {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 980.292905] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-535f7073-9abd-4c14-ba99-dd1bed285517 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.298811] env[65726]: DEBUG oslo_vmware.api [None req-d903724e-08f6-4863-a079-5fadb490f6c5 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Waiting for the task: (returnval){ [ 980.298811] env[65726]: value = "task-5116391" [ 980.298811] env[65726]: _type = "Task" [ 980.298811] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 980.302909] env[65726]: DEBUG oslo_vmware.api [None req-9f65b40f-615d-4bae-ac4c-c122fff748fd tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Task: {'id': task-5116390, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.308623] env[65726]: DEBUG oslo_vmware.api [None req-c5289ae6-5b56-421a-a0e1-7e74e3494692 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Waiting for the task: (returnval){ [ 980.308623] env[65726]: value = "task-5116392" [ 980.308623] env[65726]: _type = "Task" [ 980.308623] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 980.318583] env[65726]: DEBUG oslo_vmware.api [None req-d903724e-08f6-4863-a079-5fadb490f6c5 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': task-5116391, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.322124] env[65726]: DEBUG oslo_vmware.api [None req-c5289ae6-5b56-421a-a0e1-7e74e3494692 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Task: {'id': task-5116392, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.759769] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c75c9ed0-c9ca-4fd9-9695-87538dcf1f78 tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.409s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 980.760414] env[65726]: DEBUG nova.compute.manager [None req-c75c9ed0-c9ca-4fd9-9695-87538dcf1f78 tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] [instance: 406c59e7-8eb8-4cfd-8e1e-e151ab908571] Start building networks asynchronously for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 980.764237] env[65726]: DEBUG oslo_concurrency.lockutils [None req-83256413-d207-4cc8-b904-cfb2a5ea5628 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.502s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 980.765586] env[65726]: INFO nova.compute.claims [None req-83256413-d207-4cc8-b904-cfb2a5ea5628 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] [instance: f9a40d36-dda9-454c-8774-ed011f3f50ae] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 980.791514] env[65726]: DEBUG oslo_vmware.api [None req-9f65b40f-615d-4bae-ac4c-c122fff748fd tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Task: {'id': task-5116390, 'name': PowerOnVM_Task} progress is 88%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.793829] env[65726]: DEBUG oslo_concurrency.lockutils [None req-378e66fe-caea-4dff-a5bd-614d0f24e5ba tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] Releasing lock "refresh_cache-73f5b23e-4463-4068-8994-dd2752a2abc9" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 980.793829] env[65726]: DEBUG nova.compute.manager [None req-378e66fe-caea-4dff-a5bd-614d0f24e5ba tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] [instance: 73f5b23e-4463-4068-8994-dd2752a2abc9] Start destroying the instance on the hypervisor. {{(pid=65726) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 980.794084] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-378e66fe-caea-4dff-a5bd-614d0f24e5ba tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] [instance: 73f5b23e-4463-4068-8994-dd2752a2abc9] Destroying instance {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 980.796113] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8cd28ef-3891-42b8-b2d1-bab9b8b4a4af {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.805732] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-378e66fe-caea-4dff-a5bd-614d0f24e5ba tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] [instance: 73f5b23e-4463-4068-8994-dd2752a2abc9] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 980.808622] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-eaf29ba6-b66c-4497-9408-2096e22d136b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.819305] env[65726]: DEBUG oslo_vmware.api [None req-d903724e-08f6-4863-a079-5fadb490f6c5 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': task-5116391, 'name': ReconfigVM_Task, 'duration_secs': 0.334349} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 980.821339] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-d903724e-08f6-4863-a079-5fadb490f6c5 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 54e85161-5f63-405f-83f8-490c70645a3b] Reconfigured VM instance instance-0000004f to attach disk [datastore1] 54e85161-5f63-405f-83f8-490c70645a3b/54e85161-5f63-405f-83f8-490c70645a3b.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 980.822538] env[65726]: DEBUG oslo_vmware.api [None req-378e66fe-caea-4dff-a5bd-614d0f24e5ba tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] Waiting for the task: (returnval){ [ 980.822538] env[65726]: value = "task-5116393" [ 980.822538] env[65726]: _type = "Task" [ 980.822538] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 980.822777] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-34203c46-bd82-4f20-84c9-be6ca8ff060b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.833718] env[65726]: DEBUG oslo_vmware.api [None req-c5289ae6-5b56-421a-a0e1-7e74e3494692 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Task: {'id': task-5116392, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.173874} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 980.834868] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-c5289ae6-5b56-421a-a0e1-7e74e3494692 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Deleted the datastore file {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 980.835082] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-c5289ae6-5b56-421a-a0e1-7e74e3494692 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: f2d9090c-988f-43f4-9c81-7aa718a3438a] Deleted contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 980.835375] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-c5289ae6-5b56-421a-a0e1-7e74e3494692 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: f2d9090c-988f-43f4-9c81-7aa718a3438a] Instance destroyed {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 980.835462] env[65726]: INFO nova.compute.manager [None req-c5289ae6-5b56-421a-a0e1-7e74e3494692 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] [instance: f2d9090c-988f-43f4-9c81-7aa718a3438a] Took 1.20 seconds to destroy the instance on the hypervisor. [ 980.835706] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-c5289ae6-5b56-421a-a0e1-7e74e3494692 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 980.836312] env[65726]: DEBUG nova.compute.manager [-] [instance: f2d9090c-988f-43f4-9c81-7aa718a3438a] Deallocating network for instance {{(pid=65726) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 980.836415] env[65726]: DEBUG nova.network.neutron [-] [instance: f2d9090c-988f-43f4-9c81-7aa718a3438a] deallocate_for_instance() {{(pid=65726) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 980.836980] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 980.837263] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 980.849742] env[65726]: DEBUG oslo_vmware.api [None req-d903724e-08f6-4863-a079-5fadb490f6c5 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Waiting for the task: (returnval){ [ 980.849742] env[65726]: value = "task-5116394" [ 980.849742] env[65726]: _type = "Task" [ 980.849742] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 980.850040] env[65726]: DEBUG oslo_vmware.api [None req-378e66fe-caea-4dff-a5bd-614d0f24e5ba tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] Task: {'id': task-5116393, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.860720] env[65726]: DEBUG oslo_vmware.api [None req-d903724e-08f6-4863-a079-5fadb490f6c5 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': task-5116394, 'name': Rename_Task} progress is 10%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.885987] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 980.886203] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 981.272964] env[65726]: DEBUG nova.compute.utils [None req-c75c9ed0-c9ca-4fd9-9695-87538dcf1f78 tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Using /dev/sd instead of None {{(pid=65726) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 981.275319] env[65726]: DEBUG nova.compute.manager [None req-c75c9ed0-c9ca-4fd9-9695-87538dcf1f78 tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] [instance: 406c59e7-8eb8-4cfd-8e1e-e151ab908571] Allocating IP information in the background. {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 981.275598] env[65726]: DEBUG nova.network.neutron [None req-c75c9ed0-c9ca-4fd9-9695-87538dcf1f78 tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] [instance: 406c59e7-8eb8-4cfd-8e1e-e151ab908571] allocate_for_instance() {{(pid=65726) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 981.275915] env[65726]: WARNING neutronclient.v2_0.client [None req-c75c9ed0-c9ca-4fd9-9695-87538dcf1f78 tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 981.276267] env[65726]: WARNING neutronclient.v2_0.client [None req-c75c9ed0-c9ca-4fd9-9695-87538dcf1f78 tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 981.277129] env[65726]: WARNING openstack [None req-c75c9ed0-c9ca-4fd9-9695-87538dcf1f78 tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 981.277272] env[65726]: WARNING openstack [None req-c75c9ed0-c9ca-4fd9-9695-87538dcf1f78 tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 981.284479] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5d32d102-744f-471d-8592-b9e461d3322c tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Acquiring lock "060646cc-f0fb-4e6c-9f67-7b2daa6b6fd6" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 981.284714] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5d32d102-744f-471d-8592-b9e461d3322c tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Lock "060646cc-f0fb-4e6c-9f67-7b2daa6b6fd6" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 981.284920] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5d32d102-744f-471d-8592-b9e461d3322c tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Acquiring lock "060646cc-f0fb-4e6c-9f67-7b2daa6b6fd6-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 981.285110] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5d32d102-744f-471d-8592-b9e461d3322c tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Lock "060646cc-f0fb-4e6c-9f67-7b2daa6b6fd6-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 981.285270] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5d32d102-744f-471d-8592-b9e461d3322c tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Lock "060646cc-f0fb-4e6c-9f67-7b2daa6b6fd6-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 981.288768] env[65726]: INFO nova.compute.manager [None req-5d32d102-744f-471d-8592-b9e461d3322c tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: 060646cc-f0fb-4e6c-9f67-7b2daa6b6fd6] Terminating instance [ 981.303943] env[65726]: DEBUG oslo_vmware.api [None req-9f65b40f-615d-4bae-ac4c-c122fff748fd tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Task: {'id': task-5116390, 'name': PowerOnVM_Task, 'duration_secs': 0.960495} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 981.304358] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-9f65b40f-615d-4bae-ac4c-c122fff748fd tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] [instance: aa6ce489-c62f-4481-87b7-e74242aeb8ca] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 981.304659] env[65726]: INFO nova.compute.manager [None req-9f65b40f-615d-4bae-ac4c-c122fff748fd tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] [instance: aa6ce489-c62f-4481-87b7-e74242aeb8ca] Took 9.81 seconds to spawn the instance on the hypervisor. [ 981.304850] env[65726]: DEBUG nova.compute.manager [None req-9f65b40f-615d-4bae-ac4c-c122fff748fd tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] [instance: aa6ce489-c62f-4481-87b7-e74242aeb8ca] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 981.305742] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4322470e-f773-481c-bfe4-05ea505f746c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.339796] env[65726]: DEBUG oslo_vmware.api [None req-378e66fe-caea-4dff-a5bd-614d0f24e5ba tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] Task: {'id': task-5116393, 'name': PowerOffVM_Task, 'duration_secs': 0.339962} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 981.340308] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-378e66fe-caea-4dff-a5bd-614d0f24e5ba tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] [instance: 73f5b23e-4463-4068-8994-dd2752a2abc9] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 981.340479] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-378e66fe-caea-4dff-a5bd-614d0f24e5ba tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] [instance: 73f5b23e-4463-4068-8994-dd2752a2abc9] Unregistering the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 981.340775] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7abb89f6-9337-4532-9425-3fe1461bd9d0 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.363063] env[65726]: DEBUG oslo_vmware.api [None req-d903724e-08f6-4863-a079-5fadb490f6c5 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': task-5116394, 'name': Rename_Task, 'duration_secs': 0.178522} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 981.363434] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-d903724e-08f6-4863-a079-5fadb490f6c5 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 54e85161-5f63-405f-83f8-490c70645a3b] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 981.363970] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-467a64bd-0ef0-44a5-ad9c-97cbc67357ca {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.372240] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-378e66fe-caea-4dff-a5bd-614d0f24e5ba tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] [instance: 73f5b23e-4463-4068-8994-dd2752a2abc9] Unregistered the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 981.372240] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-378e66fe-caea-4dff-a5bd-614d0f24e5ba tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] [instance: 73f5b23e-4463-4068-8994-dd2752a2abc9] Deleting contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 981.372240] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-378e66fe-caea-4dff-a5bd-614d0f24e5ba tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] Deleting the datastore file [datastore1] 73f5b23e-4463-4068-8994-dd2752a2abc9 {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 981.374138] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8ede9040-5b81-4839-ab82-a28094cab1c3 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.376988] env[65726]: DEBUG oslo_vmware.api [None req-d903724e-08f6-4863-a079-5fadb490f6c5 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Waiting for the task: (returnval){ [ 981.376988] env[65726]: value = "task-5116396" [ 981.376988] env[65726]: _type = "Task" [ 981.376988] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 981.382766] env[65726]: DEBUG nova.policy [None req-c75c9ed0-c9ca-4fd9-9695-87538dcf1f78 tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b647b94d0d1848abaee17370c2bedc6f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6961c16aa6a346679bbde25930b2a5b7', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65726) authorize /opt/stack/nova/nova/policy.py:192}} [ 981.392833] env[65726]: DEBUG oslo_vmware.api [None req-378e66fe-caea-4dff-a5bd-614d0f24e5ba tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] Waiting for the task: (returnval){ [ 981.392833] env[65726]: value = "task-5116397" [ 981.392833] env[65726]: _type = "Task" [ 981.392833] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 981.401309] env[65726]: DEBUG oslo_vmware.api [None req-d903724e-08f6-4863-a079-5fadb490f6c5 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': task-5116396, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.409196] env[65726]: DEBUG oslo_vmware.api [None req-378e66fe-caea-4dff-a5bd-614d0f24e5ba tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] Task: {'id': task-5116397, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.538117] env[65726]: DEBUG nova.network.neutron [None req-ce025067-bba1-472f-842e-5d842d2328d6 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: dcde25df-87a5-47fb-94ce-334f68894e04] Successfully updated port: b48313f9-a0d3-4fdd-88c6-d103aef0b4f2 {{(pid=65726) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 981.640139] env[65726]: DEBUG nova.network.neutron [None req-c75c9ed0-c9ca-4fd9-9695-87538dcf1f78 tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] [instance: 406c59e7-8eb8-4cfd-8e1e-e151ab908571] Successfully created port: eb554e53-6a49-4f06-8271-81ad5c918184 {{(pid=65726) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 981.643883] env[65726]: DEBUG nova.network.neutron [-] [instance: f2d9090c-988f-43f4-9c81-7aa718a3438a] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 981.764177] env[65726]: DEBUG nova.compute.manager [req-5dbb652f-2a59-4ad5-b266-e5ae56f1698f req-0be5ac1c-dcbc-4171-95d5-2115ac04f6e3 service nova] [instance: f2d9090c-988f-43f4-9c81-7aa718a3438a] Received event network-vif-deleted-7117cd21-ee2a-497d-b789-65e43b068258 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 981.776274] env[65726]: DEBUG nova.compute.manager [None req-c75c9ed0-c9ca-4fd9-9695-87538dcf1f78 tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] [instance: 406c59e7-8eb8-4cfd-8e1e-e151ab908571] Start building block device mappings for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 981.798908] env[65726]: DEBUG nova.compute.manager [None req-5d32d102-744f-471d-8592-b9e461d3322c tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: 060646cc-f0fb-4e6c-9f67-7b2daa6b6fd6] Start destroying the instance on the hypervisor. {{(pid=65726) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 981.799184] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-5d32d102-744f-471d-8592-b9e461d3322c tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: 060646cc-f0fb-4e6c-9f67-7b2daa6b6fd6] Destroying instance {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 981.800358] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc152ab2-65f7-4d8b-8848-3b30493b1e04 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.812631] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-5d32d102-744f-471d-8592-b9e461d3322c tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: 060646cc-f0fb-4e6c-9f67-7b2daa6b6fd6] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 981.812963] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a47f4719-6483-40ba-b518-88e640274bd1 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.829732] env[65726]: DEBUG oslo_vmware.api [None req-5d32d102-744f-471d-8592-b9e461d3322c tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Waiting for the task: (returnval){ [ 981.829732] env[65726]: value = "task-5116398" [ 981.829732] env[65726]: _type = "Task" [ 981.829732] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 981.830503] env[65726]: INFO nova.compute.manager [None req-9f65b40f-615d-4bae-ac4c-c122fff748fd tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] [instance: aa6ce489-c62f-4481-87b7-e74242aeb8ca] Took 31.81 seconds to build instance. [ 981.849025] env[65726]: DEBUG oslo_vmware.api [None req-5d32d102-744f-471d-8592-b9e461d3322c tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Task: {'id': task-5116398, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.900527] env[65726]: DEBUG oslo_vmware.api [None req-d903724e-08f6-4863-a079-5fadb490f6c5 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': task-5116396, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.907447] env[65726]: DEBUG oslo_vmware.api [None req-378e66fe-caea-4dff-a5bd-614d0f24e5ba tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] Task: {'id': task-5116397, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.191886} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 981.912640] env[65726]: DEBUG nova.network.neutron [None req-c75c9ed0-c9ca-4fd9-9695-87538dcf1f78 tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] [instance: 406c59e7-8eb8-4cfd-8e1e-e151ab908571] Successfully created port: 903e092f-430c-4ecb-9cc6-1f2adbcd949f {{(pid=65726) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 981.915588] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-378e66fe-caea-4dff-a5bd-614d0f24e5ba tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] Deleted the datastore file {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 981.916611] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-378e66fe-caea-4dff-a5bd-614d0f24e5ba tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] [instance: 73f5b23e-4463-4068-8994-dd2752a2abc9] Deleted contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 981.916611] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-378e66fe-caea-4dff-a5bd-614d0f24e5ba tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] [instance: 73f5b23e-4463-4068-8994-dd2752a2abc9] Instance destroyed {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 981.916611] env[65726]: INFO nova.compute.manager [None req-378e66fe-caea-4dff-a5bd-614d0f24e5ba tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] [instance: 73f5b23e-4463-4068-8994-dd2752a2abc9] Took 1.12 seconds to destroy the instance on the hypervisor. [ 981.916611] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-378e66fe-caea-4dff-a5bd-614d0f24e5ba tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 981.917042] env[65726]: DEBUG nova.compute.manager [-] [instance: 73f5b23e-4463-4068-8994-dd2752a2abc9] Deallocating network for instance {{(pid=65726) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 981.917176] env[65726]: DEBUG nova.network.neutron [-] [instance: 73f5b23e-4463-4068-8994-dd2752a2abc9] deallocate_for_instance() {{(pid=65726) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 981.917762] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 981.917899] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 981.950751] env[65726]: DEBUG nova.network.neutron [-] [instance: 73f5b23e-4463-4068-8994-dd2752a2abc9] Instance cache missing network info. {{(pid=65726) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 981.951529] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 981.952280] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 982.043215] env[65726]: DEBUG oslo_concurrency.lockutils [None req-ce025067-bba1-472f-842e-5d842d2328d6 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Acquiring lock "refresh_cache-dcde25df-87a5-47fb-94ce-334f68894e04" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 982.043547] env[65726]: DEBUG oslo_concurrency.lockutils [None req-ce025067-bba1-472f-842e-5d842d2328d6 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Acquired lock "refresh_cache-dcde25df-87a5-47fb-94ce-334f68894e04" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 982.043906] env[65726]: DEBUG nova.network.neutron [None req-ce025067-bba1-472f-842e-5d842d2328d6 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: dcde25df-87a5-47fb-94ce-334f68894e04] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 982.067769] env[65726]: DEBUG nova.compute.manager [req-7150f28c-5935-4e64-a462-acfd452a9997 req-ffed0d4e-aa43-4c8d-91a9-e24e367b3f62 service nova] [instance: dcde25df-87a5-47fb-94ce-334f68894e04] Received event network-vif-plugged-b48313f9-a0d3-4fdd-88c6-d103aef0b4f2 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 982.067769] env[65726]: DEBUG oslo_concurrency.lockutils [req-7150f28c-5935-4e64-a462-acfd452a9997 req-ffed0d4e-aa43-4c8d-91a9-e24e367b3f62 service nova] Acquiring lock "dcde25df-87a5-47fb-94ce-334f68894e04-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 982.068026] env[65726]: DEBUG oslo_concurrency.lockutils [req-7150f28c-5935-4e64-a462-acfd452a9997 req-ffed0d4e-aa43-4c8d-91a9-e24e367b3f62 service nova] Lock "dcde25df-87a5-47fb-94ce-334f68894e04-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 982.069257] env[65726]: DEBUG oslo_concurrency.lockutils [req-7150f28c-5935-4e64-a462-acfd452a9997 req-ffed0d4e-aa43-4c8d-91a9-e24e367b3f62 service nova] Lock "dcde25df-87a5-47fb-94ce-334f68894e04-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 982.069257] env[65726]: DEBUG nova.compute.manager [req-7150f28c-5935-4e64-a462-acfd452a9997 req-ffed0d4e-aa43-4c8d-91a9-e24e367b3f62 service nova] [instance: dcde25df-87a5-47fb-94ce-334f68894e04] No waiting events found dispatching network-vif-plugged-b48313f9-a0d3-4fdd-88c6-d103aef0b4f2 {{(pid=65726) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 982.069257] env[65726]: WARNING nova.compute.manager [req-7150f28c-5935-4e64-a462-acfd452a9997 req-ffed0d4e-aa43-4c8d-91a9-e24e367b3f62 service nova] [instance: dcde25df-87a5-47fb-94ce-334f68894e04] Received unexpected event network-vif-plugged-b48313f9-a0d3-4fdd-88c6-d103aef0b4f2 for instance with vm_state active and task_state None. [ 982.147509] env[65726]: INFO nova.compute.manager [-] [instance: f2d9090c-988f-43f4-9c81-7aa718a3438a] Took 1.31 seconds to deallocate network for instance. [ 982.211647] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e003bcc5-dde1-4de3-bff0-3d22fe230593 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.220645] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-198994b8-8e3b-4ef0-9105-8d2b72648d14 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.258444] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-804cf262-7122-48a7-aec1-f09da8628dcd {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.270073] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3a3adb9-5b21-49d8-bfac-94a9281bb4ec {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.294922] env[65726]: DEBUG nova.compute.provider_tree [None req-83256413-d207-4cc8-b904-cfb2a5ea5628 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Updating inventory in ProviderTree for provider 07c4692f-bdb4-4058-9173-ff9664830295 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 982.341159] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9f65b40f-615d-4bae-ac4c-c122fff748fd tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Lock "aa6ce489-c62f-4481-87b7-e74242aeb8ca" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 33.336s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 982.351195] env[65726]: DEBUG oslo_vmware.api [None req-5d32d102-744f-471d-8592-b9e461d3322c tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Task: {'id': task-5116398, 'name': PowerOffVM_Task, 'duration_secs': 0.227716} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 982.352475] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-5d32d102-744f-471d-8592-b9e461d3322c tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: 060646cc-f0fb-4e6c-9f67-7b2daa6b6fd6] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 982.352475] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-5d32d102-744f-471d-8592-b9e461d3322c tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: 060646cc-f0fb-4e6c-9f67-7b2daa6b6fd6] Unregistering the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 982.352718] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-66f7493e-b85c-4526-8b0e-65aab1912a32 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.397931] env[65726]: DEBUG oslo_vmware.api [None req-d903724e-08f6-4863-a079-5fadb490f6c5 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': task-5116396, 'name': PowerOnVM_Task, 'duration_secs': 0.767697} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 982.398863] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-d903724e-08f6-4863-a079-5fadb490f6c5 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 54e85161-5f63-405f-83f8-490c70645a3b] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 982.399083] env[65726]: INFO nova.compute.manager [None req-d903724e-08f6-4863-a079-5fadb490f6c5 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 54e85161-5f63-405f-83f8-490c70645a3b] Took 8.27 seconds to spawn the instance on the hypervisor. [ 982.399398] env[65726]: DEBUG nova.compute.manager [None req-d903724e-08f6-4863-a079-5fadb490f6c5 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 54e85161-5f63-405f-83f8-490c70645a3b] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 982.400284] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67ad9991-b2e4-45a0-9f25-8879928d02e2 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.442986] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-5d32d102-744f-471d-8592-b9e461d3322c tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: 060646cc-f0fb-4e6c-9f67-7b2daa6b6fd6] Unregistered the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 982.443232] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-5d32d102-744f-471d-8592-b9e461d3322c tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: 060646cc-f0fb-4e6c-9f67-7b2daa6b6fd6] Deleting contents of the VM from datastore datastore2 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 982.443408] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-5d32d102-744f-471d-8592-b9e461d3322c tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Deleting the datastore file [datastore2] 060646cc-f0fb-4e6c-9f67-7b2daa6b6fd6 {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 982.443703] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-cda869ea-58c3-4005-a3bb-75641766b0ad {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.451793] env[65726]: DEBUG oslo_vmware.api [None req-5d32d102-744f-471d-8592-b9e461d3322c tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Waiting for the task: (returnval){ [ 982.451793] env[65726]: value = "task-5116400" [ 982.451793] env[65726]: _type = "Task" [ 982.451793] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 982.463250] env[65726]: DEBUG nova.network.neutron [-] [instance: 73f5b23e-4463-4068-8994-dd2752a2abc9] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 982.464714] env[65726]: DEBUG oslo_vmware.api [None req-5d32d102-744f-471d-8592-b9e461d3322c tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Task: {'id': task-5116400, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.548756] env[65726]: WARNING openstack [None req-ce025067-bba1-472f-842e-5d842d2328d6 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 982.549159] env[65726]: WARNING openstack [None req-ce025067-bba1-472f-842e-5d842d2328d6 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 982.604227] env[65726]: WARNING nova.network.neutron [None req-ce025067-bba1-472f-842e-5d842d2328d6 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: dcde25df-87a5-47fb-94ce-334f68894e04] 51a66448-0f72-4266-8868-856b48be839f already exists in list: networks containing: ['51a66448-0f72-4266-8868-856b48be839f']. ignoring it [ 982.646931] env[65726]: DEBUG nova.network.neutron [None req-c75c9ed0-c9ca-4fd9-9695-87538dcf1f78 tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] [instance: 406c59e7-8eb8-4cfd-8e1e-e151ab908571] Successfully created port: b0ae20cc-cb0a-450d-96a3-4726b64254db {{(pid=65726) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 982.657957] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c5289ae6-5b56-421a-a0e1-7e74e3494692 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 982.664923] env[65726]: WARNING openstack [None req-ce025067-bba1-472f-842e-5d842d2328d6 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 982.665309] env[65726]: WARNING openstack [None req-ce025067-bba1-472f-842e-5d842d2328d6 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 982.798591] env[65726]: DEBUG nova.compute.manager [None req-c75c9ed0-c9ca-4fd9-9695-87538dcf1f78 tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] [instance: 406c59e7-8eb8-4cfd-8e1e-e151ab908571] Start spawning the instance on the hypervisor. {{(pid=65726) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 982.830285] env[65726]: DEBUG nova.virt.hardware [None req-c75c9ed0-c9ca-4fd9-9695-87538dcf1f78 tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T19:28:34Z,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1539f06fa3534e90acbc3a60c4b8bd3f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T19:28:35Z,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 982.830552] env[65726]: DEBUG nova.virt.hardware [None req-c75c9ed0-c9ca-4fd9-9695-87538dcf1f78 tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 982.830969] env[65726]: DEBUG nova.virt.hardware [None req-c75c9ed0-c9ca-4fd9-9695-87538dcf1f78 tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 982.830969] env[65726]: DEBUG nova.virt.hardware [None req-c75c9ed0-c9ca-4fd9-9695-87538dcf1f78 tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 982.831119] env[65726]: DEBUG nova.virt.hardware [None req-c75c9ed0-c9ca-4fd9-9695-87538dcf1f78 tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 982.831157] env[65726]: DEBUG nova.virt.hardware [None req-c75c9ed0-c9ca-4fd9-9695-87538dcf1f78 tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 982.831606] env[65726]: DEBUG nova.virt.hardware [None req-c75c9ed0-c9ca-4fd9-9695-87538dcf1f78 tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 982.831771] env[65726]: DEBUG nova.virt.hardware [None req-c75c9ed0-c9ca-4fd9-9695-87538dcf1f78 tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 982.831957] env[65726]: DEBUG nova.virt.hardware [None req-c75c9ed0-c9ca-4fd9-9695-87538dcf1f78 tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 982.832185] env[65726]: DEBUG nova.virt.hardware [None req-c75c9ed0-c9ca-4fd9-9695-87538dcf1f78 tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 982.832379] env[65726]: DEBUG nova.virt.hardware [None req-c75c9ed0-c9ca-4fd9-9695-87538dcf1f78 tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 982.833505] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96354f2d-40c8-479c-8754-6b1804c8e679 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.837713] env[65726]: DEBUG nova.scheduler.client.report [None req-83256413-d207-4cc8-b904-cfb2a5ea5628 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Updated inventory for provider 07c4692f-bdb4-4058-9173-ff9664830295 with generation 118 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 982.837968] env[65726]: DEBUG nova.compute.provider_tree [None req-83256413-d207-4cc8-b904-cfb2a5ea5628 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Updating resource provider 07c4692f-bdb4-4058-9173-ff9664830295 generation from 118 to 119 during operation: update_inventory {{(pid=65726) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 982.838187] env[65726]: DEBUG nova.compute.provider_tree [None req-83256413-d207-4cc8-b904-cfb2a5ea5628 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Updating inventory in ProviderTree for provider 07c4692f-bdb4-4058-9173-ff9664830295 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 982.850015] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b859f96f-6fb2-4a85-a26e-84413de8db62 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.922458] env[65726]: INFO nova.compute.manager [None req-d903724e-08f6-4863-a079-5fadb490f6c5 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 54e85161-5f63-405f-83f8-490c70645a3b] Took 27.27 seconds to build instance. [ 982.963017] env[65726]: DEBUG oslo_vmware.api [None req-5d32d102-744f-471d-8592-b9e461d3322c tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Task: {'id': task-5116400, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.437638} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 982.963408] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-5d32d102-744f-471d-8592-b9e461d3322c tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Deleted the datastore file {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 982.963631] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-5d32d102-744f-471d-8592-b9e461d3322c tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: 060646cc-f0fb-4e6c-9f67-7b2daa6b6fd6] Deleted contents of the VM from datastore datastore2 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 982.963837] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-5d32d102-744f-471d-8592-b9e461d3322c tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: 060646cc-f0fb-4e6c-9f67-7b2daa6b6fd6] Instance destroyed {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 982.964055] env[65726]: INFO nova.compute.manager [None req-5d32d102-744f-471d-8592-b9e461d3322c tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: 060646cc-f0fb-4e6c-9f67-7b2daa6b6fd6] Took 1.16 seconds to destroy the instance on the hypervisor. [ 982.964349] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-5d32d102-744f-471d-8592-b9e461d3322c tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 982.964564] env[65726]: DEBUG nova.compute.manager [-] [instance: 060646cc-f0fb-4e6c-9f67-7b2daa6b6fd6] Deallocating network for instance {{(pid=65726) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 982.964667] env[65726]: DEBUG nova.network.neutron [-] [instance: 060646cc-f0fb-4e6c-9f67-7b2daa6b6fd6] deallocate_for_instance() {{(pid=65726) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 982.965196] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 982.965455] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 982.973024] env[65726]: INFO nova.compute.manager [-] [instance: 73f5b23e-4463-4068-8994-dd2752a2abc9] Took 1.06 seconds to deallocate network for instance. [ 983.208982] env[65726]: WARNING openstack [None req-ce025067-bba1-472f-842e-5d842d2328d6 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 983.209599] env[65726]: WARNING openstack [None req-ce025067-bba1-472f-842e-5d842d2328d6 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 983.254384] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 983.254666] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 983.344442] env[65726]: DEBUG oslo_concurrency.lockutils [None req-83256413-d207-4cc8-b904-cfb2a5ea5628 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.581s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 983.345042] env[65726]: DEBUG nova.compute.manager [None req-83256413-d207-4cc8-b904-cfb2a5ea5628 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] [instance: f9a40d36-dda9-454c-8774-ed011f3f50ae] Start building networks asynchronously for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 983.347867] env[65726]: DEBUG oslo_concurrency.lockutils [None req-ed96d5ac-cdb1-4748-8154-6accb1ae5e65 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.569s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 983.349505] env[65726]: INFO nova.compute.claims [None req-ed96d5ac-cdb1-4748-8154-6accb1ae5e65 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: 067c9332-fa2c-47fe-93ae-643de1fe77cf] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 983.424570] env[65726]: DEBUG oslo_concurrency.lockutils [None req-d903724e-08f6-4863-a079-5fadb490f6c5 tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Lock "54e85161-5f63-405f-83f8-490c70645a3b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 28.787s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 983.480891] env[65726]: DEBUG oslo_concurrency.lockutils [None req-378e66fe-caea-4dff-a5bd-614d0f24e5ba tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 983.857610] env[65726]: DEBUG nova.compute.utils [None req-83256413-d207-4cc8-b904-cfb2a5ea5628 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Using /dev/sd instead of None {{(pid=65726) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 983.862243] env[65726]: DEBUG nova.compute.manager [None req-83256413-d207-4cc8-b904-cfb2a5ea5628 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] [instance: f9a40d36-dda9-454c-8774-ed011f3f50ae] Allocating IP information in the background. {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 983.862588] env[65726]: DEBUG nova.network.neutron [None req-83256413-d207-4cc8-b904-cfb2a5ea5628 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] [instance: f9a40d36-dda9-454c-8774-ed011f3f50ae] allocate_for_instance() {{(pid=65726) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 983.862969] env[65726]: WARNING neutronclient.v2_0.client [None req-83256413-d207-4cc8-b904-cfb2a5ea5628 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 983.863359] env[65726]: WARNING neutronclient.v2_0.client [None req-83256413-d207-4cc8-b904-cfb2a5ea5628 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 983.863996] env[65726]: WARNING openstack [None req-83256413-d207-4cc8-b904-cfb2a5ea5628 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 983.864411] env[65726]: WARNING openstack [None req-83256413-d207-4cc8-b904-cfb2a5ea5628 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 983.883837] env[65726]: WARNING openstack [None req-ce025067-bba1-472f-842e-5d842d2328d6 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 983.884106] env[65726]: WARNING openstack [None req-ce025067-bba1-472f-842e-5d842d2328d6 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 983.923269] env[65726]: DEBUG nova.policy [None req-83256413-d207-4cc8-b904-cfb2a5ea5628 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8ebf00cfba6e4d4c8a72c463707b7acd', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '06e221dc693640929cac7bc5af45948f', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65726) authorize /opt/stack/nova/nova/policy.py:192}} [ 984.235273] env[65726]: DEBUG nova.network.neutron [None req-c75c9ed0-c9ca-4fd9-9695-87538dcf1f78 tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] [instance: 406c59e7-8eb8-4cfd-8e1e-e151ab908571] Successfully updated port: eb554e53-6a49-4f06-8271-81ad5c918184 {{(pid=65726) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 984.243169] env[65726]: DEBUG nova.network.neutron [None req-83256413-d207-4cc8-b904-cfb2a5ea5628 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] [instance: f9a40d36-dda9-454c-8774-ed011f3f50ae] Successfully created port: 4a5e689f-4a77-4c32-81bf-4f86a5c702cb {{(pid=65726) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 984.247913] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05bbd1df-c19d-4094-9d6c-ea5747220dc0 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.261221] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9efadfd1-ea11-4b60-a73e-104b3dc24925 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.305697] env[65726]: WARNING openstack [None req-ce025067-bba1-472f-842e-5d842d2328d6 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 984.306110] env[65726]: WARNING openstack [None req-ce025067-bba1-472f-842e-5d842d2328d6 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 984.314620] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8e774b9-e47f-4905-946d-03685d99b742 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.317286] env[65726]: DEBUG nova.network.neutron [-] [instance: 060646cc-f0fb-4e6c-9f67-7b2daa6b6fd6] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 984.326183] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfdbb600-7cf2-4c85-94ac-ee98633f7c3b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.344926] env[65726]: DEBUG nova.compute.provider_tree [None req-ed96d5ac-cdb1-4748-8154-6accb1ae5e65 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 984.363019] env[65726]: DEBUG nova.compute.manager [None req-83256413-d207-4cc8-b904-cfb2a5ea5628 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] [instance: f9a40d36-dda9-454c-8774-ed011f3f50ae] Start building block device mappings for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 984.450747] env[65726]: DEBUG nova.network.neutron [None req-ce025067-bba1-472f-842e-5d842d2328d6 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: dcde25df-87a5-47fb-94ce-334f68894e04] Updating instance_info_cache with network_info: [{"id": "bf51c96c-63c5-46de-a40f-58e6769c5fcf", "address": "fa:16:3e:4c:88:87", "network": {"id": "51a66448-0f72-4266-8868-856b48be839f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-155658961-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.146", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d46ccb00794f458b85da4a93879139ee", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b1f3e6c3-5584-4852-9017-476ab8ac4946", "external-id": "nsx-vlan-transportzone-304", "segmentation_id": 304, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbf51c96c-63", "ovs_interfaceid": "bf51c96c-63c5-46de-a40f-58e6769c5fcf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "b48313f9-a0d3-4fdd-88c6-d103aef0b4f2", "address": "fa:16:3e:e7:cd:41", "network": {"id": "51a66448-0f72-4266-8868-856b48be839f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-155658961-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d46ccb00794f458b85da4a93879139ee", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b1f3e6c3-5584-4852-9017-476ab8ac4946", "external-id": "nsx-vlan-transportzone-304", "segmentation_id": 304, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb48313f9-a0", "ovs_interfaceid": "b48313f9-a0d3-4fdd-88c6-d103aef0b4f2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 984.820315] env[65726]: INFO nova.compute.manager [-] [instance: 060646cc-f0fb-4e6c-9f67-7b2daa6b6fd6] Took 1.86 seconds to deallocate network for instance. [ 984.850061] env[65726]: DEBUG nova.scheduler.client.report [None req-ed96d5ac-cdb1-4748-8154-6accb1ae5e65 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 984.954824] env[65726]: DEBUG oslo_concurrency.lockutils [None req-ce025067-bba1-472f-842e-5d842d2328d6 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Releasing lock "refresh_cache-dcde25df-87a5-47fb-94ce-334f68894e04" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 984.955654] env[65726]: DEBUG oslo_concurrency.lockutils [None req-ce025067-bba1-472f-842e-5d842d2328d6 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Acquiring lock "dcde25df-87a5-47fb-94ce-334f68894e04" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 984.955769] env[65726]: DEBUG oslo_concurrency.lockutils [None req-ce025067-bba1-472f-842e-5d842d2328d6 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Acquired lock "dcde25df-87a5-47fb-94ce-334f68894e04" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 984.956725] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfecbeea-eab6-4a44-9b5c-4e4a5ea6a51b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.976156] env[65726]: DEBUG nova.virt.hardware [None req-ce025067-bba1-472f-842e-5d842d2328d6 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 984.976383] env[65726]: DEBUG nova.virt.hardware [None req-ce025067-bba1-472f-842e-5d842d2328d6 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 984.976545] env[65726]: DEBUG nova.virt.hardware [None req-ce025067-bba1-472f-842e-5d842d2328d6 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 984.976951] env[65726]: DEBUG nova.virt.hardware [None req-ce025067-bba1-472f-842e-5d842d2328d6 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 984.976951] env[65726]: DEBUG nova.virt.hardware [None req-ce025067-bba1-472f-842e-5d842d2328d6 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 984.977171] env[65726]: DEBUG nova.virt.hardware [None req-ce025067-bba1-472f-842e-5d842d2328d6 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 984.977414] env[65726]: DEBUG nova.virt.hardware [None req-ce025067-bba1-472f-842e-5d842d2328d6 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 984.977614] env[65726]: DEBUG nova.virt.hardware [None req-ce025067-bba1-472f-842e-5d842d2328d6 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 984.977905] env[65726]: DEBUG nova.virt.hardware [None req-ce025067-bba1-472f-842e-5d842d2328d6 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 984.978126] env[65726]: DEBUG nova.virt.hardware [None req-ce025067-bba1-472f-842e-5d842d2328d6 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 984.978364] env[65726]: DEBUG nova.virt.hardware [None req-ce025067-bba1-472f-842e-5d842d2328d6 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 984.985773] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-ce025067-bba1-472f-842e-5d842d2328d6 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: dcde25df-87a5-47fb-94ce-334f68894e04] Reconfiguring VM to attach interface {{(pid=65726) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 984.986291] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-41b8bd00-42c9-4c90-a727-d5df352d4323 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.006725] env[65726]: DEBUG oslo_vmware.api [None req-ce025067-bba1-472f-842e-5d842d2328d6 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Waiting for the task: (returnval){ [ 985.006725] env[65726]: value = "task-5116401" [ 985.006725] env[65726]: _type = "Task" [ 985.006725] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.016756] env[65726]: DEBUG oslo_vmware.api [None req-ce025067-bba1-472f-842e-5d842d2328d6 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': task-5116401, 'name': ReconfigVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.329721] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5d32d102-744f-471d-8592-b9e461d3322c tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 985.357201] env[65726]: DEBUG oslo_concurrency.lockutils [None req-ed96d5ac-cdb1-4748-8154-6accb1ae5e65 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.009s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 985.357911] env[65726]: DEBUG nova.compute.manager [None req-ed96d5ac-cdb1-4748-8154-6accb1ae5e65 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: 067c9332-fa2c-47fe-93ae-643de1fe77cf] Start building networks asynchronously for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 985.361997] env[65726]: DEBUG oslo_concurrency.lockutils [None req-775eaf93-375d-4ff3-9c57-60f5cc32d4ea tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.919s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 985.364332] env[65726]: INFO nova.compute.claims [None req-775eaf93-375d-4ff3-9c57-60f5cc32d4ea tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: a56be6fa-6027-46ee-9aa8-332fe4222d68] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 985.373617] env[65726]: DEBUG nova.compute.manager [None req-83256413-d207-4cc8-b904-cfb2a5ea5628 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] [instance: f9a40d36-dda9-454c-8774-ed011f3f50ae] Start spawning the instance on the hypervisor. {{(pid=65726) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 985.407264] env[65726]: DEBUG nova.virt.hardware [None req-83256413-d207-4cc8-b904-cfb2a5ea5628 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T19:28:34Z,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1539f06fa3534e90acbc3a60c4b8bd3f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T19:28:35Z,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 985.407582] env[65726]: DEBUG nova.virt.hardware [None req-83256413-d207-4cc8-b904-cfb2a5ea5628 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 985.407582] env[65726]: DEBUG nova.virt.hardware [None req-83256413-d207-4cc8-b904-cfb2a5ea5628 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 985.407783] env[65726]: DEBUG nova.virt.hardware [None req-83256413-d207-4cc8-b904-cfb2a5ea5628 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 985.408039] env[65726]: DEBUG nova.virt.hardware [None req-83256413-d207-4cc8-b904-cfb2a5ea5628 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 985.408538] env[65726]: DEBUG nova.virt.hardware [None req-83256413-d207-4cc8-b904-cfb2a5ea5628 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 985.408670] env[65726]: DEBUG nova.virt.hardware [None req-83256413-d207-4cc8-b904-cfb2a5ea5628 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 985.408826] env[65726]: DEBUG nova.virt.hardware [None req-83256413-d207-4cc8-b904-cfb2a5ea5628 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 985.409065] env[65726]: DEBUG nova.virt.hardware [None req-83256413-d207-4cc8-b904-cfb2a5ea5628 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 985.409612] env[65726]: DEBUG nova.virt.hardware [None req-83256413-d207-4cc8-b904-cfb2a5ea5628 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 985.409612] env[65726]: DEBUG nova.virt.hardware [None req-83256413-d207-4cc8-b904-cfb2a5ea5628 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 985.411662] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e3afe62-0aa8-4dc8-bc21-a71f48684e13 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.421720] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7beda210-7d84-4367-aa17-a5c5eb6aa2d5 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.518242] env[65726]: DEBUG oslo_vmware.api [None req-ce025067-bba1-472f-842e-5d842d2328d6 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': task-5116401, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.870082] env[65726]: DEBUG nova.network.neutron [None req-83256413-d207-4cc8-b904-cfb2a5ea5628 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] [instance: f9a40d36-dda9-454c-8774-ed011f3f50ae] Successfully updated port: 4a5e689f-4a77-4c32-81bf-4f86a5c702cb {{(pid=65726) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 985.872449] env[65726]: DEBUG nova.compute.utils [None req-ed96d5ac-cdb1-4748-8154-6accb1ae5e65 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Using /dev/sd instead of None {{(pid=65726) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 985.877447] env[65726]: DEBUG nova.compute.manager [None req-ed96d5ac-cdb1-4748-8154-6accb1ae5e65 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: 067c9332-fa2c-47fe-93ae-643de1fe77cf] Allocating IP information in the background. {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 985.877447] env[65726]: DEBUG nova.network.neutron [None req-ed96d5ac-cdb1-4748-8154-6accb1ae5e65 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: 067c9332-fa2c-47fe-93ae-643de1fe77cf] allocate_for_instance() {{(pid=65726) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 985.877447] env[65726]: WARNING neutronclient.v2_0.client [None req-ed96d5ac-cdb1-4748-8154-6accb1ae5e65 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 985.877735] env[65726]: WARNING neutronclient.v2_0.client [None req-ed96d5ac-cdb1-4748-8154-6accb1ae5e65 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 985.880026] env[65726]: WARNING openstack [None req-ed96d5ac-cdb1-4748-8154-6accb1ae5e65 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 985.880026] env[65726]: WARNING openstack [None req-ed96d5ac-cdb1-4748-8154-6accb1ae5e65 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 985.957895] env[65726]: DEBUG nova.policy [None req-ed96d5ac-cdb1-4748-8154-6accb1ae5e65 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4cd1ed63db174f21be78f86554fada0a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'cbb17a09b35c4c22ade5c4082b10bda9', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65726) authorize /opt/stack/nova/nova/policy.py:192}} [ 986.018475] env[65726]: DEBUG oslo_vmware.api [None req-ce025067-bba1-472f-842e-5d842d2328d6 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': task-5116401, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.220139] env[65726]: DEBUG nova.network.neutron [None req-ed96d5ac-cdb1-4748-8154-6accb1ae5e65 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: 067c9332-fa2c-47fe-93ae-643de1fe77cf] Successfully created port: 498a4766-79c8-471b-b6fc-e73c9eca5c5c {{(pid=65726) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 986.304287] env[65726]: DEBUG nova.network.neutron [None req-c75c9ed0-c9ca-4fd9-9695-87538dcf1f78 tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] [instance: 406c59e7-8eb8-4cfd-8e1e-e151ab908571] Successfully updated port: 903e092f-430c-4ecb-9cc6-1f2adbcd949f {{(pid=65726) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 986.377120] env[65726]: DEBUG nova.compute.manager [None req-ed96d5ac-cdb1-4748-8154-6accb1ae5e65 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: 067c9332-fa2c-47fe-93ae-643de1fe77cf] Start building block device mappings for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 986.380605] env[65726]: DEBUG oslo_concurrency.lockutils [None req-83256413-d207-4cc8-b904-cfb2a5ea5628 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Acquiring lock "refresh_cache-f9a40d36-dda9-454c-8774-ed011f3f50ae" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 986.380793] env[65726]: DEBUG oslo_concurrency.lockutils [None req-83256413-d207-4cc8-b904-cfb2a5ea5628 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Acquired lock "refresh_cache-f9a40d36-dda9-454c-8774-ed011f3f50ae" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 986.380922] env[65726]: DEBUG nova.network.neutron [None req-83256413-d207-4cc8-b904-cfb2a5ea5628 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] [instance: f9a40d36-dda9-454c-8774-ed011f3f50ae] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 986.520226] env[65726]: DEBUG oslo_vmware.api [None req-ce025067-bba1-472f-842e-5d842d2328d6 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': task-5116401, 'name': ReconfigVM_Task, 'duration_secs': 1.028619} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.524507] env[65726]: WARNING openstack [None req-ce025067-bba1-472f-842e-5d842d2328d6 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 986.524912] env[65726]: WARNING openstack [None req-ce025067-bba1-472f-842e-5d842d2328d6 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 986.530127] env[65726]: DEBUG oslo_concurrency.lockutils [None req-ce025067-bba1-472f-842e-5d842d2328d6 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Releasing lock "dcde25df-87a5-47fb-94ce-334f68894e04" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 986.530364] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-ce025067-bba1-472f-842e-5d842d2328d6 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: dcde25df-87a5-47fb-94ce-334f68894e04] Reconfigured VM to attach interface {{(pid=65726) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 986.555500] env[65726]: DEBUG nova.compute.manager [req-30ca328e-963c-4bcd-85c3-27f6f8fa45f1 req-994e98a5-9ff8-4757-b18a-09f30709700e service nova] [instance: dcde25df-87a5-47fb-94ce-334f68894e04] Received event network-changed-b48313f9-a0d3-4fdd-88c6-d103aef0b4f2 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 986.556180] env[65726]: DEBUG nova.compute.manager [req-30ca328e-963c-4bcd-85c3-27f6f8fa45f1 req-994e98a5-9ff8-4757-b18a-09f30709700e service nova] [instance: dcde25df-87a5-47fb-94ce-334f68894e04] Refreshing instance network info cache due to event network-changed-b48313f9-a0d3-4fdd-88c6-d103aef0b4f2. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 986.556590] env[65726]: DEBUG oslo_concurrency.lockutils [req-30ca328e-963c-4bcd-85c3-27f6f8fa45f1 req-994e98a5-9ff8-4757-b18a-09f30709700e service nova] Acquiring lock "refresh_cache-dcde25df-87a5-47fb-94ce-334f68894e04" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 986.556654] env[65726]: DEBUG oslo_concurrency.lockutils [req-30ca328e-963c-4bcd-85c3-27f6f8fa45f1 req-994e98a5-9ff8-4757-b18a-09f30709700e service nova] Acquired lock "refresh_cache-dcde25df-87a5-47fb-94ce-334f68894e04" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 986.556819] env[65726]: DEBUG nova.network.neutron [req-30ca328e-963c-4bcd-85c3-27f6f8fa45f1 req-994e98a5-9ff8-4757-b18a-09f30709700e service nova] [instance: dcde25df-87a5-47fb-94ce-334f68894e04] Refreshing network info cache for port b48313f9-a0d3-4fdd-88c6-d103aef0b4f2 {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 986.573920] env[65726]: DEBUG nova.compute.manager [req-6f4adb83-5202-46e8-b0c9-13fe0430929e req-980d871b-3e3f-43fe-b662-46b3f641f122 service nova] [instance: 406c59e7-8eb8-4cfd-8e1e-e151ab908571] Received event network-vif-plugged-eb554e53-6a49-4f06-8271-81ad5c918184 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 986.573920] env[65726]: DEBUG oslo_concurrency.lockutils [req-6f4adb83-5202-46e8-b0c9-13fe0430929e req-980d871b-3e3f-43fe-b662-46b3f641f122 service nova] Acquiring lock "406c59e7-8eb8-4cfd-8e1e-e151ab908571-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 986.573920] env[65726]: DEBUG oslo_concurrency.lockutils [req-6f4adb83-5202-46e8-b0c9-13fe0430929e req-980d871b-3e3f-43fe-b662-46b3f641f122 service nova] Lock "406c59e7-8eb8-4cfd-8e1e-e151ab908571-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 986.573920] env[65726]: DEBUG oslo_concurrency.lockutils [req-6f4adb83-5202-46e8-b0c9-13fe0430929e req-980d871b-3e3f-43fe-b662-46b3f641f122 service nova] Lock "406c59e7-8eb8-4cfd-8e1e-e151ab908571-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 986.573920] env[65726]: DEBUG nova.compute.manager [req-6f4adb83-5202-46e8-b0c9-13fe0430929e req-980d871b-3e3f-43fe-b662-46b3f641f122 service nova] [instance: 406c59e7-8eb8-4cfd-8e1e-e151ab908571] No waiting events found dispatching network-vif-plugged-eb554e53-6a49-4f06-8271-81ad5c918184 {{(pid=65726) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 986.573920] env[65726]: WARNING nova.compute.manager [req-6f4adb83-5202-46e8-b0c9-13fe0430929e req-980d871b-3e3f-43fe-b662-46b3f641f122 service nova] [instance: 406c59e7-8eb8-4cfd-8e1e-e151ab908571] Received unexpected event network-vif-plugged-eb554e53-6a49-4f06-8271-81ad5c918184 for instance with vm_state building and task_state spawning. [ 986.744540] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b559d4e0-841c-4360-90da-a2f6ad678c50 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.753354] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5c1df5f-cb24-4471-9d16-6c056e1986e2 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.787788] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b630fda-ae5a-46be-9129-4b3d264b671d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.797394] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc681f75-3b0a-49ce-a3cb-b8176ad93a0c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.817315] env[65726]: DEBUG nova.compute.provider_tree [None req-775eaf93-375d-4ff3-9c57-60f5cc32d4ea tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 986.883730] env[65726]: WARNING openstack [None req-83256413-d207-4cc8-b904-cfb2a5ea5628 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 986.884296] env[65726]: WARNING openstack [None req-83256413-d207-4cc8-b904-cfb2a5ea5628 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 986.890381] env[65726]: DEBUG nova.network.neutron [None req-83256413-d207-4cc8-b904-cfb2a5ea5628 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] [instance: f9a40d36-dda9-454c-8774-ed011f3f50ae] Instance cache missing network info. {{(pid=65726) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 987.035603] env[65726]: DEBUG oslo_concurrency.lockutils [None req-ce025067-bba1-472f-842e-5d842d2328d6 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Lock "interface-dcde25df-87a5-47fb-94ce-334f68894e04-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 9.328s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 987.064911] env[65726]: DEBUG oslo_vmware.rw_handles [None req-8940e544-9cc2-4a18-9dd0-e7f396520ac6 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52e1447a-555d-c02a-cc2d-96e5d4ae2965/disk-0.vmdk. {{(pid=65726) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 987.065840] env[65726]: WARNING openstack [req-30ca328e-963c-4bcd-85c3-27f6f8fa45f1 req-994e98a5-9ff8-4757-b18a-09f30709700e service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 987.066200] env[65726]: WARNING openstack [req-30ca328e-963c-4bcd-85c3-27f6f8fa45f1 req-994e98a5-9ff8-4757-b18a-09f30709700e service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 987.073380] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c820db0-d393-454c-be55-348f5299449b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.080359] env[65726]: DEBUG oslo_vmware.rw_handles [None req-8940e544-9cc2-4a18-9dd0-e7f396520ac6 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52e1447a-555d-c02a-cc2d-96e5d4ae2965/disk-0.vmdk is in state: ready. {{(pid=65726) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 987.080521] env[65726]: ERROR oslo_vmware.rw_handles [None req-8940e544-9cc2-4a18-9dd0-e7f396520ac6 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52e1447a-555d-c02a-cc2d-96e5d4ae2965/disk-0.vmdk due to incomplete transfer. [ 987.081053] env[65726]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-5af11da2-b49e-41be-aaf9-dc817a9d84e7 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.090665] env[65726]: DEBUG oslo_vmware.rw_handles [None req-8940e544-9cc2-4a18-9dd0-e7f396520ac6 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52e1447a-555d-c02a-cc2d-96e5d4ae2965/disk-0.vmdk. {{(pid=65726) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 987.090988] env[65726]: DEBUG nova.virt.vmwareapi.images [None req-8940e544-9cc2-4a18-9dd0-e7f396520ac6 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04] Uploaded image 81f5ce39-0bc0-4d0e-ab15-42a80a4f614b to the Glance image server {{(pid=65726) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 987.095197] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-8940e544-9cc2-4a18-9dd0-e7f396520ac6 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04] Destroying the VM {{(pid=65726) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 987.096818] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-ae6396ce-0ce8-463c-97b8-e152c393eb32 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.110945] env[65726]: DEBUG oslo_vmware.api [None req-8940e544-9cc2-4a18-9dd0-e7f396520ac6 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Waiting for the task: (returnval){ [ 987.110945] env[65726]: value = "task-5116402" [ 987.110945] env[65726]: _type = "Task" [ 987.110945] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.121769] env[65726]: DEBUG oslo_vmware.api [None req-8940e544-9cc2-4a18-9dd0-e7f396520ac6 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Task: {'id': task-5116402, 'name': Destroy_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.321548] env[65726]: DEBUG nova.scheduler.client.report [None req-775eaf93-375d-4ff3-9c57-60f5cc32d4ea tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 987.398264] env[65726]: DEBUG nova.compute.manager [None req-ed96d5ac-cdb1-4748-8154-6accb1ae5e65 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: 067c9332-fa2c-47fe-93ae-643de1fe77cf] Start spawning the instance on the hypervisor. {{(pid=65726) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 987.425580] env[65726]: DEBUG nova.virt.hardware [None req-ed96d5ac-cdb1-4748-8154-6accb1ae5e65 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T19:28:34Z,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1539f06fa3534e90acbc3a60c4b8bd3f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T19:28:35Z,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 987.425962] env[65726]: DEBUG nova.virt.hardware [None req-ed96d5ac-cdb1-4748-8154-6accb1ae5e65 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 987.426123] env[65726]: DEBUG nova.virt.hardware [None req-ed96d5ac-cdb1-4748-8154-6accb1ae5e65 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 987.426421] env[65726]: DEBUG nova.virt.hardware [None req-ed96d5ac-cdb1-4748-8154-6accb1ae5e65 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 987.426565] env[65726]: DEBUG nova.virt.hardware [None req-ed96d5ac-cdb1-4748-8154-6accb1ae5e65 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 987.426763] env[65726]: DEBUG nova.virt.hardware [None req-ed96d5ac-cdb1-4748-8154-6accb1ae5e65 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 987.427030] env[65726]: DEBUG nova.virt.hardware [None req-ed96d5ac-cdb1-4748-8154-6accb1ae5e65 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 987.427247] env[65726]: DEBUG nova.virt.hardware [None req-ed96d5ac-cdb1-4748-8154-6accb1ae5e65 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 987.427628] env[65726]: DEBUG nova.virt.hardware [None req-ed96d5ac-cdb1-4748-8154-6accb1ae5e65 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 987.427685] env[65726]: DEBUG nova.virt.hardware [None req-ed96d5ac-cdb1-4748-8154-6accb1ae5e65 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 987.427894] env[65726]: DEBUG nova.virt.hardware [None req-ed96d5ac-cdb1-4748-8154-6accb1ae5e65 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 987.429016] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56f4868e-2a6f-4d00-9c06-9064b9be6f77 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.439285] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92dc4cd1-2c69-4556-84c2-de932099bcd0 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.625105] env[65726]: DEBUG oslo_vmware.api [None req-8940e544-9cc2-4a18-9dd0-e7f396520ac6 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Task: {'id': task-5116402, 'name': Destroy_Task} progress is 33%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.829450] env[65726]: DEBUG oslo_concurrency.lockutils [None req-775eaf93-375d-4ff3-9c57-60f5cc32d4ea tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.467s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 987.830154] env[65726]: DEBUG nova.compute.manager [None req-775eaf93-375d-4ff3-9c57-60f5cc32d4ea tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: a56be6fa-6027-46ee-9aa8-332fe4222d68] Start building networks asynchronously for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 987.833577] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 16.200s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 987.833657] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 987.833814] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=65726) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:937}} [ 987.834129] env[65726]: DEBUG oslo_concurrency.lockutils [None req-618baf85-0896-4fc2-b1ef-24618b3b2fb7 tempest-FloatingIPsAssociationNegativeTestJSON-737087291 tempest-FloatingIPsAssociationNegativeTestJSON-737087291-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.042s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 987.836039] env[65726]: INFO nova.compute.claims [None req-618baf85-0896-4fc2-b1ef-24618b3b2fb7 tempest-FloatingIPsAssociationNegativeTestJSON-737087291 tempest-FloatingIPsAssociationNegativeTestJSON-737087291-project-member] [instance: d8468efa-c2d3-4dce-ab89-fc077011e3d8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 987.839255] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84a5d565-9acc-441b-b4ed-0c589d766996 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.849342] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c49be36-0a85-4c53-b644-63f7b5c15cbc {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.855941] env[65726]: DEBUG nova.network.neutron [None req-ed96d5ac-cdb1-4748-8154-6accb1ae5e65 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: 067c9332-fa2c-47fe-93ae-643de1fe77cf] Successfully updated port: 498a4766-79c8-471b-b6fc-e73c9eca5c5c {{(pid=65726) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 987.868786] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f26a2f33-bb58-4903-a211-d7e59e5285fd {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.878566] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45df102b-35eb-46be-b94c-4ca776470f41 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.915145] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=177937MB free_disk=95GB free_vcpus=48 pci_devices=None {{(pid=65726) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1136}} [ 987.915312] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 987.974656] env[65726]: WARNING openstack [None req-83256413-d207-4cc8-b904-cfb2a5ea5628 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 987.974656] env[65726]: WARNING openstack [None req-83256413-d207-4cc8-b904-cfb2a5ea5628 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 988.064108] env[65726]: WARNING openstack [req-30ca328e-963c-4bcd-85c3-27f6f8fa45f1 req-994e98a5-9ff8-4757-b18a-09f30709700e service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 988.064741] env[65726]: WARNING openstack [req-30ca328e-963c-4bcd-85c3-27f6f8fa45f1 req-994e98a5-9ff8-4757-b18a-09f30709700e service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 988.092363] env[65726]: WARNING openstack [None req-83256413-d207-4cc8-b904-cfb2a5ea5628 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 988.092896] env[65726]: WARNING openstack [None req-83256413-d207-4cc8-b904-cfb2a5ea5628 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 988.120636] env[65726]: DEBUG nova.compute.manager [None req-4e63de1e-de7d-4372-bcfb-4a4336e247fa tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 54e85161-5f63-405f-83f8-490c70645a3b] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 988.120771] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-232eda53-9326-4f1a-9bc6-e4e085c21918 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.135348] env[65726]: DEBUG oslo_vmware.api [None req-8940e544-9cc2-4a18-9dd0-e7f396520ac6 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Task: {'id': task-5116402, 'name': Destroy_Task} progress is 33%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.249869] env[65726]: WARNING openstack [req-30ca328e-963c-4bcd-85c3-27f6f8fa45f1 req-994e98a5-9ff8-4757-b18a-09f30709700e service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 988.250481] env[65726]: WARNING openstack [req-30ca328e-963c-4bcd-85c3-27f6f8fa45f1 req-994e98a5-9ff8-4757-b18a-09f30709700e service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 988.340785] env[65726]: DEBUG nova.compute.utils [None req-775eaf93-375d-4ff3-9c57-60f5cc32d4ea tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Using /dev/sd instead of None {{(pid=65726) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 988.342810] env[65726]: DEBUG nova.compute.manager [None req-775eaf93-375d-4ff3-9c57-60f5cc32d4ea tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: a56be6fa-6027-46ee-9aa8-332fe4222d68] Allocating IP information in the background. {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 988.343052] env[65726]: DEBUG nova.network.neutron [None req-775eaf93-375d-4ff3-9c57-60f5cc32d4ea tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: a56be6fa-6027-46ee-9aa8-332fe4222d68] allocate_for_instance() {{(pid=65726) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 988.343392] env[65726]: WARNING neutronclient.v2_0.client [None req-775eaf93-375d-4ff3-9c57-60f5cc32d4ea tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 988.343695] env[65726]: WARNING neutronclient.v2_0.client [None req-775eaf93-375d-4ff3-9c57-60f5cc32d4ea tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 988.344392] env[65726]: WARNING openstack [None req-775eaf93-375d-4ff3-9c57-60f5cc32d4ea tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 988.344653] env[65726]: WARNING openstack [None req-775eaf93-375d-4ff3-9c57-60f5cc32d4ea tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 988.358942] env[65726]: DEBUG nova.network.neutron [None req-83256413-d207-4cc8-b904-cfb2a5ea5628 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] [instance: f9a40d36-dda9-454c-8774-ed011f3f50ae] Updating instance_info_cache with network_info: [{"id": "4a5e689f-4a77-4c32-81bf-4f86a5c702cb", "address": "fa:16:3e:be:de:b1", "network": {"id": "93e38d7e-b4d7-4d4a-8393-64ab2f6769de", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1445539922-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "06e221dc693640929cac7bc5af45948f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "399f3826-705c-45f7-9fe0-3a08a945151a", "external-id": "nsx-vlan-transportzone-936", "segmentation_id": 936, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4a5e689f-4a", "ovs_interfaceid": "4a5e689f-4a77-4c32-81bf-4f86a5c702cb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 988.360339] env[65726]: DEBUG oslo_concurrency.lockutils [None req-ed96d5ac-cdb1-4748-8154-6accb1ae5e65 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Acquiring lock "refresh_cache-067c9332-fa2c-47fe-93ae-643de1fe77cf" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 988.360496] env[65726]: DEBUG oslo_concurrency.lockutils [None req-ed96d5ac-cdb1-4748-8154-6accb1ae5e65 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Acquired lock "refresh_cache-067c9332-fa2c-47fe-93ae-643de1fe77cf" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 988.360659] env[65726]: DEBUG nova.network.neutron [None req-ed96d5ac-cdb1-4748-8154-6accb1ae5e65 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: 067c9332-fa2c-47fe-93ae-643de1fe77cf] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 988.406629] env[65726]: DEBUG nova.network.neutron [None req-c75c9ed0-c9ca-4fd9-9695-87538dcf1f78 tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] [instance: 406c59e7-8eb8-4cfd-8e1e-e151ab908571] Successfully updated port: b0ae20cc-cb0a-450d-96a3-4726b64254db {{(pid=65726) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 988.447413] env[65726]: DEBUG nova.policy [None req-775eaf93-375d-4ff3-9c57-60f5cc32d4ea tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b251fcbd04044fcda192bc3febac1a90', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4eece77569624f90bf64e5c51974173f', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65726) authorize /opt/stack/nova/nova/policy.py:192}} [ 988.472730] env[65726]: DEBUG nova.network.neutron [req-30ca328e-963c-4bcd-85c3-27f6f8fa45f1 req-994e98a5-9ff8-4757-b18a-09f30709700e service nova] [instance: dcde25df-87a5-47fb-94ce-334f68894e04] Updated VIF entry in instance network info cache for port b48313f9-a0d3-4fdd-88c6-d103aef0b4f2. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 988.473170] env[65726]: DEBUG nova.network.neutron [req-30ca328e-963c-4bcd-85c3-27f6f8fa45f1 req-994e98a5-9ff8-4757-b18a-09f30709700e service nova] [instance: dcde25df-87a5-47fb-94ce-334f68894e04] Updating instance_info_cache with network_info: [{"id": "bf51c96c-63c5-46de-a40f-58e6769c5fcf", "address": "fa:16:3e:4c:88:87", "network": {"id": "51a66448-0f72-4266-8868-856b48be839f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-155658961-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.146", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d46ccb00794f458b85da4a93879139ee", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b1f3e6c3-5584-4852-9017-476ab8ac4946", "external-id": "nsx-vlan-transportzone-304", "segmentation_id": 304, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbf51c96c-63", "ovs_interfaceid": "bf51c96c-63c5-46de-a40f-58e6769c5fcf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "b48313f9-a0d3-4fdd-88c6-d103aef0b4f2", "address": "fa:16:3e:e7:cd:41", "network": {"id": "51a66448-0f72-4266-8868-856b48be839f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-155658961-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d46ccb00794f458b85da4a93879139ee", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b1f3e6c3-5584-4852-9017-476ab8ac4946", "external-id": "nsx-vlan-transportzone-304", "segmentation_id": 304, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb48313f9-a0", "ovs_interfaceid": "b48313f9-a0d3-4fdd-88c6-d103aef0b4f2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 988.626464] env[65726]: DEBUG oslo_vmware.api [None req-8940e544-9cc2-4a18-9dd0-e7f396520ac6 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Task: {'id': task-5116402, 'name': Destroy_Task, 'duration_secs': 1.236979} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.627480] env[65726]: INFO nova.virt.vmwareapi.vm_util [None req-8940e544-9cc2-4a18-9dd0-e7f396520ac6 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04] Destroyed the VM [ 988.627480] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-8940e544-9cc2-4a18-9dd0-e7f396520ac6 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04] Deleting Snapshot of the VM instance {{(pid=65726) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 988.627480] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-57d158a5-a2f0-4b1f-82aa-5e5ec467f23f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.635173] env[65726]: DEBUG oslo_vmware.api [None req-8940e544-9cc2-4a18-9dd0-e7f396520ac6 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Waiting for the task: (returnval){ [ 988.635173] env[65726]: value = "task-5116403" [ 988.635173] env[65726]: _type = "Task" [ 988.635173] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 988.640861] env[65726]: INFO nova.compute.manager [None req-4e63de1e-de7d-4372-bcfb-4a4336e247fa tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 54e85161-5f63-405f-83f8-490c70645a3b] instance snapshotting [ 988.646089] env[65726]: DEBUG oslo_vmware.api [None req-8940e544-9cc2-4a18-9dd0-e7f396520ac6 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Task: {'id': task-5116403, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.647563] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-832eba79-e7af-48d2-8341-942555ffb621 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.672722] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3909ff2-bd75-40bf-acf7-f1ca2e5d6cfb {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.759333] env[65726]: DEBUG nova.network.neutron [None req-775eaf93-375d-4ff3-9c57-60f5cc32d4ea tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: a56be6fa-6027-46ee-9aa8-332fe4222d68] Successfully created port: 4e2dfbcd-41c3-4aaa-a2ec-50c855c12307 {{(pid=65726) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 988.782712] env[65726]: DEBUG nova.compute.manager [req-4618e1ae-ab3f-4f85-99b5-06f2d52ca99b req-6b1ef7a1-7299-445a-8fac-55ecea913381 service nova] [instance: 406c59e7-8eb8-4cfd-8e1e-e151ab908571] Received event network-changed-eb554e53-6a49-4f06-8271-81ad5c918184 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 988.782902] env[65726]: DEBUG nova.compute.manager [req-4618e1ae-ab3f-4f85-99b5-06f2d52ca99b req-6b1ef7a1-7299-445a-8fac-55ecea913381 service nova] [instance: 406c59e7-8eb8-4cfd-8e1e-e151ab908571] Refreshing instance network info cache due to event network-changed-eb554e53-6a49-4f06-8271-81ad5c918184. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 988.783262] env[65726]: DEBUG oslo_concurrency.lockutils [req-4618e1ae-ab3f-4f85-99b5-06f2d52ca99b req-6b1ef7a1-7299-445a-8fac-55ecea913381 service nova] Acquiring lock "refresh_cache-406c59e7-8eb8-4cfd-8e1e-e151ab908571" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 988.783759] env[65726]: DEBUG oslo_concurrency.lockutils [req-4618e1ae-ab3f-4f85-99b5-06f2d52ca99b req-6b1ef7a1-7299-445a-8fac-55ecea913381 service nova] Acquired lock "refresh_cache-406c59e7-8eb8-4cfd-8e1e-e151ab908571" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 988.783759] env[65726]: DEBUG nova.network.neutron [req-4618e1ae-ab3f-4f85-99b5-06f2d52ca99b req-6b1ef7a1-7299-445a-8fac-55ecea913381 service nova] [instance: 406c59e7-8eb8-4cfd-8e1e-e151ab908571] Refreshing network info cache for port eb554e53-6a49-4f06-8271-81ad5c918184 {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 988.790628] env[65726]: DEBUG nova.compute.manager [req-2eee699f-12e2-486b-9b3a-bb9917f139d6 req-d5dc1313-04fd-4c83-b260-7d864ae121eb service nova] [instance: f9a40d36-dda9-454c-8774-ed011f3f50ae] Received event network-vif-plugged-4a5e689f-4a77-4c32-81bf-4f86a5c702cb {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 988.790628] env[65726]: DEBUG oslo_concurrency.lockutils [req-2eee699f-12e2-486b-9b3a-bb9917f139d6 req-d5dc1313-04fd-4c83-b260-7d864ae121eb service nova] Acquiring lock "f9a40d36-dda9-454c-8774-ed011f3f50ae-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 988.790628] env[65726]: DEBUG oslo_concurrency.lockutils [req-2eee699f-12e2-486b-9b3a-bb9917f139d6 req-d5dc1313-04fd-4c83-b260-7d864ae121eb service nova] Lock "f9a40d36-dda9-454c-8774-ed011f3f50ae-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 988.790628] env[65726]: DEBUG oslo_concurrency.lockutils [req-2eee699f-12e2-486b-9b3a-bb9917f139d6 req-d5dc1313-04fd-4c83-b260-7d864ae121eb service nova] Lock "f9a40d36-dda9-454c-8774-ed011f3f50ae-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 988.790628] env[65726]: DEBUG nova.compute.manager [req-2eee699f-12e2-486b-9b3a-bb9917f139d6 req-d5dc1313-04fd-4c83-b260-7d864ae121eb service nova] [instance: f9a40d36-dda9-454c-8774-ed011f3f50ae] No waiting events found dispatching network-vif-plugged-4a5e689f-4a77-4c32-81bf-4f86a5c702cb {{(pid=65726) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 988.790628] env[65726]: WARNING nova.compute.manager [req-2eee699f-12e2-486b-9b3a-bb9917f139d6 req-d5dc1313-04fd-4c83-b260-7d864ae121eb service nova] [instance: f9a40d36-dda9-454c-8774-ed011f3f50ae] Received unexpected event network-vif-plugged-4a5e689f-4a77-4c32-81bf-4f86a5c702cb for instance with vm_state building and task_state spawning. [ 988.790628] env[65726]: DEBUG nova.compute.manager [req-2eee699f-12e2-486b-9b3a-bb9917f139d6 req-d5dc1313-04fd-4c83-b260-7d864ae121eb service nova] [instance: f9a40d36-dda9-454c-8774-ed011f3f50ae] Received event network-changed-4a5e689f-4a77-4c32-81bf-4f86a5c702cb {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 988.790628] env[65726]: DEBUG nova.compute.manager [req-2eee699f-12e2-486b-9b3a-bb9917f139d6 req-d5dc1313-04fd-4c83-b260-7d864ae121eb service nova] [instance: f9a40d36-dda9-454c-8774-ed011f3f50ae] Refreshing instance network info cache due to event network-changed-4a5e689f-4a77-4c32-81bf-4f86a5c702cb. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 988.791141] env[65726]: DEBUG oslo_concurrency.lockutils [req-2eee699f-12e2-486b-9b3a-bb9917f139d6 req-d5dc1313-04fd-4c83-b260-7d864ae121eb service nova] Acquiring lock "refresh_cache-f9a40d36-dda9-454c-8774-ed011f3f50ae" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 988.852669] env[65726]: DEBUG nova.compute.manager [None req-775eaf93-375d-4ff3-9c57-60f5cc32d4ea tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: a56be6fa-6027-46ee-9aa8-332fe4222d68] Start building block device mappings for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 988.863552] env[65726]: WARNING openstack [None req-ed96d5ac-cdb1-4748-8154-6accb1ae5e65 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 988.864194] env[65726]: WARNING openstack [None req-ed96d5ac-cdb1-4748-8154-6accb1ae5e65 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 988.869643] env[65726]: DEBUG nova.network.neutron [None req-ed96d5ac-cdb1-4748-8154-6accb1ae5e65 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: 067c9332-fa2c-47fe-93ae-643de1fe77cf] Instance cache missing network info. {{(pid=65726) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 988.871657] env[65726]: DEBUG oslo_concurrency.lockutils [None req-83256413-d207-4cc8-b904-cfb2a5ea5628 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Releasing lock "refresh_cache-f9a40d36-dda9-454c-8774-ed011f3f50ae" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 988.872070] env[65726]: DEBUG nova.compute.manager [None req-83256413-d207-4cc8-b904-cfb2a5ea5628 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] [instance: f9a40d36-dda9-454c-8774-ed011f3f50ae] Instance network_info: |[{"id": "4a5e689f-4a77-4c32-81bf-4f86a5c702cb", "address": "fa:16:3e:be:de:b1", "network": {"id": "93e38d7e-b4d7-4d4a-8393-64ab2f6769de", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1445539922-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "06e221dc693640929cac7bc5af45948f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "399f3826-705c-45f7-9fe0-3a08a945151a", "external-id": "nsx-vlan-transportzone-936", "segmentation_id": 936, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4a5e689f-4a", "ovs_interfaceid": "4a5e689f-4a77-4c32-81bf-4f86a5c702cb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 988.873367] env[65726]: DEBUG oslo_concurrency.lockutils [req-2eee699f-12e2-486b-9b3a-bb9917f139d6 req-d5dc1313-04fd-4c83-b260-7d864ae121eb service nova] Acquired lock "refresh_cache-f9a40d36-dda9-454c-8774-ed011f3f50ae" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 988.873551] env[65726]: DEBUG nova.network.neutron [req-2eee699f-12e2-486b-9b3a-bb9917f139d6 req-d5dc1313-04fd-4c83-b260-7d864ae121eb service nova] [instance: f9a40d36-dda9-454c-8774-ed011f3f50ae] Refreshing network info cache for port 4a5e689f-4a77-4c32-81bf-4f86a5c702cb {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 988.874794] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-83256413-d207-4cc8-b904-cfb2a5ea5628 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] [instance: f9a40d36-dda9-454c-8774-ed011f3f50ae] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:be:de:b1', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '399f3826-705c-45f7-9fe0-3a08a945151a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4a5e689f-4a77-4c32-81bf-4f86a5c702cb', 'vif_model': 'vmxnet3'}] {{(pid=65726) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 988.884373] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-83256413-d207-4cc8-b904-cfb2a5ea5628 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 988.884955] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f9a40d36-dda9-454c-8774-ed011f3f50ae] Creating VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 988.885164] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-27928b7b-3b25-4827-ac86-778fab832831 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.910868] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c75c9ed0-c9ca-4fd9-9695-87538dcf1f78 tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Acquiring lock "refresh_cache-406c59e7-8eb8-4cfd-8e1e-e151ab908571" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 988.915782] env[65726]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 988.915782] env[65726]: value = "task-5116404" [ 988.915782] env[65726]: _type = "Task" [ 988.915782] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 988.932621] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116404, 'name': CreateVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.977472] env[65726]: DEBUG oslo_concurrency.lockutils [req-30ca328e-963c-4bcd-85c3-27f6f8fa45f1 req-994e98a5-9ff8-4757-b18a-09f30709700e service nova] Releasing lock "refresh_cache-dcde25df-87a5-47fb-94ce-334f68894e04" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 988.977860] env[65726]: DEBUG nova.compute.manager [req-30ca328e-963c-4bcd-85c3-27f6f8fa45f1 req-994e98a5-9ff8-4757-b18a-09f30709700e service nova] [instance: aa6ce489-c62f-4481-87b7-e74242aeb8ca] Received event network-changed-d5a805f5-b6a8-42ec-92a0-4a652b51c1e6 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 988.978044] env[65726]: DEBUG nova.compute.manager [req-30ca328e-963c-4bcd-85c3-27f6f8fa45f1 req-994e98a5-9ff8-4757-b18a-09f30709700e service nova] [instance: aa6ce489-c62f-4481-87b7-e74242aeb8ca] Refreshing instance network info cache due to event network-changed-d5a805f5-b6a8-42ec-92a0-4a652b51c1e6. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 988.978269] env[65726]: DEBUG oslo_concurrency.lockutils [req-30ca328e-963c-4bcd-85c3-27f6f8fa45f1 req-994e98a5-9ff8-4757-b18a-09f30709700e service nova] Acquiring lock "refresh_cache-aa6ce489-c62f-4481-87b7-e74242aeb8ca" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 988.978410] env[65726]: DEBUG oslo_concurrency.lockutils [req-30ca328e-963c-4bcd-85c3-27f6f8fa45f1 req-994e98a5-9ff8-4757-b18a-09f30709700e service nova] Acquired lock "refresh_cache-aa6ce489-c62f-4481-87b7-e74242aeb8ca" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 988.978845] env[65726]: DEBUG nova.network.neutron [req-30ca328e-963c-4bcd-85c3-27f6f8fa45f1 req-994e98a5-9ff8-4757-b18a-09f30709700e service nova] [instance: aa6ce489-c62f-4481-87b7-e74242aeb8ca] Refreshing network info cache for port d5a805f5-b6a8-42ec-92a0-4a652b51c1e6 {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 989.147009] env[65726]: DEBUG oslo_vmware.api [None req-8940e544-9cc2-4a18-9dd0-e7f396520ac6 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Task: {'id': task-5116403, 'name': RemoveSnapshot_Task, 'duration_secs': 0.449124} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 989.150937] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-8940e544-9cc2-4a18-9dd0-e7f396520ac6 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04] Deleted Snapshot of the VM instance {{(pid=65726) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 989.151424] env[65726]: DEBUG nova.compute.manager [None req-8940e544-9cc2-4a18-9dd0-e7f396520ac6 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 989.153093] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6eab67a6-f172-44d5-912c-1f85ea3fb10d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.187208] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-4e63de1e-de7d-4372-bcfb-4a4336e247fa tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 54e85161-5f63-405f-83f8-490c70645a3b] Creating Snapshot of the VM instance {{(pid=65726) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 989.192515] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-f4a0005a-adb4-483a-9333-93eae9604efa {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.201040] env[65726]: DEBUG oslo_vmware.api [None req-4e63de1e-de7d-4372-bcfb-4a4336e247fa tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Waiting for the task: (returnval){ [ 989.201040] env[65726]: value = "task-5116405" [ 989.201040] env[65726]: _type = "Task" [ 989.201040] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 989.216428] env[65726]: DEBUG oslo_vmware.api [None req-4e63de1e-de7d-4372-bcfb-4a4336e247fa tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': task-5116405, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.271844] env[65726]: WARNING openstack [None req-ed96d5ac-cdb1-4748-8154-6accb1ae5e65 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 989.272722] env[65726]: WARNING openstack [None req-ed96d5ac-cdb1-4748-8154-6accb1ae5e65 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 989.283270] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b47f7e32-b19a-4f34-a750-d376fea42fd9 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.291250] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b63eca95-49d1-448e-b27f-68467494c5dd {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.295703] env[65726]: WARNING openstack [req-4618e1ae-ab3f-4f85-99b5-06f2d52ca99b req-6b1ef7a1-7299-445a-8fac-55ecea913381 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 989.296163] env[65726]: WARNING openstack [req-4618e1ae-ab3f-4f85-99b5-06f2d52ca99b req-6b1ef7a1-7299-445a-8fac-55ecea913381 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 989.301160] env[65726]: DEBUG nova.network.neutron [req-4618e1ae-ab3f-4f85-99b5-06f2d52ca99b req-6b1ef7a1-7299-445a-8fac-55ecea913381 service nova] [instance: 406c59e7-8eb8-4cfd-8e1e-e151ab908571] Instance cache missing network info. {{(pid=65726) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 989.331027] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c35d764f-9848-450a-b597-90a73140bd02 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.339544] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a6711af-6eaf-4efd-b1ad-a4096cd75462 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.356109] env[65726]: DEBUG nova.compute.provider_tree [None req-618baf85-0896-4fc2-b1ef-24618b3b2fb7 tempest-FloatingIPsAssociationNegativeTestJSON-737087291 tempest-FloatingIPsAssociationNegativeTestJSON-737087291-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 989.386804] env[65726]: WARNING openstack [req-2eee699f-12e2-486b-9b3a-bb9917f139d6 req-d5dc1313-04fd-4c83-b260-7d864ae121eb service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 989.386804] env[65726]: WARNING openstack [req-2eee699f-12e2-486b-9b3a-bb9917f139d6 req-d5dc1313-04fd-4c83-b260-7d864ae121eb service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 989.428040] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116404, 'name': CreateVM_Task, 'duration_secs': 0.370496} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 989.428040] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f9a40d36-dda9-454c-8774-ed011f3f50ae] Created VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 989.428549] env[65726]: WARNING openstack [None req-83256413-d207-4cc8-b904-cfb2a5ea5628 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 989.428904] env[65726]: WARNING openstack [None req-83256413-d207-4cc8-b904-cfb2a5ea5628 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 989.434500] env[65726]: DEBUG oslo_concurrency.lockutils [None req-83256413-d207-4cc8-b904-cfb2a5ea5628 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 989.434672] env[65726]: DEBUG oslo_concurrency.lockutils [None req-83256413-d207-4cc8-b904-cfb2a5ea5628 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 989.434988] env[65726]: DEBUG oslo_concurrency.lockutils [None req-83256413-d207-4cc8-b904-cfb2a5ea5628 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 989.435632] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ebf12481-860a-4c89-81d4-6b9fbdf971b7 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.441305] env[65726]: DEBUG oslo_vmware.api [None req-83256413-d207-4cc8-b904-cfb2a5ea5628 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Waiting for the task: (returnval){ [ 989.441305] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52041b30-6dce-1884-de79-3112b6f77e45" [ 989.441305] env[65726]: _type = "Task" [ 989.441305] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 989.461992] env[65726]: DEBUG oslo_vmware.api [None req-83256413-d207-4cc8-b904-cfb2a5ea5628 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52041b30-6dce-1884-de79-3112b6f77e45, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.486173] env[65726]: WARNING openstack [req-30ca328e-963c-4bcd-85c3-27f6f8fa45f1 req-994e98a5-9ff8-4757-b18a-09f30709700e service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 989.487492] env[65726]: WARNING openstack [req-30ca328e-963c-4bcd-85c3-27f6f8fa45f1 req-994e98a5-9ff8-4757-b18a-09f30709700e service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 989.506877] env[65726]: WARNING openstack [None req-ed96d5ac-cdb1-4748-8154-6accb1ae5e65 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 989.507282] env[65726]: WARNING openstack [None req-ed96d5ac-cdb1-4748-8154-6accb1ae5e65 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 989.671029] env[65726]: INFO nova.compute.manager [None req-8940e544-9cc2-4a18-9dd0-e7f396520ac6 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04] Shelve offloading [ 989.712234] env[65726]: DEBUG oslo_vmware.api [None req-4e63de1e-de7d-4372-bcfb-4a4336e247fa tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': task-5116405, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.778489] env[65726]: DEBUG nova.network.neutron [req-4618e1ae-ab3f-4f85-99b5-06f2d52ca99b req-6b1ef7a1-7299-445a-8fac-55ecea913381 service nova] [instance: 406c59e7-8eb8-4cfd-8e1e-e151ab908571] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 989.792603] env[65726]: WARNING openstack [req-2eee699f-12e2-486b-9b3a-bb9917f139d6 req-d5dc1313-04fd-4c83-b260-7d864ae121eb service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 989.793009] env[65726]: WARNING openstack [req-2eee699f-12e2-486b-9b3a-bb9917f139d6 req-d5dc1313-04fd-4c83-b260-7d864ae121eb service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 989.857084] env[65726]: DEBUG nova.network.neutron [None req-ed96d5ac-cdb1-4748-8154-6accb1ae5e65 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: 067c9332-fa2c-47fe-93ae-643de1fe77cf] Updating instance_info_cache with network_info: [{"id": "498a4766-79c8-471b-b6fc-e73c9eca5c5c", "address": "fa:16:3e:fc:03:ad", "network": {"id": "4c972d9d-db3b-4779-ac39-39d9a2d141dd", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1418913861-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cbb17a09b35c4c22ade5c4082b10bda9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "50cf0a70-948d-4611-af05-94c1483064ed", "external-id": "nsx-vlan-transportzone-536", "segmentation_id": 536, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap498a4766-79", "ovs_interfaceid": "498a4766-79c8-471b-b6fc-e73c9eca5c5c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 989.861801] env[65726]: DEBUG nova.scheduler.client.report [None req-618baf85-0896-4fc2-b1ef-24618b3b2fb7 tempest-FloatingIPsAssociationNegativeTestJSON-737087291 tempest-FloatingIPsAssociationNegativeTestJSON-737087291-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 989.866028] env[65726]: DEBUG nova.compute.manager [None req-775eaf93-375d-4ff3-9c57-60f5cc32d4ea tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: a56be6fa-6027-46ee-9aa8-332fe4222d68] Start spawning the instance on the hypervisor. {{(pid=65726) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 989.898212] env[65726]: DEBUG nova.virt.hardware [None req-775eaf93-375d-4ff3-9c57-60f5cc32d4ea tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T19:28:34Z,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1539f06fa3534e90acbc3a60c4b8bd3f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T19:28:35Z,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 989.898572] env[65726]: DEBUG nova.virt.hardware [None req-775eaf93-375d-4ff3-9c57-60f5cc32d4ea tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 989.898657] env[65726]: DEBUG nova.virt.hardware [None req-775eaf93-375d-4ff3-9c57-60f5cc32d4ea tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 989.898829] env[65726]: DEBUG nova.virt.hardware [None req-775eaf93-375d-4ff3-9c57-60f5cc32d4ea tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 989.898977] env[65726]: DEBUG nova.virt.hardware [None req-775eaf93-375d-4ff3-9c57-60f5cc32d4ea tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 989.899150] env[65726]: DEBUG nova.virt.hardware [None req-775eaf93-375d-4ff3-9c57-60f5cc32d4ea tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 989.899533] env[65726]: DEBUG nova.virt.hardware [None req-775eaf93-375d-4ff3-9c57-60f5cc32d4ea tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 989.899533] env[65726]: DEBUG nova.virt.hardware [None req-775eaf93-375d-4ff3-9c57-60f5cc32d4ea tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 989.900179] env[65726]: DEBUG nova.virt.hardware [None req-775eaf93-375d-4ff3-9c57-60f5cc32d4ea tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 989.900179] env[65726]: DEBUG nova.virt.hardware [None req-775eaf93-375d-4ff3-9c57-60f5cc32d4ea tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 989.900179] env[65726]: DEBUG nova.virt.hardware [None req-775eaf93-375d-4ff3-9c57-60f5cc32d4ea tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 989.900972] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56fc51f1-fbba-4e9e-ac4d-63b5ecca912b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.914856] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84212f07-8b1c-4dd0-b1bb-34b6a857e856 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.943362] env[65726]: WARNING openstack [req-30ca328e-963c-4bcd-85c3-27f6f8fa45f1 req-994e98a5-9ff8-4757-b18a-09f30709700e service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 989.943362] env[65726]: WARNING openstack [req-30ca328e-963c-4bcd-85c3-27f6f8fa45f1 req-994e98a5-9ff8-4757-b18a-09f30709700e service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 989.966232] env[65726]: DEBUG oslo_vmware.api [None req-83256413-d207-4cc8-b904-cfb2a5ea5628 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52041b30-6dce-1884-de79-3112b6f77e45, 'name': SearchDatastore_Task, 'duration_secs': 0.015937} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 989.968563] env[65726]: DEBUG oslo_concurrency.lockutils [None req-83256413-d207-4cc8-b904-cfb2a5ea5628 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 989.968563] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-83256413-d207-4cc8-b904-cfb2a5ea5628 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] [instance: f9a40d36-dda9-454c-8774-ed011f3f50ae] Processing image b52362a3-ee8a-4cbf-b06f-513b0cc8f95c {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 989.968563] env[65726]: DEBUG oslo_concurrency.lockutils [None req-83256413-d207-4cc8-b904-cfb2a5ea5628 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 989.968563] env[65726]: DEBUG oslo_concurrency.lockutils [None req-83256413-d207-4cc8-b904-cfb2a5ea5628 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 989.968563] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-83256413-d207-4cc8-b904-cfb2a5ea5628 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 989.968563] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-08c4b522-b7c2-4d7f-ab6e-64a74008ce42 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.978200] env[65726]: WARNING openstack [req-2eee699f-12e2-486b-9b3a-bb9917f139d6 req-d5dc1313-04fd-4c83-b260-7d864ae121eb service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 989.978577] env[65726]: WARNING openstack [req-2eee699f-12e2-486b-9b3a-bb9917f139d6 req-d5dc1313-04fd-4c83-b260-7d864ae121eb service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 989.989726] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-83256413-d207-4cc8-b904-cfb2a5ea5628 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 989.989954] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-83256413-d207-4cc8-b904-cfb2a5ea5628 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65726) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 989.991506] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-69679518-74e1-4252-8292-82be34da1072 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.998485] env[65726]: DEBUG oslo_vmware.api [None req-83256413-d207-4cc8-b904-cfb2a5ea5628 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Waiting for the task: (returnval){ [ 989.998485] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]525f8c3c-da6b-7a37-5859-c08505933eae" [ 989.998485] env[65726]: _type = "Task" [ 989.998485] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 990.008056] env[65726]: DEBUG oslo_vmware.api [None req-83256413-d207-4cc8-b904-cfb2a5ea5628 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]525f8c3c-da6b-7a37-5859-c08505933eae, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.136203] env[65726]: WARNING openstack [req-30ca328e-963c-4bcd-85c3-27f6f8fa45f1 req-994e98a5-9ff8-4757-b18a-09f30709700e service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 990.136661] env[65726]: WARNING openstack [req-30ca328e-963c-4bcd-85c3-27f6f8fa45f1 req-994e98a5-9ff8-4757-b18a-09f30709700e service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 990.171153] env[65726]: DEBUG nova.network.neutron [req-2eee699f-12e2-486b-9b3a-bb9917f139d6 req-d5dc1313-04fd-4c83-b260-7d864ae121eb service nova] [instance: f9a40d36-dda9-454c-8774-ed011f3f50ae] Updated VIF entry in instance network info cache for port 4a5e689f-4a77-4c32-81bf-4f86a5c702cb. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 990.171683] env[65726]: DEBUG nova.network.neutron [req-2eee699f-12e2-486b-9b3a-bb9917f139d6 req-d5dc1313-04fd-4c83-b260-7d864ae121eb service nova] [instance: f9a40d36-dda9-454c-8774-ed011f3f50ae] Updating instance_info_cache with network_info: [{"id": "4a5e689f-4a77-4c32-81bf-4f86a5c702cb", "address": "fa:16:3e:be:de:b1", "network": {"id": "93e38d7e-b4d7-4d4a-8393-64ab2f6769de", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1445539922-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "06e221dc693640929cac7bc5af45948f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "399f3826-705c-45f7-9fe0-3a08a945151a", "external-id": "nsx-vlan-transportzone-936", "segmentation_id": 936, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4a5e689f-4a", "ovs_interfaceid": "4a5e689f-4a77-4c32-81bf-4f86a5c702cb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 990.177895] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-8940e544-9cc2-4a18-9dd0-e7f396520ac6 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 990.178221] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ae14cb77-184a-428a-a066-b8693d5780cf {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.185554] env[65726]: DEBUG oslo_vmware.api [None req-8940e544-9cc2-4a18-9dd0-e7f396520ac6 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Waiting for the task: (returnval){ [ 990.185554] env[65726]: value = "task-5116406" [ 990.185554] env[65726]: _type = "Task" [ 990.185554] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 990.197596] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-8940e544-9cc2-4a18-9dd0-e7f396520ac6 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04] VM already powered off {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 990.197889] env[65726]: DEBUG nova.compute.manager [None req-8940e544-9cc2-4a18-9dd0-e7f396520ac6 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 990.198852] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-520c3755-df22-4963-90aa-b68b82e45f00 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.213184] env[65726]: DEBUG oslo_concurrency.lockutils [None req-8940e544-9cc2-4a18-9dd0-e7f396520ac6 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Acquiring lock "refresh_cache-2fe6fc5a-f5c7-4f8e-96df-4e621a252f04" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 990.213463] env[65726]: DEBUG oslo_concurrency.lockutils [None req-8940e544-9cc2-4a18-9dd0-e7f396520ac6 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Acquired lock "refresh_cache-2fe6fc5a-f5c7-4f8e-96df-4e621a252f04" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 990.213537] env[65726]: DEBUG nova.network.neutron [None req-8940e544-9cc2-4a18-9dd0-e7f396520ac6 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 990.218511] env[65726]: DEBUG oslo_vmware.api [None req-4e63de1e-de7d-4372-bcfb-4a4336e247fa tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': task-5116405, 'name': CreateSnapshot_Task, 'duration_secs': 0.986508} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 990.218758] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-4e63de1e-de7d-4372-bcfb-4a4336e247fa tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 54e85161-5f63-405f-83f8-490c70645a3b] Created Snapshot of the VM instance {{(pid=65726) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 990.219746] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8c3d654-ef48-4d51-adb8-af7c16c5f914 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.250476] env[65726]: DEBUG nova.network.neutron [req-30ca328e-963c-4bcd-85c3-27f6f8fa45f1 req-994e98a5-9ff8-4757-b18a-09f30709700e service nova] [instance: aa6ce489-c62f-4481-87b7-e74242aeb8ca] Updated VIF entry in instance network info cache for port d5a805f5-b6a8-42ec-92a0-4a652b51c1e6. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 990.251033] env[65726]: DEBUG nova.network.neutron [req-30ca328e-963c-4bcd-85c3-27f6f8fa45f1 req-994e98a5-9ff8-4757-b18a-09f30709700e service nova] [instance: aa6ce489-c62f-4481-87b7-e74242aeb8ca] Updating instance_info_cache with network_info: [{"id": "d5a805f5-b6a8-42ec-92a0-4a652b51c1e6", "address": "fa:16:3e:f5:d5:d7", "network": {"id": "547d7b16-4591-4d80-9e9b-c8a9a3a78dab", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-2077874536-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.242", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4ffd45f4a7a041199a4fc7f69f5e7e9b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2ed91b7b-b4ec-486d-ab34-af0afb7ec691", "external-id": "nsx-vlan-transportzone-75", "segmentation_id": 75, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd5a805f5-b6", "ovs_interfaceid": "d5a805f5-b6a8-42ec-92a0-4a652b51c1e6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 990.281579] env[65726]: DEBUG oslo_concurrency.lockutils [req-4618e1ae-ab3f-4f85-99b5-06f2d52ca99b req-6b1ef7a1-7299-445a-8fac-55ecea913381 service nova] Releasing lock "refresh_cache-406c59e7-8eb8-4cfd-8e1e-e151ab908571" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 990.281579] env[65726]: DEBUG nova.compute.manager [req-4618e1ae-ab3f-4f85-99b5-06f2d52ca99b req-6b1ef7a1-7299-445a-8fac-55ecea913381 service nova] [instance: 406c59e7-8eb8-4cfd-8e1e-e151ab908571] Received event network-vif-plugged-903e092f-430c-4ecb-9cc6-1f2adbcd949f {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 990.281960] env[65726]: DEBUG oslo_concurrency.lockutils [req-4618e1ae-ab3f-4f85-99b5-06f2d52ca99b req-6b1ef7a1-7299-445a-8fac-55ecea913381 service nova] Acquiring lock "406c59e7-8eb8-4cfd-8e1e-e151ab908571-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 990.281960] env[65726]: DEBUG oslo_concurrency.lockutils [req-4618e1ae-ab3f-4f85-99b5-06f2d52ca99b req-6b1ef7a1-7299-445a-8fac-55ecea913381 service nova] Lock "406c59e7-8eb8-4cfd-8e1e-e151ab908571-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 990.282151] env[65726]: DEBUG oslo_concurrency.lockutils [req-4618e1ae-ab3f-4f85-99b5-06f2d52ca99b req-6b1ef7a1-7299-445a-8fac-55ecea913381 service nova] Lock "406c59e7-8eb8-4cfd-8e1e-e151ab908571-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 990.282338] env[65726]: DEBUG nova.compute.manager [req-4618e1ae-ab3f-4f85-99b5-06f2d52ca99b req-6b1ef7a1-7299-445a-8fac-55ecea913381 service nova] [instance: 406c59e7-8eb8-4cfd-8e1e-e151ab908571] No waiting events found dispatching network-vif-plugged-903e092f-430c-4ecb-9cc6-1f2adbcd949f {{(pid=65726) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 990.282587] env[65726]: WARNING nova.compute.manager [req-4618e1ae-ab3f-4f85-99b5-06f2d52ca99b req-6b1ef7a1-7299-445a-8fac-55ecea913381 service nova] [instance: 406c59e7-8eb8-4cfd-8e1e-e151ab908571] Received unexpected event network-vif-plugged-903e092f-430c-4ecb-9cc6-1f2adbcd949f for instance with vm_state building and task_state spawning. [ 990.282742] env[65726]: DEBUG nova.compute.manager [req-4618e1ae-ab3f-4f85-99b5-06f2d52ca99b req-6b1ef7a1-7299-445a-8fac-55ecea913381 service nova] [instance: 406c59e7-8eb8-4cfd-8e1e-e151ab908571] Received event network-changed-903e092f-430c-4ecb-9cc6-1f2adbcd949f {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 990.282948] env[65726]: DEBUG nova.compute.manager [req-4618e1ae-ab3f-4f85-99b5-06f2d52ca99b req-6b1ef7a1-7299-445a-8fac-55ecea913381 service nova] [instance: 406c59e7-8eb8-4cfd-8e1e-e151ab908571] Refreshing instance network info cache due to event network-changed-903e092f-430c-4ecb-9cc6-1f2adbcd949f. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 990.283324] env[65726]: DEBUG oslo_concurrency.lockutils [req-4618e1ae-ab3f-4f85-99b5-06f2d52ca99b req-6b1ef7a1-7299-445a-8fac-55ecea913381 service nova] Acquiring lock "refresh_cache-406c59e7-8eb8-4cfd-8e1e-e151ab908571" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 990.283324] env[65726]: DEBUG oslo_concurrency.lockutils [req-4618e1ae-ab3f-4f85-99b5-06f2d52ca99b req-6b1ef7a1-7299-445a-8fac-55ecea913381 service nova] Acquired lock "refresh_cache-406c59e7-8eb8-4cfd-8e1e-e151ab908571" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 990.283474] env[65726]: DEBUG nova.network.neutron [req-4618e1ae-ab3f-4f85-99b5-06f2d52ca99b req-6b1ef7a1-7299-445a-8fac-55ecea913381 service nova] [instance: 406c59e7-8eb8-4cfd-8e1e-e151ab908571] Refreshing network info cache for port 903e092f-430c-4ecb-9cc6-1f2adbcd949f {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 990.363849] env[65726]: DEBUG oslo_concurrency.lockutils [None req-ed96d5ac-cdb1-4748-8154-6accb1ae5e65 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Releasing lock "refresh_cache-067c9332-fa2c-47fe-93ae-643de1fe77cf" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 990.364193] env[65726]: DEBUG nova.compute.manager [None req-ed96d5ac-cdb1-4748-8154-6accb1ae5e65 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: 067c9332-fa2c-47fe-93ae-643de1fe77cf] Instance network_info: |[{"id": "498a4766-79c8-471b-b6fc-e73c9eca5c5c", "address": "fa:16:3e:fc:03:ad", "network": {"id": "4c972d9d-db3b-4779-ac39-39d9a2d141dd", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1418913861-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cbb17a09b35c4c22ade5c4082b10bda9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "50cf0a70-948d-4611-af05-94c1483064ed", "external-id": "nsx-vlan-transportzone-536", "segmentation_id": 536, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap498a4766-79", "ovs_interfaceid": "498a4766-79c8-471b-b6fc-e73c9eca5c5c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 990.364971] env[65726]: DEBUG oslo_concurrency.lockutils [None req-618baf85-0896-4fc2-b1ef-24618b3b2fb7 tempest-FloatingIPsAssociationNegativeTestJSON-737087291 tempest-FloatingIPsAssociationNegativeTestJSON-737087291-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.531s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 990.365462] env[65726]: DEBUG nova.compute.manager [None req-618baf85-0896-4fc2-b1ef-24618b3b2fb7 tempest-FloatingIPsAssociationNegativeTestJSON-737087291 tempest-FloatingIPsAssociationNegativeTestJSON-737087291-project-member] [instance: d8468efa-c2d3-4dce-ab89-fc077011e3d8] Start building networks asynchronously for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 990.368172] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-ed96d5ac-cdb1-4748-8154-6accb1ae5e65 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: 067c9332-fa2c-47fe-93ae-643de1fe77cf] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:fc:03:ad', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '50cf0a70-948d-4611-af05-94c1483064ed', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '498a4766-79c8-471b-b6fc-e73c9eca5c5c', 'vif_model': 'vmxnet3'}] {{(pid=65726) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 990.375965] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-ed96d5ac-cdb1-4748-8154-6accb1ae5e65 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 990.375965] env[65726]: DEBUG oslo_concurrency.lockutils [None req-4d042142-5ccf-4c86-bcb5-fb6e60309d3b tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 16.271s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 990.376249] env[65726]: DEBUG nova.objects.instance [None req-4d042142-5ccf-4c86-bcb5-fb6e60309d3b tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Lazy-loading 'resources' on Instance uuid df46141c-b545-4e03-b3a3-fd9f5feda0d2 {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 990.379435] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 067c9332-fa2c-47fe-93ae-643de1fe77cf] Creating VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 990.379435] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d25cbb59-82d9-4c90-aea2-73c4cd14143b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.400294] env[65726]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 990.400294] env[65726]: value = "task-5116407" [ 990.400294] env[65726]: _type = "Task" [ 990.400294] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 990.414025] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116407, 'name': CreateVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.423380] env[65726]: DEBUG nova.network.neutron [None req-775eaf93-375d-4ff3-9c57-60f5cc32d4ea tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: a56be6fa-6027-46ee-9aa8-332fe4222d68] Successfully updated port: 4e2dfbcd-41c3-4aaa-a2ec-50c855c12307 {{(pid=65726) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 990.459918] env[65726]: DEBUG oslo_concurrency.lockutils [None req-239cc3b9-c000-4612-951a-e583f6a67697 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Acquiring lock "interface-dcde25df-87a5-47fb-94ce-334f68894e04-280bd1f3-a22e-4f59-8574-d1783c74f8f3" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 990.460660] env[65726]: DEBUG oslo_concurrency.lockutils [None req-239cc3b9-c000-4612-951a-e583f6a67697 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Lock "interface-dcde25df-87a5-47fb-94ce-334f68894e04-280bd1f3-a22e-4f59-8574-d1783c74f8f3" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 990.460660] env[65726]: DEBUG nova.objects.instance [None req-239cc3b9-c000-4612-951a-e583f6a67697 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Lazy-loading 'flavor' on Instance uuid dcde25df-87a5-47fb-94ce-334f68894e04 {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 990.509929] env[65726]: DEBUG oslo_vmware.api [None req-83256413-d207-4cc8-b904-cfb2a5ea5628 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]525f8c3c-da6b-7a37-5859-c08505933eae, 'name': SearchDatastore_Task, 'duration_secs': 0.011195} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 990.510844] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a355a69e-a4ce-449a-acb0-096f73e9fa2a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.517530] env[65726]: DEBUG oslo_vmware.api [None req-83256413-d207-4cc8-b904-cfb2a5ea5628 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Waiting for the task: (returnval){ [ 990.517530] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]522d2e1a-4fcb-e147-0e92-bfba71bf5c87" [ 990.517530] env[65726]: _type = "Task" [ 990.517530] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 990.527999] env[65726]: DEBUG oslo_vmware.api [None req-83256413-d207-4cc8-b904-cfb2a5ea5628 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]522d2e1a-4fcb-e147-0e92-bfba71bf5c87, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.674161] env[65726]: DEBUG oslo_concurrency.lockutils [req-2eee699f-12e2-486b-9b3a-bb9917f139d6 req-d5dc1313-04fd-4c83-b260-7d864ae121eb service nova] Releasing lock "refresh_cache-f9a40d36-dda9-454c-8774-ed011f3f50ae" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 990.674406] env[65726]: DEBUG nova.compute.manager [req-2eee699f-12e2-486b-9b3a-bb9917f139d6 req-d5dc1313-04fd-4c83-b260-7d864ae121eb service nova] [instance: 067c9332-fa2c-47fe-93ae-643de1fe77cf] Received event network-vif-plugged-498a4766-79c8-471b-b6fc-e73c9eca5c5c {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 990.674598] env[65726]: DEBUG oslo_concurrency.lockutils [req-2eee699f-12e2-486b-9b3a-bb9917f139d6 req-d5dc1313-04fd-4c83-b260-7d864ae121eb service nova] Acquiring lock "067c9332-fa2c-47fe-93ae-643de1fe77cf-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 990.674794] env[65726]: DEBUG oslo_concurrency.lockutils [req-2eee699f-12e2-486b-9b3a-bb9917f139d6 req-d5dc1313-04fd-4c83-b260-7d864ae121eb service nova] Lock "067c9332-fa2c-47fe-93ae-643de1fe77cf-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 990.674946] env[65726]: DEBUG oslo_concurrency.lockutils [req-2eee699f-12e2-486b-9b3a-bb9917f139d6 req-d5dc1313-04fd-4c83-b260-7d864ae121eb service nova] Lock "067c9332-fa2c-47fe-93ae-643de1fe77cf-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 990.675124] env[65726]: DEBUG nova.compute.manager [req-2eee699f-12e2-486b-9b3a-bb9917f139d6 req-d5dc1313-04fd-4c83-b260-7d864ae121eb service nova] [instance: 067c9332-fa2c-47fe-93ae-643de1fe77cf] No waiting events found dispatching network-vif-plugged-498a4766-79c8-471b-b6fc-e73c9eca5c5c {{(pid=65726) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 990.675300] env[65726]: WARNING nova.compute.manager [req-2eee699f-12e2-486b-9b3a-bb9917f139d6 req-d5dc1313-04fd-4c83-b260-7d864ae121eb service nova] [instance: 067c9332-fa2c-47fe-93ae-643de1fe77cf] Received unexpected event network-vif-plugged-498a4766-79c8-471b-b6fc-e73c9eca5c5c for instance with vm_state building and task_state spawning. [ 990.675465] env[65726]: DEBUG nova.compute.manager [req-2eee699f-12e2-486b-9b3a-bb9917f139d6 req-d5dc1313-04fd-4c83-b260-7d864ae121eb service nova] [instance: 067c9332-fa2c-47fe-93ae-643de1fe77cf] Received event network-changed-498a4766-79c8-471b-b6fc-e73c9eca5c5c {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 990.675613] env[65726]: DEBUG nova.compute.manager [req-2eee699f-12e2-486b-9b3a-bb9917f139d6 req-d5dc1313-04fd-4c83-b260-7d864ae121eb service nova] [instance: 067c9332-fa2c-47fe-93ae-643de1fe77cf] Refreshing instance network info cache due to event network-changed-498a4766-79c8-471b-b6fc-e73c9eca5c5c. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 990.675789] env[65726]: DEBUG oslo_concurrency.lockutils [req-2eee699f-12e2-486b-9b3a-bb9917f139d6 req-d5dc1313-04fd-4c83-b260-7d864ae121eb service nova] Acquiring lock "refresh_cache-067c9332-fa2c-47fe-93ae-643de1fe77cf" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 990.675916] env[65726]: DEBUG oslo_concurrency.lockutils [req-2eee699f-12e2-486b-9b3a-bb9917f139d6 req-d5dc1313-04fd-4c83-b260-7d864ae121eb service nova] Acquired lock "refresh_cache-067c9332-fa2c-47fe-93ae-643de1fe77cf" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 990.676076] env[65726]: DEBUG nova.network.neutron [req-2eee699f-12e2-486b-9b3a-bb9917f139d6 req-d5dc1313-04fd-4c83-b260-7d864ae121eb service nova] [instance: 067c9332-fa2c-47fe-93ae-643de1fe77cf] Refreshing network info cache for port 498a4766-79c8-471b-b6fc-e73c9eca5c5c {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 990.716781] env[65726]: WARNING openstack [None req-8940e544-9cc2-4a18-9dd0-e7f396520ac6 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 990.717197] env[65726]: WARNING openstack [None req-8940e544-9cc2-4a18-9dd0-e7f396520ac6 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 990.741874] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-4e63de1e-de7d-4372-bcfb-4a4336e247fa tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 54e85161-5f63-405f-83f8-490c70645a3b] Creating linked-clone VM from snapshot {{(pid=65726) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 990.742273] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-f8f22e1b-6ef8-43bb-9c41-8b47ab0f27ea {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.754375] env[65726]: DEBUG oslo_concurrency.lockutils [req-30ca328e-963c-4bcd-85c3-27f6f8fa45f1 req-994e98a5-9ff8-4757-b18a-09f30709700e service nova] Releasing lock "refresh_cache-aa6ce489-c62f-4481-87b7-e74242aeb8ca" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 990.754643] env[65726]: DEBUG nova.compute.manager [req-30ca328e-963c-4bcd-85c3-27f6f8fa45f1 req-994e98a5-9ff8-4757-b18a-09f30709700e service nova] [instance: 060646cc-f0fb-4e6c-9f67-7b2daa6b6fd6] Received event network-vif-deleted-f923178f-ee20-4582-a2e8-6bfd87020d4b {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 990.755132] env[65726]: DEBUG oslo_vmware.api [None req-4e63de1e-de7d-4372-bcfb-4a4336e247fa tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Waiting for the task: (returnval){ [ 990.755132] env[65726]: value = "task-5116408" [ 990.755132] env[65726]: _type = "Task" [ 990.755132] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 990.768195] env[65726]: DEBUG oslo_vmware.api [None req-4e63de1e-de7d-4372-bcfb-4a4336e247fa tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': task-5116408, 'name': CloneVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.786419] env[65726]: WARNING openstack [req-4618e1ae-ab3f-4f85-99b5-06f2d52ca99b req-6b1ef7a1-7299-445a-8fac-55ecea913381 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 990.786825] env[65726]: WARNING openstack [req-4618e1ae-ab3f-4f85-99b5-06f2d52ca99b req-6b1ef7a1-7299-445a-8fac-55ecea913381 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 990.792282] env[65726]: DEBUG nova.network.neutron [req-4618e1ae-ab3f-4f85-99b5-06f2d52ca99b req-6b1ef7a1-7299-445a-8fac-55ecea913381 service nova] [instance: 406c59e7-8eb8-4cfd-8e1e-e151ab908571] Instance cache missing network info. {{(pid=65726) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 990.879634] env[65726]: WARNING openstack [None req-8940e544-9cc2-4a18-9dd0-e7f396520ac6 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 990.879976] env[65726]: WARNING openstack [None req-8940e544-9cc2-4a18-9dd0-e7f396520ac6 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 990.890736] env[65726]: DEBUG nova.compute.utils [None req-618baf85-0896-4fc2-b1ef-24618b3b2fb7 tempest-FloatingIPsAssociationNegativeTestJSON-737087291 tempest-FloatingIPsAssociationNegativeTestJSON-737087291-project-member] Using /dev/sd instead of None {{(pid=65726) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 990.892668] env[65726]: DEBUG nova.compute.manager [None req-618baf85-0896-4fc2-b1ef-24618b3b2fb7 tempest-FloatingIPsAssociationNegativeTestJSON-737087291 tempest-FloatingIPsAssociationNegativeTestJSON-737087291-project-member] [instance: d8468efa-c2d3-4dce-ab89-fc077011e3d8] Allocating IP information in the background. {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 990.893042] env[65726]: DEBUG nova.network.neutron [None req-618baf85-0896-4fc2-b1ef-24618b3b2fb7 tempest-FloatingIPsAssociationNegativeTestJSON-737087291 tempest-FloatingIPsAssociationNegativeTestJSON-737087291-project-member] [instance: d8468efa-c2d3-4dce-ab89-fc077011e3d8] allocate_for_instance() {{(pid=65726) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 990.893367] env[65726]: WARNING neutronclient.v2_0.client [None req-618baf85-0896-4fc2-b1ef-24618b3b2fb7 tempest-FloatingIPsAssociationNegativeTestJSON-737087291 tempest-FloatingIPsAssociationNegativeTestJSON-737087291-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 990.893510] env[65726]: WARNING neutronclient.v2_0.client [None req-618baf85-0896-4fc2-b1ef-24618b3b2fb7 tempest-FloatingIPsAssociationNegativeTestJSON-737087291 tempest-FloatingIPsAssociationNegativeTestJSON-737087291-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 990.894143] env[65726]: WARNING openstack [None req-618baf85-0896-4fc2-b1ef-24618b3b2fb7 tempest-FloatingIPsAssociationNegativeTestJSON-737087291 tempest-FloatingIPsAssociationNegativeTestJSON-737087291-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 990.894500] env[65726]: WARNING openstack [None req-618baf85-0896-4fc2-b1ef-24618b3b2fb7 tempest-FloatingIPsAssociationNegativeTestJSON-737087291 tempest-FloatingIPsAssociationNegativeTestJSON-737087291-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 990.919622] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116407, 'name': CreateVM_Task} progress is 99%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.926973] env[65726]: DEBUG oslo_concurrency.lockutils [None req-775eaf93-375d-4ff3-9c57-60f5cc32d4ea tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Acquiring lock "refresh_cache-a56be6fa-6027-46ee-9aa8-332fe4222d68" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 990.927187] env[65726]: DEBUG oslo_concurrency.lockutils [None req-775eaf93-375d-4ff3-9c57-60f5cc32d4ea tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Acquired lock "refresh_cache-a56be6fa-6027-46ee-9aa8-332fe4222d68" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 990.927406] env[65726]: DEBUG nova.network.neutron [None req-775eaf93-375d-4ff3-9c57-60f5cc32d4ea tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: a56be6fa-6027-46ee-9aa8-332fe4222d68] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 990.959319] env[65726]: DEBUG nova.policy [None req-618baf85-0896-4fc2-b1ef-24618b3b2fb7 tempest-FloatingIPsAssociationNegativeTestJSON-737087291 tempest-FloatingIPsAssociationNegativeTestJSON-737087291-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'cf5619f1c530428aafd0a4ab1c457bac', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'db6b64230abc42069f5587ef94c82f5a', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65726) authorize /opt/stack/nova/nova/policy.py:192}} [ 990.966245] env[65726]: WARNING openstack [None req-239cc3b9-c000-4612-951a-e583f6a67697 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 990.966656] env[65726]: WARNING openstack [None req-239cc3b9-c000-4612-951a-e583f6a67697 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 990.991492] env[65726]: DEBUG nova.network.neutron [req-4618e1ae-ab3f-4f85-99b5-06f2d52ca99b req-6b1ef7a1-7299-445a-8fac-55ecea913381 service nova] [instance: 406c59e7-8eb8-4cfd-8e1e-e151ab908571] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 990.999901] env[65726]: WARNING openstack [None req-8940e544-9cc2-4a18-9dd0-e7f396520ac6 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 990.999901] env[65726]: WARNING openstack [None req-8940e544-9cc2-4a18-9dd0-e7f396520ac6 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 991.008530] env[65726]: DEBUG oslo_concurrency.lockutils [req-4618e1ae-ab3f-4f85-99b5-06f2d52ca99b req-6b1ef7a1-7299-445a-8fac-55ecea913381 service nova] Releasing lock "refresh_cache-406c59e7-8eb8-4cfd-8e1e-e151ab908571" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 991.008809] env[65726]: DEBUG nova.compute.manager [req-4618e1ae-ab3f-4f85-99b5-06f2d52ca99b req-6b1ef7a1-7299-445a-8fac-55ecea913381 service nova] [instance: 406c59e7-8eb8-4cfd-8e1e-e151ab908571] Received event network-vif-plugged-b0ae20cc-cb0a-450d-96a3-4726b64254db {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 991.008997] env[65726]: DEBUG oslo_concurrency.lockutils [req-4618e1ae-ab3f-4f85-99b5-06f2d52ca99b req-6b1ef7a1-7299-445a-8fac-55ecea913381 service nova] Acquiring lock "406c59e7-8eb8-4cfd-8e1e-e151ab908571-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 991.009242] env[65726]: DEBUG oslo_concurrency.lockutils [req-4618e1ae-ab3f-4f85-99b5-06f2d52ca99b req-6b1ef7a1-7299-445a-8fac-55ecea913381 service nova] Lock "406c59e7-8eb8-4cfd-8e1e-e151ab908571-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 991.009460] env[65726]: DEBUG oslo_concurrency.lockutils [req-4618e1ae-ab3f-4f85-99b5-06f2d52ca99b req-6b1ef7a1-7299-445a-8fac-55ecea913381 service nova] Lock "406c59e7-8eb8-4cfd-8e1e-e151ab908571-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 991.009588] env[65726]: DEBUG nova.compute.manager [req-4618e1ae-ab3f-4f85-99b5-06f2d52ca99b req-6b1ef7a1-7299-445a-8fac-55ecea913381 service nova] [instance: 406c59e7-8eb8-4cfd-8e1e-e151ab908571] No waiting events found dispatching network-vif-plugged-b0ae20cc-cb0a-450d-96a3-4726b64254db {{(pid=65726) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 991.009754] env[65726]: WARNING nova.compute.manager [req-4618e1ae-ab3f-4f85-99b5-06f2d52ca99b req-6b1ef7a1-7299-445a-8fac-55ecea913381 service nova] [instance: 406c59e7-8eb8-4cfd-8e1e-e151ab908571] Received unexpected event network-vif-plugged-b0ae20cc-cb0a-450d-96a3-4726b64254db for instance with vm_state building and task_state spawning. [ 991.009908] env[65726]: DEBUG nova.compute.manager [req-4618e1ae-ab3f-4f85-99b5-06f2d52ca99b req-6b1ef7a1-7299-445a-8fac-55ecea913381 service nova] [instance: 406c59e7-8eb8-4cfd-8e1e-e151ab908571] Received event network-changed-b0ae20cc-cb0a-450d-96a3-4726b64254db {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 991.010073] env[65726]: DEBUG nova.compute.manager [req-4618e1ae-ab3f-4f85-99b5-06f2d52ca99b req-6b1ef7a1-7299-445a-8fac-55ecea913381 service nova] [instance: 406c59e7-8eb8-4cfd-8e1e-e151ab908571] Refreshing instance network info cache due to event network-changed-b0ae20cc-cb0a-450d-96a3-4726b64254db. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 991.010244] env[65726]: DEBUG oslo_concurrency.lockutils [req-4618e1ae-ab3f-4f85-99b5-06f2d52ca99b req-6b1ef7a1-7299-445a-8fac-55ecea913381 service nova] Acquiring lock "refresh_cache-406c59e7-8eb8-4cfd-8e1e-e151ab908571" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 991.010378] env[65726]: DEBUG oslo_concurrency.lockutils [req-4618e1ae-ab3f-4f85-99b5-06f2d52ca99b req-6b1ef7a1-7299-445a-8fac-55ecea913381 service nova] Acquired lock "refresh_cache-406c59e7-8eb8-4cfd-8e1e-e151ab908571" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 991.010535] env[65726]: DEBUG nova.network.neutron [req-4618e1ae-ab3f-4f85-99b5-06f2d52ca99b req-6b1ef7a1-7299-445a-8fac-55ecea913381 service nova] [instance: 406c59e7-8eb8-4cfd-8e1e-e151ab908571] Refreshing network info cache for port b0ae20cc-cb0a-450d-96a3-4726b64254db {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 991.033702] env[65726]: DEBUG oslo_vmware.api [None req-83256413-d207-4cc8-b904-cfb2a5ea5628 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]522d2e1a-4fcb-e147-0e92-bfba71bf5c87, 'name': SearchDatastore_Task, 'duration_secs': 0.014066} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.038021] env[65726]: DEBUG oslo_concurrency.lockutils [None req-83256413-d207-4cc8-b904-cfb2a5ea5628 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 991.038021] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-83256413-d207-4cc8-b904-cfb2a5ea5628 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] f9a40d36-dda9-454c-8774-ed011f3f50ae/f9a40d36-dda9-454c-8774-ed011f3f50ae.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 991.038021] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0dba38f8-8b4a-425a-b5e4-76ca5888aaed {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.045288] env[65726]: DEBUG oslo_vmware.api [None req-83256413-d207-4cc8-b904-cfb2a5ea5628 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Waiting for the task: (returnval){ [ 991.045288] env[65726]: value = "task-5116409" [ 991.045288] env[65726]: _type = "Task" [ 991.045288] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.060900] env[65726]: DEBUG oslo_vmware.api [None req-83256413-d207-4cc8-b904-cfb2a5ea5628 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Task: {'id': task-5116409, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.180265] env[65726]: WARNING openstack [req-2eee699f-12e2-486b-9b3a-bb9917f139d6 req-d5dc1313-04fd-4c83-b260-7d864ae121eb service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 991.180741] env[65726]: WARNING openstack [req-2eee699f-12e2-486b-9b3a-bb9917f139d6 req-d5dc1313-04fd-4c83-b260-7d864ae121eb service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 991.194955] env[65726]: DEBUG nova.network.neutron [None req-8940e544-9cc2-4a18-9dd0-e7f396520ac6 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04] Updating instance_info_cache with network_info: [{"id": "25b9d108-0b43-4459-b9db-7bd90a495bb3", "address": "fa:16:3e:5a:1e:6b", "network": {"id": "eb3278b7-c190-441b-84d2-305f9b896382", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1754571022-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.149", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "96149159e18e44f9bf3453e67681f224", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f77ff7a1-209c-4f3f-b2a0-fd817741e739", "external-id": "nsx-vlan-transportzone-935", "segmentation_id": 935, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap25b9d108-0b", "ovs_interfaceid": "25b9d108-0b43-4459-b9db-7bd90a495bb3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 991.226447] env[65726]: DEBUG nova.objects.instance [None req-239cc3b9-c000-4612-951a-e583f6a67697 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Lazy-loading 'pci_requests' on Instance uuid dcde25df-87a5-47fb-94ce-334f68894e04 {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 991.254029] env[65726]: DEBUG nova.network.neutron [None req-618baf85-0896-4fc2-b1ef-24618b3b2fb7 tempest-FloatingIPsAssociationNegativeTestJSON-737087291 tempest-FloatingIPsAssociationNegativeTestJSON-737087291-project-member] [instance: d8468efa-c2d3-4dce-ab89-fc077011e3d8] Successfully created port: 44639a8b-881d-41d3-a114-96a6bccf4064 {{(pid=65726) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 991.270189] env[65726]: DEBUG oslo_vmware.api [None req-4e63de1e-de7d-4372-bcfb-4a4336e247fa tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': task-5116408, 'name': CloneVM_Task} progress is 94%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.315624] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0e4df7a-5c28-46b6-b1f9-22dd25542f0e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.324784] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12728194-ce50-47b1-8803-74aa2e0b97d2 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.367141] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d64b9bd0-58f9-40b9-9e82-afd72b7a904a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.376995] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f2c34a0-c8ed-4d22-a4fe-0ce9cd8c28e2 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.398219] env[65726]: DEBUG nova.compute.provider_tree [None req-4d042142-5ccf-4c86-bcb5-fb6e60309d3b tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 991.403677] env[65726]: DEBUG nova.compute.manager [None req-618baf85-0896-4fc2-b1ef-24618b3b2fb7 tempest-FloatingIPsAssociationNegativeTestJSON-737087291 tempest-FloatingIPsAssociationNegativeTestJSON-737087291-project-member] [instance: d8468efa-c2d3-4dce-ab89-fc077011e3d8] Start building block device mappings for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 991.407761] env[65726]: WARNING openstack [req-2eee699f-12e2-486b-9b3a-bb9917f139d6 req-d5dc1313-04fd-4c83-b260-7d864ae121eb service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 991.408130] env[65726]: WARNING openstack [req-2eee699f-12e2-486b-9b3a-bb9917f139d6 req-d5dc1313-04fd-4c83-b260-7d864ae121eb service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 991.426899] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116407, 'name': CreateVM_Task, 'duration_secs': 0.628525} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.427163] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 067c9332-fa2c-47fe-93ae-643de1fe77cf] Created VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 991.427837] env[65726]: WARNING openstack [None req-ed96d5ac-cdb1-4748-8154-6accb1ae5e65 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 991.428207] env[65726]: WARNING openstack [None req-ed96d5ac-cdb1-4748-8154-6accb1ae5e65 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 991.434371] env[65726]: DEBUG oslo_concurrency.lockutils [None req-ed96d5ac-cdb1-4748-8154-6accb1ae5e65 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 991.434597] env[65726]: DEBUG oslo_concurrency.lockutils [None req-ed96d5ac-cdb1-4748-8154-6accb1ae5e65 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 991.435213] env[65726]: DEBUG oslo_concurrency.lockutils [None req-ed96d5ac-cdb1-4748-8154-6accb1ae5e65 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 991.435798] env[65726]: WARNING openstack [None req-775eaf93-375d-4ff3-9c57-60f5cc32d4ea tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 991.436235] env[65726]: WARNING openstack [None req-775eaf93-375d-4ff3-9c57-60f5cc32d4ea tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 991.442640] env[65726]: DEBUG nova.network.neutron [None req-775eaf93-375d-4ff3-9c57-60f5cc32d4ea tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: a56be6fa-6027-46ee-9aa8-332fe4222d68] Instance cache missing network info. {{(pid=65726) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 991.444901] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-162ca2ef-0caf-4355-ade0-fb421b8a538d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.451363] env[65726]: DEBUG oslo_vmware.api [None req-ed96d5ac-cdb1-4748-8154-6accb1ae5e65 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Waiting for the task: (returnval){ [ 991.451363] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52c2beaa-12f3-3407-b02e-25c6354dd545" [ 991.451363] env[65726]: _type = "Task" [ 991.451363] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.461806] env[65726]: DEBUG oslo_vmware.api [None req-ed96d5ac-cdb1-4748-8154-6accb1ae5e65 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52c2beaa-12f3-3407-b02e-25c6354dd545, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.515360] env[65726]: WARNING openstack [req-4618e1ae-ab3f-4f85-99b5-06f2d52ca99b req-6b1ef7a1-7299-445a-8fac-55ecea913381 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 991.515706] env[65726]: WARNING openstack [req-4618e1ae-ab3f-4f85-99b5-06f2d52ca99b req-6b1ef7a1-7299-445a-8fac-55ecea913381 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 991.521188] env[65726]: DEBUG nova.network.neutron [req-4618e1ae-ab3f-4f85-99b5-06f2d52ca99b req-6b1ef7a1-7299-445a-8fac-55ecea913381 service nova] [instance: 406c59e7-8eb8-4cfd-8e1e-e151ab908571] Instance cache missing network info. {{(pid=65726) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 991.556043] env[65726]: DEBUG oslo_vmware.api [None req-83256413-d207-4cc8-b904-cfb2a5ea5628 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Task: {'id': task-5116409, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.559644] env[65726]: WARNING openstack [req-2eee699f-12e2-486b-9b3a-bb9917f139d6 req-d5dc1313-04fd-4c83-b260-7d864ae121eb service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 991.560066] env[65726]: WARNING openstack [req-2eee699f-12e2-486b-9b3a-bb9917f139d6 req-d5dc1313-04fd-4c83-b260-7d864ae121eb service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 991.610093] env[65726]: WARNING openstack [None req-775eaf93-375d-4ff3-9c57-60f5cc32d4ea tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 991.610487] env[65726]: WARNING openstack [None req-775eaf93-375d-4ff3-9c57-60f5cc32d4ea tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 991.698165] env[65726]: DEBUG oslo_concurrency.lockutils [None req-8940e544-9cc2-4a18-9dd0-e7f396520ac6 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Releasing lock "refresh_cache-2fe6fc5a-f5c7-4f8e-96df-4e621a252f04" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 991.698973] env[65726]: WARNING openstack [None req-8940e544-9cc2-4a18-9dd0-e7f396520ac6 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 991.699387] env[65726]: WARNING openstack [None req-8940e544-9cc2-4a18-9dd0-e7f396520ac6 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 991.705174] env[65726]: WARNING openstack [None req-8940e544-9cc2-4a18-9dd0-e7f396520ac6 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 991.705525] env[65726]: WARNING openstack [None req-8940e544-9cc2-4a18-9dd0-e7f396520ac6 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 991.733799] env[65726]: DEBUG nova.objects.base [None req-239cc3b9-c000-4612-951a-e583f6a67697 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Object Instance lazy-loaded attributes: flavor,pci_requests {{(pid=65726) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 991.733799] env[65726]: DEBUG nova.network.neutron [None req-239cc3b9-c000-4612-951a-e583f6a67697 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: dcde25df-87a5-47fb-94ce-334f68894e04] allocate_for_instance() {{(pid=65726) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 991.735272] env[65726]: WARNING neutronclient.v2_0.client [None req-239cc3b9-c000-4612-951a-e583f6a67697 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 991.735272] env[65726]: WARNING neutronclient.v2_0.client [None req-239cc3b9-c000-4612-951a-e583f6a67697 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 991.737285] env[65726]: WARNING openstack [None req-239cc3b9-c000-4612-951a-e583f6a67697 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 991.737707] env[65726]: WARNING openstack [None req-239cc3b9-c000-4612-951a-e583f6a67697 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 991.766955] env[65726]: DEBUG oslo_vmware.api [None req-4e63de1e-de7d-4372-bcfb-4a4336e247fa tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': task-5116408, 'name': CloneVM_Task} progress is 94%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.805232] env[65726]: DEBUG nova.network.neutron [req-2eee699f-12e2-486b-9b3a-bb9917f139d6 req-d5dc1313-04fd-4c83-b260-7d864ae121eb service nova] [instance: 067c9332-fa2c-47fe-93ae-643de1fe77cf] Updated VIF entry in instance network info cache for port 498a4766-79c8-471b-b6fc-e73c9eca5c5c. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 991.805636] env[65726]: DEBUG nova.network.neutron [req-2eee699f-12e2-486b-9b3a-bb9917f139d6 req-d5dc1313-04fd-4c83-b260-7d864ae121eb service nova] [instance: 067c9332-fa2c-47fe-93ae-643de1fe77cf] Updating instance_info_cache with network_info: [{"id": "498a4766-79c8-471b-b6fc-e73c9eca5c5c", "address": "fa:16:3e:fc:03:ad", "network": {"id": "4c972d9d-db3b-4779-ac39-39d9a2d141dd", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1418913861-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cbb17a09b35c4c22ade5c4082b10bda9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "50cf0a70-948d-4611-af05-94c1483064ed", "external-id": "nsx-vlan-transportzone-536", "segmentation_id": 536, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap498a4766-79", "ovs_interfaceid": "498a4766-79c8-471b-b6fc-e73c9eca5c5c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 991.893556] env[65726]: DEBUG nova.network.neutron [req-4618e1ae-ab3f-4f85-99b5-06f2d52ca99b req-6b1ef7a1-7299-445a-8fac-55ecea913381 service nova] [instance: 406c59e7-8eb8-4cfd-8e1e-e151ab908571] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 991.901342] env[65726]: DEBUG nova.scheduler.client.report [None req-4d042142-5ccf-4c86-bcb5-fb6e60309d3b tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 991.927388] env[65726]: WARNING openstack [None req-775eaf93-375d-4ff3-9c57-60f5cc32d4ea tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 991.927388] env[65726]: WARNING openstack [None req-775eaf93-375d-4ff3-9c57-60f5cc32d4ea tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 991.962479] env[65726]: DEBUG oslo_vmware.api [None req-ed96d5ac-cdb1-4748-8154-6accb1ae5e65 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52c2beaa-12f3-3407-b02e-25c6354dd545, 'name': SearchDatastore_Task, 'duration_secs': 0.073165} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.963040] env[65726]: DEBUG oslo_concurrency.lockutils [None req-ed96d5ac-cdb1-4748-8154-6accb1ae5e65 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 991.963040] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-ed96d5ac-cdb1-4748-8154-6accb1ae5e65 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: 067c9332-fa2c-47fe-93ae-643de1fe77cf] Processing image b52362a3-ee8a-4cbf-b06f-513b0cc8f95c {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 991.963314] env[65726]: DEBUG oslo_concurrency.lockutils [None req-ed96d5ac-cdb1-4748-8154-6accb1ae5e65 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 991.963407] env[65726]: DEBUG oslo_concurrency.lockutils [None req-ed96d5ac-cdb1-4748-8154-6accb1ae5e65 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 991.963585] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-ed96d5ac-cdb1-4748-8154-6accb1ae5e65 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 991.963888] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3e29f64e-7283-4824-be6d-f993474333e2 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.976855] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-ed96d5ac-cdb1-4748-8154-6accb1ae5e65 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 991.977102] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-ed96d5ac-cdb1-4748-8154-6accb1ae5e65 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65726) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 991.977832] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7f780074-1dca-4b5d-bb9d-451e969186d5 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.993295] env[65726]: DEBUG oslo_vmware.api [None req-ed96d5ac-cdb1-4748-8154-6accb1ae5e65 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Waiting for the task: (returnval){ [ 991.993295] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5240033c-c510-f2fb-e639-d9c098267943" [ 991.993295] env[65726]: _type = "Task" [ 991.993295] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.003023] env[65726]: DEBUG oslo_vmware.api [None req-ed96d5ac-cdb1-4748-8154-6accb1ae5e65 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5240033c-c510-f2fb-e639-d9c098267943, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.022942] env[65726]: DEBUG nova.policy [None req-239cc3b9-c000-4612-951a-e583f6a67697 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9b65b9194933469c908c7cc04478444b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd46ccb00794f458b85da4a93879139ee', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65726) authorize /opt/stack/nova/nova/policy.py:192}} [ 992.057852] env[65726]: DEBUG oslo_vmware.api [None req-83256413-d207-4cc8-b904-cfb2a5ea5628 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Task: {'id': task-5116409, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.734826} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.062875] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-83256413-d207-4cc8-b904-cfb2a5ea5628 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] f9a40d36-dda9-454c-8774-ed011f3f50ae/f9a40d36-dda9-454c-8774-ed011f3f50ae.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 992.063156] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-83256413-d207-4cc8-b904-cfb2a5ea5628 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] [instance: f9a40d36-dda9-454c-8774-ed011f3f50ae] Extending root virtual disk to 1048576 {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 992.064359] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a45925cf-0d02-48e8-892f-af36f5d89749 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.073087] env[65726]: DEBUG oslo_vmware.api [None req-83256413-d207-4cc8-b904-cfb2a5ea5628 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Waiting for the task: (returnval){ [ 992.073087] env[65726]: value = "task-5116410" [ 992.073087] env[65726]: _type = "Task" [ 992.073087] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.084666] env[65726]: DEBUG oslo_vmware.api [None req-83256413-d207-4cc8-b904-cfb2a5ea5628 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Task: {'id': task-5116410, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.268358] env[65726]: DEBUG oslo_vmware.api [None req-4e63de1e-de7d-4372-bcfb-4a4336e247fa tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': task-5116408, 'name': CloneVM_Task} progress is 94%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.308535] env[65726]: DEBUG oslo_concurrency.lockutils [req-2eee699f-12e2-486b-9b3a-bb9917f139d6 req-d5dc1313-04fd-4c83-b260-7d864ae121eb service nova] Releasing lock "refresh_cache-067c9332-fa2c-47fe-93ae-643de1fe77cf" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 992.361755] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-8940e544-9cc2-4a18-9dd0-e7f396520ac6 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04] Destroying instance {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 992.363020] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95525d23-07b0-4d73-a1b3-5a3465ecdafd {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.371989] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-8940e544-9cc2-4a18-9dd0-e7f396520ac6 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04] Unregistering the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 992.372295] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-30bff102-4d93-4e7d-8fef-cf1ce97834e0 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.399083] env[65726]: DEBUG oslo_concurrency.lockutils [req-4618e1ae-ab3f-4f85-99b5-06f2d52ca99b req-6b1ef7a1-7299-445a-8fac-55ecea913381 service nova] Releasing lock "refresh_cache-406c59e7-8eb8-4cfd-8e1e-e151ab908571" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 992.399590] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c75c9ed0-c9ca-4fd9-9695-87538dcf1f78 tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Acquired lock "refresh_cache-406c59e7-8eb8-4cfd-8e1e-e151ab908571" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 992.399796] env[65726]: DEBUG nova.network.neutron [None req-c75c9ed0-c9ca-4fd9-9695-87538dcf1f78 tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] [instance: 406c59e7-8eb8-4cfd-8e1e-e151ab908571] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 992.405660] env[65726]: DEBUG nova.compute.manager [req-1c82e1e5-83c3-4c10-86eb-4cf74a8eb602 req-0e74702e-7b54-470f-ab58-5d002d5b6eaf service nova] [instance: a56be6fa-6027-46ee-9aa8-332fe4222d68] Received event network-vif-plugged-4e2dfbcd-41c3-4aaa-a2ec-50c855c12307 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 992.405882] env[65726]: DEBUG oslo_concurrency.lockutils [req-1c82e1e5-83c3-4c10-86eb-4cf74a8eb602 req-0e74702e-7b54-470f-ab58-5d002d5b6eaf service nova] Acquiring lock "a56be6fa-6027-46ee-9aa8-332fe4222d68-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 992.406100] env[65726]: DEBUG oslo_concurrency.lockutils [req-1c82e1e5-83c3-4c10-86eb-4cf74a8eb602 req-0e74702e-7b54-470f-ab58-5d002d5b6eaf service nova] Lock "a56be6fa-6027-46ee-9aa8-332fe4222d68-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 992.406321] env[65726]: DEBUG oslo_concurrency.lockutils [req-1c82e1e5-83c3-4c10-86eb-4cf74a8eb602 req-0e74702e-7b54-470f-ab58-5d002d5b6eaf service nova] Lock "a56be6fa-6027-46ee-9aa8-332fe4222d68-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 992.406429] env[65726]: DEBUG nova.compute.manager [req-1c82e1e5-83c3-4c10-86eb-4cf74a8eb602 req-0e74702e-7b54-470f-ab58-5d002d5b6eaf service nova] [instance: a56be6fa-6027-46ee-9aa8-332fe4222d68] No waiting events found dispatching network-vif-plugged-4e2dfbcd-41c3-4aaa-a2ec-50c855c12307 {{(pid=65726) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 992.406586] env[65726]: WARNING nova.compute.manager [req-1c82e1e5-83c3-4c10-86eb-4cf74a8eb602 req-0e74702e-7b54-470f-ab58-5d002d5b6eaf service nova] [instance: a56be6fa-6027-46ee-9aa8-332fe4222d68] Received unexpected event network-vif-plugged-4e2dfbcd-41c3-4aaa-a2ec-50c855c12307 for instance with vm_state building and task_state spawning. [ 992.409197] env[65726]: DEBUG nova.compute.manager [req-1c82e1e5-83c3-4c10-86eb-4cf74a8eb602 req-0e74702e-7b54-470f-ab58-5d002d5b6eaf service nova] [instance: a56be6fa-6027-46ee-9aa8-332fe4222d68] Received event network-changed-4e2dfbcd-41c3-4aaa-a2ec-50c855c12307 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 992.409197] env[65726]: DEBUG nova.compute.manager [req-1c82e1e5-83c3-4c10-86eb-4cf74a8eb602 req-0e74702e-7b54-470f-ab58-5d002d5b6eaf service nova] [instance: a56be6fa-6027-46ee-9aa8-332fe4222d68] Refreshing instance network info cache due to event network-changed-4e2dfbcd-41c3-4aaa-a2ec-50c855c12307. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 992.409197] env[65726]: DEBUG oslo_concurrency.lockutils [req-1c82e1e5-83c3-4c10-86eb-4cf74a8eb602 req-0e74702e-7b54-470f-ab58-5d002d5b6eaf service nova] Acquiring lock "refresh_cache-a56be6fa-6027-46ee-9aa8-332fe4222d68" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 992.409197] env[65726]: DEBUG oslo_concurrency.lockutils [None req-4d042142-5ccf-4c86-bcb5-fb6e60309d3b tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.033s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 992.411956] env[65726]: DEBUG oslo_concurrency.lockutils [None req-7ab7520f-2b6e-4be5-84c4-f04aa6f3af4f tempest-ServersTestManualDisk-1753452859 tempest-ServersTestManualDisk-1753452859-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 17.346s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 992.412289] env[65726]: DEBUG nova.objects.instance [None req-7ab7520f-2b6e-4be5-84c4-f04aa6f3af4f tempest-ServersTestManualDisk-1753452859 tempest-ServersTestManualDisk-1753452859-project-member] Lazy-loading 'resources' on Instance uuid 3088ee5e-0d27-4058-a94b-2e04c3b52add {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 992.441080] env[65726]: DEBUG nova.compute.manager [None req-618baf85-0896-4fc2-b1ef-24618b3b2fb7 tempest-FloatingIPsAssociationNegativeTestJSON-737087291 tempest-FloatingIPsAssociationNegativeTestJSON-737087291-project-member] [instance: d8468efa-c2d3-4dce-ab89-fc077011e3d8] Start spawning the instance on the hypervisor. {{(pid=65726) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 992.446944] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-8940e544-9cc2-4a18-9dd0-e7f396520ac6 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04] Unregistered the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 992.446944] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-8940e544-9cc2-4a18-9dd0-e7f396520ac6 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04] Deleting contents of the VM from datastore datastore2 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 992.446944] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-8940e544-9cc2-4a18-9dd0-e7f396520ac6 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Deleting the datastore file [datastore2] 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04 {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 992.447749] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f4dae9f8-0340-4797-a113-630e09d5a55f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.455686] env[65726]: INFO nova.scheduler.client.report [None req-4d042142-5ccf-4c86-bcb5-fb6e60309d3b tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Deleted allocations for instance df46141c-b545-4e03-b3a3-fd9f5feda0d2 [ 992.463954] env[65726]: DEBUG oslo_vmware.api [None req-8940e544-9cc2-4a18-9dd0-e7f396520ac6 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Waiting for the task: (returnval){ [ 992.463954] env[65726]: value = "task-5116412" [ 992.463954] env[65726]: _type = "Task" [ 992.463954] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.472790] env[65726]: DEBUG oslo_vmware.api [None req-8940e544-9cc2-4a18-9dd0-e7f396520ac6 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Task: {'id': task-5116412, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.484043] env[65726]: DEBUG nova.virt.hardware [None req-618baf85-0896-4fc2-b1ef-24618b3b2fb7 tempest-FloatingIPsAssociationNegativeTestJSON-737087291 tempest-FloatingIPsAssociationNegativeTestJSON-737087291-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T19:28:34Z,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1539f06fa3534e90acbc3a60c4b8bd3f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T19:28:35Z,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 992.484575] env[65726]: DEBUG nova.virt.hardware [None req-618baf85-0896-4fc2-b1ef-24618b3b2fb7 tempest-FloatingIPsAssociationNegativeTestJSON-737087291 tempest-FloatingIPsAssociationNegativeTestJSON-737087291-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 992.484575] env[65726]: DEBUG nova.virt.hardware [None req-618baf85-0896-4fc2-b1ef-24618b3b2fb7 tempest-FloatingIPsAssociationNegativeTestJSON-737087291 tempest-FloatingIPsAssociationNegativeTestJSON-737087291-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 992.484977] env[65726]: DEBUG nova.virt.hardware [None req-618baf85-0896-4fc2-b1ef-24618b3b2fb7 tempest-FloatingIPsAssociationNegativeTestJSON-737087291 tempest-FloatingIPsAssociationNegativeTestJSON-737087291-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 992.485272] env[65726]: DEBUG nova.virt.hardware [None req-618baf85-0896-4fc2-b1ef-24618b3b2fb7 tempest-FloatingIPsAssociationNegativeTestJSON-737087291 tempest-FloatingIPsAssociationNegativeTestJSON-737087291-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 992.485569] env[65726]: DEBUG nova.virt.hardware [None req-618baf85-0896-4fc2-b1ef-24618b3b2fb7 tempest-FloatingIPsAssociationNegativeTestJSON-737087291 tempest-FloatingIPsAssociationNegativeTestJSON-737087291-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 992.485958] env[65726]: DEBUG nova.virt.hardware [None req-618baf85-0896-4fc2-b1ef-24618b3b2fb7 tempest-FloatingIPsAssociationNegativeTestJSON-737087291 tempest-FloatingIPsAssociationNegativeTestJSON-737087291-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 992.486295] env[65726]: DEBUG nova.virt.hardware [None req-618baf85-0896-4fc2-b1ef-24618b3b2fb7 tempest-FloatingIPsAssociationNegativeTestJSON-737087291 tempest-FloatingIPsAssociationNegativeTestJSON-737087291-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 992.486617] env[65726]: DEBUG nova.virt.hardware [None req-618baf85-0896-4fc2-b1ef-24618b3b2fb7 tempest-FloatingIPsAssociationNegativeTestJSON-737087291 tempest-FloatingIPsAssociationNegativeTestJSON-737087291-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 992.486956] env[65726]: DEBUG nova.virt.hardware [None req-618baf85-0896-4fc2-b1ef-24618b3b2fb7 tempest-FloatingIPsAssociationNegativeTestJSON-737087291 tempest-FloatingIPsAssociationNegativeTestJSON-737087291-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 992.488029] env[65726]: DEBUG nova.virt.hardware [None req-618baf85-0896-4fc2-b1ef-24618b3b2fb7 tempest-FloatingIPsAssociationNegativeTestJSON-737087291 tempest-FloatingIPsAssociationNegativeTestJSON-737087291-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 992.488844] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b236f7bd-05d4-4bda-bdb9-141d08140916 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.512571] env[65726]: DEBUG oslo_vmware.api [None req-ed96d5ac-cdb1-4748-8154-6accb1ae5e65 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5240033c-c510-f2fb-e639-d9c098267943, 'name': SearchDatastore_Task, 'duration_secs': 0.057121} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.515716] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66c936d4-f180-4491-a2eb-15f049435d33 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.522298] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-60ac2e7e-3150-4fe7-b21d-b72867017a33 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.531593] env[65726]: DEBUG oslo_vmware.api [None req-ed96d5ac-cdb1-4748-8154-6accb1ae5e65 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Waiting for the task: (returnval){ [ 992.531593] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]526292dd-82f1-941a-65d5-d4dbd307c883" [ 992.531593] env[65726]: _type = "Task" [ 992.531593] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.559450] env[65726]: DEBUG oslo_vmware.api [None req-ed96d5ac-cdb1-4748-8154-6accb1ae5e65 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]526292dd-82f1-941a-65d5-d4dbd307c883, 'name': SearchDatastore_Task, 'duration_secs': 0.015684} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.559780] env[65726]: DEBUG oslo_concurrency.lockutils [None req-ed96d5ac-cdb1-4748-8154-6accb1ae5e65 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 992.560248] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-ed96d5ac-cdb1-4748-8154-6accb1ae5e65 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] 067c9332-fa2c-47fe-93ae-643de1fe77cf/067c9332-fa2c-47fe-93ae-643de1fe77cf.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 992.560652] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d184afdf-8e7e-4d2c-8983-0c92d0863e2e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.571059] env[65726]: DEBUG oslo_vmware.api [None req-ed96d5ac-cdb1-4748-8154-6accb1ae5e65 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Waiting for the task: (returnval){ [ 992.571059] env[65726]: value = "task-5116413" [ 992.571059] env[65726]: _type = "Task" [ 992.571059] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.590503] env[65726]: DEBUG oslo_vmware.api [None req-ed96d5ac-cdb1-4748-8154-6accb1ae5e65 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': task-5116413, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.594199] env[65726]: DEBUG oslo_vmware.api [None req-83256413-d207-4cc8-b904-cfb2a5ea5628 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Task: {'id': task-5116410, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.08483} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.594519] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-83256413-d207-4cc8-b904-cfb2a5ea5628 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] [instance: f9a40d36-dda9-454c-8774-ed011f3f50ae] Extended root virtual disk {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 992.595372] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf8a24bf-15c1-4383-8fc5-d24db6dbfa37 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.620781] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-83256413-d207-4cc8-b904-cfb2a5ea5628 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] [instance: f9a40d36-dda9-454c-8774-ed011f3f50ae] Reconfiguring VM instance instance-00000051 to attach disk [datastore1] f9a40d36-dda9-454c-8774-ed011f3f50ae/f9a40d36-dda9-454c-8774-ed011f3f50ae.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 992.621144] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7dadce8b-ffcb-49ed-8112-f46173f6c7c0 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.645170] env[65726]: DEBUG oslo_vmware.api [None req-83256413-d207-4cc8-b904-cfb2a5ea5628 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Waiting for the task: (returnval){ [ 992.645170] env[65726]: value = "task-5116414" [ 992.645170] env[65726]: _type = "Task" [ 992.645170] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.656729] env[65726]: DEBUG oslo_vmware.api [None req-83256413-d207-4cc8-b904-cfb2a5ea5628 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Task: {'id': task-5116414, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.771234] env[65726]: DEBUG oslo_vmware.api [None req-4e63de1e-de7d-4372-bcfb-4a4336e247fa tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': task-5116408, 'name': CloneVM_Task} progress is 95%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.896654] env[65726]: DEBUG nova.network.neutron [None req-618baf85-0896-4fc2-b1ef-24618b3b2fb7 tempest-FloatingIPsAssociationNegativeTestJSON-737087291 tempest-FloatingIPsAssociationNegativeTestJSON-737087291-project-member] [instance: d8468efa-c2d3-4dce-ab89-fc077011e3d8] Successfully updated port: 44639a8b-881d-41d3-a114-96a6bccf4064 {{(pid=65726) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 992.906242] env[65726]: WARNING openstack [None req-c75c9ed0-c9ca-4fd9-9695-87538dcf1f78 tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 992.906866] env[65726]: WARNING openstack [None req-c75c9ed0-c9ca-4fd9-9695-87538dcf1f78 tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 992.913929] env[65726]: DEBUG nova.network.neutron [None req-c75c9ed0-c9ca-4fd9-9695-87538dcf1f78 tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] [instance: 406c59e7-8eb8-4cfd-8e1e-e151ab908571] Instance cache missing network info. {{(pid=65726) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 992.934937] env[65726]: DEBUG nova.network.neutron [None req-775eaf93-375d-4ff3-9c57-60f5cc32d4ea tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: a56be6fa-6027-46ee-9aa8-332fe4222d68] Updating instance_info_cache with network_info: [{"id": "4e2dfbcd-41c3-4aaa-a2ec-50c855c12307", "address": "fa:16:3e:1d:82:59", "network": {"id": "0d0f561d-ac41-4ef5-9926-c707409799f7", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1134321217-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4eece77569624f90bf64e5c51974173f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2ce62383-8e84-4e26-955b-74c11392f4c9", "external-id": "nsx-vlan-transportzone-215", "segmentation_id": 215, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4e2dfbcd-41", "ovs_interfaceid": "4e2dfbcd-41c3-4aaa-a2ec-50c855c12307", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 992.970460] env[65726]: DEBUG oslo_concurrency.lockutils [None req-4d042142-5ccf-4c86-bcb5-fb6e60309d3b tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Lock "df46141c-b545-4e03-b3a3-fd9f5feda0d2" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 22.469s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 992.986476] env[65726]: DEBUG oslo_vmware.api [None req-8940e544-9cc2-4a18-9dd0-e7f396520ac6 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Task: {'id': task-5116412, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.224301} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.986750] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-8940e544-9cc2-4a18-9dd0-e7f396520ac6 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Deleted the datastore file {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 992.986914] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-8940e544-9cc2-4a18-9dd0-e7f396520ac6 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04] Deleted contents of the VM from datastore datastore2 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 992.987270] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-8940e544-9cc2-4a18-9dd0-e7f396520ac6 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04] Instance destroyed {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 993.011572] env[65726]: INFO nova.scheduler.client.report [None req-8940e544-9cc2-4a18-9dd0-e7f396520ac6 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Deleted allocations for instance 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04 [ 993.085863] env[65726]: DEBUG oslo_vmware.api [None req-ed96d5ac-cdb1-4748-8154-6accb1ae5e65 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': task-5116413, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.122385] env[65726]: WARNING openstack [None req-c75c9ed0-c9ca-4fd9-9695-87538dcf1f78 tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 993.122855] env[65726]: WARNING openstack [None req-c75c9ed0-c9ca-4fd9-9695-87538dcf1f78 tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 993.155461] env[65726]: DEBUG oslo_vmware.api [None req-83256413-d207-4cc8-b904-cfb2a5ea5628 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Task: {'id': task-5116414, 'name': ReconfigVM_Task, 'duration_secs': 0.401779} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 993.162325] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-83256413-d207-4cc8-b904-cfb2a5ea5628 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] [instance: f9a40d36-dda9-454c-8774-ed011f3f50ae] Reconfigured VM instance instance-00000051 to attach disk [datastore1] f9a40d36-dda9-454c-8774-ed011f3f50ae/f9a40d36-dda9-454c-8774-ed011f3f50ae.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 993.163245] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1f145f20-d66d-432c-b7ca-2b01be97c3c1 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.170027] env[65726]: DEBUG oslo_vmware.api [None req-83256413-d207-4cc8-b904-cfb2a5ea5628 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Waiting for the task: (returnval){ [ 993.170027] env[65726]: value = "task-5116415" [ 993.170027] env[65726]: _type = "Task" [ 993.170027] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 993.181725] env[65726]: DEBUG oslo_vmware.api [None req-83256413-d207-4cc8-b904-cfb2a5ea5628 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Task: {'id': task-5116415, 'name': Rename_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.217091] env[65726]: WARNING openstack [None req-c75c9ed0-c9ca-4fd9-9695-87538dcf1f78 tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 993.217611] env[65726]: WARNING openstack [None req-c75c9ed0-c9ca-4fd9-9695-87538dcf1f78 tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 993.270809] env[65726]: DEBUG oslo_vmware.api [None req-4e63de1e-de7d-4372-bcfb-4a4336e247fa tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': task-5116408, 'name': CloneVM_Task, 'duration_secs': 2.363613} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 993.273990] env[65726]: INFO nova.virt.vmwareapi.vmops [None req-4e63de1e-de7d-4372-bcfb-4a4336e247fa tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 54e85161-5f63-405f-83f8-490c70645a3b] Created linked-clone VM from snapshot [ 993.276239] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f9dc6e3-90f3-4917-bfe2-992cc2d5a028 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.293998] env[65726]: DEBUG nova.virt.vmwareapi.images [None req-4e63de1e-de7d-4372-bcfb-4a4336e247fa tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 54e85161-5f63-405f-83f8-490c70645a3b] Uploading image 320fc78b-3215-4cc6-8f1c-936597aeb974 {{(pid=65726) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 993.298122] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f64ed826-2431-48b9-90f7-e05776f33ef1 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.307564] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4af591e8-8181-4e90-9e62-2c73af5af152 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.312721] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-4e63de1e-de7d-4372-bcfb-4a4336e247fa tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 54e85161-5f63-405f-83f8-490c70645a3b] Destroying the VM {{(pid=65726) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 993.313080] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-6093b3cb-23c8-4e51-8a94-5a032f175de2 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.347661] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-972b025f-de92-47ed-8714-6b8cc1361eba {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.351000] env[65726]: DEBUG oslo_vmware.api [None req-4e63de1e-de7d-4372-bcfb-4a4336e247fa tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Waiting for the task: (returnval){ [ 993.351000] env[65726]: value = "task-5116416" [ 993.351000] env[65726]: _type = "Task" [ 993.351000] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 993.362665] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ce71eb3-af9a-4600-8b82-ab2555f118ed {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.369870] env[65726]: DEBUG oslo_vmware.api [None req-4e63de1e-de7d-4372-bcfb-4a4336e247fa tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': task-5116416, 'name': Destroy_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.380430] env[65726]: DEBUG nova.compute.provider_tree [None req-7ab7520f-2b6e-4be5-84c4-f04aa6f3af4f tempest-ServersTestManualDisk-1753452859 tempest-ServersTestManualDisk-1753452859-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 993.385955] env[65726]: WARNING openstack [None req-c75c9ed0-c9ca-4fd9-9695-87538dcf1f78 tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 993.386503] env[65726]: WARNING openstack [None req-c75c9ed0-c9ca-4fd9-9695-87538dcf1f78 tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 993.407029] env[65726]: DEBUG oslo_concurrency.lockutils [None req-618baf85-0896-4fc2-b1ef-24618b3b2fb7 tempest-FloatingIPsAssociationNegativeTestJSON-737087291 tempest-FloatingIPsAssociationNegativeTestJSON-737087291-project-member] Acquiring lock "refresh_cache-d8468efa-c2d3-4dce-ab89-fc077011e3d8" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 993.407235] env[65726]: DEBUG oslo_concurrency.lockutils [None req-618baf85-0896-4fc2-b1ef-24618b3b2fb7 tempest-FloatingIPsAssociationNegativeTestJSON-737087291 tempest-FloatingIPsAssociationNegativeTestJSON-737087291-project-member] Acquired lock "refresh_cache-d8468efa-c2d3-4dce-ab89-fc077011e3d8" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 993.408469] env[65726]: DEBUG nova.network.neutron [None req-618baf85-0896-4fc2-b1ef-24618b3b2fb7 tempest-FloatingIPsAssociationNegativeTestJSON-737087291 tempest-FloatingIPsAssociationNegativeTestJSON-737087291-project-member] [instance: d8468efa-c2d3-4dce-ab89-fc077011e3d8] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 993.438475] env[65726]: DEBUG oslo_concurrency.lockutils [None req-775eaf93-375d-4ff3-9c57-60f5cc32d4ea tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Releasing lock "refresh_cache-a56be6fa-6027-46ee-9aa8-332fe4222d68" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 993.439153] env[65726]: DEBUG nova.compute.manager [None req-775eaf93-375d-4ff3-9c57-60f5cc32d4ea tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: a56be6fa-6027-46ee-9aa8-332fe4222d68] Instance network_info: |[{"id": "4e2dfbcd-41c3-4aaa-a2ec-50c855c12307", "address": "fa:16:3e:1d:82:59", "network": {"id": "0d0f561d-ac41-4ef5-9926-c707409799f7", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1134321217-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4eece77569624f90bf64e5c51974173f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2ce62383-8e84-4e26-955b-74c11392f4c9", "external-id": "nsx-vlan-transportzone-215", "segmentation_id": 215, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4e2dfbcd-41", "ovs_interfaceid": "4e2dfbcd-41c3-4aaa-a2ec-50c855c12307", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 993.439664] env[65726]: DEBUG oslo_concurrency.lockutils [req-1c82e1e5-83c3-4c10-86eb-4cf74a8eb602 req-0e74702e-7b54-470f-ab58-5d002d5b6eaf service nova] Acquired lock "refresh_cache-a56be6fa-6027-46ee-9aa8-332fe4222d68" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 993.439974] env[65726]: DEBUG nova.network.neutron [req-1c82e1e5-83c3-4c10-86eb-4cf74a8eb602 req-0e74702e-7b54-470f-ab58-5d002d5b6eaf service nova] [instance: a56be6fa-6027-46ee-9aa8-332fe4222d68] Refreshing network info cache for port 4e2dfbcd-41c3-4aaa-a2ec-50c855c12307 {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 993.441955] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-775eaf93-375d-4ff3-9c57-60f5cc32d4ea tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: a56be6fa-6027-46ee-9aa8-332fe4222d68] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:1d:82:59', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '2ce62383-8e84-4e26-955b-74c11392f4c9', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4e2dfbcd-41c3-4aaa-a2ec-50c855c12307', 'vif_model': 'vmxnet3'}] {{(pid=65726) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 993.450362] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-775eaf93-375d-4ff3-9c57-60f5cc32d4ea tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 993.452137] env[65726]: WARNING openstack [req-1c82e1e5-83c3-4c10-86eb-4cf74a8eb602 req-0e74702e-7b54-470f-ab58-5d002d5b6eaf service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 993.452505] env[65726]: WARNING openstack [req-1c82e1e5-83c3-4c10-86eb-4cf74a8eb602 req-0e74702e-7b54-470f-ab58-5d002d5b6eaf service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 993.460907] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a56be6fa-6027-46ee-9aa8-332fe4222d68] Creating VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 993.461975] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5aab4e49-a3bf-4fac-8b6e-9c8df56c0f96 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.488056] env[65726]: WARNING openstack [None req-c75c9ed0-c9ca-4fd9-9695-87538dcf1f78 tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 993.488479] env[65726]: WARNING openstack [None req-c75c9ed0-c9ca-4fd9-9695-87538dcf1f78 tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 993.503740] env[65726]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 993.503740] env[65726]: value = "task-5116417" [ 993.503740] env[65726]: _type = "Task" [ 993.503740] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 993.513610] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116417, 'name': CreateVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.516047] env[65726]: DEBUG oslo_concurrency.lockutils [None req-8940e544-9cc2-4a18-9dd0-e7f396520ac6 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 993.587648] env[65726]: DEBUG oslo_vmware.api [None req-ed96d5ac-cdb1-4748-8154-6accb1ae5e65 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': task-5116413, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.53131} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 993.587968] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-ed96d5ac-cdb1-4748-8154-6accb1ae5e65 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] 067c9332-fa2c-47fe-93ae-643de1fe77cf/067c9332-fa2c-47fe-93ae-643de1fe77cf.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 993.588189] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-ed96d5ac-cdb1-4748-8154-6accb1ae5e65 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: 067c9332-fa2c-47fe-93ae-643de1fe77cf] Extending root virtual disk to 1048576 {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 993.588523] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4efbb38b-0c5d-49e1-9db3-928a5498a964 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.596983] env[65726]: DEBUG oslo_vmware.api [None req-ed96d5ac-cdb1-4748-8154-6accb1ae5e65 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Waiting for the task: (returnval){ [ 993.596983] env[65726]: value = "task-5116418" [ 993.596983] env[65726]: _type = "Task" [ 993.596983] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 993.612305] env[65726]: DEBUG oslo_vmware.api [None req-ed96d5ac-cdb1-4748-8154-6accb1ae5e65 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': task-5116418, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.680282] env[65726]: DEBUG oslo_vmware.api [None req-83256413-d207-4cc8-b904-cfb2a5ea5628 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Task: {'id': task-5116415, 'name': Rename_Task, 'duration_secs': 0.161617} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 993.684147] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-83256413-d207-4cc8-b904-cfb2a5ea5628 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] [instance: f9a40d36-dda9-454c-8774-ed011f3f50ae] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 993.684934] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f47d029b-a1a1-44d9-8f15-d7452783b8f0 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.691345] env[65726]: DEBUG oslo_vmware.api [None req-83256413-d207-4cc8-b904-cfb2a5ea5628 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Waiting for the task: (returnval){ [ 993.691345] env[65726]: value = "task-5116419" [ 993.691345] env[65726]: _type = "Task" [ 993.691345] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 993.696209] env[65726]: WARNING openstack [req-1c82e1e5-83c3-4c10-86eb-4cf74a8eb602 req-0e74702e-7b54-470f-ab58-5d002d5b6eaf service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 993.696646] env[65726]: WARNING openstack [req-1c82e1e5-83c3-4c10-86eb-4cf74a8eb602 req-0e74702e-7b54-470f-ab58-5d002d5b6eaf service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 993.709305] env[65726]: DEBUG oslo_vmware.api [None req-83256413-d207-4cc8-b904-cfb2a5ea5628 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Task: {'id': task-5116419, 'name': PowerOnVM_Task} progress is 33%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.723301] env[65726]: WARNING openstack [None req-c75c9ed0-c9ca-4fd9-9695-87538dcf1f78 tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 993.723673] env[65726]: WARNING openstack [None req-c75c9ed0-c9ca-4fd9-9695-87538dcf1f78 tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 993.788667] env[65726]: WARNING openstack [req-1c82e1e5-83c3-4c10-86eb-4cf74a8eb602 req-0e74702e-7b54-470f-ab58-5d002d5b6eaf service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 993.789120] env[65726]: WARNING openstack [req-1c82e1e5-83c3-4c10-86eb-4cf74a8eb602 req-0e74702e-7b54-470f-ab58-5d002d5b6eaf service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 993.847681] env[65726]: WARNING openstack [None req-c75c9ed0-c9ca-4fd9-9695-87538dcf1f78 tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 993.848142] env[65726]: WARNING openstack [None req-c75c9ed0-c9ca-4fd9-9695-87538dcf1f78 tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 993.864846] env[65726]: DEBUG oslo_vmware.api [None req-4e63de1e-de7d-4372-bcfb-4a4336e247fa tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': task-5116416, 'name': Destroy_Task} progress is 100%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.883806] env[65726]: DEBUG nova.scheduler.client.report [None req-7ab7520f-2b6e-4be5-84c4-f04aa6f3af4f tempest-ServersTestManualDisk-1753452859 tempest-ServersTestManualDisk-1753452859-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 993.912332] env[65726]: WARNING openstack [None req-618baf85-0896-4fc2-b1ef-24618b3b2fb7 tempest-FloatingIPsAssociationNegativeTestJSON-737087291 tempest-FloatingIPsAssociationNegativeTestJSON-737087291-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 993.912806] env[65726]: WARNING openstack [None req-618baf85-0896-4fc2-b1ef-24618b3b2fb7 tempest-FloatingIPsAssociationNegativeTestJSON-737087291 tempest-FloatingIPsAssociationNegativeTestJSON-737087291-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 993.918351] env[65726]: DEBUG nova.network.neutron [None req-618baf85-0896-4fc2-b1ef-24618b3b2fb7 tempest-FloatingIPsAssociationNegativeTestJSON-737087291 tempest-FloatingIPsAssociationNegativeTestJSON-737087291-project-member] [instance: d8468efa-c2d3-4dce-ab89-fc077011e3d8] Instance cache missing network info. {{(pid=65726) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 993.949413] env[65726]: DEBUG nova.network.neutron [None req-239cc3b9-c000-4612-951a-e583f6a67697 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: dcde25df-87a5-47fb-94ce-334f68894e04] Successfully updated port: 280bd1f3-a22e-4f59-8574-d1783c74f8f3 {{(pid=65726) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 993.968736] env[65726]: DEBUG nova.network.neutron [req-1c82e1e5-83c3-4c10-86eb-4cf74a8eb602 req-0e74702e-7b54-470f-ab58-5d002d5b6eaf service nova] [instance: a56be6fa-6027-46ee-9aa8-332fe4222d68] Updated VIF entry in instance network info cache for port 4e2dfbcd-41c3-4aaa-a2ec-50c855c12307. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 993.969100] env[65726]: DEBUG nova.network.neutron [req-1c82e1e5-83c3-4c10-86eb-4cf74a8eb602 req-0e74702e-7b54-470f-ab58-5d002d5b6eaf service nova] [instance: a56be6fa-6027-46ee-9aa8-332fe4222d68] Updating instance_info_cache with network_info: [{"id": "4e2dfbcd-41c3-4aaa-a2ec-50c855c12307", "address": "fa:16:3e:1d:82:59", "network": {"id": "0d0f561d-ac41-4ef5-9926-c707409799f7", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1134321217-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4eece77569624f90bf64e5c51974173f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2ce62383-8e84-4e26-955b-74c11392f4c9", "external-id": "nsx-vlan-transportzone-215", "segmentation_id": 215, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4e2dfbcd-41", "ovs_interfaceid": "4e2dfbcd-41c3-4aaa-a2ec-50c855c12307", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 994.012845] env[65726]: DEBUG nova.network.neutron [None req-c75c9ed0-c9ca-4fd9-9695-87538dcf1f78 tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] [instance: 406c59e7-8eb8-4cfd-8e1e-e151ab908571] Updating instance_info_cache with network_info: [{"id": "eb554e53-6a49-4f06-8271-81ad5c918184", "address": "fa:16:3e:63:4d:c3", "network": {"id": "9487fcf4-a539-4618-8f17-26d849a62be3", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1172384567", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.91", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6961c16aa6a346679bbde25930b2a5b7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7adad264-9276-43ef-9b03-07dc27d3f81e", "external-id": "nsx-vlan-transportzone-329", "segmentation_id": 329, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapeb554e53-6a", "ovs_interfaceid": "eb554e53-6a49-4f06-8271-81ad5c918184", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "903e092f-430c-4ecb-9cc6-1f2adbcd949f", "address": "fa:16:3e:16:c1:49", "network": {"id": "ef1db6dc-35f7-4b9a-beea-bb899fd757dd", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-444962096", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.228", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "6961c16aa6a346679bbde25930b2a5b7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "72781990-3cb3-42eb-9eb1-4040dedbf66f", "external-id": "cl2-zone-812", "segmentation_id": 812, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap903e092f-43", "ovs_interfaceid": "903e092f-430c-4ecb-9cc6-1f2adbcd949f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "b0ae20cc-cb0a-450d-96a3-4726b64254db", "address": "fa:16:3e:bd:0d:c9", "network": {"id": "9487fcf4-a539-4618-8f17-26d849a62be3", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1172384567", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.241", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6961c16aa6a346679bbde25930b2a5b7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7adad264-9276-43ef-9b03-07dc27d3f81e", "external-id": "nsx-vlan-transportzone-329", "segmentation_id": 329, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb0ae20cc-cb", "ovs_interfaceid": "b0ae20cc-cb0a-450d-96a3-4726b64254db", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 994.017529] env[65726]: WARNING openstack [None req-618baf85-0896-4fc2-b1ef-24618b3b2fb7 tempest-FloatingIPsAssociationNegativeTestJSON-737087291 tempest-FloatingIPsAssociationNegativeTestJSON-737087291-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 994.017529] env[65726]: WARNING openstack [None req-618baf85-0896-4fc2-b1ef-24618b3b2fb7 tempest-FloatingIPsAssociationNegativeTestJSON-737087291 tempest-FloatingIPsAssociationNegativeTestJSON-737087291-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 994.027075] env[65726]: DEBUG nova.compute.manager [req-5b8b47a3-2062-460b-af28-4179f433700f req-6e12eee7-2066-41c5-a26a-56d397a728c8 service nova] [instance: 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04] Received event network-vif-unplugged-25b9d108-0b43-4459-b9db-7bd90a495bb3 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 994.027410] env[65726]: DEBUG oslo_concurrency.lockutils [req-5b8b47a3-2062-460b-af28-4179f433700f req-6e12eee7-2066-41c5-a26a-56d397a728c8 service nova] Acquiring lock "2fe6fc5a-f5c7-4f8e-96df-4e621a252f04-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 994.027804] env[65726]: DEBUG oslo_concurrency.lockutils [req-5b8b47a3-2062-460b-af28-4179f433700f req-6e12eee7-2066-41c5-a26a-56d397a728c8 service nova] Lock "2fe6fc5a-f5c7-4f8e-96df-4e621a252f04-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 994.028171] env[65726]: DEBUG oslo_concurrency.lockutils [req-5b8b47a3-2062-460b-af28-4179f433700f req-6e12eee7-2066-41c5-a26a-56d397a728c8 service nova] Lock "2fe6fc5a-f5c7-4f8e-96df-4e621a252f04-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 994.028433] env[65726]: DEBUG nova.compute.manager [req-5b8b47a3-2062-460b-af28-4179f433700f req-6e12eee7-2066-41c5-a26a-56d397a728c8 service nova] [instance: 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04] No waiting events found dispatching network-vif-unplugged-25b9d108-0b43-4459-b9db-7bd90a495bb3 {{(pid=65726) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 994.028667] env[65726]: WARNING nova.compute.manager [req-5b8b47a3-2062-460b-af28-4179f433700f req-6e12eee7-2066-41c5-a26a-56d397a728c8 service nova] [instance: 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04] Received unexpected event network-vif-unplugged-25b9d108-0b43-4459-b9db-7bd90a495bb3 for instance with vm_state shelved_offloaded and task_state None. [ 994.035316] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116417, 'name': CreateVM_Task} progress is 25%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.096614] env[65726]: WARNING openstack [None req-618baf85-0896-4fc2-b1ef-24618b3b2fb7 tempest-FloatingIPsAssociationNegativeTestJSON-737087291 tempest-FloatingIPsAssociationNegativeTestJSON-737087291-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 994.097148] env[65726]: WARNING openstack [None req-618baf85-0896-4fc2-b1ef-24618b3b2fb7 tempest-FloatingIPsAssociationNegativeTestJSON-737087291 tempest-FloatingIPsAssociationNegativeTestJSON-737087291-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 994.118831] env[65726]: DEBUG oslo_vmware.api [None req-ed96d5ac-cdb1-4748-8154-6accb1ae5e65 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': task-5116418, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.076036} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.119174] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-ed96d5ac-cdb1-4748-8154-6accb1ae5e65 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: 067c9332-fa2c-47fe-93ae-643de1fe77cf] Extended root virtual disk {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 994.120059] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-173ec6df-3155-4c77-b18c-cdfe1443de00 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.143852] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-ed96d5ac-cdb1-4748-8154-6accb1ae5e65 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: 067c9332-fa2c-47fe-93ae-643de1fe77cf] Reconfiguring VM instance instance-00000052 to attach disk [datastore1] 067c9332-fa2c-47fe-93ae-643de1fe77cf/067c9332-fa2c-47fe-93ae-643de1fe77cf.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 994.144194] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3719108a-548d-497f-b439-6368ffb8b600 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.168986] env[65726]: DEBUG oslo_vmware.api [None req-ed96d5ac-cdb1-4748-8154-6accb1ae5e65 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Waiting for the task: (returnval){ [ 994.168986] env[65726]: value = "task-5116420" [ 994.168986] env[65726]: _type = "Task" [ 994.168986] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.178340] env[65726]: DEBUG oslo_vmware.api [None req-ed96d5ac-cdb1-4748-8154-6accb1ae5e65 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': task-5116420, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.200101] env[65726]: DEBUG nova.network.neutron [None req-618baf85-0896-4fc2-b1ef-24618b3b2fb7 tempest-FloatingIPsAssociationNegativeTestJSON-737087291 tempest-FloatingIPsAssociationNegativeTestJSON-737087291-project-member] [instance: d8468efa-c2d3-4dce-ab89-fc077011e3d8] Updating instance_info_cache with network_info: [{"id": "44639a8b-881d-41d3-a114-96a6bccf4064", "address": "fa:16:3e:85:bb:61", "network": {"id": "0bcec3df-912b-4dad-9c7b-4b1b24cce6eb", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationNegativeTestJSON-754860692-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "db6b64230abc42069f5587ef94c82f5a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "01fe2e08-46f6-4cee-aefd-934461f8077d", "external-id": "nsx-vlan-transportzone-806", "segmentation_id": 806, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap44639a8b-88", "ovs_interfaceid": "44639a8b-881d-41d3-a114-96a6bccf4064", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 994.205458] env[65726]: DEBUG oslo_vmware.api [None req-83256413-d207-4cc8-b904-cfb2a5ea5628 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Task: {'id': task-5116419, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.370566] env[65726]: DEBUG oslo_vmware.api [None req-4e63de1e-de7d-4372-bcfb-4a4336e247fa tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': task-5116416, 'name': Destroy_Task} progress is 100%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.388672] env[65726]: DEBUG oslo_concurrency.lockutils [None req-7ab7520f-2b6e-4be5-84c4-f04aa6f3af4f tempest-ServersTestManualDisk-1753452859 tempest-ServersTestManualDisk-1753452859-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.977s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 994.391146] env[65726]: DEBUG oslo_concurrency.lockutils [None req-4d7e3d4f-40bd-4bd5-8d6c-4c6d705b54c9 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 17.757s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 994.391394] env[65726]: DEBUG nova.objects.instance [None req-4d7e3d4f-40bd-4bd5-8d6c-4c6d705b54c9 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Lazy-loading 'resources' on Instance uuid 29823bc7-3909-4ab7-8119-91ee59b289c5 {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 994.414624] env[65726]: INFO nova.scheduler.client.report [None req-7ab7520f-2b6e-4be5-84c4-f04aa6f3af4f tempest-ServersTestManualDisk-1753452859 tempest-ServersTestManualDisk-1753452859-project-member] Deleted allocations for instance 3088ee5e-0d27-4058-a94b-2e04c3b52add [ 994.453296] env[65726]: DEBUG oslo_concurrency.lockutils [None req-239cc3b9-c000-4612-951a-e583f6a67697 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Acquiring lock "refresh_cache-dcde25df-87a5-47fb-94ce-334f68894e04" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 994.453296] env[65726]: DEBUG oslo_concurrency.lockutils [None req-239cc3b9-c000-4612-951a-e583f6a67697 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Acquired lock "refresh_cache-dcde25df-87a5-47fb-94ce-334f68894e04" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 994.453523] env[65726]: DEBUG nova.network.neutron [None req-239cc3b9-c000-4612-951a-e583f6a67697 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: dcde25df-87a5-47fb-94ce-334f68894e04] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 994.475576] env[65726]: DEBUG oslo_concurrency.lockutils [req-1c82e1e5-83c3-4c10-86eb-4cf74a8eb602 req-0e74702e-7b54-470f-ab58-5d002d5b6eaf service nova] Releasing lock "refresh_cache-a56be6fa-6027-46ee-9aa8-332fe4222d68" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 994.516774] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116417, 'name': CreateVM_Task} progress is 25%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.525058] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c75c9ed0-c9ca-4fd9-9695-87538dcf1f78 tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Releasing lock "refresh_cache-406c59e7-8eb8-4cfd-8e1e-e151ab908571" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 994.525403] env[65726]: DEBUG nova.compute.manager [None req-c75c9ed0-c9ca-4fd9-9695-87538dcf1f78 tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] [instance: 406c59e7-8eb8-4cfd-8e1e-e151ab908571] Instance network_info: |[{"id": "eb554e53-6a49-4f06-8271-81ad5c918184", "address": "fa:16:3e:63:4d:c3", "network": {"id": "9487fcf4-a539-4618-8f17-26d849a62be3", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1172384567", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.91", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6961c16aa6a346679bbde25930b2a5b7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7adad264-9276-43ef-9b03-07dc27d3f81e", "external-id": "nsx-vlan-transportzone-329", "segmentation_id": 329, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapeb554e53-6a", "ovs_interfaceid": "eb554e53-6a49-4f06-8271-81ad5c918184", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "903e092f-430c-4ecb-9cc6-1f2adbcd949f", "address": "fa:16:3e:16:c1:49", "network": {"id": "ef1db6dc-35f7-4b9a-beea-bb899fd757dd", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-444962096", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.228", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "6961c16aa6a346679bbde25930b2a5b7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "72781990-3cb3-42eb-9eb1-4040dedbf66f", "external-id": "cl2-zone-812", "segmentation_id": 812, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap903e092f-43", "ovs_interfaceid": "903e092f-430c-4ecb-9cc6-1f2adbcd949f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "b0ae20cc-cb0a-450d-96a3-4726b64254db", "address": "fa:16:3e:bd:0d:c9", "network": {"id": "9487fcf4-a539-4618-8f17-26d849a62be3", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1172384567", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.241", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6961c16aa6a346679bbde25930b2a5b7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7adad264-9276-43ef-9b03-07dc27d3f81e", "external-id": "nsx-vlan-transportzone-329", "segmentation_id": 329, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb0ae20cc-cb", "ovs_interfaceid": "b0ae20cc-cb0a-450d-96a3-4726b64254db", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 994.526172] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-c75c9ed0-c9ca-4fd9-9695-87538dcf1f78 tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] [instance: 406c59e7-8eb8-4cfd-8e1e-e151ab908571] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:63:4d:c3', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '7adad264-9276-43ef-9b03-07dc27d3f81e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'eb554e53-6a49-4f06-8271-81ad5c918184', 'vif_model': 'vmxnet3'}, {'network_name': 'br-int', 'mac_address': 'fa:16:3e:16:c1:49', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '72781990-3cb3-42eb-9eb1-4040dedbf66f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '903e092f-430c-4ecb-9cc6-1f2adbcd949f', 'vif_model': 'vmxnet3'}, {'network_name': 'br-int', 'mac_address': 'fa:16:3e:bd:0d:c9', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '7adad264-9276-43ef-9b03-07dc27d3f81e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b0ae20cc-cb0a-450d-96a3-4726b64254db', 'vif_model': 'vmxnet3'}] {{(pid=65726) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 994.539723] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-c75c9ed0-c9ca-4fd9-9695-87538dcf1f78 tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Creating folder: Project (6961c16aa6a346679bbde25930b2a5b7). Parent ref: group-v995008. {{(pid=65726) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 994.540746] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-8ddd454a-b2de-44cf-a0e0-a8868af50f39 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.558877] env[65726]: INFO nova.virt.vmwareapi.vm_util [None req-c75c9ed0-c9ca-4fd9-9695-87538dcf1f78 tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Created folder: Project (6961c16aa6a346679bbde25930b2a5b7) in parent group-v995008. [ 994.559134] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-c75c9ed0-c9ca-4fd9-9695-87538dcf1f78 tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Creating folder: Instances. Parent ref: group-v995246. {{(pid=65726) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 994.559450] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7a1da5be-ea15-43f1-b340-e41d172e6b6a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.573988] env[65726]: INFO nova.virt.vmwareapi.vm_util [None req-c75c9ed0-c9ca-4fd9-9695-87538dcf1f78 tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Created folder: Instances in parent group-v995246. [ 994.574307] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-c75c9ed0-c9ca-4fd9-9695-87538dcf1f78 tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 994.574552] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 406c59e7-8eb8-4cfd-8e1e-e151ab908571] Creating VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 994.574830] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-71ea6944-e7e2-45a5-9c2c-1ad665c9b285 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.604138] env[65726]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 994.604138] env[65726]: value = "task-5116423" [ 994.604138] env[65726]: _type = "Task" [ 994.604138] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.615646] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116423, 'name': CreateVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.683186] env[65726]: DEBUG oslo_vmware.api [None req-ed96d5ac-cdb1-4748-8154-6accb1ae5e65 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': task-5116420, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.704869] env[65726]: DEBUG oslo_vmware.api [None req-83256413-d207-4cc8-b904-cfb2a5ea5628 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Task: {'id': task-5116419, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.709036] env[65726]: DEBUG oslo_concurrency.lockutils [None req-618baf85-0896-4fc2-b1ef-24618b3b2fb7 tempest-FloatingIPsAssociationNegativeTestJSON-737087291 tempest-FloatingIPsAssociationNegativeTestJSON-737087291-project-member] Releasing lock "refresh_cache-d8468efa-c2d3-4dce-ab89-fc077011e3d8" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 994.709036] env[65726]: DEBUG nova.compute.manager [None req-618baf85-0896-4fc2-b1ef-24618b3b2fb7 tempest-FloatingIPsAssociationNegativeTestJSON-737087291 tempest-FloatingIPsAssociationNegativeTestJSON-737087291-project-member] [instance: d8468efa-c2d3-4dce-ab89-fc077011e3d8] Instance network_info: |[{"id": "44639a8b-881d-41d3-a114-96a6bccf4064", "address": "fa:16:3e:85:bb:61", "network": {"id": "0bcec3df-912b-4dad-9c7b-4b1b24cce6eb", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationNegativeTestJSON-754860692-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "db6b64230abc42069f5587ef94c82f5a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "01fe2e08-46f6-4cee-aefd-934461f8077d", "external-id": "nsx-vlan-transportzone-806", "segmentation_id": 806, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap44639a8b-88", "ovs_interfaceid": "44639a8b-881d-41d3-a114-96a6bccf4064", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 994.709036] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-618baf85-0896-4fc2-b1ef-24618b3b2fb7 tempest-FloatingIPsAssociationNegativeTestJSON-737087291 tempest-FloatingIPsAssociationNegativeTestJSON-737087291-project-member] [instance: d8468efa-c2d3-4dce-ab89-fc077011e3d8] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:85:bb:61', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '01fe2e08-46f6-4cee-aefd-934461f8077d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '44639a8b-881d-41d3-a114-96a6bccf4064', 'vif_model': 'vmxnet3'}] {{(pid=65726) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 994.717922] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-618baf85-0896-4fc2-b1ef-24618b3b2fb7 tempest-FloatingIPsAssociationNegativeTestJSON-737087291 tempest-FloatingIPsAssociationNegativeTestJSON-737087291-project-member] Creating folder: Project (db6b64230abc42069f5587ef94c82f5a). Parent ref: group-v995008. {{(pid=65726) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 994.718351] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a9e5fc57-b344-4270-80bb-94361c659e7e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.731059] env[65726]: INFO nova.virt.vmwareapi.vm_util [None req-618baf85-0896-4fc2-b1ef-24618b3b2fb7 tempest-FloatingIPsAssociationNegativeTestJSON-737087291 tempest-FloatingIPsAssociationNegativeTestJSON-737087291-project-member] Created folder: Project (db6b64230abc42069f5587ef94c82f5a) in parent group-v995008. [ 994.731279] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-618baf85-0896-4fc2-b1ef-24618b3b2fb7 tempest-FloatingIPsAssociationNegativeTestJSON-737087291 tempest-FloatingIPsAssociationNegativeTestJSON-737087291-project-member] Creating folder: Instances. Parent ref: group-v995249. {{(pid=65726) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 994.731556] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4bc53db5-0996-454c-a2bf-ca9915666d83 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.745132] env[65726]: INFO nova.virt.vmwareapi.vm_util [None req-618baf85-0896-4fc2-b1ef-24618b3b2fb7 tempest-FloatingIPsAssociationNegativeTestJSON-737087291 tempest-FloatingIPsAssociationNegativeTestJSON-737087291-project-member] Created folder: Instances in parent group-v995249. [ 994.745548] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-618baf85-0896-4fc2-b1ef-24618b3b2fb7 tempest-FloatingIPsAssociationNegativeTestJSON-737087291 tempest-FloatingIPsAssociationNegativeTestJSON-737087291-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 994.745695] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d8468efa-c2d3-4dce-ab89-fc077011e3d8] Creating VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 994.745840] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b005c038-e7e8-45aa-97ac-f66fec0aa199 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.768712] env[65726]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 994.768712] env[65726]: value = "task-5116426" [ 994.768712] env[65726]: _type = "Task" [ 994.768712] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.778234] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116426, 'name': CreateVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.867431] env[65726]: DEBUG oslo_vmware.api [None req-4e63de1e-de7d-4372-bcfb-4a4336e247fa tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': task-5116416, 'name': Destroy_Task, 'duration_secs': 1.375012} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.867712] env[65726]: INFO nova.virt.vmwareapi.vm_util [None req-4e63de1e-de7d-4372-bcfb-4a4336e247fa tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 54e85161-5f63-405f-83f8-490c70645a3b] Destroyed the VM [ 994.867948] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-4e63de1e-de7d-4372-bcfb-4a4336e247fa tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 54e85161-5f63-405f-83f8-490c70645a3b] Deleting Snapshot of the VM instance {{(pid=65726) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 994.868236] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-2db4c629-52bf-4d51-a7ad-b80838810c0e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.876574] env[65726]: DEBUG oslo_vmware.api [None req-4e63de1e-de7d-4372-bcfb-4a4336e247fa tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Waiting for the task: (returnval){ [ 994.876574] env[65726]: value = "task-5116427" [ 994.876574] env[65726]: _type = "Task" [ 994.876574] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.886685] env[65726]: DEBUG oslo_vmware.api [None req-4e63de1e-de7d-4372-bcfb-4a4336e247fa tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': task-5116427, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.926392] env[65726]: DEBUG oslo_concurrency.lockutils [None req-7ab7520f-2b6e-4be5-84c4-f04aa6f3af4f tempest-ServersTestManualDisk-1753452859 tempest-ServersTestManualDisk-1753452859-project-member] Lock "3088ee5e-0d27-4058-a94b-2e04c3b52add" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 23.550s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 994.958878] env[65726]: WARNING openstack [None req-239cc3b9-c000-4612-951a-e583f6a67697 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 994.959536] env[65726]: WARNING openstack [None req-239cc3b9-c000-4612-951a-e583f6a67697 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 994.987956] env[65726]: DEBUG nova.compute.manager [req-56e376ee-d74c-468f-bcfe-21ca6757e7d0 req-d7cb3fe5-a550-4792-8101-4c62ef71e81a service nova] [instance: d8468efa-c2d3-4dce-ab89-fc077011e3d8] Received event network-vif-plugged-44639a8b-881d-41d3-a114-96a6bccf4064 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 994.988292] env[65726]: DEBUG oslo_concurrency.lockutils [req-56e376ee-d74c-468f-bcfe-21ca6757e7d0 req-d7cb3fe5-a550-4792-8101-4c62ef71e81a service nova] Acquiring lock "d8468efa-c2d3-4dce-ab89-fc077011e3d8-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 994.988620] env[65726]: DEBUG oslo_concurrency.lockutils [req-56e376ee-d74c-468f-bcfe-21ca6757e7d0 req-d7cb3fe5-a550-4792-8101-4c62ef71e81a service nova] Lock "d8468efa-c2d3-4dce-ab89-fc077011e3d8-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 994.988898] env[65726]: DEBUG oslo_concurrency.lockutils [req-56e376ee-d74c-468f-bcfe-21ca6757e7d0 req-d7cb3fe5-a550-4792-8101-4c62ef71e81a service nova] Lock "d8468efa-c2d3-4dce-ab89-fc077011e3d8-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 994.989174] env[65726]: DEBUG nova.compute.manager [req-56e376ee-d74c-468f-bcfe-21ca6757e7d0 req-d7cb3fe5-a550-4792-8101-4c62ef71e81a service nova] [instance: d8468efa-c2d3-4dce-ab89-fc077011e3d8] No waiting events found dispatching network-vif-plugged-44639a8b-881d-41d3-a114-96a6bccf4064 {{(pid=65726) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 994.989352] env[65726]: WARNING nova.compute.manager [req-56e376ee-d74c-468f-bcfe-21ca6757e7d0 req-d7cb3fe5-a550-4792-8101-4c62ef71e81a service nova] [instance: d8468efa-c2d3-4dce-ab89-fc077011e3d8] Received unexpected event network-vif-plugged-44639a8b-881d-41d3-a114-96a6bccf4064 for instance with vm_state building and task_state spawning. [ 994.989513] env[65726]: DEBUG nova.compute.manager [req-56e376ee-d74c-468f-bcfe-21ca6757e7d0 req-d7cb3fe5-a550-4792-8101-4c62ef71e81a service nova] [instance: d8468efa-c2d3-4dce-ab89-fc077011e3d8] Received event network-changed-44639a8b-881d-41d3-a114-96a6bccf4064 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 994.989662] env[65726]: DEBUG nova.compute.manager [req-56e376ee-d74c-468f-bcfe-21ca6757e7d0 req-d7cb3fe5-a550-4792-8101-4c62ef71e81a service nova] [instance: d8468efa-c2d3-4dce-ab89-fc077011e3d8] Refreshing instance network info cache due to event network-changed-44639a8b-881d-41d3-a114-96a6bccf4064. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 994.989843] env[65726]: DEBUG oslo_concurrency.lockutils [req-56e376ee-d74c-468f-bcfe-21ca6757e7d0 req-d7cb3fe5-a550-4792-8101-4c62ef71e81a service nova] Acquiring lock "refresh_cache-d8468efa-c2d3-4dce-ab89-fc077011e3d8" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 994.989972] env[65726]: DEBUG oslo_concurrency.lockutils [req-56e376ee-d74c-468f-bcfe-21ca6757e7d0 req-d7cb3fe5-a550-4792-8101-4c62ef71e81a service nova] Acquired lock "refresh_cache-d8468efa-c2d3-4dce-ab89-fc077011e3d8" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 994.990165] env[65726]: DEBUG nova.network.neutron [req-56e376ee-d74c-468f-bcfe-21ca6757e7d0 req-d7cb3fe5-a550-4792-8101-4c62ef71e81a service nova] [instance: d8468efa-c2d3-4dce-ab89-fc077011e3d8] Refreshing network info cache for port 44639a8b-881d-41d3-a114-96a6bccf4064 {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 995.017936] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116417, 'name': CreateVM_Task, 'duration_secs': 1.219866} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.018159] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a56be6fa-6027-46ee-9aa8-332fe4222d68] Created VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 995.021699] env[65726]: WARNING openstack [None req-775eaf93-375d-4ff3-9c57-60f5cc32d4ea tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 995.022200] env[65726]: WARNING openstack [None req-775eaf93-375d-4ff3-9c57-60f5cc32d4ea tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 995.029152] env[65726]: DEBUG oslo_concurrency.lockutils [None req-775eaf93-375d-4ff3-9c57-60f5cc32d4ea tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 995.029395] env[65726]: DEBUG oslo_concurrency.lockutils [None req-775eaf93-375d-4ff3-9c57-60f5cc32d4ea tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 995.029796] env[65726]: DEBUG oslo_concurrency.lockutils [None req-775eaf93-375d-4ff3-9c57-60f5cc32d4ea tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 995.037565] env[65726]: WARNING nova.network.neutron [None req-239cc3b9-c000-4612-951a-e583f6a67697 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: dcde25df-87a5-47fb-94ce-334f68894e04] 51a66448-0f72-4266-8868-856b48be839f already exists in list: networks containing: ['51a66448-0f72-4266-8868-856b48be839f']. ignoring it [ 995.037823] env[65726]: WARNING nova.network.neutron [None req-239cc3b9-c000-4612-951a-e583f6a67697 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: dcde25df-87a5-47fb-94ce-334f68894e04] 51a66448-0f72-4266-8868-856b48be839f already exists in list: networks containing: ['51a66448-0f72-4266-8868-856b48be839f']. ignoring it [ 995.040307] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3ada00e2-b3c0-4da3-b570-fbe43044ab02 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.050903] env[65726]: DEBUG oslo_vmware.api [None req-775eaf93-375d-4ff3-9c57-60f5cc32d4ea tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Waiting for the task: (returnval){ [ 995.050903] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5290d9ed-4a06-aea5-f9f7-22b084ec804e" [ 995.050903] env[65726]: _type = "Task" [ 995.050903] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 995.063245] env[65726]: DEBUG oslo_vmware.api [None req-775eaf93-375d-4ff3-9c57-60f5cc32d4ea tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5290d9ed-4a06-aea5-f9f7-22b084ec804e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.090117] env[65726]: WARNING openstack [None req-239cc3b9-c000-4612-951a-e583f6a67697 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 995.090576] env[65726]: WARNING openstack [None req-239cc3b9-c000-4612-951a-e583f6a67697 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 995.117725] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116423, 'name': CreateVM_Task, 'duration_secs': 0.478839} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.117725] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 406c59e7-8eb8-4cfd-8e1e-e151ab908571] Created VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 995.118549] env[65726]: WARNING openstack [None req-c75c9ed0-c9ca-4fd9-9695-87538dcf1f78 tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 995.121144] env[65726]: WARNING openstack [None req-c75c9ed0-c9ca-4fd9-9695-87538dcf1f78 tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 995.126894] env[65726]: WARNING openstack [None req-c75c9ed0-c9ca-4fd9-9695-87538dcf1f78 tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 995.127120] env[65726]: WARNING openstack [None req-c75c9ed0-c9ca-4fd9-9695-87538dcf1f78 tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 995.133432] env[65726]: WARNING openstack [None req-c75c9ed0-c9ca-4fd9-9695-87538dcf1f78 tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 995.133840] env[65726]: WARNING openstack [None req-c75c9ed0-c9ca-4fd9-9695-87538dcf1f78 tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 995.138582] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c75c9ed0-c9ca-4fd9-9695-87538dcf1f78 tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 995.183976] env[65726]: DEBUG oslo_vmware.api [None req-ed96d5ac-cdb1-4748-8154-6accb1ae5e65 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': task-5116420, 'name': ReconfigVM_Task, 'duration_secs': 0.583462} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.184392] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-ed96d5ac-cdb1-4748-8154-6accb1ae5e65 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: 067c9332-fa2c-47fe-93ae-643de1fe77cf] Reconfigured VM instance instance-00000052 to attach disk [datastore1] 067c9332-fa2c-47fe-93ae-643de1fe77cf/067c9332-fa2c-47fe-93ae-643de1fe77cf.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 995.184975] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a3257e7f-7746-4d54-ba9b-53040e49818a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.195525] env[65726]: DEBUG oslo_vmware.api [None req-ed96d5ac-cdb1-4748-8154-6accb1ae5e65 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Waiting for the task: (returnval){ [ 995.195525] env[65726]: value = "task-5116428" [ 995.195525] env[65726]: _type = "Task" [ 995.195525] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 995.199076] env[65726]: WARNING openstack [None req-239cc3b9-c000-4612-951a-e583f6a67697 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 995.199460] env[65726]: WARNING openstack [None req-239cc3b9-c000-4612-951a-e583f6a67697 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 995.219190] env[65726]: DEBUG oslo_vmware.api [None req-ed96d5ac-cdb1-4748-8154-6accb1ae5e65 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': task-5116428, 'name': Rename_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.226221] env[65726]: DEBUG oslo_vmware.api [None req-83256413-d207-4cc8-b904-cfb2a5ea5628 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Task: {'id': task-5116419, 'name': PowerOnVM_Task, 'duration_secs': 1.153313} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.226963] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-83256413-d207-4cc8-b904-cfb2a5ea5628 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] [instance: f9a40d36-dda9-454c-8774-ed011f3f50ae] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 995.226963] env[65726]: INFO nova.compute.manager [None req-83256413-d207-4cc8-b904-cfb2a5ea5628 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] [instance: f9a40d36-dda9-454c-8774-ed011f3f50ae] Took 9.85 seconds to spawn the instance on the hypervisor. [ 995.227067] env[65726]: DEBUG nova.compute.manager [None req-83256413-d207-4cc8-b904-cfb2a5ea5628 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] [instance: f9a40d36-dda9-454c-8774-ed011f3f50ae] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 995.230963] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d9cd0ed-34e9-4cac-9fd3-859ae5bae2cd {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.281216] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116426, 'name': CreateVM_Task} progress is 99%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.285092] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6bd6f39-f0e7-4608-a5ef-0c8996f1d735 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.295868] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d585020-2ea6-4117-a51c-6a0554ae5a2a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.331636] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69b8ece7-b31b-419c-ac3e-d7b95a7b430e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.345365] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-809e1d99-724e-4c20-9c7d-7db126eba9bd {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.367625] env[65726]: DEBUG nova.compute.provider_tree [None req-4d7e3d4f-40bd-4bd5-8d6c-4c6d705b54c9 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 995.370789] env[65726]: WARNING openstack [None req-239cc3b9-c000-4612-951a-e583f6a67697 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 995.371188] env[65726]: WARNING openstack [None req-239cc3b9-c000-4612-951a-e583f6a67697 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 995.390023] env[65726]: DEBUG oslo_vmware.api [None req-4e63de1e-de7d-4372-bcfb-4a4336e247fa tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': task-5116427, 'name': RemoveSnapshot_Task} progress is 80%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.449027] env[65726]: WARNING openstack [None req-239cc3b9-c000-4612-951a-e583f6a67697 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 995.449027] env[65726]: WARNING openstack [None req-239cc3b9-c000-4612-951a-e583f6a67697 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 995.493782] env[65726]: WARNING openstack [req-56e376ee-d74c-468f-bcfe-21ca6757e7d0 req-d7cb3fe5-a550-4792-8101-4c62ef71e81a service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 995.494289] env[65726]: WARNING openstack [req-56e376ee-d74c-468f-bcfe-21ca6757e7d0 req-d7cb3fe5-a550-4792-8101-4c62ef71e81a service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 995.572242] env[65726]: DEBUG oslo_vmware.api [None req-775eaf93-375d-4ff3-9c57-60f5cc32d4ea tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5290d9ed-4a06-aea5-f9f7-22b084ec804e, 'name': SearchDatastore_Task, 'duration_secs': 0.011293} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.572725] env[65726]: DEBUG oslo_concurrency.lockutils [None req-775eaf93-375d-4ff3-9c57-60f5cc32d4ea tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 995.573019] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-775eaf93-375d-4ff3-9c57-60f5cc32d4ea tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: a56be6fa-6027-46ee-9aa8-332fe4222d68] Processing image b52362a3-ee8a-4cbf-b06f-513b0cc8f95c {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 995.573288] env[65726]: DEBUG oslo_concurrency.lockutils [None req-775eaf93-375d-4ff3-9c57-60f5cc32d4ea tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 995.573460] env[65726]: DEBUG oslo_concurrency.lockutils [None req-775eaf93-375d-4ff3-9c57-60f5cc32d4ea tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 995.574464] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-775eaf93-375d-4ff3-9c57-60f5cc32d4ea tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 995.574816] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c75c9ed0-c9ca-4fd9-9695-87538dcf1f78 tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 995.575155] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c75c9ed0-c9ca-4fd9-9695-87538dcf1f78 tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 995.575404] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-912d4f29-0b95-4a7a-b2d2-84ec91c2c2b9 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.577559] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0ef05ac4-a6e6-4696-a499-97c03f64a9f2 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.584204] env[65726]: DEBUG oslo_vmware.api [None req-c75c9ed0-c9ca-4fd9-9695-87538dcf1f78 tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Waiting for the task: (returnval){ [ 995.584204] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52758885-88ba-c7cc-f0da-b7f4aef1e36b" [ 995.584204] env[65726]: _type = "Task" [ 995.584204] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 995.592283] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-775eaf93-375d-4ff3-9c57-60f5cc32d4ea tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 995.592283] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-775eaf93-375d-4ff3-9c57-60f5cc32d4ea tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65726) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 995.592283] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c09a1090-725b-487d-863c-a46248fe002d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.598603] env[65726]: DEBUG oslo_vmware.api [None req-c75c9ed0-c9ca-4fd9-9695-87538dcf1f78 tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52758885-88ba-c7cc-f0da-b7f4aef1e36b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.604528] env[65726]: DEBUG oslo_vmware.api [None req-775eaf93-375d-4ff3-9c57-60f5cc32d4ea tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Waiting for the task: (returnval){ [ 995.604528] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5237c7bd-fc90-5070-0c3b-e6d910446047" [ 995.604528] env[65726]: _type = "Task" [ 995.604528] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 995.613766] env[65726]: DEBUG oslo_vmware.api [None req-775eaf93-375d-4ff3-9c57-60f5cc32d4ea tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5237c7bd-fc90-5070-0c3b-e6d910446047, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.674558] env[65726]: WARNING openstack [None req-239cc3b9-c000-4612-951a-e583f6a67697 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 995.674963] env[65726]: WARNING openstack [None req-239cc3b9-c000-4612-951a-e583f6a67697 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 995.703509] env[65726]: DEBUG oslo_vmware.api [None req-ed96d5ac-cdb1-4748-8154-6accb1ae5e65 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': task-5116428, 'name': Rename_Task, 'duration_secs': 0.176106} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.704943] env[65726]: WARNING openstack [req-56e376ee-d74c-468f-bcfe-21ca6757e7d0 req-d7cb3fe5-a550-4792-8101-4c62ef71e81a service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 995.705319] env[65726]: WARNING openstack [req-56e376ee-d74c-468f-bcfe-21ca6757e7d0 req-d7cb3fe5-a550-4792-8101-4c62ef71e81a service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 995.715329] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-ed96d5ac-cdb1-4748-8154-6accb1ae5e65 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: 067c9332-fa2c-47fe-93ae-643de1fe77cf] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 995.716210] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d1e34eb4-dcf1-4929-9886-fb9ea1851316 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.723969] env[65726]: DEBUG oslo_vmware.api [None req-ed96d5ac-cdb1-4748-8154-6accb1ae5e65 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Waiting for the task: (returnval){ [ 995.723969] env[65726]: value = "task-5116429" [ 995.723969] env[65726]: _type = "Task" [ 995.723969] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 995.733444] env[65726]: DEBUG oslo_vmware.api [None req-ed96d5ac-cdb1-4748-8154-6accb1ae5e65 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': task-5116429, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.754081] env[65726]: WARNING openstack [None req-239cc3b9-c000-4612-951a-e583f6a67697 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 995.754544] env[65726]: WARNING openstack [None req-239cc3b9-c000-4612-951a-e583f6a67697 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 995.763110] env[65726]: INFO nova.compute.manager [None req-83256413-d207-4cc8-b904-cfb2a5ea5628 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] [instance: f9a40d36-dda9-454c-8774-ed011f3f50ae] Took 32.54 seconds to build instance. [ 995.780253] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116426, 'name': CreateVM_Task, 'duration_secs': 0.614181} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.784219] env[65726]: WARNING openstack [req-56e376ee-d74c-468f-bcfe-21ca6757e7d0 req-d7cb3fe5-a550-4792-8101-4c62ef71e81a service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 995.784706] env[65726]: WARNING openstack [req-56e376ee-d74c-468f-bcfe-21ca6757e7d0 req-d7cb3fe5-a550-4792-8101-4c62ef71e81a service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 995.792477] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d8468efa-c2d3-4dce-ab89-fc077011e3d8] Created VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 995.794083] env[65726]: WARNING openstack [None req-618baf85-0896-4fc2-b1ef-24618b3b2fb7 tempest-FloatingIPsAssociationNegativeTestJSON-737087291 tempest-FloatingIPsAssociationNegativeTestJSON-737087291-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 995.794422] env[65726]: WARNING openstack [None req-618baf85-0896-4fc2-b1ef-24618b3b2fb7 tempest-FloatingIPsAssociationNegativeTestJSON-737087291 tempest-FloatingIPsAssociationNegativeTestJSON-737087291-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 995.799324] env[65726]: DEBUG oslo_concurrency.lockutils [None req-618baf85-0896-4fc2-b1ef-24618b3b2fb7 tempest-FloatingIPsAssociationNegativeTestJSON-737087291 tempest-FloatingIPsAssociationNegativeTestJSON-737087291-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 995.799500] env[65726]: DEBUG oslo_concurrency.lockutils [None req-618baf85-0896-4fc2-b1ef-24618b3b2fb7 tempest-FloatingIPsAssociationNegativeTestJSON-737087291 tempest-FloatingIPsAssociationNegativeTestJSON-737087291-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 995.799807] env[65726]: DEBUG oslo_concurrency.lockutils [None req-618baf85-0896-4fc2-b1ef-24618b3b2fb7 tempest-FloatingIPsAssociationNegativeTestJSON-737087291 tempest-FloatingIPsAssociationNegativeTestJSON-737087291-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 995.800473] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1a84ef82-4443-4da2-8589-2a6336de0872 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.811687] env[65726]: DEBUG oslo_vmware.api [None req-618baf85-0896-4fc2-b1ef-24618b3b2fb7 tempest-FloatingIPsAssociationNegativeTestJSON-737087291 tempest-FloatingIPsAssociationNegativeTestJSON-737087291-project-member] Waiting for the task: (returnval){ [ 995.811687] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]520ddbf5-6c52-50bb-d167-56b93ca5ec48" [ 995.811687] env[65726]: _type = "Task" [ 995.811687] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 995.821746] env[65726]: DEBUG oslo_vmware.api [None req-618baf85-0896-4fc2-b1ef-24618b3b2fb7 tempest-FloatingIPsAssociationNegativeTestJSON-737087291 tempest-FloatingIPsAssociationNegativeTestJSON-737087291-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]520ddbf5-6c52-50bb-d167-56b93ca5ec48, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.845844] env[65726]: DEBUG nova.network.neutron [None req-239cc3b9-c000-4612-951a-e583f6a67697 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: dcde25df-87a5-47fb-94ce-334f68894e04] Updating instance_info_cache with network_info: [{"id": "bf51c96c-63c5-46de-a40f-58e6769c5fcf", "address": "fa:16:3e:4c:88:87", "network": {"id": "51a66448-0f72-4266-8868-856b48be839f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-155658961-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.146", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d46ccb00794f458b85da4a93879139ee", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b1f3e6c3-5584-4852-9017-476ab8ac4946", "external-id": "nsx-vlan-transportzone-304", "segmentation_id": 304, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbf51c96c-63", "ovs_interfaceid": "bf51c96c-63c5-46de-a40f-58e6769c5fcf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "b48313f9-a0d3-4fdd-88c6-d103aef0b4f2", "address": "fa:16:3e:e7:cd:41", "network": {"id": "51a66448-0f72-4266-8868-856b48be839f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-155658961-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d46ccb00794f458b85da4a93879139ee", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b1f3e6c3-5584-4852-9017-476ab8ac4946", "external-id": "nsx-vlan-transportzone-304", "segmentation_id": 304, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb48313f9-a0", "ovs_interfaceid": "b48313f9-a0d3-4fdd-88c6-d103aef0b4f2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "280bd1f3-a22e-4f59-8574-d1783c74f8f3", "address": "fa:16:3e:2b:59:bd", "network": {"id": "51a66448-0f72-4266-8868-856b48be839f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-155658961-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d46ccb00794f458b85da4a93879139ee", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b1f3e6c3-5584-4852-9017-476ab8ac4946", "external-id": "nsx-vlan-transportzone-304", "segmentation_id": 304, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap280bd1f3-a2", "ovs_interfaceid": "280bd1f3-a22e-4f59-8574-d1783c74f8f3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 995.878568] env[65726]: DEBUG nova.network.neutron [req-56e376ee-d74c-468f-bcfe-21ca6757e7d0 req-d7cb3fe5-a550-4792-8101-4c62ef71e81a service nova] [instance: d8468efa-c2d3-4dce-ab89-fc077011e3d8] Updated VIF entry in instance network info cache for port 44639a8b-881d-41d3-a114-96a6bccf4064. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 995.879013] env[65726]: DEBUG nova.network.neutron [req-56e376ee-d74c-468f-bcfe-21ca6757e7d0 req-d7cb3fe5-a550-4792-8101-4c62ef71e81a service nova] [instance: d8468efa-c2d3-4dce-ab89-fc077011e3d8] Updating instance_info_cache with network_info: [{"id": "44639a8b-881d-41d3-a114-96a6bccf4064", "address": "fa:16:3e:85:bb:61", "network": {"id": "0bcec3df-912b-4dad-9c7b-4b1b24cce6eb", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationNegativeTestJSON-754860692-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "db6b64230abc42069f5587ef94c82f5a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "01fe2e08-46f6-4cee-aefd-934461f8077d", "external-id": "nsx-vlan-transportzone-806", "segmentation_id": 806, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap44639a8b-88", "ovs_interfaceid": "44639a8b-881d-41d3-a114-96a6bccf4064", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 995.880831] env[65726]: DEBUG nova.scheduler.client.report [None req-4d7e3d4f-40bd-4bd5-8d6c-4c6d705b54c9 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 995.897077] env[65726]: DEBUG oslo_vmware.api [None req-4e63de1e-de7d-4372-bcfb-4a4336e247fa tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': task-5116427, 'name': RemoveSnapshot_Task, 'duration_secs': 0.705855} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.897471] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-4e63de1e-de7d-4372-bcfb-4a4336e247fa tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 54e85161-5f63-405f-83f8-490c70645a3b] Deleted Snapshot of the VM instance {{(pid=65726) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 996.098824] env[65726]: DEBUG oslo_vmware.api [None req-c75c9ed0-c9ca-4fd9-9695-87538dcf1f78 tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52758885-88ba-c7cc-f0da-b7f4aef1e36b, 'name': SearchDatastore_Task, 'duration_secs': 0.012418} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 996.099436] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c75c9ed0-c9ca-4fd9-9695-87538dcf1f78 tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 996.099827] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-c75c9ed0-c9ca-4fd9-9695-87538dcf1f78 tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] [instance: 406c59e7-8eb8-4cfd-8e1e-e151ab908571] Processing image b52362a3-ee8a-4cbf-b06f-513b0cc8f95c {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 996.100183] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c75c9ed0-c9ca-4fd9-9695-87538dcf1f78 tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 996.112878] env[65726]: DEBUG oslo_vmware.api [None req-775eaf93-375d-4ff3-9c57-60f5cc32d4ea tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5237c7bd-fc90-5070-0c3b-e6d910446047, 'name': SearchDatastore_Task, 'duration_secs': 0.019472} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 996.120588] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9a8aef3f-fb1d-4c0c-9a41-c37f8d741aeb {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.128520] env[65726]: DEBUG oslo_vmware.api [None req-775eaf93-375d-4ff3-9c57-60f5cc32d4ea tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Waiting for the task: (returnval){ [ 996.128520] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52d85f58-1cc2-223e-8cc7-5ffec683e2ab" [ 996.128520] env[65726]: _type = "Task" [ 996.128520] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 996.138747] env[65726]: DEBUG oslo_vmware.api [None req-775eaf93-375d-4ff3-9c57-60f5cc32d4ea tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52d85f58-1cc2-223e-8cc7-5ffec683e2ab, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.235335] env[65726]: DEBUG oslo_vmware.api [None req-ed96d5ac-cdb1-4748-8154-6accb1ae5e65 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': task-5116429, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.266795] env[65726]: DEBUG oslo_concurrency.lockutils [None req-83256413-d207-4cc8-b904-cfb2a5ea5628 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Lock "f9a40d36-dda9-454c-8774-ed011f3f50ae" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 34.056s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 996.324438] env[65726]: DEBUG oslo_vmware.api [None req-618baf85-0896-4fc2-b1ef-24618b3b2fb7 tempest-FloatingIPsAssociationNegativeTestJSON-737087291 tempest-FloatingIPsAssociationNegativeTestJSON-737087291-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]520ddbf5-6c52-50bb-d167-56b93ca5ec48, 'name': SearchDatastore_Task, 'duration_secs': 0.011475} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 996.324801] env[65726]: DEBUG oslo_concurrency.lockutils [None req-618baf85-0896-4fc2-b1ef-24618b3b2fb7 tempest-FloatingIPsAssociationNegativeTestJSON-737087291 tempest-FloatingIPsAssociationNegativeTestJSON-737087291-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 996.325178] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-618baf85-0896-4fc2-b1ef-24618b3b2fb7 tempest-FloatingIPsAssociationNegativeTestJSON-737087291 tempest-FloatingIPsAssociationNegativeTestJSON-737087291-project-member] [instance: d8468efa-c2d3-4dce-ab89-fc077011e3d8] Processing image b52362a3-ee8a-4cbf-b06f-513b0cc8f95c {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 996.325469] env[65726]: DEBUG oslo_concurrency.lockutils [None req-618baf85-0896-4fc2-b1ef-24618b3b2fb7 tempest-FloatingIPsAssociationNegativeTestJSON-737087291 tempest-FloatingIPsAssociationNegativeTestJSON-737087291-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 996.325635] env[65726]: DEBUG oslo_concurrency.lockutils [None req-618baf85-0896-4fc2-b1ef-24618b3b2fb7 tempest-FloatingIPsAssociationNegativeTestJSON-737087291 tempest-FloatingIPsAssociationNegativeTestJSON-737087291-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 996.325841] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-618baf85-0896-4fc2-b1ef-24618b3b2fb7 tempest-FloatingIPsAssociationNegativeTestJSON-737087291 tempest-FloatingIPsAssociationNegativeTestJSON-737087291-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 996.326174] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-38b1eda9-fb61-411e-a3ea-2787cf766df8 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.337143] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-618baf85-0896-4fc2-b1ef-24618b3b2fb7 tempest-FloatingIPsAssociationNegativeTestJSON-737087291 tempest-FloatingIPsAssociationNegativeTestJSON-737087291-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 996.337319] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-618baf85-0896-4fc2-b1ef-24618b3b2fb7 tempest-FloatingIPsAssociationNegativeTestJSON-737087291 tempest-FloatingIPsAssociationNegativeTestJSON-737087291-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65726) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 996.338168] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c45ba56a-b874-4364-96a8-7dd0a3b1fd3a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.345692] env[65726]: DEBUG oslo_vmware.api [None req-618baf85-0896-4fc2-b1ef-24618b3b2fb7 tempest-FloatingIPsAssociationNegativeTestJSON-737087291 tempest-FloatingIPsAssociationNegativeTestJSON-737087291-project-member] Waiting for the task: (returnval){ [ 996.345692] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52203409-87df-59e3-3a4e-6f9a5d655323" [ 996.345692] env[65726]: _type = "Task" [ 996.345692] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 996.350080] env[65726]: DEBUG oslo_concurrency.lockutils [None req-239cc3b9-c000-4612-951a-e583f6a67697 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Releasing lock "refresh_cache-dcde25df-87a5-47fb-94ce-334f68894e04" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 996.350796] env[65726]: DEBUG oslo_concurrency.lockutils [None req-239cc3b9-c000-4612-951a-e583f6a67697 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Acquiring lock "dcde25df-87a5-47fb-94ce-334f68894e04" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 996.350954] env[65726]: DEBUG oslo_concurrency.lockutils [None req-239cc3b9-c000-4612-951a-e583f6a67697 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Acquired lock "dcde25df-87a5-47fb-94ce-334f68894e04" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 996.351878] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9cef9c8-8fa5-45da-a5cd-621940781867 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.360898] env[65726]: DEBUG oslo_vmware.api [None req-618baf85-0896-4fc2-b1ef-24618b3b2fb7 tempest-FloatingIPsAssociationNegativeTestJSON-737087291 tempest-FloatingIPsAssociationNegativeTestJSON-737087291-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52203409-87df-59e3-3a4e-6f9a5d655323, 'name': SearchDatastore_Task, 'duration_secs': 0.013257} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 996.374506] env[65726]: DEBUG nova.virt.hardware [None req-239cc3b9-c000-4612-951a-e583f6a67697 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 996.375109] env[65726]: DEBUG nova.virt.hardware [None req-239cc3b9-c000-4612-951a-e583f6a67697 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 996.375109] env[65726]: DEBUG nova.virt.hardware [None req-239cc3b9-c000-4612-951a-e583f6a67697 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 996.375109] env[65726]: DEBUG nova.virt.hardware [None req-239cc3b9-c000-4612-951a-e583f6a67697 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 996.375409] env[65726]: DEBUG nova.virt.hardware [None req-239cc3b9-c000-4612-951a-e583f6a67697 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 996.375409] env[65726]: DEBUG nova.virt.hardware [None req-239cc3b9-c000-4612-951a-e583f6a67697 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 996.375568] env[65726]: DEBUG nova.virt.hardware [None req-239cc3b9-c000-4612-951a-e583f6a67697 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 996.375698] env[65726]: DEBUG nova.virt.hardware [None req-239cc3b9-c000-4612-951a-e583f6a67697 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 996.375856] env[65726]: DEBUG nova.virt.hardware [None req-239cc3b9-c000-4612-951a-e583f6a67697 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 996.376023] env[65726]: DEBUG nova.virt.hardware [None req-239cc3b9-c000-4612-951a-e583f6a67697 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 996.376204] env[65726]: DEBUG nova.virt.hardware [None req-239cc3b9-c000-4612-951a-e583f6a67697 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 996.382925] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-239cc3b9-c000-4612-951a-e583f6a67697 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: dcde25df-87a5-47fb-94ce-334f68894e04] Reconfiguring VM to attach interface {{(pid=65726) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 996.383330] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-080363d9-02aa-48bb-b1b6-4bf8e271a524 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.385911] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cb7460a9-1db5-4436-8542-dedad16f0236 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.399890] env[65726]: DEBUG oslo_concurrency.lockutils [req-56e376ee-d74c-468f-bcfe-21ca6757e7d0 req-d7cb3fe5-a550-4792-8101-4c62ef71e81a service nova] Releasing lock "refresh_cache-d8468efa-c2d3-4dce-ab89-fc077011e3d8" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 996.401043] env[65726]: DEBUG oslo_concurrency.lockutils [None req-4d7e3d4f-40bd-4bd5-8d6c-4c6d705b54c9 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.010s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 996.404956] env[65726]: WARNING nova.compute.manager [None req-4e63de1e-de7d-4372-bcfb-4a4336e247fa tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 54e85161-5f63-405f-83f8-490c70645a3b] Image not found during snapshot: nova.exception.ImageNotFound: Image 320fc78b-3215-4cc6-8f1c-936597aeb974 could not be found. [ 996.406231] env[65726]: DEBUG oslo_concurrency.lockutils [None req-de3a85d2-cc71-4c75-a1b0-fe6db6e57022 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 19.004s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 996.406997] env[65726]: DEBUG nova.objects.instance [None req-de3a85d2-cc71-4c75-a1b0-fe6db6e57022 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] [instance: 73f5b23e-4463-4068-8994-dd2752a2abc9] Trying to apply a migration context that does not seem to be set for this instance {{(pid=65726) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 996.412644] env[65726]: DEBUG oslo_vmware.api [None req-618baf85-0896-4fc2-b1ef-24618b3b2fb7 tempest-FloatingIPsAssociationNegativeTestJSON-737087291 tempest-FloatingIPsAssociationNegativeTestJSON-737087291-project-member] Waiting for the task: (returnval){ [ 996.412644] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]521fab20-0d04-7407-d111-12555efbe198" [ 996.412644] env[65726]: _type = "Task" [ 996.412644] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 996.417737] env[65726]: DEBUG oslo_vmware.api [None req-239cc3b9-c000-4612-951a-e583f6a67697 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Waiting for the task: (returnval){ [ 996.417737] env[65726]: value = "task-5116430" [ 996.417737] env[65726]: _type = "Task" [ 996.417737] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 996.424991] env[65726]: DEBUG oslo_vmware.api [None req-618baf85-0896-4fc2-b1ef-24618b3b2fb7 tempest-FloatingIPsAssociationNegativeTestJSON-737087291 tempest-FloatingIPsAssociationNegativeTestJSON-737087291-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]521fab20-0d04-7407-d111-12555efbe198, 'name': SearchDatastore_Task, 'duration_secs': 0.011525} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 996.425634] env[65726]: DEBUG oslo_concurrency.lockutils [None req-618baf85-0896-4fc2-b1ef-24618b3b2fb7 tempest-FloatingIPsAssociationNegativeTestJSON-737087291 tempest-FloatingIPsAssociationNegativeTestJSON-737087291-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 996.426047] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-618baf85-0896-4fc2-b1ef-24618b3b2fb7 tempest-FloatingIPsAssociationNegativeTestJSON-737087291 tempest-FloatingIPsAssociationNegativeTestJSON-737087291-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore2] d8468efa-c2d3-4dce-ab89-fc077011e3d8/d8468efa-c2d3-4dce-ab89-fc077011e3d8.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 996.426235] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0fdcf85f-2780-4c23-a5cd-5af9df325ca9 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.432212] env[65726]: DEBUG oslo_vmware.api [None req-239cc3b9-c000-4612-951a-e583f6a67697 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': task-5116430, 'name': ReconfigVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.436634] env[65726]: INFO nova.scheduler.client.report [None req-4d7e3d4f-40bd-4bd5-8d6c-4c6d705b54c9 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Deleted allocations for instance 29823bc7-3909-4ab7-8119-91ee59b289c5 [ 996.439658] env[65726]: DEBUG oslo_vmware.api [None req-618baf85-0896-4fc2-b1ef-24618b3b2fb7 tempest-FloatingIPsAssociationNegativeTestJSON-737087291 tempest-FloatingIPsAssociationNegativeTestJSON-737087291-project-member] Waiting for the task: (returnval){ [ 996.439658] env[65726]: value = "task-5116431" [ 996.439658] env[65726]: _type = "Task" [ 996.439658] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 996.452485] env[65726]: DEBUG oslo_vmware.api [None req-618baf85-0896-4fc2-b1ef-24618b3b2fb7 tempest-FloatingIPsAssociationNegativeTestJSON-737087291 tempest-FloatingIPsAssociationNegativeTestJSON-737087291-project-member] Task: {'id': task-5116431, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.624607] env[65726]: DEBUG oslo_concurrency.lockutils [None req-1d2791c0-c341-4061-b914-188004153713 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Acquiring lock "2fe6fc5a-f5c7-4f8e-96df-4e621a252f04" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 996.640539] env[65726]: DEBUG oslo_vmware.api [None req-775eaf93-375d-4ff3-9c57-60f5cc32d4ea tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52d85f58-1cc2-223e-8cc7-5ffec683e2ab, 'name': SearchDatastore_Task, 'duration_secs': 0.013702} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 996.644025] env[65726]: DEBUG oslo_concurrency.lockutils [None req-775eaf93-375d-4ff3-9c57-60f5cc32d4ea tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 996.644025] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-775eaf93-375d-4ff3-9c57-60f5cc32d4ea tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] a56be6fa-6027-46ee-9aa8-332fe4222d68/a56be6fa-6027-46ee-9aa8-332fe4222d68.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 996.644025] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c75c9ed0-c9ca-4fd9-9695-87538dcf1f78 tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 996.644025] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-c75c9ed0-c9ca-4fd9-9695-87538dcf1f78 tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 996.644025] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1493aaaf-02af-4f96-92e1-11d552ec032f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.644542] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f61228cf-31fa-435e-b516-9587d2856638 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.657182] env[65726]: DEBUG oslo_vmware.api [None req-775eaf93-375d-4ff3-9c57-60f5cc32d4ea tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Waiting for the task: (returnval){ [ 996.657182] env[65726]: value = "task-5116432" [ 996.657182] env[65726]: _type = "Task" [ 996.657182] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 996.663040] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-c75c9ed0-c9ca-4fd9-9695-87538dcf1f78 tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 996.663196] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-c75c9ed0-c9ca-4fd9-9695-87538dcf1f78 tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65726) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 996.664420] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-72c1ad8c-7693-48ee-9548-93ba0cd129be {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.670539] env[65726]: DEBUG oslo_vmware.api [None req-775eaf93-375d-4ff3-9c57-60f5cc32d4ea tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5116432, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.674693] env[65726]: DEBUG oslo_vmware.api [None req-c75c9ed0-c9ca-4fd9-9695-87538dcf1f78 tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Waiting for the task: (returnval){ [ 996.674693] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52eaeee5-ff1f-2978-48d2-501e1b329b8c" [ 996.674693] env[65726]: _type = "Task" [ 996.674693] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 996.683309] env[65726]: DEBUG oslo_vmware.api [None req-c75c9ed0-c9ca-4fd9-9695-87538dcf1f78 tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52eaeee5-ff1f-2978-48d2-501e1b329b8c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.735319] env[65726]: DEBUG oslo_vmware.api [None req-ed96d5ac-cdb1-4748-8154-6accb1ae5e65 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': task-5116429, 'name': PowerOnVM_Task, 'duration_secs': 0.947936} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 996.735600] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-ed96d5ac-cdb1-4748-8154-6accb1ae5e65 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: 067c9332-fa2c-47fe-93ae-643de1fe77cf] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 996.735801] env[65726]: INFO nova.compute.manager [None req-ed96d5ac-cdb1-4748-8154-6accb1ae5e65 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: 067c9332-fa2c-47fe-93ae-643de1fe77cf] Took 9.34 seconds to spawn the instance on the hypervisor. [ 996.736042] env[65726]: DEBUG nova.compute.manager [None req-ed96d5ac-cdb1-4748-8154-6accb1ae5e65 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: 067c9332-fa2c-47fe-93ae-643de1fe77cf] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 996.736890] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be494d20-ab3f-4af9-8b12-4be6361da9a5 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.937425] env[65726]: DEBUG oslo_vmware.api [None req-239cc3b9-c000-4612-951a-e583f6a67697 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': task-5116430, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.953272] env[65726]: DEBUG oslo_concurrency.lockutils [None req-4d7e3d4f-40bd-4bd5-8d6c-4c6d705b54c9 tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Lock "29823bc7-3909-4ab7-8119-91ee59b289c5" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 23.922s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 996.969851] env[65726]: DEBUG oslo_vmware.api [None req-618baf85-0896-4fc2-b1ef-24618b3b2fb7 tempest-FloatingIPsAssociationNegativeTestJSON-737087291 tempest-FloatingIPsAssociationNegativeTestJSON-737087291-project-member] Task: {'id': task-5116431, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.172266] env[65726]: DEBUG oslo_vmware.api [None req-775eaf93-375d-4ff3-9c57-60f5cc32d4ea tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5116432, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.185895] env[65726]: DEBUG oslo_vmware.api [None req-c75c9ed0-c9ca-4fd9-9695-87538dcf1f78 tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52eaeee5-ff1f-2978-48d2-501e1b329b8c, 'name': SearchDatastore_Task, 'duration_secs': 0.011683} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.186206] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-45146af9-21f7-4f5f-8fad-00c1e613529f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.195610] env[65726]: DEBUG oslo_vmware.api [None req-c75c9ed0-c9ca-4fd9-9695-87538dcf1f78 tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Waiting for the task: (returnval){ [ 997.195610] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]523bac86-a220-5bdb-a05a-da34104c4f2a" [ 997.195610] env[65726]: _type = "Task" [ 997.195610] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.205062] env[65726]: DEBUG oslo_vmware.api [None req-c75c9ed0-c9ca-4fd9-9695-87538dcf1f78 tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]523bac86-a220-5bdb-a05a-da34104c4f2a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.254754] env[65726]: INFO nova.compute.manager [None req-ed96d5ac-cdb1-4748-8154-6accb1ae5e65 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: 067c9332-fa2c-47fe-93ae-643de1fe77cf] Took 33.50 seconds to build instance. [ 997.367929] env[65726]: DEBUG nova.compute.manager [req-688629ea-98e2-4cec-895d-ef952731929d req-adf5fa76-a785-4e72-961e-c03ff76e70f1 service nova] [instance: 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04] Received event network-changed-25b9d108-0b43-4459-b9db-7bd90a495bb3 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 997.368165] env[65726]: DEBUG nova.compute.manager [req-688629ea-98e2-4cec-895d-ef952731929d req-adf5fa76-a785-4e72-961e-c03ff76e70f1 service nova] [instance: 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04] Refreshing instance network info cache due to event network-changed-25b9d108-0b43-4459-b9db-7bd90a495bb3. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 997.368459] env[65726]: DEBUG oslo_concurrency.lockutils [req-688629ea-98e2-4cec-895d-ef952731929d req-adf5fa76-a785-4e72-961e-c03ff76e70f1 service nova] Acquiring lock "refresh_cache-2fe6fc5a-f5c7-4f8e-96df-4e621a252f04" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 997.368660] env[65726]: DEBUG oslo_concurrency.lockutils [req-688629ea-98e2-4cec-895d-ef952731929d req-adf5fa76-a785-4e72-961e-c03ff76e70f1 service nova] Acquired lock "refresh_cache-2fe6fc5a-f5c7-4f8e-96df-4e621a252f04" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 997.369346] env[65726]: DEBUG nova.network.neutron [req-688629ea-98e2-4cec-895d-ef952731929d req-adf5fa76-a785-4e72-961e-c03ff76e70f1 service nova] [instance: 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04] Refreshing network info cache for port 25b9d108-0b43-4459-b9db-7bd90a495bb3 {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 997.420688] env[65726]: DEBUG oslo_concurrency.lockutils [None req-de3a85d2-cc71-4c75-a1b0-fe6db6e57022 tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.014s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 997.422481] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c5289ae6-5b56-421a-a0e1-7e74e3494692 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 14.764s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 997.424181] env[65726]: DEBUG nova.objects.instance [None req-c5289ae6-5b56-421a-a0e1-7e74e3494692 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Lazy-loading 'resources' on Instance uuid f2d9090c-988f-43f4-9c81-7aa718a3438a {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 997.440354] env[65726]: DEBUG oslo_vmware.api [None req-239cc3b9-c000-4612-951a-e583f6a67697 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': task-5116430, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.466224] env[65726]: DEBUG oslo_vmware.api [None req-618baf85-0896-4fc2-b1ef-24618b3b2fb7 tempest-FloatingIPsAssociationNegativeTestJSON-737087291 tempest-FloatingIPsAssociationNegativeTestJSON-737087291-project-member] Task: {'id': task-5116431, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.701028} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.466224] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-618baf85-0896-4fc2-b1ef-24618b3b2fb7 tempest-FloatingIPsAssociationNegativeTestJSON-737087291 tempest-FloatingIPsAssociationNegativeTestJSON-737087291-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore2] d8468efa-c2d3-4dce-ab89-fc077011e3d8/d8468efa-c2d3-4dce-ab89-fc077011e3d8.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 997.466563] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-618baf85-0896-4fc2-b1ef-24618b3b2fb7 tempest-FloatingIPsAssociationNegativeTestJSON-737087291 tempest-FloatingIPsAssociationNegativeTestJSON-737087291-project-member] [instance: d8468efa-c2d3-4dce-ab89-fc077011e3d8] Extending root virtual disk to 1048576 {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 997.466666] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ed550b96-41a8-42b8-863f-8be712878872 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.481034] env[65726]: DEBUG oslo_vmware.api [None req-618baf85-0896-4fc2-b1ef-24618b3b2fb7 tempest-FloatingIPsAssociationNegativeTestJSON-737087291 tempest-FloatingIPsAssociationNegativeTestJSON-737087291-project-member] Waiting for the task: (returnval){ [ 997.481034] env[65726]: value = "task-5116433" [ 997.481034] env[65726]: _type = "Task" [ 997.481034] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.498861] env[65726]: DEBUG oslo_vmware.api [None req-618baf85-0896-4fc2-b1ef-24618b3b2fb7 tempest-FloatingIPsAssociationNegativeTestJSON-737087291 tempest-FloatingIPsAssociationNegativeTestJSON-737087291-project-member] Task: {'id': task-5116433, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.670721] env[65726]: DEBUG oslo_vmware.api [None req-775eaf93-375d-4ff3-9c57-60f5cc32d4ea tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5116432, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.880193} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.671105] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-775eaf93-375d-4ff3-9c57-60f5cc32d4ea tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] a56be6fa-6027-46ee-9aa8-332fe4222d68/a56be6fa-6027-46ee-9aa8-332fe4222d68.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 997.671757] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-775eaf93-375d-4ff3-9c57-60f5cc32d4ea tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: a56be6fa-6027-46ee-9aa8-332fe4222d68] Extending root virtual disk to 1048576 {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 997.672091] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c2577251-7083-49c6-ae85-44ecee0d66f4 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.681910] env[65726]: DEBUG oslo_vmware.api [None req-775eaf93-375d-4ff3-9c57-60f5cc32d4ea tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Waiting for the task: (returnval){ [ 997.681910] env[65726]: value = "task-5116434" [ 997.681910] env[65726]: _type = "Task" [ 997.681910] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.690339] env[65726]: DEBUG oslo_vmware.api [None req-775eaf93-375d-4ff3-9c57-60f5cc32d4ea tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5116434, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.706182] env[65726]: DEBUG oslo_vmware.api [None req-c75c9ed0-c9ca-4fd9-9695-87538dcf1f78 tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]523bac86-a220-5bdb-a05a-da34104c4f2a, 'name': SearchDatastore_Task, 'duration_secs': 0.082405} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.706488] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c75c9ed0-c9ca-4fd9-9695-87538dcf1f78 tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 997.706770] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-c75c9ed0-c9ca-4fd9-9695-87538dcf1f78 tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] 406c59e7-8eb8-4cfd-8e1e-e151ab908571/406c59e7-8eb8-4cfd-8e1e-e151ab908571.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 997.707068] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-807e1ee9-12aa-4c77-ac41-2a21f801f67c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.716528] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a27b5bca-2338-4542-8dd7-9cac7e4f3b3a tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Acquiring lock "a286ba4f-a83e-4c30-8079-ba88e3ba5696" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 997.716955] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a27b5bca-2338-4542-8dd7-9cac7e4f3b3a tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Lock "a286ba4f-a83e-4c30-8079-ba88e3ba5696" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 997.720446] env[65726]: DEBUG oslo_vmware.api [None req-c75c9ed0-c9ca-4fd9-9695-87538dcf1f78 tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Waiting for the task: (returnval){ [ 997.720446] env[65726]: value = "task-5116435" [ 997.720446] env[65726]: _type = "Task" [ 997.720446] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.731653] env[65726]: DEBUG oslo_vmware.api [None req-c75c9ed0-c9ca-4fd9-9695-87538dcf1f78 tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Task: {'id': task-5116435, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.757276] env[65726]: DEBUG oslo_concurrency.lockutils [None req-ed96d5ac-cdb1-4748-8154-6accb1ae5e65 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Lock "067c9332-fa2c-47fe-93ae-643de1fe77cf" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 35.022s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 997.873673] env[65726]: WARNING openstack [req-688629ea-98e2-4cec-895d-ef952731929d req-adf5fa76-a785-4e72-961e-c03ff76e70f1 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 997.874152] env[65726]: WARNING openstack [req-688629ea-98e2-4cec-895d-ef952731929d req-adf5fa76-a785-4e72-961e-c03ff76e70f1 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 997.935563] env[65726]: DEBUG oslo_vmware.api [None req-239cc3b9-c000-4612-951a-e583f6a67697 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': task-5116430, 'name': ReconfigVM_Task, 'duration_secs': 1.218286} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.936910] env[65726]: WARNING openstack [None req-239cc3b9-c000-4612-951a-e583f6a67697 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 997.936910] env[65726]: WARNING openstack [None req-239cc3b9-c000-4612-951a-e583f6a67697 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 997.942278] env[65726]: DEBUG oslo_concurrency.lockutils [None req-239cc3b9-c000-4612-951a-e583f6a67697 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Releasing lock "dcde25df-87a5-47fb-94ce-334f68894e04" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 997.942278] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-239cc3b9-c000-4612-951a-e583f6a67697 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: dcde25df-87a5-47fb-94ce-334f68894e04] Reconfigured VM to attach interface {{(pid=65726) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 997.991665] env[65726]: DEBUG oslo_vmware.api [None req-618baf85-0896-4fc2-b1ef-24618b3b2fb7 tempest-FloatingIPsAssociationNegativeTestJSON-737087291 tempest-FloatingIPsAssociationNegativeTestJSON-737087291-project-member] Task: {'id': task-5116433, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.079698} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.994664] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-618baf85-0896-4fc2-b1ef-24618b3b2fb7 tempest-FloatingIPsAssociationNegativeTestJSON-737087291 tempest-FloatingIPsAssociationNegativeTestJSON-737087291-project-member] [instance: d8468efa-c2d3-4dce-ab89-fc077011e3d8] Extended root virtual disk {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 997.995698] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-964f8467-87e0-431d-9695-bf662bf567f6 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.022539] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-618baf85-0896-4fc2-b1ef-24618b3b2fb7 tempest-FloatingIPsAssociationNegativeTestJSON-737087291 tempest-FloatingIPsAssociationNegativeTestJSON-737087291-project-member] [instance: d8468efa-c2d3-4dce-ab89-fc077011e3d8] Reconfiguring VM instance instance-00000054 to attach disk [datastore2] d8468efa-c2d3-4dce-ab89-fc077011e3d8/d8468efa-c2d3-4dce-ab89-fc077011e3d8.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 998.025567] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5bb2bef5-8cc7-4586-be39-d5067bb75ef1 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.044937] env[65726]: WARNING openstack [req-688629ea-98e2-4cec-895d-ef952731929d req-adf5fa76-a785-4e72-961e-c03ff76e70f1 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 998.046303] env[65726]: WARNING openstack [req-688629ea-98e2-4cec-895d-ef952731929d req-adf5fa76-a785-4e72-961e-c03ff76e70f1 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 998.058571] env[65726]: DEBUG oslo_vmware.api [None req-618baf85-0896-4fc2-b1ef-24618b3b2fb7 tempest-FloatingIPsAssociationNegativeTestJSON-737087291 tempest-FloatingIPsAssociationNegativeTestJSON-737087291-project-member] Waiting for the task: (returnval){ [ 998.058571] env[65726]: value = "task-5116436" [ 998.058571] env[65726]: _type = "Task" [ 998.058571] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 998.067682] env[65726]: DEBUG oslo_vmware.api [None req-618baf85-0896-4fc2-b1ef-24618b3b2fb7 tempest-FloatingIPsAssociationNegativeTestJSON-737087291 tempest-FloatingIPsAssociationNegativeTestJSON-737087291-project-member] Task: {'id': task-5116436, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.136958] env[65726]: WARNING openstack [req-688629ea-98e2-4cec-895d-ef952731929d req-adf5fa76-a785-4e72-961e-c03ff76e70f1 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 998.136958] env[65726]: WARNING openstack [req-688629ea-98e2-4cec-895d-ef952731929d req-adf5fa76-a785-4e72-961e-c03ff76e70f1 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 998.196064] env[65726]: DEBUG oslo_vmware.api [None req-775eaf93-375d-4ff3-9c57-60f5cc32d4ea tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5116434, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.221866] env[65726]: DEBUG nova.compute.manager [None req-a27b5bca-2338-4542-8dd7-9cac7e4f3b3a tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] [instance: a286ba4f-a83e-4c30-8079-ba88e3ba5696] Starting instance... {{(pid=65726) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 998.235648] env[65726]: DEBUG oslo_vmware.api [None req-c75c9ed0-c9ca-4fd9-9695-87538dcf1f78 tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Task: {'id': task-5116435, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.255247] env[65726]: DEBUG nova.network.neutron [req-688629ea-98e2-4cec-895d-ef952731929d req-adf5fa76-a785-4e72-961e-c03ff76e70f1 service nova] [instance: 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04] Updated VIF entry in instance network info cache for port 25b9d108-0b43-4459-b9db-7bd90a495bb3. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 998.255885] env[65726]: DEBUG nova.network.neutron [req-688629ea-98e2-4cec-895d-ef952731929d req-adf5fa76-a785-4e72-961e-c03ff76e70f1 service nova] [instance: 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04] Updating instance_info_cache with network_info: [{"id": "25b9d108-0b43-4459-b9db-7bd90a495bb3", "address": "fa:16:3e:5a:1e:6b", "network": {"id": "eb3278b7-c190-441b-84d2-305f9b896382", "bridge": null, "label": "tempest-AttachVolumeShelveTestJSON-1754571022-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.149", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "96149159e18e44f9bf3453e67681f224", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap25b9d108-0b", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 998.345312] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c0f5780-43d1-4999-bb38-ce13f80fbe1e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.356470] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1802f6b4-4677-498c-ad2b-31eabb856453 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.395593] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0fbad25-25c7-4ea0-9881-3224592edccb {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.406950] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e703345-9a96-4cd4-afa5-70397ba72df1 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.422486] env[65726]: DEBUG nova.compute.provider_tree [None req-c5289ae6-5b56-421a-a0e1-7e74e3494692 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 998.448146] env[65726]: DEBUG oslo_concurrency.lockutils [None req-239cc3b9-c000-4612-951a-e583f6a67697 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Lock "interface-dcde25df-87a5-47fb-94ce-334f68894e04-280bd1f3-a22e-4f59-8574-d1783c74f8f3" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 7.988s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 998.571812] env[65726]: DEBUG oslo_vmware.api [None req-618baf85-0896-4fc2-b1ef-24618b3b2fb7 tempest-FloatingIPsAssociationNegativeTestJSON-737087291 tempest-FloatingIPsAssociationNegativeTestJSON-737087291-project-member] Task: {'id': task-5116436, 'name': ReconfigVM_Task, 'duration_secs': 0.296254} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 998.572123] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-618baf85-0896-4fc2-b1ef-24618b3b2fb7 tempest-FloatingIPsAssociationNegativeTestJSON-737087291 tempest-FloatingIPsAssociationNegativeTestJSON-737087291-project-member] [instance: d8468efa-c2d3-4dce-ab89-fc077011e3d8] Reconfigured VM instance instance-00000054 to attach disk [datastore2] d8468efa-c2d3-4dce-ab89-fc077011e3d8/d8468efa-c2d3-4dce-ab89-fc077011e3d8.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 998.572827] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-13d48b0d-dbf5-4718-a756-9f0bdbdff6f0 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.581055] env[65726]: DEBUG oslo_vmware.api [None req-618baf85-0896-4fc2-b1ef-24618b3b2fb7 tempest-FloatingIPsAssociationNegativeTestJSON-737087291 tempest-FloatingIPsAssociationNegativeTestJSON-737087291-project-member] Waiting for the task: (returnval){ [ 998.581055] env[65726]: value = "task-5116437" [ 998.581055] env[65726]: _type = "Task" [ 998.581055] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 998.591790] env[65726]: DEBUG oslo_vmware.api [None req-618baf85-0896-4fc2-b1ef-24618b3b2fb7 tempest-FloatingIPsAssociationNegativeTestJSON-737087291 tempest-FloatingIPsAssociationNegativeTestJSON-737087291-project-member] Task: {'id': task-5116437, 'name': Rename_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.692730] env[65726]: DEBUG oslo_vmware.api [None req-775eaf93-375d-4ff3-9c57-60f5cc32d4ea tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5116434, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.950463} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 998.692996] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-775eaf93-375d-4ff3-9c57-60f5cc32d4ea tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: a56be6fa-6027-46ee-9aa8-332fe4222d68] Extended root virtual disk {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 998.693998] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7048d38-5619-4218-82f3-5a2d9e43a5f9 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.722446] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-775eaf93-375d-4ff3-9c57-60f5cc32d4ea tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: a56be6fa-6027-46ee-9aa8-332fe4222d68] Reconfiguring VM instance instance-00000053 to attach disk [datastore1] a56be6fa-6027-46ee-9aa8-332fe4222d68/a56be6fa-6027-46ee-9aa8-332fe4222d68.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 998.725685] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5981b818-90a1-4e41-b942-514d5ae635ba {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.762772] env[65726]: DEBUG oslo_concurrency.lockutils [req-688629ea-98e2-4cec-895d-ef952731929d req-adf5fa76-a785-4e72-961e-c03ff76e70f1 service nova] Releasing lock "refresh_cache-2fe6fc5a-f5c7-4f8e-96df-4e621a252f04" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 998.762772] env[65726]: DEBUG nova.compute.manager [req-688629ea-98e2-4cec-895d-ef952731929d req-adf5fa76-a785-4e72-961e-c03ff76e70f1 service nova] [instance: dcde25df-87a5-47fb-94ce-334f68894e04] Received event network-vif-plugged-280bd1f3-a22e-4f59-8574-d1783c74f8f3 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 998.762908] env[65726]: DEBUG oslo_concurrency.lockutils [req-688629ea-98e2-4cec-895d-ef952731929d req-adf5fa76-a785-4e72-961e-c03ff76e70f1 service nova] Acquiring lock "dcde25df-87a5-47fb-94ce-334f68894e04-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 998.763130] env[65726]: DEBUG oslo_concurrency.lockutils [req-688629ea-98e2-4cec-895d-ef952731929d req-adf5fa76-a785-4e72-961e-c03ff76e70f1 service nova] Lock "dcde25df-87a5-47fb-94ce-334f68894e04-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 998.763286] env[65726]: DEBUG oslo_concurrency.lockutils [req-688629ea-98e2-4cec-895d-ef952731929d req-adf5fa76-a785-4e72-961e-c03ff76e70f1 service nova] Lock "dcde25df-87a5-47fb-94ce-334f68894e04-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 998.763462] env[65726]: DEBUG nova.compute.manager [req-688629ea-98e2-4cec-895d-ef952731929d req-adf5fa76-a785-4e72-961e-c03ff76e70f1 service nova] [instance: dcde25df-87a5-47fb-94ce-334f68894e04] No waiting events found dispatching network-vif-plugged-280bd1f3-a22e-4f59-8574-d1783c74f8f3 {{(pid=65726) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 998.763620] env[65726]: WARNING nova.compute.manager [req-688629ea-98e2-4cec-895d-ef952731929d req-adf5fa76-a785-4e72-961e-c03ff76e70f1 service nova] [instance: dcde25df-87a5-47fb-94ce-334f68894e04] Received unexpected event network-vif-plugged-280bd1f3-a22e-4f59-8574-d1783c74f8f3 for instance with vm_state active and task_state None. [ 998.763782] env[65726]: DEBUG nova.compute.manager [req-688629ea-98e2-4cec-895d-ef952731929d req-adf5fa76-a785-4e72-961e-c03ff76e70f1 service nova] [instance: dcde25df-87a5-47fb-94ce-334f68894e04] Received event network-changed-280bd1f3-a22e-4f59-8574-d1783c74f8f3 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 998.763933] env[65726]: DEBUG nova.compute.manager [req-688629ea-98e2-4cec-895d-ef952731929d req-adf5fa76-a785-4e72-961e-c03ff76e70f1 service nova] [instance: dcde25df-87a5-47fb-94ce-334f68894e04] Refreshing instance network info cache due to event network-changed-280bd1f3-a22e-4f59-8574-d1783c74f8f3. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 998.764126] env[65726]: DEBUG oslo_concurrency.lockutils [req-688629ea-98e2-4cec-895d-ef952731929d req-adf5fa76-a785-4e72-961e-c03ff76e70f1 service nova] Acquiring lock "refresh_cache-dcde25df-87a5-47fb-94ce-334f68894e04" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 998.764264] env[65726]: DEBUG oslo_concurrency.lockutils [req-688629ea-98e2-4cec-895d-ef952731929d req-adf5fa76-a785-4e72-961e-c03ff76e70f1 service nova] Acquired lock "refresh_cache-dcde25df-87a5-47fb-94ce-334f68894e04" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 998.764417] env[65726]: DEBUG nova.network.neutron [req-688629ea-98e2-4cec-895d-ef952731929d req-adf5fa76-a785-4e72-961e-c03ff76e70f1 service nova] [instance: dcde25df-87a5-47fb-94ce-334f68894e04] Refreshing network info cache for port 280bd1f3-a22e-4f59-8574-d1783c74f8f3 {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 998.772021] env[65726]: DEBUG oslo_vmware.api [None req-775eaf93-375d-4ff3-9c57-60f5cc32d4ea tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Waiting for the task: (returnval){ [ 998.772021] env[65726]: value = "task-5116438" [ 998.772021] env[65726]: _type = "Task" [ 998.772021] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 998.772281] env[65726]: DEBUG oslo_vmware.api [None req-c75c9ed0-c9ca-4fd9-9695-87538dcf1f78 tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Task: {'id': task-5116435, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.784296] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a27b5bca-2338-4542-8dd7-9cac7e4f3b3a tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 998.785087] env[65726]: DEBUG oslo_vmware.api [None req-775eaf93-375d-4ff3-9c57-60f5cc32d4ea tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5116438, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.906175] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3a484f95-3cdc-4dca-a553-0dec45f0913e tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Acquiring lock "815ee20b-cb07-4514-bd43-1b74972508a5" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 998.906466] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3a484f95-3cdc-4dca-a553-0dec45f0913e tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Lock "815ee20b-cb07-4514-bd43-1b74972508a5" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 998.906679] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3a484f95-3cdc-4dca-a553-0dec45f0913e tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Acquiring lock "815ee20b-cb07-4514-bd43-1b74972508a5-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 998.906919] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3a484f95-3cdc-4dca-a553-0dec45f0913e tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Lock "815ee20b-cb07-4514-bd43-1b74972508a5-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 998.907125] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3a484f95-3cdc-4dca-a553-0dec45f0913e tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Lock "815ee20b-cb07-4514-bd43-1b74972508a5-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 998.909308] env[65726]: INFO nova.compute.manager [None req-3a484f95-3cdc-4dca-a553-0dec45f0913e tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] [instance: 815ee20b-cb07-4514-bd43-1b74972508a5] Terminating instance [ 998.925722] env[65726]: DEBUG nova.scheduler.client.report [None req-c5289ae6-5b56-421a-a0e1-7e74e3494692 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 999.092954] env[65726]: DEBUG oslo_vmware.api [None req-618baf85-0896-4fc2-b1ef-24618b3b2fb7 tempest-FloatingIPsAssociationNegativeTestJSON-737087291 tempest-FloatingIPsAssociationNegativeTestJSON-737087291-project-member] Task: {'id': task-5116437, 'name': Rename_Task, 'duration_secs': 0.156273} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 999.092954] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-618baf85-0896-4fc2-b1ef-24618b3b2fb7 tempest-FloatingIPsAssociationNegativeTestJSON-737087291 tempest-FloatingIPsAssociationNegativeTestJSON-737087291-project-member] [instance: d8468efa-c2d3-4dce-ab89-fc077011e3d8] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 999.093322] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-def87dc1-0b4e-4ba3-b768-fa50cf6b391d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.101558] env[65726]: DEBUG oslo_vmware.api [None req-618baf85-0896-4fc2-b1ef-24618b3b2fb7 tempest-FloatingIPsAssociationNegativeTestJSON-737087291 tempest-FloatingIPsAssociationNegativeTestJSON-737087291-project-member] Waiting for the task: (returnval){ [ 999.101558] env[65726]: value = "task-5116439" [ 999.101558] env[65726]: _type = "Task" [ 999.101558] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 999.111769] env[65726]: DEBUG oslo_vmware.api [None req-618baf85-0896-4fc2-b1ef-24618b3b2fb7 tempest-FloatingIPsAssociationNegativeTestJSON-737087291 tempest-FloatingIPsAssociationNegativeTestJSON-737087291-project-member] Task: {'id': task-5116439, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.251729] env[65726]: DEBUG oslo_concurrency.lockutils [None req-22e29168-0ae7-43a0-9b85-be2977371fed tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Acquiring lock "f9a40d36-dda9-454c-8774-ed011f3f50ae" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 999.252071] env[65726]: DEBUG oslo_concurrency.lockutils [None req-22e29168-0ae7-43a0-9b85-be2977371fed tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Lock "f9a40d36-dda9-454c-8774-ed011f3f50ae" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 999.252301] env[65726]: DEBUG oslo_concurrency.lockutils [None req-22e29168-0ae7-43a0-9b85-be2977371fed tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Acquiring lock "f9a40d36-dda9-454c-8774-ed011f3f50ae-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 999.252574] env[65726]: DEBUG oslo_concurrency.lockutils [None req-22e29168-0ae7-43a0-9b85-be2977371fed tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Lock "f9a40d36-dda9-454c-8774-ed011f3f50ae-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 999.253294] env[65726]: DEBUG oslo_concurrency.lockutils [None req-22e29168-0ae7-43a0-9b85-be2977371fed tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Lock "f9a40d36-dda9-454c-8774-ed011f3f50ae-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 999.256121] env[65726]: INFO nova.compute.manager [None req-22e29168-0ae7-43a0-9b85-be2977371fed tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] [instance: f9a40d36-dda9-454c-8774-ed011f3f50ae] Terminating instance [ 999.271059] env[65726]: WARNING openstack [req-688629ea-98e2-4cec-895d-ef952731929d req-adf5fa76-a785-4e72-961e-c03ff76e70f1 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 999.271059] env[65726]: WARNING openstack [req-688629ea-98e2-4cec-895d-ef952731929d req-adf5fa76-a785-4e72-961e-c03ff76e70f1 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 999.277396] env[65726]: DEBUG oslo_vmware.api [None req-c75c9ed0-c9ca-4fd9-9695-87538dcf1f78 tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Task: {'id': task-5116435, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.290730] env[65726]: DEBUG oslo_vmware.api [None req-775eaf93-375d-4ff3-9c57-60f5cc32d4ea tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5116438, 'name': ReconfigVM_Task, 'duration_secs': 0.50017} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 999.291050] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-775eaf93-375d-4ff3-9c57-60f5cc32d4ea tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: a56be6fa-6027-46ee-9aa8-332fe4222d68] Reconfigured VM instance instance-00000053 to attach disk [datastore1] a56be6fa-6027-46ee-9aa8-332fe4222d68/a56be6fa-6027-46ee-9aa8-332fe4222d68.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 999.291854] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2345e80e-3bfc-4f3a-b400-7daebdd8ae00 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.299870] env[65726]: DEBUG oslo_vmware.api [None req-775eaf93-375d-4ff3-9c57-60f5cc32d4ea tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Waiting for the task: (returnval){ [ 999.299870] env[65726]: value = "task-5116440" [ 999.299870] env[65726]: _type = "Task" [ 999.299870] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 999.311037] env[65726]: DEBUG oslo_vmware.api [None req-775eaf93-375d-4ff3-9c57-60f5cc32d4ea tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5116440, 'name': Rename_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.413895] env[65726]: DEBUG nova.compute.manager [None req-3a484f95-3cdc-4dca-a553-0dec45f0913e tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] [instance: 815ee20b-cb07-4514-bd43-1b74972508a5] Start destroying the instance on the hypervisor. {{(pid=65726) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 999.414367] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-3a484f95-3cdc-4dca-a553-0dec45f0913e tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] [instance: 815ee20b-cb07-4514-bd43-1b74972508a5] Destroying instance {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 999.415819] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67cb3e53-2ce8-432d-ba60-c2fdb232d61d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.429602] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-3a484f95-3cdc-4dca-a553-0dec45f0913e tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] [instance: 815ee20b-cb07-4514-bd43-1b74972508a5] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 999.429955] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3232a598-1f59-446e-bc00-a4bbce3e3761 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.436834] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c5289ae6-5b56-421a-a0e1-7e74e3494692 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.015s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 999.444051] env[65726]: DEBUG oslo_concurrency.lockutils [None req-378e66fe-caea-4dff-a5bd-614d0f24e5ba tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 15.961s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 999.444615] env[65726]: DEBUG nova.objects.instance [None req-378e66fe-caea-4dff-a5bd-614d0f24e5ba tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] Lazy-loading 'resources' on Instance uuid 73f5b23e-4463-4068-8994-dd2752a2abc9 {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 999.456425] env[65726]: DEBUG oslo_vmware.api [None req-3a484f95-3cdc-4dca-a553-0dec45f0913e tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Waiting for the task: (returnval){ [ 999.456425] env[65726]: value = "task-5116441" [ 999.456425] env[65726]: _type = "Task" [ 999.456425] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 999.467582] env[65726]: DEBUG oslo_vmware.api [None req-3a484f95-3cdc-4dca-a553-0dec45f0913e tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Task: {'id': task-5116441, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.475099] env[65726]: INFO nova.scheduler.client.report [None req-c5289ae6-5b56-421a-a0e1-7e74e3494692 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Deleted allocations for instance f2d9090c-988f-43f4-9c81-7aa718a3438a [ 999.568883] env[65726]: WARNING openstack [req-688629ea-98e2-4cec-895d-ef952731929d req-adf5fa76-a785-4e72-961e-c03ff76e70f1 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 999.569737] env[65726]: WARNING openstack [req-688629ea-98e2-4cec-895d-ef952731929d req-adf5fa76-a785-4e72-961e-c03ff76e70f1 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 999.615447] env[65726]: DEBUG oslo_vmware.api [None req-618baf85-0896-4fc2-b1ef-24618b3b2fb7 tempest-FloatingIPsAssociationNegativeTestJSON-737087291 tempest-FloatingIPsAssociationNegativeTestJSON-737087291-project-member] Task: {'id': task-5116439, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.618199] env[65726]: DEBUG oslo_concurrency.lockutils [None req-69b4acd5-8968-4363-a895-1312ce324cfb tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Acquiring lock "54e85161-5f63-405f-83f8-490c70645a3b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 999.618934] env[65726]: DEBUG oslo_concurrency.lockutils [None req-69b4acd5-8968-4363-a895-1312ce324cfb tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Lock "54e85161-5f63-405f-83f8-490c70645a3b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 999.619276] env[65726]: DEBUG oslo_concurrency.lockutils [None req-69b4acd5-8968-4363-a895-1312ce324cfb tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Acquiring lock "54e85161-5f63-405f-83f8-490c70645a3b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 999.619683] env[65726]: DEBUG oslo_concurrency.lockutils [None req-69b4acd5-8968-4363-a895-1312ce324cfb tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Lock "54e85161-5f63-405f-83f8-490c70645a3b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 999.620038] env[65726]: DEBUG oslo_concurrency.lockutils [None req-69b4acd5-8968-4363-a895-1312ce324cfb tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Lock "54e85161-5f63-405f-83f8-490c70645a3b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 999.624836] env[65726]: INFO nova.compute.manager [None req-69b4acd5-8968-4363-a895-1312ce324cfb tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 54e85161-5f63-405f-83f8-490c70645a3b] Terminating instance [ 999.703125] env[65726]: WARNING openstack [req-688629ea-98e2-4cec-895d-ef952731929d req-adf5fa76-a785-4e72-961e-c03ff76e70f1 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 999.703421] env[65726]: WARNING openstack [req-688629ea-98e2-4cec-895d-ef952731929d req-adf5fa76-a785-4e72-961e-c03ff76e70f1 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 999.751665] env[65726]: DEBUG oslo_concurrency.lockutils [None req-bd9cdf9a-c493-4e4b-996c-7dba718501da tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Acquiring lock "067c9332-fa2c-47fe-93ae-643de1fe77cf" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 999.752477] env[65726]: DEBUG oslo_concurrency.lockutils [None req-bd9cdf9a-c493-4e4b-996c-7dba718501da tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Lock "067c9332-fa2c-47fe-93ae-643de1fe77cf" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.001s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 999.752672] env[65726]: INFO nova.compute.manager [None req-bd9cdf9a-c493-4e4b-996c-7dba718501da tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: 067c9332-fa2c-47fe-93ae-643de1fe77cf] Shelving [ 999.764020] env[65726]: DEBUG oslo_vmware.api [None req-c75c9ed0-c9ca-4fd9-9695-87538dcf1f78 tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Task: {'id': task-5116435, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.579756} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 999.765703] env[65726]: DEBUG nova.compute.manager [None req-22e29168-0ae7-43a0-9b85-be2977371fed tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] [instance: f9a40d36-dda9-454c-8774-ed011f3f50ae] Start destroying the instance on the hypervisor. {{(pid=65726) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 999.765905] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-22e29168-0ae7-43a0-9b85-be2977371fed tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] [instance: f9a40d36-dda9-454c-8774-ed011f3f50ae] Destroying instance {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 999.766191] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-c75c9ed0-c9ca-4fd9-9695-87538dcf1f78 tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] 406c59e7-8eb8-4cfd-8e1e-e151ab908571/406c59e7-8eb8-4cfd-8e1e-e151ab908571.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 999.766388] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-c75c9ed0-c9ca-4fd9-9695-87538dcf1f78 tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] [instance: 406c59e7-8eb8-4cfd-8e1e-e151ab908571] Extending root virtual disk to 1048576 {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 999.767305] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8dbf364c-beb5-4047-a665-f19265d03fdf {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.770599] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2e192dcc-2327-4cd3-a659-e16eb7346978 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.778917] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-22e29168-0ae7-43a0-9b85-be2977371fed tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] [instance: f9a40d36-dda9-454c-8774-ed011f3f50ae] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 999.780257] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fc9f2c8d-5943-49dc-834d-86bee7407c4c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.782021] env[65726]: DEBUG oslo_vmware.api [None req-c75c9ed0-c9ca-4fd9-9695-87538dcf1f78 tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Waiting for the task: (returnval){ [ 999.782021] env[65726]: value = "task-5116442" [ 999.782021] env[65726]: _type = "Task" [ 999.782021] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 999.789388] env[65726]: DEBUG oslo_vmware.api [None req-22e29168-0ae7-43a0-9b85-be2977371fed tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Waiting for the task: (returnval){ [ 999.789388] env[65726]: value = "task-5116443" [ 999.789388] env[65726]: _type = "Task" [ 999.789388] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 999.793525] env[65726]: DEBUG oslo_vmware.api [None req-c75c9ed0-c9ca-4fd9-9695-87538dcf1f78 tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Task: {'id': task-5116442, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.802540] env[65726]: DEBUG oslo_vmware.api [None req-22e29168-0ae7-43a0-9b85-be2977371fed tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Task: {'id': task-5116443, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.817237] env[65726]: DEBUG oslo_vmware.api [None req-775eaf93-375d-4ff3-9c57-60f5cc32d4ea tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5116440, 'name': Rename_Task, 'duration_secs': 0.2985} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 999.817735] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-775eaf93-375d-4ff3-9c57-60f5cc32d4ea tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: a56be6fa-6027-46ee-9aa8-332fe4222d68] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 999.818145] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f8927322-2f33-4aa1-800f-d8a8362421cd {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.827373] env[65726]: DEBUG oslo_vmware.api [None req-775eaf93-375d-4ff3-9c57-60f5cc32d4ea tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Waiting for the task: (returnval){ [ 999.827373] env[65726]: value = "task-5116444" [ 999.827373] env[65726]: _type = "Task" [ 999.827373] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 999.836251] env[65726]: DEBUG oslo_vmware.api [None req-775eaf93-375d-4ff3-9c57-60f5cc32d4ea tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5116444, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.965367] env[65726]: DEBUG oslo_vmware.api [None req-3a484f95-3cdc-4dca-a553-0dec45f0913e tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Task: {'id': task-5116441, 'name': PowerOffVM_Task, 'duration_secs': 0.265291} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 999.965677] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-3a484f95-3cdc-4dca-a553-0dec45f0913e tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] [instance: 815ee20b-cb07-4514-bd43-1b74972508a5] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 999.965843] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-3a484f95-3cdc-4dca-a553-0dec45f0913e tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] [instance: 815ee20b-cb07-4514-bd43-1b74972508a5] Unregistering the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 999.966125] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9aedfdf6-8b9e-4fcb-ad3f-0bb11015c204 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.990444] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c5289ae6-5b56-421a-a0e1-7e74e3494692 tempest-ServersAdminTestJSON-1403998232 tempest-ServersAdminTestJSON-1403998232-project-member] Lock "f2d9090c-988f-43f4-9c81-7aa718a3438a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 21.282s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1000.039906] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-3a484f95-3cdc-4dca-a553-0dec45f0913e tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] [instance: 815ee20b-cb07-4514-bd43-1b74972508a5] Unregistered the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1000.040202] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-3a484f95-3cdc-4dca-a553-0dec45f0913e tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] [instance: 815ee20b-cb07-4514-bd43-1b74972508a5] Deleting contents of the VM from datastore datastore2 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1000.040441] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-3a484f95-3cdc-4dca-a553-0dec45f0913e tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Deleting the datastore file [datastore2] 815ee20b-cb07-4514-bd43-1b74972508a5 {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1000.040774] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-33f917f1-1d62-451c-a163-9aec9ace87ea {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.049983] env[65726]: DEBUG oslo_vmware.api [None req-3a484f95-3cdc-4dca-a553-0dec45f0913e tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Waiting for the task: (returnval){ [ 1000.049983] env[65726]: value = "task-5116446" [ 1000.049983] env[65726]: _type = "Task" [ 1000.049983] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1000.067807] env[65726]: DEBUG oslo_vmware.api [None req-3a484f95-3cdc-4dca-a553-0dec45f0913e tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Task: {'id': task-5116446, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.113499] env[65726]: DEBUG oslo_vmware.api [None req-618baf85-0896-4fc2-b1ef-24618b3b2fb7 tempest-FloatingIPsAssociationNegativeTestJSON-737087291 tempest-FloatingIPsAssociationNegativeTestJSON-737087291-project-member] Task: {'id': task-5116439, 'name': PowerOnVM_Task, 'duration_secs': 0.521173} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.113783] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-618baf85-0896-4fc2-b1ef-24618b3b2fb7 tempest-FloatingIPsAssociationNegativeTestJSON-737087291 tempest-FloatingIPsAssociationNegativeTestJSON-737087291-project-member] [instance: d8468efa-c2d3-4dce-ab89-fc077011e3d8] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1000.113985] env[65726]: INFO nova.compute.manager [None req-618baf85-0896-4fc2-b1ef-24618b3b2fb7 tempest-FloatingIPsAssociationNegativeTestJSON-737087291 tempest-FloatingIPsAssociationNegativeTestJSON-737087291-project-member] [instance: d8468efa-c2d3-4dce-ab89-fc077011e3d8] Took 7.67 seconds to spawn the instance on the hypervisor. [ 1000.114178] env[65726]: DEBUG nova.compute.manager [None req-618baf85-0896-4fc2-b1ef-24618b3b2fb7 tempest-FloatingIPsAssociationNegativeTestJSON-737087291 tempest-FloatingIPsAssociationNegativeTestJSON-737087291-project-member] [instance: d8468efa-c2d3-4dce-ab89-fc077011e3d8] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1000.115012] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8b2a651-4a9b-4840-9b6f-22aeb09020fb {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.130432] env[65726]: DEBUG nova.compute.manager [None req-69b4acd5-8968-4363-a895-1312ce324cfb tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 54e85161-5f63-405f-83f8-490c70645a3b] Start destroying the instance on the hypervisor. {{(pid=65726) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1000.130432] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-69b4acd5-8968-4363-a895-1312ce324cfb tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 54e85161-5f63-405f-83f8-490c70645a3b] Destroying instance {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1000.131920] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15e18780-b909-45c6-8472-128cc89889c1 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.142861] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-69b4acd5-8968-4363-a895-1312ce324cfb tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 54e85161-5f63-405f-83f8-490c70645a3b] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1000.143644] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6785cd30-9534-49ca-ae93-58bb8b47aa22 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.152186] env[65726]: DEBUG oslo_vmware.api [None req-69b4acd5-8968-4363-a895-1312ce324cfb tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Waiting for the task: (returnval){ [ 1000.152186] env[65726]: value = "task-5116447" [ 1000.152186] env[65726]: _type = "Task" [ 1000.152186] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1000.167160] env[65726]: DEBUG oslo_vmware.api [None req-69b4acd5-8968-4363-a895-1312ce324cfb tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': task-5116447, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.293685] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5886046-180e-404c-bf8a-bac57d3c2ab9 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.304042] env[65726]: DEBUG oslo_vmware.api [None req-c75c9ed0-c9ca-4fd9-9695-87538dcf1f78 tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Task: {'id': task-5116442, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.11946} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.308609] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-c75c9ed0-c9ca-4fd9-9695-87538dcf1f78 tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] [instance: 406c59e7-8eb8-4cfd-8e1e-e151ab908571] Extended root virtual disk {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1000.309937] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-234b735f-823f-4c47-aae5-14572bd88af8 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.318467] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-edd7470a-4614-486a-860b-1fe79fd4b096 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.359111] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-c75c9ed0-c9ca-4fd9-9695-87538dcf1f78 tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] [instance: 406c59e7-8eb8-4cfd-8e1e-e151ab908571] Reconfiguring VM instance instance-00000050 to attach disk [datastore1] 406c59e7-8eb8-4cfd-8e1e-e151ab908571/406c59e7-8eb8-4cfd-8e1e-e151ab908571.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1000.359395] env[65726]: DEBUG oslo_vmware.api [None req-22e29168-0ae7-43a0-9b85-be2977371fed tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Task: {'id': task-5116443, 'name': PowerOffVM_Task, 'duration_secs': 0.51276} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.364124] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-99379c43-aca3-45db-8454-b83403a4f4a5 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.378380] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-22e29168-0ae7-43a0-9b85-be2977371fed tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] [instance: f9a40d36-dda9-454c-8774-ed011f3f50ae] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1000.378574] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-22e29168-0ae7-43a0-9b85-be2977371fed tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] [instance: f9a40d36-dda9-454c-8774-ed011f3f50ae] Unregistering the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1000.405970] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-da95fefa-4c10-4f44-99a5-1886b23a1ac6 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.409243] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70755b33-8c6b-4c23-861f-e5b8da4690c9 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.415987] env[65726]: DEBUG oslo_vmware.api [None req-775eaf93-375d-4ff3-9c57-60f5cc32d4ea tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5116444, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.418130] env[65726]: DEBUG oslo_vmware.api [None req-c75c9ed0-c9ca-4fd9-9695-87538dcf1f78 tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Waiting for the task: (returnval){ [ 1000.418130] env[65726]: value = "task-5116448" [ 1000.418130] env[65726]: _type = "Task" [ 1000.418130] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1000.426175] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c1213ab-d3cf-4d58-b94d-32f78f791d17 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.434428] env[65726]: DEBUG oslo_vmware.api [None req-c75c9ed0-c9ca-4fd9-9695-87538dcf1f78 tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Task: {'id': task-5116448, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.445827] env[65726]: DEBUG nova.compute.provider_tree [None req-378e66fe-caea-4dff-a5bd-614d0f24e5ba tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1000.508254] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-22e29168-0ae7-43a0-9b85-be2977371fed tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] [instance: f9a40d36-dda9-454c-8774-ed011f3f50ae] Unregistered the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1000.508254] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-22e29168-0ae7-43a0-9b85-be2977371fed tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] [instance: f9a40d36-dda9-454c-8774-ed011f3f50ae] Deleting contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1000.508254] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-22e29168-0ae7-43a0-9b85-be2977371fed tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Deleting the datastore file [datastore1] f9a40d36-dda9-454c-8774-ed011f3f50ae {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1000.508254] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f93b0cfe-35fc-43f7-93c8-fca661e001a6 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.521653] env[65726]: DEBUG oslo_vmware.api [None req-22e29168-0ae7-43a0-9b85-be2977371fed tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Waiting for the task: (returnval){ [ 1000.521653] env[65726]: value = "task-5116450" [ 1000.521653] env[65726]: _type = "Task" [ 1000.521653] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1000.532400] env[65726]: DEBUG oslo_vmware.api [None req-22e29168-0ae7-43a0-9b85-be2977371fed tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Task: {'id': task-5116450, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.564348] env[65726]: DEBUG oslo_vmware.api [None req-3a484f95-3cdc-4dca-a553-0dec45f0913e tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Task: {'id': task-5116446, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.412267} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.564910] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-3a484f95-3cdc-4dca-a553-0dec45f0913e tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Deleted the datastore file {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1000.564910] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-3a484f95-3cdc-4dca-a553-0dec45f0913e tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] [instance: 815ee20b-cb07-4514-bd43-1b74972508a5] Deleted contents of the VM from datastore datastore2 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1000.565087] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-3a484f95-3cdc-4dca-a553-0dec45f0913e tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] [instance: 815ee20b-cb07-4514-bd43-1b74972508a5] Instance destroyed {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1000.565197] env[65726]: INFO nova.compute.manager [None req-3a484f95-3cdc-4dca-a553-0dec45f0913e tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] [instance: 815ee20b-cb07-4514-bd43-1b74972508a5] Took 1.15 seconds to destroy the instance on the hypervisor. [ 1000.565450] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-3a484f95-3cdc-4dca-a553-0dec45f0913e tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1000.565651] env[65726]: DEBUG nova.compute.manager [-] [instance: 815ee20b-cb07-4514-bd43-1b74972508a5] Deallocating network for instance {{(pid=65726) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1000.565749] env[65726]: DEBUG nova.network.neutron [-] [instance: 815ee20b-cb07-4514-bd43-1b74972508a5] deallocate_for_instance() {{(pid=65726) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 1000.567112] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1000.567112] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1000.647843] env[65726]: INFO nova.compute.manager [None req-618baf85-0896-4fc2-b1ef-24618b3b2fb7 tempest-FloatingIPsAssociationNegativeTestJSON-737087291 tempest-FloatingIPsAssociationNegativeTestJSON-737087291-project-member] [instance: d8468efa-c2d3-4dce-ab89-fc077011e3d8] Took 28.88 seconds to build instance. [ 1000.662954] env[65726]: DEBUG oslo_vmware.api [None req-69b4acd5-8968-4363-a895-1312ce324cfb tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': task-5116447, 'name': PowerOffVM_Task, 'duration_secs': 0.340148} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.663843] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-69b4acd5-8968-4363-a895-1312ce324cfb tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 54e85161-5f63-405f-83f8-490c70645a3b] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1000.664105] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-69b4acd5-8968-4363-a895-1312ce324cfb tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 54e85161-5f63-405f-83f8-490c70645a3b] Unregistering the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1000.664390] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-efb3798e-1c54-414a-9e8e-0d832fdfa25d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.742169] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-69b4acd5-8968-4363-a895-1312ce324cfb tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 54e85161-5f63-405f-83f8-490c70645a3b] Unregistered the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1000.742545] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-69b4acd5-8968-4363-a895-1312ce324cfb tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 54e85161-5f63-405f-83f8-490c70645a3b] Deleting contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1000.743090] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-69b4acd5-8968-4363-a895-1312ce324cfb tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Deleting the datastore file [datastore1] 54e85161-5f63-405f-83f8-490c70645a3b {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1000.743447] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8f769d68-dd5a-42c3-8468-702882d1c04b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.752837] env[65726]: DEBUG oslo_vmware.api [None req-69b4acd5-8968-4363-a895-1312ce324cfb tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Waiting for the task: (returnval){ [ 1000.752837] env[65726]: value = "task-5116452" [ 1000.752837] env[65726]: _type = "Task" [ 1000.752837] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1000.764859] env[65726]: DEBUG oslo_vmware.api [None req-69b4acd5-8968-4363-a895-1312ce324cfb tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': task-5116452, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.773289] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-bd9cdf9a-c493-4e4b-996c-7dba718501da tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: 067c9332-fa2c-47fe-93ae-643de1fe77cf] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1000.773289] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-960378ef-60f1-4e95-a6e0-0f29ac438b7b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.779167] env[65726]: DEBUG oslo_vmware.api [None req-bd9cdf9a-c493-4e4b-996c-7dba718501da tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Waiting for the task: (returnval){ [ 1000.779167] env[65726]: value = "task-5116453" [ 1000.779167] env[65726]: _type = "Task" [ 1000.779167] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1000.788860] env[65726]: DEBUG oslo_vmware.api [None req-bd9cdf9a-c493-4e4b-996c-7dba718501da tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': task-5116453, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.861478] env[65726]: DEBUG oslo_vmware.api [None req-775eaf93-375d-4ff3-9c57-60f5cc32d4ea tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5116444, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.933122] env[65726]: DEBUG oslo_vmware.api [None req-c75c9ed0-c9ca-4fd9-9695-87538dcf1f78 tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Task: {'id': task-5116448, 'name': ReconfigVM_Task, 'duration_secs': 0.469568} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.933522] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-c75c9ed0-c9ca-4fd9-9695-87538dcf1f78 tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] [instance: 406c59e7-8eb8-4cfd-8e1e-e151ab908571] Reconfigured VM instance instance-00000050 to attach disk [datastore1] 406c59e7-8eb8-4cfd-8e1e-e151ab908571/406c59e7-8eb8-4cfd-8e1e-e151ab908571.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1000.934516] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6a831dec-2f70-4978-8c6c-40407f93de5e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.947531] env[65726]: DEBUG oslo_vmware.api [None req-c75c9ed0-c9ca-4fd9-9695-87538dcf1f78 tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Waiting for the task: (returnval){ [ 1000.947531] env[65726]: value = "task-5116454" [ 1000.947531] env[65726]: _type = "Task" [ 1000.947531] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1000.953223] env[65726]: DEBUG nova.scheduler.client.report [None req-378e66fe-caea-4dff-a5bd-614d0f24e5ba tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1000.963778] env[65726]: DEBUG oslo_vmware.api [None req-c75c9ed0-c9ca-4fd9-9695-87538dcf1f78 tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Task: {'id': task-5116454, 'name': Rename_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.012662] env[65726]: DEBUG nova.network.neutron [req-688629ea-98e2-4cec-895d-ef952731929d req-adf5fa76-a785-4e72-961e-c03ff76e70f1 service nova] [instance: dcde25df-87a5-47fb-94ce-334f68894e04] Updated VIF entry in instance network info cache for port 280bd1f3-a22e-4f59-8574-d1783c74f8f3. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 1001.013498] env[65726]: DEBUG nova.network.neutron [req-688629ea-98e2-4cec-895d-ef952731929d req-adf5fa76-a785-4e72-961e-c03ff76e70f1 service nova] [instance: dcde25df-87a5-47fb-94ce-334f68894e04] Updating instance_info_cache with network_info: [{"id": "bf51c96c-63c5-46de-a40f-58e6769c5fcf", "address": "fa:16:3e:4c:88:87", "network": {"id": "51a66448-0f72-4266-8868-856b48be839f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-155658961-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.146", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d46ccb00794f458b85da4a93879139ee", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b1f3e6c3-5584-4852-9017-476ab8ac4946", "external-id": "nsx-vlan-transportzone-304", "segmentation_id": 304, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbf51c96c-63", "ovs_interfaceid": "bf51c96c-63c5-46de-a40f-58e6769c5fcf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "b48313f9-a0d3-4fdd-88c6-d103aef0b4f2", "address": "fa:16:3e:e7:cd:41", "network": {"id": "51a66448-0f72-4266-8868-856b48be839f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-155658961-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d46ccb00794f458b85da4a93879139ee", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b1f3e6c3-5584-4852-9017-476ab8ac4946", "external-id": "nsx-vlan-transportzone-304", "segmentation_id": 304, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb48313f9-a0", "ovs_interfaceid": "b48313f9-a0d3-4fdd-88c6-d103aef0b4f2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "280bd1f3-a22e-4f59-8574-d1783c74f8f3", "address": "fa:16:3e:2b:59:bd", "network": {"id": "51a66448-0f72-4266-8868-856b48be839f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-155658961-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d46ccb00794f458b85da4a93879139ee", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b1f3e6c3-5584-4852-9017-476ab8ac4946", "external-id": "nsx-vlan-transportzone-304", "segmentation_id": 304, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap280bd1f3-a2", "ovs_interfaceid": "280bd1f3-a22e-4f59-8574-d1783c74f8f3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1001.035659] env[65726]: DEBUG oslo_vmware.api [None req-22e29168-0ae7-43a0-9b85-be2977371fed tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Task: {'id': task-5116450, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.266009} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.036080] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-22e29168-0ae7-43a0-9b85-be2977371fed tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Deleted the datastore file {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1001.036364] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-22e29168-0ae7-43a0-9b85-be2977371fed tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] [instance: f9a40d36-dda9-454c-8774-ed011f3f50ae] Deleted contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1001.036605] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-22e29168-0ae7-43a0-9b85-be2977371fed tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] [instance: f9a40d36-dda9-454c-8774-ed011f3f50ae] Instance destroyed {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1001.036865] env[65726]: INFO nova.compute.manager [None req-22e29168-0ae7-43a0-9b85-be2977371fed tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] [instance: f9a40d36-dda9-454c-8774-ed011f3f50ae] Took 1.27 seconds to destroy the instance on the hypervisor. [ 1001.037263] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-22e29168-0ae7-43a0-9b85-be2977371fed tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1001.037591] env[65726]: DEBUG nova.compute.manager [-] [instance: f9a40d36-dda9-454c-8774-ed011f3f50ae] Deallocating network for instance {{(pid=65726) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1001.037750] env[65726]: DEBUG nova.network.neutron [-] [instance: f9a40d36-dda9-454c-8774-ed011f3f50ae] deallocate_for_instance() {{(pid=65726) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 1001.038453] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1001.038938] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1001.101795] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1001.102360] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1001.148481] env[65726]: DEBUG oslo_concurrency.lockutils [None req-618baf85-0896-4fc2-b1ef-24618b3b2fb7 tempest-FloatingIPsAssociationNegativeTestJSON-737087291 tempest-FloatingIPsAssociationNegativeTestJSON-737087291-project-member] Lock "d8468efa-c2d3-4dce-ab89-fc077011e3d8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 30.399s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1001.265425] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1001.265976] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1001.293924] env[65726]: DEBUG oslo_vmware.api [None req-69b4acd5-8968-4363-a895-1312ce324cfb tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Task: {'id': task-5116452, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.20342} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.294823] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-69b4acd5-8968-4363-a895-1312ce324cfb tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Deleted the datastore file {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1001.295313] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-69b4acd5-8968-4363-a895-1312ce324cfb tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 54e85161-5f63-405f-83f8-490c70645a3b] Deleted contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1001.295588] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-69b4acd5-8968-4363-a895-1312ce324cfb tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 54e85161-5f63-405f-83f8-490c70645a3b] Instance destroyed {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1001.295863] env[65726]: INFO nova.compute.manager [None req-69b4acd5-8968-4363-a895-1312ce324cfb tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] [instance: 54e85161-5f63-405f-83f8-490c70645a3b] Took 1.17 seconds to destroy the instance on the hypervisor. [ 1001.299040] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-69b4acd5-8968-4363-a895-1312ce324cfb tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1001.300416] env[65726]: DEBUG nova.compute.manager [-] [instance: 54e85161-5f63-405f-83f8-490c70645a3b] Deallocating network for instance {{(pid=65726) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1001.300632] env[65726]: DEBUG nova.network.neutron [-] [instance: 54e85161-5f63-405f-83f8-490c70645a3b] deallocate_for_instance() {{(pid=65726) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 1001.301529] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1001.302087] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1001.313765] env[65726]: DEBUG oslo_vmware.api [None req-bd9cdf9a-c493-4e4b-996c-7dba718501da tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': task-5116453, 'name': PowerOffVM_Task, 'duration_secs': 0.211684} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.315253] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-bd9cdf9a-c493-4e4b-996c-7dba718501da tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: 067c9332-fa2c-47fe-93ae-643de1fe77cf] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1001.316736] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd5376fc-eb80-4fc8-bbe6-ce4cc88375b7 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.352924] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e96c836-c032-4d71-a470-22b44b3f030a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.363535] env[65726]: DEBUG oslo_vmware.api [None req-775eaf93-375d-4ff3-9c57-60f5cc32d4ea tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5116444, 'name': PowerOnVM_Task, 'duration_secs': 1.121213} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.365760] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-775eaf93-375d-4ff3-9c57-60f5cc32d4ea tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: a56be6fa-6027-46ee-9aa8-332fe4222d68] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1001.366023] env[65726]: INFO nova.compute.manager [None req-775eaf93-375d-4ff3-9c57-60f5cc32d4ea tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: a56be6fa-6027-46ee-9aa8-332fe4222d68] Took 11.50 seconds to spawn the instance on the hypervisor. [ 1001.366258] env[65726]: DEBUG nova.compute.manager [None req-775eaf93-375d-4ff3-9c57-60f5cc32d4ea tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: a56be6fa-6027-46ee-9aa8-332fe4222d68] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1001.369493] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7305da6-e083-4ae2-a557-d056d578ef48 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.463431] env[65726]: DEBUG oslo_concurrency.lockutils [None req-378e66fe-caea-4dff-a5bd-614d0f24e5ba tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.021s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1001.466182] env[65726]: DEBUG oslo_vmware.api [None req-c75c9ed0-c9ca-4fd9-9695-87538dcf1f78 tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Task: {'id': task-5116454, 'name': Rename_Task, 'duration_secs': 0.180018} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.466182] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5d32d102-744f-471d-8592-b9e461d3322c tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 16.137s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1001.466535] env[65726]: DEBUG nova.objects.instance [None req-5d32d102-744f-471d-8592-b9e461d3322c tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Lazy-loading 'resources' on Instance uuid 060646cc-f0fb-4e6c-9f67-7b2daa6b6fd6 {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1001.467572] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-c75c9ed0-c9ca-4fd9-9695-87538dcf1f78 tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] [instance: 406c59e7-8eb8-4cfd-8e1e-e151ab908571] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1001.468276] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-43fae4eb-1f1a-4183-92e1-9876206e66b2 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.476166] env[65726]: DEBUG oslo_vmware.api [None req-c75c9ed0-c9ca-4fd9-9695-87538dcf1f78 tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Waiting for the task: (returnval){ [ 1001.476166] env[65726]: value = "task-5116455" [ 1001.476166] env[65726]: _type = "Task" [ 1001.476166] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.487348] env[65726]: DEBUG oslo_vmware.api [None req-c75c9ed0-c9ca-4fd9-9695-87538dcf1f78 tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Task: {'id': task-5116455, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.505329] env[65726]: INFO nova.scheduler.client.report [None req-378e66fe-caea-4dff-a5bd-614d0f24e5ba tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] Deleted allocations for instance 73f5b23e-4463-4068-8994-dd2752a2abc9 [ 1001.516961] env[65726]: DEBUG oslo_concurrency.lockutils [req-688629ea-98e2-4cec-895d-ef952731929d req-adf5fa76-a785-4e72-961e-c03ff76e70f1 service nova] Releasing lock "refresh_cache-dcde25df-87a5-47fb-94ce-334f68894e04" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1001.740294] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1001.740582] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1001.874789] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-bd9cdf9a-c493-4e4b-996c-7dba718501da tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: 067c9332-fa2c-47fe-93ae-643de1fe77cf] Creating Snapshot of the VM instance {{(pid=65726) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1001.875166] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-8fc03264-c935-4f9a-972f-20dc3f25c09f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.889199] env[65726]: DEBUG oslo_vmware.api [None req-bd9cdf9a-c493-4e4b-996c-7dba718501da tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Waiting for the task: (returnval){ [ 1001.889199] env[65726]: value = "task-5116456" [ 1001.889199] env[65726]: _type = "Task" [ 1001.889199] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.896207] env[65726]: INFO nova.compute.manager [None req-775eaf93-375d-4ff3-9c57-60f5cc32d4ea tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: a56be6fa-6027-46ee-9aa8-332fe4222d68] Took 33.47 seconds to build instance. [ 1001.905961] env[65726]: DEBUG oslo_vmware.api [None req-bd9cdf9a-c493-4e4b-996c-7dba718501da tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': task-5116456, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.989305] env[65726]: DEBUG oslo_vmware.api [None req-c75c9ed0-c9ca-4fd9-9695-87538dcf1f78 tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Task: {'id': task-5116455, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.015698] env[65726]: DEBUG oslo_concurrency.lockutils [None req-378e66fe-caea-4dff-a5bd-614d0f24e5ba tempest-ServerShowV257Test-188382960 tempest-ServerShowV257Test-188382960-project-member] Lock "73f5b23e-4463-4068-8994-dd2752a2abc9" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 23.582s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1002.089557] env[65726]: DEBUG nova.network.neutron [-] [instance: 815ee20b-cb07-4514-bd43-1b74972508a5] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1002.322675] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-969c736d-91cb-4c0a-aeab-1c04850cbf5c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.333325] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb9591ae-04d4-47c6-80e0-49066b3baca1 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.372896] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf27859a-9bf2-40fe-9e14-8d8a8638fdda {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.382526] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89c04703-af52-493a-9934-51b61a3950f8 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.398929] env[65726]: DEBUG nova.compute.provider_tree [None req-5d32d102-744f-471d-8592-b9e461d3322c tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1002.403067] env[65726]: DEBUG oslo_concurrency.lockutils [None req-775eaf93-375d-4ff3-9c57-60f5cc32d4ea tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Lock "a56be6fa-6027-46ee-9aa8-332fe4222d68" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 34.993s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1002.409824] env[65726]: DEBUG oslo_vmware.api [None req-bd9cdf9a-c493-4e4b-996c-7dba718501da tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': task-5116456, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.427650] env[65726]: DEBUG nova.network.neutron [-] [instance: f9a40d36-dda9-454c-8774-ed011f3f50ae] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1002.488867] env[65726]: DEBUG oslo_vmware.api [None req-c75c9ed0-c9ca-4fd9-9695-87538dcf1f78 tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Task: {'id': task-5116455, 'name': PowerOnVM_Task, 'duration_secs': 0.80197} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1002.490049] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-c75c9ed0-c9ca-4fd9-9695-87538dcf1f78 tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] [instance: 406c59e7-8eb8-4cfd-8e1e-e151ab908571] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1002.490049] env[65726]: INFO nova.compute.manager [None req-c75c9ed0-c9ca-4fd9-9695-87538dcf1f78 tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] [instance: 406c59e7-8eb8-4cfd-8e1e-e151ab908571] Took 19.69 seconds to spawn the instance on the hypervisor. [ 1002.490049] env[65726]: DEBUG nova.compute.manager [None req-c75c9ed0-c9ca-4fd9-9695-87538dcf1f78 tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] [instance: 406c59e7-8eb8-4cfd-8e1e-e151ab908571] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1002.490340] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96ed4fc2-a27a-492f-a7bf-d933b4de7355 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.593735] env[65726]: INFO nova.compute.manager [-] [instance: 815ee20b-cb07-4514-bd43-1b74972508a5] Took 2.03 seconds to deallocate network for instance. [ 1002.709263] env[65726]: DEBUG nova.network.neutron [-] [instance: 54e85161-5f63-405f-83f8-490c70645a3b] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1002.834422] env[65726]: DEBUG nova.compute.manager [req-f8ab8e72-2f6f-4b57-821e-0e9fdc393924 req-8467ceb7-6047-4e4c-9c58-2ccfd09dc770 service nova] [instance: 815ee20b-cb07-4514-bd43-1b74972508a5] Received event network-vif-deleted-c5f4ccbb-2fbd-43ef-979c-d7538b388e9a {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1002.904741] env[65726]: DEBUG nova.scheduler.client.report [None req-5d32d102-744f-471d-8592-b9e461d3322c tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1002.913914] env[65726]: DEBUG oslo_vmware.api [None req-bd9cdf9a-c493-4e4b-996c-7dba718501da tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': task-5116456, 'name': CreateSnapshot_Task, 'duration_secs': 0.60944} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1002.915027] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-bd9cdf9a-c493-4e4b-996c-7dba718501da tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: 067c9332-fa2c-47fe-93ae-643de1fe77cf] Created Snapshot of the VM instance {{(pid=65726) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1002.915703] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b77c5555-5414-43cd-b92d-5055aeda172b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.932515] env[65726]: INFO nova.compute.manager [-] [instance: f9a40d36-dda9-454c-8774-ed011f3f50ae] Took 1.89 seconds to deallocate network for instance. [ 1003.014819] env[65726]: INFO nova.compute.manager [None req-c75c9ed0-c9ca-4fd9-9695-87538dcf1f78 tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] [instance: 406c59e7-8eb8-4cfd-8e1e-e151ab908571] Took 43.23 seconds to build instance. [ 1003.103219] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3a484f95-3cdc-4dca-a553-0dec45f0913e tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1003.214324] env[65726]: INFO nova.compute.manager [-] [instance: 54e85161-5f63-405f-83f8-490c70645a3b] Took 1.91 seconds to deallocate network for instance. [ 1003.416986] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5d32d102-744f-471d-8592-b9e461d3322c tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.951s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1003.419688] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 15.504s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1003.444596] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-bd9cdf9a-c493-4e4b-996c-7dba718501da tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: 067c9332-fa2c-47fe-93ae-643de1fe77cf] Creating linked-clone VM from snapshot {{(pid=65726) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1003.444896] env[65726]: DEBUG oslo_concurrency.lockutils [None req-22e29168-0ae7-43a0-9b85-be2977371fed tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1003.446244] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-79ffa011-1f28-47bc-988c-989df1210ce1 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.456253] env[65726]: DEBUG oslo_vmware.api [None req-bd9cdf9a-c493-4e4b-996c-7dba718501da tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Waiting for the task: (returnval){ [ 1003.456253] env[65726]: value = "task-5116457" [ 1003.456253] env[65726]: _type = "Task" [ 1003.456253] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.457701] env[65726]: INFO nova.scheduler.client.report [None req-5d32d102-744f-471d-8592-b9e461d3322c tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Deleted allocations for instance 060646cc-f0fb-4e6c-9f67-7b2daa6b6fd6 [ 1003.475657] env[65726]: DEBUG oslo_vmware.api [None req-bd9cdf9a-c493-4e4b-996c-7dba718501da tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': task-5116457, 'name': CloneVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.522320] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c75c9ed0-c9ca-4fd9-9695-87538dcf1f78 tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Lock "406c59e7-8eb8-4cfd-8e1e-e151ab908571" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 44.769s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1003.724017] env[65726]: DEBUG oslo_concurrency.lockutils [None req-69b4acd5-8968-4363-a895-1312ce324cfb tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1003.972573] env[65726]: DEBUG oslo_vmware.api [None req-bd9cdf9a-c493-4e4b-996c-7dba718501da tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': task-5116457, 'name': CloneVM_Task} progress is 94%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.973112] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5d32d102-744f-471d-8592-b9e461d3322c tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Lock "060646cc-f0fb-4e6c-9f67-7b2daa6b6fd6" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 22.688s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1004.464032] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Instance 895797e4-2941-44cd-aab1-67afa4fac02d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65726) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1004.464196] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Instance 6d2b86c2-0a44-433f-8b3f-2e9c7f693baa actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65726) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1004.464313] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Instance 9fc0c9ee-9dc7-4059-b8d2-a5db965e00c2 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65726) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1004.464658] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Instance cf950ec3-9914-4eb9-99db-048b3969bb21 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65726) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1004.464826] env[65726]: WARNING nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Instance 815ee20b-cb07-4514-bd43-1b74972508a5 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 1004.464993] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Instance dcde25df-87a5-47fb-94ce-334f68894e04 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65726) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1004.465099] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Instance 7aa4c0d9-f7ca-42ba-9fd0-4075cd76752b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65726) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1004.465157] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Instance 6fd70b4a-63e4-4258-9cff-f3b582500b9f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65726) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1004.465257] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Instance aa6ce489-c62f-4481-87b7-e74242aeb8ca actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65726) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1004.465483] env[65726]: WARNING nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Instance 54e85161-5f63-405f-83f8-490c70645a3b is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 1004.465483] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Instance 406c59e7-8eb8-4cfd-8e1e-e151ab908571 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65726) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1004.465576] env[65726]: WARNING nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Instance f9a40d36-dda9-454c-8774-ed011f3f50ae is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 1004.465685] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Instance 067c9332-fa2c-47fe-93ae-643de1fe77cf actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65726) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1004.465840] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Instance a56be6fa-6027-46ee-9aa8-332fe4222d68 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65726) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1004.465905] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Instance d8468efa-c2d3-4dce-ab89-fc077011e3d8 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65726) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1004.466035] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] [instance: 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04] Instance with task_state "unshelving" is not being actively managed by this compute host but has allocations referencing this compute node (07c4692f-bdb4-4058-9173-ff9664830295): {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocations during the task state transition. {{(pid=65726) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1813}} [ 1004.473988] env[65726]: DEBUG oslo_vmware.api [None req-bd9cdf9a-c493-4e4b-996c-7dba718501da tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': task-5116457, 'name': CloneVM_Task} progress is 94%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.545383] env[65726]: DEBUG oslo_concurrency.lockutils [None req-59619dfd-192b-4bfb-9fd7-b4adaf4a925f tempest-ServerMetadataTestJSON-637854735 tempest-ServerMetadataTestJSON-637854735-project-member] Acquiring lock "699127a9-f40f-4c1c-ba4a-625af097350b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1004.545680] env[65726]: DEBUG oslo_concurrency.lockutils [None req-59619dfd-192b-4bfb-9fd7-b4adaf4a925f tempest-ServerMetadataTestJSON-637854735 tempest-ServerMetadataTestJSON-637854735-project-member] Lock "699127a9-f40f-4c1c-ba4a-625af097350b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1004.972788] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Instance a286ba4f-a83e-4c30-8079-ba88e3ba5696 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65726) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 1004.974106] env[65726]: DEBUG oslo_vmware.api [None req-bd9cdf9a-c493-4e4b-996c-7dba718501da tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': task-5116457, 'name': CloneVM_Task} progress is 95%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.049113] env[65726]: DEBUG nova.compute.manager [None req-59619dfd-192b-4bfb-9fd7-b4adaf4a925f tempest-ServerMetadataTestJSON-637854735 tempest-ServerMetadataTestJSON-637854735-project-member] [instance: 699127a9-f40f-4c1c-ba4a-625af097350b] Starting instance... {{(pid=65726) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 1005.056159] env[65726]: DEBUG oslo_concurrency.lockutils [None req-4ed5fc19-fb1a-414f-a0bd-edcdbeccf1c1 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Acquiring lock "interface-dcde25df-87a5-47fb-94ce-334f68894e04-b48313f9-a0d3-4fdd-88c6-d103aef0b4f2" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1005.056159] env[65726]: DEBUG oslo_concurrency.lockutils [None req-4ed5fc19-fb1a-414f-a0bd-edcdbeccf1c1 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Lock "interface-dcde25df-87a5-47fb-94ce-334f68894e04-b48313f9-a0d3-4fdd-88c6-d103aef0b4f2" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1005.149923] env[65726]: DEBUG oslo_concurrency.lockutils [None req-d8ec8f57-04a2-4fd5-b8fa-981fbeb154b5 tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Acquiring lock "406c59e7-8eb8-4cfd-8e1e-e151ab908571" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1005.149923] env[65726]: DEBUG oslo_concurrency.lockutils [None req-d8ec8f57-04a2-4fd5-b8fa-981fbeb154b5 tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Lock "406c59e7-8eb8-4cfd-8e1e-e151ab908571" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1005.150152] env[65726]: DEBUG oslo_concurrency.lockutils [None req-d8ec8f57-04a2-4fd5-b8fa-981fbeb154b5 tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Acquiring lock "406c59e7-8eb8-4cfd-8e1e-e151ab908571-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1005.150297] env[65726]: DEBUG oslo_concurrency.lockutils [None req-d8ec8f57-04a2-4fd5-b8fa-981fbeb154b5 tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Lock "406c59e7-8eb8-4cfd-8e1e-e151ab908571-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1005.150464] env[65726]: DEBUG oslo_concurrency.lockutils [None req-d8ec8f57-04a2-4fd5-b8fa-981fbeb154b5 tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Lock "406c59e7-8eb8-4cfd-8e1e-e151ab908571-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1005.155130] env[65726]: INFO nova.compute.manager [None req-d8ec8f57-04a2-4fd5-b8fa-981fbeb154b5 tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] [instance: 406c59e7-8eb8-4cfd-8e1e-e151ab908571] Terminating instance [ 1005.473682] env[65726]: DEBUG oslo_vmware.api [None req-bd9cdf9a-c493-4e4b-996c-7dba718501da tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': task-5116457, 'name': CloneVM_Task, 'duration_secs': 1.668332} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1005.474027] env[65726]: INFO nova.virt.vmwareapi.vmops [None req-bd9cdf9a-c493-4e4b-996c-7dba718501da tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: 067c9332-fa2c-47fe-93ae-643de1fe77cf] Created linked-clone VM from snapshot [ 1005.475259] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a84a9ee-994e-4489-ba80-e6c814332447 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.481270] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Instance 699127a9-f40f-4c1c-ba4a-625af097350b has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65726) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 1005.482111] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Total usable vcpus: 48, total allocated vcpus: 12 {{(pid=65726) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1159}} [ 1005.482111] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2816MB phys_disk=100GB used_disk=12GB total_vcpus=48 used_vcpus=12 pci_stats=[] stats={'failed_builds': '0', 'num_instances': '12', 'num_vm_active': '11', 'num_task_None': '11', 'num_os_type_None': '12', 'num_proj_a090d553766847e58e3231b966c92565': '1', 'io_workload': '0', 'num_proj_a8362295a62f4b51bae719a7ef5a4656': '3', 'num_vm_rescued': '1', 'num_proj_d46ccb00794f458b85da4a93879139ee': '1', 'num_proj_534461fc8c62483d875b8e165ef0d5c0': '1', 'num_proj_06e221dc693640929cac7bc5af45948f': '1', 'num_proj_4ffd45f4a7a041199a4fc7f69f5e7e9b': '1', 'num_proj_6961c16aa6a346679bbde25930b2a5b7': '1', 'num_task_shelving_image_uploading': '1', 'num_proj_cbb17a09b35c4c22ade5c4082b10bda9': '1', 'num_proj_4eece77569624f90bf64e5c51974173f': '1', 'num_proj_db6b64230abc42069f5587ef94c82f5a': '1'} {{(pid=65726) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1168}} [ 1005.490786] env[65726]: DEBUG nova.virt.vmwareapi.images [None req-bd9cdf9a-c493-4e4b-996c-7dba718501da tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: 067c9332-fa2c-47fe-93ae-643de1fe77cf] Uploading image 54e60667-7017-48b6-80e1-6e4cfed9ff74 {{(pid=65726) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 1005.530249] env[65726]: DEBUG oslo_vmware.rw_handles [None req-bd9cdf9a-c493-4e4b-996c-7dba718501da tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1005.530249] env[65726]: value = "vm-995253" [ 1005.530249] env[65726]: _type = "VirtualMachine" [ 1005.530249] env[65726]: }. {{(pid=65726) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1005.530603] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-f0599543-f4e9-4e4d-9114-4a540b20b2b0 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.539887] env[65726]: DEBUG oslo_vmware.rw_handles [None req-bd9cdf9a-c493-4e4b-996c-7dba718501da tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Lease: (returnval){ [ 1005.539887] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52bae809-c879-4031-7742-33fe5a99e2a8" [ 1005.539887] env[65726]: _type = "HttpNfcLease" [ 1005.539887] env[65726]: } obtained for exporting VM: (result){ [ 1005.539887] env[65726]: value = "vm-995253" [ 1005.539887] env[65726]: _type = "VirtualMachine" [ 1005.539887] env[65726]: }. {{(pid=65726) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1005.540199] env[65726]: DEBUG oslo_vmware.api [None req-bd9cdf9a-c493-4e4b-996c-7dba718501da tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Waiting for the lease: (returnval){ [ 1005.540199] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52bae809-c879-4031-7742-33fe5a99e2a8" [ 1005.540199] env[65726]: _type = "HttpNfcLease" [ 1005.540199] env[65726]: } to be ready. {{(pid=65726) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1005.558168] env[65726]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1005.558168] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52bae809-c879-4031-7742-33fe5a99e2a8" [ 1005.558168] env[65726]: _type = "HttpNfcLease" [ 1005.558168] env[65726]: } is initializing. {{(pid=65726) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1005.561895] env[65726]: DEBUG oslo_concurrency.lockutils [None req-4ed5fc19-fb1a-414f-a0bd-edcdbeccf1c1 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Acquiring lock "dcde25df-87a5-47fb-94ce-334f68894e04" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1005.562067] env[65726]: DEBUG oslo_concurrency.lockutils [None req-4ed5fc19-fb1a-414f-a0bd-edcdbeccf1c1 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Acquired lock "dcde25df-87a5-47fb-94ce-334f68894e04" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1005.565537] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed56b393-2f8e-43e2-bb21-fafaba9e83eb {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.589987] env[65726]: DEBUG oslo_concurrency.lockutils [None req-59619dfd-192b-4bfb-9fd7-b4adaf4a925f tempest-ServerMetadataTestJSON-637854735 tempest-ServerMetadataTestJSON-637854735-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1005.591038] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c5ff18f-0255-43ad-801d-af32b61fc320 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.618041] env[65726]: WARNING openstack [None req-4ed5fc19-fb1a-414f-a0bd-edcdbeccf1c1 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1005.618450] env[65726]: WARNING openstack [None req-4ed5fc19-fb1a-414f-a0bd-edcdbeccf1c1 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1005.629025] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-4ed5fc19-fb1a-414f-a0bd-edcdbeccf1c1 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: dcde25df-87a5-47fb-94ce-334f68894e04] Reconfiguring VM to detach interface {{(pid=65726) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1968}} [ 1005.632356] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3b7b0ba6-6ece-4a98-8da8-91732ad4c91c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.654532] env[65726]: DEBUG oslo_vmware.api [None req-4ed5fc19-fb1a-414f-a0bd-edcdbeccf1c1 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Waiting for the task: (returnval){ [ 1005.654532] env[65726]: value = "task-5116459" [ 1005.654532] env[65726]: _type = "Task" [ 1005.654532] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1005.657976] env[65726]: DEBUG nova.compute.manager [req-b46153d2-4a8e-44d0-9e8c-1fb1709a2b09 req-29befb7f-c695-47f7-aad5-aab8a46086b3 service nova] [instance: f9a40d36-dda9-454c-8774-ed011f3f50ae] Received event network-vif-deleted-4a5e689f-4a77-4c32-81bf-4f86a5c702cb {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1005.658305] env[65726]: DEBUG nova.compute.manager [req-b46153d2-4a8e-44d0-9e8c-1fb1709a2b09 req-29befb7f-c695-47f7-aad5-aab8a46086b3 service nova] [instance: 54e85161-5f63-405f-83f8-490c70645a3b] Received event network-vif-deleted-8adb72c3-8e2f-48d7-8197-301c8958dbfb {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1005.668071] env[65726]: DEBUG nova.compute.manager [None req-d8ec8f57-04a2-4fd5-b8fa-981fbeb154b5 tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] [instance: 406c59e7-8eb8-4cfd-8e1e-e151ab908571] Start destroying the instance on the hypervisor. {{(pid=65726) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1005.668372] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-d8ec8f57-04a2-4fd5-b8fa-981fbeb154b5 tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] [instance: 406c59e7-8eb8-4cfd-8e1e-e151ab908571] Destroying instance {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1005.673224] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-220b6495-1595-4eea-87c3-0b5cdd2d0e06 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.684779] env[65726]: DEBUG nova.compute.manager [req-4a253f3e-8adc-4535-90f2-527dd4cdabf4 req-4aab06af-27d1-4012-9122-0b7c9c4717aa service nova] [instance: a56be6fa-6027-46ee-9aa8-332fe4222d68] Received event network-changed-4e2dfbcd-41c3-4aaa-a2ec-50c855c12307 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1005.684836] env[65726]: DEBUG nova.compute.manager [req-4a253f3e-8adc-4535-90f2-527dd4cdabf4 req-4aab06af-27d1-4012-9122-0b7c9c4717aa service nova] [instance: a56be6fa-6027-46ee-9aa8-332fe4222d68] Refreshing instance network info cache due to event network-changed-4e2dfbcd-41c3-4aaa-a2ec-50c855c12307. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 1005.685159] env[65726]: DEBUG oslo_concurrency.lockutils [req-4a253f3e-8adc-4535-90f2-527dd4cdabf4 req-4aab06af-27d1-4012-9122-0b7c9c4717aa service nova] Acquiring lock "refresh_cache-a56be6fa-6027-46ee-9aa8-332fe4222d68" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1005.685305] env[65726]: DEBUG oslo_concurrency.lockutils [req-4a253f3e-8adc-4535-90f2-527dd4cdabf4 req-4aab06af-27d1-4012-9122-0b7c9c4717aa service nova] Acquired lock "refresh_cache-a56be6fa-6027-46ee-9aa8-332fe4222d68" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1005.685718] env[65726]: DEBUG nova.network.neutron [req-4a253f3e-8adc-4535-90f2-527dd4cdabf4 req-4aab06af-27d1-4012-9122-0b7c9c4717aa service nova] [instance: a56be6fa-6027-46ee-9aa8-332fe4222d68] Refreshing network info cache for port 4e2dfbcd-41c3-4aaa-a2ec-50c855c12307 {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 1005.688437] env[65726]: DEBUG oslo_vmware.api [None req-4ed5fc19-fb1a-414f-a0bd-edcdbeccf1c1 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': task-5116459, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.696722] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-d8ec8f57-04a2-4fd5-b8fa-981fbeb154b5 tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] [instance: 406c59e7-8eb8-4cfd-8e1e-e151ab908571] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1005.697526] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-15ee83c6-8350-47f8-afc0-a6d5b0871af1 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.708263] env[65726]: DEBUG oslo_vmware.api [None req-d8ec8f57-04a2-4fd5-b8fa-981fbeb154b5 tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Waiting for the task: (returnval){ [ 1005.708263] env[65726]: value = "task-5116460" [ 1005.708263] env[65726]: _type = "Task" [ 1005.708263] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1005.723972] env[65726]: DEBUG oslo_vmware.api [None req-d8ec8f57-04a2-4fd5-b8fa-981fbeb154b5 tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Task: {'id': task-5116460, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.857664] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb955335-5f90-46c4-a9ce-ea173abe4e4f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.869234] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9764e6b-482d-4b89-9516-0282bc8c8a3b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.912748] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2b80f49-5758-403a-8913-867a6926bd1b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.922731] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c561d83f-5531-405d-95c4-63d911b6465b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.939493] env[65726]: DEBUG nova.compute.provider_tree [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1006.052751] env[65726]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1006.052751] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52bae809-c879-4031-7742-33fe5a99e2a8" [ 1006.052751] env[65726]: _type = "HttpNfcLease" [ 1006.052751] env[65726]: } is ready. {{(pid=65726) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1006.053741] env[65726]: DEBUG oslo_vmware.rw_handles [None req-bd9cdf9a-c493-4e4b-996c-7dba718501da tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1006.053741] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52bae809-c879-4031-7742-33fe5a99e2a8" [ 1006.053741] env[65726]: _type = "HttpNfcLease" [ 1006.053741] env[65726]: }. {{(pid=65726) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1006.055347] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52c933e9-a333-4476-bc30-cb09f53273f5 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.067028] env[65726]: DEBUG oslo_vmware.rw_handles [None req-bd9cdf9a-c493-4e4b-996c-7dba718501da tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52dc880d-b914-d12a-301e-6991c37c09ef/disk-0.vmdk from lease info. {{(pid=65726) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1006.067028] env[65726]: DEBUG oslo_vmware.rw_handles [None req-bd9cdf9a-c493-4e4b-996c-7dba718501da tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52dc880d-b914-d12a-301e-6991c37c09ef/disk-0.vmdk for reading. {{(pid=65726) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1006.168562] env[65726]: DEBUG oslo_vmware.api [None req-4ed5fc19-fb1a-414f-a0bd-edcdbeccf1c1 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': task-5116459, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.193722] env[65726]: WARNING openstack [req-4a253f3e-8adc-4535-90f2-527dd4cdabf4 req-4aab06af-27d1-4012-9122-0b7c9c4717aa service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1006.194119] env[65726]: WARNING openstack [req-4a253f3e-8adc-4535-90f2-527dd4cdabf4 req-4aab06af-27d1-4012-9122-0b7c9c4717aa service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1006.220019] env[65726]: DEBUG oslo_vmware.api [None req-d8ec8f57-04a2-4fd5-b8fa-981fbeb154b5 tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Task: {'id': task-5116460, 'name': PowerOffVM_Task, 'duration_secs': 0.235476} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1006.221289] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-d8ec8f57-04a2-4fd5-b8fa-981fbeb154b5 tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] [instance: 406c59e7-8eb8-4cfd-8e1e-e151ab908571] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1006.221447] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-d8ec8f57-04a2-4fd5-b8fa-981fbeb154b5 tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] [instance: 406c59e7-8eb8-4cfd-8e1e-e151ab908571] Unregistering the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1006.223645] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e391d98a-8890-4c73-abdb-e8be575b3bbd {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.275592] env[65726]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-f4248c33-4344-43c2-b42d-5e85eb28e23c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.367747] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-d8ec8f57-04a2-4fd5-b8fa-981fbeb154b5 tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] [instance: 406c59e7-8eb8-4cfd-8e1e-e151ab908571] Unregistered the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1006.367747] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-d8ec8f57-04a2-4fd5-b8fa-981fbeb154b5 tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] [instance: 406c59e7-8eb8-4cfd-8e1e-e151ab908571] Deleting contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1006.367747] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-d8ec8f57-04a2-4fd5-b8fa-981fbeb154b5 tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Deleting the datastore file [datastore1] 406c59e7-8eb8-4cfd-8e1e-e151ab908571 {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1006.367747] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c92d9050-90a3-4bf9-81b6-480a18afb06c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.378637] env[65726]: DEBUG oslo_vmware.api [None req-d8ec8f57-04a2-4fd5-b8fa-981fbeb154b5 tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Waiting for the task: (returnval){ [ 1006.378637] env[65726]: value = "task-5116462" [ 1006.378637] env[65726]: _type = "Task" [ 1006.378637] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1006.381497] env[65726]: DEBUG oslo_concurrency.lockutils [None req-091e79d3-e236-42d2-ae84-50c5a8fed075 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Acquiring lock "895797e4-2941-44cd-aab1-67afa4fac02d" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1006.381828] env[65726]: DEBUG oslo_concurrency.lockutils [None req-091e79d3-e236-42d2-ae84-50c5a8fed075 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Lock "895797e4-2941-44cd-aab1-67afa4fac02d" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1006.390014] env[65726]: DEBUG oslo_vmware.api [None req-d8ec8f57-04a2-4fd5-b8fa-981fbeb154b5 tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Task: {'id': task-5116462, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.442993] env[65726]: DEBUG nova.scheduler.client.report [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1006.668691] env[65726]: DEBUG oslo_vmware.api [None req-4ed5fc19-fb1a-414f-a0bd-edcdbeccf1c1 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': task-5116459, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.754232] env[65726]: WARNING openstack [req-4a253f3e-8adc-4535-90f2-527dd4cdabf4 req-4aab06af-27d1-4012-9122-0b7c9c4717aa service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1006.755417] env[65726]: WARNING openstack [req-4a253f3e-8adc-4535-90f2-527dd4cdabf4 req-4aab06af-27d1-4012-9122-0b7c9c4717aa service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1006.887199] env[65726]: INFO nova.compute.manager [None req-091e79d3-e236-42d2-ae84-50c5a8fed075 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: 895797e4-2941-44cd-aab1-67afa4fac02d] Detaching volume 3c728bbf-472e-4a76-86f7-6fd547117357 [ 1006.896207] env[65726]: DEBUG oslo_vmware.api [None req-d8ec8f57-04a2-4fd5-b8fa-981fbeb154b5 tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Task: {'id': task-5116462, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.162379} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1006.903961] env[65726]: WARNING openstack [req-4a253f3e-8adc-4535-90f2-527dd4cdabf4 req-4aab06af-27d1-4012-9122-0b7c9c4717aa service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1006.904382] env[65726]: WARNING openstack [req-4a253f3e-8adc-4535-90f2-527dd4cdabf4 req-4aab06af-27d1-4012-9122-0b7c9c4717aa service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1006.914719] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-d8ec8f57-04a2-4fd5-b8fa-981fbeb154b5 tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Deleted the datastore file {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1006.915099] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-d8ec8f57-04a2-4fd5-b8fa-981fbeb154b5 tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] [instance: 406c59e7-8eb8-4cfd-8e1e-e151ab908571] Deleted contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1006.915377] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-d8ec8f57-04a2-4fd5-b8fa-981fbeb154b5 tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] [instance: 406c59e7-8eb8-4cfd-8e1e-e151ab908571] Instance destroyed {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1006.915633] env[65726]: INFO nova.compute.manager [None req-d8ec8f57-04a2-4fd5-b8fa-981fbeb154b5 tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] [instance: 406c59e7-8eb8-4cfd-8e1e-e151ab908571] Took 1.25 seconds to destroy the instance on the hypervisor. [ 1006.915951] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-d8ec8f57-04a2-4fd5-b8fa-981fbeb154b5 tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1006.916721] env[65726]: DEBUG nova.compute.manager [-] [instance: 406c59e7-8eb8-4cfd-8e1e-e151ab908571] Deallocating network for instance {{(pid=65726) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1006.918437] env[65726]: DEBUG nova.network.neutron [-] [instance: 406c59e7-8eb8-4cfd-8e1e-e151ab908571] deallocate_for_instance() {{(pid=65726) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 1006.919056] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1006.919399] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1006.947834] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=65726) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1097}} [ 1006.948122] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 3.528s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1006.948397] env[65726]: DEBUG oslo_concurrency.lockutils [None req-8940e544-9cc2-4a18-9dd0-e7f396520ac6 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 13.433s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1006.948582] env[65726]: DEBUG oslo_concurrency.lockutils [None req-8940e544-9cc2-4a18-9dd0-e7f396520ac6 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1006.955131] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a27b5bca-2338-4542-8dd7-9cac7e4f3b3a tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 8.171s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1006.960019] env[65726]: INFO nova.compute.claims [None req-a27b5bca-2338-4542-8dd7-9cac7e4f3b3a tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] [instance: a286ba4f-a83e-4c30-8079-ba88e3ba5696] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1006.961521] env[65726]: INFO nova.virt.block_device [None req-091e79d3-e236-42d2-ae84-50c5a8fed075 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: 895797e4-2941-44cd-aab1-67afa4fac02d] Attempting to driver detach volume 3c728bbf-472e-4a76-86f7-6fd547117357 from mountpoint /dev/sdb [ 1006.961778] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-091e79d3-e236-42d2-ae84-50c5a8fed075 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: 895797e4-2941-44cd-aab1-67afa4fac02d] Volume detach. Driver type: vmdk {{(pid=65726) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1006.961956] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-091e79d3-e236-42d2-ae84-50c5a8fed075 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: 895797e4-2941-44cd-aab1-67afa4fac02d] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-995185', 'volume_id': '3c728bbf-472e-4a76-86f7-6fd547117357', 'name': 'volume-3c728bbf-472e-4a76-86f7-6fd547117357', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '895797e4-2941-44cd-aab1-67afa4fac02d', 'attached_at': '', 'detached_at': '', 'volume_id': '3c728bbf-472e-4a76-86f7-6fd547117357', 'serial': '3c728bbf-472e-4a76-86f7-6fd547117357'} {{(pid=65726) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1006.966027] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9178e97-c7b7-4d43-bcfa-e702451de9e6 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.993147] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d910ecf4-eaf2-42af-bca7-1201bc570a91 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.009815] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f657111a-75bf-46a4-96b6-a93507ca71c1 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.040021] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d73f24f-0af3-407b-b2ea-7632b3df7cdf {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.048864] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1007.049592] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1007.076352] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-091e79d3-e236-42d2-ae84-50c5a8fed075 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] The volume has not been displaced from its original location: [datastore1] volume-3c728bbf-472e-4a76-86f7-6fd547117357/volume-3c728bbf-472e-4a76-86f7-6fd547117357.vmdk. No consolidation needed. {{(pid=65726) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1007.082401] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-091e79d3-e236-42d2-ae84-50c5a8fed075 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: 895797e4-2941-44cd-aab1-67afa4fac02d] Reconfiguring VM instance instance-00000028 to detach disk 2001 {{(pid=65726) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1007.083244] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-51cad7b4-425f-4d6b-b576-c66ed58c8120 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.104911] env[65726]: DEBUG oslo_vmware.api [None req-091e79d3-e236-42d2-ae84-50c5a8fed075 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Waiting for the task: (returnval){ [ 1007.104911] env[65726]: value = "task-5116463" [ 1007.104911] env[65726]: _type = "Task" [ 1007.104911] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1007.114685] env[65726]: DEBUG oslo_vmware.api [None req-091e79d3-e236-42d2-ae84-50c5a8fed075 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Task: {'id': task-5116463, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.132480] env[65726]: DEBUG nova.network.neutron [req-4a253f3e-8adc-4535-90f2-527dd4cdabf4 req-4aab06af-27d1-4012-9122-0b7c9c4717aa service nova] [instance: a56be6fa-6027-46ee-9aa8-332fe4222d68] Updated VIF entry in instance network info cache for port 4e2dfbcd-41c3-4aaa-a2ec-50c855c12307. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 1007.132480] env[65726]: DEBUG nova.network.neutron [req-4a253f3e-8adc-4535-90f2-527dd4cdabf4 req-4aab06af-27d1-4012-9122-0b7c9c4717aa service nova] [instance: a56be6fa-6027-46ee-9aa8-332fe4222d68] Updating instance_info_cache with network_info: [{"id": "4e2dfbcd-41c3-4aaa-a2ec-50c855c12307", "address": "fa:16:3e:1d:82:59", "network": {"id": "0d0f561d-ac41-4ef5-9926-c707409799f7", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1134321217-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.160", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4eece77569624f90bf64e5c51974173f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2ce62383-8e84-4e26-955b-74c11392f4c9", "external-id": "nsx-vlan-transportzone-215", "segmentation_id": 215, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4e2dfbcd-41", "ovs_interfaceid": "4e2dfbcd-41c3-4aaa-a2ec-50c855c12307", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1007.176051] env[65726]: DEBUG oslo_vmware.api [None req-4ed5fc19-fb1a-414f-a0bd-edcdbeccf1c1 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': task-5116459, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.469094] env[65726]: DEBUG oslo_concurrency.lockutils [None req-8940e544-9cc2-4a18-9dd0-e7f396520ac6 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Lock "2fe6fc5a-f5c7-4f8e-96df-4e621a252f04" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 37.522s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1007.470577] env[65726]: DEBUG oslo_concurrency.lockutils [None req-1d2791c0-c341-4061-b914-188004153713 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Lock "2fe6fc5a-f5c7-4f8e-96df-4e621a252f04" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 10.846s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1007.470837] env[65726]: INFO nova.compute.manager [None req-1d2791c0-c341-4061-b914-188004153713 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04] Unshelving [ 1007.619442] env[65726]: DEBUG oslo_vmware.api [None req-091e79d3-e236-42d2-ae84-50c5a8fed075 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Task: {'id': task-5116463, 'name': ReconfigVM_Task, 'duration_secs': 0.362482} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1007.619746] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-091e79d3-e236-42d2-ae84-50c5a8fed075 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: 895797e4-2941-44cd-aab1-67afa4fac02d] Reconfigured VM instance instance-00000028 to detach disk 2001 {{(pid=65726) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1007.624839] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ad6d2d77-f9b0-407c-9c43-21feb5f44b18 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.644620] env[65726]: DEBUG oslo_concurrency.lockutils [req-4a253f3e-8adc-4535-90f2-527dd4cdabf4 req-4aab06af-27d1-4012-9122-0b7c9c4717aa service nova] Releasing lock "refresh_cache-a56be6fa-6027-46ee-9aa8-332fe4222d68" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1007.653428] env[65726]: DEBUG oslo_vmware.api [None req-091e79d3-e236-42d2-ae84-50c5a8fed075 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Waiting for the task: (returnval){ [ 1007.653428] env[65726]: value = "task-5116464" [ 1007.653428] env[65726]: _type = "Task" [ 1007.653428] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1007.666517] env[65726]: DEBUG oslo_vmware.api [None req-091e79d3-e236-42d2-ae84-50c5a8fed075 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Task: {'id': task-5116464, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.673204] env[65726]: DEBUG oslo_vmware.api [None req-4ed5fc19-fb1a-414f-a0bd-edcdbeccf1c1 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': task-5116459, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.169509] env[65726]: DEBUG oslo_vmware.api [None req-4ed5fc19-fb1a-414f-a0bd-edcdbeccf1c1 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': task-5116459, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.173157] env[65726]: DEBUG oslo_vmware.api [None req-091e79d3-e236-42d2-ae84-50c5a8fed075 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Task: {'id': task-5116464, 'name': ReconfigVM_Task, 'duration_secs': 0.181781} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1008.175989] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-091e79d3-e236-42d2-ae84-50c5a8fed075 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: 895797e4-2941-44cd-aab1-67afa4fac02d] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-995185', 'volume_id': '3c728bbf-472e-4a76-86f7-6fd547117357', 'name': 'volume-3c728bbf-472e-4a76-86f7-6fd547117357', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '895797e4-2941-44cd-aab1-67afa4fac02d', 'attached_at': '', 'detached_at': '', 'volume_id': '3c728bbf-472e-4a76-86f7-6fd547117357', 'serial': '3c728bbf-472e-4a76-86f7-6fd547117357'} {{(pid=65726) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1008.208541] env[65726]: DEBUG nova.compute.manager [req-4ef5d6a8-8b1a-4c43-bed7-1a8c1ef598c6 req-2160e483-7988-4a43-ac84-ae6e54258f32 service nova] [instance: d8468efa-c2d3-4dce-ab89-fc077011e3d8] Received event network-changed-44639a8b-881d-41d3-a114-96a6bccf4064 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1008.208718] env[65726]: DEBUG nova.compute.manager [req-4ef5d6a8-8b1a-4c43-bed7-1a8c1ef598c6 req-2160e483-7988-4a43-ac84-ae6e54258f32 service nova] [instance: d8468efa-c2d3-4dce-ab89-fc077011e3d8] Refreshing instance network info cache due to event network-changed-44639a8b-881d-41d3-a114-96a6bccf4064. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 1008.209070] env[65726]: DEBUG oslo_concurrency.lockutils [req-4ef5d6a8-8b1a-4c43-bed7-1a8c1ef598c6 req-2160e483-7988-4a43-ac84-ae6e54258f32 service nova] Acquiring lock "refresh_cache-d8468efa-c2d3-4dce-ab89-fc077011e3d8" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1008.209202] env[65726]: DEBUG oslo_concurrency.lockutils [req-4ef5d6a8-8b1a-4c43-bed7-1a8c1ef598c6 req-2160e483-7988-4a43-ac84-ae6e54258f32 service nova] Acquired lock "refresh_cache-d8468efa-c2d3-4dce-ab89-fc077011e3d8" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1008.209385] env[65726]: DEBUG nova.network.neutron [req-4ef5d6a8-8b1a-4c43-bed7-1a8c1ef598c6 req-2160e483-7988-4a43-ac84-ae6e54258f32 service nova] [instance: d8468efa-c2d3-4dce-ab89-fc077011e3d8] Refreshing network info cache for port 44639a8b-881d-41d3-a114-96a6bccf4064 {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 1008.323514] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bff576a7-333b-4c88-a5f6-9199735bfcf2 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.333825] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ab36a02-d881-4c94-8dca-43b87a4e8b9c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.373653] env[65726]: DEBUG nova.network.neutron [-] [instance: 406c59e7-8eb8-4cfd-8e1e-e151ab908571] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1008.378282] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b74890f-cef4-4beb-959d-944ccd3d2a2c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.387878] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fcdef500-cb1c-477a-9248-8bb5a5aa4b80 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.402777] env[65726]: DEBUG nova.compute.provider_tree [None req-a27b5bca-2338-4542-8dd7-9cac7e4f3b3a tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1008.492808] env[65726]: DEBUG nova.compute.utils [None req-1d2791c0-c341-4061-b914-188004153713 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Using /dev/sd instead of None {{(pid=65726) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1008.677298] env[65726]: DEBUG oslo_vmware.api [None req-4ed5fc19-fb1a-414f-a0bd-edcdbeccf1c1 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': task-5116459, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.721708] env[65726]: WARNING openstack [req-4ef5d6a8-8b1a-4c43-bed7-1a8c1ef598c6 req-2160e483-7988-4a43-ac84-ae6e54258f32 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1008.721708] env[65726]: WARNING openstack [req-4ef5d6a8-8b1a-4c43-bed7-1a8c1ef598c6 req-2160e483-7988-4a43-ac84-ae6e54258f32 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1008.764176] env[65726]: DEBUG nova.objects.instance [None req-091e79d3-e236-42d2-ae84-50c5a8fed075 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Lazy-loading 'flavor' on Instance uuid 895797e4-2941-44cd-aab1-67afa4fac02d {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1008.879799] env[65726]: INFO nova.compute.manager [-] [instance: 406c59e7-8eb8-4cfd-8e1e-e151ab908571] Took 1.96 seconds to deallocate network for instance. [ 1008.904023] env[65726]: WARNING openstack [req-4ef5d6a8-8b1a-4c43-bed7-1a8c1ef598c6 req-2160e483-7988-4a43-ac84-ae6e54258f32 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1008.904514] env[65726]: WARNING openstack [req-4ef5d6a8-8b1a-4c43-bed7-1a8c1ef598c6 req-2160e483-7988-4a43-ac84-ae6e54258f32 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1008.912436] env[65726]: DEBUG nova.scheduler.client.report [None req-a27b5bca-2338-4542-8dd7-9cac7e4f3b3a tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1008.980735] env[65726]: WARNING openstack [req-4ef5d6a8-8b1a-4c43-bed7-1a8c1ef598c6 req-2160e483-7988-4a43-ac84-ae6e54258f32 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1008.981136] env[65726]: WARNING openstack [req-4ef5d6a8-8b1a-4c43-bed7-1a8c1ef598c6 req-2160e483-7988-4a43-ac84-ae6e54258f32 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1008.997015] env[65726]: INFO nova.virt.block_device [None req-1d2791c0-c341-4061-b914-188004153713 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04] Booting with volume 98eb7918-3cfa-4296-9855-bce47b982be1 at /dev/sdb [ 1009.047776] env[65726]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-33acb7ae-3f40-4e51-bfcd-029f6510212b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.060464] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4fbde920-26de-4e25-9ac6-f550031654cf {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.074180] env[65726]: DEBUG nova.network.neutron [req-4ef5d6a8-8b1a-4c43-bed7-1a8c1ef598c6 req-2160e483-7988-4a43-ac84-ae6e54258f32 service nova] [instance: d8468efa-c2d3-4dce-ab89-fc077011e3d8] Updated VIF entry in instance network info cache for port 44639a8b-881d-41d3-a114-96a6bccf4064. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 1009.074600] env[65726]: DEBUG nova.network.neutron [req-4ef5d6a8-8b1a-4c43-bed7-1a8c1ef598c6 req-2160e483-7988-4a43-ac84-ae6e54258f32 service nova] [instance: d8468efa-c2d3-4dce-ab89-fc077011e3d8] Updating instance_info_cache with network_info: [{"id": "44639a8b-881d-41d3-a114-96a6bccf4064", "address": "fa:16:3e:85:bb:61", "network": {"id": "0bcec3df-912b-4dad-9c7b-4b1b24cce6eb", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationNegativeTestJSON-754860692-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.214", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "db6b64230abc42069f5587ef94c82f5a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "01fe2e08-46f6-4cee-aefd-934461f8077d", "external-id": "nsx-vlan-transportzone-806", "segmentation_id": 806, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap44639a8b-88", "ovs_interfaceid": "44639a8b-881d-41d3-a114-96a6bccf4064", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1009.078870] env[65726]: DEBUG oslo_concurrency.lockutils [None req-12727f4f-3b97-4bba-895f-741fee8ce27a tempest-ServersTestFqdnHostnames-1699296007 tempest-ServersTestFqdnHostnames-1699296007-project-member] Acquiring lock "7aa4c0d9-f7ca-42ba-9fd0-4075cd76752b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1009.078870] env[65726]: DEBUG oslo_concurrency.lockutils [None req-12727f4f-3b97-4bba-895f-741fee8ce27a tempest-ServersTestFqdnHostnames-1699296007 tempest-ServersTestFqdnHostnames-1699296007-project-member] Lock "7aa4c0d9-f7ca-42ba-9fd0-4075cd76752b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1009.078870] env[65726]: DEBUG oslo_concurrency.lockutils [None req-12727f4f-3b97-4bba-895f-741fee8ce27a tempest-ServersTestFqdnHostnames-1699296007 tempest-ServersTestFqdnHostnames-1699296007-project-member] Acquiring lock "7aa4c0d9-f7ca-42ba-9fd0-4075cd76752b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1009.078870] env[65726]: DEBUG oslo_concurrency.lockutils [None req-12727f4f-3b97-4bba-895f-741fee8ce27a tempest-ServersTestFqdnHostnames-1699296007 tempest-ServersTestFqdnHostnames-1699296007-project-member] Lock "7aa4c0d9-f7ca-42ba-9fd0-4075cd76752b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1009.078870] env[65726]: DEBUG oslo_concurrency.lockutils [None req-12727f4f-3b97-4bba-895f-741fee8ce27a tempest-ServersTestFqdnHostnames-1699296007 tempest-ServersTestFqdnHostnames-1699296007-project-member] Lock "7aa4c0d9-f7ca-42ba-9fd0-4075cd76752b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1009.080648] env[65726]: INFO nova.compute.manager [None req-12727f4f-3b97-4bba-895f-741fee8ce27a tempest-ServersTestFqdnHostnames-1699296007 tempest-ServersTestFqdnHostnames-1699296007-project-member] [instance: 7aa4c0d9-f7ca-42ba-9fd0-4075cd76752b] Terminating instance [ 1009.115592] env[65726]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-66801873-d09d-4547-95fe-a3c135c16357 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.125588] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e570c44-d8d2-48ea-be03-70facdc7beab {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.167537] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08594639-dad7-4f6e-a91c-5c24c5199702 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.183067] env[65726]: DEBUG oslo_vmware.api [None req-4ed5fc19-fb1a-414f-a0bd-edcdbeccf1c1 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': task-5116459, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.183685] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc1c1817-db47-4872-9065-7cb16c37ab6e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.201997] env[65726]: DEBUG nova.virt.block_device [None req-1d2791c0-c341-4061-b914-188004153713 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04] Updating existing volume attachment record: f06d8ae1-535c-4f84-9791-3988884e7159 {{(pid=65726) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1009.222882] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5f1cb876-8e7c-4680-9e78-4abda3ac2850 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Acquiring lock "fb8eac12-aedd-426f-9dd3-6c52f9d4c71d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1009.223092] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5f1cb876-8e7c-4680-9e78-4abda3ac2850 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Lock "fb8eac12-aedd-426f-9dd3-6c52f9d4c71d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1009.389136] env[65726]: DEBUG oslo_concurrency.lockutils [None req-d8ec8f57-04a2-4fd5-b8fa-981fbeb154b5 tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1009.418212] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a27b5bca-2338-4542-8dd7-9cac7e4f3b3a tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.463s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1009.418277] env[65726]: DEBUG nova.compute.manager [None req-a27b5bca-2338-4542-8dd7-9cac7e4f3b3a tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] [instance: a286ba4f-a83e-4c30-8079-ba88e3ba5696] Start building networks asynchronously for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1009.421290] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3a484f95-3cdc-4dca-a553-0dec45f0913e tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 6.319s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1009.421515] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3a484f95-3cdc-4dca-a553-0dec45f0913e tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1009.424425] env[65726]: DEBUG oslo_concurrency.lockutils [None req-22e29168-0ae7-43a0-9b85-be2977371fed tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 5.980s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1009.424775] env[65726]: DEBUG oslo_concurrency.lockutils [None req-22e29168-0ae7-43a0-9b85-be2977371fed tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1009.427348] env[65726]: DEBUG oslo_concurrency.lockutils [None req-69b4acd5-8968-4363-a895-1312ce324cfb tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 5.704s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1009.427549] env[65726]: DEBUG oslo_concurrency.lockutils [None req-69b4acd5-8968-4363-a895-1312ce324cfb tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1009.429401] env[65726]: DEBUG oslo_concurrency.lockutils [None req-59619dfd-192b-4bfb-9fd7-b4adaf4a925f tempest-ServerMetadataTestJSON-637854735 tempest-ServerMetadataTestJSON-637854735-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.840s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1009.431186] env[65726]: INFO nova.compute.claims [None req-59619dfd-192b-4bfb-9fd7-b4adaf4a925f tempest-ServerMetadataTestJSON-637854735 tempest-ServerMetadataTestJSON-637854735-project-member] [instance: 699127a9-f40f-4c1c-ba4a-625af097350b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1009.470930] env[65726]: INFO nova.scheduler.client.report [None req-3a484f95-3cdc-4dca-a553-0dec45f0913e tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Deleted allocations for instance 815ee20b-cb07-4514-bd43-1b74972508a5 [ 1009.475575] env[65726]: INFO nova.scheduler.client.report [None req-22e29168-0ae7-43a0-9b85-be2977371fed tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Deleted allocations for instance f9a40d36-dda9-454c-8774-ed011f3f50ae [ 1009.506469] env[65726]: INFO nova.scheduler.client.report [None req-69b4acd5-8968-4363-a895-1312ce324cfb tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Deleted allocations for instance 54e85161-5f63-405f-83f8-490c70645a3b [ 1009.580439] env[65726]: DEBUG oslo_concurrency.lockutils [req-4ef5d6a8-8b1a-4c43-bed7-1a8c1ef598c6 req-2160e483-7988-4a43-ac84-ae6e54258f32 service nova] Releasing lock "refresh_cache-d8468efa-c2d3-4dce-ab89-fc077011e3d8" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1009.581245] env[65726]: DEBUG nova.compute.manager [req-4ef5d6a8-8b1a-4c43-bed7-1a8c1ef598c6 req-2160e483-7988-4a43-ac84-ae6e54258f32 service nova] [instance: 406c59e7-8eb8-4cfd-8e1e-e151ab908571] Received event network-vif-deleted-903e092f-430c-4ecb-9cc6-1f2adbcd949f {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1009.581245] env[65726]: INFO nova.compute.manager [req-4ef5d6a8-8b1a-4c43-bed7-1a8c1ef598c6 req-2160e483-7988-4a43-ac84-ae6e54258f32 service nova] [instance: 406c59e7-8eb8-4cfd-8e1e-e151ab908571] Neutron deleted interface 903e092f-430c-4ecb-9cc6-1f2adbcd949f; detaching it from the instance and deleting it from the info cache [ 1009.581540] env[65726]: DEBUG nova.network.neutron [req-4ef5d6a8-8b1a-4c43-bed7-1a8c1ef598c6 req-2160e483-7988-4a43-ac84-ae6e54258f32 service nova] [instance: 406c59e7-8eb8-4cfd-8e1e-e151ab908571] Updating instance_info_cache with network_info: [{"id": "eb554e53-6a49-4f06-8271-81ad5c918184", "address": "fa:16:3e:63:4d:c3", "network": {"id": "9487fcf4-a539-4618-8f17-26d849a62be3", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1172384567", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.91", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6961c16aa6a346679bbde25930b2a5b7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7adad264-9276-43ef-9b03-07dc27d3f81e", "external-id": "nsx-vlan-transportzone-329", "segmentation_id": 329, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapeb554e53-6a", "ovs_interfaceid": "eb554e53-6a49-4f06-8271-81ad5c918184", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "b0ae20cc-cb0a-450d-96a3-4726b64254db", "address": "fa:16:3e:bd:0d:c9", "network": {"id": "9487fcf4-a539-4618-8f17-26d849a62be3", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1172384567", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.241", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6961c16aa6a346679bbde25930b2a5b7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7adad264-9276-43ef-9b03-07dc27d3f81e", "external-id": "nsx-vlan-transportzone-329", "segmentation_id": 329, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb0ae20cc-cb", "ovs_interfaceid": "b0ae20cc-cb0a-450d-96a3-4726b64254db", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1009.588206] env[65726]: DEBUG nova.compute.manager [None req-12727f4f-3b97-4bba-895f-741fee8ce27a tempest-ServersTestFqdnHostnames-1699296007 tempest-ServersTestFqdnHostnames-1699296007-project-member] [instance: 7aa4c0d9-f7ca-42ba-9fd0-4075cd76752b] Start destroying the instance on the hypervisor. {{(pid=65726) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1009.588433] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-12727f4f-3b97-4bba-895f-741fee8ce27a tempest-ServersTestFqdnHostnames-1699296007 tempest-ServersTestFqdnHostnames-1699296007-project-member] [instance: 7aa4c0d9-f7ca-42ba-9fd0-4075cd76752b] Destroying instance {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1009.589656] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce7946b1-dceb-451f-ae58-2803a46f2ad0 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.600937] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-12727f4f-3b97-4bba-895f-741fee8ce27a tempest-ServersTestFqdnHostnames-1699296007 tempest-ServersTestFqdnHostnames-1699296007-project-member] [instance: 7aa4c0d9-f7ca-42ba-9fd0-4075cd76752b] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1009.601375] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d1f96ffe-7090-4ffa-a48b-fa446951ba12 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.609723] env[65726]: DEBUG oslo_vmware.api [None req-12727f4f-3b97-4bba-895f-741fee8ce27a tempest-ServersTestFqdnHostnames-1699296007 tempest-ServersTestFqdnHostnames-1699296007-project-member] Waiting for the task: (returnval){ [ 1009.609723] env[65726]: value = "task-5116466" [ 1009.609723] env[65726]: _type = "Task" [ 1009.609723] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1009.620377] env[65726]: DEBUG oslo_vmware.api [None req-12727f4f-3b97-4bba-895f-741fee8ce27a tempest-ServersTestFqdnHostnames-1699296007 tempest-ServersTestFqdnHostnames-1699296007-project-member] Task: {'id': task-5116466, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.675597] env[65726]: DEBUG oslo_vmware.api [None req-4ed5fc19-fb1a-414f-a0bd-edcdbeccf1c1 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': task-5116459, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.726856] env[65726]: DEBUG nova.compute.manager [None req-5f1cb876-8e7c-4680-9e78-4abda3ac2850 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: fb8eac12-aedd-426f-9dd3-6c52f9d4c71d] Starting instance... {{(pid=65726) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 1009.779947] env[65726]: DEBUG oslo_concurrency.lockutils [None req-091e79d3-e236-42d2-ae84-50c5a8fed075 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Lock "895797e4-2941-44cd-aab1-67afa4fac02d" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.397s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1009.937727] env[65726]: DEBUG nova.compute.utils [None req-a27b5bca-2338-4542-8dd7-9cac7e4f3b3a tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Using /dev/sd instead of None {{(pid=65726) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1009.942237] env[65726]: DEBUG nova.compute.manager [None req-a27b5bca-2338-4542-8dd7-9cac7e4f3b3a tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] [instance: a286ba4f-a83e-4c30-8079-ba88e3ba5696] Allocating IP information in the background. {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1009.942924] env[65726]: DEBUG nova.network.neutron [None req-a27b5bca-2338-4542-8dd7-9cac7e4f3b3a tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] [instance: a286ba4f-a83e-4c30-8079-ba88e3ba5696] allocate_for_instance() {{(pid=65726) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 1009.942924] env[65726]: WARNING neutronclient.v2_0.client [None req-a27b5bca-2338-4542-8dd7-9cac7e4f3b3a tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1009.943525] env[65726]: WARNING neutronclient.v2_0.client [None req-a27b5bca-2338-4542-8dd7-9cac7e4f3b3a tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1009.943903] env[65726]: WARNING openstack [None req-a27b5bca-2338-4542-8dd7-9cac7e4f3b3a tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1009.944322] env[65726]: WARNING openstack [None req-a27b5bca-2338-4542-8dd7-9cac7e4f3b3a tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1009.984065] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3a484f95-3cdc-4dca-a553-0dec45f0913e tempest-ListImageFiltersTestJSON-1018059865 tempest-ListImageFiltersTestJSON-1018059865-project-member] Lock "815ee20b-cb07-4514-bd43-1b74972508a5" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 11.077s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1009.987587] env[65726]: DEBUG oslo_concurrency.lockutils [None req-22e29168-0ae7-43a0-9b85-be2977371fed tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Lock "f9a40d36-dda9-454c-8774-ed011f3f50ae" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 10.735s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1010.005381] env[65726]: DEBUG nova.policy [None req-a27b5bca-2338-4542-8dd7-9cac7e4f3b3a tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2614400ef47b4d20ae684c79641ec2a6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1d5cb47905144da49aaf07ddb7d263af', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65726) authorize /opt/stack/nova/nova/policy.py:192}} [ 1010.015903] env[65726]: DEBUG oslo_concurrency.lockutils [None req-69b4acd5-8968-4363-a895-1312ce324cfb tempest-ImagesTestJSON-2081200251 tempest-ImagesTestJSON-2081200251-project-member] Lock "54e85161-5f63-405f-83f8-490c70645a3b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 10.397s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1010.090408] env[65726]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-4bac3118-26ad-42e6-b89c-049a8323d21c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.105232] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1bedee6d-1f23-4359-b44a-f2be7b8c01e8 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.128828] env[65726]: DEBUG oslo_vmware.api [None req-12727f4f-3b97-4bba-895f-741fee8ce27a tempest-ServersTestFqdnHostnames-1699296007 tempest-ServersTestFqdnHostnames-1699296007-project-member] Task: {'id': task-5116466, 'name': PowerOffVM_Task, 'duration_secs': 0.285641} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1010.129411] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-12727f4f-3b97-4bba-895f-741fee8ce27a tempest-ServersTestFqdnHostnames-1699296007 tempest-ServersTestFqdnHostnames-1699296007-project-member] [instance: 7aa4c0d9-f7ca-42ba-9fd0-4075cd76752b] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1010.130173] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-12727f4f-3b97-4bba-895f-741fee8ce27a tempest-ServersTestFqdnHostnames-1699296007 tempest-ServersTestFqdnHostnames-1699296007-project-member] [instance: 7aa4c0d9-f7ca-42ba-9fd0-4075cd76752b] Unregistering the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1010.130173] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d810e9a5-d221-4aa0-bbc8-562b0f47f40d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.149788] env[65726]: DEBUG nova.compute.manager [req-4ef5d6a8-8b1a-4c43-bed7-1a8c1ef598c6 req-2160e483-7988-4a43-ac84-ae6e54258f32 service nova] [instance: 406c59e7-8eb8-4cfd-8e1e-e151ab908571] Detach interface failed, port_id=903e092f-430c-4ecb-9cc6-1f2adbcd949f, reason: Instance 406c59e7-8eb8-4cfd-8e1e-e151ab908571 could not be found. {{(pid=65726) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11669}} [ 1010.149788] env[65726]: DEBUG nova.compute.manager [req-4ef5d6a8-8b1a-4c43-bed7-1a8c1ef598c6 req-2160e483-7988-4a43-ac84-ae6e54258f32 service nova] [instance: 406c59e7-8eb8-4cfd-8e1e-e151ab908571] Received event network-vif-deleted-b0ae20cc-cb0a-450d-96a3-4726b64254db {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1010.149931] env[65726]: INFO nova.compute.manager [req-4ef5d6a8-8b1a-4c43-bed7-1a8c1ef598c6 req-2160e483-7988-4a43-ac84-ae6e54258f32 service nova] [instance: 406c59e7-8eb8-4cfd-8e1e-e151ab908571] Neutron deleted interface b0ae20cc-cb0a-450d-96a3-4726b64254db; detaching it from the instance and deleting it from the info cache [ 1010.150195] env[65726]: DEBUG nova.network.neutron [req-4ef5d6a8-8b1a-4c43-bed7-1a8c1ef598c6 req-2160e483-7988-4a43-ac84-ae6e54258f32 service nova] [instance: 406c59e7-8eb8-4cfd-8e1e-e151ab908571] Updating instance_info_cache with network_info: [{"id": "eb554e53-6a49-4f06-8271-81ad5c918184", "address": "fa:16:3e:63:4d:c3", "network": {"id": "9487fcf4-a539-4618-8f17-26d849a62be3", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1172384567", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.91", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6961c16aa6a346679bbde25930b2a5b7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7adad264-9276-43ef-9b03-07dc27d3f81e", "external-id": "nsx-vlan-transportzone-329", "segmentation_id": 329, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapeb554e53-6a", "ovs_interfaceid": "eb554e53-6a49-4f06-8271-81ad5c918184", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1010.179087] env[65726]: DEBUG oslo_vmware.api [None req-4ed5fc19-fb1a-414f-a0bd-edcdbeccf1c1 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': task-5116459, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.213094] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-12727f4f-3b97-4bba-895f-741fee8ce27a tempest-ServersTestFqdnHostnames-1699296007 tempest-ServersTestFqdnHostnames-1699296007-project-member] [instance: 7aa4c0d9-f7ca-42ba-9fd0-4075cd76752b] Unregistered the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1010.213331] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-12727f4f-3b97-4bba-895f-741fee8ce27a tempest-ServersTestFqdnHostnames-1699296007 tempest-ServersTestFqdnHostnames-1699296007-project-member] [instance: 7aa4c0d9-f7ca-42ba-9fd0-4075cd76752b] Deleting contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1010.213573] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-12727f4f-3b97-4bba-895f-741fee8ce27a tempest-ServersTestFqdnHostnames-1699296007 tempest-ServersTestFqdnHostnames-1699296007-project-member] Deleting the datastore file [datastore1] 7aa4c0d9-f7ca-42ba-9fd0-4075cd76752b {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1010.213874] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1fbecb76-1930-4d3f-bd36-425c97335134 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.222137] env[65726]: DEBUG oslo_vmware.api [None req-12727f4f-3b97-4bba-895f-741fee8ce27a tempest-ServersTestFqdnHostnames-1699296007 tempest-ServersTestFqdnHostnames-1699296007-project-member] Waiting for the task: (returnval){ [ 1010.222137] env[65726]: value = "task-5116470" [ 1010.222137] env[65726]: _type = "Task" [ 1010.222137] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1010.237283] env[65726]: DEBUG oslo_vmware.api [None req-12727f4f-3b97-4bba-895f-741fee8ce27a tempest-ServersTestFqdnHostnames-1699296007 tempest-ServersTestFqdnHostnames-1699296007-project-member] Task: {'id': task-5116470, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.259212] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5f1cb876-8e7c-4680-9e78-4abda3ac2850 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1010.381161] env[65726]: DEBUG nova.network.neutron [None req-a27b5bca-2338-4542-8dd7-9cac7e4f3b3a tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] [instance: a286ba4f-a83e-4c30-8079-ba88e3ba5696] Successfully created port: bae2338e-4be5-4ac9-b3b7-0f31ce3e0b37 {{(pid=65726) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1010.426223] env[65726]: DEBUG nova.compute.manager [req-cb947eba-a8a1-4290-995e-ddce430f1644 req-a8968651-90de-490e-b08c-7f679a48135b service nova] [instance: 406c59e7-8eb8-4cfd-8e1e-e151ab908571] Received event network-vif-deleted-eb554e53-6a49-4f06-8271-81ad5c918184 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1010.426429] env[65726]: INFO nova.compute.manager [req-cb947eba-a8a1-4290-995e-ddce430f1644 req-a8968651-90de-490e-b08c-7f679a48135b service nova] [instance: 406c59e7-8eb8-4cfd-8e1e-e151ab908571] Neutron deleted interface eb554e53-6a49-4f06-8271-81ad5c918184; detaching it from the instance and deleting it from the info cache [ 1010.426595] env[65726]: DEBUG nova.network.neutron [req-cb947eba-a8a1-4290-995e-ddce430f1644 req-a8968651-90de-490e-b08c-7f679a48135b service nova] [instance: 406c59e7-8eb8-4cfd-8e1e-e151ab908571] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1010.442719] env[65726]: DEBUG nova.compute.manager [None req-a27b5bca-2338-4542-8dd7-9cac7e4f3b3a tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] [instance: a286ba4f-a83e-4c30-8079-ba88e3ba5696] Start building block device mappings for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1010.655048] env[65726]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-42047d1c-473b-4858-927e-548e348f9354 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.671660] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f672e98f-6064-4dcb-ab2c-8e045f47d23d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.694860] env[65726]: DEBUG oslo_vmware.api [None req-4ed5fc19-fb1a-414f-a0bd-edcdbeccf1c1 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': task-5116459, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.715741] env[65726]: DEBUG nova.compute.manager [req-4ef5d6a8-8b1a-4c43-bed7-1a8c1ef598c6 req-2160e483-7988-4a43-ac84-ae6e54258f32 service nova] [instance: 406c59e7-8eb8-4cfd-8e1e-e151ab908571] Detach interface failed, port_id=b0ae20cc-cb0a-450d-96a3-4726b64254db, reason: Instance 406c59e7-8eb8-4cfd-8e1e-e151ab908571 could not be found. {{(pid=65726) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11669}} [ 1010.720359] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a137ea89-0570-4b3c-9ec4-419c7b4ba995 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.734803] env[65726]: DEBUG oslo_vmware.api [None req-12727f4f-3b97-4bba-895f-741fee8ce27a tempest-ServersTestFqdnHostnames-1699296007 tempest-ServersTestFqdnHostnames-1699296007-project-member] Task: {'id': task-5116470, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.171106} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1010.735839] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f38acbc-6064-4fa7-a065-6a550ec83052 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.739340] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-12727f4f-3b97-4bba-895f-741fee8ce27a tempest-ServersTestFqdnHostnames-1699296007 tempest-ServersTestFqdnHostnames-1699296007-project-member] Deleted the datastore file {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1010.739594] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-12727f4f-3b97-4bba-895f-741fee8ce27a tempest-ServersTestFqdnHostnames-1699296007 tempest-ServersTestFqdnHostnames-1699296007-project-member] [instance: 7aa4c0d9-f7ca-42ba-9fd0-4075cd76752b] Deleted contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1010.739844] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-12727f4f-3b97-4bba-895f-741fee8ce27a tempest-ServersTestFqdnHostnames-1699296007 tempest-ServersTestFqdnHostnames-1699296007-project-member] [instance: 7aa4c0d9-f7ca-42ba-9fd0-4075cd76752b] Instance destroyed {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1010.740061] env[65726]: INFO nova.compute.manager [None req-12727f4f-3b97-4bba-895f-741fee8ce27a tempest-ServersTestFqdnHostnames-1699296007 tempest-ServersTestFqdnHostnames-1699296007-project-member] [instance: 7aa4c0d9-f7ca-42ba-9fd0-4075cd76752b] Took 1.15 seconds to destroy the instance on the hypervisor. [ 1010.740352] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-12727f4f-3b97-4bba-895f-741fee8ce27a tempest-ServersTestFqdnHostnames-1699296007 tempest-ServersTestFqdnHostnames-1699296007-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1010.740599] env[65726]: DEBUG nova.compute.manager [-] [instance: 7aa4c0d9-f7ca-42ba-9fd0-4075cd76752b] Deallocating network for instance {{(pid=65726) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1010.742815] env[65726]: DEBUG nova.network.neutron [-] [instance: 7aa4c0d9-f7ca-42ba-9fd0-4075cd76752b] deallocate_for_instance() {{(pid=65726) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 1010.742815] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1010.742815] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1010.782495] env[65726]: DEBUG oslo_concurrency.lockutils [None req-db012f3d-acd5-42b8-b506-baca1980cf7e tempest-FloatingIPsAssociationNegativeTestJSON-737087291 tempest-FloatingIPsAssociationNegativeTestJSON-737087291-project-member] Acquiring lock "d8468efa-c2d3-4dce-ab89-fc077011e3d8" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1010.782913] env[65726]: DEBUG oslo_concurrency.lockutils [None req-db012f3d-acd5-42b8-b506-baca1980cf7e tempest-FloatingIPsAssociationNegativeTestJSON-737087291 tempest-FloatingIPsAssociationNegativeTestJSON-737087291-project-member] Lock "d8468efa-c2d3-4dce-ab89-fc077011e3d8" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1010.783047] env[65726]: DEBUG oslo_concurrency.lockutils [None req-db012f3d-acd5-42b8-b506-baca1980cf7e tempest-FloatingIPsAssociationNegativeTestJSON-737087291 tempest-FloatingIPsAssociationNegativeTestJSON-737087291-project-member] Acquiring lock "d8468efa-c2d3-4dce-ab89-fc077011e3d8-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1010.783921] env[65726]: DEBUG oslo_concurrency.lockutils [None req-db012f3d-acd5-42b8-b506-baca1980cf7e tempest-FloatingIPsAssociationNegativeTestJSON-737087291 tempest-FloatingIPsAssociationNegativeTestJSON-737087291-project-member] Lock "d8468efa-c2d3-4dce-ab89-fc077011e3d8-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1010.783921] env[65726]: DEBUG oslo_concurrency.lockutils [None req-db012f3d-acd5-42b8-b506-baca1980cf7e tempest-FloatingIPsAssociationNegativeTestJSON-737087291 tempest-FloatingIPsAssociationNegativeTestJSON-737087291-project-member] Lock "d8468efa-c2d3-4dce-ab89-fc077011e3d8-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1010.785515] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a188cb6-e8fb-473a-a749-39f224ce4ddc {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.789046] env[65726]: INFO nova.compute.manager [None req-db012f3d-acd5-42b8-b506-baca1980cf7e tempest-FloatingIPsAssociationNegativeTestJSON-737087291 tempest-FloatingIPsAssociationNegativeTestJSON-737087291-project-member] [instance: d8468efa-c2d3-4dce-ab89-fc077011e3d8] Terminating instance [ 1010.798124] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffa44290-585b-4c71-b542-7f1cbec119d5 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.818011] env[65726]: DEBUG nova.compute.provider_tree [None req-59619dfd-192b-4bfb-9fd7-b4adaf4a925f tempest-ServerMetadataTestJSON-637854735 tempest-ServerMetadataTestJSON-637854735-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1010.824666] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1010.824936] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1010.837782] env[65726]: DEBUG nova.scheduler.client.report [None req-59619dfd-192b-4bfb-9fd7-b4adaf4a925f tempest-ServerMetadataTestJSON-637854735 tempest-ServerMetadataTestJSON-637854735-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1010.932118] env[65726]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-4615b465-065b-4682-b7a5-11008b4b6e63 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.956367] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08fdba1a-9ca9-4f72-9a51-d1f0c18d3d8d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.995235] env[65726]: DEBUG nova.compute.manager [req-cb947eba-a8a1-4290-995e-ddce430f1644 req-a8968651-90de-490e-b08c-7f679a48135b service nova] [instance: 406c59e7-8eb8-4cfd-8e1e-e151ab908571] Detach interface failed, port_id=eb554e53-6a49-4f06-8271-81ad5c918184, reason: Instance 406c59e7-8eb8-4cfd-8e1e-e151ab908571 could not be found. {{(pid=65726) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11669}} [ 1010.995235] env[65726]: DEBUG nova.compute.manager [req-cb947eba-a8a1-4290-995e-ddce430f1644 req-a8968651-90de-490e-b08c-7f679a48135b service nova] [instance: d8468efa-c2d3-4dce-ab89-fc077011e3d8] Received event network-changed-44639a8b-881d-41d3-a114-96a6bccf4064 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1010.995235] env[65726]: DEBUG nova.compute.manager [req-cb947eba-a8a1-4290-995e-ddce430f1644 req-a8968651-90de-490e-b08c-7f679a48135b service nova] [instance: d8468efa-c2d3-4dce-ab89-fc077011e3d8] Refreshing instance network info cache due to event network-changed-44639a8b-881d-41d3-a114-96a6bccf4064. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 1010.995452] env[65726]: DEBUG oslo_concurrency.lockutils [req-cb947eba-a8a1-4290-995e-ddce430f1644 req-a8968651-90de-490e-b08c-7f679a48135b service nova] Acquiring lock "refresh_cache-d8468efa-c2d3-4dce-ab89-fc077011e3d8" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1010.995496] env[65726]: DEBUG oslo_concurrency.lockutils [req-cb947eba-a8a1-4290-995e-ddce430f1644 req-a8968651-90de-490e-b08c-7f679a48135b service nova] Acquired lock "refresh_cache-d8468efa-c2d3-4dce-ab89-fc077011e3d8" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1010.995629] env[65726]: DEBUG nova.network.neutron [req-cb947eba-a8a1-4290-995e-ddce430f1644 req-a8968651-90de-490e-b08c-7f679a48135b service nova] [instance: d8468efa-c2d3-4dce-ab89-fc077011e3d8] Refreshing network info cache for port 44639a8b-881d-41d3-a114-96a6bccf4064 {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 1011.176485] env[65726]: DEBUG oslo_vmware.api [None req-4ed5fc19-fb1a-414f-a0bd-edcdbeccf1c1 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': task-5116459, 'name': ReconfigVM_Task} progress is 18%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.293808] env[65726]: DEBUG nova.compute.manager [None req-db012f3d-acd5-42b8-b506-baca1980cf7e tempest-FloatingIPsAssociationNegativeTestJSON-737087291 tempest-FloatingIPsAssociationNegativeTestJSON-737087291-project-member] [instance: d8468efa-c2d3-4dce-ab89-fc077011e3d8] Start destroying the instance on the hypervisor. {{(pid=65726) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1011.294070] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-db012f3d-acd5-42b8-b506-baca1980cf7e tempest-FloatingIPsAssociationNegativeTestJSON-737087291 tempest-FloatingIPsAssociationNegativeTestJSON-737087291-project-member] [instance: d8468efa-c2d3-4dce-ab89-fc077011e3d8] Destroying instance {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1011.295077] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cdf27062-06cc-4b5f-86a5-2750fa7067ce {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.305693] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-db012f3d-acd5-42b8-b506-baca1980cf7e tempest-FloatingIPsAssociationNegativeTestJSON-737087291 tempest-FloatingIPsAssociationNegativeTestJSON-737087291-project-member] [instance: d8468efa-c2d3-4dce-ab89-fc077011e3d8] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1011.305957] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-be9961f6-ff98-4537-8a0e-97d56bf1ddb5 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.315614] env[65726]: DEBUG oslo_vmware.api [None req-db012f3d-acd5-42b8-b506-baca1980cf7e tempest-FloatingIPsAssociationNegativeTestJSON-737087291 tempest-FloatingIPsAssociationNegativeTestJSON-737087291-project-member] Waiting for the task: (returnval){ [ 1011.315614] env[65726]: value = "task-5116471" [ 1011.315614] env[65726]: _type = "Task" [ 1011.315614] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1011.330322] env[65726]: DEBUG oslo_vmware.api [None req-db012f3d-acd5-42b8-b506-baca1980cf7e tempest-FloatingIPsAssociationNegativeTestJSON-737087291 tempest-FloatingIPsAssociationNegativeTestJSON-737087291-project-member] Task: {'id': task-5116471, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.344330] env[65726]: DEBUG oslo_concurrency.lockutils [None req-59619dfd-192b-4bfb-9fd7-b4adaf4a925f tempest-ServerMetadataTestJSON-637854735 tempest-ServerMetadataTestJSON-637854735-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 1.915s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1011.345414] env[65726]: DEBUG nova.compute.manager [None req-59619dfd-192b-4bfb-9fd7-b4adaf4a925f tempest-ServerMetadataTestJSON-637854735 tempest-ServerMetadataTestJSON-637854735-project-member] [instance: 699127a9-f40f-4c1c-ba4a-625af097350b] Start building networks asynchronously for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1011.349334] env[65726]: DEBUG oslo_concurrency.lockutils [None req-d8ec8f57-04a2-4fd5-b8fa-981fbeb154b5 tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.960s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1011.349605] env[65726]: DEBUG nova.objects.instance [None req-d8ec8f57-04a2-4fd5-b8fa-981fbeb154b5 tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Lazy-loading 'resources' on Instance uuid 406c59e7-8eb8-4cfd-8e1e-e151ab908571 {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1011.469614] env[65726]: DEBUG nova.compute.manager [None req-a27b5bca-2338-4542-8dd7-9cac7e4f3b3a tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] [instance: a286ba4f-a83e-4c30-8079-ba88e3ba5696] Start spawning the instance on the hypervisor. {{(pid=65726) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1011.502947] env[65726]: DEBUG nova.virt.hardware [None req-a27b5bca-2338-4542-8dd7-9cac7e4f3b3a tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T19:28:34Z,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1539f06fa3534e90acbc3a60c4b8bd3f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T19:28:35Z,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1011.503110] env[65726]: DEBUG nova.virt.hardware [None req-a27b5bca-2338-4542-8dd7-9cac7e4f3b3a tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1011.503220] env[65726]: DEBUG nova.virt.hardware [None req-a27b5bca-2338-4542-8dd7-9cac7e4f3b3a tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1011.503372] env[65726]: DEBUG nova.virt.hardware [None req-a27b5bca-2338-4542-8dd7-9cac7e4f3b3a tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1011.503535] env[65726]: DEBUG nova.virt.hardware [None req-a27b5bca-2338-4542-8dd7-9cac7e4f3b3a tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1011.503683] env[65726]: DEBUG nova.virt.hardware [None req-a27b5bca-2338-4542-8dd7-9cac7e4f3b3a tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1011.503929] env[65726]: DEBUG nova.virt.hardware [None req-a27b5bca-2338-4542-8dd7-9cac7e4f3b3a tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1011.504146] env[65726]: DEBUG nova.virt.hardware [None req-a27b5bca-2338-4542-8dd7-9cac7e4f3b3a tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1011.504503] env[65726]: DEBUG nova.virt.hardware [None req-a27b5bca-2338-4542-8dd7-9cac7e4f3b3a tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1011.504503] env[65726]: DEBUG nova.virt.hardware [None req-a27b5bca-2338-4542-8dd7-9cac7e4f3b3a tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1011.504672] env[65726]: DEBUG nova.virt.hardware [None req-a27b5bca-2338-4542-8dd7-9cac7e4f3b3a tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1011.505592] env[65726]: WARNING openstack [req-cb947eba-a8a1-4290-995e-ddce430f1644 req-a8968651-90de-490e-b08c-7f679a48135b service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1011.505955] env[65726]: WARNING openstack [req-cb947eba-a8a1-4290-995e-ddce430f1644 req-a8968651-90de-490e-b08c-7f679a48135b service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1011.515096] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d27ada2f-b100-4947-bb10-3aec9d932852 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.527847] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14902344-8cd0-4292-bd7e-e8baeabc0e65 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.575154] env[65726]: DEBUG oslo_concurrency.lockutils [None req-cb64949f-89ac-453e-b54b-6d8bfee8a2ad tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Acquiring lock "cf950ec3-9914-4eb9-99db-048b3969bb21" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1011.575420] env[65726]: DEBUG oslo_concurrency.lockutils [None req-cb64949f-89ac-453e-b54b-6d8bfee8a2ad tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Lock "cf950ec3-9914-4eb9-99db-048b3969bb21" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1011.576978] env[65726]: DEBUG oslo_concurrency.lockutils [None req-12289131-1b6e-4176-844f-57bf1f466d9e tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Acquiring lock "895797e4-2941-44cd-aab1-67afa4fac02d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1011.577119] env[65726]: DEBUG oslo_concurrency.lockutils [None req-12289131-1b6e-4176-844f-57bf1f466d9e tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Lock "895797e4-2941-44cd-aab1-67afa4fac02d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1011.577312] env[65726]: DEBUG oslo_concurrency.lockutils [None req-12289131-1b6e-4176-844f-57bf1f466d9e tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Acquiring lock "895797e4-2941-44cd-aab1-67afa4fac02d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1011.577497] env[65726]: DEBUG oslo_concurrency.lockutils [None req-12289131-1b6e-4176-844f-57bf1f466d9e tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Lock "895797e4-2941-44cd-aab1-67afa4fac02d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1011.577660] env[65726]: DEBUG oslo_concurrency.lockutils [None req-12289131-1b6e-4176-844f-57bf1f466d9e tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Lock "895797e4-2941-44cd-aab1-67afa4fac02d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1011.579787] env[65726]: INFO nova.compute.manager [None req-12289131-1b6e-4176-844f-57bf1f466d9e tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: 895797e4-2941-44cd-aab1-67afa4fac02d] Terminating instance [ 1011.677527] env[65726]: DEBUG oslo_vmware.api [None req-4ed5fc19-fb1a-414f-a0bd-edcdbeccf1c1 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': task-5116459, 'name': ReconfigVM_Task, 'duration_secs': 5.822378} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1011.677724] env[65726]: DEBUG oslo_concurrency.lockutils [None req-4ed5fc19-fb1a-414f-a0bd-edcdbeccf1c1 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Releasing lock "dcde25df-87a5-47fb-94ce-334f68894e04" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1011.677923] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-4ed5fc19-fb1a-414f-a0bd-edcdbeccf1c1 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: dcde25df-87a5-47fb-94ce-334f68894e04] Reconfigured VM to detach interface {{(pid=65726) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1978}} [ 1011.678668] env[65726]: WARNING openstack [None req-4ed5fc19-fb1a-414f-a0bd-edcdbeccf1c1 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1011.679063] env[65726]: WARNING openstack [None req-4ed5fc19-fb1a-414f-a0bd-edcdbeccf1c1 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1011.827032] env[65726]: DEBUG oslo_vmware.api [None req-db012f3d-acd5-42b8-b506-baca1980cf7e tempest-FloatingIPsAssociationNegativeTestJSON-737087291 tempest-FloatingIPsAssociationNegativeTestJSON-737087291-project-member] Task: {'id': task-5116471, 'name': PowerOffVM_Task, 'duration_secs': 0.219853} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1011.835554] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-db012f3d-acd5-42b8-b506-baca1980cf7e tempest-FloatingIPsAssociationNegativeTestJSON-737087291 tempest-FloatingIPsAssociationNegativeTestJSON-737087291-project-member] [instance: d8468efa-c2d3-4dce-ab89-fc077011e3d8] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1011.835554] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-db012f3d-acd5-42b8-b506-baca1980cf7e tempest-FloatingIPsAssociationNegativeTestJSON-737087291 tempest-FloatingIPsAssociationNegativeTestJSON-737087291-project-member] [instance: d8468efa-c2d3-4dce-ab89-fc077011e3d8] Unregistering the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1011.835554] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c63fbe34-1acd-4b1b-abf4-c51e1929664e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.854279] env[65726]: DEBUG nova.compute.utils [None req-59619dfd-192b-4bfb-9fd7-b4adaf4a925f tempest-ServerMetadataTestJSON-637854735 tempest-ServerMetadataTestJSON-637854735-project-member] Using /dev/sd instead of None {{(pid=65726) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1011.861211] env[65726]: DEBUG nova.compute.manager [None req-59619dfd-192b-4bfb-9fd7-b4adaf4a925f tempest-ServerMetadataTestJSON-637854735 tempest-ServerMetadataTestJSON-637854735-project-member] [instance: 699127a9-f40f-4c1c-ba4a-625af097350b] Allocating IP information in the background. {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1011.861211] env[65726]: DEBUG nova.network.neutron [None req-59619dfd-192b-4bfb-9fd7-b4adaf4a925f tempest-ServerMetadataTestJSON-637854735 tempest-ServerMetadataTestJSON-637854735-project-member] [instance: 699127a9-f40f-4c1c-ba4a-625af097350b] allocate_for_instance() {{(pid=65726) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 1011.861211] env[65726]: WARNING neutronclient.v2_0.client [None req-59619dfd-192b-4bfb-9fd7-b4adaf4a925f tempest-ServerMetadataTestJSON-637854735 tempest-ServerMetadataTestJSON-637854735-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1011.861211] env[65726]: WARNING neutronclient.v2_0.client [None req-59619dfd-192b-4bfb-9fd7-b4adaf4a925f tempest-ServerMetadataTestJSON-637854735 tempest-ServerMetadataTestJSON-637854735-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1011.862143] env[65726]: WARNING openstack [None req-59619dfd-192b-4bfb-9fd7-b4adaf4a925f tempest-ServerMetadataTestJSON-637854735 tempest-ServerMetadataTestJSON-637854735-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1011.862973] env[65726]: WARNING openstack [None req-59619dfd-192b-4bfb-9fd7-b4adaf4a925f tempest-ServerMetadataTestJSON-637854735 tempest-ServerMetadataTestJSON-637854735-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1011.907615] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-db012f3d-acd5-42b8-b506-baca1980cf7e tempest-FloatingIPsAssociationNegativeTestJSON-737087291 tempest-FloatingIPsAssociationNegativeTestJSON-737087291-project-member] [instance: d8468efa-c2d3-4dce-ab89-fc077011e3d8] Unregistered the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1011.908023] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-db012f3d-acd5-42b8-b506-baca1980cf7e tempest-FloatingIPsAssociationNegativeTestJSON-737087291 tempest-FloatingIPsAssociationNegativeTestJSON-737087291-project-member] [instance: d8468efa-c2d3-4dce-ab89-fc077011e3d8] Deleting contents of the VM from datastore datastore2 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1011.908980] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-db012f3d-acd5-42b8-b506-baca1980cf7e tempest-FloatingIPsAssociationNegativeTestJSON-737087291 tempest-FloatingIPsAssociationNegativeTestJSON-737087291-project-member] Deleting the datastore file [datastore2] d8468efa-c2d3-4dce-ab89-fc077011e3d8 {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1011.909452] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c91ca1c3-dc15-46fb-b44b-c6898d59f494 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.918159] env[65726]: DEBUG oslo_vmware.api [None req-db012f3d-acd5-42b8-b506-baca1980cf7e tempest-FloatingIPsAssociationNegativeTestJSON-737087291 tempest-FloatingIPsAssociationNegativeTestJSON-737087291-project-member] Waiting for the task: (returnval){ [ 1011.918159] env[65726]: value = "task-5116474" [ 1011.918159] env[65726]: _type = "Task" [ 1011.918159] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1011.931122] env[65726]: DEBUG oslo_vmware.api [None req-db012f3d-acd5-42b8-b506-baca1980cf7e tempest-FloatingIPsAssociationNegativeTestJSON-737087291 tempest-FloatingIPsAssociationNegativeTestJSON-737087291-project-member] Task: {'id': task-5116474, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.003220] env[65726]: DEBUG nova.network.neutron [None req-a27b5bca-2338-4542-8dd7-9cac7e4f3b3a tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] [instance: a286ba4f-a83e-4c30-8079-ba88e3ba5696] Successfully updated port: bae2338e-4be5-4ac9-b3b7-0f31ce3e0b37 {{(pid=65726) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 1012.023330] env[65726]: DEBUG nova.network.neutron [-] [instance: 7aa4c0d9-f7ca-42ba-9fd0-4075cd76752b] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1012.082645] env[65726]: DEBUG nova.policy [None req-59619dfd-192b-4bfb-9fd7-b4adaf4a925f tempest-ServerMetadataTestJSON-637854735 tempest-ServerMetadataTestJSON-637854735-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '119225ecb49f4932ae8364391cb01a08', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7befc792f09a46749a3f917ea0008953', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65726) authorize /opt/stack/nova/nova/policy.py:192}} [ 1012.086532] env[65726]: INFO nova.compute.manager [None req-cb64949f-89ac-453e-b54b-6d8bfee8a2ad tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: cf950ec3-9914-4eb9-99db-048b3969bb21] Detaching volume dbb0735d-409f-439d-a32a-4de84d10f7a9 [ 1012.089815] env[65726]: DEBUG nova.compute.manager [None req-12289131-1b6e-4176-844f-57bf1f466d9e tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: 895797e4-2941-44cd-aab1-67afa4fac02d] Start destroying the instance on the hypervisor. {{(pid=65726) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1012.090233] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-12289131-1b6e-4176-844f-57bf1f466d9e tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: 895797e4-2941-44cd-aab1-67afa4fac02d] Destroying instance {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1012.094269] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-812ddc76-b319-4ba8-ba65-e32293f2a841 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.107392] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-12289131-1b6e-4176-844f-57bf1f466d9e tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: 895797e4-2941-44cd-aab1-67afa4fac02d] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1012.107392] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a360f933-3bb6-40ee-a2a9-afccb76d3cf8 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.118967] env[65726]: DEBUG oslo_vmware.api [None req-12289131-1b6e-4176-844f-57bf1f466d9e tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Waiting for the task: (returnval){ [ 1012.118967] env[65726]: value = "task-5116475" [ 1012.118967] env[65726]: _type = "Task" [ 1012.118967] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1012.133919] env[65726]: DEBUG oslo_vmware.api [None req-12289131-1b6e-4176-844f-57bf1f466d9e tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Task: {'id': task-5116475, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.151052] env[65726]: INFO nova.virt.block_device [None req-cb64949f-89ac-453e-b54b-6d8bfee8a2ad tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: cf950ec3-9914-4eb9-99db-048b3969bb21] Attempting to driver detach volume dbb0735d-409f-439d-a32a-4de84d10f7a9 from mountpoint /dev/sdb [ 1012.151052] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-cb64949f-89ac-453e-b54b-6d8bfee8a2ad tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: cf950ec3-9914-4eb9-99db-048b3969bb21] Volume detach. Driver type: vmdk {{(pid=65726) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1012.151166] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-cb64949f-89ac-453e-b54b-6d8bfee8a2ad tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: cf950ec3-9914-4eb9-99db-048b3969bb21] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-995220', 'volume_id': 'dbb0735d-409f-439d-a32a-4de84d10f7a9', 'name': 'volume-dbb0735d-409f-439d-a32a-4de84d10f7a9', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'cf950ec3-9914-4eb9-99db-048b3969bb21', 'attached_at': '', 'detached_at': '', 'volume_id': 'dbb0735d-409f-439d-a32a-4de84d10f7a9', 'serial': 'dbb0735d-409f-439d-a32a-4de84d10f7a9'} {{(pid=65726) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1012.152276] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99d85800-ca2e-4d6a-bb01-b74481457bbf {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.190811] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c91902a3-9d41-4f26-a380-2ca42f6bcf59 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.199801] env[65726]: WARNING openstack [req-cb947eba-a8a1-4290-995e-ddce430f1644 req-a8968651-90de-490e-b08c-7f679a48135b service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1012.200382] env[65726]: WARNING openstack [req-cb947eba-a8a1-4290-995e-ddce430f1644 req-a8968651-90de-490e-b08c-7f679a48135b service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1012.214716] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef68c77c-db09-4e9b-a374-3f337540d68b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.242951] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ae554c6-ff6b-4595-8c97-5599209229c4 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.260846] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-cb64949f-89ac-453e-b54b-6d8bfee8a2ad tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] The volume has not been displaced from its original location: [datastore1] volume-dbb0735d-409f-439d-a32a-4de84d10f7a9/volume-dbb0735d-409f-439d-a32a-4de84d10f7a9.vmdk. No consolidation needed. {{(pid=65726) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1012.266900] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-cb64949f-89ac-453e-b54b-6d8bfee8a2ad tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: cf950ec3-9914-4eb9-99db-048b3969bb21] Reconfiguring VM instance instance-0000003e to detach disk 2001 {{(pid=65726) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1012.271691] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-daaa8054-0954-4e36-bd8a-bdc65c46cce8 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.299929] env[65726]: DEBUG oslo_vmware.api [None req-cb64949f-89ac-453e-b54b-6d8bfee8a2ad tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Waiting for the task: (returnval){ [ 1012.299929] env[65726]: value = "task-5116476" [ 1012.299929] env[65726]: _type = "Task" [ 1012.299929] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1012.308519] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1510ef99-f4fe-455d-afee-aeca283c51e9 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.321533] env[65726]: DEBUG oslo_vmware.api [None req-cb64949f-89ac-453e-b54b-6d8bfee8a2ad tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Task: {'id': task-5116476, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.322906] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d25b71f-a9f9-4f3c-abdd-f75291a0ba4f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.364474] env[65726]: DEBUG nova.compute.manager [None req-59619dfd-192b-4bfb-9fd7-b4adaf4a925f tempest-ServerMetadataTestJSON-637854735 tempest-ServerMetadataTestJSON-637854735-project-member] [instance: 699127a9-f40f-4c1c-ba4a-625af097350b] Start building block device mappings for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1012.369229] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6935566-a068-4b86-939f-3ad1558fb5d1 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.379194] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08f55c91-9139-4c34-87ab-20f0141dd80d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.395087] env[65726]: DEBUG nova.compute.provider_tree [None req-d8ec8f57-04a2-4fd5-b8fa-981fbeb154b5 tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1012.430902] env[65726]: DEBUG oslo_vmware.api [None req-db012f3d-acd5-42b8-b506-baca1980cf7e tempest-FloatingIPsAssociationNegativeTestJSON-737087291 tempest-FloatingIPsAssociationNegativeTestJSON-737087291-project-member] Task: {'id': task-5116474, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.176213} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1012.430902] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-db012f3d-acd5-42b8-b506-baca1980cf7e tempest-FloatingIPsAssociationNegativeTestJSON-737087291 tempest-FloatingIPsAssociationNegativeTestJSON-737087291-project-member] Deleted the datastore file {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1012.431108] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-db012f3d-acd5-42b8-b506-baca1980cf7e tempest-FloatingIPsAssociationNegativeTestJSON-737087291 tempest-FloatingIPsAssociationNegativeTestJSON-737087291-project-member] [instance: d8468efa-c2d3-4dce-ab89-fc077011e3d8] Deleted contents of the VM from datastore datastore2 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1012.431282] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-db012f3d-acd5-42b8-b506-baca1980cf7e tempest-FloatingIPsAssociationNegativeTestJSON-737087291 tempest-FloatingIPsAssociationNegativeTestJSON-737087291-project-member] [instance: d8468efa-c2d3-4dce-ab89-fc077011e3d8] Instance destroyed {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1012.431456] env[65726]: INFO nova.compute.manager [None req-db012f3d-acd5-42b8-b506-baca1980cf7e tempest-FloatingIPsAssociationNegativeTestJSON-737087291 tempest-FloatingIPsAssociationNegativeTestJSON-737087291-project-member] [instance: d8468efa-c2d3-4dce-ab89-fc077011e3d8] Took 1.14 seconds to destroy the instance on the hypervisor. [ 1012.431781] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-db012f3d-acd5-42b8-b506-baca1980cf7e tempest-FloatingIPsAssociationNegativeTestJSON-737087291 tempest-FloatingIPsAssociationNegativeTestJSON-737087291-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1012.431987] env[65726]: DEBUG nova.compute.manager [-] [instance: d8468efa-c2d3-4dce-ab89-fc077011e3d8] Deallocating network for instance {{(pid=65726) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1012.432114] env[65726]: DEBUG nova.network.neutron [-] [instance: d8468efa-c2d3-4dce-ab89-fc077011e3d8] deallocate_for_instance() {{(pid=65726) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 1012.432690] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1012.432954] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1012.506063] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a27b5bca-2338-4542-8dd7-9cac7e4f3b3a tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Acquiring lock "refresh_cache-a286ba4f-a83e-4c30-8079-ba88e3ba5696" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1012.506063] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a27b5bca-2338-4542-8dd7-9cac7e4f3b3a tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Acquired lock "refresh_cache-a286ba4f-a83e-4c30-8079-ba88e3ba5696" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1012.506279] env[65726]: DEBUG nova.network.neutron [None req-a27b5bca-2338-4542-8dd7-9cac7e4f3b3a tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] [instance: a286ba4f-a83e-4c30-8079-ba88e3ba5696] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1012.526061] env[65726]: DEBUG nova.network.neutron [None req-59619dfd-192b-4bfb-9fd7-b4adaf4a925f tempest-ServerMetadataTestJSON-637854735 tempest-ServerMetadataTestJSON-637854735-project-member] [instance: 699127a9-f40f-4c1c-ba4a-625af097350b] Successfully created port: 544abfef-23be-48c2-932e-1bf64094c763 {{(pid=65726) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1012.530191] env[65726]: INFO nova.compute.manager [-] [instance: 7aa4c0d9-f7ca-42ba-9fd0-4075cd76752b] Took 1.79 seconds to deallocate network for instance. [ 1012.632260] env[65726]: DEBUG oslo_vmware.api [None req-12289131-1b6e-4176-844f-57bf1f466d9e tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Task: {'id': task-5116475, 'name': PowerOffVM_Task, 'duration_secs': 0.238882} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1012.632582] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-12289131-1b6e-4176-844f-57bf1f466d9e tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: 895797e4-2941-44cd-aab1-67afa4fac02d] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1012.632778] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-12289131-1b6e-4176-844f-57bf1f466d9e tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: 895797e4-2941-44cd-aab1-67afa4fac02d] Unregistering the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1012.633158] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-52950fd7-b360-4323-b558-d941cb1e4c9d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.648564] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1012.648935] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1012.717051] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-12289131-1b6e-4176-844f-57bf1f466d9e tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: 895797e4-2941-44cd-aab1-67afa4fac02d] Unregistered the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1012.717399] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-12289131-1b6e-4176-844f-57bf1f466d9e tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: 895797e4-2941-44cd-aab1-67afa4fac02d] Deleting contents of the VM from datastore datastore2 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1012.717609] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-12289131-1b6e-4176-844f-57bf1f466d9e tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Deleting the datastore file [datastore2] 895797e4-2941-44cd-aab1-67afa4fac02d {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1012.717915] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-010d8a1c-2938-44c7-8f97-66672f503862 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.727022] env[65726]: DEBUG oslo_vmware.api [None req-12289131-1b6e-4176-844f-57bf1f466d9e tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Waiting for the task: (returnval){ [ 1012.727022] env[65726]: value = "task-5116478" [ 1012.727022] env[65726]: _type = "Task" [ 1012.727022] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1012.739687] env[65726]: DEBUG oslo_vmware.api [None req-12289131-1b6e-4176-844f-57bf1f466d9e tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Task: {'id': task-5116478, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.811658] env[65726]: DEBUG oslo_vmware.api [None req-cb64949f-89ac-453e-b54b-6d8bfee8a2ad tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Task: {'id': task-5116476, 'name': ReconfigVM_Task, 'duration_secs': 0.321382} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1012.812039] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-cb64949f-89ac-453e-b54b-6d8bfee8a2ad tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: cf950ec3-9914-4eb9-99db-048b3969bb21] Reconfigured VM instance instance-0000003e to detach disk 2001 {{(pid=65726) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1012.817366] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8c37d494-d8f5-4285-ad72-4f7ac907c9d7 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.835210] env[65726]: DEBUG oslo_vmware.api [None req-cb64949f-89ac-453e-b54b-6d8bfee8a2ad tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Waiting for the task: (returnval){ [ 1012.835210] env[65726]: value = "task-5116479" [ 1012.835210] env[65726]: _type = "Task" [ 1012.835210] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1012.844924] env[65726]: DEBUG oslo_vmware.api [None req-cb64949f-89ac-453e-b54b-6d8bfee8a2ad tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Task: {'id': task-5116479, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.898355] env[65726]: DEBUG nova.scheduler.client.report [None req-d8ec8f57-04a2-4fd5-b8fa-981fbeb154b5 tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1013.009177] env[65726]: WARNING openstack [None req-a27b5bca-2338-4542-8dd7-9cac7e4f3b3a tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1013.009646] env[65726]: WARNING openstack [None req-a27b5bca-2338-4542-8dd7-9cac7e4f3b3a tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1013.014884] env[65726]: DEBUG nova.network.neutron [None req-a27b5bca-2338-4542-8dd7-9cac7e4f3b3a tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] [instance: a286ba4f-a83e-4c30-8079-ba88e3ba5696] Instance cache missing network info. {{(pid=65726) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 1013.039368] env[65726]: DEBUG oslo_concurrency.lockutils [None req-12727f4f-3b97-4bba-895f-741fee8ce27a tempest-ServersTestFqdnHostnames-1699296007 tempest-ServersTestFqdnHostnames-1699296007-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1013.190270] env[65726]: WARNING openstack [req-cb947eba-a8a1-4290-995e-ddce430f1644 req-a8968651-90de-490e-b08c-7f679a48135b service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1013.190660] env[65726]: WARNING openstack [req-cb947eba-a8a1-4290-995e-ddce430f1644 req-a8968651-90de-490e-b08c-7f679a48135b service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1013.239840] env[65726]: DEBUG oslo_vmware.api [None req-12289131-1b6e-4176-844f-57bf1f466d9e tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Task: {'id': task-5116478, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.191001} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1013.240181] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-12289131-1b6e-4176-844f-57bf1f466d9e tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Deleted the datastore file {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1013.240737] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-12289131-1b6e-4176-844f-57bf1f466d9e tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: 895797e4-2941-44cd-aab1-67afa4fac02d] Deleted contents of the VM from datastore datastore2 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1013.242455] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-12289131-1b6e-4176-844f-57bf1f466d9e tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: 895797e4-2941-44cd-aab1-67afa4fac02d] Instance destroyed {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1013.242455] env[65726]: INFO nova.compute.manager [None req-12289131-1b6e-4176-844f-57bf1f466d9e tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: 895797e4-2941-44cd-aab1-67afa4fac02d] Took 1.15 seconds to destroy the instance on the hypervisor. [ 1013.242455] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-12289131-1b6e-4176-844f-57bf1f466d9e tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1013.242455] env[65726]: DEBUG nova.compute.manager [-] [instance: 895797e4-2941-44cd-aab1-67afa4fac02d] Deallocating network for instance {{(pid=65726) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1013.242455] env[65726]: DEBUG nova.network.neutron [-] [instance: 895797e4-2941-44cd-aab1-67afa4fac02d] deallocate_for_instance() {{(pid=65726) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 1013.242983] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1013.243267] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1013.351703] env[65726]: DEBUG oslo_vmware.api [None req-cb64949f-89ac-453e-b54b-6d8bfee8a2ad tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Task: {'id': task-5116479, 'name': ReconfigVM_Task, 'duration_secs': 0.178897} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1013.352147] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-cb64949f-89ac-453e-b54b-6d8bfee8a2ad tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: cf950ec3-9914-4eb9-99db-048b3969bb21] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-995220', 'volume_id': 'dbb0735d-409f-439d-a32a-4de84d10f7a9', 'name': 'volume-dbb0735d-409f-439d-a32a-4de84d10f7a9', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'cf950ec3-9914-4eb9-99db-048b3969bb21', 'attached_at': '', 'detached_at': '', 'volume_id': 'dbb0735d-409f-439d-a32a-4de84d10f7a9', 'serial': 'dbb0735d-409f-439d-a32a-4de84d10f7a9'} {{(pid=65726) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1013.375388] env[65726]: DEBUG nova.compute.manager [None req-59619dfd-192b-4bfb-9fd7-b4adaf4a925f tempest-ServerMetadataTestJSON-637854735 tempest-ServerMetadataTestJSON-637854735-project-member] [instance: 699127a9-f40f-4c1c-ba4a-625af097350b] Start spawning the instance on the hypervisor. {{(pid=65726) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1013.404025] env[65726]: DEBUG oslo_concurrency.lockutils [None req-d8ec8f57-04a2-4fd5-b8fa-981fbeb154b5 tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.055s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1013.407915] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5f1cb876-8e7c-4680-9e78-4abda3ac2850 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.148s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1013.410145] env[65726]: INFO nova.compute.claims [None req-5f1cb876-8e7c-4680-9e78-4abda3ac2850 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: fb8eac12-aedd-426f-9dd3-6c52f9d4c71d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1013.420553] env[65726]: DEBUG nova.virt.hardware [None req-59619dfd-192b-4bfb-9fd7-b4adaf4a925f tempest-ServerMetadataTestJSON-637854735 tempest-ServerMetadataTestJSON-637854735-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T19:28:34Z,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1539f06fa3534e90acbc3a60c4b8bd3f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T19:28:35Z,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1013.420553] env[65726]: DEBUG nova.virt.hardware [None req-59619dfd-192b-4bfb-9fd7-b4adaf4a925f tempest-ServerMetadataTestJSON-637854735 tempest-ServerMetadataTestJSON-637854735-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1013.420553] env[65726]: DEBUG nova.virt.hardware [None req-59619dfd-192b-4bfb-9fd7-b4adaf4a925f tempest-ServerMetadataTestJSON-637854735 tempest-ServerMetadataTestJSON-637854735-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1013.420758] env[65726]: DEBUG nova.virt.hardware [None req-59619dfd-192b-4bfb-9fd7-b4adaf4a925f tempest-ServerMetadataTestJSON-637854735 tempest-ServerMetadataTestJSON-637854735-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1013.420941] env[65726]: DEBUG nova.virt.hardware [None req-59619dfd-192b-4bfb-9fd7-b4adaf4a925f tempest-ServerMetadataTestJSON-637854735 tempest-ServerMetadataTestJSON-637854735-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1013.421039] env[65726]: DEBUG nova.virt.hardware [None req-59619dfd-192b-4bfb-9fd7-b4adaf4a925f tempest-ServerMetadataTestJSON-637854735 tempest-ServerMetadataTestJSON-637854735-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1013.421253] env[65726]: DEBUG nova.virt.hardware [None req-59619dfd-192b-4bfb-9fd7-b4adaf4a925f tempest-ServerMetadataTestJSON-637854735 tempest-ServerMetadataTestJSON-637854735-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1013.421406] env[65726]: DEBUG nova.virt.hardware [None req-59619dfd-192b-4bfb-9fd7-b4adaf4a925f tempest-ServerMetadataTestJSON-637854735 tempest-ServerMetadataTestJSON-637854735-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1013.421667] env[65726]: DEBUG nova.virt.hardware [None req-59619dfd-192b-4bfb-9fd7-b4adaf4a925f tempest-ServerMetadataTestJSON-637854735 tempest-ServerMetadataTestJSON-637854735-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1013.421938] env[65726]: DEBUG nova.virt.hardware [None req-59619dfd-192b-4bfb-9fd7-b4adaf4a925f tempest-ServerMetadataTestJSON-637854735 tempest-ServerMetadataTestJSON-637854735-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1013.422141] env[65726]: DEBUG nova.virt.hardware [None req-59619dfd-192b-4bfb-9fd7-b4adaf4a925f tempest-ServerMetadataTestJSON-637854735 tempest-ServerMetadataTestJSON-637854735-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1013.422847] env[65726]: DEBUG oslo_concurrency.lockutils [None req-4ed5fc19-fb1a-414f-a0bd-edcdbeccf1c1 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Acquiring lock "refresh_cache-dcde25df-87a5-47fb-94ce-334f68894e04" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1013.423014] env[65726]: DEBUG oslo_concurrency.lockutils [None req-4ed5fc19-fb1a-414f-a0bd-edcdbeccf1c1 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Acquired lock "refresh_cache-dcde25df-87a5-47fb-94ce-334f68894e04" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1013.423885] env[65726]: DEBUG nova.network.neutron [None req-4ed5fc19-fb1a-414f-a0bd-edcdbeccf1c1 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: dcde25df-87a5-47fb-94ce-334f68894e04] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1013.426451] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bf3b3ee-23a6-44a0-9c06-cb70b5fda42b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.439847] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53a5212d-0941-494b-ad89-318119b6406f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.448563] env[65726]: INFO nova.scheduler.client.report [None req-d8ec8f57-04a2-4fd5-b8fa-981fbeb154b5 tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Deleted allocations for instance 406c59e7-8eb8-4cfd-8e1e-e151ab908571 [ 1013.909700] env[65726]: DEBUG nova.objects.instance [None req-cb64949f-89ac-453e-b54b-6d8bfee8a2ad tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Lazy-loading 'flavor' on Instance uuid cf950ec3-9914-4eb9-99db-048b3969bb21 {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1013.935020] env[65726]: WARNING openstack [None req-4ed5fc19-fb1a-414f-a0bd-edcdbeccf1c1 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1013.935414] env[65726]: WARNING openstack [None req-4ed5fc19-fb1a-414f-a0bd-edcdbeccf1c1 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1013.970078] env[65726]: DEBUG nova.network.neutron [-] [instance: d8468efa-c2d3-4dce-ab89-fc077011e3d8] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1013.972831] env[65726]: DEBUG oslo_concurrency.lockutils [None req-d8ec8f57-04a2-4fd5-b8fa-981fbeb154b5 tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Lock "406c59e7-8eb8-4cfd-8e1e-e151ab908571" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.822s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1014.231159] env[65726]: DEBUG nova.network.neutron [None req-59619dfd-192b-4bfb-9fd7-b4adaf4a925f tempest-ServerMetadataTestJSON-637854735 tempest-ServerMetadataTestJSON-637854735-project-member] [instance: 699127a9-f40f-4c1c-ba4a-625af097350b] Successfully updated port: 544abfef-23be-48c2-932e-1bf64094c763 {{(pid=65726) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 1014.367646] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1014.367973] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1014.438801] env[65726]: WARNING openstack [None req-a27b5bca-2338-4542-8dd7-9cac7e4f3b3a tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1014.439235] env[65726]: WARNING openstack [None req-a27b5bca-2338-4542-8dd7-9cac7e4f3b3a tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1014.475800] env[65726]: INFO nova.compute.manager [-] [instance: d8468efa-c2d3-4dce-ab89-fc077011e3d8] Took 2.04 seconds to deallocate network for instance. [ 1014.500388] env[65726]: DEBUG nova.network.neutron [req-cb947eba-a8a1-4290-995e-ddce430f1644 req-a8968651-90de-490e-b08c-7f679a48135b service nova] [instance: d8468efa-c2d3-4dce-ab89-fc077011e3d8] Updated VIF entry in instance network info cache for port 44639a8b-881d-41d3-a114-96a6bccf4064. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 1014.500795] env[65726]: DEBUG nova.network.neutron [req-cb947eba-a8a1-4290-995e-ddce430f1644 req-a8968651-90de-490e-b08c-7f679a48135b service nova] [instance: d8468efa-c2d3-4dce-ab89-fc077011e3d8] Updating instance_info_cache with network_info: [{"id": "44639a8b-881d-41d3-a114-96a6bccf4064", "address": "fa:16:3e:85:bb:61", "network": {"id": "0bcec3df-912b-4dad-9c7b-4b1b24cce6eb", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationNegativeTestJSON-754860692-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "db6b64230abc42069f5587ef94c82f5a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "01fe2e08-46f6-4cee-aefd-934461f8077d", "external-id": "nsx-vlan-transportzone-806", "segmentation_id": 806, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap44639a8b-88", "ovs_interfaceid": "44639a8b-881d-41d3-a114-96a6bccf4064", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1014.515888] env[65726]: DEBUG oslo_vmware.rw_handles [None req-bd9cdf9a-c493-4e4b-996c-7dba718501da tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52dc880d-b914-d12a-301e-6991c37c09ef/disk-0.vmdk. {{(pid=65726) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1014.520558] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6a40ea0-5169-4fef-a513-08bdd85c290d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.530530] env[65726]: DEBUG oslo_vmware.rw_handles [None req-bd9cdf9a-c493-4e4b-996c-7dba718501da tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52dc880d-b914-d12a-301e-6991c37c09ef/disk-0.vmdk is in state: ready. {{(pid=65726) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1014.531030] env[65726]: ERROR oslo_vmware.rw_handles [None req-bd9cdf9a-c493-4e4b-996c-7dba718501da tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52dc880d-b914-d12a-301e-6991c37c09ef/disk-0.vmdk due to incomplete transfer. [ 1014.531120] env[65726]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-b38f0abe-3404-40d6-bb01-94beb4120c3c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.550263] env[65726]: DEBUG nova.compute.manager [req-d6496508-3e63-47fd-9235-77f839b245e0 req-ebeee55f-ffaf-4757-9092-5ba4644ebc02 service nova] [instance: 7aa4c0d9-f7ca-42ba-9fd0-4075cd76752b] Received event network-vif-deleted-47831805-14e7-41d0-91a6-54a9da3ff27c {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1014.561666] env[65726]: DEBUG oslo_vmware.rw_handles [None req-bd9cdf9a-c493-4e4b-996c-7dba718501da tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52dc880d-b914-d12a-301e-6991c37c09ef/disk-0.vmdk. {{(pid=65726) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1014.561666] env[65726]: DEBUG nova.virt.vmwareapi.images [None req-bd9cdf9a-c493-4e4b-996c-7dba718501da tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: 067c9332-fa2c-47fe-93ae-643de1fe77cf] Uploaded image 54e60667-7017-48b6-80e1-6e4cfed9ff74 to the Glance image server {{(pid=65726) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 1014.562817] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-bd9cdf9a-c493-4e4b-996c-7dba718501da tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: 067c9332-fa2c-47fe-93ae-643de1fe77cf] Destroying the VM {{(pid=65726) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 1014.563502] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-14b0ba25-c05e-4c73-8be6-edc6b914a5e7 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.573662] env[65726]: DEBUG oslo_vmware.api [None req-bd9cdf9a-c493-4e4b-996c-7dba718501da tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Waiting for the task: (returnval){ [ 1014.573662] env[65726]: value = "task-5116480" [ 1014.573662] env[65726]: _type = "Task" [ 1014.573662] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.588262] env[65726]: DEBUG oslo_vmware.api [None req-bd9cdf9a-c493-4e4b-996c-7dba718501da tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': task-5116480, 'name': Destroy_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.669023] env[65726]: WARNING openstack [None req-a27b5bca-2338-4542-8dd7-9cac7e4f3b3a tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1014.669023] env[65726]: WARNING openstack [None req-a27b5bca-2338-4542-8dd7-9cac7e4f3b3a tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1014.738695] env[65726]: DEBUG oslo_concurrency.lockutils [None req-59619dfd-192b-4bfb-9fd7-b4adaf4a925f tempest-ServerMetadataTestJSON-637854735 tempest-ServerMetadataTestJSON-637854735-project-member] Acquiring lock "refresh_cache-699127a9-f40f-4c1c-ba4a-625af097350b" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1014.738695] env[65726]: DEBUG oslo_concurrency.lockutils [None req-59619dfd-192b-4bfb-9fd7-b4adaf4a925f tempest-ServerMetadataTestJSON-637854735 tempest-ServerMetadataTestJSON-637854735-project-member] Acquired lock "refresh_cache-699127a9-f40f-4c1c-ba4a-625af097350b" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1014.738695] env[65726]: DEBUG nova.network.neutron [None req-59619dfd-192b-4bfb-9fd7-b4adaf4a925f tempest-ServerMetadataTestJSON-637854735 tempest-ServerMetadataTestJSON-637854735-project-member] [instance: 699127a9-f40f-4c1c-ba4a-625af097350b] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1014.758590] env[65726]: WARNING openstack [None req-4ed5fc19-fb1a-414f-a0bd-edcdbeccf1c1 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1014.758965] env[65726]: WARNING openstack [None req-4ed5fc19-fb1a-414f-a0bd-edcdbeccf1c1 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1014.799497] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b22671f5-6246-47c6-baf8-1f22d64124a0 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.819689] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11ab5f14-587b-4f70-bf38-488b6ba896d3 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.868980] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c20b2c9-b75d-44a5-854c-2968dbe22a5d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.883027] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc8233fe-36dc-43d3-8fae-3c20f62eaef7 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.897951] env[65726]: DEBUG nova.compute.provider_tree [None req-5f1cb876-8e7c-4680-9e78-4abda3ac2850 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1014.901994] env[65726]: DEBUG nova.compute.manager [req-3f2bf7e6-ad41-4f51-ba71-ec94c7c4013a req-9ace82fb-8fbd-465b-9f3a-2b8eff988426 service nova] [instance: a286ba4f-a83e-4c30-8079-ba88e3ba5696] Received event network-vif-plugged-bae2338e-4be5-4ac9-b3b7-0f31ce3e0b37 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1014.902240] env[65726]: DEBUG oslo_concurrency.lockutils [req-3f2bf7e6-ad41-4f51-ba71-ec94c7c4013a req-9ace82fb-8fbd-465b-9f3a-2b8eff988426 service nova] Acquiring lock "a286ba4f-a83e-4c30-8079-ba88e3ba5696-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1014.902470] env[65726]: DEBUG oslo_concurrency.lockutils [req-3f2bf7e6-ad41-4f51-ba71-ec94c7c4013a req-9ace82fb-8fbd-465b-9f3a-2b8eff988426 service nova] Lock "a286ba4f-a83e-4c30-8079-ba88e3ba5696-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1014.902660] env[65726]: DEBUG oslo_concurrency.lockutils [req-3f2bf7e6-ad41-4f51-ba71-ec94c7c4013a req-9ace82fb-8fbd-465b-9f3a-2b8eff988426 service nova] Lock "a286ba4f-a83e-4c30-8079-ba88e3ba5696-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1014.902814] env[65726]: DEBUG nova.compute.manager [req-3f2bf7e6-ad41-4f51-ba71-ec94c7c4013a req-9ace82fb-8fbd-465b-9f3a-2b8eff988426 service nova] [instance: a286ba4f-a83e-4c30-8079-ba88e3ba5696] No waiting events found dispatching network-vif-plugged-bae2338e-4be5-4ac9-b3b7-0f31ce3e0b37 {{(pid=65726) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1014.906087] env[65726]: WARNING nova.compute.manager [req-3f2bf7e6-ad41-4f51-ba71-ec94c7c4013a req-9ace82fb-8fbd-465b-9f3a-2b8eff988426 service nova] [instance: a286ba4f-a83e-4c30-8079-ba88e3ba5696] Received unexpected event network-vif-plugged-bae2338e-4be5-4ac9-b3b7-0f31ce3e0b37 for instance with vm_state building and task_state spawning. [ 1014.906087] env[65726]: DEBUG nova.compute.manager [req-3f2bf7e6-ad41-4f51-ba71-ec94c7c4013a req-9ace82fb-8fbd-465b-9f3a-2b8eff988426 service nova] [instance: a286ba4f-a83e-4c30-8079-ba88e3ba5696] Received event network-changed-bae2338e-4be5-4ac9-b3b7-0f31ce3e0b37 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1014.906087] env[65726]: DEBUG nova.compute.manager [req-3f2bf7e6-ad41-4f51-ba71-ec94c7c4013a req-9ace82fb-8fbd-465b-9f3a-2b8eff988426 service nova] [instance: a286ba4f-a83e-4c30-8079-ba88e3ba5696] Refreshing instance network info cache due to event network-changed-bae2338e-4be5-4ac9-b3b7-0f31ce3e0b37. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 1014.906087] env[65726]: DEBUG oslo_concurrency.lockutils [req-3f2bf7e6-ad41-4f51-ba71-ec94c7c4013a req-9ace82fb-8fbd-465b-9f3a-2b8eff988426 service nova] Acquiring lock "refresh_cache-a286ba4f-a83e-4c30-8079-ba88e3ba5696" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1014.923255] env[65726]: DEBUG oslo_concurrency.lockutils [None req-cb64949f-89ac-453e-b54b-6d8bfee8a2ad tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Lock "cf950ec3-9914-4eb9-99db-048b3969bb21" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.347s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1014.996372] env[65726]: DEBUG oslo_concurrency.lockutils [None req-db012f3d-acd5-42b8-b506-baca1980cf7e tempest-FloatingIPsAssociationNegativeTestJSON-737087291 tempest-FloatingIPsAssociationNegativeTestJSON-737087291-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1015.006312] env[65726]: DEBUG oslo_concurrency.lockutils [req-cb947eba-a8a1-4290-995e-ddce430f1644 req-a8968651-90de-490e-b08c-7f679a48135b service nova] Releasing lock "refresh_cache-d8468efa-c2d3-4dce-ab89-fc077011e3d8" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1015.087986] env[65726]: DEBUG oslo_vmware.api [None req-bd9cdf9a-c493-4e4b-996c-7dba718501da tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': task-5116480, 'name': Destroy_Task} progress is 33%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.114946] env[65726]: DEBUG nova.network.neutron [None req-a27b5bca-2338-4542-8dd7-9cac7e4f3b3a tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] [instance: a286ba4f-a83e-4c30-8079-ba88e3ba5696] Updating instance_info_cache with network_info: [{"id": "bae2338e-4be5-4ac9-b3b7-0f31ce3e0b37", "address": "fa:16:3e:5a:d0:da", "network": {"id": "a2c8f0c5-7fad-4b72-8e1a-376e03d3c0ff", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-2135463722-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1d5cb47905144da49aaf07ddb7d263af", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "98f447de-d71e-41ef-bc37-ed97b4a1f58f", "external-id": "nsx-vlan-transportzone-904", "segmentation_id": 904, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbae2338e-4b", "ovs_interfaceid": "bae2338e-4be5-4ac9-b3b7-0f31ce3e0b37", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1015.159992] env[65726]: WARNING openstack [None req-4ed5fc19-fb1a-414f-a0bd-edcdbeccf1c1 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1015.161296] env[65726]: WARNING openstack [None req-4ed5fc19-fb1a-414f-a0bd-edcdbeccf1c1 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1015.241946] env[65726]: WARNING openstack [None req-59619dfd-192b-4bfb-9fd7-b4adaf4a925f tempest-ServerMetadataTestJSON-637854735 tempest-ServerMetadataTestJSON-637854735-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1015.242391] env[65726]: WARNING openstack [None req-59619dfd-192b-4bfb-9fd7-b4adaf4a925f tempest-ServerMetadataTestJSON-637854735 tempest-ServerMetadataTestJSON-637854735-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1015.252953] env[65726]: DEBUG nova.network.neutron [None req-59619dfd-192b-4bfb-9fd7-b4adaf4a925f tempest-ServerMetadataTestJSON-637854735 tempest-ServerMetadataTestJSON-637854735-project-member] [instance: 699127a9-f40f-4c1c-ba4a-625af097350b] Instance cache missing network info. {{(pid=65726) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 1015.398993] env[65726]: DEBUG oslo_concurrency.lockutils [None req-1d2791c0-c341-4061-b914-188004153713 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1015.406999] env[65726]: DEBUG nova.scheduler.client.report [None req-5f1cb876-8e7c-4680-9e78-4abda3ac2850 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1015.543156] env[65726]: DEBUG nova.network.neutron [-] [instance: 895797e4-2941-44cd-aab1-67afa4fac02d] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1015.586237] env[65726]: DEBUG oslo_vmware.api [None req-bd9cdf9a-c493-4e4b-996c-7dba718501da tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': task-5116480, 'name': Destroy_Task, 'duration_secs': 0.792614} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1015.586374] env[65726]: INFO nova.virt.vmwareapi.vm_util [None req-bd9cdf9a-c493-4e4b-996c-7dba718501da tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: 067c9332-fa2c-47fe-93ae-643de1fe77cf] Destroyed the VM [ 1015.586616] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-bd9cdf9a-c493-4e4b-996c-7dba718501da tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: 067c9332-fa2c-47fe-93ae-643de1fe77cf] Deleting Snapshot of the VM instance {{(pid=65726) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1015.586888] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-c4d075eb-55e9-45d1-9c11-5919596d0609 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.598444] env[65726]: DEBUG oslo_vmware.api [None req-bd9cdf9a-c493-4e4b-996c-7dba718501da tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Waiting for the task: (returnval){ [ 1015.598444] env[65726]: value = "task-5116481" [ 1015.598444] env[65726]: _type = "Task" [ 1015.598444] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1015.613198] env[65726]: DEBUG oslo_vmware.api [None req-bd9cdf9a-c493-4e4b-996c-7dba718501da tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': task-5116481, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.623026] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a27b5bca-2338-4542-8dd7-9cac7e4f3b3a tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Releasing lock "refresh_cache-a286ba4f-a83e-4c30-8079-ba88e3ba5696" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1015.623026] env[65726]: DEBUG nova.compute.manager [None req-a27b5bca-2338-4542-8dd7-9cac7e4f3b3a tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] [instance: a286ba4f-a83e-4c30-8079-ba88e3ba5696] Instance network_info: |[{"id": "bae2338e-4be5-4ac9-b3b7-0f31ce3e0b37", "address": "fa:16:3e:5a:d0:da", "network": {"id": "a2c8f0c5-7fad-4b72-8e1a-376e03d3c0ff", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-2135463722-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1d5cb47905144da49aaf07ddb7d263af", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "98f447de-d71e-41ef-bc37-ed97b4a1f58f", "external-id": "nsx-vlan-transportzone-904", "segmentation_id": 904, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbae2338e-4b", "ovs_interfaceid": "bae2338e-4be5-4ac9-b3b7-0f31ce3e0b37", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1015.623026] env[65726]: DEBUG oslo_concurrency.lockutils [req-3f2bf7e6-ad41-4f51-ba71-ec94c7c4013a req-9ace82fb-8fbd-465b-9f3a-2b8eff988426 service nova] Acquired lock "refresh_cache-a286ba4f-a83e-4c30-8079-ba88e3ba5696" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1015.623026] env[65726]: DEBUG nova.network.neutron [req-3f2bf7e6-ad41-4f51-ba71-ec94c7c4013a req-9ace82fb-8fbd-465b-9f3a-2b8eff988426 service nova] [instance: a286ba4f-a83e-4c30-8079-ba88e3ba5696] Refreshing network info cache for port bae2338e-4be5-4ac9-b3b7-0f31ce3e0b37 {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 1015.624086] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-a27b5bca-2338-4542-8dd7-9cac7e4f3b3a tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] [instance: a286ba4f-a83e-4c30-8079-ba88e3ba5696] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:5a:d0:da', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '98f447de-d71e-41ef-bc37-ed97b4a1f58f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'bae2338e-4be5-4ac9-b3b7-0f31ce3e0b37', 'vif_model': 'vmxnet3'}] {{(pid=65726) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1015.636019] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-a27b5bca-2338-4542-8dd7-9cac7e4f3b3a tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1015.637578] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a286ba4f-a83e-4c30-8079-ba88e3ba5696] Creating VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1015.637578] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-24228bff-cd6b-4315-8902-da481b1d3726 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.666517] env[65726]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1015.666517] env[65726]: value = "task-5116482" [ 1015.666517] env[65726]: _type = "Task" [ 1015.666517] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1015.676316] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116482, 'name': CreateVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.690868] env[65726]: WARNING openstack [None req-59619dfd-192b-4bfb-9fd7-b4adaf4a925f tempest-ServerMetadataTestJSON-637854735 tempest-ServerMetadataTestJSON-637854735-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1015.691282] env[65726]: WARNING openstack [None req-59619dfd-192b-4bfb-9fd7-b4adaf4a925f tempest-ServerMetadataTestJSON-637854735 tempest-ServerMetadataTestJSON-637854735-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1015.728316] env[65726]: INFO nova.network.neutron [None req-4ed5fc19-fb1a-414f-a0bd-edcdbeccf1c1 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: dcde25df-87a5-47fb-94ce-334f68894e04] Port b48313f9-a0d3-4fdd-88c6-d103aef0b4f2 from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 1015.910989] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5f1cb876-8e7c-4680-9e78-4abda3ac2850 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.503s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1015.911655] env[65726]: DEBUG nova.compute.manager [None req-5f1cb876-8e7c-4680-9e78-4abda3ac2850 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: fb8eac12-aedd-426f-9dd3-6c52f9d4c71d] Start building networks asynchronously for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1015.914655] env[65726]: DEBUG oslo_concurrency.lockutils [None req-12727f4f-3b97-4bba-895f-741fee8ce27a tempest-ServersTestFqdnHostnames-1699296007 tempest-ServersTestFqdnHostnames-1699296007-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.875s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1015.914933] env[65726]: DEBUG nova.objects.instance [None req-12727f4f-3b97-4bba-895f-741fee8ce27a tempest-ServersTestFqdnHostnames-1699296007 tempest-ServersTestFqdnHostnames-1699296007-project-member] Lazy-loading 'resources' on Instance uuid 7aa4c0d9-f7ca-42ba-9fd0-4075cd76752b {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1015.975441] env[65726]: WARNING openstack [None req-4ed5fc19-fb1a-414f-a0bd-edcdbeccf1c1 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1015.975953] env[65726]: WARNING openstack [None req-4ed5fc19-fb1a-414f-a0bd-edcdbeccf1c1 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1016.048327] env[65726]: INFO nova.compute.manager [-] [instance: 895797e4-2941-44cd-aab1-67afa4fac02d] Took 2.80 seconds to deallocate network for instance. [ 1016.115440] env[65726]: DEBUG oslo_vmware.api [None req-bd9cdf9a-c493-4e4b-996c-7dba718501da tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': task-5116481, 'name': RemoveSnapshot_Task} progress is 100%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.139550] env[65726]: WARNING openstack [req-3f2bf7e6-ad41-4f51-ba71-ec94c7c4013a req-9ace82fb-8fbd-465b-9f3a-2b8eff988426 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1016.140522] env[65726]: WARNING openstack [req-3f2bf7e6-ad41-4f51-ba71-ec94c7c4013a req-9ace82fb-8fbd-465b-9f3a-2b8eff988426 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1016.159257] env[65726]: WARNING openstack [None req-59619dfd-192b-4bfb-9fd7-b4adaf4a925f tempest-ServerMetadataTestJSON-637854735 tempest-ServerMetadataTestJSON-637854735-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1016.159975] env[65726]: WARNING openstack [None req-59619dfd-192b-4bfb-9fd7-b4adaf4a925f tempest-ServerMetadataTestJSON-637854735 tempest-ServerMetadataTestJSON-637854735-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1016.181317] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116482, 'name': CreateVM_Task, 'duration_secs': 0.485211} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1016.187229] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a286ba4f-a83e-4c30-8079-ba88e3ba5696] Created VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1016.188323] env[65726]: WARNING openstack [None req-a27b5bca-2338-4542-8dd7-9cac7e4f3b3a tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1016.188694] env[65726]: WARNING openstack [None req-a27b5bca-2338-4542-8dd7-9cac7e4f3b3a tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1016.197856] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a27b5bca-2338-4542-8dd7-9cac7e4f3b3a tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1016.198043] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a27b5bca-2338-4542-8dd7-9cac7e4f3b3a tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1016.198376] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a27b5bca-2338-4542-8dd7-9cac7e4f3b3a tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1016.198687] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3b6621bb-f0fa-4c03-801c-e147dad417ed {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.205295] env[65726]: DEBUG oslo_vmware.api [None req-a27b5bca-2338-4542-8dd7-9cac7e4f3b3a tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Waiting for the task: (returnval){ [ 1016.205295] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52acc554-9707-b359-6924-39b848bb0c4b" [ 1016.205295] env[65726]: _type = "Task" [ 1016.205295] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1016.217095] env[65726]: DEBUG oslo_vmware.api [None req-a27b5bca-2338-4542-8dd7-9cac7e4f3b3a tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52acc554-9707-b359-6924-39b848bb0c4b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.365280] env[65726]: WARNING openstack [None req-4ed5fc19-fb1a-414f-a0bd-edcdbeccf1c1 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1016.365738] env[65726]: WARNING openstack [None req-4ed5fc19-fb1a-414f-a0bd-edcdbeccf1c1 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1016.418676] env[65726]: DEBUG nova.compute.utils [None req-5f1cb876-8e7c-4680-9e78-4abda3ac2850 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Using /dev/sd instead of None {{(pid=65726) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1016.423513] env[65726]: DEBUG nova.compute.manager [None req-5f1cb876-8e7c-4680-9e78-4abda3ac2850 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: fb8eac12-aedd-426f-9dd3-6c52f9d4c71d] Allocating IP information in the background. {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1016.424517] env[65726]: DEBUG nova.network.neutron [None req-5f1cb876-8e7c-4680-9e78-4abda3ac2850 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: fb8eac12-aedd-426f-9dd3-6c52f9d4c71d] allocate_for_instance() {{(pid=65726) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 1016.424517] env[65726]: WARNING neutronclient.v2_0.client [None req-5f1cb876-8e7c-4680-9e78-4abda3ac2850 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1016.424517] env[65726]: WARNING neutronclient.v2_0.client [None req-5f1cb876-8e7c-4680-9e78-4abda3ac2850 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1016.425042] env[65726]: WARNING openstack [None req-5f1cb876-8e7c-4680-9e78-4abda3ac2850 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1016.425386] env[65726]: WARNING openstack [None req-5f1cb876-8e7c-4680-9e78-4abda3ac2850 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1016.480483] env[65726]: DEBUG nova.network.neutron [None req-59619dfd-192b-4bfb-9fd7-b4adaf4a925f tempest-ServerMetadataTestJSON-637854735 tempest-ServerMetadataTestJSON-637854735-project-member] [instance: 699127a9-f40f-4c1c-ba4a-625af097350b] Updating instance_info_cache with network_info: [{"id": "544abfef-23be-48c2-932e-1bf64094c763", "address": "fa:16:3e:76:c8:11", "network": {"id": "78a42c4c-9842-4dad-a5e3-15c74b9367af", "bridge": "br-int", "label": "tempest-ServerMetadataTestJSON-2015529683-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7befc792f09a46749a3f917ea0008953", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f5f49a7d-c6e5-404f-b71a-91d8c070cd18", "external-id": "nsx-vlan-transportzone-120", "segmentation_id": 120, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap544abfef-23", "ovs_interfaceid": "544abfef-23be-48c2-932e-1bf64094c763", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1016.554678] env[65726]: DEBUG oslo_concurrency.lockutils [None req-12289131-1b6e-4176-844f-57bf1f466d9e tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1016.558259] env[65726]: WARNING openstack [req-3f2bf7e6-ad41-4f51-ba71-ec94c7c4013a req-9ace82fb-8fbd-465b-9f3a-2b8eff988426 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1016.558551] env[65726]: WARNING openstack [req-3f2bf7e6-ad41-4f51-ba71-ec94c7c4013a req-9ace82fb-8fbd-465b-9f3a-2b8eff988426 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1016.588016] env[65726]: DEBUG nova.policy [None req-5f1cb876-8e7c-4680-9e78-4abda3ac2850 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a82797497789484a8fbf88a23a449c95', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f93d852e2c904f42981cbfff882558e9', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65726) authorize /opt/stack/nova/nova/policy.py:192}} [ 1016.618602] env[65726]: DEBUG oslo_vmware.api [None req-bd9cdf9a-c493-4e4b-996c-7dba718501da tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': task-5116481, 'name': RemoveSnapshot_Task, 'duration_secs': 0.624307} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1016.618889] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-bd9cdf9a-c493-4e4b-996c-7dba718501da tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: 067c9332-fa2c-47fe-93ae-643de1fe77cf] Deleted Snapshot of the VM instance {{(pid=65726) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1016.619175] env[65726]: DEBUG nova.compute.manager [None req-bd9cdf9a-c493-4e4b-996c-7dba718501da tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: 067c9332-fa2c-47fe-93ae-643de1fe77cf] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1016.620122] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c532a6ff-0e21-4b95-989c-ea5cceb0f2b2 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.629740] env[65726]: DEBUG nova.network.neutron [None req-4ed5fc19-fb1a-414f-a0bd-edcdbeccf1c1 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: dcde25df-87a5-47fb-94ce-334f68894e04] Updating instance_info_cache with network_info: [{"id": "bf51c96c-63c5-46de-a40f-58e6769c5fcf", "address": "fa:16:3e:4c:88:87", "network": {"id": "51a66448-0f72-4266-8868-856b48be839f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-155658961-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.146", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d46ccb00794f458b85da4a93879139ee", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b1f3e6c3-5584-4852-9017-476ab8ac4946", "external-id": "nsx-vlan-transportzone-304", "segmentation_id": 304, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbf51c96c-63", "ovs_interfaceid": "bf51c96c-63c5-46de-a40f-58e6769c5fcf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "280bd1f3-a22e-4f59-8574-d1783c74f8f3", "address": "fa:16:3e:2b:59:bd", "network": {"id": "51a66448-0f72-4266-8868-856b48be839f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-155658961-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d46ccb00794f458b85da4a93879139ee", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b1f3e6c3-5584-4852-9017-476ab8ac4946", "external-id": "nsx-vlan-transportzone-304", "segmentation_id": 304, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap280bd1f3-a2", "ovs_interfaceid": "280bd1f3-a22e-4f59-8574-d1783c74f8f3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1016.719230] env[65726]: DEBUG oslo_vmware.api [None req-a27b5bca-2338-4542-8dd7-9cac7e4f3b3a tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52acc554-9707-b359-6924-39b848bb0c4b, 'name': SearchDatastore_Task, 'duration_secs': 0.014293} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1016.719626] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a27b5bca-2338-4542-8dd7-9cac7e4f3b3a tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1016.719890] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-a27b5bca-2338-4542-8dd7-9cac7e4f3b3a tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] [instance: a286ba4f-a83e-4c30-8079-ba88e3ba5696] Processing image b52362a3-ee8a-4cbf-b06f-513b0cc8f95c {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1016.720175] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a27b5bca-2338-4542-8dd7-9cac7e4f3b3a tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1016.720347] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a27b5bca-2338-4542-8dd7-9cac7e4f3b3a tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1016.720556] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-a27b5bca-2338-4542-8dd7-9cac7e4f3b3a tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1016.720851] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-93091468-ff0f-4cae-9b71-a0104309f1e1 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.724440] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7041efb4-8b39-438b-aa86-f160cfd6f3e6 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.732425] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-786d2738-7a82-4d4f-9857-2c3a0f5306d1 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.740364] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-a27b5bca-2338-4542-8dd7-9cac7e4f3b3a tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1016.740364] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-a27b5bca-2338-4542-8dd7-9cac7e4f3b3a tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65726) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1016.740364] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-48bf426e-b194-49f8-b4eb-cca500833813 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.768531] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f883c4a-b08c-42df-87bf-e21122825ffa {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.772815] env[65726]: DEBUG oslo_vmware.api [None req-a27b5bca-2338-4542-8dd7-9cac7e4f3b3a tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Waiting for the task: (returnval){ [ 1016.772815] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52e41325-59bf-3c9d-ba66-f6409322ad7b" [ 1016.772815] env[65726]: _type = "Task" [ 1016.772815] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1016.780099] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c885121-947e-4f25-bc25-8b9b5e861671 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.791361] env[65726]: DEBUG oslo_vmware.api [None req-a27b5bca-2338-4542-8dd7-9cac7e4f3b3a tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52e41325-59bf-3c9d-ba66-f6409322ad7b, 'name': SearchDatastore_Task, 'duration_secs': 0.010399} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1016.792624] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-30eac306-dc39-4e18-a356-ce35f698ec00 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.804662] env[65726]: DEBUG nova.compute.provider_tree [None req-12727f4f-3b97-4bba-895f-741fee8ce27a tempest-ServersTestFqdnHostnames-1699296007 tempest-ServersTestFqdnHostnames-1699296007-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1016.809576] env[65726]: DEBUG oslo_vmware.api [None req-a27b5bca-2338-4542-8dd7-9cac7e4f3b3a tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Waiting for the task: (returnval){ [ 1016.809576] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]528341b9-4db1-8f67-d6e2-5e626fcd7d9b" [ 1016.809576] env[65726]: _type = "Task" [ 1016.809576] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1016.818282] env[65726]: DEBUG oslo_vmware.api [None req-a27b5bca-2338-4542-8dd7-9cac7e4f3b3a tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]528341b9-4db1-8f67-d6e2-5e626fcd7d9b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.924549] env[65726]: DEBUG nova.compute.manager [None req-5f1cb876-8e7c-4680-9e78-4abda3ac2850 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: fb8eac12-aedd-426f-9dd3-6c52f9d4c71d] Start building block device mappings for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1016.983923] env[65726]: DEBUG oslo_concurrency.lockutils [None req-59619dfd-192b-4bfb-9fd7-b4adaf4a925f tempest-ServerMetadataTestJSON-637854735 tempest-ServerMetadataTestJSON-637854735-project-member] Releasing lock "refresh_cache-699127a9-f40f-4c1c-ba4a-625af097350b" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1016.984333] env[65726]: DEBUG nova.compute.manager [None req-59619dfd-192b-4bfb-9fd7-b4adaf4a925f tempest-ServerMetadataTestJSON-637854735 tempest-ServerMetadataTestJSON-637854735-project-member] [instance: 699127a9-f40f-4c1c-ba4a-625af097350b] Instance network_info: |[{"id": "544abfef-23be-48c2-932e-1bf64094c763", "address": "fa:16:3e:76:c8:11", "network": {"id": "78a42c4c-9842-4dad-a5e3-15c74b9367af", "bridge": "br-int", "label": "tempest-ServerMetadataTestJSON-2015529683-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7befc792f09a46749a3f917ea0008953", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f5f49a7d-c6e5-404f-b71a-91d8c070cd18", "external-id": "nsx-vlan-transportzone-120", "segmentation_id": 120, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap544abfef-23", "ovs_interfaceid": "544abfef-23be-48c2-932e-1bf64094c763", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1016.984952] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-59619dfd-192b-4bfb-9fd7-b4adaf4a925f tempest-ServerMetadataTestJSON-637854735 tempest-ServerMetadataTestJSON-637854735-project-member] [instance: 699127a9-f40f-4c1c-ba4a-625af097350b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:76:c8:11', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f5f49a7d-c6e5-404f-b71a-91d8c070cd18', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '544abfef-23be-48c2-932e-1bf64094c763', 'vif_model': 'vmxnet3'}] {{(pid=65726) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1016.994265] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-59619dfd-192b-4bfb-9fd7-b4adaf4a925f tempest-ServerMetadataTestJSON-637854735 tempest-ServerMetadataTestJSON-637854735-project-member] Creating folder: Project (7befc792f09a46749a3f917ea0008953). Parent ref: group-v995008. {{(pid=65726) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1016.994671] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-707f13bf-5d48-4776-8afc-7eab71261fc0 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.008290] env[65726]: INFO nova.virt.vmwareapi.vm_util [None req-59619dfd-192b-4bfb-9fd7-b4adaf4a925f tempest-ServerMetadataTestJSON-637854735 tempest-ServerMetadataTestJSON-637854735-project-member] Created folder: Project (7befc792f09a46749a3f917ea0008953) in parent group-v995008. [ 1017.008628] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-59619dfd-192b-4bfb-9fd7-b4adaf4a925f tempest-ServerMetadataTestJSON-637854735 tempest-ServerMetadataTestJSON-637854735-project-member] Creating folder: Instances. Parent ref: group-v995257. {{(pid=65726) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1017.008910] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-32ef4b04-fde1-4099-b855-8d58c23d1a48 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.019852] env[65726]: INFO nova.virt.vmwareapi.vm_util [None req-59619dfd-192b-4bfb-9fd7-b4adaf4a925f tempest-ServerMetadataTestJSON-637854735 tempest-ServerMetadataTestJSON-637854735-project-member] Created folder: Instances in parent group-v995257. [ 1017.019852] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-59619dfd-192b-4bfb-9fd7-b4adaf4a925f tempest-ServerMetadataTestJSON-637854735 tempest-ServerMetadataTestJSON-637854735-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1017.020082] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 699127a9-f40f-4c1c-ba4a-625af097350b] Creating VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1017.020306] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5e827dc1-eac4-433e-9656-cd38bedeb824 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.041339] env[65726]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1017.041339] env[65726]: value = "task-5116485" [ 1017.041339] env[65726]: _type = "Task" [ 1017.041339] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1017.049369] env[65726]: WARNING openstack [req-3f2bf7e6-ad41-4f51-ba71-ec94c7c4013a req-9ace82fb-8fbd-465b-9f3a-2b8eff988426 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1017.049896] env[65726]: WARNING openstack [req-3f2bf7e6-ad41-4f51-ba71-ec94c7c4013a req-9ace82fb-8fbd-465b-9f3a-2b8eff988426 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1017.062505] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116485, 'name': CreateVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.101559] env[65726]: DEBUG nova.network.neutron [None req-5f1cb876-8e7c-4680-9e78-4abda3ac2850 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: fb8eac12-aedd-426f-9dd3-6c52f9d4c71d] Successfully created port: 839bf297-99c8-450e-9cd5-546932e1bea1 {{(pid=65726) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1017.133772] env[65726]: DEBUG oslo_concurrency.lockutils [None req-4ed5fc19-fb1a-414f-a0bd-edcdbeccf1c1 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Releasing lock "refresh_cache-dcde25df-87a5-47fb-94ce-334f68894e04" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1017.143581] env[65726]: INFO nova.compute.manager [None req-bd9cdf9a-c493-4e4b-996c-7dba718501da tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: 067c9332-fa2c-47fe-93ae-643de1fe77cf] Shelve offloading [ 1017.309999] env[65726]: DEBUG nova.scheduler.client.report [None req-12727f4f-3b97-4bba-895f-741fee8ce27a tempest-ServersTestFqdnHostnames-1699296007 tempest-ServersTestFqdnHostnames-1699296007-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1017.323780] env[65726]: DEBUG oslo_vmware.api [None req-a27b5bca-2338-4542-8dd7-9cac7e4f3b3a tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]528341b9-4db1-8f67-d6e2-5e626fcd7d9b, 'name': SearchDatastore_Task, 'duration_secs': 0.010768} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1017.324031] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a27b5bca-2338-4542-8dd7-9cac7e4f3b3a tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1017.324285] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-a27b5bca-2338-4542-8dd7-9cac7e4f3b3a tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore2] a286ba4f-a83e-4c30-8079-ba88e3ba5696/a286ba4f-a83e-4c30-8079-ba88e3ba5696.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1017.325385] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0ac21fc5-c7b7-44ae-9b5b-ed7cf918ca17 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.333823] env[65726]: DEBUG oslo_vmware.api [None req-a27b5bca-2338-4542-8dd7-9cac7e4f3b3a tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Waiting for the task: (returnval){ [ 1017.333823] env[65726]: value = "task-5116486" [ 1017.333823] env[65726]: _type = "Task" [ 1017.333823] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1017.344224] env[65726]: DEBUG oslo_vmware.api [None req-a27b5bca-2338-4542-8dd7-9cac7e4f3b3a tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Task: {'id': task-5116486, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.429157] env[65726]: DEBUG nova.network.neutron [req-3f2bf7e6-ad41-4f51-ba71-ec94c7c4013a req-9ace82fb-8fbd-465b-9f3a-2b8eff988426 service nova] [instance: a286ba4f-a83e-4c30-8079-ba88e3ba5696] Updated VIF entry in instance network info cache for port bae2338e-4be5-4ac9-b3b7-0f31ce3e0b37. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 1017.432524] env[65726]: DEBUG nova.network.neutron [req-3f2bf7e6-ad41-4f51-ba71-ec94c7c4013a req-9ace82fb-8fbd-465b-9f3a-2b8eff988426 service nova] [instance: a286ba4f-a83e-4c30-8079-ba88e3ba5696] Updating instance_info_cache with network_info: [{"id": "bae2338e-4be5-4ac9-b3b7-0f31ce3e0b37", "address": "fa:16:3e:5a:d0:da", "network": {"id": "a2c8f0c5-7fad-4b72-8e1a-376e03d3c0ff", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-2135463722-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1d5cb47905144da49aaf07ddb7d263af", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "98f447de-d71e-41ef-bc37-ed97b4a1f58f", "external-id": "nsx-vlan-transportzone-904", "segmentation_id": 904, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbae2338e-4b", "ovs_interfaceid": "bae2338e-4be5-4ac9-b3b7-0f31ce3e0b37", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1017.557034] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116485, 'name': CreateVM_Task, 'duration_secs': 0.344261} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1017.557529] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 699127a9-f40f-4c1c-ba4a-625af097350b] Created VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1017.558376] env[65726]: WARNING openstack [None req-59619dfd-192b-4bfb-9fd7-b4adaf4a925f tempest-ServerMetadataTestJSON-637854735 tempest-ServerMetadataTestJSON-637854735-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1017.558795] env[65726]: WARNING openstack [None req-59619dfd-192b-4bfb-9fd7-b4adaf4a925f tempest-ServerMetadataTestJSON-637854735 tempest-ServerMetadataTestJSON-637854735-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1017.565403] env[65726]: DEBUG oslo_concurrency.lockutils [None req-59619dfd-192b-4bfb-9fd7-b4adaf4a925f tempest-ServerMetadataTestJSON-637854735 tempest-ServerMetadataTestJSON-637854735-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1017.565740] env[65726]: DEBUG oslo_concurrency.lockutils [None req-59619dfd-192b-4bfb-9fd7-b4adaf4a925f tempest-ServerMetadataTestJSON-637854735 tempest-ServerMetadataTestJSON-637854735-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1017.566203] env[65726]: DEBUG oslo_concurrency.lockutils [None req-59619dfd-192b-4bfb-9fd7-b4adaf4a925f tempest-ServerMetadataTestJSON-637854735 tempest-ServerMetadataTestJSON-637854735-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1017.567673] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a2db2bfc-99cc-42e1-a55d-74432f7d3047 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.575485] env[65726]: DEBUG oslo_vmware.api [None req-59619dfd-192b-4bfb-9fd7-b4adaf4a925f tempest-ServerMetadataTestJSON-637854735 tempest-ServerMetadataTestJSON-637854735-project-member] Waiting for the task: (returnval){ [ 1017.575485] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]521ed347-e822-14c5-85ad-200afa64fcd1" [ 1017.575485] env[65726]: _type = "Task" [ 1017.575485] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1017.587219] env[65726]: DEBUG oslo_vmware.api [None req-59619dfd-192b-4bfb-9fd7-b4adaf4a925f tempest-ServerMetadataTestJSON-637854735 tempest-ServerMetadataTestJSON-637854735-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]521ed347-e822-14c5-85ad-200afa64fcd1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.645411] env[65726]: DEBUG oslo_concurrency.lockutils [None req-4ed5fc19-fb1a-414f-a0bd-edcdbeccf1c1 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Lock "interface-dcde25df-87a5-47fb-94ce-334f68894e04-b48313f9-a0d3-4fdd-88c6-d103aef0b4f2" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 12.589s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1017.651245] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-bd9cdf9a-c493-4e4b-996c-7dba718501da tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: 067c9332-fa2c-47fe-93ae-643de1fe77cf] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1017.652904] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-66022973-1643-4b5a-9940-d87bd9edd077 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.661779] env[65726]: DEBUG oslo_vmware.api [None req-bd9cdf9a-c493-4e4b-996c-7dba718501da tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Waiting for the task: (returnval){ [ 1017.661779] env[65726]: value = "task-5116487" [ 1017.661779] env[65726]: _type = "Task" [ 1017.661779] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1017.674804] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-bd9cdf9a-c493-4e4b-996c-7dba718501da tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: 067c9332-fa2c-47fe-93ae-643de1fe77cf] VM already powered off {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1017.675067] env[65726]: DEBUG nova.compute.manager [None req-bd9cdf9a-c493-4e4b-996c-7dba718501da tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: 067c9332-fa2c-47fe-93ae-643de1fe77cf] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1017.676064] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50be4e5a-b1a0-4f27-a946-e60dac4c8790 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.686201] env[65726]: DEBUG oslo_concurrency.lockutils [None req-bd9cdf9a-c493-4e4b-996c-7dba718501da tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Acquiring lock "refresh_cache-067c9332-fa2c-47fe-93ae-643de1fe77cf" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1017.686564] env[65726]: DEBUG oslo_concurrency.lockutils [None req-bd9cdf9a-c493-4e4b-996c-7dba718501da tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Acquired lock "refresh_cache-067c9332-fa2c-47fe-93ae-643de1fe77cf" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1017.686564] env[65726]: DEBUG nova.network.neutron [None req-bd9cdf9a-c493-4e4b-996c-7dba718501da tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: 067c9332-fa2c-47fe-93ae-643de1fe77cf] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1017.711935] env[65726]: DEBUG nova.compute.manager [req-81f65902-9a9f-4e96-9418-c4f581566ee4 req-b63b638c-1bb0-4bda-ae05-238c37d48441 service nova] [instance: dcde25df-87a5-47fb-94ce-334f68894e04] Received event network-vif-deleted-b48313f9-a0d3-4fdd-88c6-d103aef0b4f2 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1017.712150] env[65726]: DEBUG nova.compute.manager [req-81f65902-9a9f-4e96-9418-c4f581566ee4 req-b63b638c-1bb0-4bda-ae05-238c37d48441 service nova] [instance: d8468efa-c2d3-4dce-ab89-fc077011e3d8] Received event network-vif-deleted-44639a8b-881d-41d3-a114-96a6bccf4064 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1017.713758] env[65726]: INFO nova.compute.manager [req-81f65902-9a9f-4e96-9418-c4f581566ee4 req-b63b638c-1bb0-4bda-ae05-238c37d48441 service nova] [instance: d8468efa-c2d3-4dce-ab89-fc077011e3d8] Neutron deleted interface 44639a8b-881d-41d3-a114-96a6bccf4064; detaching it from the instance and deleting it from the info cache [ 1017.713758] env[65726]: DEBUG nova.network.neutron [req-81f65902-9a9f-4e96-9418-c4f581566ee4 req-b63b638c-1bb0-4bda-ae05-238c37d48441 service nova] [instance: d8468efa-c2d3-4dce-ab89-fc077011e3d8] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1017.817771] env[65726]: DEBUG oslo_concurrency.lockutils [None req-12727f4f-3b97-4bba-895f-741fee8ce27a tempest-ServersTestFqdnHostnames-1699296007 tempest-ServersTestFqdnHostnames-1699296007-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.903s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1017.823271] env[65726]: DEBUG oslo_concurrency.lockutils [None req-db012f3d-acd5-42b8-b506-baca1980cf7e tempest-FloatingIPsAssociationNegativeTestJSON-737087291 tempest-FloatingIPsAssociationNegativeTestJSON-737087291-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.826s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1017.823271] env[65726]: DEBUG nova.objects.instance [None req-db012f3d-acd5-42b8-b506-baca1980cf7e tempest-FloatingIPsAssociationNegativeTestJSON-737087291 tempest-FloatingIPsAssociationNegativeTestJSON-737087291-project-member] Lazy-loading 'resources' on Instance uuid d8468efa-c2d3-4dce-ab89-fc077011e3d8 {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1017.844685] env[65726]: DEBUG oslo_vmware.api [None req-a27b5bca-2338-4542-8dd7-9cac7e4f3b3a tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Task: {'id': task-5116486, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.845822] env[65726]: INFO nova.scheduler.client.report [None req-12727f4f-3b97-4bba-895f-741fee8ce27a tempest-ServersTestFqdnHostnames-1699296007 tempest-ServersTestFqdnHostnames-1699296007-project-member] Deleted allocations for instance 7aa4c0d9-f7ca-42ba-9fd0-4075cd76752b [ 1017.886706] env[65726]: DEBUG nova.compute.manager [req-6eac64e1-5151-44b2-b54c-172e273851b6 req-5db2a803-e44e-442b-a454-a869d7950055 service nova] [instance: 699127a9-f40f-4c1c-ba4a-625af097350b] Received event network-vif-plugged-544abfef-23be-48c2-932e-1bf64094c763 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1017.886826] env[65726]: DEBUG oslo_concurrency.lockutils [req-6eac64e1-5151-44b2-b54c-172e273851b6 req-5db2a803-e44e-442b-a454-a869d7950055 service nova] Acquiring lock "699127a9-f40f-4c1c-ba4a-625af097350b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1017.886969] env[65726]: DEBUG oslo_concurrency.lockutils [req-6eac64e1-5151-44b2-b54c-172e273851b6 req-5db2a803-e44e-442b-a454-a869d7950055 service nova] Lock "699127a9-f40f-4c1c-ba4a-625af097350b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1017.887469] env[65726]: DEBUG oslo_concurrency.lockutils [req-6eac64e1-5151-44b2-b54c-172e273851b6 req-5db2a803-e44e-442b-a454-a869d7950055 service nova] Lock "699127a9-f40f-4c1c-ba4a-625af097350b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1017.887585] env[65726]: DEBUG nova.compute.manager [req-6eac64e1-5151-44b2-b54c-172e273851b6 req-5db2a803-e44e-442b-a454-a869d7950055 service nova] [instance: 699127a9-f40f-4c1c-ba4a-625af097350b] No waiting events found dispatching network-vif-plugged-544abfef-23be-48c2-932e-1bf64094c763 {{(pid=65726) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1017.887733] env[65726]: WARNING nova.compute.manager [req-6eac64e1-5151-44b2-b54c-172e273851b6 req-5db2a803-e44e-442b-a454-a869d7950055 service nova] [instance: 699127a9-f40f-4c1c-ba4a-625af097350b] Received unexpected event network-vif-plugged-544abfef-23be-48c2-932e-1bf64094c763 for instance with vm_state building and task_state spawning. [ 1017.887896] env[65726]: DEBUG nova.compute.manager [req-6eac64e1-5151-44b2-b54c-172e273851b6 req-5db2a803-e44e-442b-a454-a869d7950055 service nova] [instance: 699127a9-f40f-4c1c-ba4a-625af097350b] Received event network-changed-544abfef-23be-48c2-932e-1bf64094c763 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1017.888062] env[65726]: DEBUG nova.compute.manager [req-6eac64e1-5151-44b2-b54c-172e273851b6 req-5db2a803-e44e-442b-a454-a869d7950055 service nova] [instance: 699127a9-f40f-4c1c-ba4a-625af097350b] Refreshing instance network info cache due to event network-changed-544abfef-23be-48c2-932e-1bf64094c763. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 1017.888335] env[65726]: DEBUG oslo_concurrency.lockutils [req-6eac64e1-5151-44b2-b54c-172e273851b6 req-5db2a803-e44e-442b-a454-a869d7950055 service nova] Acquiring lock "refresh_cache-699127a9-f40f-4c1c-ba4a-625af097350b" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1017.888457] env[65726]: DEBUG oslo_concurrency.lockutils [req-6eac64e1-5151-44b2-b54c-172e273851b6 req-5db2a803-e44e-442b-a454-a869d7950055 service nova] Acquired lock "refresh_cache-699127a9-f40f-4c1c-ba4a-625af097350b" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1017.889813] env[65726]: DEBUG nova.network.neutron [req-6eac64e1-5151-44b2-b54c-172e273851b6 req-5db2a803-e44e-442b-a454-a869d7950055 service nova] [instance: 699127a9-f40f-4c1c-ba4a-625af097350b] Refreshing network info cache for port 544abfef-23be-48c2-932e-1bf64094c763 {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 1017.941029] env[65726]: DEBUG oslo_concurrency.lockutils [req-3f2bf7e6-ad41-4f51-ba71-ec94c7c4013a req-9ace82fb-8fbd-465b-9f3a-2b8eff988426 service nova] Releasing lock "refresh_cache-a286ba4f-a83e-4c30-8079-ba88e3ba5696" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1017.941029] env[65726]: DEBUG nova.compute.manager [None req-5f1cb876-8e7c-4680-9e78-4abda3ac2850 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: fb8eac12-aedd-426f-9dd3-6c52f9d4c71d] Start spawning the instance on the hypervisor. {{(pid=65726) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1017.965147] env[65726]: DEBUG nova.virt.hardware [None req-5f1cb876-8e7c-4680-9e78-4abda3ac2850 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T19:28:34Z,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1539f06fa3534e90acbc3a60c4b8bd3f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T19:28:35Z,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1017.965388] env[65726]: DEBUG nova.virt.hardware [None req-5f1cb876-8e7c-4680-9e78-4abda3ac2850 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1017.965537] env[65726]: DEBUG nova.virt.hardware [None req-5f1cb876-8e7c-4680-9e78-4abda3ac2850 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1017.965761] env[65726]: DEBUG nova.virt.hardware [None req-5f1cb876-8e7c-4680-9e78-4abda3ac2850 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1017.965960] env[65726]: DEBUG nova.virt.hardware [None req-5f1cb876-8e7c-4680-9e78-4abda3ac2850 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1017.966171] env[65726]: DEBUG nova.virt.hardware [None req-5f1cb876-8e7c-4680-9e78-4abda3ac2850 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1017.966424] env[65726]: DEBUG nova.virt.hardware [None req-5f1cb876-8e7c-4680-9e78-4abda3ac2850 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1017.966637] env[65726]: DEBUG nova.virt.hardware [None req-5f1cb876-8e7c-4680-9e78-4abda3ac2850 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1017.966857] env[65726]: DEBUG nova.virt.hardware [None req-5f1cb876-8e7c-4680-9e78-4abda3ac2850 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1017.967052] env[65726]: DEBUG nova.virt.hardware [None req-5f1cb876-8e7c-4680-9e78-4abda3ac2850 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1017.967230] env[65726]: DEBUG nova.virt.hardware [None req-5f1cb876-8e7c-4680-9e78-4abda3ac2850 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1017.968596] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9e06212-3312-4da8-b4a5-41cede1d9250 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.977557] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c8553ee-6221-4b36-a4dc-168f0fcb1eb1 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.058496] env[65726]: DEBUG oslo_concurrency.lockutils [None req-7aacd0be-ee58-4b95-8abe-402ca2fa658f tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Acquiring lock "dcde25df-87a5-47fb-94ce-334f68894e04" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1018.059115] env[65726]: DEBUG oslo_concurrency.lockutils [None req-7aacd0be-ee58-4b95-8abe-402ca2fa658f tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Lock "dcde25df-87a5-47fb-94ce-334f68894e04" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1018.059115] env[65726]: DEBUG oslo_concurrency.lockutils [None req-7aacd0be-ee58-4b95-8abe-402ca2fa658f tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Acquiring lock "dcde25df-87a5-47fb-94ce-334f68894e04-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1018.059267] env[65726]: DEBUG oslo_concurrency.lockutils [None req-7aacd0be-ee58-4b95-8abe-402ca2fa658f tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Lock "dcde25df-87a5-47fb-94ce-334f68894e04-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1018.059433] env[65726]: DEBUG oslo_concurrency.lockutils [None req-7aacd0be-ee58-4b95-8abe-402ca2fa658f tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Lock "dcde25df-87a5-47fb-94ce-334f68894e04-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1018.063993] env[65726]: INFO nova.compute.manager [None req-7aacd0be-ee58-4b95-8abe-402ca2fa658f tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: dcde25df-87a5-47fb-94ce-334f68894e04] Terminating instance [ 1018.072067] env[65726]: DEBUG oslo_concurrency.lockutils [None req-f1e7868e-b3d9-4342-8972-f27e4cb4ec45 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] Acquiring lock "2b8eb60b-dc9b-47a4-9ddc-205f2af69888" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1018.072294] env[65726]: DEBUG oslo_concurrency.lockutils [None req-f1e7868e-b3d9-4342-8972-f27e4cb4ec45 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] Lock "2b8eb60b-dc9b-47a4-9ddc-205f2af69888" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1018.086808] env[65726]: DEBUG oslo_vmware.api [None req-59619dfd-192b-4bfb-9fd7-b4adaf4a925f tempest-ServerMetadataTestJSON-637854735 tempest-ServerMetadataTestJSON-637854735-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]521ed347-e822-14c5-85ad-200afa64fcd1, 'name': SearchDatastore_Task, 'duration_secs': 0.059744} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1018.087874] env[65726]: DEBUG oslo_concurrency.lockutils [None req-59619dfd-192b-4bfb-9fd7-b4adaf4a925f tempest-ServerMetadataTestJSON-637854735 tempest-ServerMetadataTestJSON-637854735-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1018.087874] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-59619dfd-192b-4bfb-9fd7-b4adaf4a925f tempest-ServerMetadataTestJSON-637854735 tempest-ServerMetadataTestJSON-637854735-project-member] [instance: 699127a9-f40f-4c1c-ba4a-625af097350b] Processing image b52362a3-ee8a-4cbf-b06f-513b0cc8f95c {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1018.088133] env[65726]: DEBUG oslo_concurrency.lockutils [None req-59619dfd-192b-4bfb-9fd7-b4adaf4a925f tempest-ServerMetadataTestJSON-637854735 tempest-ServerMetadataTestJSON-637854735-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1018.088282] env[65726]: DEBUG oslo_concurrency.lockutils [None req-59619dfd-192b-4bfb-9fd7-b4adaf4a925f tempest-ServerMetadataTestJSON-637854735 tempest-ServerMetadataTestJSON-637854735-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1018.088458] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-59619dfd-192b-4bfb-9fd7-b4adaf4a925f tempest-ServerMetadataTestJSON-637854735 tempest-ServerMetadataTestJSON-637854735-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1018.089130] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f12042fe-7a38-4a06-9134-1a7d36d51ead {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.099166] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-59619dfd-192b-4bfb-9fd7-b4adaf4a925f tempest-ServerMetadataTestJSON-637854735 tempest-ServerMetadataTestJSON-637854735-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1018.099379] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-59619dfd-192b-4bfb-9fd7-b4adaf4a925f tempest-ServerMetadataTestJSON-637854735 tempest-ServerMetadataTestJSON-637854735-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65726) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1018.100189] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4aee3181-0476-49ae-a1a2-fd46f5ffa14e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.109571] env[65726]: DEBUG oslo_vmware.api [None req-59619dfd-192b-4bfb-9fd7-b4adaf4a925f tempest-ServerMetadataTestJSON-637854735 tempest-ServerMetadataTestJSON-637854735-project-member] Waiting for the task: (returnval){ [ 1018.109571] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5219bea8-c278-7484-c31e-edd15397dc56" [ 1018.109571] env[65726]: _type = "Task" [ 1018.109571] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.118500] env[65726]: DEBUG oslo_vmware.api [None req-59619dfd-192b-4bfb-9fd7-b4adaf4a925f tempest-ServerMetadataTestJSON-637854735 tempest-ServerMetadataTestJSON-637854735-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5219bea8-c278-7484-c31e-edd15397dc56, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.191171] env[65726]: WARNING openstack [None req-bd9cdf9a-c493-4e4b-996c-7dba718501da tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1018.191555] env[65726]: WARNING openstack [None req-bd9cdf9a-c493-4e4b-996c-7dba718501da tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1018.215929] env[65726]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-36d074be-fca0-43a1-9f72-c2e89e42e4c9 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.226951] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53638c16-1cbb-4d4e-9265-c5a7f324ed5a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.267132] env[65726]: DEBUG nova.compute.manager [req-81f65902-9a9f-4e96-9418-c4f581566ee4 req-b63b638c-1bb0-4bda-ae05-238c37d48441 service nova] [instance: d8468efa-c2d3-4dce-ab89-fc077011e3d8] Detach interface failed, port_id=44639a8b-881d-41d3-a114-96a6bccf4064, reason: Instance d8468efa-c2d3-4dce-ab89-fc077011e3d8 could not be found. {{(pid=65726) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11669}} [ 1018.267132] env[65726]: DEBUG nova.compute.manager [req-81f65902-9a9f-4e96-9418-c4f581566ee4 req-b63b638c-1bb0-4bda-ae05-238c37d48441 service nova] [instance: 895797e4-2941-44cd-aab1-67afa4fac02d] Received event network-vif-deleted-a1dec2db-0e1e-45dd-8587-a84e4a5e2ef4 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1018.267132] env[65726]: DEBUG nova.compute.manager [req-81f65902-9a9f-4e96-9418-c4f581566ee4 req-b63b638c-1bb0-4bda-ae05-238c37d48441 service nova] [instance: dcde25df-87a5-47fb-94ce-334f68894e04] Received event network-vif-deleted-280bd1f3-a22e-4f59-8574-d1783c74f8f3 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1018.267132] env[65726]: INFO nova.compute.manager [req-81f65902-9a9f-4e96-9418-c4f581566ee4 req-b63b638c-1bb0-4bda-ae05-238c37d48441 service nova] [instance: dcde25df-87a5-47fb-94ce-334f68894e04] Neutron deleted interface 280bd1f3-a22e-4f59-8574-d1783c74f8f3; detaching it from the instance and deleting it from the info cache [ 1018.267132] env[65726]: DEBUG nova.network.neutron [req-81f65902-9a9f-4e96-9418-c4f581566ee4 req-b63b638c-1bb0-4bda-ae05-238c37d48441 service nova] [instance: dcde25df-87a5-47fb-94ce-334f68894e04] Updating instance_info_cache with network_info: [{"id": "bf51c96c-63c5-46de-a40f-58e6769c5fcf", "address": "fa:16:3e:4c:88:87", "network": {"id": "51a66448-0f72-4266-8868-856b48be839f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-155658961-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.146", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d46ccb00794f458b85da4a93879139ee", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b1f3e6c3-5584-4852-9017-476ab8ac4946", "external-id": "nsx-vlan-transportzone-304", "segmentation_id": 304, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbf51c96c-63", "ovs_interfaceid": "bf51c96c-63c5-46de-a40f-58e6769c5fcf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1018.345456] env[65726]: DEBUG oslo_vmware.api [None req-a27b5bca-2338-4542-8dd7-9cac7e4f3b3a tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Task: {'id': task-5116486, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.515658} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1018.345943] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-a27b5bca-2338-4542-8dd7-9cac7e4f3b3a tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore2] a286ba4f-a83e-4c30-8079-ba88e3ba5696/a286ba4f-a83e-4c30-8079-ba88e3ba5696.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1018.346205] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-a27b5bca-2338-4542-8dd7-9cac7e4f3b3a tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] [instance: a286ba4f-a83e-4c30-8079-ba88e3ba5696] Extending root virtual disk to 1048576 {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1018.346510] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-10068d20-7369-4da4-b486-62613d9dd871 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.357813] env[65726]: DEBUG oslo_vmware.api [None req-a27b5bca-2338-4542-8dd7-9cac7e4f3b3a tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Waiting for the task: (returnval){ [ 1018.357813] env[65726]: value = "task-5116488" [ 1018.357813] env[65726]: _type = "Task" [ 1018.357813] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.358370] env[65726]: DEBUG oslo_concurrency.lockutils [None req-12727f4f-3b97-4bba-895f-741fee8ce27a tempest-ServersTestFqdnHostnames-1699296007 tempest-ServersTestFqdnHostnames-1699296007-project-member] Lock "7aa4c0d9-f7ca-42ba-9fd0-4075cd76752b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.282s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1018.370025] env[65726]: DEBUG oslo_vmware.api [None req-a27b5bca-2338-4542-8dd7-9cac7e4f3b3a tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Task: {'id': task-5116488, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.393117] env[65726]: WARNING openstack [req-6eac64e1-5151-44b2-b54c-172e273851b6 req-5db2a803-e44e-442b-a454-a869d7950055 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1018.393653] env[65726]: WARNING openstack [req-6eac64e1-5151-44b2-b54c-172e273851b6 req-5db2a803-e44e-442b-a454-a869d7950055 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1018.565300] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7fda8eb-8908-4011-8b9a-0e4e4fef5d33 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.568306] env[65726]: DEBUG nova.compute.manager [None req-7aacd0be-ee58-4b95-8abe-402ca2fa658f tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: dcde25df-87a5-47fb-94ce-334f68894e04] Start destroying the instance on the hypervisor. {{(pid=65726) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1018.568406] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-7aacd0be-ee58-4b95-8abe-402ca2fa658f tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: dcde25df-87a5-47fb-94ce-334f68894e04] Destroying instance {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1018.569383] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90f2c4b2-f086-4cd3-9bce-bc39eca3ce52 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.574610] env[65726]: DEBUG nova.compute.manager [None req-f1e7868e-b3d9-4342-8972-f27e4cb4ec45 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] [instance: 2b8eb60b-dc9b-47a4-9ddc-205f2af69888] Starting instance... {{(pid=65726) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 1018.585238] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61e6d491-ef12-4221-81e8-55b13e1d0acb {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.590638] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-7aacd0be-ee58-4b95-8abe-402ca2fa658f tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: dcde25df-87a5-47fb-94ce-334f68894e04] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1018.591228] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-28d00a85-a2a5-4ba0-a2a7-88aba744a4fc {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.632762] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1b92fd1-f8ce-46c4-94e2-3a13b0f3b16d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.635603] env[65726]: DEBUG oslo_vmware.api [None req-7aacd0be-ee58-4b95-8abe-402ca2fa658f tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Waiting for the task: (returnval){ [ 1018.635603] env[65726]: value = "task-5116489" [ 1018.635603] env[65726]: _type = "Task" [ 1018.635603] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.639859] env[65726]: WARNING openstack [None req-bd9cdf9a-c493-4e4b-996c-7dba718501da tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1018.639859] env[65726]: WARNING openstack [None req-bd9cdf9a-c493-4e4b-996c-7dba718501da tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1018.656020] env[65726]: DEBUG oslo_vmware.api [None req-59619dfd-192b-4bfb-9fd7-b4adaf4a925f tempest-ServerMetadataTestJSON-637854735 tempest-ServerMetadataTestJSON-637854735-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5219bea8-c278-7484-c31e-edd15397dc56, 'name': SearchDatastore_Task, 'duration_secs': 0.010891} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1018.657258] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ec03fea-92c9-4ef2-99a6-b1e74e374a6a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.666794] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5b9cdaf2-b738-486e-8bd3-ba4704ba2af4 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.668211] env[65726]: DEBUG oslo_vmware.api [None req-7aacd0be-ee58-4b95-8abe-402ca2fa658f tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': task-5116489, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.676031] env[65726]: DEBUG oslo_vmware.api [None req-59619dfd-192b-4bfb-9fd7-b4adaf4a925f tempest-ServerMetadataTestJSON-637854735 tempest-ServerMetadataTestJSON-637854735-project-member] Waiting for the task: (returnval){ [ 1018.676031] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5241384c-9850-a3e6-0e5c-1e7c464f52b8" [ 1018.676031] env[65726]: _type = "Task" [ 1018.676031] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.681482] env[65726]: DEBUG nova.compute.provider_tree [None req-db012f3d-acd5-42b8-b506-baca1980cf7e tempest-FloatingIPsAssociationNegativeTestJSON-737087291 tempest-FloatingIPsAssociationNegativeTestJSON-737087291-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1018.692691] env[65726]: DEBUG oslo_vmware.api [None req-59619dfd-192b-4bfb-9fd7-b4adaf4a925f tempest-ServerMetadataTestJSON-637854735 tempest-ServerMetadataTestJSON-637854735-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5241384c-9850-a3e6-0e5c-1e7c464f52b8, 'name': SearchDatastore_Task, 'duration_secs': 0.017309} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1018.693485] env[65726]: DEBUG oslo_concurrency.lockutils [None req-59619dfd-192b-4bfb-9fd7-b4adaf4a925f tempest-ServerMetadataTestJSON-637854735 tempest-ServerMetadataTestJSON-637854735-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1018.693754] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-59619dfd-192b-4bfb-9fd7-b4adaf4a925f tempest-ServerMetadataTestJSON-637854735 tempest-ServerMetadataTestJSON-637854735-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore2] 699127a9-f40f-4c1c-ba4a-625af097350b/699127a9-f40f-4c1c-ba4a-625af097350b.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1018.694057] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-638608b5-b4f5-4e9d-ad1f-3d9887a0bf2e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.699707] env[65726]: DEBUG oslo_concurrency.lockutils [None req-345ea84c-77bf-4393-9db3-8f6989cf9c1c tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Acquiring lock "cf950ec3-9914-4eb9-99db-048b3969bb21" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1018.700117] env[65726]: DEBUG oslo_concurrency.lockutils [None req-345ea84c-77bf-4393-9db3-8f6989cf9c1c tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Lock "cf950ec3-9914-4eb9-99db-048b3969bb21" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1018.700409] env[65726]: DEBUG oslo_concurrency.lockutils [None req-345ea84c-77bf-4393-9db3-8f6989cf9c1c tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Acquiring lock "cf950ec3-9914-4eb9-99db-048b3969bb21-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1018.700655] env[65726]: DEBUG oslo_concurrency.lockutils [None req-345ea84c-77bf-4393-9db3-8f6989cf9c1c tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Lock "cf950ec3-9914-4eb9-99db-048b3969bb21-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1018.700908] env[65726]: DEBUG oslo_concurrency.lockutils [None req-345ea84c-77bf-4393-9db3-8f6989cf9c1c tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Lock "cf950ec3-9914-4eb9-99db-048b3969bb21-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1018.705099] env[65726]: DEBUG oslo_vmware.api [None req-59619dfd-192b-4bfb-9fd7-b4adaf4a925f tempest-ServerMetadataTestJSON-637854735 tempest-ServerMetadataTestJSON-637854735-project-member] Waiting for the task: (returnval){ [ 1018.705099] env[65726]: value = "task-5116490" [ 1018.705099] env[65726]: _type = "Task" [ 1018.705099] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.705704] env[65726]: INFO nova.compute.manager [None req-345ea84c-77bf-4393-9db3-8f6989cf9c1c tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: cf950ec3-9914-4eb9-99db-048b3969bb21] Terminating instance [ 1018.720434] env[65726]: DEBUG oslo_vmware.api [None req-59619dfd-192b-4bfb-9fd7-b4adaf4a925f tempest-ServerMetadataTestJSON-637854735 tempest-ServerMetadataTestJSON-637854735-project-member] Task: {'id': task-5116490, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.747844] env[65726]: WARNING openstack [req-6eac64e1-5151-44b2-b54c-172e273851b6 req-5db2a803-e44e-442b-a454-a869d7950055 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1018.749013] env[65726]: WARNING openstack [req-6eac64e1-5151-44b2-b54c-172e273851b6 req-5db2a803-e44e-442b-a454-a869d7950055 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1018.762118] env[65726]: DEBUG nova.network.neutron [None req-5f1cb876-8e7c-4680-9e78-4abda3ac2850 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: fb8eac12-aedd-426f-9dd3-6c52f9d4c71d] Successfully updated port: 839bf297-99c8-450e-9cd5-546932e1bea1 {{(pid=65726) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 1018.770101] env[65726]: DEBUG oslo_concurrency.lockutils [req-81f65902-9a9f-4e96-9418-c4f581566ee4 req-b63b638c-1bb0-4bda-ae05-238c37d48441 service nova] Acquiring lock "dcde25df-87a5-47fb-94ce-334f68894e04" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1018.870288] env[65726]: DEBUG oslo_vmware.api [None req-a27b5bca-2338-4542-8dd7-9cac7e4f3b3a tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Task: {'id': task-5116488, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.07226} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1018.870288] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-a27b5bca-2338-4542-8dd7-9cac7e4f3b3a tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] [instance: a286ba4f-a83e-4c30-8079-ba88e3ba5696] Extended root virtual disk {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1018.870774] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ce06035-bf99-4894-9634-2417510b3777 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.894932] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-a27b5bca-2338-4542-8dd7-9cac7e4f3b3a tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] [instance: a286ba4f-a83e-4c30-8079-ba88e3ba5696] Reconfiguring VM instance instance-00000055 to attach disk [datastore2] a286ba4f-a83e-4c30-8079-ba88e3ba5696/a286ba4f-a83e-4c30-8079-ba88e3ba5696.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1018.898463] env[65726]: WARNING openstack [None req-bd9cdf9a-c493-4e4b-996c-7dba718501da tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1018.898816] env[65726]: WARNING openstack [None req-bd9cdf9a-c493-4e4b-996c-7dba718501da tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1018.906548] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e7d6570b-dc9c-4697-85d7-73ea38aaec1c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.927299] env[65726]: WARNING openstack [req-6eac64e1-5151-44b2-b54c-172e273851b6 req-5db2a803-e44e-442b-a454-a869d7950055 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1018.927711] env[65726]: WARNING openstack [req-6eac64e1-5151-44b2-b54c-172e273851b6 req-5db2a803-e44e-442b-a454-a869d7950055 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1018.947244] env[65726]: DEBUG oslo_vmware.api [None req-a27b5bca-2338-4542-8dd7-9cac7e4f3b3a tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Waiting for the task: (returnval){ [ 1018.947244] env[65726]: value = "task-5116491" [ 1018.947244] env[65726]: _type = "Task" [ 1018.947244] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.960891] env[65726]: DEBUG oslo_vmware.api [None req-a27b5bca-2338-4542-8dd7-9cac7e4f3b3a tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Task: {'id': task-5116491, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.116734] env[65726]: DEBUG oslo_concurrency.lockutils [None req-f1e7868e-b3d9-4342-8972-f27e4cb4ec45 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1019.127829] env[65726]: DEBUG nova.network.neutron [None req-bd9cdf9a-c493-4e4b-996c-7dba718501da tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: 067c9332-fa2c-47fe-93ae-643de1fe77cf] Updating instance_info_cache with network_info: [{"id": "498a4766-79c8-471b-b6fc-e73c9eca5c5c", "address": "fa:16:3e:fc:03:ad", "network": {"id": "4c972d9d-db3b-4779-ac39-39d9a2d141dd", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1418913861-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cbb17a09b35c4c22ade5c4082b10bda9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "50cf0a70-948d-4611-af05-94c1483064ed", "external-id": "nsx-vlan-transportzone-536", "segmentation_id": 536, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap498a4766-79", "ovs_interfaceid": "498a4766-79c8-471b-b6fc-e73c9eca5c5c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1019.162608] env[65726]: DEBUG oslo_vmware.api [None req-7aacd0be-ee58-4b95-8abe-402ca2fa658f tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': task-5116489, 'name': PowerOffVM_Task, 'duration_secs': 0.251128} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.163230] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-7aacd0be-ee58-4b95-8abe-402ca2fa658f tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: dcde25df-87a5-47fb-94ce-334f68894e04] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1019.163230] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-7aacd0be-ee58-4b95-8abe-402ca2fa658f tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: dcde25df-87a5-47fb-94ce-334f68894e04] Unregistering the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1019.163814] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a39eacf1-0420-4e75-8820-72c02701c02e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.171192] env[65726]: DEBUG nova.network.neutron [req-6eac64e1-5151-44b2-b54c-172e273851b6 req-5db2a803-e44e-442b-a454-a869d7950055 service nova] [instance: 699127a9-f40f-4c1c-ba4a-625af097350b] Updated VIF entry in instance network info cache for port 544abfef-23be-48c2-932e-1bf64094c763. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 1019.171711] env[65726]: DEBUG nova.network.neutron [req-6eac64e1-5151-44b2-b54c-172e273851b6 req-5db2a803-e44e-442b-a454-a869d7950055 service nova] [instance: 699127a9-f40f-4c1c-ba4a-625af097350b] Updating instance_info_cache with network_info: [{"id": "544abfef-23be-48c2-932e-1bf64094c763", "address": "fa:16:3e:76:c8:11", "network": {"id": "78a42c4c-9842-4dad-a5e3-15c74b9367af", "bridge": "br-int", "label": "tempest-ServerMetadataTestJSON-2015529683-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7befc792f09a46749a3f917ea0008953", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f5f49a7d-c6e5-404f-b71a-91d8c070cd18", "external-id": "nsx-vlan-transportzone-120", "segmentation_id": 120, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap544abfef-23", "ovs_interfaceid": "544abfef-23be-48c2-932e-1bf64094c763", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1019.188112] env[65726]: DEBUG nova.scheduler.client.report [None req-db012f3d-acd5-42b8-b506-baca1980cf7e tempest-FloatingIPsAssociationNegativeTestJSON-737087291 tempest-FloatingIPsAssociationNegativeTestJSON-737087291-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1019.214051] env[65726]: DEBUG nova.compute.manager [None req-345ea84c-77bf-4393-9db3-8f6989cf9c1c tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: cf950ec3-9914-4eb9-99db-048b3969bb21] Start destroying the instance on the hypervisor. {{(pid=65726) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1019.214388] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-345ea84c-77bf-4393-9db3-8f6989cf9c1c tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: cf950ec3-9914-4eb9-99db-048b3969bb21] Destroying instance {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1019.215368] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e960f971-3b98-466b-9cdb-b3ca7a82484c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.222751] env[65726]: DEBUG oslo_vmware.api [None req-59619dfd-192b-4bfb-9fd7-b4adaf4a925f tempest-ServerMetadataTestJSON-637854735 tempest-ServerMetadataTestJSON-637854735-project-member] Task: {'id': task-5116490, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.229401] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-345ea84c-77bf-4393-9db3-8f6989cf9c1c tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: cf950ec3-9914-4eb9-99db-048b3969bb21] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1019.229549] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6e9a0db5-f486-4ae5-90c1-80a66896367d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.239102] env[65726]: DEBUG oslo_vmware.api [None req-345ea84c-77bf-4393-9db3-8f6989cf9c1c tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Waiting for the task: (returnval){ [ 1019.239102] env[65726]: value = "task-5116493" [ 1019.239102] env[65726]: _type = "Task" [ 1019.239102] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1019.252895] env[65726]: DEBUG oslo_vmware.api [None req-345ea84c-77bf-4393-9db3-8f6989cf9c1c tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Task: {'id': task-5116493, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.265648] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5f1cb876-8e7c-4680-9e78-4abda3ac2850 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Acquiring lock "refresh_cache-fb8eac12-aedd-426f-9dd3-6c52f9d4c71d" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1019.266813] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5f1cb876-8e7c-4680-9e78-4abda3ac2850 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Acquired lock "refresh_cache-fb8eac12-aedd-426f-9dd3-6c52f9d4c71d" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1019.266813] env[65726]: DEBUG nova.network.neutron [None req-5f1cb876-8e7c-4680-9e78-4abda3ac2850 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: fb8eac12-aedd-426f-9dd3-6c52f9d4c71d] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1019.311997] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-7aacd0be-ee58-4b95-8abe-402ca2fa658f tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: dcde25df-87a5-47fb-94ce-334f68894e04] Unregistered the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1019.312416] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-7aacd0be-ee58-4b95-8abe-402ca2fa658f tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: dcde25df-87a5-47fb-94ce-334f68894e04] Deleting contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1019.312698] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-7aacd0be-ee58-4b95-8abe-402ca2fa658f tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Deleting the datastore file [datastore1] dcde25df-87a5-47fb-94ce-334f68894e04 {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1019.313874] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9e98ea77-d074-43f8-aa0d-226792f7edb0 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.323870] env[65726]: DEBUG oslo_vmware.api [None req-7aacd0be-ee58-4b95-8abe-402ca2fa658f tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Waiting for the task: (returnval){ [ 1019.323870] env[65726]: value = "task-5116494" [ 1019.323870] env[65726]: _type = "Task" [ 1019.323870] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1019.336355] env[65726]: DEBUG oslo_vmware.api [None req-7aacd0be-ee58-4b95-8abe-402ca2fa658f tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': task-5116494, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.460505] env[65726]: DEBUG oslo_vmware.api [None req-a27b5bca-2338-4542-8dd7-9cac7e4f3b3a tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Task: {'id': task-5116491, 'name': ReconfigVM_Task, 'duration_secs': 0.394795} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.462026] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-a27b5bca-2338-4542-8dd7-9cac7e4f3b3a tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] [instance: a286ba4f-a83e-4c30-8079-ba88e3ba5696] Reconfigured VM instance instance-00000055 to attach disk [datastore2] a286ba4f-a83e-4c30-8079-ba88e3ba5696/a286ba4f-a83e-4c30-8079-ba88e3ba5696.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1019.462026] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-82947875-509b-421a-85a6-09e3d3f95442 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.472131] env[65726]: DEBUG oslo_vmware.api [None req-a27b5bca-2338-4542-8dd7-9cac7e4f3b3a tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Waiting for the task: (returnval){ [ 1019.472131] env[65726]: value = "task-5116495" [ 1019.472131] env[65726]: _type = "Task" [ 1019.472131] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1019.484425] env[65726]: DEBUG oslo_vmware.api [None req-a27b5bca-2338-4542-8dd7-9cac7e4f3b3a tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Task: {'id': task-5116495, 'name': Rename_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.632666] env[65726]: DEBUG oslo_concurrency.lockutils [None req-bd9cdf9a-c493-4e4b-996c-7dba718501da tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Releasing lock "refresh_cache-067c9332-fa2c-47fe-93ae-643de1fe77cf" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1019.633595] env[65726]: WARNING openstack [None req-bd9cdf9a-c493-4e4b-996c-7dba718501da tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1019.633872] env[65726]: WARNING openstack [None req-bd9cdf9a-c493-4e4b-996c-7dba718501da tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1019.640706] env[65726]: WARNING openstack [None req-bd9cdf9a-c493-4e4b-996c-7dba718501da tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1019.641316] env[65726]: WARNING openstack [None req-bd9cdf9a-c493-4e4b-996c-7dba718501da tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1019.675048] env[65726]: DEBUG oslo_concurrency.lockutils [req-6eac64e1-5151-44b2-b54c-172e273851b6 req-5db2a803-e44e-442b-a454-a869d7950055 service nova] Releasing lock "refresh_cache-699127a9-f40f-4c1c-ba4a-625af097350b" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1019.696598] env[65726]: DEBUG oslo_concurrency.lockutils [None req-2a9dd832-d188-4542-a7fe-bbdea8173b2d tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Acquiring lock "2c812714-dccc-4d1e-bdb5-c11e446949c3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1019.697025] env[65726]: DEBUG oslo_concurrency.lockutils [None req-2a9dd832-d188-4542-a7fe-bbdea8173b2d tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Lock "2c812714-dccc-4d1e-bdb5-c11e446949c3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1019.698787] env[65726]: DEBUG oslo_concurrency.lockutils [None req-db012f3d-acd5-42b8-b506-baca1980cf7e tempest-FloatingIPsAssociationNegativeTestJSON-737087291 tempest-FloatingIPsAssociationNegativeTestJSON-737087291-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.877s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1019.704770] env[65726]: DEBUG oslo_concurrency.lockutils [None req-1d2791c0-c341-4061-b914-188004153713 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.307s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1019.705142] env[65726]: DEBUG nova.objects.instance [None req-1d2791c0-c341-4061-b914-188004153713 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Lazy-loading 'pci_requests' on Instance uuid 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04 {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1019.724427] env[65726]: DEBUG oslo_vmware.api [None req-59619dfd-192b-4bfb-9fd7-b4adaf4a925f tempest-ServerMetadataTestJSON-637854735 tempest-ServerMetadataTestJSON-637854735-project-member] Task: {'id': task-5116490, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.595191} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.724788] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-59619dfd-192b-4bfb-9fd7-b4adaf4a925f tempest-ServerMetadataTestJSON-637854735 tempest-ServerMetadataTestJSON-637854735-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore2] 699127a9-f40f-4c1c-ba4a-625af097350b/699127a9-f40f-4c1c-ba4a-625af097350b.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1019.725083] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-59619dfd-192b-4bfb-9fd7-b4adaf4a925f tempest-ServerMetadataTestJSON-637854735 tempest-ServerMetadataTestJSON-637854735-project-member] [instance: 699127a9-f40f-4c1c-ba4a-625af097350b] Extending root virtual disk to 1048576 {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1019.725620] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ee781aa1-6da7-4cc1-9ae6-69b2029fc683 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.734655] env[65726]: DEBUG oslo_vmware.api [None req-59619dfd-192b-4bfb-9fd7-b4adaf4a925f tempest-ServerMetadataTestJSON-637854735 tempest-ServerMetadataTestJSON-637854735-project-member] Waiting for the task: (returnval){ [ 1019.734655] env[65726]: value = "task-5116496" [ 1019.734655] env[65726]: _type = "Task" [ 1019.734655] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1019.739908] env[65726]: INFO nova.scheduler.client.report [None req-db012f3d-acd5-42b8-b506-baca1980cf7e tempest-FloatingIPsAssociationNegativeTestJSON-737087291 tempest-FloatingIPsAssociationNegativeTestJSON-737087291-project-member] Deleted allocations for instance d8468efa-c2d3-4dce-ab89-fc077011e3d8 [ 1019.758979] env[65726]: DEBUG oslo_vmware.api [None req-59619dfd-192b-4bfb-9fd7-b4adaf4a925f tempest-ServerMetadataTestJSON-637854735 tempest-ServerMetadataTestJSON-637854735-project-member] Task: {'id': task-5116496, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.765808] env[65726]: DEBUG oslo_vmware.api [None req-345ea84c-77bf-4393-9db3-8f6989cf9c1c tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Task: {'id': task-5116493, 'name': PowerOffVM_Task, 'duration_secs': 0.284418} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.766195] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-345ea84c-77bf-4393-9db3-8f6989cf9c1c tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: cf950ec3-9914-4eb9-99db-048b3969bb21] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1019.766418] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-345ea84c-77bf-4393-9db3-8f6989cf9c1c tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: cf950ec3-9914-4eb9-99db-048b3969bb21] Unregistering the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1019.766773] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-cff047dd-5db4-48b4-adeb-cfd1e3f13d68 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.770233] env[65726]: WARNING openstack [None req-5f1cb876-8e7c-4680-9e78-4abda3ac2850 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1019.770678] env[65726]: WARNING openstack [None req-5f1cb876-8e7c-4680-9e78-4abda3ac2850 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1019.776247] env[65726]: DEBUG nova.network.neutron [None req-5f1cb876-8e7c-4680-9e78-4abda3ac2850 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: fb8eac12-aedd-426f-9dd3-6c52f9d4c71d] Instance cache missing network info. {{(pid=65726) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 1019.834449] env[65726]: DEBUG oslo_vmware.api [None req-7aacd0be-ee58-4b95-8abe-402ca2fa658f tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': task-5116494, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.206896} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.834784] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-7aacd0be-ee58-4b95-8abe-402ca2fa658f tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Deleted the datastore file {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1019.834953] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-7aacd0be-ee58-4b95-8abe-402ca2fa658f tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: dcde25df-87a5-47fb-94ce-334f68894e04] Deleted contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1019.835143] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-7aacd0be-ee58-4b95-8abe-402ca2fa658f tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: dcde25df-87a5-47fb-94ce-334f68894e04] Instance destroyed {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1019.835316] env[65726]: INFO nova.compute.manager [None req-7aacd0be-ee58-4b95-8abe-402ca2fa658f tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: dcde25df-87a5-47fb-94ce-334f68894e04] Took 1.27 seconds to destroy the instance on the hypervisor. [ 1019.835571] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-7aacd0be-ee58-4b95-8abe-402ca2fa658f tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1019.835809] env[65726]: DEBUG nova.compute.manager [-] [instance: dcde25df-87a5-47fb-94ce-334f68894e04] Deallocating network for instance {{(pid=65726) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1019.835875] env[65726]: DEBUG nova.network.neutron [-] [instance: dcde25df-87a5-47fb-94ce-334f68894e04] deallocate_for_instance() {{(pid=65726) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 1019.836479] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1019.836758] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1019.848372] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-345ea84c-77bf-4393-9db3-8f6989cf9c1c tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: cf950ec3-9914-4eb9-99db-048b3969bb21] Unregistered the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1019.852026] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-345ea84c-77bf-4393-9db3-8f6989cf9c1c tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: cf950ec3-9914-4eb9-99db-048b3969bb21] Deleting contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1019.852026] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-345ea84c-77bf-4393-9db3-8f6989cf9c1c tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Deleting the datastore file [datastore1] cf950ec3-9914-4eb9-99db-048b3969bb21 {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1019.852026] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-929b84db-7366-402d-be12-ad7e756fa55a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.857351] env[65726]: DEBUG oslo_vmware.api [None req-345ea84c-77bf-4393-9db3-8f6989cf9c1c tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Waiting for the task: (returnval){ [ 1019.857351] env[65726]: value = "task-5116498" [ 1019.857351] env[65726]: _type = "Task" [ 1019.857351] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1019.866757] env[65726]: DEBUG oslo_vmware.api [None req-345ea84c-77bf-4393-9db3-8f6989cf9c1c tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Task: {'id': task-5116498, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.925180] env[65726]: DEBUG nova.compute.manager [req-6ba1a87e-aec7-4177-baba-8f351c1b93f6 req-8ff9acf5-89e9-42fd-9978-c56fc3740918 service nova] [instance: fb8eac12-aedd-426f-9dd3-6c52f9d4c71d] Received event network-vif-plugged-839bf297-99c8-450e-9cd5-546932e1bea1 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1019.925180] env[65726]: DEBUG oslo_concurrency.lockutils [req-6ba1a87e-aec7-4177-baba-8f351c1b93f6 req-8ff9acf5-89e9-42fd-9978-c56fc3740918 service nova] Acquiring lock "fb8eac12-aedd-426f-9dd3-6c52f9d4c71d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1019.925432] env[65726]: DEBUG oslo_concurrency.lockutils [req-6ba1a87e-aec7-4177-baba-8f351c1b93f6 req-8ff9acf5-89e9-42fd-9978-c56fc3740918 service nova] Lock "fb8eac12-aedd-426f-9dd3-6c52f9d4c71d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1019.925669] env[65726]: DEBUG oslo_concurrency.lockutils [req-6ba1a87e-aec7-4177-baba-8f351c1b93f6 req-8ff9acf5-89e9-42fd-9978-c56fc3740918 service nova] Lock "fb8eac12-aedd-426f-9dd3-6c52f9d4c71d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1019.925893] env[65726]: DEBUG nova.compute.manager [req-6ba1a87e-aec7-4177-baba-8f351c1b93f6 req-8ff9acf5-89e9-42fd-9978-c56fc3740918 service nova] [instance: fb8eac12-aedd-426f-9dd3-6c52f9d4c71d] No waiting events found dispatching network-vif-plugged-839bf297-99c8-450e-9cd5-546932e1bea1 {{(pid=65726) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1019.926098] env[65726]: WARNING nova.compute.manager [req-6ba1a87e-aec7-4177-baba-8f351c1b93f6 req-8ff9acf5-89e9-42fd-9978-c56fc3740918 service nova] [instance: fb8eac12-aedd-426f-9dd3-6c52f9d4c71d] Received unexpected event network-vif-plugged-839bf297-99c8-450e-9cd5-546932e1bea1 for instance with vm_state building and task_state spawning. [ 1019.926296] env[65726]: DEBUG nova.compute.manager [req-6ba1a87e-aec7-4177-baba-8f351c1b93f6 req-8ff9acf5-89e9-42fd-9978-c56fc3740918 service nova] [instance: fb8eac12-aedd-426f-9dd3-6c52f9d4c71d] Received event network-changed-839bf297-99c8-450e-9cd5-546932e1bea1 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1019.926478] env[65726]: DEBUG nova.compute.manager [req-6ba1a87e-aec7-4177-baba-8f351c1b93f6 req-8ff9acf5-89e9-42fd-9978-c56fc3740918 service nova] [instance: fb8eac12-aedd-426f-9dd3-6c52f9d4c71d] Refreshing instance network info cache due to event network-changed-839bf297-99c8-450e-9cd5-546932e1bea1. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 1019.926677] env[65726]: DEBUG oslo_concurrency.lockutils [req-6ba1a87e-aec7-4177-baba-8f351c1b93f6 req-8ff9acf5-89e9-42fd-9978-c56fc3740918 service nova] Acquiring lock "refresh_cache-fb8eac12-aedd-426f-9dd3-6c52f9d4c71d" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1019.984915] env[65726]: DEBUG oslo_vmware.api [None req-a27b5bca-2338-4542-8dd7-9cac7e4f3b3a tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Task: {'id': task-5116495, 'name': Rename_Task, 'duration_secs': 0.202778} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.985346] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-a27b5bca-2338-4542-8dd7-9cac7e4f3b3a tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] [instance: a286ba4f-a83e-4c30-8079-ba88e3ba5696] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1019.985669] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d6d8a6c7-3c27-4703-b4be-b0e6bd2770ee {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.994137] env[65726]: DEBUG oslo_vmware.api [None req-a27b5bca-2338-4542-8dd7-9cac7e4f3b3a tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Waiting for the task: (returnval){ [ 1019.994137] env[65726]: value = "task-5116499" [ 1019.994137] env[65726]: _type = "Task" [ 1019.994137] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1020.003646] env[65726]: DEBUG oslo_vmware.api [None req-a27b5bca-2338-4542-8dd7-9cac7e4f3b3a tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Task: {'id': task-5116499, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.064194] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1020.064664] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1020.081059] env[65726]: WARNING openstack [None req-5f1cb876-8e7c-4680-9e78-4abda3ac2850 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1020.081550] env[65726]: WARNING openstack [None req-5f1cb876-8e7c-4680-9e78-4abda3ac2850 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1020.157820] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-bd9cdf9a-c493-4e4b-996c-7dba718501da tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: 067c9332-fa2c-47fe-93ae-643de1fe77cf] Destroying instance {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1020.158902] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10a5411f-95e8-4176-a14c-529719f552ae {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.167652] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-bd9cdf9a-c493-4e4b-996c-7dba718501da tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: 067c9332-fa2c-47fe-93ae-643de1fe77cf] Unregistering the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1020.168108] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b7bb21f7-e5d0-4699-928d-3fc369a6a40b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.170475] env[65726]: DEBUG nova.network.neutron [-] Unable to show port 280bd1f3-a22e-4f59-8574-d1783c74f8f3 as it no longer exists. {{(pid=65726) _unbind_ports /opt/stack/nova/nova/network/neutron.py:678}} [ 1020.204797] env[65726]: DEBUG nova.compute.manager [None req-2a9dd832-d188-4542-a7fe-bbdea8173b2d tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] [instance: 2c812714-dccc-4d1e-bdb5-c11e446949c3] Starting instance... {{(pid=65726) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 1020.209821] env[65726]: DEBUG nova.objects.instance [None req-1d2791c0-c341-4061-b914-188004153713 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Lazy-loading 'numa_topology' on Instance uuid 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04 {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1020.242772] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-bd9cdf9a-c493-4e4b-996c-7dba718501da tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: 067c9332-fa2c-47fe-93ae-643de1fe77cf] Unregistered the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1020.243077] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-bd9cdf9a-c493-4e4b-996c-7dba718501da tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: 067c9332-fa2c-47fe-93ae-643de1fe77cf] Deleting contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1020.243286] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-bd9cdf9a-c493-4e4b-996c-7dba718501da tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Deleting the datastore file [datastore1] 067c9332-fa2c-47fe-93ae-643de1fe77cf {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1020.248113] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7d659b3c-f282-45c3-823e-f7740ca17251 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.251151] env[65726]: DEBUG oslo_vmware.api [None req-59619dfd-192b-4bfb-9fd7-b4adaf4a925f tempest-ServerMetadataTestJSON-637854735 tempest-ServerMetadataTestJSON-637854735-project-member] Task: {'id': task-5116496, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.080594} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1020.252360] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-59619dfd-192b-4bfb-9fd7-b4adaf4a925f tempest-ServerMetadataTestJSON-637854735 tempest-ServerMetadataTestJSON-637854735-project-member] [instance: 699127a9-f40f-4c1c-ba4a-625af097350b] Extended root virtual disk {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1020.256225] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbd21529-bd00-4381-b308-901bc07ec8d3 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.264025] env[65726]: DEBUG oslo_vmware.api [None req-bd9cdf9a-c493-4e4b-996c-7dba718501da tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Waiting for the task: (returnval){ [ 1020.264025] env[65726]: value = "task-5116501" [ 1020.264025] env[65726]: _type = "Task" [ 1020.264025] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1020.265435] env[65726]: DEBUG oslo_concurrency.lockutils [None req-db012f3d-acd5-42b8-b506-baca1980cf7e tempest-FloatingIPsAssociationNegativeTestJSON-737087291 tempest-FloatingIPsAssociationNegativeTestJSON-737087291-project-member] Lock "d8468efa-c2d3-4dce-ab89-fc077011e3d8" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.482s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1020.301044] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-59619dfd-192b-4bfb-9fd7-b4adaf4a925f tempest-ServerMetadataTestJSON-637854735 tempest-ServerMetadataTestJSON-637854735-project-member] [instance: 699127a9-f40f-4c1c-ba4a-625af097350b] Reconfiguring VM instance instance-00000056 to attach disk [datastore2] 699127a9-f40f-4c1c-ba4a-625af097350b/699127a9-f40f-4c1c-ba4a-625af097350b.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1020.304803] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-acbd11ff-7990-46e3-af77-07f865fc2f03 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.324896] env[65726]: DEBUG nova.compute.manager [req-15090263-df59-45b6-8e44-7a0fc7aadc27 req-7177802f-da97-479f-9493-66838602270e service nova] [instance: 067c9332-fa2c-47fe-93ae-643de1fe77cf] Received event network-vif-unplugged-498a4766-79c8-471b-b6fc-e73c9eca5c5c {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1020.324896] env[65726]: DEBUG oslo_concurrency.lockutils [req-15090263-df59-45b6-8e44-7a0fc7aadc27 req-7177802f-da97-479f-9493-66838602270e service nova] Acquiring lock "067c9332-fa2c-47fe-93ae-643de1fe77cf-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1020.325227] env[65726]: DEBUG oslo_concurrency.lockutils [req-15090263-df59-45b6-8e44-7a0fc7aadc27 req-7177802f-da97-479f-9493-66838602270e service nova] Lock "067c9332-fa2c-47fe-93ae-643de1fe77cf-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1020.325389] env[65726]: DEBUG oslo_concurrency.lockutils [req-15090263-df59-45b6-8e44-7a0fc7aadc27 req-7177802f-da97-479f-9493-66838602270e service nova] Lock "067c9332-fa2c-47fe-93ae-643de1fe77cf-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1020.325617] env[65726]: DEBUG nova.compute.manager [req-15090263-df59-45b6-8e44-7a0fc7aadc27 req-7177802f-da97-479f-9493-66838602270e service nova] [instance: 067c9332-fa2c-47fe-93ae-643de1fe77cf] No waiting events found dispatching network-vif-unplugged-498a4766-79c8-471b-b6fc-e73c9eca5c5c {{(pid=65726) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1020.326270] env[65726]: WARNING nova.compute.manager [req-15090263-df59-45b6-8e44-7a0fc7aadc27 req-7177802f-da97-479f-9493-66838602270e service nova] [instance: 067c9332-fa2c-47fe-93ae-643de1fe77cf] Received unexpected event network-vif-unplugged-498a4766-79c8-471b-b6fc-e73c9eca5c5c for instance with vm_state shelved and task_state shelving_offloading. [ 1020.331105] env[65726]: DEBUG oslo_vmware.api [None req-bd9cdf9a-c493-4e4b-996c-7dba718501da tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': task-5116501, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.337616] env[65726]: DEBUG oslo_vmware.api [None req-59619dfd-192b-4bfb-9fd7-b4adaf4a925f tempest-ServerMetadataTestJSON-637854735 tempest-ServerMetadataTestJSON-637854735-project-member] Waiting for the task: (returnval){ [ 1020.337616] env[65726]: value = "task-5116502" [ 1020.337616] env[65726]: _type = "Task" [ 1020.337616] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1020.356174] env[65726]: DEBUG oslo_vmware.api [None req-59619dfd-192b-4bfb-9fd7-b4adaf4a925f tempest-ServerMetadataTestJSON-637854735 tempest-ServerMetadataTestJSON-637854735-project-member] Task: {'id': task-5116502, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.373087] env[65726]: DEBUG oslo_vmware.api [None req-345ea84c-77bf-4393-9db3-8f6989cf9c1c tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Task: {'id': task-5116498, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.468146} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1020.373388] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-345ea84c-77bf-4393-9db3-8f6989cf9c1c tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Deleted the datastore file {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1020.373598] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-345ea84c-77bf-4393-9db3-8f6989cf9c1c tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: cf950ec3-9914-4eb9-99db-048b3969bb21] Deleted contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1020.373790] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-345ea84c-77bf-4393-9db3-8f6989cf9c1c tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: cf950ec3-9914-4eb9-99db-048b3969bb21] Instance destroyed {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1020.373997] env[65726]: INFO nova.compute.manager [None req-345ea84c-77bf-4393-9db3-8f6989cf9c1c tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: cf950ec3-9914-4eb9-99db-048b3969bb21] Took 1.16 seconds to destroy the instance on the hypervisor. [ 1020.374295] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-345ea84c-77bf-4393-9db3-8f6989cf9c1c tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1020.374528] env[65726]: DEBUG nova.compute.manager [-] [instance: cf950ec3-9914-4eb9-99db-048b3969bb21] Deallocating network for instance {{(pid=65726) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1020.374619] env[65726]: DEBUG nova.network.neutron [-] [instance: cf950ec3-9914-4eb9-99db-048b3969bb21] deallocate_for_instance() {{(pid=65726) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 1020.375200] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1020.375475] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1020.505785] env[65726]: DEBUG oslo_vmware.api [None req-a27b5bca-2338-4542-8dd7-9cac7e4f3b3a tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Task: {'id': task-5116499, 'name': PowerOnVM_Task, 'duration_secs': 0.491264} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1020.506080] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-a27b5bca-2338-4542-8dd7-9cac7e4f3b3a tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] [instance: a286ba4f-a83e-4c30-8079-ba88e3ba5696] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1020.506298] env[65726]: INFO nova.compute.manager [None req-a27b5bca-2338-4542-8dd7-9cac7e4f3b3a tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] [instance: a286ba4f-a83e-4c30-8079-ba88e3ba5696] Took 9.04 seconds to spawn the instance on the hypervisor. [ 1020.506497] env[65726]: DEBUG nova.compute.manager [None req-a27b5bca-2338-4542-8dd7-9cac7e4f3b3a tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] [instance: a286ba4f-a83e-4c30-8079-ba88e3ba5696] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1020.507332] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-413aeb59-ed54-4b39-9424-5fda86cc7782 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.666671] env[65726]: WARNING openstack [None req-5f1cb876-8e7c-4680-9e78-4abda3ac2850 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1020.667209] env[65726]: WARNING openstack [None req-5f1cb876-8e7c-4680-9e78-4abda3ac2850 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1020.715395] env[65726]: INFO nova.compute.claims [None req-1d2791c0-c341-4061-b914-188004153713 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1020.748261] env[65726]: DEBUG oslo_concurrency.lockutils [None req-2a9dd832-d188-4542-a7fe-bbdea8173b2d tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1020.756932] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1020.757320] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1020.778014] env[65726]: DEBUG oslo_vmware.api [None req-bd9cdf9a-c493-4e4b-996c-7dba718501da tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': task-5116501, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.17975} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1020.778278] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-bd9cdf9a-c493-4e4b-996c-7dba718501da tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Deleted the datastore file {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1020.778459] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-bd9cdf9a-c493-4e4b-996c-7dba718501da tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: 067c9332-fa2c-47fe-93ae-643de1fe77cf] Deleted contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1020.778663] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-bd9cdf9a-c493-4e4b-996c-7dba718501da tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: 067c9332-fa2c-47fe-93ae-643de1fe77cf] Instance destroyed {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1020.801808] env[65726]: INFO nova.scheduler.client.report [None req-bd9cdf9a-c493-4e4b-996c-7dba718501da tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Deleted allocations for instance 067c9332-fa2c-47fe-93ae-643de1fe77cf [ 1020.856218] env[65726]: DEBUG oslo_vmware.api [None req-59619dfd-192b-4bfb-9fd7-b4adaf4a925f tempest-ServerMetadataTestJSON-637854735 tempest-ServerMetadataTestJSON-637854735-project-member] Task: {'id': task-5116502, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.031240] env[65726]: INFO nova.compute.manager [None req-a27b5bca-2338-4542-8dd7-9cac7e4f3b3a tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] [instance: a286ba4f-a83e-4c30-8079-ba88e3ba5696] Took 22.28 seconds to build instance. [ 1021.200338] env[65726]: DEBUG nova.network.neutron [-] [instance: dcde25df-87a5-47fb-94ce-334f68894e04] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1021.308363] env[65726]: DEBUG oslo_concurrency.lockutils [None req-bd9cdf9a-c493-4e4b-996c-7dba718501da tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1021.357103] env[65726]: DEBUG oslo_vmware.api [None req-59619dfd-192b-4bfb-9fd7-b4adaf4a925f tempest-ServerMetadataTestJSON-637854735 tempest-ServerMetadataTestJSON-637854735-project-member] Task: {'id': task-5116502, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.492794] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11291a49-9077-44a5-bb1e-804fb54ad735 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.501092] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17e04f8d-3325-4cd9-89d3-6065777a8d1c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.506859] env[65726]: DEBUG nova.network.neutron [None req-5f1cb876-8e7c-4680-9e78-4abda3ac2850 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: fb8eac12-aedd-426f-9dd3-6c52f9d4c71d] Updating instance_info_cache with network_info: [{"id": "839bf297-99c8-450e-9cd5-546932e1bea1", "address": "fa:16:3e:3d:2a:cb", "network": {"id": "9dd55fb6-0dd2-484d-b94d-fbdccb958c79", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1005598846-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "f93d852e2c904f42981cbfff882558e9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "74f30339-6421-4654-bddb-81d7f34db9d7", "external-id": "nsx-vlan-transportzone-899", "segmentation_id": 899, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap839bf297-99", "ovs_interfaceid": "839bf297-99c8-450e-9cd5-546932e1bea1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1021.537727] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a27b5bca-2338-4542-8dd7-9cac7e4f3b3a tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Lock "a286ba4f-a83e-4c30-8079-ba88e3ba5696" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 23.820s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1021.539557] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a3ec4fb-6d61-41af-ba11-33b73e5b4f99 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.548627] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c52b6822-ef66-496a-b143-9f4f833f705e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.569954] env[65726]: DEBUG nova.compute.provider_tree [None req-1d2791c0-c341-4061-b914-188004153713 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1021.705165] env[65726]: INFO nova.compute.manager [-] [instance: dcde25df-87a5-47fb-94ce-334f68894e04] Took 1.87 seconds to deallocate network for instance. [ 1021.852152] env[65726]: DEBUG oslo_vmware.api [None req-59619dfd-192b-4bfb-9fd7-b4adaf4a925f tempest-ServerMetadataTestJSON-637854735 tempest-ServerMetadataTestJSON-637854735-project-member] Task: {'id': task-5116502, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.899509] env[65726]: DEBUG nova.network.neutron [-] [instance: cf950ec3-9914-4eb9-99db-048b3969bb21] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1022.011574] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5f1cb876-8e7c-4680-9e78-4abda3ac2850 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Releasing lock "refresh_cache-fb8eac12-aedd-426f-9dd3-6c52f9d4c71d" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1022.011574] env[65726]: DEBUG nova.compute.manager [None req-5f1cb876-8e7c-4680-9e78-4abda3ac2850 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: fb8eac12-aedd-426f-9dd3-6c52f9d4c71d] Instance network_info: |[{"id": "839bf297-99c8-450e-9cd5-546932e1bea1", "address": "fa:16:3e:3d:2a:cb", "network": {"id": "9dd55fb6-0dd2-484d-b94d-fbdccb958c79", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1005598846-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "f93d852e2c904f42981cbfff882558e9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "74f30339-6421-4654-bddb-81d7f34db9d7", "external-id": "nsx-vlan-transportzone-899", "segmentation_id": 899, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap839bf297-99", "ovs_interfaceid": "839bf297-99c8-450e-9cd5-546932e1bea1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1022.011574] env[65726]: DEBUG oslo_concurrency.lockutils [req-6ba1a87e-aec7-4177-baba-8f351c1b93f6 req-8ff9acf5-89e9-42fd-9978-c56fc3740918 service nova] Acquired lock "refresh_cache-fb8eac12-aedd-426f-9dd3-6c52f9d4c71d" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1022.011574] env[65726]: DEBUG nova.network.neutron [req-6ba1a87e-aec7-4177-baba-8f351c1b93f6 req-8ff9acf5-89e9-42fd-9978-c56fc3740918 service nova] [instance: fb8eac12-aedd-426f-9dd3-6c52f9d4c71d] Refreshing network info cache for port 839bf297-99c8-450e-9cd5-546932e1bea1 {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 1022.012418] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-5f1cb876-8e7c-4680-9e78-4abda3ac2850 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: fb8eac12-aedd-426f-9dd3-6c52f9d4c71d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:3d:2a:cb', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '74f30339-6421-4654-bddb-81d7f34db9d7', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '839bf297-99c8-450e-9cd5-546932e1bea1', 'vif_model': 'vmxnet3'}] {{(pid=65726) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1022.023905] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-5f1cb876-8e7c-4680-9e78-4abda3ac2850 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Creating folder: Project (f93d852e2c904f42981cbfff882558e9). Parent ref: group-v995008. {{(pid=65726) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1022.025596] env[65726]: WARNING openstack [req-6ba1a87e-aec7-4177-baba-8f351c1b93f6 req-8ff9acf5-89e9-42fd-9978-c56fc3740918 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1022.025942] env[65726]: WARNING openstack [req-6ba1a87e-aec7-4177-baba-8f351c1b93f6 req-8ff9acf5-89e9-42fd-9978-c56fc3740918 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1022.036832] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b5cc4e03-8710-4ca6-b01f-99e83ce93ffe {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.052027] env[65726]: INFO nova.virt.vmwareapi.vm_util [None req-5f1cb876-8e7c-4680-9e78-4abda3ac2850 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Created folder: Project (f93d852e2c904f42981cbfff882558e9) in parent group-v995008. [ 1022.052027] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-5f1cb876-8e7c-4680-9e78-4abda3ac2850 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Creating folder: Instances. Parent ref: group-v995260. {{(pid=65726) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1022.052027] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-965db4e4-158c-4372-b863-ee14956afbe9 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.062414] env[65726]: INFO nova.virt.vmwareapi.vm_util [None req-5f1cb876-8e7c-4680-9e78-4abda3ac2850 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Created folder: Instances in parent group-v995260. [ 1022.062679] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-5f1cb876-8e7c-4680-9e78-4abda3ac2850 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1022.066020] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fb8eac12-aedd-426f-9dd3-6c52f9d4c71d] Creating VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1022.066020] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-4ece688c-c330-49b6-82ad-1995ae25e408 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.081203] env[65726]: DEBUG nova.scheduler.client.report [None req-1d2791c0-c341-4061-b914-188004153713 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1022.090522] env[65726]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1022.090522] env[65726]: value = "task-5116505" [ 1022.090522] env[65726]: _type = "Task" [ 1022.090522] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1022.101443] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116505, 'name': CreateVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.213411] env[65726]: DEBUG oslo_concurrency.lockutils [None req-7aacd0be-ee58-4b95-8abe-402ca2fa658f tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1022.354448] env[65726]: DEBUG oslo_vmware.api [None req-59619dfd-192b-4bfb-9fd7-b4adaf4a925f tempest-ServerMetadataTestJSON-637854735 tempest-ServerMetadataTestJSON-637854735-project-member] Task: {'id': task-5116502, 'name': ReconfigVM_Task, 'duration_secs': 1.87732} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1022.354822] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-59619dfd-192b-4bfb-9fd7-b4adaf4a925f tempest-ServerMetadataTestJSON-637854735 tempest-ServerMetadataTestJSON-637854735-project-member] [instance: 699127a9-f40f-4c1c-ba4a-625af097350b] Reconfigured VM instance instance-00000056 to attach disk [datastore2] 699127a9-f40f-4c1c-ba4a-625af097350b/699127a9-f40f-4c1c-ba4a-625af097350b.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1022.355612] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-279244e9-2fa9-4c09-8cb3-160691ac95dd {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.365231] env[65726]: DEBUG oslo_vmware.api [None req-59619dfd-192b-4bfb-9fd7-b4adaf4a925f tempest-ServerMetadataTestJSON-637854735 tempest-ServerMetadataTestJSON-637854735-project-member] Waiting for the task: (returnval){ [ 1022.365231] env[65726]: value = "task-5116506" [ 1022.365231] env[65726]: _type = "Task" [ 1022.365231] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1022.375843] env[65726]: DEBUG oslo_vmware.api [None req-59619dfd-192b-4bfb-9fd7-b4adaf4a925f tempest-ServerMetadataTestJSON-637854735 tempest-ServerMetadataTestJSON-637854735-project-member] Task: {'id': task-5116506, 'name': Rename_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.403952] env[65726]: INFO nova.compute.manager [-] [instance: cf950ec3-9914-4eb9-99db-048b3969bb21] Took 2.03 seconds to deallocate network for instance. [ 1022.588601] env[65726]: DEBUG oslo_concurrency.lockutils [None req-1d2791c0-c341-4061-b914-188004153713 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.884s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1022.589687] env[65726]: WARNING openstack [None req-1d2791c0-c341-4061-b914-188004153713 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1022.590074] env[65726]: WARNING openstack [None req-1d2791c0-c341-4061-b914-188004153713 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1022.600787] env[65726]: DEBUG oslo_concurrency.lockutils [None req-12289131-1b6e-4176-844f-57bf1f466d9e tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 6.045s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1022.600787] env[65726]: DEBUG nova.objects.instance [None req-12289131-1b6e-4176-844f-57bf1f466d9e tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Lazy-loading 'resources' on Instance uuid 895797e4-2941-44cd-aab1-67afa4fac02d {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1022.617755] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116505, 'name': CreateVM_Task, 'duration_secs': 0.38364} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1022.618964] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fb8eac12-aedd-426f-9dd3-6c52f9d4c71d] Created VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1022.619719] env[65726]: WARNING openstack [None req-5f1cb876-8e7c-4680-9e78-4abda3ac2850 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1022.620151] env[65726]: WARNING openstack [None req-5f1cb876-8e7c-4680-9e78-4abda3ac2850 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1022.627065] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5f1cb876-8e7c-4680-9e78-4abda3ac2850 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1022.627065] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5f1cb876-8e7c-4680-9e78-4abda3ac2850 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1022.627065] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5f1cb876-8e7c-4680-9e78-4abda3ac2850 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1022.627410] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e8558245-a3a6-4cf5-8686-579b465f163a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.636813] env[65726]: DEBUG oslo_vmware.api [None req-5f1cb876-8e7c-4680-9e78-4abda3ac2850 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Waiting for the task: (returnval){ [ 1022.636813] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5250f492-084f-fff9-f0a9-6f96f1b00382" [ 1022.636813] env[65726]: _type = "Task" [ 1022.636813] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1022.645599] env[65726]: DEBUG oslo_vmware.api [None req-5f1cb876-8e7c-4680-9e78-4abda3ac2850 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5250f492-084f-fff9-f0a9-6f96f1b00382, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.764255] env[65726]: WARNING openstack [req-6ba1a87e-aec7-4177-baba-8f351c1b93f6 req-8ff9acf5-89e9-42fd-9978-c56fc3740918 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1022.764654] env[65726]: WARNING openstack [req-6ba1a87e-aec7-4177-baba-8f351c1b93f6 req-8ff9acf5-89e9-42fd-9978-c56fc3740918 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1022.858901] env[65726]: INFO nova.network.neutron [None req-1d2791c0-c341-4061-b914-188004153713 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04] Updating port 25b9d108-0b43-4459-b9db-7bd90a495bb3 with attributes {'binding_host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 1022.884440] env[65726]: DEBUG oslo_vmware.api [None req-59619dfd-192b-4bfb-9fd7-b4adaf4a925f tempest-ServerMetadataTestJSON-637854735 tempest-ServerMetadataTestJSON-637854735-project-member] Task: {'id': task-5116506, 'name': Rename_Task, 'duration_secs': 0.176269} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1022.884800] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-59619dfd-192b-4bfb-9fd7-b4adaf4a925f tempest-ServerMetadataTestJSON-637854735 tempest-ServerMetadataTestJSON-637854735-project-member] [instance: 699127a9-f40f-4c1c-ba4a-625af097350b] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1022.885139] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b590e988-eb99-42ec-bea8-da274119b75b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.899221] env[65726]: DEBUG oslo_vmware.api [None req-59619dfd-192b-4bfb-9fd7-b4adaf4a925f tempest-ServerMetadataTestJSON-637854735 tempest-ServerMetadataTestJSON-637854735-project-member] Waiting for the task: (returnval){ [ 1022.899221] env[65726]: value = "task-5116507" [ 1022.899221] env[65726]: _type = "Task" [ 1022.899221] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1022.908600] env[65726]: DEBUG oslo_vmware.api [None req-59619dfd-192b-4bfb-9fd7-b4adaf4a925f tempest-ServerMetadataTestJSON-637854735 tempest-ServerMetadataTestJSON-637854735-project-member] Task: {'id': task-5116507, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.911274] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a8820945-d2e3-406b-a110-724bc285e8cb tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Acquiring lock "aa6ce489-c62f-4481-87b7-e74242aeb8ca" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1022.911512] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a8820945-d2e3-406b-a110-724bc285e8cb tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Lock "aa6ce489-c62f-4481-87b7-e74242aeb8ca" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1022.920995] env[65726]: DEBUG oslo_concurrency.lockutils [None req-345ea84c-77bf-4393-9db3-8f6989cf9c1c tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1023.090326] env[65726]: WARNING openstack [req-6ba1a87e-aec7-4177-baba-8f351c1b93f6 req-8ff9acf5-89e9-42fd-9978-c56fc3740918 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1023.090326] env[65726]: WARNING openstack [req-6ba1a87e-aec7-4177-baba-8f351c1b93f6 req-8ff9acf5-89e9-42fd-9978-c56fc3740918 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1023.155042] env[65726]: DEBUG oslo_vmware.api [None req-5f1cb876-8e7c-4680-9e78-4abda3ac2850 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5250f492-084f-fff9-f0a9-6f96f1b00382, 'name': SearchDatastore_Task, 'duration_secs': 0.011654} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1023.155042] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5f1cb876-8e7c-4680-9e78-4abda3ac2850 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1023.155042] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-5f1cb876-8e7c-4680-9e78-4abda3ac2850 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: fb8eac12-aedd-426f-9dd3-6c52f9d4c71d] Processing image b52362a3-ee8a-4cbf-b06f-513b0cc8f95c {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1023.155554] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5f1cb876-8e7c-4680-9e78-4abda3ac2850 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1023.155554] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5f1cb876-8e7c-4680-9e78-4abda3ac2850 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1023.155554] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-5f1cb876-8e7c-4680-9e78-4abda3ac2850 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1023.155777] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6853a048-535a-4473-8afa-f892d12c3f46 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.166308] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-5f1cb876-8e7c-4680-9e78-4abda3ac2850 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1023.166545] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-5f1cb876-8e7c-4680-9e78-4abda3ac2850 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65726) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1023.167320] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8f82aff1-d2f4-4371-8f04-a9d6d0f93fb9 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.175999] env[65726]: DEBUG oslo_vmware.api [None req-5f1cb876-8e7c-4680-9e78-4abda3ac2850 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Waiting for the task: (returnval){ [ 1023.175999] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52662ee1-9127-48b7-9ab2-d91bdaf4d478" [ 1023.175999] env[65726]: _type = "Task" [ 1023.175999] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1023.187807] env[65726]: DEBUG oslo_vmware.api [None req-5f1cb876-8e7c-4680-9e78-4abda3ac2850 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52662ee1-9127-48b7-9ab2-d91bdaf4d478, 'name': SearchDatastore_Task, 'duration_secs': 0.010185} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1023.191567] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c817d894-fa9f-4f42-964b-4d0c6fe41bb5 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.199283] env[65726]: DEBUG oslo_vmware.api [None req-5f1cb876-8e7c-4680-9e78-4abda3ac2850 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Waiting for the task: (returnval){ [ 1023.199283] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]523ca3c6-d61c-569e-07cb-9081c10dca4f" [ 1023.199283] env[65726]: _type = "Task" [ 1023.199283] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1023.214325] env[65726]: DEBUG oslo_vmware.api [None req-5f1cb876-8e7c-4680-9e78-4abda3ac2850 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]523ca3c6-d61c-569e-07cb-9081c10dca4f, 'name': SearchDatastore_Task, 'duration_secs': 0.011719} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1023.214601] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5f1cb876-8e7c-4680-9e78-4abda3ac2850 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1023.215310] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-5f1cb876-8e7c-4680-9e78-4abda3ac2850 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore2] fb8eac12-aedd-426f-9dd3-6c52f9d4c71d/fb8eac12-aedd-426f-9dd3-6c52f9d4c71d.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1023.215426] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2c986ed4-1df4-4c32-9e54-56cd9126beab {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.226010] env[65726]: DEBUG oslo_vmware.api [None req-5f1cb876-8e7c-4680-9e78-4abda3ac2850 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Waiting for the task: (returnval){ [ 1023.226010] env[65726]: value = "task-5116508" [ 1023.226010] env[65726]: _type = "Task" [ 1023.226010] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1023.239557] env[65726]: DEBUG oslo_vmware.api [None req-5f1cb876-8e7c-4680-9e78-4abda3ac2850 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': task-5116508, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.357287] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc231e29-2f36-48d6-96e9-8c2831cfa25f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.366693] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a375401-7af8-47fb-93ac-53398186b446 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.404891] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-807102d3-0aff-4fc1-905f-33541a8c5684 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.417515] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eab01163-8a8b-4cab-8727-2cd968db0574 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.422677] env[65726]: DEBUG nova.compute.utils [None req-a8820945-d2e3-406b-a110-724bc285e8cb tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Using /dev/sd instead of None {{(pid=65726) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1023.424424] env[65726]: DEBUG oslo_vmware.api [None req-59619dfd-192b-4bfb-9fd7-b4adaf4a925f tempest-ServerMetadataTestJSON-637854735 tempest-ServerMetadataTestJSON-637854735-project-member] Task: {'id': task-5116507, 'name': PowerOnVM_Task} progress is 94%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.437481] env[65726]: DEBUG nova.compute.provider_tree [None req-12289131-1b6e-4176-844f-57bf1f466d9e tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1023.653278] env[65726]: DEBUG nova.network.neutron [req-6ba1a87e-aec7-4177-baba-8f351c1b93f6 req-8ff9acf5-89e9-42fd-9978-c56fc3740918 service nova] [instance: fb8eac12-aedd-426f-9dd3-6c52f9d4c71d] Updated VIF entry in instance network info cache for port 839bf297-99c8-450e-9cd5-546932e1bea1. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 1023.653694] env[65726]: DEBUG nova.network.neutron [req-6ba1a87e-aec7-4177-baba-8f351c1b93f6 req-8ff9acf5-89e9-42fd-9978-c56fc3740918 service nova] [instance: fb8eac12-aedd-426f-9dd3-6c52f9d4c71d] Updating instance_info_cache with network_info: [{"id": "839bf297-99c8-450e-9cd5-546932e1bea1", "address": "fa:16:3e:3d:2a:cb", "network": {"id": "9dd55fb6-0dd2-484d-b94d-fbdccb958c79", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1005598846-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "f93d852e2c904f42981cbfff882558e9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "74f30339-6421-4654-bddb-81d7f34db9d7", "external-id": "nsx-vlan-transportzone-899", "segmentation_id": 899, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap839bf297-99", "ovs_interfaceid": "839bf297-99c8-450e-9cd5-546932e1bea1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1023.739504] env[65726]: DEBUG oslo_vmware.api [None req-5f1cb876-8e7c-4680-9e78-4abda3ac2850 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': task-5116508, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.912643] env[65726]: DEBUG oslo_vmware.api [None req-59619dfd-192b-4bfb-9fd7-b4adaf4a925f tempest-ServerMetadataTestJSON-637854735 tempest-ServerMetadataTestJSON-637854735-project-member] Task: {'id': task-5116507, 'name': PowerOnVM_Task, 'duration_secs': 0.609745} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1023.912945] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-59619dfd-192b-4bfb-9fd7-b4adaf4a925f tempest-ServerMetadataTestJSON-637854735 tempest-ServerMetadataTestJSON-637854735-project-member] [instance: 699127a9-f40f-4c1c-ba4a-625af097350b] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1023.913170] env[65726]: INFO nova.compute.manager [None req-59619dfd-192b-4bfb-9fd7-b4adaf4a925f tempest-ServerMetadataTestJSON-637854735 tempest-ServerMetadataTestJSON-637854735-project-member] [instance: 699127a9-f40f-4c1c-ba4a-625af097350b] Took 10.54 seconds to spawn the instance on the hypervisor. [ 1023.913350] env[65726]: DEBUG nova.compute.manager [None req-59619dfd-192b-4bfb-9fd7-b4adaf4a925f tempest-ServerMetadataTestJSON-637854735 tempest-ServerMetadataTestJSON-637854735-project-member] [instance: 699127a9-f40f-4c1c-ba4a-625af097350b] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1023.914338] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71291560-70cf-4526-8cbf-08fc4b546408 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.925904] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a8820945-d2e3-406b-a110-724bc285e8cb tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Lock "aa6ce489-c62f-4481-87b7-e74242aeb8ca" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.014s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1023.941304] env[65726]: DEBUG nova.scheduler.client.report [None req-12289131-1b6e-4176-844f-57bf1f466d9e tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1023.975828] env[65726]: DEBUG nova.compute.manager [req-99bd2908-a271-4c68-86a6-f6f641fde043 req-ceecc245-aa64-418d-9745-7a991ba5a98d service nova] [instance: 067c9332-fa2c-47fe-93ae-643de1fe77cf] Received event network-changed-498a4766-79c8-471b-b6fc-e73c9eca5c5c {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1023.975970] env[65726]: DEBUG nova.compute.manager [req-99bd2908-a271-4c68-86a6-f6f641fde043 req-ceecc245-aa64-418d-9745-7a991ba5a98d service nova] [instance: 067c9332-fa2c-47fe-93ae-643de1fe77cf] Refreshing instance network info cache due to event network-changed-498a4766-79c8-471b-b6fc-e73c9eca5c5c. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 1023.976403] env[65726]: DEBUG oslo_concurrency.lockutils [req-99bd2908-a271-4c68-86a6-f6f641fde043 req-ceecc245-aa64-418d-9745-7a991ba5a98d service nova] Acquiring lock "refresh_cache-067c9332-fa2c-47fe-93ae-643de1fe77cf" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1023.976403] env[65726]: DEBUG oslo_concurrency.lockutils [req-99bd2908-a271-4c68-86a6-f6f641fde043 req-ceecc245-aa64-418d-9745-7a991ba5a98d service nova] Acquired lock "refresh_cache-067c9332-fa2c-47fe-93ae-643de1fe77cf" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1023.976561] env[65726]: DEBUG nova.network.neutron [req-99bd2908-a271-4c68-86a6-f6f641fde043 req-ceecc245-aa64-418d-9745-7a991ba5a98d service nova] [instance: 067c9332-fa2c-47fe-93ae-643de1fe77cf] Refreshing network info cache for port 498a4766-79c8-471b-b6fc-e73c9eca5c5c {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 1024.156754] env[65726]: DEBUG oslo_concurrency.lockutils [req-6ba1a87e-aec7-4177-baba-8f351c1b93f6 req-8ff9acf5-89e9-42fd-9978-c56fc3740918 service nova] Releasing lock "refresh_cache-fb8eac12-aedd-426f-9dd3-6c52f9d4c71d" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1024.240845] env[65726]: DEBUG oslo_vmware.api [None req-5f1cb876-8e7c-4680-9e78-4abda3ac2850 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': task-5116508, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.53681} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1024.241155] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-5f1cb876-8e7c-4680-9e78-4abda3ac2850 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore2] fb8eac12-aedd-426f-9dd3-6c52f9d4c71d/fb8eac12-aedd-426f-9dd3-6c52f9d4c71d.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1024.241431] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-5f1cb876-8e7c-4680-9e78-4abda3ac2850 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: fb8eac12-aedd-426f-9dd3-6c52f9d4c71d] Extending root virtual disk to 1048576 {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1024.241968] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2f541a00-c933-4b23-9ea6-103213285e60 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.250498] env[65726]: DEBUG oslo_vmware.api [None req-5f1cb876-8e7c-4680-9e78-4abda3ac2850 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Waiting for the task: (returnval){ [ 1024.250498] env[65726]: value = "task-5116509" [ 1024.250498] env[65726]: _type = "Task" [ 1024.250498] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.260575] env[65726]: DEBUG oslo_vmware.api [None req-5f1cb876-8e7c-4680-9e78-4abda3ac2850 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': task-5116509, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.437942] env[65726]: INFO nova.compute.manager [None req-59619dfd-192b-4bfb-9fd7-b4adaf4a925f tempest-ServerMetadataTestJSON-637854735 tempest-ServerMetadataTestJSON-637854735-project-member] [instance: 699127a9-f40f-4c1c-ba4a-625af097350b] Took 18.88 seconds to build instance. [ 1024.447200] env[65726]: DEBUG oslo_concurrency.lockutils [None req-12289131-1b6e-4176-844f-57bf1f466d9e tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.847s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1024.449260] env[65726]: DEBUG oslo_concurrency.lockutils [None req-f1e7868e-b3d9-4342-8972-f27e4cb4ec45 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.333s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1024.451051] env[65726]: INFO nova.compute.claims [None req-f1e7868e-b3d9-4342-8972-f27e4cb4ec45 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] [instance: 2b8eb60b-dc9b-47a4-9ddc-205f2af69888] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1024.483193] env[65726]: WARNING openstack [req-99bd2908-a271-4c68-86a6-f6f641fde043 req-ceecc245-aa64-418d-9745-7a991ba5a98d service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1024.483645] env[65726]: WARNING openstack [req-99bd2908-a271-4c68-86a6-f6f641fde043 req-ceecc245-aa64-418d-9745-7a991ba5a98d service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1024.493102] env[65726]: INFO nova.scheduler.client.report [None req-12289131-1b6e-4176-844f-57bf1f466d9e tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Deleted allocations for instance 895797e4-2941-44cd-aab1-67afa4fac02d [ 1024.529747] env[65726]: DEBUG oslo_concurrency.lockutils [None req-1d2791c0-c341-4061-b914-188004153713 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Acquiring lock "refresh_cache-2fe6fc5a-f5c7-4f8e-96df-4e621a252f04" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1024.529747] env[65726]: DEBUG oslo_concurrency.lockutils [None req-1d2791c0-c341-4061-b914-188004153713 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Acquired lock "refresh_cache-2fe6fc5a-f5c7-4f8e-96df-4e621a252f04" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1024.529747] env[65726]: DEBUG nova.network.neutron [None req-1d2791c0-c341-4061-b914-188004153713 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1024.764823] env[65726]: DEBUG oslo_vmware.api [None req-5f1cb876-8e7c-4680-9e78-4abda3ac2850 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': task-5116509, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.079052} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1024.765968] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-5f1cb876-8e7c-4680-9e78-4abda3ac2850 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: fb8eac12-aedd-426f-9dd3-6c52f9d4c71d] Extended root virtual disk {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1024.767145] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b28bb811-1160-4b1a-b29e-ef9ff6fe2bb8 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.803992] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-5f1cb876-8e7c-4680-9e78-4abda3ac2850 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: fb8eac12-aedd-426f-9dd3-6c52f9d4c71d] Reconfiguring VM instance instance-00000057 to attach disk [datastore2] fb8eac12-aedd-426f-9dd3-6c52f9d4c71d/fb8eac12-aedd-426f-9dd3-6c52f9d4c71d.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1024.804619] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9a28dc6f-0cbb-4337-a8b9-45bb45881cc8 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.826540] env[65726]: DEBUG oslo_vmware.api [None req-5f1cb876-8e7c-4680-9e78-4abda3ac2850 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Waiting for the task: (returnval){ [ 1024.826540] env[65726]: value = "task-5116510" [ 1024.826540] env[65726]: _type = "Task" [ 1024.826540] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.835877] env[65726]: DEBUG oslo_vmware.api [None req-5f1cb876-8e7c-4680-9e78-4abda3ac2850 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': task-5116510, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.943572] env[65726]: DEBUG oslo_concurrency.lockutils [None req-59619dfd-192b-4bfb-9fd7-b4adaf4a925f tempest-ServerMetadataTestJSON-637854735 tempest-ServerMetadataTestJSON-637854735-project-member] Lock "699127a9-f40f-4c1c-ba4a-625af097350b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 20.398s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1025.003461] env[65726]: DEBUG oslo_concurrency.lockutils [None req-12289131-1b6e-4176-844f-57bf1f466d9e tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Lock "895797e4-2941-44cd-aab1-67afa4fac02d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 13.426s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1025.008109] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a8820945-d2e3-406b-a110-724bc285e8cb tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Acquiring lock "aa6ce489-c62f-4481-87b7-e74242aeb8ca" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1025.009420] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a8820945-d2e3-406b-a110-724bc285e8cb tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Lock "aa6ce489-c62f-4481-87b7-e74242aeb8ca" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1025.009420] env[65726]: INFO nova.compute.manager [None req-a8820945-d2e3-406b-a110-724bc285e8cb tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] [instance: aa6ce489-c62f-4481-87b7-e74242aeb8ca] Attaching volume 3547dfa9-73e6-4baa-b777-e22891db153f to /dev/sdb [ 1025.039024] env[65726]: WARNING openstack [None req-1d2791c0-c341-4061-b914-188004153713 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1025.039024] env[65726]: WARNING openstack [None req-1d2791c0-c341-4061-b914-188004153713 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1025.053038] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3f4ecee-bd57-472a-a441-cbb4e5e6be89 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.059798] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4c7b654-7de7-4836-9322-46fac71b69b1 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.077215] env[65726]: DEBUG nova.virt.block_device [None req-a8820945-d2e3-406b-a110-724bc285e8cb tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] [instance: aa6ce489-c62f-4481-87b7-e74242aeb8ca] Updating existing volume attachment record: 7e52c97b-5136-4824-a8f7-a37ca4548c09 {{(pid=65726) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1025.341338] env[65726]: DEBUG oslo_vmware.api [None req-5f1cb876-8e7c-4680-9e78-4abda3ac2850 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': task-5116510, 'name': ReconfigVM_Task, 'duration_secs': 0.298735} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1025.341666] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-5f1cb876-8e7c-4680-9e78-4abda3ac2850 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: fb8eac12-aedd-426f-9dd3-6c52f9d4c71d] Reconfigured VM instance instance-00000057 to attach disk [datastore2] fb8eac12-aedd-426f-9dd3-6c52f9d4c71d/fb8eac12-aedd-426f-9dd3-6c52f9d4c71d.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1025.342482] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9595c7de-037e-4f5d-ac5f-017e53a80f45 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.354728] env[65726]: DEBUG oslo_vmware.api [None req-5f1cb876-8e7c-4680-9e78-4abda3ac2850 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Waiting for the task: (returnval){ [ 1025.354728] env[65726]: value = "task-5116513" [ 1025.354728] env[65726]: _type = "Task" [ 1025.354728] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1025.368911] env[65726]: DEBUG oslo_vmware.api [None req-5f1cb876-8e7c-4680-9e78-4abda3ac2850 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': task-5116513, 'name': Rename_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.594017] env[65726]: WARNING openstack [req-99bd2908-a271-4c68-86a6-f6f641fde043 req-ceecc245-aa64-418d-9745-7a991ba5a98d service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1025.594607] env[65726]: WARNING openstack [req-99bd2908-a271-4c68-86a6-f6f641fde043 req-ceecc245-aa64-418d-9745-7a991ba5a98d service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1025.726821] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-849ddc3f-c178-4aa3-8cf9-595cb998161d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.737805] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bec29316-05dd-4d66-b5b2-32b3f03eda78 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.781047] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5812a539-606b-4e41-9a6e-dcbb87ee56ed {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.790101] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5fb2398e-341e-4fa4-89df-dd261dab1505 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.806371] env[65726]: DEBUG nova.compute.provider_tree [None req-f1e7868e-b3d9-4342-8972-f27e4cb4ec45 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1025.867362] env[65726]: DEBUG oslo_vmware.api [None req-5f1cb876-8e7c-4680-9e78-4abda3ac2850 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': task-5116513, 'name': Rename_Task, 'duration_secs': 0.214173} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1025.867695] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-5f1cb876-8e7c-4680-9e78-4abda3ac2850 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: fb8eac12-aedd-426f-9dd3-6c52f9d4c71d] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1025.867976] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-557eea93-741d-4a61-9077-a52d9fb5cff6 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.876918] env[65726]: DEBUG oslo_vmware.api [None req-5f1cb876-8e7c-4680-9e78-4abda3ac2850 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Waiting for the task: (returnval){ [ 1025.876918] env[65726]: value = "task-5116515" [ 1025.876918] env[65726]: _type = "Task" [ 1025.876918] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1025.887331] env[65726]: DEBUG oslo_vmware.api [None req-5f1cb876-8e7c-4680-9e78-4abda3ac2850 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': task-5116515, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.192867] env[65726]: DEBUG nova.compute.manager [req-cd57fd15-0826-4f69-a60c-6a8a9b83af8c req-93e2b9e8-c9f2-4563-ab0d-96810eedbf18 service nova] [instance: 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04] Received event network-vif-plugged-25b9d108-0b43-4459-b9db-7bd90a495bb3 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1026.193848] env[65726]: DEBUG oslo_concurrency.lockutils [req-cd57fd15-0826-4f69-a60c-6a8a9b83af8c req-93e2b9e8-c9f2-4563-ab0d-96810eedbf18 service nova] Acquiring lock "2fe6fc5a-f5c7-4f8e-96df-4e621a252f04-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1026.193848] env[65726]: DEBUG oslo_concurrency.lockutils [req-cd57fd15-0826-4f69-a60c-6a8a9b83af8c req-93e2b9e8-c9f2-4563-ab0d-96810eedbf18 service nova] Lock "2fe6fc5a-f5c7-4f8e-96df-4e621a252f04-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1026.193848] env[65726]: DEBUG oslo_concurrency.lockutils [req-cd57fd15-0826-4f69-a60c-6a8a9b83af8c req-93e2b9e8-c9f2-4563-ab0d-96810eedbf18 service nova] Lock "2fe6fc5a-f5c7-4f8e-96df-4e621a252f04-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1026.193848] env[65726]: DEBUG nova.compute.manager [req-cd57fd15-0826-4f69-a60c-6a8a9b83af8c req-93e2b9e8-c9f2-4563-ab0d-96810eedbf18 service nova] [instance: 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04] No waiting events found dispatching network-vif-plugged-25b9d108-0b43-4459-b9db-7bd90a495bb3 {{(pid=65726) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1026.193848] env[65726]: WARNING nova.compute.manager [req-cd57fd15-0826-4f69-a60c-6a8a9b83af8c req-93e2b9e8-c9f2-4563-ab0d-96810eedbf18 service nova] [instance: 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04] Received unexpected event network-vif-plugged-25b9d108-0b43-4459-b9db-7bd90a495bb3 for instance with vm_state shelved_offloaded and task_state spawning. [ 1026.309701] env[65726]: DEBUG nova.scheduler.client.report [None req-f1e7868e-b3d9-4342-8972-f27e4cb4ec45 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1026.388230] env[65726]: DEBUG oslo_vmware.api [None req-5f1cb876-8e7c-4680-9e78-4abda3ac2850 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': task-5116515, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.645553] env[65726]: DEBUG oslo_concurrency.lockutils [None req-8ffe452d-09ce-47ef-a349-6236507a111f tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Acquiring lock "067c9332-fa2c-47fe-93ae-643de1fe77cf" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1026.683561] env[65726]: WARNING openstack [req-99bd2908-a271-4c68-86a6-f6f641fde043 req-ceecc245-aa64-418d-9745-7a991ba5a98d service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1026.683986] env[65726]: WARNING openstack [req-99bd2908-a271-4c68-86a6-f6f641fde043 req-ceecc245-aa64-418d-9745-7a991ba5a98d service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1026.723302] env[65726]: WARNING openstack [None req-1d2791c0-c341-4061-b914-188004153713 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1026.723751] env[65726]: WARNING openstack [None req-1d2791c0-c341-4061-b914-188004153713 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1026.816654] env[65726]: DEBUG oslo_concurrency.lockutils [None req-f1e7868e-b3d9-4342-8972-f27e4cb4ec45 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.367s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1026.817463] env[65726]: DEBUG nova.compute.manager [None req-f1e7868e-b3d9-4342-8972-f27e4cb4ec45 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] [instance: 2b8eb60b-dc9b-47a4-9ddc-205f2af69888] Start building networks asynchronously for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1026.820708] env[65726]: DEBUG oslo_concurrency.lockutils [None req-2a9dd832-d188-4542-a7fe-bbdea8173b2d tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.073s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1026.824492] env[65726]: INFO nova.compute.claims [None req-2a9dd832-d188-4542-a7fe-bbdea8173b2d tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] [instance: 2c812714-dccc-4d1e-bdb5-c11e446949c3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1026.897760] env[65726]: DEBUG oslo_vmware.api [None req-5f1cb876-8e7c-4680-9e78-4abda3ac2850 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': task-5116515, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.909821] env[65726]: DEBUG nova.network.neutron [req-99bd2908-a271-4c68-86a6-f6f641fde043 req-ceecc245-aa64-418d-9745-7a991ba5a98d service nova] [instance: 067c9332-fa2c-47fe-93ae-643de1fe77cf] Updated VIF entry in instance network info cache for port 498a4766-79c8-471b-b6fc-e73c9eca5c5c. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 1026.910142] env[65726]: DEBUG nova.network.neutron [req-99bd2908-a271-4c68-86a6-f6f641fde043 req-ceecc245-aa64-418d-9745-7a991ba5a98d service nova] [instance: 067c9332-fa2c-47fe-93ae-643de1fe77cf] Updating instance_info_cache with network_info: [{"id": "498a4766-79c8-471b-b6fc-e73c9eca5c5c", "address": "fa:16:3e:fc:03:ad", "network": {"id": "4c972d9d-db3b-4779-ac39-39d9a2d141dd", "bridge": null, "label": "tempest-DeleteServersTestJSON-1418913861-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cbb17a09b35c4c22ade5c4082b10bda9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap498a4766-79", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1026.947903] env[65726]: WARNING openstack [None req-1d2791c0-c341-4061-b914-188004153713 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1026.948421] env[65726]: WARNING openstack [None req-1d2791c0-c341-4061-b914-188004153713 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1027.071879] env[65726]: DEBUG nova.network.neutron [None req-1d2791c0-c341-4061-b914-188004153713 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04] Updating instance_info_cache with network_info: [{"id": "25b9d108-0b43-4459-b9db-7bd90a495bb3", "address": "fa:16:3e:5a:1e:6b", "network": {"id": "eb3278b7-c190-441b-84d2-305f9b896382", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1754571022-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.149", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "96149159e18e44f9bf3453e67681f224", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f77ff7a1-209c-4f3f-b2a0-fd817741e739", "external-id": "nsx-vlan-transportzone-935", "segmentation_id": 935, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap25b9d108-0b", "ovs_interfaceid": "25b9d108-0b43-4459-b9db-7bd90a495bb3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1027.243726] env[65726]: DEBUG nova.compute.manager [None req-9ee0c86a-59b9-4c49-9232-0da097c0fa28 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] [instance: a286ba4f-a83e-4c30-8079-ba88e3ba5696] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1027.244744] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef59a27f-d3b6-4ad1-9fc1-4842f39819dc {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.323553] env[65726]: DEBUG nova.compute.utils [None req-f1e7868e-b3d9-4342-8972-f27e4cb4ec45 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] Using /dev/sd instead of None {{(pid=65726) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1027.325494] env[65726]: DEBUG nova.compute.manager [None req-f1e7868e-b3d9-4342-8972-f27e4cb4ec45 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] [instance: 2b8eb60b-dc9b-47a4-9ddc-205f2af69888] Allocating IP information in the background. {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1027.325938] env[65726]: DEBUG nova.network.neutron [None req-f1e7868e-b3d9-4342-8972-f27e4cb4ec45 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] [instance: 2b8eb60b-dc9b-47a4-9ddc-205f2af69888] allocate_for_instance() {{(pid=65726) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 1027.326392] env[65726]: WARNING neutronclient.v2_0.client [None req-f1e7868e-b3d9-4342-8972-f27e4cb4ec45 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1027.328324] env[65726]: WARNING neutronclient.v2_0.client [None req-f1e7868e-b3d9-4342-8972-f27e4cb4ec45 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1027.328324] env[65726]: WARNING openstack [None req-f1e7868e-b3d9-4342-8972-f27e4cb4ec45 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1027.328665] env[65726]: WARNING openstack [None req-f1e7868e-b3d9-4342-8972-f27e4cb4ec45 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1027.391076] env[65726]: DEBUG oslo_vmware.api [None req-5f1cb876-8e7c-4680-9e78-4abda3ac2850 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': task-5116515, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.418025] env[65726]: DEBUG oslo_concurrency.lockutils [req-99bd2908-a271-4c68-86a6-f6f641fde043 req-ceecc245-aa64-418d-9745-7a991ba5a98d service nova] Releasing lock "refresh_cache-067c9332-fa2c-47fe-93ae-643de1fe77cf" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1027.418025] env[65726]: DEBUG nova.compute.manager [req-99bd2908-a271-4c68-86a6-f6f641fde043 req-ceecc245-aa64-418d-9745-7a991ba5a98d service nova] [instance: dcde25df-87a5-47fb-94ce-334f68894e04] Received event network-vif-deleted-bf51c96c-63c5-46de-a40f-58e6769c5fcf {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1027.418025] env[65726]: DEBUG nova.compute.manager [req-99bd2908-a271-4c68-86a6-f6f641fde043 req-ceecc245-aa64-418d-9745-7a991ba5a98d service nova] [instance: cf950ec3-9914-4eb9-99db-048b3969bb21] Received event network-vif-deleted-58accf41-587f-4367-854e-fc571a6a3424 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1027.429073] env[65726]: DEBUG nova.policy [None req-f1e7868e-b3d9-4342-8972-f27e4cb4ec45 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'cceeaf88846a47a7a9a47d301a1b1870', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '878eef0927ed47dbb370522e93a5453f', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65726) authorize /opt/stack/nova/nova/policy.py:192}} [ 1027.577804] env[65726]: DEBUG oslo_concurrency.lockutils [None req-1d2791c0-c341-4061-b914-188004153713 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Releasing lock "refresh_cache-2fe6fc5a-f5c7-4f8e-96df-4e621a252f04" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1027.608662] env[65726]: DEBUG nova.virt.hardware [None req-1d2791c0-c341-4061-b914-188004153713 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='c47319a54754a76a2ae8cc1ca0c8c5f5',container_format='bare',created_at=2025-12-12T19:37:13Z,direct_url=,disk_format='vmdk',id=81f5ce39-0bc0-4d0e-ab15-42a80a4f614b,min_disk=1,min_ram=0,name='tempest-AttachVolumeShelveTestJSON-server-867361236-shelved',owner='96149159e18e44f9bf3453e67681f224',properties=ImageMetaProps,protected=,size=31673344,status='active',tags=,updated_at=2025-12-12T19:37:30Z,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1027.609436] env[65726]: DEBUG nova.virt.hardware [None req-1d2791c0-c341-4061-b914-188004153713 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1027.609436] env[65726]: DEBUG nova.virt.hardware [None req-1d2791c0-c341-4061-b914-188004153713 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1027.609640] env[65726]: DEBUG nova.virt.hardware [None req-1d2791c0-c341-4061-b914-188004153713 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1027.609727] env[65726]: DEBUG nova.virt.hardware [None req-1d2791c0-c341-4061-b914-188004153713 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1027.609927] env[65726]: DEBUG nova.virt.hardware [None req-1d2791c0-c341-4061-b914-188004153713 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1027.610196] env[65726]: DEBUG nova.virt.hardware [None req-1d2791c0-c341-4061-b914-188004153713 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1027.610398] env[65726]: DEBUG nova.virt.hardware [None req-1d2791c0-c341-4061-b914-188004153713 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1027.610609] env[65726]: DEBUG nova.virt.hardware [None req-1d2791c0-c341-4061-b914-188004153713 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1027.610832] env[65726]: DEBUG nova.virt.hardware [None req-1d2791c0-c341-4061-b914-188004153713 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1027.611082] env[65726]: DEBUG nova.virt.hardware [None req-1d2791c0-c341-4061-b914-188004153713 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1027.612212] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4bf6fd69-45e5-4871-ac97-7ff3e0c34045 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.623214] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57e081fa-04ca-4f88-9488-78f1621071a8 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.640582] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-1d2791c0-c341-4061-b914-188004153713 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:5a:1e:6b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f77ff7a1-209c-4f3f-b2a0-fd817741e739', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '25b9d108-0b43-4459-b9db-7bd90a495bb3', 'vif_model': 'vmxnet3'}] {{(pid=65726) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1027.653351] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-1d2791c0-c341-4061-b914-188004153713 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1027.653789] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04] Creating VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1027.654075] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3a7a09f6-9112-4f1b-9140-774ffd50597e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.688935] env[65726]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1027.688935] env[65726]: value = "task-5116517" [ 1027.688935] env[65726]: _type = "Task" [ 1027.688935] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1027.700268] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116517, 'name': CreateVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.751448] env[65726]: DEBUG nova.network.neutron [None req-f1e7868e-b3d9-4342-8972-f27e4cb4ec45 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] [instance: 2b8eb60b-dc9b-47a4-9ddc-205f2af69888] Successfully created port: f210c8b9-8838-4774-838c-aeb3860a63ca {{(pid=65726) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1027.758552] env[65726]: INFO nova.compute.manager [None req-9ee0c86a-59b9-4c49-9232-0da097c0fa28 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] [instance: a286ba4f-a83e-4c30-8079-ba88e3ba5696] instance snapshotting [ 1027.762246] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7762f59a-9aa8-4390-b4f9-b03d5cea9b5e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.787377] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60d51dc5-f9e1-4aa3-bd27-ce4007c9d4d1 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.840144] env[65726]: DEBUG nova.compute.manager [None req-f1e7868e-b3d9-4342-8972-f27e4cb4ec45 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] [instance: 2b8eb60b-dc9b-47a4-9ddc-205f2af69888] Start building block device mappings for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1027.896528] env[65726]: DEBUG oslo_vmware.api [None req-5f1cb876-8e7c-4680-9e78-4abda3ac2850 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': task-5116515, 'name': PowerOnVM_Task, 'duration_secs': 1.565628} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1027.896778] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-5f1cb876-8e7c-4680-9e78-4abda3ac2850 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: fb8eac12-aedd-426f-9dd3-6c52f9d4c71d] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1027.896950] env[65726]: INFO nova.compute.manager [None req-5f1cb876-8e7c-4680-9e78-4abda3ac2850 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: fb8eac12-aedd-426f-9dd3-6c52f9d4c71d] Took 9.96 seconds to spawn the instance on the hypervisor. [ 1027.897144] env[65726]: DEBUG nova.compute.manager [None req-5f1cb876-8e7c-4680-9e78-4abda3ac2850 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: fb8eac12-aedd-426f-9dd3-6c52f9d4c71d] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1027.900679] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e65ad3a-aca2-4613-82d4-3e1daa9d6789 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.112239] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4dc1f313-cc3f-4b0f-a334-0f4251be54e9 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.120450] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5a816f7-a498-497d-9b1c-368601fd26cc {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.155818] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3c3aa2b-88a1-42b4-9056-da8a56b93f0c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.165309] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ed9eee8-9b11-4fe7-b865-201ff4a0667a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.180993] env[65726]: DEBUG nova.compute.provider_tree [None req-2a9dd832-d188-4542-a7fe-bbdea8173b2d tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1028.202501] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116517, 'name': CreateVM_Task, 'duration_secs': 0.386738} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1028.202501] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04] Created VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1028.203283] env[65726]: WARNING openstack [None req-1d2791c0-c341-4061-b914-188004153713 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1028.203643] env[65726]: WARNING openstack [None req-1d2791c0-c341-4061-b914-188004153713 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1028.210146] env[65726]: DEBUG oslo_concurrency.lockutils [None req-1d2791c0-c341-4061-b914-188004153713 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/81f5ce39-0bc0-4d0e-ab15-42a80a4f614b" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1028.210368] env[65726]: DEBUG oslo_concurrency.lockutils [None req-1d2791c0-c341-4061-b914-188004153713 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Acquired lock "[datastore1] devstack-image-cache_base/81f5ce39-0bc0-4d0e-ab15-42a80a4f614b" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1028.210761] env[65726]: DEBUG oslo_concurrency.lockutils [None req-1d2791c0-c341-4061-b914-188004153713 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/81f5ce39-0bc0-4d0e-ab15-42a80a4f614b" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1028.211464] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ddd27166-aa60-4262-86ba-53599fcd73ea {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.218215] env[65726]: DEBUG oslo_vmware.api [None req-1d2791c0-c341-4061-b914-188004153713 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Waiting for the task: (returnval){ [ 1028.218215] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52784409-dcb6-e09c-fe88-325a12b81ab9" [ 1028.218215] env[65726]: _type = "Task" [ 1028.218215] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1028.229087] env[65726]: DEBUG oslo_vmware.api [None req-1d2791c0-c341-4061-b914-188004153713 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52784409-dcb6-e09c-fe88-325a12b81ab9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.304460] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-9ee0c86a-59b9-4c49-9232-0da097c0fa28 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] [instance: a286ba4f-a83e-4c30-8079-ba88e3ba5696] Creating Snapshot of the VM instance {{(pid=65726) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1028.305131] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-04c7edbe-c412-4391-87ed-28274004086a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.314546] env[65726]: DEBUG oslo_vmware.api [None req-9ee0c86a-59b9-4c49-9232-0da097c0fa28 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Waiting for the task: (returnval){ [ 1028.314546] env[65726]: value = "task-5116518" [ 1028.314546] env[65726]: _type = "Task" [ 1028.314546] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1028.328772] env[65726]: DEBUG oslo_vmware.api [None req-9ee0c86a-59b9-4c49-9232-0da097c0fa28 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Task: {'id': task-5116518, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.422386] env[65726]: INFO nova.compute.manager [None req-5f1cb876-8e7c-4680-9e78-4abda3ac2850 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: fb8eac12-aedd-426f-9dd3-6c52f9d4c71d] Took 18.19 seconds to build instance. [ 1028.685500] env[65726]: DEBUG nova.scheduler.client.report [None req-2a9dd832-d188-4542-a7fe-bbdea8173b2d tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1028.729756] env[65726]: DEBUG oslo_concurrency.lockutils [None req-1d2791c0-c341-4061-b914-188004153713 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Releasing lock "[datastore1] devstack-image-cache_base/81f5ce39-0bc0-4d0e-ab15-42a80a4f614b" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1028.730068] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-1d2791c0-c341-4061-b914-188004153713 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04] Processing image 81f5ce39-0bc0-4d0e-ab15-42a80a4f614b {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1028.730497] env[65726]: DEBUG oslo_concurrency.lockutils [None req-1d2791c0-c341-4061-b914-188004153713 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/81f5ce39-0bc0-4d0e-ab15-42a80a4f614b/81f5ce39-0bc0-4d0e-ab15-42a80a4f614b.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1028.730611] env[65726]: DEBUG oslo_concurrency.lockutils [None req-1d2791c0-c341-4061-b914-188004153713 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Acquired lock "[datastore1] devstack-image-cache_base/81f5ce39-0bc0-4d0e-ab15-42a80a4f614b/81f5ce39-0bc0-4d0e-ab15-42a80a4f614b.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1028.730793] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-1d2791c0-c341-4061-b914-188004153713 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1028.732041] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4ea204ed-c998-489c-9741-18f6106677d8 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.742268] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-1d2791c0-c341-4061-b914-188004153713 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1028.742538] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-1d2791c0-c341-4061-b914-188004153713 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65726) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1028.743848] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-587c74e5-8b59-4db0-a08b-809c3922d467 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.749414] env[65726]: DEBUG oslo_vmware.api [None req-1d2791c0-c341-4061-b914-188004153713 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Waiting for the task: (returnval){ [ 1028.749414] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5266e2ff-f91e-d8a1-b327-f1801625bf99" [ 1028.749414] env[65726]: _type = "Task" [ 1028.749414] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1028.759371] env[65726]: DEBUG oslo_vmware.api [None req-1d2791c0-c341-4061-b914-188004153713 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5266e2ff-f91e-d8a1-b327-f1801625bf99, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.825778] env[65726]: DEBUG oslo_vmware.api [None req-9ee0c86a-59b9-4c49-9232-0da097c0fa28 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Task: {'id': task-5116518, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.850938] env[65726]: DEBUG nova.compute.manager [None req-f1e7868e-b3d9-4342-8972-f27e4cb4ec45 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] [instance: 2b8eb60b-dc9b-47a4-9ddc-205f2af69888] Start spawning the instance on the hypervisor. {{(pid=65726) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1028.892719] env[65726]: DEBUG nova.virt.hardware [None req-f1e7868e-b3d9-4342-8972-f27e4cb4ec45 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T19:28:34Z,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1539f06fa3534e90acbc3a60c4b8bd3f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T19:28:35Z,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1028.892719] env[65726]: DEBUG nova.virt.hardware [None req-f1e7868e-b3d9-4342-8972-f27e4cb4ec45 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1028.893320] env[65726]: DEBUG nova.virt.hardware [None req-f1e7868e-b3d9-4342-8972-f27e4cb4ec45 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1028.893320] env[65726]: DEBUG nova.virt.hardware [None req-f1e7868e-b3d9-4342-8972-f27e4cb4ec45 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1028.893320] env[65726]: DEBUG nova.virt.hardware [None req-f1e7868e-b3d9-4342-8972-f27e4cb4ec45 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1028.893426] env[65726]: DEBUG nova.virt.hardware [None req-f1e7868e-b3d9-4342-8972-f27e4cb4ec45 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1028.893662] env[65726]: DEBUG nova.virt.hardware [None req-f1e7868e-b3d9-4342-8972-f27e4cb4ec45 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1028.895382] env[65726]: DEBUG nova.virt.hardware [None req-f1e7868e-b3d9-4342-8972-f27e4cb4ec45 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1028.895382] env[65726]: DEBUG nova.virt.hardware [None req-f1e7868e-b3d9-4342-8972-f27e4cb4ec45 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1028.895566] env[65726]: DEBUG nova.virt.hardware [None req-f1e7868e-b3d9-4342-8972-f27e4cb4ec45 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1028.896051] env[65726]: DEBUG nova.virt.hardware [None req-f1e7868e-b3d9-4342-8972-f27e4cb4ec45 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1028.897509] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65974e10-59e4-4d9c-8e04-be245a29e471 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.908119] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c91229a3-0bae-4f84-bba8-5995054dfff3 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.926835] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5f1cb876-8e7c-4680-9e78-4abda3ac2850 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Lock "fb8eac12-aedd-426f-9dd3-6c52f9d4c71d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 19.703s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1029.192384] env[65726]: DEBUG oslo_concurrency.lockutils [None req-2a9dd832-d188-4542-a7fe-bbdea8173b2d tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.371s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1029.192640] env[65726]: DEBUG nova.compute.manager [None req-2a9dd832-d188-4542-a7fe-bbdea8173b2d tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] [instance: 2c812714-dccc-4d1e-bdb5-c11e446949c3] Start building networks asynchronously for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1029.195518] env[65726]: DEBUG oslo_concurrency.lockutils [None req-bd9cdf9a-c493-4e4b-996c-7dba718501da tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 7.888s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1029.195769] env[65726]: DEBUG nova.objects.instance [None req-bd9cdf9a-c493-4e4b-996c-7dba718501da tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Lazy-loading 'resources' on Instance uuid 067c9332-fa2c-47fe-93ae-643de1fe77cf {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1029.263673] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-1d2791c0-c341-4061-b914-188004153713 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04] Preparing fetch location {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 1029.264212] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-1d2791c0-c341-4061-b914-188004153713 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04] Fetch image to [datastore1] OSTACK_IMG_866b99ef-d360-4d77-8ef7-e8f0b892d507/OSTACK_IMG_866b99ef-d360-4d77-8ef7-e8f0b892d507.vmdk {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 1029.264590] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-1d2791c0-c341-4061-b914-188004153713 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04] Downloading stream optimized image 81f5ce39-0bc0-4d0e-ab15-42a80a4f614b to [datastore1] OSTACK_IMG_866b99ef-d360-4d77-8ef7-e8f0b892d507/OSTACK_IMG_866b99ef-d360-4d77-8ef7-e8f0b892d507.vmdk on the data store datastore1 as vApp {{(pid=65726) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 1029.266098] env[65726]: DEBUG nova.virt.vmwareapi.images [None req-1d2791c0-c341-4061-b914-188004153713 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04] Downloading image file data 81f5ce39-0bc0-4d0e-ab15-42a80a4f614b to the ESX as VM named 'OSTACK_IMG_866b99ef-d360-4d77-8ef7-e8f0b892d507' {{(pid=65726) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 1029.301368] env[65726]: DEBUG nova.compute.manager [req-dee74ac9-fce0-4d54-b1dd-829320e193f2 req-67afb67f-ab82-44c0-b3e0-67237866c7d9 service nova] [instance: 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04] Received event network-changed-25b9d108-0b43-4459-b9db-7bd90a495bb3 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1029.301368] env[65726]: DEBUG nova.compute.manager [req-dee74ac9-fce0-4d54-b1dd-829320e193f2 req-67afb67f-ab82-44c0-b3e0-67237866c7d9 service nova] [instance: 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04] Refreshing instance network info cache due to event network-changed-25b9d108-0b43-4459-b9db-7bd90a495bb3. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 1029.301368] env[65726]: DEBUG oslo_concurrency.lockutils [req-dee74ac9-fce0-4d54-b1dd-829320e193f2 req-67afb67f-ab82-44c0-b3e0-67237866c7d9 service nova] Acquiring lock "refresh_cache-2fe6fc5a-f5c7-4f8e-96df-4e621a252f04" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1029.301663] env[65726]: DEBUG oslo_concurrency.lockutils [req-dee74ac9-fce0-4d54-b1dd-829320e193f2 req-67afb67f-ab82-44c0-b3e0-67237866c7d9 service nova] Acquired lock "refresh_cache-2fe6fc5a-f5c7-4f8e-96df-4e621a252f04" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1029.301663] env[65726]: DEBUG nova.network.neutron [req-dee74ac9-fce0-4d54-b1dd-829320e193f2 req-67afb67f-ab82-44c0-b3e0-67237866c7d9 service nova] [instance: 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04] Refreshing network info cache for port 25b9d108-0b43-4459-b9db-7bd90a495bb3 {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 1029.359164] env[65726]: DEBUG oslo_vmware.api [None req-9ee0c86a-59b9-4c49-9232-0da097c0fa28 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Task: {'id': task-5116518, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.431609] env[65726]: DEBUG oslo_vmware.rw_handles [None req-1d2791c0-c341-4061-b914-188004153713 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 1029.431609] env[65726]: value = "resgroup-9" [ 1029.431609] env[65726]: _type = "ResourcePool" [ 1029.431609] env[65726]: }. {{(pid=65726) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 1029.432063] env[65726]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-603f9d31-07b3-43c2-989f-69f0e0cb7226 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.467094] env[65726]: DEBUG nova.network.neutron [None req-f1e7868e-b3d9-4342-8972-f27e4cb4ec45 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] [instance: 2b8eb60b-dc9b-47a4-9ddc-205f2af69888] Successfully updated port: f210c8b9-8838-4774-838c-aeb3860a63ca {{(pid=65726) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 1029.477046] env[65726]: DEBUG oslo_vmware.rw_handles [None req-1d2791c0-c341-4061-b914-188004153713 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Lease: (returnval){ [ 1029.477046] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]529732e3-90b7-0880-e42f-c2edd4f43a6a" [ 1029.477046] env[65726]: _type = "HttpNfcLease" [ 1029.477046] env[65726]: } obtained for vApp import into resource pool (val){ [ 1029.477046] env[65726]: value = "resgroup-9" [ 1029.477046] env[65726]: _type = "ResourcePool" [ 1029.477046] env[65726]: }. {{(pid=65726) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 1029.477046] env[65726]: DEBUG oslo_vmware.api [None req-1d2791c0-c341-4061-b914-188004153713 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Waiting for the lease: (returnval){ [ 1029.477046] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]529732e3-90b7-0880-e42f-c2edd4f43a6a" [ 1029.477046] env[65726]: _type = "HttpNfcLease" [ 1029.477046] env[65726]: } to be ready. {{(pid=65726) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1029.487143] env[65726]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1029.487143] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]529732e3-90b7-0880-e42f-c2edd4f43a6a" [ 1029.487143] env[65726]: _type = "HttpNfcLease" [ 1029.487143] env[65726]: } is initializing. {{(pid=65726) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1029.652678] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-a8820945-d2e3-406b-a110-724bc285e8cb tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] [instance: aa6ce489-c62f-4481-87b7-e74242aeb8ca] Volume attach. Driver type: vmdk {{(pid=65726) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1029.652678] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-a8820945-d2e3-406b-a110-724bc285e8cb tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] [instance: aa6ce489-c62f-4481-87b7-e74242aeb8ca] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-995264', 'volume_id': '3547dfa9-73e6-4baa-b777-e22891db153f', 'name': 'volume-3547dfa9-73e6-4baa-b777-e22891db153f', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'aa6ce489-c62f-4481-87b7-e74242aeb8ca', 'attached_at': '', 'detached_at': '', 'volume_id': '3547dfa9-73e6-4baa-b777-e22891db153f', 'serial': '3547dfa9-73e6-4baa-b777-e22891db153f'} {{(pid=65726) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1029.652678] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e732fe5b-ea26-4ae3-be92-7f47c8a8e5b1 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.670924] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83cf9292-5a25-4253-9771-6d998cd2e30d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.700515] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-a8820945-d2e3-406b-a110-724bc285e8cb tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] [instance: aa6ce489-c62f-4481-87b7-e74242aeb8ca] Reconfiguring VM instance instance-0000004e to attach disk [datastore1] volume-3547dfa9-73e6-4baa-b777-e22891db153f/volume-3547dfa9-73e6-4baa-b777-e22891db153f.vmdk or device None with type thin {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1029.702040] env[65726]: DEBUG nova.compute.utils [None req-2a9dd832-d188-4542-a7fe-bbdea8173b2d tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Using /dev/sd instead of None {{(pid=65726) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1029.703889] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ed959727-70c7-41e0-90b1-262196bcdb3f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.718409] env[65726]: DEBUG nova.objects.instance [None req-bd9cdf9a-c493-4e4b-996c-7dba718501da tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Lazy-loading 'numa_topology' on Instance uuid 067c9332-fa2c-47fe-93ae-643de1fe77cf {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1029.719621] env[65726]: DEBUG nova.compute.manager [None req-2a9dd832-d188-4542-a7fe-bbdea8173b2d tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] [instance: 2c812714-dccc-4d1e-bdb5-c11e446949c3] Allocating IP information in the background. {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1029.723021] env[65726]: DEBUG nova.network.neutron [None req-2a9dd832-d188-4542-a7fe-bbdea8173b2d tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] [instance: 2c812714-dccc-4d1e-bdb5-c11e446949c3] allocate_for_instance() {{(pid=65726) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 1029.723021] env[65726]: WARNING neutronclient.v2_0.client [None req-2a9dd832-d188-4542-a7fe-bbdea8173b2d tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1029.723021] env[65726]: WARNING neutronclient.v2_0.client [None req-2a9dd832-d188-4542-a7fe-bbdea8173b2d tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1029.723021] env[65726]: WARNING openstack [None req-2a9dd832-d188-4542-a7fe-bbdea8173b2d tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1029.723021] env[65726]: WARNING openstack [None req-2a9dd832-d188-4542-a7fe-bbdea8173b2d tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1029.737862] env[65726]: DEBUG oslo_vmware.api [None req-a8820945-d2e3-406b-a110-724bc285e8cb tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Waiting for the task: (returnval){ [ 1029.737862] env[65726]: value = "task-5116520" [ 1029.737862] env[65726]: _type = "Task" [ 1029.737862] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1029.748173] env[65726]: DEBUG oslo_vmware.api [None req-a8820945-d2e3-406b-a110-724bc285e8cb tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Task: {'id': task-5116520, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.789066] env[65726]: DEBUG nova.policy [None req-2a9dd832-d188-4542-a7fe-bbdea8173b2d tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b647b94d0d1848abaee17370c2bedc6f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6961c16aa6a346679bbde25930b2a5b7', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65726) authorize /opt/stack/nova/nova/policy.py:192}} [ 1029.808169] env[65726]: WARNING openstack [req-dee74ac9-fce0-4d54-b1dd-829320e193f2 req-67afb67f-ab82-44c0-b3e0-67237866c7d9 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1029.808169] env[65726]: WARNING openstack [req-dee74ac9-fce0-4d54-b1dd-829320e193f2 req-67afb67f-ab82-44c0-b3e0-67237866c7d9 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1029.855655] env[65726]: DEBUG oslo_vmware.api [None req-9ee0c86a-59b9-4c49-9232-0da097c0fa28 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Task: {'id': task-5116518, 'name': CreateSnapshot_Task, 'duration_secs': 1.094281} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1029.855942] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-9ee0c86a-59b9-4c49-9232-0da097c0fa28 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] [instance: a286ba4f-a83e-4c30-8079-ba88e3ba5696] Created Snapshot of the VM instance {{(pid=65726) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1029.856875] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e876afbe-c481-48e5-8b95-73e62764d47d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.971964] env[65726]: DEBUG oslo_concurrency.lockutils [None req-f1e7868e-b3d9-4342-8972-f27e4cb4ec45 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] Acquiring lock "refresh_cache-2b8eb60b-dc9b-47a4-9ddc-205f2af69888" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1029.972230] env[65726]: DEBUG oslo_concurrency.lockutils [None req-f1e7868e-b3d9-4342-8972-f27e4cb4ec45 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] Acquired lock "refresh_cache-2b8eb60b-dc9b-47a4-9ddc-205f2af69888" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1029.972426] env[65726]: DEBUG nova.network.neutron [None req-f1e7868e-b3d9-4342-8972-f27e4cb4ec45 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] [instance: 2b8eb60b-dc9b-47a4-9ddc-205f2af69888] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1029.993064] env[65726]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1029.993064] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]529732e3-90b7-0880-e42f-c2edd4f43a6a" [ 1029.993064] env[65726]: _type = "HttpNfcLease" [ 1029.993064] env[65726]: } is initializing. {{(pid=65726) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1030.196119] env[65726]: DEBUG nova.network.neutron [None req-2a9dd832-d188-4542-a7fe-bbdea8173b2d tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] [instance: 2c812714-dccc-4d1e-bdb5-c11e446949c3] Successfully created port: 203fa726-d08f-47a8-a7cd-389ed31da940 {{(pid=65726) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1030.219078] env[65726]: DEBUG nova.compute.manager [None req-2a9dd832-d188-4542-a7fe-bbdea8173b2d tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] [instance: 2c812714-dccc-4d1e-bdb5-c11e446949c3] Start building block device mappings for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1030.231599] env[65726]: DEBUG nova.objects.base [None req-bd9cdf9a-c493-4e4b-996c-7dba718501da tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Object Instance<067c9332-fa2c-47fe-93ae-643de1fe77cf> lazy-loaded attributes: resources,numa_topology {{(pid=65726) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1030.255470] env[65726]: DEBUG oslo_vmware.api [None req-a8820945-d2e3-406b-a110-724bc285e8cb tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Task: {'id': task-5116520, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.278476] env[65726]: WARNING openstack [req-dee74ac9-fce0-4d54-b1dd-829320e193f2 req-67afb67f-ab82-44c0-b3e0-67237866c7d9 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1030.278937] env[65726]: WARNING openstack [req-dee74ac9-fce0-4d54-b1dd-829320e193f2 req-67afb67f-ab82-44c0-b3e0-67237866c7d9 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1030.378876] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-9ee0c86a-59b9-4c49-9232-0da097c0fa28 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] [instance: a286ba4f-a83e-4c30-8079-ba88e3ba5696] Creating linked-clone VM from snapshot {{(pid=65726) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1030.383433] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-306b5608-3bc6-4ca6-9df5-ded8fe4fe43e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.396176] env[65726]: DEBUG oslo_vmware.api [None req-9ee0c86a-59b9-4c49-9232-0da097c0fa28 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Waiting for the task: (returnval){ [ 1030.396176] env[65726]: value = "task-5116521" [ 1030.396176] env[65726]: _type = "Task" [ 1030.396176] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1030.406686] env[65726]: DEBUG oslo_vmware.api [None req-9ee0c86a-59b9-4c49-9232-0da097c0fa28 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Task: {'id': task-5116521, 'name': CloneVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.475485] env[65726]: WARNING openstack [None req-f1e7868e-b3d9-4342-8972-f27e4cb4ec45 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1030.475956] env[65726]: WARNING openstack [None req-f1e7868e-b3d9-4342-8972-f27e4cb4ec45 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1030.482346] env[65726]: DEBUG nova.network.neutron [None req-f1e7868e-b3d9-4342-8972-f27e4cb4ec45 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] [instance: 2b8eb60b-dc9b-47a4-9ddc-205f2af69888] Instance cache missing network info. {{(pid=65726) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 1030.501385] env[65726]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1030.501385] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]529732e3-90b7-0880-e42f-c2edd4f43a6a" [ 1030.501385] env[65726]: _type = "HttpNfcLease" [ 1030.501385] env[65726]: } is ready. {{(pid=65726) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1030.501687] env[65726]: DEBUG oslo_vmware.rw_handles [None req-1d2791c0-c341-4061-b914-188004153713 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1030.501687] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]529732e3-90b7-0880-e42f-c2edd4f43a6a" [ 1030.501687] env[65726]: _type = "HttpNfcLease" [ 1030.501687] env[65726]: }. {{(pid=65726) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 1030.506157] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6667f97-d5e5-4f2b-9604-8eb5119719b4 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.517676] env[65726]: DEBUG oslo_vmware.rw_handles [None req-1d2791c0-c341-4061-b914-188004153713 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52c81e8e-ce0a-1d4f-5618-808bc72a0636/disk-0.vmdk from lease info. {{(pid=65726) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1030.517882] env[65726]: DEBUG oslo_vmware.rw_handles [None req-1d2791c0-c341-4061-b914-188004153713 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Creating HTTP connection to write to file with size = 31673344 and URL = https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52c81e8e-ce0a-1d4f-5618-808bc72a0636/disk-0.vmdk. {{(pid=65726) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 1030.595483] env[65726]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-0479c8e5-1a5a-4727-b532-811c83dc2b57 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.602721] env[65726]: DEBUG nova.network.neutron [None req-2a9dd832-d188-4542-a7fe-bbdea8173b2d tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] [instance: 2c812714-dccc-4d1e-bdb5-c11e446949c3] Successfully created port: 8addcb43-85bf-481e-92bc-5bbdf48c0ac4 {{(pid=65726) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1030.619584] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28db1a67-7320-4f37-ba30-9dc7f9c2bdba {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.628433] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93f53f4f-28f9-4062-9857-4e63338c7a46 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.669603] env[65726]: WARNING openstack [req-dee74ac9-fce0-4d54-b1dd-829320e193f2 req-67afb67f-ab82-44c0-b3e0-67237866c7d9 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1030.670035] env[65726]: WARNING openstack [req-dee74ac9-fce0-4d54-b1dd-829320e193f2 req-67afb67f-ab82-44c0-b3e0-67237866c7d9 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1030.678354] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a1d6fde-6da4-46de-bf67-cc94f0de02cd {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.688089] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6923052d-9c8d-4b0e-86d3-b2f45aec9ead {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.703478] env[65726]: DEBUG nova.compute.provider_tree [None req-bd9cdf9a-c493-4e4b-996c-7dba718501da tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1030.752838] env[65726]: DEBUG oslo_vmware.api [None req-a8820945-d2e3-406b-a110-724bc285e8cb tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Task: {'id': task-5116520, 'name': ReconfigVM_Task, 'duration_secs': 0.658687} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1030.753312] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-a8820945-d2e3-406b-a110-724bc285e8cb tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] [instance: aa6ce489-c62f-4481-87b7-e74242aeb8ca] Reconfigured VM instance instance-0000004e to attach disk [datastore1] volume-3547dfa9-73e6-4baa-b777-e22891db153f/volume-3547dfa9-73e6-4baa-b777-e22891db153f.vmdk or device None with type thin {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1030.759374] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-65fde611-73c8-4fce-b1e8-b1cf2fb867c9 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.777585] env[65726]: DEBUG oslo_vmware.api [None req-a8820945-d2e3-406b-a110-724bc285e8cb tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Waiting for the task: (returnval){ [ 1030.777585] env[65726]: value = "task-5116522" [ 1030.777585] env[65726]: _type = "Task" [ 1030.777585] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1030.791993] env[65726]: DEBUG oslo_vmware.api [None req-a8820945-d2e3-406b-a110-724bc285e8cb tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Task: {'id': task-5116522, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.907076] env[65726]: DEBUG oslo_vmware.api [None req-9ee0c86a-59b9-4c49-9232-0da097c0fa28 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Task: {'id': task-5116521, 'name': CloneVM_Task} progress is 94%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.967633] env[65726]: WARNING openstack [None req-f1e7868e-b3d9-4342-8972-f27e4cb4ec45 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1030.968126] env[65726]: WARNING openstack [None req-f1e7868e-b3d9-4342-8972-f27e4cb4ec45 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1031.099548] env[65726]: DEBUG oslo_concurrency.lockutils [None req-10701ce3-ca7d-4146-b0df-5677afb5eba0 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] Acquiring lock "df08d166-7811-49b5-9c16-a3434229d623" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1031.099891] env[65726]: DEBUG oslo_concurrency.lockutils [None req-10701ce3-ca7d-4146-b0df-5677afb5eba0 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] Lock "df08d166-7811-49b5-9c16-a3434229d623" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1031.186871] env[65726]: DEBUG nova.network.neutron [req-dee74ac9-fce0-4d54-b1dd-829320e193f2 req-67afb67f-ab82-44c0-b3e0-67237866c7d9 service nova] [instance: 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04] Updated VIF entry in instance network info cache for port 25b9d108-0b43-4459-b9db-7bd90a495bb3. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 1031.187280] env[65726]: DEBUG nova.network.neutron [req-dee74ac9-fce0-4d54-b1dd-829320e193f2 req-67afb67f-ab82-44c0-b3e0-67237866c7d9 service nova] [instance: 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04] Updating instance_info_cache with network_info: [{"id": "25b9d108-0b43-4459-b9db-7bd90a495bb3", "address": "fa:16:3e:5a:1e:6b", "network": {"id": "eb3278b7-c190-441b-84d2-305f9b896382", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1754571022-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.149", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "96149159e18e44f9bf3453e67681f224", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f77ff7a1-209c-4f3f-b2a0-fd817741e739", "external-id": "nsx-vlan-transportzone-935", "segmentation_id": 935, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap25b9d108-0b", "ovs_interfaceid": "25b9d108-0b43-4459-b9db-7bd90a495bb3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1031.210712] env[65726]: DEBUG nova.scheduler.client.report [None req-bd9cdf9a-c493-4e4b-996c-7dba718501da tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1031.237758] env[65726]: DEBUG nova.compute.manager [None req-2a9dd832-d188-4542-a7fe-bbdea8173b2d tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] [instance: 2c812714-dccc-4d1e-bdb5-c11e446949c3] Start spawning the instance on the hypervisor. {{(pid=65726) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1031.248092] env[65726]: WARNING openstack [None req-f1e7868e-b3d9-4342-8972-f27e4cb4ec45 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1031.249207] env[65726]: WARNING openstack [None req-f1e7868e-b3d9-4342-8972-f27e4cb4ec45 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1031.279021] env[65726]: DEBUG nova.virt.hardware [None req-2a9dd832-d188-4542-a7fe-bbdea8173b2d tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T19:28:34Z,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1539f06fa3534e90acbc3a60c4b8bd3f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T19:28:35Z,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1031.279021] env[65726]: DEBUG nova.virt.hardware [None req-2a9dd832-d188-4542-a7fe-bbdea8173b2d tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1031.279021] env[65726]: DEBUG nova.virt.hardware [None req-2a9dd832-d188-4542-a7fe-bbdea8173b2d tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1031.279021] env[65726]: DEBUG nova.virt.hardware [None req-2a9dd832-d188-4542-a7fe-bbdea8173b2d tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1031.279021] env[65726]: DEBUG nova.virt.hardware [None req-2a9dd832-d188-4542-a7fe-bbdea8173b2d tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1031.279021] env[65726]: DEBUG nova.virt.hardware [None req-2a9dd832-d188-4542-a7fe-bbdea8173b2d tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1031.279021] env[65726]: DEBUG nova.virt.hardware [None req-2a9dd832-d188-4542-a7fe-bbdea8173b2d tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1031.279021] env[65726]: DEBUG nova.virt.hardware [None req-2a9dd832-d188-4542-a7fe-bbdea8173b2d tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1031.279021] env[65726]: DEBUG nova.virt.hardware [None req-2a9dd832-d188-4542-a7fe-bbdea8173b2d tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1031.279021] env[65726]: DEBUG nova.virt.hardware [None req-2a9dd832-d188-4542-a7fe-bbdea8173b2d tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1031.279021] env[65726]: DEBUG nova.virt.hardware [None req-2a9dd832-d188-4542-a7fe-bbdea8173b2d tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1031.280862] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38a47a46-b06d-4620-bd36-161630100895 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.305027] env[65726]: DEBUG oslo_vmware.api [None req-a8820945-d2e3-406b-a110-724bc285e8cb tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Task: {'id': task-5116522, 'name': ReconfigVM_Task, 'duration_secs': 0.177928} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1031.309548] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c972145-79de-4af3-8ecc-100dd5c2b94f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.315722] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-a8820945-d2e3-406b-a110-724bc285e8cb tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] [instance: aa6ce489-c62f-4481-87b7-e74242aeb8ca] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-995264', 'volume_id': '3547dfa9-73e6-4baa-b777-e22891db153f', 'name': 'volume-3547dfa9-73e6-4baa-b777-e22891db153f', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'aa6ce489-c62f-4481-87b7-e74242aeb8ca', 'attached_at': '', 'detached_at': '', 'volume_id': '3547dfa9-73e6-4baa-b777-e22891db153f', 'serial': '3547dfa9-73e6-4baa-b777-e22891db153f'} {{(pid=65726) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1031.410500] env[65726]: DEBUG oslo_vmware.api [None req-9ee0c86a-59b9-4c49-9232-0da097c0fa28 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Task: {'id': task-5116521, 'name': CloneVM_Task} progress is 94%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.454033] env[65726]: DEBUG nova.network.neutron [None req-f1e7868e-b3d9-4342-8972-f27e4cb4ec45 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] [instance: 2b8eb60b-dc9b-47a4-9ddc-205f2af69888] Updating instance_info_cache with network_info: [{"id": "f210c8b9-8838-4774-838c-aeb3860a63ca", "address": "fa:16:3e:5a:3b:53", "network": {"id": "b99bfe5a-6e05-45dc-81ff-61ad5e98453a", "bridge": "br-int", "label": "tempest-InstanceActionsTestJSON-314314203-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "878eef0927ed47dbb370522e93a5453f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "65e4a2b4-fd64-4ac9-b2ec-bac768b501c5", "external-id": "nsx-vlan-transportzone-449", "segmentation_id": 449, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf210c8b9-88", "ovs_interfaceid": "f210c8b9-8838-4774-838c-aeb3860a63ca", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1031.605998] env[65726]: DEBUG nova.compute.manager [None req-10701ce3-ca7d-4146-b0df-5677afb5eba0 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] [instance: df08d166-7811-49b5-9c16-a3434229d623] Starting instance... {{(pid=65726) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 1031.693047] env[65726]: DEBUG oslo_concurrency.lockutils [req-dee74ac9-fce0-4d54-b1dd-829320e193f2 req-67afb67f-ab82-44c0-b3e0-67237866c7d9 service nova] Releasing lock "refresh_cache-2fe6fc5a-f5c7-4f8e-96df-4e621a252f04" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1031.716767] env[65726]: DEBUG oslo_concurrency.lockutils [None req-bd9cdf9a-c493-4e4b-996c-7dba718501da tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.521s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1031.721246] env[65726]: DEBUG oslo_concurrency.lockutils [None req-7aacd0be-ee58-4b95-8abe-402ca2fa658f tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 9.508s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1031.721588] env[65726]: DEBUG nova.objects.instance [None req-7aacd0be-ee58-4b95-8abe-402ca2fa658f tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Lazy-loading 'resources' on Instance uuid dcde25df-87a5-47fb-94ce-334f68894e04 {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1031.770682] env[65726]: INFO nova.compute.manager [None req-a3acab1e-0b06-47ad-b154-32b257cd5af7 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: fb8eac12-aedd-426f-9dd3-6c52f9d4c71d] Rebuilding instance [ 1031.830122] env[65726]: DEBUG nova.compute.manager [None req-a3acab1e-0b06-47ad-b154-32b257cd5af7 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: fb8eac12-aedd-426f-9dd3-6c52f9d4c71d] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1031.831146] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cac54ce0-7cce-49ab-bc23-e59fa798bff8 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.909316] env[65726]: DEBUG oslo_vmware.api [None req-9ee0c86a-59b9-4c49-9232-0da097c0fa28 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Task: {'id': task-5116521, 'name': CloneVM_Task} progress is 95%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.957226] env[65726]: DEBUG oslo_concurrency.lockutils [None req-f1e7868e-b3d9-4342-8972-f27e4cb4ec45 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] Releasing lock "refresh_cache-2b8eb60b-dc9b-47a4-9ddc-205f2af69888" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1031.957605] env[65726]: DEBUG nova.compute.manager [None req-f1e7868e-b3d9-4342-8972-f27e4cb4ec45 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] [instance: 2b8eb60b-dc9b-47a4-9ddc-205f2af69888] Instance network_info: |[{"id": "f210c8b9-8838-4774-838c-aeb3860a63ca", "address": "fa:16:3e:5a:3b:53", "network": {"id": "b99bfe5a-6e05-45dc-81ff-61ad5e98453a", "bridge": "br-int", "label": "tempest-InstanceActionsTestJSON-314314203-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "878eef0927ed47dbb370522e93a5453f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "65e4a2b4-fd64-4ac9-b2ec-bac768b501c5", "external-id": "nsx-vlan-transportzone-449", "segmentation_id": 449, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf210c8b9-88", "ovs_interfaceid": "f210c8b9-8838-4774-838c-aeb3860a63ca", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1031.958198] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-f1e7868e-b3d9-4342-8972-f27e4cb4ec45 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] [instance: 2b8eb60b-dc9b-47a4-9ddc-205f2af69888] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:5a:3b:53', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '65e4a2b4-fd64-4ac9-b2ec-bac768b501c5', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f210c8b9-8838-4774-838c-aeb3860a63ca', 'vif_model': 'vmxnet3'}] {{(pid=65726) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1031.967666] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-f1e7868e-b3d9-4342-8972-f27e4cb4ec45 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] Creating folder: Project (878eef0927ed47dbb370522e93a5453f). Parent ref: group-v995008. {{(pid=65726) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1031.968062] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-3bc4b2f4-1681-4011-9eba-f10cf7eb535b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.982071] env[65726]: INFO nova.virt.vmwareapi.vm_util [None req-f1e7868e-b3d9-4342-8972-f27e4cb4ec45 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] Created folder: Project (878eef0927ed47dbb370522e93a5453f) in parent group-v995008. [ 1031.984917] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-f1e7868e-b3d9-4342-8972-f27e4cb4ec45 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] Creating folder: Instances. Parent ref: group-v995269. {{(pid=65726) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1031.984917] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d6f5bf55-5856-42b3-879d-4a5199537cdd {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.994934] env[65726]: INFO nova.virt.vmwareapi.vm_util [None req-f1e7868e-b3d9-4342-8972-f27e4cb4ec45 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] Created folder: Instances in parent group-v995269. [ 1031.998593] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-f1e7868e-b3d9-4342-8972-f27e4cb4ec45 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1031.998593] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2b8eb60b-dc9b-47a4-9ddc-205f2af69888] Creating VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1031.998593] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a7018565-f53a-4048-8142-bc42ec5ecee8 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.019757] env[65726]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1032.019757] env[65726]: value = "task-5116525" [ 1032.019757] env[65726]: _type = "Task" [ 1032.019757] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1032.029888] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116525, 'name': CreateVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.132536] env[65726]: DEBUG oslo_concurrency.lockutils [None req-10701ce3-ca7d-4146-b0df-5677afb5eba0 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1032.161951] env[65726]: DEBUG nova.compute.manager [req-ca08a4f2-4ea6-42fb-b87f-1c44fe2ea345 req-17642d9f-2a6c-4f06-8465-38c759b667f2 service nova] [instance: 2b8eb60b-dc9b-47a4-9ddc-205f2af69888] Received event network-vif-plugged-f210c8b9-8838-4774-838c-aeb3860a63ca {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1032.162538] env[65726]: DEBUG oslo_concurrency.lockutils [req-ca08a4f2-4ea6-42fb-b87f-1c44fe2ea345 req-17642d9f-2a6c-4f06-8465-38c759b667f2 service nova] Acquiring lock "2b8eb60b-dc9b-47a4-9ddc-205f2af69888-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1032.162793] env[65726]: DEBUG oslo_concurrency.lockutils [req-ca08a4f2-4ea6-42fb-b87f-1c44fe2ea345 req-17642d9f-2a6c-4f06-8465-38c759b667f2 service nova] Lock "2b8eb60b-dc9b-47a4-9ddc-205f2af69888-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1032.162967] env[65726]: DEBUG oslo_concurrency.lockutils [req-ca08a4f2-4ea6-42fb-b87f-1c44fe2ea345 req-17642d9f-2a6c-4f06-8465-38c759b667f2 service nova] Lock "2b8eb60b-dc9b-47a4-9ddc-205f2af69888-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1032.163193] env[65726]: DEBUG nova.compute.manager [req-ca08a4f2-4ea6-42fb-b87f-1c44fe2ea345 req-17642d9f-2a6c-4f06-8465-38c759b667f2 service nova] [instance: 2b8eb60b-dc9b-47a4-9ddc-205f2af69888] No waiting events found dispatching network-vif-plugged-f210c8b9-8838-4774-838c-aeb3860a63ca {{(pid=65726) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1032.163332] env[65726]: WARNING nova.compute.manager [req-ca08a4f2-4ea6-42fb-b87f-1c44fe2ea345 req-17642d9f-2a6c-4f06-8465-38c759b667f2 service nova] [instance: 2b8eb60b-dc9b-47a4-9ddc-205f2af69888] Received unexpected event network-vif-plugged-f210c8b9-8838-4774-838c-aeb3860a63ca for instance with vm_state building and task_state spawning. [ 1032.163493] env[65726]: DEBUG nova.compute.manager [req-ca08a4f2-4ea6-42fb-b87f-1c44fe2ea345 req-17642d9f-2a6c-4f06-8465-38c759b667f2 service nova] [instance: 2b8eb60b-dc9b-47a4-9ddc-205f2af69888] Received event network-changed-f210c8b9-8838-4774-838c-aeb3860a63ca {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1032.163715] env[65726]: DEBUG nova.compute.manager [req-ca08a4f2-4ea6-42fb-b87f-1c44fe2ea345 req-17642d9f-2a6c-4f06-8465-38c759b667f2 service nova] [instance: 2b8eb60b-dc9b-47a4-9ddc-205f2af69888] Refreshing instance network info cache due to event network-changed-f210c8b9-8838-4774-838c-aeb3860a63ca. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 1032.163911] env[65726]: DEBUG oslo_concurrency.lockutils [req-ca08a4f2-4ea6-42fb-b87f-1c44fe2ea345 req-17642d9f-2a6c-4f06-8465-38c759b667f2 service nova] Acquiring lock "refresh_cache-2b8eb60b-dc9b-47a4-9ddc-205f2af69888" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1032.164294] env[65726]: DEBUG oslo_concurrency.lockutils [req-ca08a4f2-4ea6-42fb-b87f-1c44fe2ea345 req-17642d9f-2a6c-4f06-8465-38c759b667f2 service nova] Acquired lock "refresh_cache-2b8eb60b-dc9b-47a4-9ddc-205f2af69888" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1032.165386] env[65726]: DEBUG nova.network.neutron [req-ca08a4f2-4ea6-42fb-b87f-1c44fe2ea345 req-17642d9f-2a6c-4f06-8465-38c759b667f2 service nova] [instance: 2b8eb60b-dc9b-47a4-9ddc-205f2af69888] Refreshing network info cache for port f210c8b9-8838-4774-838c-aeb3860a63ca {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 1032.243577] env[65726]: DEBUG oslo_concurrency.lockutils [None req-bd9cdf9a-c493-4e4b-996c-7dba718501da tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Lock "067c9332-fa2c-47fe-93ae-643de1fe77cf" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 32.491s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1032.247910] env[65726]: DEBUG oslo_concurrency.lockutils [None req-8ffe452d-09ce-47ef-a349-6236507a111f tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Lock "067c9332-fa2c-47fe-93ae-643de1fe77cf" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 5.600s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1032.247910] env[65726]: DEBUG oslo_concurrency.lockutils [None req-8ffe452d-09ce-47ef-a349-6236507a111f tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Acquiring lock "067c9332-fa2c-47fe-93ae-643de1fe77cf-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1032.247910] env[65726]: DEBUG oslo_concurrency.lockutils [None req-8ffe452d-09ce-47ef-a349-6236507a111f tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Lock "067c9332-fa2c-47fe-93ae-643de1fe77cf-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1032.247910] env[65726]: DEBUG oslo_concurrency.lockutils [None req-8ffe452d-09ce-47ef-a349-6236507a111f tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Lock "067c9332-fa2c-47fe-93ae-643de1fe77cf-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1032.249538] env[65726]: INFO nova.compute.manager [None req-8ffe452d-09ce-47ef-a349-6236507a111f tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: 067c9332-fa2c-47fe-93ae-643de1fe77cf] Terminating instance [ 1032.369362] env[65726]: DEBUG nova.objects.instance [None req-a8820945-d2e3-406b-a110-724bc285e8cb tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Lazy-loading 'flavor' on Instance uuid aa6ce489-c62f-4481-87b7-e74242aeb8ca {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1032.376429] env[65726]: DEBUG oslo_vmware.rw_handles [None req-1d2791c0-c341-4061-b914-188004153713 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Completed reading data from the image iterator. {{(pid=65726) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 1032.376686] env[65726]: DEBUG oslo_vmware.rw_handles [None req-1d2791c0-c341-4061-b914-188004153713 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52c81e8e-ce0a-1d4f-5618-808bc72a0636/disk-0.vmdk. {{(pid=65726) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1032.382140] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c48a5cb8-64eb-4c8f-ae72-f3d9e1a96a58 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.388668] env[65726]: DEBUG oslo_vmware.rw_handles [None req-1d2791c0-c341-4061-b914-188004153713 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52c81e8e-ce0a-1d4f-5618-808bc72a0636/disk-0.vmdk is in state: ready. {{(pid=65726) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1032.388846] env[65726]: DEBUG oslo_vmware.rw_handles [None req-1d2791c0-c341-4061-b914-188004153713 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Releasing lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52c81e8e-ce0a-1d4f-5618-808bc72a0636/disk-0.vmdk. {{(pid=65726) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 1032.389754] env[65726]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-599d3cc5-e675-49b9-9e1e-63bb4538295b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.411893] env[65726]: DEBUG nova.network.neutron [None req-2a9dd832-d188-4542-a7fe-bbdea8173b2d tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] [instance: 2c812714-dccc-4d1e-bdb5-c11e446949c3] Successfully updated port: 203fa726-d08f-47a8-a7cd-389ed31da940 {{(pid=65726) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 1032.420113] env[65726]: DEBUG oslo_vmware.api [None req-9ee0c86a-59b9-4c49-9232-0da097c0fa28 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Task: {'id': task-5116521, 'name': CloneVM_Task, 'duration_secs': 1.762679} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1032.420796] env[65726]: INFO nova.virt.vmwareapi.vmops [None req-9ee0c86a-59b9-4c49-9232-0da097c0fa28 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] [instance: a286ba4f-a83e-4c30-8079-ba88e3ba5696] Created linked-clone VM from snapshot [ 1032.422114] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3b42618-52af-414c-b6dc-534d5c859c38 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.432879] env[65726]: DEBUG nova.virt.vmwareapi.images [None req-9ee0c86a-59b9-4c49-9232-0da097c0fa28 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] [instance: a286ba4f-a83e-4c30-8079-ba88e3ba5696] Uploading image 46a5e33a-19cd-4fd5-8455-a08423252044 {{(pid=65726) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 1032.452315] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-9ee0c86a-59b9-4c49-9232-0da097c0fa28 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] [instance: a286ba4f-a83e-4c30-8079-ba88e3ba5696] Destroying the VM {{(pid=65726) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 1032.452609] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-6a1ea61e-6702-4a71-92ec-cd286ebfa248 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.461373] env[65726]: DEBUG oslo_vmware.api [None req-9ee0c86a-59b9-4c49-9232-0da097c0fa28 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Waiting for the task: (returnval){ [ 1032.461373] env[65726]: value = "task-5116526" [ 1032.461373] env[65726]: _type = "Task" [ 1032.461373] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1032.475076] env[65726]: DEBUG oslo_vmware.api [None req-9ee0c86a-59b9-4c49-9232-0da097c0fa28 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Task: {'id': task-5116526, 'name': Destroy_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.525742] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c32c7ad3-4d13-4bd7-a2d4-4f638876078e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.535238] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116525, 'name': CreateVM_Task, 'duration_secs': 0.399284} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1032.537040] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2b8eb60b-dc9b-47a4-9ddc-205f2af69888] Created VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1032.537942] env[65726]: WARNING openstack [None req-f1e7868e-b3d9-4342-8972-f27e4cb4ec45 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1032.538588] env[65726]: WARNING openstack [None req-f1e7868e-b3d9-4342-8972-f27e4cb4ec45 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1032.543965] env[65726]: DEBUG oslo_concurrency.lockutils [None req-f1e7868e-b3d9-4342-8972-f27e4cb4ec45 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1032.544185] env[65726]: DEBUG oslo_concurrency.lockutils [None req-f1e7868e-b3d9-4342-8972-f27e4cb4ec45 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1032.544563] env[65726]: DEBUG oslo_concurrency.lockutils [None req-f1e7868e-b3d9-4342-8972-f27e4cb4ec45 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1032.545713] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d999e2de-2b73-45e3-9e95-8ff79a6155c5 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.549272] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-38ca1195-c437-411c-b50c-875c94b0cbd3 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.555389] env[65726]: DEBUG oslo_vmware.api [None req-f1e7868e-b3d9-4342-8972-f27e4cb4ec45 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] Waiting for the task: (returnval){ [ 1032.555389] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]526c5aea-f989-26f6-40cb-342b2f0292bf" [ 1032.555389] env[65726]: _type = "Task" [ 1032.555389] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1032.585336] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-165a73a0-31f2-4303-998e-6f2baf6aeae7 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.594073] env[65726]: DEBUG oslo_vmware.api [None req-f1e7868e-b3d9-4342-8972-f27e4cb4ec45 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]526c5aea-f989-26f6-40cb-342b2f0292bf, 'name': SearchDatastore_Task, 'duration_secs': 0.010869} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1032.596409] env[65726]: DEBUG oslo_concurrency.lockutils [None req-f1e7868e-b3d9-4342-8972-f27e4cb4ec45 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1032.596695] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-f1e7868e-b3d9-4342-8972-f27e4cb4ec45 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] [instance: 2b8eb60b-dc9b-47a4-9ddc-205f2af69888] Processing image b52362a3-ee8a-4cbf-b06f-513b0cc8f95c {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1032.596938] env[65726]: DEBUG oslo_concurrency.lockutils [None req-f1e7868e-b3d9-4342-8972-f27e4cb4ec45 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1032.597114] env[65726]: DEBUG oslo_concurrency.lockutils [None req-f1e7868e-b3d9-4342-8972-f27e4cb4ec45 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1032.597337] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-f1e7868e-b3d9-4342-8972-f27e4cb4ec45 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1032.597674] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7ea50981-639a-4a28-b62a-0abb592d651d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.600732] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9f9910c-41c3-46c4-a522-4d9e7467fdea {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.616862] env[65726]: DEBUG nova.compute.provider_tree [None req-7aacd0be-ee58-4b95-8abe-402ca2fa658f tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1032.619902] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-f1e7868e-b3d9-4342-8972-f27e4cb4ec45 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1032.620087] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-f1e7868e-b3d9-4342-8972-f27e4cb4ec45 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65726) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1032.621043] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ea91d7ca-93f1-4159-9094-27da4fb23282 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.627618] env[65726]: DEBUG oslo_vmware.api [None req-f1e7868e-b3d9-4342-8972-f27e4cb4ec45 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] Waiting for the task: (returnval){ [ 1032.627618] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52e44172-0e25-2b0b-70f8-66a5b2cd073e" [ 1032.627618] env[65726]: _type = "Task" [ 1032.627618] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1032.637306] env[65726]: DEBUG oslo_vmware.api [None req-f1e7868e-b3d9-4342-8972-f27e4cb4ec45 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52e44172-0e25-2b0b-70f8-66a5b2cd073e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.668355] env[65726]: WARNING openstack [req-ca08a4f2-4ea6-42fb-b87f-1c44fe2ea345 req-17642d9f-2a6c-4f06-8465-38c759b667f2 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1032.668760] env[65726]: WARNING openstack [req-ca08a4f2-4ea6-42fb-b87f-1c44fe2ea345 req-17642d9f-2a6c-4f06-8465-38c759b667f2 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1032.758793] env[65726]: DEBUG nova.compute.manager [None req-8ffe452d-09ce-47ef-a349-6236507a111f tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: 067c9332-fa2c-47fe-93ae-643de1fe77cf] Start destroying the instance on the hypervisor. {{(pid=65726) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1032.759145] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-8ffe452d-09ce-47ef-a349-6236507a111f tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: 067c9332-fa2c-47fe-93ae-643de1fe77cf] Destroying instance {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1032.764936] env[65726]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1d9289f8-8530-4a75-89b9-aa3a21bddcd6 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.767704] env[65726]: DEBUG oslo_vmware.rw_handles [None req-1d2791c0-c341-4061-b914-188004153713 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Closed VMDK write handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52c81e8e-ce0a-1d4f-5618-808bc72a0636/disk-0.vmdk. {{(pid=65726) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 1032.768111] env[65726]: INFO nova.virt.vmwareapi.images [None req-1d2791c0-c341-4061-b914-188004153713 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04] Downloaded image file data 81f5ce39-0bc0-4d0e-ab15-42a80a4f614b [ 1032.769198] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7aadec4-8bac-4995-9094-cb0c3f1332b5 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.796215] env[65726]: WARNING openstack [req-ca08a4f2-4ea6-42fb-b87f-1c44fe2ea345 req-17642d9f-2a6c-4f06-8465-38c759b667f2 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1032.796708] env[65726]: WARNING openstack [req-ca08a4f2-4ea6-42fb-b87f-1c44fe2ea345 req-17642d9f-2a6c-4f06-8465-38c759b667f2 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1032.805015] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a36dbde4-7d1d-479a-8c67-6642f3a150a2 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.809785] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6af4046d-b86a-4ac2-ac95-7b82cfc0b6a6 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.854657] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-a3acab1e-0b06-47ad-b154-32b257cd5af7 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: fb8eac12-aedd-426f-9dd3-6c52f9d4c71d] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1032.855105] env[65726]: WARNING nova.virt.vmwareapi.vmops [None req-8ffe452d-09ce-47ef-a349-6236507a111f tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: 067c9332-fa2c-47fe-93ae-643de1fe77cf] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 067c9332-fa2c-47fe-93ae-643de1fe77cf could not be found. [ 1032.855305] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-8ffe452d-09ce-47ef-a349-6236507a111f tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: 067c9332-fa2c-47fe-93ae-643de1fe77cf] Instance destroyed {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1032.855844] env[65726]: INFO nova.compute.manager [None req-8ffe452d-09ce-47ef-a349-6236507a111f tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: 067c9332-fa2c-47fe-93ae-643de1fe77cf] Took 0.10 seconds to destroy the instance on the hypervisor. [ 1032.856165] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-8ffe452d-09ce-47ef-a349-6236507a111f tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1032.856527] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1238b551-a45c-447e-be8a-be1dc3ec1851 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.858727] env[65726]: DEBUG nova.compute.manager [-] [instance: 067c9332-fa2c-47fe-93ae-643de1fe77cf] Deallocating network for instance {{(pid=65726) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1032.859046] env[65726]: DEBUG nova.network.neutron [-] [instance: 067c9332-fa2c-47fe-93ae-643de1fe77cf] deallocate_for_instance() {{(pid=65726) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 1032.859386] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1032.859664] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1032.873609] env[65726]: DEBUG oslo_vmware.api [None req-a3acab1e-0b06-47ad-b154-32b257cd5af7 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Waiting for the task: (returnval){ [ 1032.873609] env[65726]: value = "task-5116528" [ 1032.873609] env[65726]: _type = "Task" [ 1032.873609] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1032.882332] env[65726]: WARNING openstack [req-ca08a4f2-4ea6-42fb-b87f-1c44fe2ea345 req-17642d9f-2a6c-4f06-8465-38c759b667f2 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1032.882332] env[65726]: WARNING openstack [req-ca08a4f2-4ea6-42fb-b87f-1c44fe2ea345 req-17642d9f-2a6c-4f06-8465-38c759b667f2 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1032.893678] env[65726]: INFO nova.virt.vmwareapi.images [None req-1d2791c0-c341-4061-b914-188004153713 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04] The imported VM was unregistered [ 1032.897014] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-1d2791c0-c341-4061-b914-188004153713 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04] Caching image {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 1032.897289] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-1d2791c0-c341-4061-b914-188004153713 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Creating directory with path [datastore1] devstack-image-cache_base/81f5ce39-0bc0-4d0e-ab15-42a80a4f614b {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1032.902036] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a8820945-d2e3-406b-a110-724bc285e8cb tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Lock "aa6ce489-c62f-4481-87b7-e74242aeb8ca" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.890s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1032.902036] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-17519f6e-3c2e-41d1-a325-312ad7378639 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.908497] env[65726]: DEBUG oslo_vmware.api [None req-a3acab1e-0b06-47ad-b154-32b257cd5af7 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': task-5116528, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.920985] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-1d2791c0-c341-4061-b914-188004153713 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Created directory with path [datastore1] devstack-image-cache_base/81f5ce39-0bc0-4d0e-ab15-42a80a4f614b {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1032.921244] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-1d2791c0-c341-4061-b914-188004153713 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Moving virtual disk from [datastore1] OSTACK_IMG_866b99ef-d360-4d77-8ef7-e8f0b892d507/OSTACK_IMG_866b99ef-d360-4d77-8ef7-e8f0b892d507.vmdk to [datastore1] devstack-image-cache_base/81f5ce39-0bc0-4d0e-ab15-42a80a4f614b/81f5ce39-0bc0-4d0e-ab15-42a80a4f614b.vmdk. {{(pid=65726) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 1032.928020] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1032.929339] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1032.937425] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-565f05c0-f968-48d7-be08-7428404c200a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.947406] env[65726]: DEBUG oslo_vmware.api [None req-1d2791c0-c341-4061-b914-188004153713 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Waiting for the task: (returnval){ [ 1032.947406] env[65726]: value = "task-5116529" [ 1032.947406] env[65726]: _type = "Task" [ 1032.947406] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1032.959489] env[65726]: DEBUG oslo_vmware.api [None req-1d2791c0-c341-4061-b914-188004153713 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Task: {'id': task-5116529, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.974905] env[65726]: DEBUG oslo_vmware.api [None req-9ee0c86a-59b9-4c49-9232-0da097c0fa28 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Task: {'id': task-5116526, 'name': Destroy_Task, 'duration_secs': 0.368619} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1032.975299] env[65726]: INFO nova.virt.vmwareapi.vm_util [None req-9ee0c86a-59b9-4c49-9232-0da097c0fa28 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] [instance: a286ba4f-a83e-4c30-8079-ba88e3ba5696] Destroyed the VM [ 1032.975616] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-9ee0c86a-59b9-4c49-9232-0da097c0fa28 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] [instance: a286ba4f-a83e-4c30-8079-ba88e3ba5696] Deleting Snapshot of the VM instance {{(pid=65726) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1032.976679] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-4b310c88-274e-49cb-a0ca-09694012bb37 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.989955] env[65726]: DEBUG oslo_vmware.api [None req-9ee0c86a-59b9-4c49-9232-0da097c0fa28 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Waiting for the task: (returnval){ [ 1032.989955] env[65726]: value = "task-5116530" [ 1032.989955] env[65726]: _type = "Task" [ 1032.989955] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1033.003465] env[65726]: DEBUG oslo_vmware.api [None req-9ee0c86a-59b9-4c49-9232-0da097c0fa28 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Task: {'id': task-5116530, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.122374] env[65726]: DEBUG nova.scheduler.client.report [None req-7aacd0be-ee58-4b95-8abe-402ca2fa658f tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1033.140706] env[65726]: DEBUG oslo_vmware.api [None req-f1e7868e-b3d9-4342-8972-f27e4cb4ec45 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52e44172-0e25-2b0b-70f8-66a5b2cd073e, 'name': SearchDatastore_Task, 'duration_secs': 0.010258} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1033.142324] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7c6be903-056d-4ea4-b980-7c2ef5862014 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.149150] env[65726]: DEBUG oslo_vmware.api [None req-f1e7868e-b3d9-4342-8972-f27e4cb4ec45 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] Waiting for the task: (returnval){ [ 1033.149150] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52f7eb81-c9c1-6858-000a-e79e28e8df03" [ 1033.149150] env[65726]: _type = "Task" [ 1033.149150] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1033.158061] env[65726]: DEBUG oslo_vmware.api [None req-f1e7868e-b3d9-4342-8972-f27e4cb4ec45 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52f7eb81-c9c1-6858-000a-e79e28e8df03, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.252024] env[65726]: DEBUG nova.network.neutron [req-ca08a4f2-4ea6-42fb-b87f-1c44fe2ea345 req-17642d9f-2a6c-4f06-8465-38c759b667f2 service nova] [instance: 2b8eb60b-dc9b-47a4-9ddc-205f2af69888] Updated VIF entry in instance network info cache for port f210c8b9-8838-4774-838c-aeb3860a63ca. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 1033.252024] env[65726]: DEBUG nova.network.neutron [req-ca08a4f2-4ea6-42fb-b87f-1c44fe2ea345 req-17642d9f-2a6c-4f06-8465-38c759b667f2 service nova] [instance: 2b8eb60b-dc9b-47a4-9ddc-205f2af69888] Updating instance_info_cache with network_info: [{"id": "f210c8b9-8838-4774-838c-aeb3860a63ca", "address": "fa:16:3e:5a:3b:53", "network": {"id": "b99bfe5a-6e05-45dc-81ff-61ad5e98453a", "bridge": "br-int", "label": "tempest-InstanceActionsTestJSON-314314203-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "878eef0927ed47dbb370522e93a5453f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "65e4a2b4-fd64-4ac9-b2ec-bac768b501c5", "external-id": "nsx-vlan-transportzone-449", "segmentation_id": 449, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf210c8b9-88", "ovs_interfaceid": "f210c8b9-8838-4774-838c-aeb3860a63ca", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1033.333155] env[65726]: DEBUG oslo_concurrency.lockutils [None req-72c48a25-f856-4f68-98a0-3ea61ba329e4 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Acquiring lock "aa6ce489-c62f-4481-87b7-e74242aeb8ca" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1033.333155] env[65726]: DEBUG oslo_concurrency.lockutils [None req-72c48a25-f856-4f68-98a0-3ea61ba329e4 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Lock "aa6ce489-c62f-4481-87b7-e74242aeb8ca" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1033.333577] env[65726]: DEBUG nova.compute.manager [None req-72c48a25-f856-4f68-98a0-3ea61ba329e4 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] [instance: aa6ce489-c62f-4481-87b7-e74242aeb8ca] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1033.334722] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89c2a976-b3ac-48cd-af7b-48cd78e8bb1a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.344018] env[65726]: DEBUG nova.compute.manager [None req-72c48a25-f856-4f68-98a0-3ea61ba329e4 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] [instance: aa6ce489-c62f-4481-87b7-e74242aeb8ca] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=65726) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3459}} [ 1033.344750] env[65726]: DEBUG nova.objects.instance [None req-72c48a25-f856-4f68-98a0-3ea61ba329e4 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Lazy-loading 'flavor' on Instance uuid aa6ce489-c62f-4481-87b7-e74242aeb8ca {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1033.388687] env[65726]: DEBUG oslo_vmware.api [None req-a3acab1e-0b06-47ad-b154-32b257cd5af7 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': task-5116528, 'name': PowerOffVM_Task, 'duration_secs': 0.225105} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1033.389351] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-a3acab1e-0b06-47ad-b154-32b257cd5af7 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: fb8eac12-aedd-426f-9dd3-6c52f9d4c71d] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1033.389588] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-a3acab1e-0b06-47ad-b154-32b257cd5af7 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: fb8eac12-aedd-426f-9dd3-6c52f9d4c71d] Destroying instance {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1033.391181] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a27a7b8f-5cad-4bdf-9289-41ece9203b35 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.406362] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-a3acab1e-0b06-47ad-b154-32b257cd5af7 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: fb8eac12-aedd-426f-9dd3-6c52f9d4c71d] Unregistering the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1033.406362] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0fe24ecb-430c-4b15-8a4b-0dbf36f07300 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.418386] env[65726]: DEBUG oslo_concurrency.lockutils [None req-0c1711f1-74cb-4ba7-a859-716c68812e37 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Acquiring lock "ad4f00b5-3765-494f-8132-8b17b29d219a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1033.418792] env[65726]: DEBUG oslo_concurrency.lockutils [None req-0c1711f1-74cb-4ba7-a859-716c68812e37 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Lock "ad4f00b5-3765-494f-8132-8b17b29d219a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1033.459586] env[65726]: DEBUG oslo_vmware.api [None req-1d2791c0-c341-4061-b914-188004153713 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Task: {'id': task-5116529, 'name': MoveVirtualDisk_Task} progress is 18%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.488071] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-a3acab1e-0b06-47ad-b154-32b257cd5af7 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: fb8eac12-aedd-426f-9dd3-6c52f9d4c71d] Unregistered the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1033.488354] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-a3acab1e-0b06-47ad-b154-32b257cd5af7 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: fb8eac12-aedd-426f-9dd3-6c52f9d4c71d] Deleting contents of the VM from datastore datastore2 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1033.488474] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-a3acab1e-0b06-47ad-b154-32b257cd5af7 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Deleting the datastore file [datastore2] fb8eac12-aedd-426f-9dd3-6c52f9d4c71d {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1033.488756] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0b44a998-0007-455b-9f5d-7da5b1393903 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.502377] env[65726]: DEBUG oslo_vmware.api [None req-9ee0c86a-59b9-4c49-9232-0da097c0fa28 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Task: {'id': task-5116530, 'name': RemoveSnapshot_Task} progress is 80%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.504716] env[65726]: DEBUG oslo_vmware.api [None req-a3acab1e-0b06-47ad-b154-32b257cd5af7 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Waiting for the task: (returnval){ [ 1033.504716] env[65726]: value = "task-5116532" [ 1033.504716] env[65726]: _type = "Task" [ 1033.504716] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1033.519370] env[65726]: DEBUG oslo_vmware.api [None req-a3acab1e-0b06-47ad-b154-32b257cd5af7 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': task-5116532, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.628988] env[65726]: DEBUG oslo_concurrency.lockutils [None req-7aacd0be-ee58-4b95-8abe-402ca2fa658f tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.908s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1033.631686] env[65726]: DEBUG oslo_concurrency.lockutils [None req-345ea84c-77bf-4393-9db3-8f6989cf9c1c tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 10.711s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1033.632021] env[65726]: DEBUG nova.objects.instance [None req-345ea84c-77bf-4393-9db3-8f6989cf9c1c tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Lazy-loading 'resources' on Instance uuid cf950ec3-9914-4eb9-99db-048b3969bb21 {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1033.655434] env[65726]: INFO nova.scheduler.client.report [None req-7aacd0be-ee58-4b95-8abe-402ca2fa658f tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Deleted allocations for instance dcde25df-87a5-47fb-94ce-334f68894e04 [ 1033.670030] env[65726]: DEBUG oslo_vmware.api [None req-f1e7868e-b3d9-4342-8972-f27e4cb4ec45 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52f7eb81-c9c1-6858-000a-e79e28e8df03, 'name': SearchDatastore_Task, 'duration_secs': 0.013907} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1033.670030] env[65726]: DEBUG oslo_concurrency.lockutils [None req-f1e7868e-b3d9-4342-8972-f27e4cb4ec45 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1033.670030] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-f1e7868e-b3d9-4342-8972-f27e4cb4ec45 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] 2b8eb60b-dc9b-47a4-9ddc-205f2af69888/2b8eb60b-dc9b-47a4-9ddc-205f2af69888.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1033.670030] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3427c3ba-af6d-4e6e-9c62-cf1edb1b6788 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.681889] env[65726]: DEBUG oslo_vmware.api [None req-f1e7868e-b3d9-4342-8972-f27e4cb4ec45 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] Waiting for the task: (returnval){ [ 1033.681889] env[65726]: value = "task-5116533" [ 1033.681889] env[65726]: _type = "Task" [ 1033.681889] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1033.698862] env[65726]: DEBUG nova.network.neutron [-] [instance: 067c9332-fa2c-47fe-93ae-643de1fe77cf] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1033.700230] env[65726]: DEBUG oslo_vmware.api [None req-f1e7868e-b3d9-4342-8972-f27e4cb4ec45 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] Task: {'id': task-5116533, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.755811] env[65726]: DEBUG oslo_concurrency.lockutils [req-ca08a4f2-4ea6-42fb-b87f-1c44fe2ea345 req-17642d9f-2a6c-4f06-8465-38c759b667f2 service nova] Releasing lock "refresh_cache-2b8eb60b-dc9b-47a4-9ddc-205f2af69888" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1033.922095] env[65726]: DEBUG nova.compute.manager [None req-0c1711f1-74cb-4ba7-a859-716c68812e37 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: ad4f00b5-3765-494f-8132-8b17b29d219a] Starting instance... {{(pid=65726) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 1033.963643] env[65726]: DEBUG oslo_vmware.api [None req-1d2791c0-c341-4061-b914-188004153713 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Task: {'id': task-5116529, 'name': MoveVirtualDisk_Task} progress is 38%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.010487] env[65726]: DEBUG oslo_vmware.api [None req-9ee0c86a-59b9-4c49-9232-0da097c0fa28 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Task: {'id': task-5116530, 'name': RemoveSnapshot_Task, 'duration_secs': 0.985014} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1034.014935] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-9ee0c86a-59b9-4c49-9232-0da097c0fa28 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] [instance: a286ba4f-a83e-4c30-8079-ba88e3ba5696] Deleted Snapshot of the VM instance {{(pid=65726) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1034.031909] env[65726]: DEBUG oslo_vmware.api [None req-a3acab1e-0b06-47ad-b154-32b257cd5af7 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': task-5116532, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.168658] env[65726]: DEBUG oslo_concurrency.lockutils [None req-7aacd0be-ee58-4b95-8abe-402ca2fa658f tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Lock "dcde25df-87a5-47fb-94ce-334f68894e04" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 16.108s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1034.168658] env[65726]: DEBUG oslo_concurrency.lockutils [req-81f65902-9a9f-4e96-9418-c4f581566ee4 req-b63b638c-1bb0-4bda-ae05-238c37d48441 service nova] Acquired lock "dcde25df-87a5-47fb-94ce-334f68894e04" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1034.169439] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f69daa41-2002-40d5-b8c9-ddf11f487bd3 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.190475] env[65726]: WARNING suds.client [-] Web service reported a SOAP processing fault using an unexpected HTTP status code 200. Reporting as an internal server error. [ 1034.190675] env[65726]: DEBUG oslo_vmware.api [-] Fault list: [ManagedObjectNotFound] {{(pid=65726) _invoke_api /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:337}} [ 1034.195785] env[65726]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-76587a4c-52cd-494a-a7db-c4490a55b1aa {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.204363] env[65726]: INFO nova.compute.manager [-] [instance: 067c9332-fa2c-47fe-93ae-643de1fe77cf] Took 1.34 seconds to deallocate network for instance. [ 1034.204363] env[65726]: DEBUG oslo_vmware.api [None req-f1e7868e-b3d9-4342-8972-f27e4cb4ec45 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] Task: {'id': task-5116533, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.219567] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b2798ac-cad1-41dd-890a-9925438bb48b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.268591] env[65726]: ERROR root [req-81f65902-9a9f-4e96-9418-c4f581566ee4 req-b63b638c-1bb0-4bda-ae05-238c37d48441 service nova] Original exception being dropped: ['Traceback (most recent call last):\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py", line 377, in request_handler\n response = request(managed_object, **kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/suds/client.py", line 586, in __call__\n return client.invoke(args, kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/suds/client.py", line 728, in invoke\n result = self.send(soapenv, timeout=timeout)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/suds/client.py", line 777, in send\n return self.process_reply(reply.message, None, None)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/suds/client.py", line 840, in process_reply\n raise WebFault(fault, replyroot)\n', "suds.WebFault: Server raised fault: 'The object 'vim.VirtualMachine:vm-995207' has already been deleted or has not been completely created'\n", '\nDuring handling of the above exception, another exception occurred:\n\n', 'Traceback (most recent call last):\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 301, in _invoke_api\n return api_method(*args, **kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/vim_util.py", line 479, in get_object_property\n props = get_object_properties(vim, moref, [property_name],\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/vim_util.py", line 360, in get_object_properties\n retrieve_result = vim.RetrievePropertiesEx(\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py", line 413, in request_handler\n raise exceptions.VimFaultException(fault_list, fault_string,\n', "oslo_vmware.exceptions.VimFaultException: The object 'vim.VirtualMachine:vm-995207' has already been deleted or has not been completely created\nCause: Server raised fault: 'The object 'vim.VirtualMachine:vm-995207' has already been deleted or has not been completely created'\nFaults: [ManagedObjectNotFound]\nDetails: {'obj': 'vm-995207'}\n", '\nDuring handling of the above exception, another exception occurred:\n\n', 'Traceback (most recent call last):\n', ' File "/opt/stack/nova/nova/virt/vmwareapi/session.py", line 123, in _call_method\n return self.invoke_api(module, method, self.vim, *args,\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 358, in invoke_api\n return _invoke_api(module, method, *args, **kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 122, in func\n return evt.wait()\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait\n result = hub.switch()\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch\n return self.greenlet.switch()\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/common/loopingcall.py", line 122, in _inner\n idle = self.f(*self.args, **self.kw)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 96, in _func\n result = f(*args, **kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 341, in _invoke_api\n raise clazz(str(excep),\n', "oslo_vmware.exceptions.ManagedObjectNotFoundException: The object 'vim.VirtualMachine:vm-995207' has already been deleted or has not been completely created\nCause: Server raised fault: 'The object 'vim.VirtualMachine:vm-995207' has already been deleted or has not been completely created'\nFaults: [ManagedObjectNotFound]\nDetails: {'obj': 'vm-995207'}\n"]: nova.exception.InstanceNotFound: Instance dcde25df-87a5-47fb-94ce-334f68894e04 could not be found. [ 1034.268929] env[65726]: DEBUG oslo_concurrency.lockutils [req-81f65902-9a9f-4e96-9418-c4f581566ee4 req-b63b638c-1bb0-4bda-ae05-238c37d48441 service nova] Releasing lock "dcde25df-87a5-47fb-94ce-334f68894e04" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1034.269037] env[65726]: DEBUG nova.compute.manager [req-81f65902-9a9f-4e96-9418-c4f581566ee4 req-b63b638c-1bb0-4bda-ae05-238c37d48441 service nova] [instance: dcde25df-87a5-47fb-94ce-334f68894e04] Detach interface failed, port_id=280bd1f3-a22e-4f59-8574-d1783c74f8f3, reason: Instance dcde25df-87a5-47fb-94ce-334f68894e04 could not be found. {{(pid=65726) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11669}} [ 1034.354693] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-72c48a25-f856-4f68-98a0-3ea61ba329e4 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] [instance: aa6ce489-c62f-4481-87b7-e74242aeb8ca] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1034.355794] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-58ef1d31-e691-4894-90b9-a8df63618002 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.367253] env[65726]: DEBUG oslo_vmware.api [None req-72c48a25-f856-4f68-98a0-3ea61ba329e4 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Waiting for the task: (returnval){ [ 1034.367253] env[65726]: value = "task-5116534" [ 1034.367253] env[65726]: _type = "Task" [ 1034.367253] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1034.389474] env[65726]: DEBUG oslo_vmware.api [None req-72c48a25-f856-4f68-98a0-3ea61ba329e4 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Task: {'id': task-5116534, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.411809] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c91b10c2-b170-4395-aeda-82a4d77adcf6 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.425233] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94e03848-3e16-40ed-8578-acf0130fef8e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.487977] env[65726]: DEBUG oslo_concurrency.lockutils [None req-0c1711f1-74cb-4ba7-a859-716c68812e37 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1034.493012] env[65726]: DEBUG nova.compute.manager [req-82a7e2c4-a5c2-4772-b1ed-b30220325995 req-425a5fad-2dab-4cdc-9c13-93447d11b81a service nova] [instance: 2c812714-dccc-4d1e-bdb5-c11e446949c3] Received event network-vif-plugged-203fa726-d08f-47a8-a7cd-389ed31da940 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1034.493314] env[65726]: DEBUG oslo_concurrency.lockutils [req-82a7e2c4-a5c2-4772-b1ed-b30220325995 req-425a5fad-2dab-4cdc-9c13-93447d11b81a service nova] Acquiring lock "2c812714-dccc-4d1e-bdb5-c11e446949c3-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1034.493895] env[65726]: DEBUG oslo_concurrency.lockutils [req-82a7e2c4-a5c2-4772-b1ed-b30220325995 req-425a5fad-2dab-4cdc-9c13-93447d11b81a service nova] Lock "2c812714-dccc-4d1e-bdb5-c11e446949c3-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1034.493895] env[65726]: DEBUG oslo_concurrency.lockutils [req-82a7e2c4-a5c2-4772-b1ed-b30220325995 req-425a5fad-2dab-4cdc-9c13-93447d11b81a service nova] Lock "2c812714-dccc-4d1e-bdb5-c11e446949c3-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1034.493895] env[65726]: DEBUG nova.compute.manager [req-82a7e2c4-a5c2-4772-b1ed-b30220325995 req-425a5fad-2dab-4cdc-9c13-93447d11b81a service nova] [instance: 2c812714-dccc-4d1e-bdb5-c11e446949c3] No waiting events found dispatching network-vif-plugged-203fa726-d08f-47a8-a7cd-389ed31da940 {{(pid=65726) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1034.494283] env[65726]: WARNING nova.compute.manager [req-82a7e2c4-a5c2-4772-b1ed-b30220325995 req-425a5fad-2dab-4cdc-9c13-93447d11b81a service nova] [instance: 2c812714-dccc-4d1e-bdb5-c11e446949c3] Received unexpected event network-vif-plugged-203fa726-d08f-47a8-a7cd-389ed31da940 for instance with vm_state building and task_state spawning. [ 1034.494283] env[65726]: DEBUG nova.compute.manager [req-82a7e2c4-a5c2-4772-b1ed-b30220325995 req-425a5fad-2dab-4cdc-9c13-93447d11b81a service nova] [instance: 2c812714-dccc-4d1e-bdb5-c11e446949c3] Received event network-changed-203fa726-d08f-47a8-a7cd-389ed31da940 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1034.494592] env[65726]: DEBUG nova.compute.manager [req-82a7e2c4-a5c2-4772-b1ed-b30220325995 req-425a5fad-2dab-4cdc-9c13-93447d11b81a service nova] [instance: 2c812714-dccc-4d1e-bdb5-c11e446949c3] Refreshing instance network info cache due to event network-changed-203fa726-d08f-47a8-a7cd-389ed31da940. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 1034.494592] env[65726]: DEBUG oslo_concurrency.lockutils [req-82a7e2c4-a5c2-4772-b1ed-b30220325995 req-425a5fad-2dab-4cdc-9c13-93447d11b81a service nova] Acquiring lock "refresh_cache-2c812714-dccc-4d1e-bdb5-c11e446949c3" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1034.494811] env[65726]: DEBUG oslo_concurrency.lockutils [req-82a7e2c4-a5c2-4772-b1ed-b30220325995 req-425a5fad-2dab-4cdc-9c13-93447d11b81a service nova] Acquired lock "refresh_cache-2c812714-dccc-4d1e-bdb5-c11e446949c3" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1034.494894] env[65726]: DEBUG nova.network.neutron [req-82a7e2c4-a5c2-4772-b1ed-b30220325995 req-425a5fad-2dab-4cdc-9c13-93447d11b81a service nova] [instance: 2c812714-dccc-4d1e-bdb5-c11e446949c3] Refreshing network info cache for port 203fa726-d08f-47a8-a7cd-389ed31da940 {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 1034.496785] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9dbe921-10fa-44d3-9781-634a8f5e23f0 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.510388] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7c6fa62-c520-4a43-9001-5ae2a6d58865 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.518663] env[65726]: DEBUG oslo_vmware.api [None req-1d2791c0-c341-4061-b914-188004153713 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Task: {'id': task-5116529, 'name': MoveVirtualDisk_Task} progress is 60%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.523339] env[65726]: WARNING nova.compute.manager [None req-9ee0c86a-59b9-4c49-9232-0da097c0fa28 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] [instance: a286ba4f-a83e-4c30-8079-ba88e3ba5696] Image not found during snapshot: nova.exception.ImageNotFound: Image 46a5e33a-19cd-4fd5-8455-a08423252044 could not be found. [ 1034.536389] env[65726]: DEBUG nova.compute.provider_tree [None req-345ea84c-77bf-4393-9db3-8f6989cf9c1c tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1034.546565] env[65726]: DEBUG oslo_vmware.api [None req-a3acab1e-0b06-47ad-b154-32b257cd5af7 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': task-5116532, 'name': DeleteDatastoreFile_Task, 'duration_secs': 1.03403} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1034.547530] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-a3acab1e-0b06-47ad-b154-32b257cd5af7 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Deleted the datastore file {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1034.547732] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-a3acab1e-0b06-47ad-b154-32b257cd5af7 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: fb8eac12-aedd-426f-9dd3-6c52f9d4c71d] Deleted contents of the VM from datastore datastore2 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1034.547933] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-a3acab1e-0b06-47ad-b154-32b257cd5af7 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: fb8eac12-aedd-426f-9dd3-6c52f9d4c71d] Instance destroyed {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1034.674163] env[65726]: DEBUG nova.network.neutron [None req-2a9dd832-d188-4542-a7fe-bbdea8173b2d tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] [instance: 2c812714-dccc-4d1e-bdb5-c11e446949c3] Successfully updated port: 8addcb43-85bf-481e-92bc-5bbdf48c0ac4 {{(pid=65726) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 1034.700937] env[65726]: DEBUG oslo_vmware.api [None req-f1e7868e-b3d9-4342-8972-f27e4cb4ec45 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] Task: {'id': task-5116533, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.883243] env[65726]: DEBUG oslo_vmware.api [None req-72c48a25-f856-4f68-98a0-3ea61ba329e4 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Task: {'id': task-5116534, 'name': PowerOffVM_Task, 'duration_secs': 0.291217} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1034.883567] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-72c48a25-f856-4f68-98a0-3ea61ba329e4 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] [instance: aa6ce489-c62f-4481-87b7-e74242aeb8ca] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1034.883761] env[65726]: DEBUG nova.compute.manager [None req-72c48a25-f856-4f68-98a0-3ea61ba329e4 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] [instance: aa6ce489-c62f-4481-87b7-e74242aeb8ca] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1034.884599] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c7a6d50-51c1-4f0c-8f63-13b898586aa4 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.998642] env[65726]: DEBUG oslo_vmware.api [None req-1d2791c0-c341-4061-b914-188004153713 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Task: {'id': task-5116529, 'name': MoveVirtualDisk_Task} progress is 80%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.001789] env[65726]: WARNING openstack [req-82a7e2c4-a5c2-4772-b1ed-b30220325995 req-425a5fad-2dab-4cdc-9c13-93447d11b81a service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1035.002303] env[65726]: WARNING openstack [req-82a7e2c4-a5c2-4772-b1ed-b30220325995 req-425a5fad-2dab-4cdc-9c13-93447d11b81a service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1035.007755] env[65726]: DEBUG nova.network.neutron [req-82a7e2c4-a5c2-4772-b1ed-b30220325995 req-425a5fad-2dab-4cdc-9c13-93447d11b81a service nova] [instance: 2c812714-dccc-4d1e-bdb5-c11e446949c3] Instance cache missing network info. {{(pid=65726) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 1035.040407] env[65726]: DEBUG nova.scheduler.client.report [None req-345ea84c-77bf-4393-9db3-8f6989cf9c1c tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1035.070464] env[65726]: DEBUG oslo_concurrency.lockutils [None req-43be9038-9e49-4e4d-82ea-842afcd9a64d tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Acquiring lock "a286ba4f-a83e-4c30-8079-ba88e3ba5696" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1035.070754] env[65726]: DEBUG oslo_concurrency.lockutils [None req-43be9038-9e49-4e4d-82ea-842afcd9a64d tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Lock "a286ba4f-a83e-4c30-8079-ba88e3ba5696" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1035.070966] env[65726]: DEBUG oslo_concurrency.lockutils [None req-43be9038-9e49-4e4d-82ea-842afcd9a64d tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Acquiring lock "a286ba4f-a83e-4c30-8079-ba88e3ba5696-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1035.071164] env[65726]: DEBUG oslo_concurrency.lockutils [None req-43be9038-9e49-4e4d-82ea-842afcd9a64d tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Lock "a286ba4f-a83e-4c30-8079-ba88e3ba5696-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1035.071395] env[65726]: DEBUG oslo_concurrency.lockutils [None req-43be9038-9e49-4e4d-82ea-842afcd9a64d tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Lock "a286ba4f-a83e-4c30-8079-ba88e3ba5696-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1035.073957] env[65726]: INFO nova.compute.manager [None req-43be9038-9e49-4e4d-82ea-842afcd9a64d tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] [instance: a286ba4f-a83e-4c30-8079-ba88e3ba5696] Terminating instance [ 1035.168052] env[65726]: DEBUG nova.network.neutron [req-82a7e2c4-a5c2-4772-b1ed-b30220325995 req-425a5fad-2dab-4cdc-9c13-93447d11b81a service nova] [instance: 2c812714-dccc-4d1e-bdb5-c11e446949c3] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1035.176598] env[65726]: DEBUG oslo_concurrency.lockutils [None req-2a9dd832-d188-4542-a7fe-bbdea8173b2d tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Acquiring lock "refresh_cache-2c812714-dccc-4d1e-bdb5-c11e446949c3" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1035.198913] env[65726]: DEBUG oslo_vmware.api [None req-f1e7868e-b3d9-4342-8972-f27e4cb4ec45 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] Task: {'id': task-5116533, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.242475] env[65726]: DEBUG oslo_concurrency.lockutils [None req-8ffe452d-09ce-47ef-a349-6236507a111f tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Lock "067c9332-fa2c-47fe-93ae-643de1fe77cf" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 2.998s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1035.403395] env[65726]: DEBUG oslo_concurrency.lockutils [None req-72c48a25-f856-4f68-98a0-3ea61ba329e4 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Lock "aa6ce489-c62f-4481-87b7-e74242aeb8ca" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.070s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1035.504836] env[65726]: DEBUG oslo_vmware.api [None req-1d2791c0-c341-4061-b914-188004153713 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Task: {'id': task-5116529, 'name': MoveVirtualDisk_Task} progress is 100%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.546165] env[65726]: DEBUG oslo_concurrency.lockutils [None req-345ea84c-77bf-4393-9db3-8f6989cf9c1c tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.914s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1035.549370] env[65726]: DEBUG oslo_concurrency.lockutils [None req-10701ce3-ca7d-4146-b0df-5677afb5eba0 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.417s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1035.551400] env[65726]: INFO nova.compute.claims [None req-10701ce3-ca7d-4146-b0df-5677afb5eba0 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] [instance: df08d166-7811-49b5-9c16-a3434229d623] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1035.576264] env[65726]: INFO nova.scheduler.client.report [None req-345ea84c-77bf-4393-9db3-8f6989cf9c1c tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Deleted allocations for instance cf950ec3-9914-4eb9-99db-048b3969bb21 [ 1035.577739] env[65726]: DEBUG nova.compute.manager [None req-43be9038-9e49-4e4d-82ea-842afcd9a64d tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] [instance: a286ba4f-a83e-4c30-8079-ba88e3ba5696] Start destroying the instance on the hypervisor. {{(pid=65726) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1035.578052] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-43be9038-9e49-4e4d-82ea-842afcd9a64d tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] [instance: a286ba4f-a83e-4c30-8079-ba88e3ba5696] Destroying instance {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1035.580569] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0b41b77-803b-44d5-84c3-347d2a7af66e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.599045] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-43be9038-9e49-4e4d-82ea-842afcd9a64d tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] [instance: a286ba4f-a83e-4c30-8079-ba88e3ba5696] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1035.599371] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9f030b31-5c54-4b8b-ac5b-e682d01c7511 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.612786] env[65726]: DEBUG oslo_vmware.api [None req-43be9038-9e49-4e4d-82ea-842afcd9a64d tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Waiting for the task: (returnval){ [ 1035.612786] env[65726]: value = "task-5116535" [ 1035.612786] env[65726]: _type = "Task" [ 1035.612786] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1035.620567] env[65726]: DEBUG nova.virt.hardware [None req-a3acab1e-0b06-47ad-b154-32b257cd5af7 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T19:28:34Z,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1539f06fa3534e90acbc3a60c4b8bd3f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T19:28:35Z,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1035.620895] env[65726]: DEBUG nova.virt.hardware [None req-a3acab1e-0b06-47ad-b154-32b257cd5af7 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1035.621117] env[65726]: DEBUG nova.virt.hardware [None req-a3acab1e-0b06-47ad-b154-32b257cd5af7 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1035.621791] env[65726]: DEBUG nova.virt.hardware [None req-a3acab1e-0b06-47ad-b154-32b257cd5af7 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1035.621791] env[65726]: DEBUG nova.virt.hardware [None req-a3acab1e-0b06-47ad-b154-32b257cd5af7 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1035.621990] env[65726]: DEBUG nova.virt.hardware [None req-a3acab1e-0b06-47ad-b154-32b257cd5af7 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1035.622421] env[65726]: DEBUG nova.virt.hardware [None req-a3acab1e-0b06-47ad-b154-32b257cd5af7 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1035.622421] env[65726]: DEBUG nova.virt.hardware [None req-a3acab1e-0b06-47ad-b154-32b257cd5af7 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1035.622569] env[65726]: DEBUG nova.virt.hardware [None req-a3acab1e-0b06-47ad-b154-32b257cd5af7 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1035.622761] env[65726]: DEBUG nova.virt.hardware [None req-a3acab1e-0b06-47ad-b154-32b257cd5af7 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1035.622989] env[65726]: DEBUG nova.virt.hardware [None req-a3acab1e-0b06-47ad-b154-32b257cd5af7 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1035.624043] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8b90bf5-3920-4cbc-bade-04169d413096 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.638296] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b03a27a-bcc5-491d-8830-10943cb429e7 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.645048] env[65726]: DEBUG oslo_vmware.api [None req-43be9038-9e49-4e4d-82ea-842afcd9a64d tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Task: {'id': task-5116535, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.662273] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-a3acab1e-0b06-47ad-b154-32b257cd5af7 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: fb8eac12-aedd-426f-9dd3-6c52f9d4c71d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:3d:2a:cb', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '74f30339-6421-4654-bddb-81d7f34db9d7', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '839bf297-99c8-450e-9cd5-546932e1bea1', 'vif_model': 'vmxnet3'}] {{(pid=65726) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1035.671868] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-a3acab1e-0b06-47ad-b154-32b257cd5af7 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1035.673111] env[65726]: DEBUG oslo_concurrency.lockutils [req-82a7e2c4-a5c2-4772-b1ed-b30220325995 req-425a5fad-2dab-4cdc-9c13-93447d11b81a service nova] Releasing lock "refresh_cache-2c812714-dccc-4d1e-bdb5-c11e446949c3" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1035.673517] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fb8eac12-aedd-426f-9dd3-6c52f9d4c71d] Creating VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1035.673871] env[65726]: DEBUG oslo_concurrency.lockutils [None req-2a9dd832-d188-4542-a7fe-bbdea8173b2d tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Acquired lock "refresh_cache-2c812714-dccc-4d1e-bdb5-c11e446949c3" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1035.674073] env[65726]: DEBUG nova.network.neutron [None req-2a9dd832-d188-4542-a7fe-bbdea8173b2d tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] [instance: 2c812714-dccc-4d1e-bdb5-c11e446949c3] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1035.675768] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9fc2d975-628d-4835-a347-a9c6495d89f0 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.700923] env[65726]: WARNING openstack [None req-2a9dd832-d188-4542-a7fe-bbdea8173b2d tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1035.701428] env[65726]: WARNING openstack [None req-2a9dd832-d188-4542-a7fe-bbdea8173b2d tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1035.709451] env[65726]: DEBUG nova.network.neutron [None req-2a9dd832-d188-4542-a7fe-bbdea8173b2d tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] [instance: 2c812714-dccc-4d1e-bdb5-c11e446949c3] Instance cache missing network info. {{(pid=65726) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 1035.726230] env[65726]: DEBUG oslo_vmware.api [None req-f1e7868e-b3d9-4342-8972-f27e4cb4ec45 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] Task: {'id': task-5116533, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.729998] env[65726]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1035.729998] env[65726]: value = "task-5116536" [ 1035.729998] env[65726]: _type = "Task" [ 1035.729998] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1035.743484] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116536, 'name': CreateVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.796753] env[65726]: DEBUG nova.objects.instance [None req-76e5406a-1ee2-4318-b0b7-91862308e4d0 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Lazy-loading 'flavor' on Instance uuid aa6ce489-c62f-4481-87b7-e74242aeb8ca {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1035.800104] env[65726]: WARNING openstack [None req-2a9dd832-d188-4542-a7fe-bbdea8173b2d tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1035.800755] env[65726]: WARNING openstack [None req-2a9dd832-d188-4542-a7fe-bbdea8173b2d tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1035.873741] env[65726]: WARNING openstack [None req-2a9dd832-d188-4542-a7fe-bbdea8173b2d tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1035.874213] env[65726]: WARNING openstack [None req-2a9dd832-d188-4542-a7fe-bbdea8173b2d tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1035.996947] env[65726]: DEBUG oslo_vmware.api [None req-1d2791c0-c341-4061-b914-188004153713 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Task: {'id': task-5116529, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.846125} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1035.997358] env[65726]: INFO nova.virt.vmwareapi.ds_util [None req-1d2791c0-c341-4061-b914-188004153713 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Moved virtual disk from [datastore1] OSTACK_IMG_866b99ef-d360-4d77-8ef7-e8f0b892d507/OSTACK_IMG_866b99ef-d360-4d77-8ef7-e8f0b892d507.vmdk to [datastore1] devstack-image-cache_base/81f5ce39-0bc0-4d0e-ab15-42a80a4f614b/81f5ce39-0bc0-4d0e-ab15-42a80a4f614b.vmdk. [ 1035.997584] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-1d2791c0-c341-4061-b914-188004153713 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04] Cleaning up location [datastore1] OSTACK_IMG_866b99ef-d360-4d77-8ef7-e8f0b892d507 {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 1035.997778] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-1d2791c0-c341-4061-b914-188004153713 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Deleting the datastore file [datastore1] OSTACK_IMG_866b99ef-d360-4d77-8ef7-e8f0b892d507 {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1035.998588] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-df15e4be-258a-47ba-a91b-bb6abf4eca14 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.008837] env[65726]: DEBUG oslo_vmware.api [None req-1d2791c0-c341-4061-b914-188004153713 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Waiting for the task: (returnval){ [ 1036.008837] env[65726]: value = "task-5116537" [ 1036.008837] env[65726]: _type = "Task" [ 1036.008837] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1036.019325] env[65726]: DEBUG oslo_vmware.api [None req-1d2791c0-c341-4061-b914-188004153713 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Task: {'id': task-5116537, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.020963] env[65726]: WARNING openstack [None req-2a9dd832-d188-4542-a7fe-bbdea8173b2d tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1036.021230] env[65726]: WARNING openstack [None req-2a9dd832-d188-4542-a7fe-bbdea8173b2d tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1036.095678] env[65726]: DEBUG oslo_concurrency.lockutils [None req-345ea84c-77bf-4393-9db3-8f6989cf9c1c tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Lock "cf950ec3-9914-4eb9-99db-048b3969bb21" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 17.395s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1036.103802] env[65726]: WARNING openstack [None req-2a9dd832-d188-4542-a7fe-bbdea8173b2d tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1036.104165] env[65726]: WARNING openstack [None req-2a9dd832-d188-4542-a7fe-bbdea8173b2d tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1036.123519] env[65726]: DEBUG oslo_vmware.api [None req-43be9038-9e49-4e4d-82ea-842afcd9a64d tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Task: {'id': task-5116535, 'name': PowerOffVM_Task, 'duration_secs': 0.242006} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1036.123784] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-43be9038-9e49-4e4d-82ea-842afcd9a64d tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] [instance: a286ba4f-a83e-4c30-8079-ba88e3ba5696] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1036.123952] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-43be9038-9e49-4e4d-82ea-842afcd9a64d tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] [instance: a286ba4f-a83e-4c30-8079-ba88e3ba5696] Unregistering the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1036.124252] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-03c11c59-1abb-41e1-93fa-0647db5ec498 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.194616] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-43be9038-9e49-4e4d-82ea-842afcd9a64d tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] [instance: a286ba4f-a83e-4c30-8079-ba88e3ba5696] Unregistered the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1036.194900] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-43be9038-9e49-4e4d-82ea-842afcd9a64d tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] [instance: a286ba4f-a83e-4c30-8079-ba88e3ba5696] Deleting contents of the VM from datastore datastore2 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1036.195339] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-43be9038-9e49-4e4d-82ea-842afcd9a64d tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Deleting the datastore file [datastore2] a286ba4f-a83e-4c30-8079-ba88e3ba5696 {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1036.195906] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-175c5df8-c476-4552-83d5-ab39be6d3661 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.205129] env[65726]: DEBUG oslo_vmware.api [None req-43be9038-9e49-4e4d-82ea-842afcd9a64d tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Waiting for the task: (returnval){ [ 1036.205129] env[65726]: value = "task-5116539" [ 1036.205129] env[65726]: _type = "Task" [ 1036.205129] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1036.228156] env[65726]: DEBUG oslo_vmware.api [None req-43be9038-9e49-4e4d-82ea-842afcd9a64d tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Task: {'id': task-5116539, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.235317] env[65726]: DEBUG oslo_vmware.api [None req-f1e7868e-b3d9-4342-8972-f27e4cb4ec45 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] Task: {'id': task-5116533, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.166487} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1036.236410] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-f1e7868e-b3d9-4342-8972-f27e4cb4ec45 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] 2b8eb60b-dc9b-47a4-9ddc-205f2af69888/2b8eb60b-dc9b-47a4-9ddc-205f2af69888.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1036.236694] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-f1e7868e-b3d9-4342-8972-f27e4cb4ec45 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] [instance: 2b8eb60b-dc9b-47a4-9ddc-205f2af69888] Extending root virtual disk to 1048576 {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1036.236954] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-aa651cd9-64a0-4ea3-a270-8adcb6f6a050 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.243936] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116536, 'name': CreateVM_Task} progress is 99%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.250213] env[65726]: DEBUG oslo_vmware.api [None req-f1e7868e-b3d9-4342-8972-f27e4cb4ec45 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] Waiting for the task: (returnval){ [ 1036.250213] env[65726]: value = "task-5116540" [ 1036.250213] env[65726]: _type = "Task" [ 1036.250213] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1036.261782] env[65726]: DEBUG oslo_vmware.api [None req-f1e7868e-b3d9-4342-8972-f27e4cb4ec45 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] Task: {'id': task-5116540, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.276133] env[65726]: DEBUG nova.network.neutron [None req-2a9dd832-d188-4542-a7fe-bbdea8173b2d tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] [instance: 2c812714-dccc-4d1e-bdb5-c11e446949c3] Updating instance_info_cache with network_info: [{"id": "203fa726-d08f-47a8-a7cd-389ed31da940", "address": "fa:16:3e:0b:bd:69", "network": {"id": "f9d433a7-98f6-4013-ba92-7bbacbef4969", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1784765821", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.175", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "6961c16aa6a346679bbde25930b2a5b7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7ab8d568-adb0-4f3b-b6cc-68413e6546ae", "external-id": "nsx-vlan-transportzone-86", "segmentation_id": 86, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap203fa726-d0", "ovs_interfaceid": "203fa726-d08f-47a8-a7cd-389ed31da940", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "8addcb43-85bf-481e-92bc-5bbdf48c0ac4", "address": "fa:16:3e:4f:bc:c3", "network": {"id": "deb5277b-48b1-472a-b940-d808da6ffa26", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1186999789", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.239", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "6961c16aa6a346679bbde25930b2a5b7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7adad264-9276-43ef-9b03-07dc27d3f81e", "external-id": "nsx-vlan-transportzone-329", "segmentation_id": 329, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8addcb43-85", "ovs_interfaceid": "8addcb43-85bf-481e-92bc-5bbdf48c0ac4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1036.313941] env[65726]: DEBUG oslo_concurrency.lockutils [None req-76e5406a-1ee2-4318-b0b7-91862308e4d0 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Acquiring lock "refresh_cache-aa6ce489-c62f-4481-87b7-e74242aeb8ca" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1036.315617] env[65726]: DEBUG oslo_concurrency.lockutils [None req-76e5406a-1ee2-4318-b0b7-91862308e4d0 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Acquired lock "refresh_cache-aa6ce489-c62f-4481-87b7-e74242aeb8ca" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1036.315792] env[65726]: DEBUG nova.network.neutron [None req-76e5406a-1ee2-4318-b0b7-91862308e4d0 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] [instance: aa6ce489-c62f-4481-87b7-e74242aeb8ca] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1036.316266] env[65726]: DEBUG nova.objects.instance [None req-76e5406a-1ee2-4318-b0b7-91862308e4d0 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Lazy-loading 'info_cache' on Instance uuid aa6ce489-c62f-4481-87b7-e74242aeb8ca {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1036.519346] env[65726]: DEBUG oslo_vmware.api [None req-1d2791c0-c341-4061-b914-188004153713 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Task: {'id': task-5116537, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.069865} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1036.519631] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-1d2791c0-c341-4061-b914-188004153713 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Deleted the datastore file {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1036.519814] env[65726]: DEBUG oslo_concurrency.lockutils [None req-1d2791c0-c341-4061-b914-188004153713 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Releasing lock "[datastore1] devstack-image-cache_base/81f5ce39-0bc0-4d0e-ab15-42a80a4f614b/81f5ce39-0bc0-4d0e-ab15-42a80a4f614b.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1036.520078] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-1d2791c0-c341-4061-b914-188004153713 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/81f5ce39-0bc0-4d0e-ab15-42a80a4f614b/81f5ce39-0bc0-4d0e-ab15-42a80a4f614b.vmdk to [datastore1] 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04/2fe6fc5a-f5c7-4f8e-96df-4e621a252f04.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1036.520389] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c2e82fb6-8312-4b60-bc14-2b37b2bed5ed {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.528080] env[65726]: DEBUG oslo_vmware.api [None req-1d2791c0-c341-4061-b914-188004153713 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Waiting for the task: (returnval){ [ 1036.528080] env[65726]: value = "task-5116541" [ 1036.528080] env[65726]: _type = "Task" [ 1036.528080] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1036.536895] env[65726]: DEBUG oslo_vmware.api [None req-1d2791c0-c341-4061-b914-188004153713 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Task: {'id': task-5116541, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.722222] env[65726]: DEBUG oslo_vmware.api [None req-43be9038-9e49-4e4d-82ea-842afcd9a64d tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Task: {'id': task-5116539, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.184106} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1036.725478] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-43be9038-9e49-4e4d-82ea-842afcd9a64d tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Deleted the datastore file {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1036.725909] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-43be9038-9e49-4e4d-82ea-842afcd9a64d tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] [instance: a286ba4f-a83e-4c30-8079-ba88e3ba5696] Deleted contents of the VM from datastore datastore2 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1036.726291] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-43be9038-9e49-4e4d-82ea-842afcd9a64d tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] [instance: a286ba4f-a83e-4c30-8079-ba88e3ba5696] Instance destroyed {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1036.726291] env[65726]: INFO nova.compute.manager [None req-43be9038-9e49-4e4d-82ea-842afcd9a64d tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] [instance: a286ba4f-a83e-4c30-8079-ba88e3ba5696] Took 1.15 seconds to destroy the instance on the hypervisor. [ 1036.726599] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-43be9038-9e49-4e4d-82ea-842afcd9a64d tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1036.727885] env[65726]: DEBUG nova.compute.manager [-] [instance: a286ba4f-a83e-4c30-8079-ba88e3ba5696] Deallocating network for instance {{(pid=65726) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1036.727885] env[65726]: DEBUG nova.network.neutron [-] [instance: a286ba4f-a83e-4c30-8079-ba88e3ba5696] deallocate_for_instance() {{(pid=65726) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 1036.727885] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1036.728145] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1036.747155] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116536, 'name': CreateVM_Task} progress is 99%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.763400] env[65726]: DEBUG oslo_vmware.api [None req-f1e7868e-b3d9-4342-8972-f27e4cb4ec45 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] Task: {'id': task-5116540, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.769347] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1036.769690] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1036.779340] env[65726]: DEBUG oslo_concurrency.lockutils [None req-2a9dd832-d188-4542-a7fe-bbdea8173b2d tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Releasing lock "refresh_cache-2c812714-dccc-4d1e-bdb5-c11e446949c3" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1036.779776] env[65726]: DEBUG nova.compute.manager [None req-2a9dd832-d188-4542-a7fe-bbdea8173b2d tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] [instance: 2c812714-dccc-4d1e-bdb5-c11e446949c3] Instance network_info: |[{"id": "203fa726-d08f-47a8-a7cd-389ed31da940", "address": "fa:16:3e:0b:bd:69", "network": {"id": "f9d433a7-98f6-4013-ba92-7bbacbef4969", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1784765821", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.175", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "6961c16aa6a346679bbde25930b2a5b7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7ab8d568-adb0-4f3b-b6cc-68413e6546ae", "external-id": "nsx-vlan-transportzone-86", "segmentation_id": 86, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap203fa726-d0", "ovs_interfaceid": "203fa726-d08f-47a8-a7cd-389ed31da940", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "8addcb43-85bf-481e-92bc-5bbdf48c0ac4", "address": "fa:16:3e:4f:bc:c3", "network": {"id": "deb5277b-48b1-472a-b940-d808da6ffa26", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1186999789", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.239", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "6961c16aa6a346679bbde25930b2a5b7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7adad264-9276-43ef-9b03-07dc27d3f81e", "external-id": "nsx-vlan-transportzone-329", "segmentation_id": 329, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8addcb43-85", "ovs_interfaceid": "8addcb43-85bf-481e-92bc-5bbdf48c0ac4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1036.781170] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-2a9dd832-d188-4542-a7fe-bbdea8173b2d tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] [instance: 2c812714-dccc-4d1e-bdb5-c11e446949c3] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:0b:bd:69', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '7ab8d568-adb0-4f3b-b6cc-68413e6546ae', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '203fa726-d08f-47a8-a7cd-389ed31da940', 'vif_model': 'vmxnet3'}, {'network_name': 'br-int', 'mac_address': 'fa:16:3e:4f:bc:c3', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '7adad264-9276-43ef-9b03-07dc27d3f81e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '8addcb43-85bf-481e-92bc-5bbdf48c0ac4', 'vif_model': 'vmxnet3'}] {{(pid=65726) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1036.790269] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-2a9dd832-d188-4542-a7fe-bbdea8173b2d tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1036.791570] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2c812714-dccc-4d1e-bdb5-c11e446949c3] Creating VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1036.792401] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-023ce70d-3d52-44ef-ae14-6363bedf0710 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.795522] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-807f794e-3331-406a-8268-bfae3f950bed {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.823397] env[65726]: DEBUG nova.objects.base [None req-76e5406a-1ee2-4318-b0b7-91862308e4d0 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Object Instance lazy-loaded attributes: flavor,info_cache {{(pid=65726) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1036.829199] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e018fae-0ce9-45e3-958f-e268031b1303 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.833905] env[65726]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1036.833905] env[65726]: value = "task-5116542" [ 1036.833905] env[65726]: _type = "Task" [ 1036.833905] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1036.870153] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10215615-5a71-4881-9d89-b1bfbf2a9a70 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.873384] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116542, 'name': CreateVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.879699] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3063d761-8eb8-42ae-88f7-b3f7c2870bc8 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.896683] env[65726]: DEBUG nova.compute.provider_tree [None req-10701ce3-ca7d-4146-b0df-5677afb5eba0 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1037.039940] env[65726]: DEBUG oslo_vmware.api [None req-1d2791c0-c341-4061-b914-188004153713 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Task: {'id': task-5116541, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.109267] env[65726]: DEBUG nova.compute.manager [req-090da2c3-13b5-4d5b-bea5-cf65b985fd32 req-b83d43d2-9504-4125-9aa4-c0aeecbc1e9f service nova] [instance: 2c812714-dccc-4d1e-bdb5-c11e446949c3] Received event network-vif-plugged-8addcb43-85bf-481e-92bc-5bbdf48c0ac4 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1037.109491] env[65726]: DEBUG oslo_concurrency.lockutils [req-090da2c3-13b5-4d5b-bea5-cf65b985fd32 req-b83d43d2-9504-4125-9aa4-c0aeecbc1e9f service nova] Acquiring lock "2c812714-dccc-4d1e-bdb5-c11e446949c3-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1037.109762] env[65726]: DEBUG oslo_concurrency.lockutils [req-090da2c3-13b5-4d5b-bea5-cf65b985fd32 req-b83d43d2-9504-4125-9aa4-c0aeecbc1e9f service nova] Lock "2c812714-dccc-4d1e-bdb5-c11e446949c3-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1037.109929] env[65726]: DEBUG oslo_concurrency.lockutils [req-090da2c3-13b5-4d5b-bea5-cf65b985fd32 req-b83d43d2-9504-4125-9aa4-c0aeecbc1e9f service nova] Lock "2c812714-dccc-4d1e-bdb5-c11e446949c3-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1037.111236] env[65726]: DEBUG nova.compute.manager [req-090da2c3-13b5-4d5b-bea5-cf65b985fd32 req-b83d43d2-9504-4125-9aa4-c0aeecbc1e9f service nova] [instance: 2c812714-dccc-4d1e-bdb5-c11e446949c3] No waiting events found dispatching network-vif-plugged-8addcb43-85bf-481e-92bc-5bbdf48c0ac4 {{(pid=65726) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1037.111560] env[65726]: WARNING nova.compute.manager [req-090da2c3-13b5-4d5b-bea5-cf65b985fd32 req-b83d43d2-9504-4125-9aa4-c0aeecbc1e9f service nova] [instance: 2c812714-dccc-4d1e-bdb5-c11e446949c3] Received unexpected event network-vif-plugged-8addcb43-85bf-481e-92bc-5bbdf48c0ac4 for instance with vm_state building and task_state spawning. [ 1037.111685] env[65726]: DEBUG nova.compute.manager [req-090da2c3-13b5-4d5b-bea5-cf65b985fd32 req-b83d43d2-9504-4125-9aa4-c0aeecbc1e9f service nova] [instance: 2c812714-dccc-4d1e-bdb5-c11e446949c3] Received event network-changed-8addcb43-85bf-481e-92bc-5bbdf48c0ac4 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1037.114615] env[65726]: DEBUG nova.compute.manager [req-090da2c3-13b5-4d5b-bea5-cf65b985fd32 req-b83d43d2-9504-4125-9aa4-c0aeecbc1e9f service nova] [instance: 2c812714-dccc-4d1e-bdb5-c11e446949c3] Refreshing instance network info cache due to event network-changed-8addcb43-85bf-481e-92bc-5bbdf48c0ac4. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 1037.114615] env[65726]: DEBUG oslo_concurrency.lockutils [req-090da2c3-13b5-4d5b-bea5-cf65b985fd32 req-b83d43d2-9504-4125-9aa4-c0aeecbc1e9f service nova] Acquiring lock "refresh_cache-2c812714-dccc-4d1e-bdb5-c11e446949c3" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1037.114615] env[65726]: DEBUG oslo_concurrency.lockutils [req-090da2c3-13b5-4d5b-bea5-cf65b985fd32 req-b83d43d2-9504-4125-9aa4-c0aeecbc1e9f service nova] Acquired lock "refresh_cache-2c812714-dccc-4d1e-bdb5-c11e446949c3" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1037.114615] env[65726]: DEBUG nova.network.neutron [req-090da2c3-13b5-4d5b-bea5-cf65b985fd32 req-b83d43d2-9504-4125-9aa4-c0aeecbc1e9f service nova] [instance: 2c812714-dccc-4d1e-bdb5-c11e446949c3] Refreshing network info cache for port 8addcb43-85bf-481e-92bc-5bbdf48c0ac4 {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 1037.250712] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116536, 'name': CreateVM_Task, 'duration_secs': 1.45382} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1037.251115] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fb8eac12-aedd-426f-9dd3-6c52f9d4c71d] Created VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1037.253040] env[65726]: WARNING openstack [None req-a3acab1e-0b06-47ad-b154-32b257cd5af7 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1037.253040] env[65726]: WARNING openstack [None req-a3acab1e-0b06-47ad-b154-32b257cd5af7 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1037.259042] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a3acab1e-0b06-47ad-b154-32b257cd5af7 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1037.259248] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a3acab1e-0b06-47ad-b154-32b257cd5af7 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1037.259606] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a3acab1e-0b06-47ad-b154-32b257cd5af7 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1037.260440] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b92cfe7c-1dfc-4bd6-9be7-fcee2d3e2d26 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.273037] env[65726]: DEBUG oslo_vmware.api [None req-f1e7868e-b3d9-4342-8972-f27e4cb4ec45 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] Task: {'id': task-5116540, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.684033} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1037.274799] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-f1e7868e-b3d9-4342-8972-f27e4cb4ec45 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] [instance: 2b8eb60b-dc9b-47a4-9ddc-205f2af69888] Extended root virtual disk {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1037.275399] env[65726]: DEBUG oslo_vmware.api [None req-a3acab1e-0b06-47ad-b154-32b257cd5af7 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Waiting for the task: (returnval){ [ 1037.275399] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]527d7969-90eb-2ed8-557c-13ca53fa16cf" [ 1037.275399] env[65726]: _type = "Task" [ 1037.275399] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1037.276712] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5997b74-7825-41b1-bca4-6e460117a9d2 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.290186] env[65726]: DEBUG oslo_vmware.api [None req-a3acab1e-0b06-47ad-b154-32b257cd5af7 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]527d7969-90eb-2ed8-557c-13ca53fa16cf, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.310032] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-f1e7868e-b3d9-4342-8972-f27e4cb4ec45 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] [instance: 2b8eb60b-dc9b-47a4-9ddc-205f2af69888] Reconfiguring VM instance instance-00000058 to attach disk [datastore1] 2b8eb60b-dc9b-47a4-9ddc-205f2af69888/2b8eb60b-dc9b-47a4-9ddc-205f2af69888.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1037.310445] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f22581ca-d6aa-4d49-939e-a1ad90a44c51 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.336851] env[65726]: WARNING openstack [None req-76e5406a-1ee2-4318-b0b7-91862308e4d0 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1037.337299] env[65726]: WARNING openstack [None req-76e5406a-1ee2-4318-b0b7-91862308e4d0 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1037.345770] env[65726]: DEBUG oslo_vmware.api [None req-f1e7868e-b3d9-4342-8972-f27e4cb4ec45 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] Waiting for the task: (returnval){ [ 1037.345770] env[65726]: value = "task-5116543" [ 1037.345770] env[65726]: _type = "Task" [ 1037.345770] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1037.364205] env[65726]: DEBUG oslo_vmware.api [None req-f1e7868e-b3d9-4342-8972-f27e4cb4ec45 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] Task: {'id': task-5116543, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.371348] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116542, 'name': CreateVM_Task} progress is 25%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.399987] env[65726]: DEBUG nova.scheduler.client.report [None req-10701ce3-ca7d-4146-b0df-5677afb5eba0 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1037.481242] env[65726]: WARNING openstack [None req-76e5406a-1ee2-4318-b0b7-91862308e4d0 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1037.481485] env[65726]: WARNING openstack [None req-76e5406a-1ee2-4318-b0b7-91862308e4d0 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1037.500043] env[65726]: DEBUG nova.network.neutron [-] [instance: a286ba4f-a83e-4c30-8079-ba88e3ba5696] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1037.546472] env[65726]: DEBUG oslo_vmware.api [None req-1d2791c0-c341-4061-b914-188004153713 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Task: {'id': task-5116541, 'name': CopyVirtualDisk_Task} progress is 18%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.618541] env[65726]: WARNING openstack [req-090da2c3-13b5-4d5b-bea5-cf65b985fd32 req-b83d43d2-9504-4125-9aa4-c0aeecbc1e9f service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1037.618541] env[65726]: WARNING openstack [req-090da2c3-13b5-4d5b-bea5-cf65b985fd32 req-b83d43d2-9504-4125-9aa4-c0aeecbc1e9f service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1037.691312] env[65726]: WARNING openstack [None req-76e5406a-1ee2-4318-b0b7-91862308e4d0 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1037.692092] env[65726]: WARNING openstack [None req-76e5406a-1ee2-4318-b0b7-91862308e4d0 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1037.712997] env[65726]: DEBUG oslo_concurrency.lockutils [None req-534e859d-9bab-4a75-8dfe-7304a9dfecfd tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Acquiring lock "f947529e-c930-4420-91a7-1e677b5ac2f1" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1037.713288] env[65726]: DEBUG oslo_concurrency.lockutils [None req-534e859d-9bab-4a75-8dfe-7304a9dfecfd tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Lock "f947529e-c930-4420-91a7-1e677b5ac2f1" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1037.796506] env[65726]: DEBUG oslo_vmware.api [None req-a3acab1e-0b06-47ad-b154-32b257cd5af7 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]527d7969-90eb-2ed8-557c-13ca53fa16cf, 'name': SearchDatastore_Task, 'duration_secs': 0.062025} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1037.797380] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a3acab1e-0b06-47ad-b154-32b257cd5af7 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1037.797682] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-a3acab1e-0b06-47ad-b154-32b257cd5af7 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: fb8eac12-aedd-426f-9dd3-6c52f9d4c71d] Processing image b52362a3-ee8a-4cbf-b06f-513b0cc8f95c {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1037.797961] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a3acab1e-0b06-47ad-b154-32b257cd5af7 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1037.798201] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a3acab1e-0b06-47ad-b154-32b257cd5af7 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1037.798826] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-a3acab1e-0b06-47ad-b154-32b257cd5af7 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1037.798826] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ca0887e3-8c95-4e11-849e-216c0f9ec832 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.816764] env[65726]: DEBUG nova.network.neutron [None req-76e5406a-1ee2-4318-b0b7-91862308e4d0 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] [instance: aa6ce489-c62f-4481-87b7-e74242aeb8ca] Updating instance_info_cache with network_info: [{"id": "d5a805f5-b6a8-42ec-92a0-4a652b51c1e6", "address": "fa:16:3e:f5:d5:d7", "network": {"id": "547d7b16-4591-4d80-9e9b-c8a9a3a78dab", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-2077874536-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.242", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4ffd45f4a7a041199a4fc7f69f5e7e9b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2ed91b7b-b4ec-486d-ab34-af0afb7ec691", "external-id": "nsx-vlan-transportzone-75", "segmentation_id": 75, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd5a805f5-b6", "ovs_interfaceid": "d5a805f5-b6a8-42ec-92a0-4a652b51c1e6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1037.822601] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-a3acab1e-0b06-47ad-b154-32b257cd5af7 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1037.822808] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-a3acab1e-0b06-47ad-b154-32b257cd5af7 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65726) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1037.823724] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8c85a170-7d06-4d5a-ab46-3596257d3332 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.832658] env[65726]: WARNING openstack [req-090da2c3-13b5-4d5b-bea5-cf65b985fd32 req-b83d43d2-9504-4125-9aa4-c0aeecbc1e9f service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1037.833158] env[65726]: WARNING openstack [req-090da2c3-13b5-4d5b-bea5-cf65b985fd32 req-b83d43d2-9504-4125-9aa4-c0aeecbc1e9f service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1037.842944] env[65726]: DEBUG oslo_vmware.api [None req-a3acab1e-0b06-47ad-b154-32b257cd5af7 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Waiting for the task: (returnval){ [ 1037.842944] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]527d3cef-4a60-6ddb-cfa1-09cb43a9fa74" [ 1037.842944] env[65726]: _type = "Task" [ 1037.842944] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1037.863983] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116542, 'name': CreateVM_Task} progress is 25%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.872032] env[65726]: DEBUG oslo_vmware.api [None req-a3acab1e-0b06-47ad-b154-32b257cd5af7 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]527d3cef-4a60-6ddb-cfa1-09cb43a9fa74, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.872431] env[65726]: DEBUG oslo_vmware.api [None req-f1e7868e-b3d9-4342-8972-f27e4cb4ec45 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] Task: {'id': task-5116543, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.909191] env[65726]: DEBUG oslo_concurrency.lockutils [None req-10701ce3-ca7d-4146-b0df-5677afb5eba0 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.360s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1037.910187] env[65726]: DEBUG nova.compute.manager [None req-10701ce3-ca7d-4146-b0df-5677afb5eba0 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] [instance: df08d166-7811-49b5-9c16-a3434229d623] Start building networks asynchronously for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1037.917045] env[65726]: DEBUG oslo_concurrency.lockutils [None req-0c1711f1-74cb-4ba7-a859-716c68812e37 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.427s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1037.917045] env[65726]: INFO nova.compute.claims [None req-0c1711f1-74cb-4ba7-a859-716c68812e37 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: ad4f00b5-3765-494f-8132-8b17b29d219a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1037.978512] env[65726]: WARNING openstack [req-090da2c3-13b5-4d5b-bea5-cf65b985fd32 req-b83d43d2-9504-4125-9aa4-c0aeecbc1e9f service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1037.978987] env[65726]: WARNING openstack [req-090da2c3-13b5-4d5b-bea5-cf65b985fd32 req-b83d43d2-9504-4125-9aa4-c0aeecbc1e9f service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1038.004819] env[65726]: INFO nova.compute.manager [-] [instance: a286ba4f-a83e-4c30-8079-ba88e3ba5696] Took 1.28 seconds to deallocate network for instance. [ 1038.045191] env[65726]: DEBUG oslo_vmware.api [None req-1d2791c0-c341-4061-b914-188004153713 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Task: {'id': task-5116541, 'name': CopyVirtualDisk_Task} progress is 38%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.080625] env[65726]: DEBUG nova.network.neutron [req-090da2c3-13b5-4d5b-bea5-cf65b985fd32 req-b83d43d2-9504-4125-9aa4-c0aeecbc1e9f service nova] [instance: 2c812714-dccc-4d1e-bdb5-c11e446949c3] Updated VIF entry in instance network info cache for port 8addcb43-85bf-481e-92bc-5bbdf48c0ac4. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 1038.081751] env[65726]: DEBUG nova.network.neutron [req-090da2c3-13b5-4d5b-bea5-cf65b985fd32 req-b83d43d2-9504-4125-9aa4-c0aeecbc1e9f service nova] [instance: 2c812714-dccc-4d1e-bdb5-c11e446949c3] Updating instance_info_cache with network_info: [{"id": "203fa726-d08f-47a8-a7cd-389ed31da940", "address": "fa:16:3e:0b:bd:69", "network": {"id": "f9d433a7-98f6-4013-ba92-7bbacbef4969", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1784765821", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.175", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "6961c16aa6a346679bbde25930b2a5b7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7ab8d568-adb0-4f3b-b6cc-68413e6546ae", "external-id": "nsx-vlan-transportzone-86", "segmentation_id": 86, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap203fa726-d0", "ovs_interfaceid": "203fa726-d08f-47a8-a7cd-389ed31da940", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "8addcb43-85bf-481e-92bc-5bbdf48c0ac4", "address": "fa:16:3e:4f:bc:c3", "network": {"id": "deb5277b-48b1-472a-b940-d808da6ffa26", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1186999789", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.239", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "6961c16aa6a346679bbde25930b2a5b7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7adad264-9276-43ef-9b03-07dc27d3f81e", "external-id": "nsx-vlan-transportzone-329", "segmentation_id": 329, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8addcb43-85", "ovs_interfaceid": "8addcb43-85bf-481e-92bc-5bbdf48c0ac4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1038.218152] env[65726]: DEBUG nova.compute.manager [None req-534e859d-9bab-4a75-8dfe-7304a9dfecfd tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: f947529e-c930-4420-91a7-1e677b5ac2f1] Starting instance... {{(pid=65726) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 1038.233640] env[65726]: DEBUG nova.compute.manager [req-e2739c97-a87c-4abb-b676-260112d8cd99 req-34d4c05d-ccdf-4f9e-a704-40f70956db72 service nova] [instance: a286ba4f-a83e-4c30-8079-ba88e3ba5696] Received event network-vif-deleted-bae2338e-4be5-4ac9-b3b7-0f31ce3e0b37 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1038.320227] env[65726]: DEBUG oslo_concurrency.lockutils [None req-76e5406a-1ee2-4318-b0b7-91862308e4d0 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Releasing lock "refresh_cache-aa6ce489-c62f-4481-87b7-e74242aeb8ca" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1038.362627] env[65726]: DEBUG oslo_vmware.api [None req-a3acab1e-0b06-47ad-b154-32b257cd5af7 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]527d3cef-4a60-6ddb-cfa1-09cb43a9fa74, 'name': SearchDatastore_Task, 'duration_secs': 0.083601} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1038.363979] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f3fbd5e9-058d-497f-be17-f0abe574e91e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.373647] env[65726]: DEBUG oslo_vmware.api [None req-f1e7868e-b3d9-4342-8972-f27e4cb4ec45 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] Task: {'id': task-5116543, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.373951] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116542, 'name': CreateVM_Task} progress is 99%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.378533] env[65726]: DEBUG oslo_vmware.api [None req-a3acab1e-0b06-47ad-b154-32b257cd5af7 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Waiting for the task: (returnval){ [ 1038.378533] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]526f6325-1d8f-d4f5-fef9-d1d1f1866e0d" [ 1038.378533] env[65726]: _type = "Task" [ 1038.378533] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1038.387834] env[65726]: DEBUG oslo_vmware.api [None req-a3acab1e-0b06-47ad-b154-32b257cd5af7 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]526f6325-1d8f-d4f5-fef9-d1d1f1866e0d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.405364] env[65726]: DEBUG oslo_concurrency.lockutils [None req-031ff252-21d5-4d9e-b884-2b2939b62196 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Acquiring lock "a56be6fa-6027-46ee-9aa8-332fe4222d68" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1038.405779] env[65726]: DEBUG oslo_concurrency.lockutils [None req-031ff252-21d5-4d9e-b884-2b2939b62196 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Lock "a56be6fa-6027-46ee-9aa8-332fe4222d68" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1038.423057] env[65726]: DEBUG nova.compute.utils [None req-10701ce3-ca7d-4146-b0df-5677afb5eba0 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] Using /dev/sd instead of None {{(pid=65726) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1038.426965] env[65726]: DEBUG nova.compute.manager [None req-10701ce3-ca7d-4146-b0df-5677afb5eba0 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] [instance: df08d166-7811-49b5-9c16-a3434229d623] Allocating IP information in the background. {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1038.427193] env[65726]: DEBUG nova.network.neutron [None req-10701ce3-ca7d-4146-b0df-5677afb5eba0 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] [instance: df08d166-7811-49b5-9c16-a3434229d623] allocate_for_instance() {{(pid=65726) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 1038.427619] env[65726]: WARNING neutronclient.v2_0.client [None req-10701ce3-ca7d-4146-b0df-5677afb5eba0 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1038.427725] env[65726]: WARNING neutronclient.v2_0.client [None req-10701ce3-ca7d-4146-b0df-5677afb5eba0 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1038.428442] env[65726]: WARNING openstack [None req-10701ce3-ca7d-4146-b0df-5677afb5eba0 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1038.428783] env[65726]: WARNING openstack [None req-10701ce3-ca7d-4146-b0df-5677afb5eba0 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1038.489964] env[65726]: DEBUG nova.policy [None req-10701ce3-ca7d-4146-b0df-5677afb5eba0 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9a104357e484455a805c751b553372dd', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '305ecace479440ad9a616e546ccd38ce', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65726) authorize /opt/stack/nova/nova/policy.py:192}} [ 1038.516758] env[65726]: DEBUG oslo_concurrency.lockutils [None req-43be9038-9e49-4e4d-82ea-842afcd9a64d tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1038.543792] env[65726]: DEBUG oslo_vmware.api [None req-1d2791c0-c341-4061-b914-188004153713 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Task: {'id': task-5116541, 'name': CopyVirtualDisk_Task} progress is 60%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.585710] env[65726]: DEBUG oslo_concurrency.lockutils [req-090da2c3-13b5-4d5b-bea5-cf65b985fd32 req-b83d43d2-9504-4125-9aa4-c0aeecbc1e9f service nova] Releasing lock "refresh_cache-2c812714-dccc-4d1e-bdb5-c11e446949c3" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1038.638849] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5c22cc7c-7dcc-417f-8f16-1f19636dd1f0 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Acquiring lock "51f5337d-2e23-4b7c-b590-76be6bd82411" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1038.639119] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5c22cc7c-7dcc-417f-8f16-1f19636dd1f0 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Lock "51f5337d-2e23-4b7c-b590-76be6bd82411" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1038.752577] env[65726]: DEBUG oslo_concurrency.lockutils [None req-534e859d-9bab-4a75-8dfe-7304a9dfecfd tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1038.847555] env[65726]: DEBUG nova.network.neutron [None req-10701ce3-ca7d-4146-b0df-5677afb5eba0 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] [instance: df08d166-7811-49b5-9c16-a3434229d623] Successfully created port: 70efdc7d-7775-4990-b851-d60195bd1504 {{(pid=65726) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1038.869539] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116542, 'name': CreateVM_Task, 'duration_secs': 1.660942} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1038.873614] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2c812714-dccc-4d1e-bdb5-c11e446949c3] Created VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1038.874066] env[65726]: DEBUG oslo_vmware.api [None req-f1e7868e-b3d9-4342-8972-f27e4cb4ec45 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] Task: {'id': task-5116543, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.874950] env[65726]: WARNING openstack [None req-2a9dd832-d188-4542-a7fe-bbdea8173b2d tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1038.875503] env[65726]: WARNING openstack [None req-2a9dd832-d188-4542-a7fe-bbdea8173b2d tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1038.882872] env[65726]: WARNING openstack [None req-2a9dd832-d188-4542-a7fe-bbdea8173b2d tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1038.882872] env[65726]: WARNING openstack [None req-2a9dd832-d188-4542-a7fe-bbdea8173b2d tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1038.889673] env[65726]: DEBUG oslo_concurrency.lockutils [None req-2a9dd832-d188-4542-a7fe-bbdea8173b2d tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1038.889847] env[65726]: DEBUG oslo_concurrency.lockutils [None req-2a9dd832-d188-4542-a7fe-bbdea8173b2d tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1038.890182] env[65726]: DEBUG oslo_concurrency.lockutils [None req-2a9dd832-d188-4542-a7fe-bbdea8173b2d tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1038.890583] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-535dc816-4a9f-4a67-b58f-9a6f4a45f258 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.907319] env[65726]: DEBUG oslo_vmware.api [None req-a3acab1e-0b06-47ad-b154-32b257cd5af7 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]526f6325-1d8f-d4f5-fef9-d1d1f1866e0d, 'name': SearchDatastore_Task, 'duration_secs': 0.08148} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1038.907319] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a3acab1e-0b06-47ad-b154-32b257cd5af7 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1038.908800] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-a3acab1e-0b06-47ad-b154-32b257cd5af7 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] fb8eac12-aedd-426f-9dd3-6c52f9d4c71d/fb8eac12-aedd-426f-9dd3-6c52f9d4c71d.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1038.908800] env[65726]: DEBUG oslo_vmware.api [None req-2a9dd832-d188-4542-a7fe-bbdea8173b2d tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Waiting for the task: (returnval){ [ 1038.908800] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52b2157e-9ec0-0fcb-937c-f43241a685c1" [ 1038.908800] env[65726]: _type = "Task" [ 1038.908800] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1038.908800] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e2427d9c-321a-4e13-a382-c4b9534e7445 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.912203] env[65726]: DEBUG nova.compute.utils [None req-031ff252-21d5-4d9e-b884-2b2939b62196 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Using /dev/sd instead of None {{(pid=65726) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1038.926141] env[65726]: DEBUG oslo_vmware.api [None req-2a9dd832-d188-4542-a7fe-bbdea8173b2d tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52b2157e-9ec0-0fcb-937c-f43241a685c1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.929243] env[65726]: DEBUG nova.compute.manager [None req-10701ce3-ca7d-4146-b0df-5677afb5eba0 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] [instance: df08d166-7811-49b5-9c16-a3434229d623] Start building block device mappings for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1038.933046] env[65726]: DEBUG oslo_vmware.api [None req-a3acab1e-0b06-47ad-b154-32b257cd5af7 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Waiting for the task: (returnval){ [ 1038.933046] env[65726]: value = "task-5116545" [ 1038.933046] env[65726]: _type = "Task" [ 1038.933046] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1038.953795] env[65726]: DEBUG oslo_vmware.api [None req-a3acab1e-0b06-47ad-b154-32b257cd5af7 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': task-5116545, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.048550] env[65726]: DEBUG oslo_vmware.api [None req-1d2791c0-c341-4061-b914-188004153713 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Task: {'id': task-5116541, 'name': CopyVirtualDisk_Task} progress is 80%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.141550] env[65726]: DEBUG nova.compute.manager [None req-5c22cc7c-7dcc-417f-8f16-1f19636dd1f0 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: 51f5337d-2e23-4b7c-b590-76be6bd82411] Starting instance... {{(pid=65726) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 1039.178718] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8837e8c6-8ebb-4de2-9647-ccc06178084d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.194403] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-357ee74b-86cc-4824-8439-5bec34b0883b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.231869] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12d37e5d-7385-4764-b519-dd4a4507d089 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.245457] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-440a5874-11f3-45d6-828b-3a0cea48c4c1 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.269076] env[65726]: DEBUG nova.compute.provider_tree [None req-0c1711f1-74cb-4ba7-a859-716c68812e37 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1039.329119] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-76e5406a-1ee2-4318-b0b7-91862308e4d0 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] [instance: aa6ce489-c62f-4481-87b7-e74242aeb8ca] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1039.329298] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-fbdec921-fc6c-43d3-9c09-d444497f720b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.341386] env[65726]: DEBUG oslo_vmware.api [None req-76e5406a-1ee2-4318-b0b7-91862308e4d0 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Waiting for the task: (returnval){ [ 1039.341386] env[65726]: value = "task-5116546" [ 1039.341386] env[65726]: _type = "Task" [ 1039.341386] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1039.357829] env[65726]: DEBUG oslo_vmware.api [None req-76e5406a-1ee2-4318-b0b7-91862308e4d0 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Task: {'id': task-5116546, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.372325] env[65726]: DEBUG oslo_vmware.api [None req-f1e7868e-b3d9-4342-8972-f27e4cb4ec45 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] Task: {'id': task-5116543, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.418756] env[65726]: DEBUG oslo_concurrency.lockutils [None req-031ff252-21d5-4d9e-b884-2b2939b62196 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Lock "a56be6fa-6027-46ee-9aa8-332fe4222d68" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.013s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1039.428204] env[65726]: DEBUG oslo_vmware.api [None req-2a9dd832-d188-4542-a7fe-bbdea8173b2d tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52b2157e-9ec0-0fcb-937c-f43241a685c1, 'name': SearchDatastore_Task, 'duration_secs': 0.08824} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1039.428602] env[65726]: DEBUG oslo_concurrency.lockutils [None req-2a9dd832-d188-4542-a7fe-bbdea8173b2d tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1039.428840] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-2a9dd832-d188-4542-a7fe-bbdea8173b2d tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] [instance: 2c812714-dccc-4d1e-bdb5-c11e446949c3] Processing image b52362a3-ee8a-4cbf-b06f-513b0cc8f95c {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1039.429097] env[65726]: DEBUG oslo_concurrency.lockutils [None req-2a9dd832-d188-4542-a7fe-bbdea8173b2d tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1039.429249] env[65726]: DEBUG oslo_concurrency.lockutils [None req-2a9dd832-d188-4542-a7fe-bbdea8173b2d tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1039.429434] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-2a9dd832-d188-4542-a7fe-bbdea8173b2d tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1039.429713] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-007680d2-0831-4cc1-829d-208fbefe4b9b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.448503] env[65726]: DEBUG oslo_vmware.api [None req-a3acab1e-0b06-47ad-b154-32b257cd5af7 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': task-5116545, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.451052] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-2a9dd832-d188-4542-a7fe-bbdea8173b2d tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1039.451186] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-2a9dd832-d188-4542-a7fe-bbdea8173b2d tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65726) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1039.452025] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-957f384a-5410-4372-b174-db4dcc7c0a74 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.463950] env[65726]: DEBUG oslo_vmware.api [None req-2a9dd832-d188-4542-a7fe-bbdea8173b2d tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Waiting for the task: (returnval){ [ 1039.463950] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]523cebfa-da5e-1a97-4b5b-0b7f54e442eb" [ 1039.463950] env[65726]: _type = "Task" [ 1039.463950] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1039.478778] env[65726]: DEBUG oslo_vmware.api [None req-2a9dd832-d188-4542-a7fe-bbdea8173b2d tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]523cebfa-da5e-1a97-4b5b-0b7f54e442eb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.546736] env[65726]: DEBUG oslo_vmware.api [None req-1d2791c0-c341-4061-b914-188004153713 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Task: {'id': task-5116541, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.977869} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1039.547024] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-1d2791c0-c341-4061-b914-188004153713 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/81f5ce39-0bc0-4d0e-ab15-42a80a4f614b/81f5ce39-0bc0-4d0e-ab15-42a80a4f614b.vmdk to [datastore1] 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04/2fe6fc5a-f5c7-4f8e-96df-4e621a252f04.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1039.547885] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1cae038e-31e5-4ebe-9ef3-9a10a338b1c1 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.574029] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-1d2791c0-c341-4061-b914-188004153713 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04] Reconfiguring VM instance instance-0000003f to attach disk [datastore1] 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04/2fe6fc5a-f5c7-4f8e-96df-4e621a252f04.vmdk or device None with type streamOptimized {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1039.574029] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a072759e-f5b3-49d8-a76a-4844d68bf88d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.595952] env[65726]: DEBUG oslo_vmware.api [None req-1d2791c0-c341-4061-b914-188004153713 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Waiting for the task: (returnval){ [ 1039.595952] env[65726]: value = "task-5116547" [ 1039.595952] env[65726]: _type = "Task" [ 1039.595952] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1039.606600] env[65726]: DEBUG oslo_vmware.api [None req-1d2791c0-c341-4061-b914-188004153713 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Task: {'id': task-5116547, 'name': ReconfigVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.667658] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5c22cc7c-7dcc-417f-8f16-1f19636dd1f0 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1039.773519] env[65726]: DEBUG nova.scheduler.client.report [None req-0c1711f1-74cb-4ba7-a859-716c68812e37 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1039.856155] env[65726]: DEBUG oslo_vmware.api [None req-76e5406a-1ee2-4318-b0b7-91862308e4d0 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Task: {'id': task-5116546, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.872747] env[65726]: DEBUG oslo_vmware.api [None req-f1e7868e-b3d9-4342-8972-f27e4cb4ec45 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] Task: {'id': task-5116543, 'name': ReconfigVM_Task, 'duration_secs': 2.287861} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1039.873210] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-f1e7868e-b3d9-4342-8972-f27e4cb4ec45 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] [instance: 2b8eb60b-dc9b-47a4-9ddc-205f2af69888] Reconfigured VM instance instance-00000058 to attach disk [datastore1] 2b8eb60b-dc9b-47a4-9ddc-205f2af69888/2b8eb60b-dc9b-47a4-9ddc-205f2af69888.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1039.874102] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ef937bc1-6f49-43c9-986a-67818126c172 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.885941] env[65726]: DEBUG oslo_vmware.api [None req-f1e7868e-b3d9-4342-8972-f27e4cb4ec45 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] Waiting for the task: (returnval){ [ 1039.885941] env[65726]: value = "task-5116548" [ 1039.885941] env[65726]: _type = "Task" [ 1039.885941] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1039.901714] env[65726]: DEBUG oslo_vmware.api [None req-f1e7868e-b3d9-4342-8972-f27e4cb4ec45 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] Task: {'id': task-5116548, 'name': Rename_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.942572] env[65726]: DEBUG nova.compute.manager [None req-10701ce3-ca7d-4146-b0df-5677afb5eba0 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] [instance: df08d166-7811-49b5-9c16-a3434229d623] Start spawning the instance on the hypervisor. {{(pid=65726) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1039.962274] env[65726]: DEBUG oslo_vmware.api [None req-a3acab1e-0b06-47ad-b154-32b257cd5af7 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': task-5116545, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.984309] env[65726]: DEBUG oslo_vmware.api [None req-2a9dd832-d188-4542-a7fe-bbdea8173b2d tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]523cebfa-da5e-1a97-4b5b-0b7f54e442eb, 'name': SearchDatastore_Task, 'duration_secs': 0.030147} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1039.985242] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-28c3e5ca-80e7-44bf-b671-bcb40872c287 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.990821] env[65726]: DEBUG nova.virt.hardware [None req-10701ce3-ca7d-4146-b0df-5677afb5eba0 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T19:28:34Z,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1539f06fa3534e90acbc3a60c4b8bd3f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T19:28:35Z,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1039.991179] env[65726]: DEBUG nova.virt.hardware [None req-10701ce3-ca7d-4146-b0df-5677afb5eba0 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1039.991385] env[65726]: DEBUG nova.virt.hardware [None req-10701ce3-ca7d-4146-b0df-5677afb5eba0 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1039.991607] env[65726]: DEBUG nova.virt.hardware [None req-10701ce3-ca7d-4146-b0df-5677afb5eba0 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1039.991763] env[65726]: DEBUG nova.virt.hardware [None req-10701ce3-ca7d-4146-b0df-5677afb5eba0 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1039.991940] env[65726]: DEBUG nova.virt.hardware [None req-10701ce3-ca7d-4146-b0df-5677afb5eba0 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1039.992272] env[65726]: DEBUG nova.virt.hardware [None req-10701ce3-ca7d-4146-b0df-5677afb5eba0 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1039.992530] env[65726]: DEBUG nova.virt.hardware [None req-10701ce3-ca7d-4146-b0df-5677afb5eba0 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1039.992805] env[65726]: DEBUG nova.virt.hardware [None req-10701ce3-ca7d-4146-b0df-5677afb5eba0 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1039.993104] env[65726]: DEBUG nova.virt.hardware [None req-10701ce3-ca7d-4146-b0df-5677afb5eba0 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1039.995208] env[65726]: DEBUG nova.virt.hardware [None req-10701ce3-ca7d-4146-b0df-5677afb5eba0 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1039.998933] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aeb9086a-270f-4b29-b141-48f574a8204b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.011679] env[65726]: DEBUG oslo_vmware.api [None req-2a9dd832-d188-4542-a7fe-bbdea8173b2d tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Waiting for the task: (returnval){ [ 1040.011679] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5244baae-3016-3e2c-d62d-69a1c0a5acd4" [ 1040.011679] env[65726]: _type = "Task" [ 1040.011679] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1040.018808] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aaf72f5c-cfcb-410c-b28b-61bf88130be7 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.032247] env[65726]: DEBUG oslo_vmware.api [None req-2a9dd832-d188-4542-a7fe-bbdea8173b2d tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5244baae-3016-3e2c-d62d-69a1c0a5acd4, 'name': SearchDatastore_Task, 'duration_secs': 0.011022} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1040.040618] env[65726]: DEBUG oslo_concurrency.lockutils [None req-2a9dd832-d188-4542-a7fe-bbdea8173b2d tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1040.040902] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-2a9dd832-d188-4542-a7fe-bbdea8173b2d tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] 2c812714-dccc-4d1e-bdb5-c11e446949c3/2c812714-dccc-4d1e-bdb5-c11e446949c3.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1040.041980] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-59f7f12b-ca35-4ad7-80a8-8bf2364e4d05 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.049813] env[65726]: DEBUG oslo_vmware.api [None req-2a9dd832-d188-4542-a7fe-bbdea8173b2d tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Waiting for the task: (returnval){ [ 1040.049813] env[65726]: value = "task-5116549" [ 1040.049813] env[65726]: _type = "Task" [ 1040.049813] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1040.059258] env[65726]: DEBUG oslo_vmware.api [None req-2a9dd832-d188-4542-a7fe-bbdea8173b2d tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Task: {'id': task-5116549, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.142815] env[65726]: DEBUG oslo_vmware.api [None req-1d2791c0-c341-4061-b914-188004153713 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Task: {'id': task-5116547, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.281562] env[65726]: DEBUG oslo_concurrency.lockutils [None req-0c1711f1-74cb-4ba7-a859-716c68812e37 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.365s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1040.281562] env[65726]: DEBUG nova.compute.manager [None req-0c1711f1-74cb-4ba7-a859-716c68812e37 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: ad4f00b5-3765-494f-8132-8b17b29d219a] Start building networks asynchronously for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1040.284035] env[65726]: DEBUG oslo_concurrency.lockutils [None req-43be9038-9e49-4e4d-82ea-842afcd9a64d tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.767s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1040.284220] env[65726]: DEBUG nova.objects.instance [None req-43be9038-9e49-4e4d-82ea-842afcd9a64d tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Lazy-loading 'resources' on Instance uuid a286ba4f-a83e-4c30-8079-ba88e3ba5696 {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1040.356237] env[65726]: DEBUG oslo_vmware.api [None req-76e5406a-1ee2-4318-b0b7-91862308e4d0 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Task: {'id': task-5116546, 'name': PowerOnVM_Task, 'duration_secs': 0.596911} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1040.356559] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-76e5406a-1ee2-4318-b0b7-91862308e4d0 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] [instance: aa6ce489-c62f-4481-87b7-e74242aeb8ca] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1040.356879] env[65726]: DEBUG nova.compute.manager [None req-76e5406a-1ee2-4318-b0b7-91862308e4d0 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] [instance: aa6ce489-c62f-4481-87b7-e74242aeb8ca] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1040.357854] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6369def3-cd3c-463a-bd8c-e5139f753e93 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.402448] env[65726]: DEBUG oslo_vmware.api [None req-f1e7868e-b3d9-4342-8972-f27e4cb4ec45 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] Task: {'id': task-5116548, 'name': Rename_Task, 'duration_secs': 0.194034} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1040.402811] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-f1e7868e-b3d9-4342-8972-f27e4cb4ec45 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] [instance: 2b8eb60b-dc9b-47a4-9ddc-205f2af69888] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1040.403135] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b5388a3d-f408-40c1-a98a-401c25b77e83 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.414352] env[65726]: DEBUG oslo_vmware.api [None req-f1e7868e-b3d9-4342-8972-f27e4cb4ec45 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] Waiting for the task: (returnval){ [ 1040.414352] env[65726]: value = "task-5116550" [ 1040.414352] env[65726]: _type = "Task" [ 1040.414352] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1040.426741] env[65726]: DEBUG oslo_vmware.api [None req-f1e7868e-b3d9-4342-8972-f27e4cb4ec45 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] Task: {'id': task-5116550, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.456577] env[65726]: DEBUG oslo_vmware.api [None req-a3acab1e-0b06-47ad-b154-32b257cd5af7 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': task-5116545, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.036981} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1040.457044] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-a3acab1e-0b06-47ad-b154-32b257cd5af7 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] fb8eac12-aedd-426f-9dd3-6c52f9d4c71d/fb8eac12-aedd-426f-9dd3-6c52f9d4c71d.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1040.457314] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-a3acab1e-0b06-47ad-b154-32b257cd5af7 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: fb8eac12-aedd-426f-9dd3-6c52f9d4c71d] Extending root virtual disk to 1048576 {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1040.457635] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3e54db7b-34d4-4a81-a1d2-5edbef111bc5 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.468152] env[65726]: DEBUG oslo_vmware.api [None req-a3acab1e-0b06-47ad-b154-32b257cd5af7 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Waiting for the task: (returnval){ [ 1040.468152] env[65726]: value = "task-5116551" [ 1040.468152] env[65726]: _type = "Task" [ 1040.468152] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1040.482208] env[65726]: DEBUG oslo_vmware.api [None req-a3acab1e-0b06-47ad-b154-32b257cd5af7 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': task-5116551, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.508990] env[65726]: DEBUG oslo_concurrency.lockutils [None req-031ff252-21d5-4d9e-b884-2b2939b62196 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Acquiring lock "a56be6fa-6027-46ee-9aa8-332fe4222d68" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1040.509410] env[65726]: DEBUG oslo_concurrency.lockutils [None req-031ff252-21d5-4d9e-b884-2b2939b62196 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Lock "a56be6fa-6027-46ee-9aa8-332fe4222d68" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1040.509715] env[65726]: INFO nova.compute.manager [None req-031ff252-21d5-4d9e-b884-2b2939b62196 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: a56be6fa-6027-46ee-9aa8-332fe4222d68] Attaching volume 30181d53-b2b9-4253-a8f5-28a3abf812f9 to /dev/sdb [ 1040.545640] env[65726]: DEBUG nova.network.neutron [None req-10701ce3-ca7d-4146-b0df-5677afb5eba0 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] [instance: df08d166-7811-49b5-9c16-a3434229d623] Successfully updated port: 70efdc7d-7775-4990-b851-d60195bd1504 {{(pid=65726) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 1040.557801] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ba7affa-02f1-4789-b213-41fef16243f9 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.568799] env[65726]: DEBUG oslo_vmware.api [None req-2a9dd832-d188-4542-a7fe-bbdea8173b2d tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Task: {'id': task-5116549, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.569845] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ddd53ea0-2e75-4d7e-9dea-acd0603ba980 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.589504] env[65726]: DEBUG nova.virt.block_device [None req-031ff252-21d5-4d9e-b884-2b2939b62196 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: a56be6fa-6027-46ee-9aa8-332fe4222d68] Updating existing volume attachment record: e06c404d-8d21-48d5-b376-ec3a89b1e52c {{(pid=65726) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1040.592406] env[65726]: DEBUG oslo_concurrency.lockutils [None req-17e357f7-df51-4538-bd2b-ea061ce92818 tempest-ServerMetadataTestJSON-637854735 tempest-ServerMetadataTestJSON-637854735-project-member] Acquiring lock "699127a9-f40f-4c1c-ba4a-625af097350b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1040.592406] env[65726]: DEBUG oslo_concurrency.lockutils [None req-17e357f7-df51-4538-bd2b-ea061ce92818 tempest-ServerMetadataTestJSON-637854735 tempest-ServerMetadataTestJSON-637854735-project-member] Lock "699127a9-f40f-4c1c-ba4a-625af097350b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1040.592638] env[65726]: DEBUG oslo_concurrency.lockutils [None req-17e357f7-df51-4538-bd2b-ea061ce92818 tempest-ServerMetadataTestJSON-637854735 tempest-ServerMetadataTestJSON-637854735-project-member] Acquiring lock "699127a9-f40f-4c1c-ba4a-625af097350b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1040.592824] env[65726]: DEBUG oslo_concurrency.lockutils [None req-17e357f7-df51-4538-bd2b-ea061ce92818 tempest-ServerMetadataTestJSON-637854735 tempest-ServerMetadataTestJSON-637854735-project-member] Lock "699127a9-f40f-4c1c-ba4a-625af097350b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1040.593000] env[65726]: DEBUG oslo_concurrency.lockutils [None req-17e357f7-df51-4538-bd2b-ea061ce92818 tempest-ServerMetadataTestJSON-637854735 tempest-ServerMetadataTestJSON-637854735-project-member] Lock "699127a9-f40f-4c1c-ba4a-625af097350b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1040.596315] env[65726]: INFO nova.compute.manager [None req-17e357f7-df51-4538-bd2b-ea061ce92818 tempest-ServerMetadataTestJSON-637854735 tempest-ServerMetadataTestJSON-637854735-project-member] [instance: 699127a9-f40f-4c1c-ba4a-625af097350b] Terminating instance [ 1040.608231] env[65726]: DEBUG oslo_vmware.api [None req-1d2791c0-c341-4061-b914-188004153713 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Task: {'id': task-5116547, 'name': ReconfigVM_Task, 'duration_secs': 0.597165} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1040.608844] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-1d2791c0-c341-4061-b914-188004153713 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04] Reconfigured VM instance instance-0000003f to attach disk [datastore1] 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04/2fe6fc5a-f5c7-4f8e-96df-4e621a252f04.vmdk or device None with type streamOptimized {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1040.610204] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-1d2791c0-c341-4061-b914-188004153713 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04] Block device information present: {'root_device_name': '/dev/sda', 'image': [{'disk_bus': None, 'encrypted': False, 'encryption_format': None, 'device_type': 'disk', 'encryption_secret_uuid': None, 'encryption_options': None, 'boot_index': 0, 'guest_format': None, 'size': 0, 'device_name': '/dev/sda', 'image_id': 'b52362a3-ee8a-4cbf-b06f-513b0cc8f95c'}], 'ephemerals': [], 'block_device_mapping': [{'disk_bus': None, 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-995255', 'volume_id': '98eb7918-3cfa-4296-9855-bce47b982be1', 'name': 'volume-98eb7918-3cfa-4296-9855-bce47b982be1', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'attached', 'instance': '2fe6fc5a-f5c7-4f8e-96df-4e621a252f04', 'attached_at': '', 'detached_at': '', 'volume_id': '98eb7918-3cfa-4296-9855-bce47b982be1', 'serial': '98eb7918-3cfa-4296-9855-bce47b982be1'}, 'delete_on_termination': False, 'device_type': None, 'boot_index': None, 'mount_device': '/dev/sdb', 'attachment_id': 'f06d8ae1-535c-4f84-9791-3988884e7159', 'guest_format': None, 'volume_type': None}], 'swap': None} {{(pid=65726) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 1040.610401] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-1d2791c0-c341-4061-b914-188004153713 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04] Volume attach. Driver type: vmdk {{(pid=65726) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1040.610587] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-1d2791c0-c341-4061-b914-188004153713 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-995255', 'volume_id': '98eb7918-3cfa-4296-9855-bce47b982be1', 'name': 'volume-98eb7918-3cfa-4296-9855-bce47b982be1', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'attached', 'instance': '2fe6fc5a-f5c7-4f8e-96df-4e621a252f04', 'attached_at': '', 'detached_at': '', 'volume_id': '98eb7918-3cfa-4296-9855-bce47b982be1', 'serial': '98eb7918-3cfa-4296-9855-bce47b982be1'} {{(pid=65726) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1040.612403] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90195a59-d4ee-43e7-935f-72db62fb5d08 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.629675] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e34d3e9-f7ac-41dc-964b-73c50f17d40e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.657230] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-1d2791c0-c341-4061-b914-188004153713 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04] Reconfiguring VM instance instance-0000003f to attach disk [datastore2] volume-98eb7918-3cfa-4296-9855-bce47b982be1/volume-98eb7918-3cfa-4296-9855-bce47b982be1.vmdk or device None with type thin {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1040.657573] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bb72c9fe-e350-4a26-80ce-7e5722b95304 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.677646] env[65726]: DEBUG oslo_vmware.api [None req-1d2791c0-c341-4061-b914-188004153713 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Waiting for the task: (returnval){ [ 1040.677646] env[65726]: value = "task-5116552" [ 1040.677646] env[65726]: _type = "Task" [ 1040.677646] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1040.690633] env[65726]: DEBUG oslo_vmware.api [None req-1d2791c0-c341-4061-b914-188004153713 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Task: {'id': task-5116552, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.728112] env[65726]: DEBUG nova.compute.manager [req-4af50c77-51df-4147-a860-eb0b1164115c req-081544f1-a635-4a76-80ce-ac66f3cfdeba service nova] [instance: df08d166-7811-49b5-9c16-a3434229d623] Received event network-vif-plugged-70efdc7d-7775-4990-b851-d60195bd1504 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1040.728112] env[65726]: DEBUG oslo_concurrency.lockutils [req-4af50c77-51df-4147-a860-eb0b1164115c req-081544f1-a635-4a76-80ce-ac66f3cfdeba service nova] Acquiring lock "df08d166-7811-49b5-9c16-a3434229d623-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1040.728112] env[65726]: DEBUG oslo_concurrency.lockutils [req-4af50c77-51df-4147-a860-eb0b1164115c req-081544f1-a635-4a76-80ce-ac66f3cfdeba service nova] Lock "df08d166-7811-49b5-9c16-a3434229d623-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1040.728203] env[65726]: DEBUG oslo_concurrency.lockutils [req-4af50c77-51df-4147-a860-eb0b1164115c req-081544f1-a635-4a76-80ce-ac66f3cfdeba service nova] Lock "df08d166-7811-49b5-9c16-a3434229d623-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1040.728364] env[65726]: DEBUG nova.compute.manager [req-4af50c77-51df-4147-a860-eb0b1164115c req-081544f1-a635-4a76-80ce-ac66f3cfdeba service nova] [instance: df08d166-7811-49b5-9c16-a3434229d623] No waiting events found dispatching network-vif-plugged-70efdc7d-7775-4990-b851-d60195bd1504 {{(pid=65726) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1040.728514] env[65726]: WARNING nova.compute.manager [req-4af50c77-51df-4147-a860-eb0b1164115c req-081544f1-a635-4a76-80ce-ac66f3cfdeba service nova] [instance: df08d166-7811-49b5-9c16-a3434229d623] Received unexpected event network-vif-plugged-70efdc7d-7775-4990-b851-d60195bd1504 for instance with vm_state building and task_state spawning. [ 1040.787768] env[65726]: DEBUG nova.compute.utils [None req-0c1711f1-74cb-4ba7-a859-716c68812e37 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Using /dev/sd instead of None {{(pid=65726) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1040.789990] env[65726]: DEBUG nova.compute.manager [None req-0c1711f1-74cb-4ba7-a859-716c68812e37 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: ad4f00b5-3765-494f-8132-8b17b29d219a] Allocating IP information in the background. {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1040.790185] env[65726]: DEBUG nova.network.neutron [None req-0c1711f1-74cb-4ba7-a859-716c68812e37 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: ad4f00b5-3765-494f-8132-8b17b29d219a] allocate_for_instance() {{(pid=65726) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 1040.790509] env[65726]: WARNING neutronclient.v2_0.client [None req-0c1711f1-74cb-4ba7-a859-716c68812e37 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1040.790791] env[65726]: WARNING neutronclient.v2_0.client [None req-0c1711f1-74cb-4ba7-a859-716c68812e37 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1040.791370] env[65726]: WARNING openstack [None req-0c1711f1-74cb-4ba7-a859-716c68812e37 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1040.791739] env[65726]: WARNING openstack [None req-0c1711f1-74cb-4ba7-a859-716c68812e37 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1040.853335] env[65726]: DEBUG nova.policy [None req-0c1711f1-74cb-4ba7-a859-716c68812e37 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f3050294b8f34a0f97d10b038b048779', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a090d553766847e58e3231b966c92565', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65726) authorize /opt/stack/nova/nova/policy.py:192}} [ 1040.930549] env[65726]: DEBUG oslo_vmware.api [None req-f1e7868e-b3d9-4342-8972-f27e4cb4ec45 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] Task: {'id': task-5116550, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.981482] env[65726]: DEBUG oslo_vmware.api [None req-a3acab1e-0b06-47ad-b154-32b257cd5af7 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': task-5116551, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.089978} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1040.981859] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-a3acab1e-0b06-47ad-b154-32b257cd5af7 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: fb8eac12-aedd-426f-9dd3-6c52f9d4c71d] Extended root virtual disk {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1040.982816] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-952f0c72-c273-4933-aefd-4e689ee523bc {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.010353] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-a3acab1e-0b06-47ad-b154-32b257cd5af7 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: fb8eac12-aedd-426f-9dd3-6c52f9d4c71d] Reconfiguring VM instance instance-00000057 to attach disk [datastore1] fb8eac12-aedd-426f-9dd3-6c52f9d4c71d/fb8eac12-aedd-426f-9dd3-6c52f9d4c71d.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1041.013688] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c34278cf-a710-4b11-9a46-3884d999e013 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.037925] env[65726]: DEBUG oslo_vmware.api [None req-a3acab1e-0b06-47ad-b154-32b257cd5af7 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Waiting for the task: (returnval){ [ 1041.037925] env[65726]: value = "task-5116554" [ 1041.037925] env[65726]: _type = "Task" [ 1041.037925] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1041.051484] env[65726]: DEBUG oslo_concurrency.lockutils [None req-10701ce3-ca7d-4146-b0df-5677afb5eba0 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] Acquiring lock "refresh_cache-df08d166-7811-49b5-9c16-a3434229d623" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1041.051551] env[65726]: DEBUG oslo_concurrency.lockutils [None req-10701ce3-ca7d-4146-b0df-5677afb5eba0 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] Acquired lock "refresh_cache-df08d166-7811-49b5-9c16-a3434229d623" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1041.051760] env[65726]: DEBUG nova.network.neutron [None req-10701ce3-ca7d-4146-b0df-5677afb5eba0 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] [instance: df08d166-7811-49b5-9c16-a3434229d623] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1041.053286] env[65726]: DEBUG oslo_vmware.api [None req-a3acab1e-0b06-47ad-b154-32b257cd5af7 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': task-5116554, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.068142] env[65726]: DEBUG oslo_vmware.api [None req-2a9dd832-d188-4542-a7fe-bbdea8173b2d tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Task: {'id': task-5116549, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.522555} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1041.069064] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-2a9dd832-d188-4542-a7fe-bbdea8173b2d tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] 2c812714-dccc-4d1e-bdb5-c11e446949c3/2c812714-dccc-4d1e-bdb5-c11e446949c3.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1041.069290] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-2a9dd832-d188-4542-a7fe-bbdea8173b2d tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] [instance: 2c812714-dccc-4d1e-bdb5-c11e446949c3] Extending root virtual disk to 1048576 {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1041.069583] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-bc1532a6-8bf3-454a-a3ad-812406a19624 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.080224] env[65726]: DEBUG oslo_vmware.api [None req-2a9dd832-d188-4542-a7fe-bbdea8173b2d tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Waiting for the task: (returnval){ [ 1041.080224] env[65726]: value = "task-5116555" [ 1041.080224] env[65726]: _type = "Task" [ 1041.080224] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1041.089600] env[65726]: DEBUG oslo_vmware.api [None req-2a9dd832-d188-4542-a7fe-bbdea8173b2d tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Task: {'id': task-5116555, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.112607] env[65726]: DEBUG nova.compute.manager [None req-17e357f7-df51-4538-bd2b-ea061ce92818 tempest-ServerMetadataTestJSON-637854735 tempest-ServerMetadataTestJSON-637854735-project-member] [instance: 699127a9-f40f-4c1c-ba4a-625af097350b] Start destroying the instance on the hypervisor. {{(pid=65726) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1041.112874] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-17e357f7-df51-4538-bd2b-ea061ce92818 tempest-ServerMetadataTestJSON-637854735 tempest-ServerMetadataTestJSON-637854735-project-member] [instance: 699127a9-f40f-4c1c-ba4a-625af097350b] Destroying instance {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1041.114121] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ade9e75b-82b6-46c7-aef9-55fb5611739b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.124899] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-17e357f7-df51-4538-bd2b-ea061ce92818 tempest-ServerMetadataTestJSON-637854735 tempest-ServerMetadataTestJSON-637854735-project-member] [instance: 699127a9-f40f-4c1c-ba4a-625af097350b] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1041.128021] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5c1d3531-b2db-49d4-bc1c-4761f4c3966f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.130157] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2314048a-9ba3-4bde-b6be-2a570e8b7d90 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.134949] env[65726]: DEBUG oslo_vmware.api [None req-17e357f7-df51-4538-bd2b-ea061ce92818 tempest-ServerMetadataTestJSON-637854735 tempest-ServerMetadataTestJSON-637854735-project-member] Waiting for the task: (returnval){ [ 1041.134949] env[65726]: value = "task-5116556" [ 1041.134949] env[65726]: _type = "Task" [ 1041.134949] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1041.142414] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df45ca66-394e-412a-a130-7bbe87f18e38 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.153509] env[65726]: DEBUG oslo_vmware.api [None req-17e357f7-df51-4538-bd2b-ea061ce92818 tempest-ServerMetadataTestJSON-637854735 tempest-ServerMetadataTestJSON-637854735-project-member] Task: {'id': task-5116556, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.191148] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6eaac59e-6551-47d8-acdc-8c6114cb5b9d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.196666] env[65726]: DEBUG oslo_vmware.api [None req-1d2791c0-c341-4061-b914-188004153713 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Task: {'id': task-5116552, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.201376] env[65726]: DEBUG nova.network.neutron [None req-0c1711f1-74cb-4ba7-a859-716c68812e37 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: ad4f00b5-3765-494f-8132-8b17b29d219a] Successfully created port: 508e0342-f4cf-409d-96e4-742e92925ba6 {{(pid=65726) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1041.205702] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ab5ed12-31d8-4b48-9419-7b44712f63b7 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.223770] env[65726]: DEBUG nova.compute.provider_tree [None req-43be9038-9e49-4e4d-82ea-842afcd9a64d tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1041.304231] env[65726]: DEBUG nova.compute.manager [None req-0c1711f1-74cb-4ba7-a859-716c68812e37 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: ad4f00b5-3765-494f-8132-8b17b29d219a] Start building block device mappings for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1041.426254] env[65726]: DEBUG oslo_vmware.api [None req-f1e7868e-b3d9-4342-8972-f27e4cb4ec45 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] Task: {'id': task-5116550, 'name': PowerOnVM_Task, 'duration_secs': 0.589835} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1041.426566] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-f1e7868e-b3d9-4342-8972-f27e4cb4ec45 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] [instance: 2b8eb60b-dc9b-47a4-9ddc-205f2af69888] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1041.426765] env[65726]: INFO nova.compute.manager [None req-f1e7868e-b3d9-4342-8972-f27e4cb4ec45 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] [instance: 2b8eb60b-dc9b-47a4-9ddc-205f2af69888] Took 12.58 seconds to spawn the instance on the hypervisor. [ 1041.426937] env[65726]: DEBUG nova.compute.manager [None req-f1e7868e-b3d9-4342-8972-f27e4cb4ec45 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] [instance: 2b8eb60b-dc9b-47a4-9ddc-205f2af69888] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1041.427895] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce63736f-b9e4-41a1-b702-d67c5daa04af {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.549501] env[65726]: DEBUG oslo_vmware.api [None req-a3acab1e-0b06-47ad-b154-32b257cd5af7 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': task-5116554, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.557519] env[65726]: WARNING openstack [None req-10701ce3-ca7d-4146-b0df-5677afb5eba0 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1041.557906] env[65726]: WARNING openstack [None req-10701ce3-ca7d-4146-b0df-5677afb5eba0 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1041.563375] env[65726]: DEBUG nova.network.neutron [None req-10701ce3-ca7d-4146-b0df-5677afb5eba0 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] [instance: df08d166-7811-49b5-9c16-a3434229d623] Instance cache missing network info. {{(pid=65726) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 1041.590972] env[65726]: DEBUG oslo_vmware.api [None req-2a9dd832-d188-4542-a7fe-bbdea8173b2d tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Task: {'id': task-5116555, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.084296} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1041.590972] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-2a9dd832-d188-4542-a7fe-bbdea8173b2d tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] [instance: 2c812714-dccc-4d1e-bdb5-c11e446949c3] Extended root virtual disk {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1041.591634] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-097fe57b-0c80-4669-9689-332b8f8e446d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.618977] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-2a9dd832-d188-4542-a7fe-bbdea8173b2d tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] [instance: 2c812714-dccc-4d1e-bdb5-c11e446949c3] Reconfiguring VM instance instance-00000059 to attach disk [datastore1] 2c812714-dccc-4d1e-bdb5-c11e446949c3/2c812714-dccc-4d1e-bdb5-c11e446949c3.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1041.619652] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e2803fee-e559-488b-83d9-d32863426738 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.645089] env[65726]: DEBUG oslo_vmware.api [None req-2a9dd832-d188-4542-a7fe-bbdea8173b2d tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Waiting for the task: (returnval){ [ 1041.645089] env[65726]: value = "task-5116559" [ 1041.645089] env[65726]: _type = "Task" [ 1041.645089] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1041.648797] env[65726]: DEBUG oslo_vmware.api [None req-17e357f7-df51-4538-bd2b-ea061ce92818 tempest-ServerMetadataTestJSON-637854735 tempest-ServerMetadataTestJSON-637854735-project-member] Task: {'id': task-5116556, 'name': PowerOffVM_Task, 'duration_secs': 0.278355} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1041.652464] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-17e357f7-df51-4538-bd2b-ea061ce92818 tempest-ServerMetadataTestJSON-637854735 tempest-ServerMetadataTestJSON-637854735-project-member] [instance: 699127a9-f40f-4c1c-ba4a-625af097350b] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1041.652708] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-17e357f7-df51-4538-bd2b-ea061ce92818 tempest-ServerMetadataTestJSON-637854735 tempest-ServerMetadataTestJSON-637854735-project-member] [instance: 699127a9-f40f-4c1c-ba4a-625af097350b] Unregistering the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1041.653080] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-cfdd0592-23c2-4fbc-9180-12a260965baa {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.660275] env[65726]: DEBUG oslo_vmware.api [None req-2a9dd832-d188-4542-a7fe-bbdea8173b2d tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Task: {'id': task-5116559, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.670259] env[65726]: WARNING openstack [None req-10701ce3-ca7d-4146-b0df-5677afb5eba0 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1041.670703] env[65726]: WARNING openstack [None req-10701ce3-ca7d-4146-b0df-5677afb5eba0 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1041.695782] env[65726]: DEBUG oslo_vmware.api [None req-1d2791c0-c341-4061-b914-188004153713 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Task: {'id': task-5116552, 'name': ReconfigVM_Task, 'duration_secs': 0.655499} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1041.696166] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-1d2791c0-c341-4061-b914-188004153713 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04] Reconfigured VM instance instance-0000003f to attach disk [datastore2] volume-98eb7918-3cfa-4296-9855-bce47b982be1/volume-98eb7918-3cfa-4296-9855-bce47b982be1.vmdk or device None with type thin {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1041.701476] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0f168e7e-b3c0-4ff9-9bd8-f0d6d0087231 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.720937] env[65726]: DEBUG oslo_vmware.api [None req-1d2791c0-c341-4061-b914-188004153713 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Waiting for the task: (returnval){ [ 1041.720937] env[65726]: value = "task-5116561" [ 1041.720937] env[65726]: _type = "Task" [ 1041.720937] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1041.725705] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-17e357f7-df51-4538-bd2b-ea061ce92818 tempest-ServerMetadataTestJSON-637854735 tempest-ServerMetadataTestJSON-637854735-project-member] [instance: 699127a9-f40f-4c1c-ba4a-625af097350b] Unregistered the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1041.725919] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-17e357f7-df51-4538-bd2b-ea061ce92818 tempest-ServerMetadataTestJSON-637854735 tempest-ServerMetadataTestJSON-637854735-project-member] [instance: 699127a9-f40f-4c1c-ba4a-625af097350b] Deleting contents of the VM from datastore datastore2 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1041.726121] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-17e357f7-df51-4538-bd2b-ea061ce92818 tempest-ServerMetadataTestJSON-637854735 tempest-ServerMetadataTestJSON-637854735-project-member] Deleting the datastore file [datastore2] 699127a9-f40f-4c1c-ba4a-625af097350b {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1041.727473] env[65726]: DEBUG nova.scheduler.client.report [None req-43be9038-9e49-4e4d-82ea-842afcd9a64d tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1041.730919] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8ac72e83-75e8-4266-99e0-baf0cd296615 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.736593] env[65726]: DEBUG oslo_vmware.api [None req-1d2791c0-c341-4061-b914-188004153713 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Task: {'id': task-5116561, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.741695] env[65726]: DEBUG oslo_vmware.api [None req-17e357f7-df51-4538-bd2b-ea061ce92818 tempest-ServerMetadataTestJSON-637854735 tempest-ServerMetadataTestJSON-637854735-project-member] Waiting for the task: (returnval){ [ 1041.741695] env[65726]: value = "task-5116562" [ 1041.741695] env[65726]: _type = "Task" [ 1041.741695] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1041.752280] env[65726]: DEBUG oslo_vmware.api [None req-17e357f7-df51-4538-bd2b-ea061ce92818 tempest-ServerMetadataTestJSON-637854735 tempest-ServerMetadataTestJSON-637854735-project-member] Task: {'id': task-5116562, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.832843] env[65726]: WARNING openstack [None req-10701ce3-ca7d-4146-b0df-5677afb5eba0 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1041.836797] env[65726]: WARNING openstack [None req-10701ce3-ca7d-4146-b0df-5677afb5eba0 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1041.947535] env[65726]: INFO nova.compute.manager [None req-f1e7868e-b3d9-4342-8972-f27e4cb4ec45 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] [instance: 2b8eb60b-dc9b-47a4-9ddc-205f2af69888] Took 22.86 seconds to build instance. [ 1042.051112] env[65726]: DEBUG oslo_vmware.api [None req-a3acab1e-0b06-47ad-b154-32b257cd5af7 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': task-5116554, 'name': ReconfigVM_Task, 'duration_secs': 0.711899} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1042.051496] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-a3acab1e-0b06-47ad-b154-32b257cd5af7 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: fb8eac12-aedd-426f-9dd3-6c52f9d4c71d] Reconfigured VM instance instance-00000057 to attach disk [datastore1] fb8eac12-aedd-426f-9dd3-6c52f9d4c71d/fb8eac12-aedd-426f-9dd3-6c52f9d4c71d.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1042.052412] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7dadda84-ce91-4f81-beb2-4366cffbee3a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.063600] env[65726]: DEBUG nova.network.neutron [None req-10701ce3-ca7d-4146-b0df-5677afb5eba0 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] [instance: df08d166-7811-49b5-9c16-a3434229d623] Updating instance_info_cache with network_info: [{"id": "70efdc7d-7775-4990-b851-d60195bd1504", "address": "fa:16:3e:be:5b:19", "network": {"id": "835ec128-d214-4f54-ab53-88f8378b7d74", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-1935476539-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "305ecace479440ad9a616e546ccd38ce", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3b107fab-ee71-47db-ad4d-3c6f05546843", "external-id": "cl2-zone-554", "segmentation_id": 554, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap70efdc7d-77", "ovs_interfaceid": "70efdc7d-7775-4990-b851-d60195bd1504", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1042.066978] env[65726]: DEBUG oslo_vmware.api [None req-a3acab1e-0b06-47ad-b154-32b257cd5af7 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Waiting for the task: (returnval){ [ 1042.066978] env[65726]: value = "task-5116563" [ 1042.066978] env[65726]: _type = "Task" [ 1042.066978] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1042.078659] env[65726]: DEBUG oslo_vmware.api [None req-a3acab1e-0b06-47ad-b154-32b257cd5af7 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': task-5116563, 'name': Rename_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.159693] env[65726]: DEBUG oslo_vmware.api [None req-2a9dd832-d188-4542-a7fe-bbdea8173b2d tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Task: {'id': task-5116559, 'name': ReconfigVM_Task, 'duration_secs': 0.513725} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1042.160018] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-2a9dd832-d188-4542-a7fe-bbdea8173b2d tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] [instance: 2c812714-dccc-4d1e-bdb5-c11e446949c3] Reconfigured VM instance instance-00000059 to attach disk [datastore1] 2c812714-dccc-4d1e-bdb5-c11e446949c3/2c812714-dccc-4d1e-bdb5-c11e446949c3.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1042.160772] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-fa86a5e9-c8cc-4cb9-aa10-547beec9f74d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.168576] env[65726]: DEBUG oslo_vmware.api [None req-2a9dd832-d188-4542-a7fe-bbdea8173b2d tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Waiting for the task: (returnval){ [ 1042.168576] env[65726]: value = "task-5116564" [ 1042.168576] env[65726]: _type = "Task" [ 1042.168576] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1042.180662] env[65726]: DEBUG oslo_vmware.api [None req-2a9dd832-d188-4542-a7fe-bbdea8173b2d tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Task: {'id': task-5116564, 'name': Rename_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.243437] env[65726]: DEBUG oslo_concurrency.lockutils [None req-43be9038-9e49-4e4d-82ea-842afcd9a64d tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.959s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1042.246926] env[65726]: DEBUG oslo_vmware.api [None req-1d2791c0-c341-4061-b914-188004153713 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Task: {'id': task-5116561, 'name': ReconfigVM_Task, 'duration_secs': 0.311803} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1042.247892] env[65726]: DEBUG oslo_concurrency.lockutils [None req-534e859d-9bab-4a75-8dfe-7304a9dfecfd tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.495s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1042.250456] env[65726]: INFO nova.compute.claims [None req-534e859d-9bab-4a75-8dfe-7304a9dfecfd tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: f947529e-c930-4420-91a7-1e677b5ac2f1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1042.262914] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-1d2791c0-c341-4061-b914-188004153713 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-995255', 'volume_id': '98eb7918-3cfa-4296-9855-bce47b982be1', 'name': 'volume-98eb7918-3cfa-4296-9855-bce47b982be1', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'attached', 'instance': '2fe6fc5a-f5c7-4f8e-96df-4e621a252f04', 'attached_at': '', 'detached_at': '', 'volume_id': '98eb7918-3cfa-4296-9855-bce47b982be1', 'serial': '98eb7918-3cfa-4296-9855-bce47b982be1'} {{(pid=65726) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1042.264374] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-163db54d-59cf-4237-8ec3-265a6a79a60c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.278735] env[65726]: DEBUG oslo_vmware.api [None req-17e357f7-df51-4538-bd2b-ea061ce92818 tempest-ServerMetadataTestJSON-637854735 tempest-ServerMetadataTestJSON-637854735-project-member] Task: {'id': task-5116562, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.308445} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1042.278735] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-17e357f7-df51-4538-bd2b-ea061ce92818 tempest-ServerMetadataTestJSON-637854735 tempest-ServerMetadataTestJSON-637854735-project-member] Deleted the datastore file {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1042.278735] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-17e357f7-df51-4538-bd2b-ea061ce92818 tempest-ServerMetadataTestJSON-637854735 tempest-ServerMetadataTestJSON-637854735-project-member] [instance: 699127a9-f40f-4c1c-ba4a-625af097350b] Deleted contents of the VM from datastore datastore2 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1042.278735] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-17e357f7-df51-4538-bd2b-ea061ce92818 tempest-ServerMetadataTestJSON-637854735 tempest-ServerMetadataTestJSON-637854735-project-member] [instance: 699127a9-f40f-4c1c-ba4a-625af097350b] Instance destroyed {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1042.278735] env[65726]: INFO nova.compute.manager [None req-17e357f7-df51-4538-bd2b-ea061ce92818 tempest-ServerMetadataTestJSON-637854735 tempest-ServerMetadataTestJSON-637854735-project-member] [instance: 699127a9-f40f-4c1c-ba4a-625af097350b] Took 1.16 seconds to destroy the instance on the hypervisor. [ 1042.278735] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-17e357f7-df51-4538-bd2b-ea061ce92818 tempest-ServerMetadataTestJSON-637854735 tempest-ServerMetadataTestJSON-637854735-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1042.278735] env[65726]: DEBUG nova.compute.manager [-] [instance: 699127a9-f40f-4c1c-ba4a-625af097350b] Deallocating network for instance {{(pid=65726) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1042.278735] env[65726]: DEBUG nova.network.neutron [-] [instance: 699127a9-f40f-4c1c-ba4a-625af097350b] deallocate_for_instance() {{(pid=65726) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 1042.279378] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1042.279729] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1042.290038] env[65726]: INFO nova.scheduler.client.report [None req-43be9038-9e49-4e4d-82ea-842afcd9a64d tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Deleted allocations for instance a286ba4f-a83e-4c30-8079-ba88e3ba5696 [ 1042.291730] env[65726]: DEBUG oslo_vmware.api [None req-1d2791c0-c341-4061-b914-188004153713 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Waiting for the task: (returnval){ [ 1042.291730] env[65726]: value = "task-5116565" [ 1042.291730] env[65726]: _type = "Task" [ 1042.291730] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1042.307763] env[65726]: DEBUG oslo_vmware.api [None req-1d2791c0-c341-4061-b914-188004153713 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Task: {'id': task-5116565, 'name': Rename_Task} progress is 10%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.317757] env[65726]: DEBUG nova.compute.manager [None req-0c1711f1-74cb-4ba7-a859-716c68812e37 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: ad4f00b5-3765-494f-8132-8b17b29d219a] Start spawning the instance on the hypervisor. {{(pid=65726) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1042.359577] env[65726]: DEBUG nova.virt.hardware [None req-0c1711f1-74cb-4ba7-a859-716c68812e37 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T19:28:34Z,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1539f06fa3534e90acbc3a60c4b8bd3f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T19:28:35Z,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1042.360124] env[65726]: DEBUG nova.virt.hardware [None req-0c1711f1-74cb-4ba7-a859-716c68812e37 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1042.360124] env[65726]: DEBUG nova.virt.hardware [None req-0c1711f1-74cb-4ba7-a859-716c68812e37 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1042.360259] env[65726]: DEBUG nova.virt.hardware [None req-0c1711f1-74cb-4ba7-a859-716c68812e37 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1042.360448] env[65726]: DEBUG nova.virt.hardware [None req-0c1711f1-74cb-4ba7-a859-716c68812e37 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1042.360773] env[65726]: DEBUG nova.virt.hardware [None req-0c1711f1-74cb-4ba7-a859-716c68812e37 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1042.360858] env[65726]: DEBUG nova.virt.hardware [None req-0c1711f1-74cb-4ba7-a859-716c68812e37 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1042.361119] env[65726]: DEBUG nova.virt.hardware [None req-0c1711f1-74cb-4ba7-a859-716c68812e37 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1042.361322] env[65726]: DEBUG nova.virt.hardware [None req-0c1711f1-74cb-4ba7-a859-716c68812e37 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1042.361497] env[65726]: DEBUG nova.virt.hardware [None req-0c1711f1-74cb-4ba7-a859-716c68812e37 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1042.361680] env[65726]: DEBUG nova.virt.hardware [None req-0c1711f1-74cb-4ba7-a859-716c68812e37 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1042.363297] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3bf73fb2-c7c6-4f5f-8ff2-f4c898629e31 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.373644] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-959b3287-8db8-4656-920a-ac59d11da555 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.449663] env[65726]: DEBUG oslo_concurrency.lockutils [None req-f1e7868e-b3d9-4342-8972-f27e4cb4ec45 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] Lock "2b8eb60b-dc9b-47a4-9ddc-205f2af69888" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 24.377s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1042.568814] env[65726]: DEBUG oslo_concurrency.lockutils [None req-10701ce3-ca7d-4146-b0df-5677afb5eba0 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] Releasing lock "refresh_cache-df08d166-7811-49b5-9c16-a3434229d623" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1042.569249] env[65726]: DEBUG nova.compute.manager [None req-10701ce3-ca7d-4146-b0df-5677afb5eba0 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] [instance: df08d166-7811-49b5-9c16-a3434229d623] Instance network_info: |[{"id": "70efdc7d-7775-4990-b851-d60195bd1504", "address": "fa:16:3e:be:5b:19", "network": {"id": "835ec128-d214-4f54-ab53-88f8378b7d74", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-1935476539-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "305ecace479440ad9a616e546ccd38ce", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3b107fab-ee71-47db-ad4d-3c6f05546843", "external-id": "cl2-zone-554", "segmentation_id": 554, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap70efdc7d-77", "ovs_interfaceid": "70efdc7d-7775-4990-b851-d60195bd1504", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1042.569925] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-10701ce3-ca7d-4146-b0df-5677afb5eba0 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] [instance: df08d166-7811-49b5-9c16-a3434229d623] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:be:5b:19', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3b107fab-ee71-47db-ad4d-3c6f05546843', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '70efdc7d-7775-4990-b851-d60195bd1504', 'vif_model': 'vmxnet3'}] {{(pid=65726) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1042.578015] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-10701ce3-ca7d-4146-b0df-5677afb5eba0 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] Creating folder: Project (305ecace479440ad9a616e546ccd38ce). Parent ref: group-v995008. {{(pid=65726) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1042.581820] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-93449934-7f83-4055-a1b9-e1226cf9ac57 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.589061] env[65726]: DEBUG oslo_vmware.api [None req-a3acab1e-0b06-47ad-b154-32b257cd5af7 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': task-5116563, 'name': Rename_Task, 'duration_secs': 0.181489} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1042.589336] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-a3acab1e-0b06-47ad-b154-32b257cd5af7 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: fb8eac12-aedd-426f-9dd3-6c52f9d4c71d] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1042.589716] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-87129dc6-760e-4638-9c0f-3e5e1de41c00 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.593739] env[65726]: INFO nova.virt.vmwareapi.vm_util [None req-10701ce3-ca7d-4146-b0df-5677afb5eba0 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] Created folder: Project (305ecace479440ad9a616e546ccd38ce) in parent group-v995008. [ 1042.593975] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-10701ce3-ca7d-4146-b0df-5677afb5eba0 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] Creating folder: Instances. Parent ref: group-v995276. {{(pid=65726) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1042.594271] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-10c35478-0e10-41c0-a268-87e987c79f0f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.597240] env[65726]: DEBUG oslo_vmware.api [None req-a3acab1e-0b06-47ad-b154-32b257cd5af7 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Waiting for the task: (returnval){ [ 1042.597240] env[65726]: value = "task-5116567" [ 1042.597240] env[65726]: _type = "Task" [ 1042.597240] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1042.605866] env[65726]: DEBUG oslo_vmware.api [None req-a3acab1e-0b06-47ad-b154-32b257cd5af7 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': task-5116567, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.607826] env[65726]: INFO nova.virt.vmwareapi.vm_util [None req-10701ce3-ca7d-4146-b0df-5677afb5eba0 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] Created folder: Instances in parent group-v995276. [ 1042.607826] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-10701ce3-ca7d-4146-b0df-5677afb5eba0 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1042.607932] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: df08d166-7811-49b5-9c16-a3434229d623] Creating VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1042.608109] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e5b220ae-fb18-4faf-97ba-7ee91d24ba43 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.630012] env[65726]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1042.630012] env[65726]: value = "task-5116569" [ 1042.630012] env[65726]: _type = "Task" [ 1042.630012] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1042.640440] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116569, 'name': CreateVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.681076] env[65726]: DEBUG oslo_vmware.api [None req-2a9dd832-d188-4542-a7fe-bbdea8173b2d tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Task: {'id': task-5116564, 'name': Rename_Task, 'duration_secs': 0.191636} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1042.681551] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-2a9dd832-d188-4542-a7fe-bbdea8173b2d tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] [instance: 2c812714-dccc-4d1e-bdb5-c11e446949c3] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1042.681933] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-194d6bef-db1b-4caa-a4f9-37830323c4df {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.689929] env[65726]: DEBUG oslo_vmware.api [None req-2a9dd832-d188-4542-a7fe-bbdea8173b2d tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Waiting for the task: (returnval){ [ 1042.689929] env[65726]: value = "task-5116570" [ 1042.689929] env[65726]: _type = "Task" [ 1042.689929] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1042.699152] env[65726]: DEBUG oslo_vmware.api [None req-2a9dd832-d188-4542-a7fe-bbdea8173b2d tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Task: {'id': task-5116570, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.805789] env[65726]: DEBUG oslo_vmware.api [None req-1d2791c0-c341-4061-b914-188004153713 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Task: {'id': task-5116565, 'name': Rename_Task, 'duration_secs': 0.228494} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1042.808105] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-1d2791c0-c341-4061-b914-188004153713 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1042.808105] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d5e81edc-34e2-4aaa-8513-8b0032de181c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.816261] env[65726]: DEBUG oslo_concurrency.lockutils [None req-43be9038-9e49-4e4d-82ea-842afcd9a64d tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Lock "a286ba4f-a83e-4c30-8079-ba88e3ba5696" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.745s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1042.822605] env[65726]: DEBUG oslo_vmware.api [None req-1d2791c0-c341-4061-b914-188004153713 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Waiting for the task: (returnval){ [ 1042.822605] env[65726]: value = "task-5116571" [ 1042.822605] env[65726]: _type = "Task" [ 1042.822605] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1042.844684] env[65726]: DEBUG oslo_vmware.api [None req-1d2791c0-c341-4061-b914-188004153713 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Task: {'id': task-5116571, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.847488] env[65726]: DEBUG nova.network.neutron [None req-0c1711f1-74cb-4ba7-a859-716c68812e37 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: ad4f00b5-3765-494f-8132-8b17b29d219a] Successfully updated port: 508e0342-f4cf-409d-96e4-742e92925ba6 {{(pid=65726) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 1042.905216] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1042.905968] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1043.118022] env[65726]: DEBUG oslo_vmware.api [None req-a3acab1e-0b06-47ad-b154-32b257cd5af7 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': task-5116567, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.143483] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116569, 'name': CreateVM_Task, 'duration_secs': 0.482719} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1043.143813] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: df08d166-7811-49b5-9c16-a3434229d623] Created VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1043.144703] env[65726]: WARNING openstack [None req-10701ce3-ca7d-4146-b0df-5677afb5eba0 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1043.145110] env[65726]: WARNING openstack [None req-10701ce3-ca7d-4146-b0df-5677afb5eba0 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1043.150484] env[65726]: DEBUG oslo_concurrency.lockutils [None req-10701ce3-ca7d-4146-b0df-5677afb5eba0 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1043.151351] env[65726]: DEBUG oslo_concurrency.lockutils [None req-10701ce3-ca7d-4146-b0df-5677afb5eba0 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1043.151351] env[65726]: DEBUG oslo_concurrency.lockutils [None req-10701ce3-ca7d-4146-b0df-5677afb5eba0 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1043.151351] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bf4b6ef5-5f30-45fa-b457-97576d038384 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.157252] env[65726]: DEBUG oslo_vmware.api [None req-10701ce3-ca7d-4146-b0df-5677afb5eba0 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] Waiting for the task: (returnval){ [ 1043.157252] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52cd9866-793b-41f7-ab62-c742459347fe" [ 1043.157252] env[65726]: _type = "Task" [ 1043.157252] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1043.166556] env[65726]: DEBUG oslo_vmware.api [None req-10701ce3-ca7d-4146-b0df-5677afb5eba0 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52cd9866-793b-41f7-ab62-c742459347fe, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.206353] env[65726]: DEBUG oslo_vmware.api [None req-2a9dd832-d188-4542-a7fe-bbdea8173b2d tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Task: {'id': task-5116570, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.342132] env[65726]: DEBUG oslo_vmware.api [None req-1d2791c0-c341-4061-b914-188004153713 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Task: {'id': task-5116571, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.354431] env[65726]: DEBUG oslo_concurrency.lockutils [None req-0c1711f1-74cb-4ba7-a859-716c68812e37 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Acquiring lock "refresh_cache-ad4f00b5-3765-494f-8132-8b17b29d219a" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1043.354431] env[65726]: DEBUG oslo_concurrency.lockutils [None req-0c1711f1-74cb-4ba7-a859-716c68812e37 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Acquired lock "refresh_cache-ad4f00b5-3765-494f-8132-8b17b29d219a" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1043.354559] env[65726]: DEBUG nova.network.neutron [None req-0c1711f1-74cb-4ba7-a859-716c68812e37 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: ad4f00b5-3765-494f-8132-8b17b29d219a] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1043.563612] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85dfb892-94da-489f-9346-3cd1b13f6356 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.573131] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-490cd9c0-0fc2-4789-9d45-3cf3463a36a8 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.609761] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b0f1a31-c88c-4561-a294-3b19657e7d10 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.622049] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea439e4b-dc2c-4e9d-930a-e0c93c7efd2b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.626945] env[65726]: DEBUG oslo_vmware.api [None req-a3acab1e-0b06-47ad-b154-32b257cd5af7 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': task-5116567, 'name': PowerOnVM_Task, 'duration_secs': 0.744993} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1043.628117] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-a3acab1e-0b06-47ad-b154-32b257cd5af7 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: fb8eac12-aedd-426f-9dd3-6c52f9d4c71d] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1043.628117] env[65726]: DEBUG nova.compute.manager [None req-a3acab1e-0b06-47ad-b154-32b257cd5af7 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: fb8eac12-aedd-426f-9dd3-6c52f9d4c71d] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1043.629207] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02ae9dbb-0371-4ee9-a534-91213250c20d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.645581] env[65726]: DEBUG nova.compute.provider_tree [None req-534e859d-9bab-4a75-8dfe-7304a9dfecfd tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1043.673370] env[65726]: DEBUG oslo_vmware.api [None req-10701ce3-ca7d-4146-b0df-5677afb5eba0 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52cd9866-793b-41f7-ab62-c742459347fe, 'name': SearchDatastore_Task, 'duration_secs': 0.031272} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1043.673790] env[65726]: DEBUG oslo_concurrency.lockutils [None req-10701ce3-ca7d-4146-b0df-5677afb5eba0 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1043.674300] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-10701ce3-ca7d-4146-b0df-5677afb5eba0 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] [instance: df08d166-7811-49b5-9c16-a3434229d623] Processing image b52362a3-ee8a-4cbf-b06f-513b0cc8f95c {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1043.674580] env[65726]: DEBUG oslo_concurrency.lockutils [None req-10701ce3-ca7d-4146-b0df-5677afb5eba0 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1043.674759] env[65726]: DEBUG oslo_concurrency.lockutils [None req-10701ce3-ca7d-4146-b0df-5677afb5eba0 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1043.675069] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-10701ce3-ca7d-4146-b0df-5677afb5eba0 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1043.675454] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b7720207-7caf-41bd-91a4-d3abedc09359 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.678426] env[65726]: DEBUG nova.network.neutron [-] [instance: 699127a9-f40f-4c1c-ba4a-625af097350b] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1043.689805] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-10701ce3-ca7d-4146-b0df-5677afb5eba0 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1043.690189] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-10701ce3-ca7d-4146-b0df-5677afb5eba0 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65726) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1043.691730] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e75d7dc2-b37a-4414-8653-5bf46851d56d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.708691] env[65726]: DEBUG oslo_vmware.api [None req-10701ce3-ca7d-4146-b0df-5677afb5eba0 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] Waiting for the task: (returnval){ [ 1043.708691] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5255372b-8901-415e-c788-f99a38dc302e" [ 1043.708691] env[65726]: _type = "Task" [ 1043.708691] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1043.715718] env[65726]: DEBUG oslo_vmware.api [None req-2a9dd832-d188-4542-a7fe-bbdea8173b2d tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Task: {'id': task-5116570, 'name': PowerOnVM_Task, 'duration_secs': 0.644271} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1043.721136] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-2a9dd832-d188-4542-a7fe-bbdea8173b2d tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] [instance: 2c812714-dccc-4d1e-bdb5-c11e446949c3] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1043.721537] env[65726]: INFO nova.compute.manager [None req-2a9dd832-d188-4542-a7fe-bbdea8173b2d tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] [instance: 2c812714-dccc-4d1e-bdb5-c11e446949c3] Took 12.48 seconds to spawn the instance on the hypervisor. [ 1043.721801] env[65726]: DEBUG nova.compute.manager [None req-2a9dd832-d188-4542-a7fe-bbdea8173b2d tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] [instance: 2c812714-dccc-4d1e-bdb5-c11e446949c3] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1043.723778] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9332d828-aa20-47c8-bdc1-5cc088cd9b08 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.746043] env[65726]: DEBUG oslo_vmware.api [None req-10701ce3-ca7d-4146-b0df-5677afb5eba0 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5255372b-8901-415e-c788-f99a38dc302e, 'name': SearchDatastore_Task, 'duration_secs': 0.013217} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1043.747787] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-85a60fe2-6212-489d-abdb-8146f6905350 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.756763] env[65726]: DEBUG oslo_vmware.api [None req-10701ce3-ca7d-4146-b0df-5677afb5eba0 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] Waiting for the task: (returnval){ [ 1043.756763] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]523619a3-a5ce-3008-8892-b2df77cf5bd9" [ 1043.756763] env[65726]: _type = "Task" [ 1043.756763] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1043.770869] env[65726]: DEBUG oslo_vmware.api [None req-10701ce3-ca7d-4146-b0df-5677afb5eba0 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]523619a3-a5ce-3008-8892-b2df77cf5bd9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.840062] env[65726]: DEBUG oslo_vmware.api [None req-1d2791c0-c341-4061-b914-188004153713 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Task: {'id': task-5116571, 'name': PowerOnVM_Task, 'duration_secs': 0.634139} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1043.840281] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-1d2791c0-c341-4061-b914-188004153713 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1043.861710] env[65726]: WARNING openstack [None req-0c1711f1-74cb-4ba7-a859-716c68812e37 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1043.862197] env[65726]: WARNING openstack [None req-0c1711f1-74cb-4ba7-a859-716c68812e37 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1043.867454] env[65726]: DEBUG nova.network.neutron [None req-0c1711f1-74cb-4ba7-a859-716c68812e37 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: ad4f00b5-3765-494f-8132-8b17b29d219a] Instance cache missing network info. {{(pid=65726) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 1043.968864] env[65726]: DEBUG nova.compute.manager [None req-1d2791c0-c341-4061-b914-188004153713 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1043.969838] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77660486-b8a2-4346-a3b3-be34e7906a06 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.997772] env[65726]: WARNING openstack [None req-0c1711f1-74cb-4ba7-a859-716c68812e37 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1043.998249] env[65726]: WARNING openstack [None req-0c1711f1-74cb-4ba7-a859-716c68812e37 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1044.042503] env[65726]: DEBUG nova.compute.manager [req-36a9dadc-c26e-4e4d-988e-bc25089e4c83 req-ee61a719-26f1-4910-a7f3-a20f87551f6c service nova] [instance: df08d166-7811-49b5-9c16-a3434229d623] Received event network-changed-70efdc7d-7775-4990-b851-d60195bd1504 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1044.042810] env[65726]: DEBUG nova.compute.manager [req-36a9dadc-c26e-4e4d-988e-bc25089e4c83 req-ee61a719-26f1-4910-a7f3-a20f87551f6c service nova] [instance: df08d166-7811-49b5-9c16-a3434229d623] Refreshing instance network info cache due to event network-changed-70efdc7d-7775-4990-b851-d60195bd1504. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 1044.043098] env[65726]: DEBUG oslo_concurrency.lockutils [req-36a9dadc-c26e-4e4d-988e-bc25089e4c83 req-ee61a719-26f1-4910-a7f3-a20f87551f6c service nova] Acquiring lock "refresh_cache-df08d166-7811-49b5-9c16-a3434229d623" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1044.043243] env[65726]: DEBUG oslo_concurrency.lockutils [req-36a9dadc-c26e-4e4d-988e-bc25089e4c83 req-ee61a719-26f1-4910-a7f3-a20f87551f6c service nova] Acquired lock "refresh_cache-df08d166-7811-49b5-9c16-a3434229d623" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1044.043401] env[65726]: DEBUG nova.network.neutron [req-36a9dadc-c26e-4e4d-988e-bc25089e4c83 req-ee61a719-26f1-4910-a7f3-a20f87551f6c service nova] [instance: df08d166-7811-49b5-9c16-a3434229d623] Refreshing network info cache for port 70efdc7d-7775-4990-b851-d60195bd1504 {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 1044.077281] env[65726]: WARNING openstack [None req-0c1711f1-74cb-4ba7-a859-716c68812e37 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1044.077851] env[65726]: WARNING openstack [None req-0c1711f1-74cb-4ba7-a859-716c68812e37 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1044.148660] env[65726]: DEBUG nova.scheduler.client.report [None req-534e859d-9bab-4a75-8dfe-7304a9dfecfd tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1044.162492] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a3acab1e-0b06-47ad-b154-32b257cd5af7 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1044.180838] env[65726]: INFO nova.compute.manager [-] [instance: 699127a9-f40f-4c1c-ba4a-625af097350b] Took 1.90 seconds to deallocate network for instance. [ 1044.227240] env[65726]: DEBUG nova.compute.manager [req-4e086789-d0ae-4c79-85af-625b9d8b5e41 req-17384f58-25d4-4a1c-af80-b47b81832398 service nova] [instance: 699127a9-f40f-4c1c-ba4a-625af097350b] Received event network-vif-deleted-544abfef-23be-48c2-932e-1bf64094c763 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1044.227681] env[65726]: DEBUG nova.network.neutron [None req-0c1711f1-74cb-4ba7-a859-716c68812e37 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: ad4f00b5-3765-494f-8132-8b17b29d219a] Updating instance_info_cache with network_info: [{"id": "508e0342-f4cf-409d-96e4-742e92925ba6", "address": "fa:16:3e:89:e2:97", "network": {"id": "ad1373b5-e0ed-4ee6-a97c-af931afbab35", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-909418417-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a090d553766847e58e3231b966c92565", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5f60c972-a72d-4c5f-a250-faadfd6eafbe", "external-id": "nsx-vlan-transportzone-932", "segmentation_id": 932, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap508e0342-f4", "ovs_interfaceid": "508e0342-f4cf-409d-96e4-742e92925ba6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1044.266085] env[65726]: INFO nova.compute.manager [None req-2a9dd832-d188-4542-a7fe-bbdea8173b2d tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] [instance: 2c812714-dccc-4d1e-bdb5-c11e446949c3] Took 23.56 seconds to build instance. [ 1044.280674] env[65726]: DEBUG oslo_vmware.api [None req-10701ce3-ca7d-4146-b0df-5677afb5eba0 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]523619a3-a5ce-3008-8892-b2df77cf5bd9, 'name': SearchDatastore_Task, 'duration_secs': 0.013614} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1044.280972] env[65726]: DEBUG oslo_concurrency.lockutils [None req-10701ce3-ca7d-4146-b0df-5677afb5eba0 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1044.281282] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-10701ce3-ca7d-4146-b0df-5677afb5eba0 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore2] df08d166-7811-49b5-9c16-a3434229d623/df08d166-7811-49b5-9c16-a3434229d623.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1044.281601] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-08c0c04f-4c79-407e-929e-0883e74a621a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.292723] env[65726]: DEBUG oslo_vmware.api [None req-10701ce3-ca7d-4146-b0df-5677afb5eba0 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] Waiting for the task: (returnval){ [ 1044.292723] env[65726]: value = "task-5116573" [ 1044.292723] env[65726]: _type = "Task" [ 1044.292723] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1044.303643] env[65726]: DEBUG oslo_vmware.api [None req-10701ce3-ca7d-4146-b0df-5677afb5eba0 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] Task: {'id': task-5116573, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1044.489785] env[65726]: DEBUG oslo_concurrency.lockutils [None req-1d2791c0-c341-4061-b914-188004153713 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Lock "2fe6fc5a-f5c7-4f8e-96df-4e621a252f04" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 37.019s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1044.548481] env[65726]: WARNING openstack [req-36a9dadc-c26e-4e4d-988e-bc25089e4c83 req-ee61a719-26f1-4910-a7f3-a20f87551f6c service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1044.549047] env[65726]: WARNING openstack [req-36a9dadc-c26e-4e4d-988e-bc25089e4c83 req-ee61a719-26f1-4910-a7f3-a20f87551f6c service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1044.661578] env[65726]: DEBUG oslo_concurrency.lockutils [None req-534e859d-9bab-4a75-8dfe-7304a9dfecfd tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.414s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1044.662382] env[65726]: DEBUG nova.compute.manager [None req-534e859d-9bab-4a75-8dfe-7304a9dfecfd tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: f947529e-c930-4420-91a7-1e677b5ac2f1] Start building networks asynchronously for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1044.665322] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5c22cc7c-7dcc-417f-8f16-1f19636dd1f0 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.998s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1044.667318] env[65726]: INFO nova.compute.claims [None req-5c22cc7c-7dcc-417f-8f16-1f19636dd1f0 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: 51f5337d-2e23-4b7c-b590-76be6bd82411] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1044.689153] env[65726]: DEBUG oslo_concurrency.lockutils [None req-17e357f7-df51-4538-bd2b-ea061ce92818 tempest-ServerMetadataTestJSON-637854735 tempest-ServerMetadataTestJSON-637854735-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1044.732125] env[65726]: WARNING openstack [req-36a9dadc-c26e-4e4d-988e-bc25089e4c83 req-ee61a719-26f1-4910-a7f3-a20f87551f6c service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1044.732736] env[65726]: WARNING openstack [req-36a9dadc-c26e-4e4d-988e-bc25089e4c83 req-ee61a719-26f1-4910-a7f3-a20f87551f6c service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1044.743251] env[65726]: DEBUG oslo_concurrency.lockutils [None req-0c1711f1-74cb-4ba7-a859-716c68812e37 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Releasing lock "refresh_cache-ad4f00b5-3765-494f-8132-8b17b29d219a" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1044.743768] env[65726]: DEBUG nova.compute.manager [None req-0c1711f1-74cb-4ba7-a859-716c68812e37 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: ad4f00b5-3765-494f-8132-8b17b29d219a] Instance network_info: |[{"id": "508e0342-f4cf-409d-96e4-742e92925ba6", "address": "fa:16:3e:89:e2:97", "network": {"id": "ad1373b5-e0ed-4ee6-a97c-af931afbab35", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-909418417-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a090d553766847e58e3231b966c92565", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5f60c972-a72d-4c5f-a250-faadfd6eafbe", "external-id": "nsx-vlan-transportzone-932", "segmentation_id": 932, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap508e0342-f4", "ovs_interfaceid": "508e0342-f4cf-409d-96e4-742e92925ba6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1044.745686] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-0c1711f1-74cb-4ba7-a859-716c68812e37 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: ad4f00b5-3765-494f-8132-8b17b29d219a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:89:e2:97', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '5f60c972-a72d-4c5f-a250-faadfd6eafbe', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '508e0342-f4cf-409d-96e4-742e92925ba6', 'vif_model': 'vmxnet3'}] {{(pid=65726) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1044.759501] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-0c1711f1-74cb-4ba7-a859-716c68812e37 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1044.760393] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ad4f00b5-3765-494f-8132-8b17b29d219a] Creating VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1044.760652] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-08ac3934-0b77-442c-8917-f189e01e4bc0 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.778235] env[65726]: DEBUG oslo_concurrency.lockutils [None req-03f877a9-8771-4f37-b985-f35105d89922 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Acquiring lock "9fc0c9ee-9dc7-4059-b8d2-a5db965e00c2" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1044.778492] env[65726]: DEBUG oslo_concurrency.lockutils [None req-03f877a9-8771-4f37-b985-f35105d89922 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Lock "9fc0c9ee-9dc7-4059-b8d2-a5db965e00c2" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1044.778697] env[65726]: DEBUG oslo_concurrency.lockutils [None req-03f877a9-8771-4f37-b985-f35105d89922 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Acquiring lock "9fc0c9ee-9dc7-4059-b8d2-a5db965e00c2-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1044.778873] env[65726]: DEBUG oslo_concurrency.lockutils [None req-03f877a9-8771-4f37-b985-f35105d89922 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Lock "9fc0c9ee-9dc7-4059-b8d2-a5db965e00c2-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1044.779268] env[65726]: DEBUG oslo_concurrency.lockutils [None req-03f877a9-8771-4f37-b985-f35105d89922 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Lock "9fc0c9ee-9dc7-4059-b8d2-a5db965e00c2-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1044.781080] env[65726]: DEBUG oslo_concurrency.lockutils [None req-2a9dd832-d188-4542-a7fe-bbdea8173b2d tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Lock "2c812714-dccc-4d1e-bdb5-c11e446949c3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 25.084s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1044.781741] env[65726]: INFO nova.compute.manager [None req-03f877a9-8771-4f37-b985-f35105d89922 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: 9fc0c9ee-9dc7-4059-b8d2-a5db965e00c2] Terminating instance [ 1044.789671] env[65726]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1044.789671] env[65726]: value = "task-5116574" [ 1044.789671] env[65726]: _type = "Task" [ 1044.789671] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1044.812021] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116574, 'name': CreateVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1044.818325] env[65726]: DEBUG oslo_vmware.api [None req-10701ce3-ca7d-4146-b0df-5677afb5eba0 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] Task: {'id': task-5116573, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1044.847863] env[65726]: WARNING openstack [req-36a9dadc-c26e-4e4d-988e-bc25089e4c83 req-ee61a719-26f1-4910-a7f3-a20f87551f6c service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1044.848313] env[65726]: WARNING openstack [req-36a9dadc-c26e-4e4d-988e-bc25089e4c83 req-ee61a719-26f1-4910-a7f3-a20f87551f6c service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1044.954053] env[65726]: DEBUG nova.network.neutron [req-36a9dadc-c26e-4e4d-988e-bc25089e4c83 req-ee61a719-26f1-4910-a7f3-a20f87551f6c service nova] [instance: df08d166-7811-49b5-9c16-a3434229d623] Updated VIF entry in instance network info cache for port 70efdc7d-7775-4990-b851-d60195bd1504. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 1044.954442] env[65726]: DEBUG nova.network.neutron [req-36a9dadc-c26e-4e4d-988e-bc25089e4c83 req-ee61a719-26f1-4910-a7f3-a20f87551f6c service nova] [instance: df08d166-7811-49b5-9c16-a3434229d623] Updating instance_info_cache with network_info: [{"id": "70efdc7d-7775-4990-b851-d60195bd1504", "address": "fa:16:3e:be:5b:19", "network": {"id": "835ec128-d214-4f54-ab53-88f8378b7d74", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-1935476539-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "305ecace479440ad9a616e546ccd38ce", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3b107fab-ee71-47db-ad4d-3c6f05546843", "external-id": "cl2-zone-554", "segmentation_id": 554, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap70efdc7d-77", "ovs_interfaceid": "70efdc7d-7775-4990-b851-d60195bd1504", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1045.172350] env[65726]: DEBUG nova.compute.utils [None req-534e859d-9bab-4a75-8dfe-7304a9dfecfd tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Using /dev/sd instead of None {{(pid=65726) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1045.180882] env[65726]: DEBUG nova.compute.manager [None req-534e859d-9bab-4a75-8dfe-7304a9dfecfd tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: f947529e-c930-4420-91a7-1e677b5ac2f1] Allocating IP information in the background. {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1045.181211] env[65726]: DEBUG nova.network.neutron [None req-534e859d-9bab-4a75-8dfe-7304a9dfecfd tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: f947529e-c930-4420-91a7-1e677b5ac2f1] allocate_for_instance() {{(pid=65726) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 1045.181550] env[65726]: WARNING neutronclient.v2_0.client [None req-534e859d-9bab-4a75-8dfe-7304a9dfecfd tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1045.182456] env[65726]: WARNING neutronclient.v2_0.client [None req-534e859d-9bab-4a75-8dfe-7304a9dfecfd tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1045.183880] env[65726]: WARNING openstack [None req-534e859d-9bab-4a75-8dfe-7304a9dfecfd tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1045.183880] env[65726]: WARNING openstack [None req-534e859d-9bab-4a75-8dfe-7304a9dfecfd tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1045.290891] env[65726]: DEBUG nova.compute.manager [None req-03f877a9-8771-4f37-b985-f35105d89922 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: 9fc0c9ee-9dc7-4059-b8d2-a5db965e00c2] Start destroying the instance on the hypervisor. {{(pid=65726) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1045.290891] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-03f877a9-8771-4f37-b985-f35105d89922 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: 9fc0c9ee-9dc7-4059-b8d2-a5db965e00c2] Destroying instance {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1045.290891] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6498f2b-6c8d-4869-80d9-78c7c7216176 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.313492] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116574, 'name': CreateVM_Task} progress is 25%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.314437] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-03f877a9-8771-4f37-b985-f35105d89922 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: 9fc0c9ee-9dc7-4059-b8d2-a5db965e00c2] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1045.319757] env[65726]: DEBUG nova.policy [None req-534e859d-9bab-4a75-8dfe-7304a9dfecfd tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4cd1ed63db174f21be78f86554fada0a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'cbb17a09b35c4c22ade5c4082b10bda9', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65726) authorize /opt/stack/nova/nova/policy.py:192}} [ 1045.323953] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-bbd007e8-174e-4cce-b6eb-1ef24aa17c0c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.331146] env[65726]: DEBUG oslo_vmware.api [None req-10701ce3-ca7d-4146-b0df-5677afb5eba0 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] Task: {'id': task-5116573, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.678365} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1045.331501] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-10701ce3-ca7d-4146-b0df-5677afb5eba0 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore2] df08d166-7811-49b5-9c16-a3434229d623/df08d166-7811-49b5-9c16-a3434229d623.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1045.331744] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-10701ce3-ca7d-4146-b0df-5677afb5eba0 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] [instance: df08d166-7811-49b5-9c16-a3434229d623] Extending root virtual disk to 1048576 {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1045.332139] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1b9b939a-2605-428c-924c-cfb75fdcbdb5 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.337129] env[65726]: DEBUG oslo_vmware.api [None req-03f877a9-8771-4f37-b985-f35105d89922 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Waiting for the task: (returnval){ [ 1045.337129] env[65726]: value = "task-5116575" [ 1045.337129] env[65726]: _type = "Task" [ 1045.337129] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1045.344494] env[65726]: DEBUG oslo_vmware.api [None req-10701ce3-ca7d-4146-b0df-5677afb5eba0 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] Waiting for the task: (returnval){ [ 1045.344494] env[65726]: value = "task-5116576" [ 1045.344494] env[65726]: _type = "Task" [ 1045.344494] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1045.354391] env[65726]: DEBUG oslo_vmware.api [None req-03f877a9-8771-4f37-b985-f35105d89922 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Task: {'id': task-5116575, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.361687] env[65726]: DEBUG oslo_vmware.api [None req-10701ce3-ca7d-4146-b0df-5677afb5eba0 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] Task: {'id': task-5116576, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.421588] env[65726]: DEBUG oslo_concurrency.lockutils [None req-4e9ea6a2-e7a5-42fb-bc17-3cf3a7f02eb9 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] Acquiring lock "2b8eb60b-dc9b-47a4-9ddc-205f2af69888" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1045.421875] env[65726]: DEBUG oslo_concurrency.lockutils [None req-4e9ea6a2-e7a5-42fb-bc17-3cf3a7f02eb9 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] Lock "2b8eb60b-dc9b-47a4-9ddc-205f2af69888" acquired by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1045.422183] env[65726]: INFO nova.compute.manager [None req-4e9ea6a2-e7a5-42fb-bc17-3cf3a7f02eb9 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] [instance: 2b8eb60b-dc9b-47a4-9ddc-205f2af69888] Rebooting instance [ 1045.457878] env[65726]: DEBUG oslo_concurrency.lockutils [req-36a9dadc-c26e-4e4d-988e-bc25089e4c83 req-ee61a719-26f1-4910-a7f3-a20f87551f6c service nova] Releasing lock "refresh_cache-df08d166-7811-49b5-9c16-a3434229d623" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1045.458228] env[65726]: DEBUG nova.compute.manager [req-36a9dadc-c26e-4e4d-988e-bc25089e4c83 req-ee61a719-26f1-4910-a7f3-a20f87551f6c service nova] [instance: ad4f00b5-3765-494f-8132-8b17b29d219a] Received event network-vif-plugged-508e0342-f4cf-409d-96e4-742e92925ba6 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1045.458463] env[65726]: DEBUG oslo_concurrency.lockutils [req-36a9dadc-c26e-4e4d-988e-bc25089e4c83 req-ee61a719-26f1-4910-a7f3-a20f87551f6c service nova] Acquiring lock "ad4f00b5-3765-494f-8132-8b17b29d219a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1045.458718] env[65726]: DEBUG oslo_concurrency.lockutils [req-36a9dadc-c26e-4e4d-988e-bc25089e4c83 req-ee61a719-26f1-4910-a7f3-a20f87551f6c service nova] Lock "ad4f00b5-3765-494f-8132-8b17b29d219a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1045.458910] env[65726]: DEBUG oslo_concurrency.lockutils [req-36a9dadc-c26e-4e4d-988e-bc25089e4c83 req-ee61a719-26f1-4910-a7f3-a20f87551f6c service nova] Lock "ad4f00b5-3765-494f-8132-8b17b29d219a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1045.459107] env[65726]: DEBUG nova.compute.manager [req-36a9dadc-c26e-4e4d-988e-bc25089e4c83 req-ee61a719-26f1-4910-a7f3-a20f87551f6c service nova] [instance: ad4f00b5-3765-494f-8132-8b17b29d219a] No waiting events found dispatching network-vif-plugged-508e0342-f4cf-409d-96e4-742e92925ba6 {{(pid=65726) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1045.459280] env[65726]: WARNING nova.compute.manager [req-36a9dadc-c26e-4e4d-988e-bc25089e4c83 req-ee61a719-26f1-4910-a7f3-a20f87551f6c service nova] [instance: ad4f00b5-3765-494f-8132-8b17b29d219a] Received unexpected event network-vif-plugged-508e0342-f4cf-409d-96e4-742e92925ba6 for instance with vm_state building and task_state spawning. [ 1045.642407] env[65726]: DEBUG oslo_concurrency.lockutils [None req-fe2eb30a-001a-49d1-91b1-ed3f11534bdf tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Acquiring lock "2c812714-dccc-4d1e-bdb5-c11e446949c3" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1045.642673] env[65726]: DEBUG oslo_concurrency.lockutils [None req-fe2eb30a-001a-49d1-91b1-ed3f11534bdf tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Lock "2c812714-dccc-4d1e-bdb5-c11e446949c3" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1045.642907] env[65726]: DEBUG oslo_concurrency.lockutils [None req-fe2eb30a-001a-49d1-91b1-ed3f11534bdf tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Acquiring lock "2c812714-dccc-4d1e-bdb5-c11e446949c3-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1045.643119] env[65726]: DEBUG oslo_concurrency.lockutils [None req-fe2eb30a-001a-49d1-91b1-ed3f11534bdf tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Lock "2c812714-dccc-4d1e-bdb5-c11e446949c3-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1045.643356] env[65726]: DEBUG oslo_concurrency.lockutils [None req-fe2eb30a-001a-49d1-91b1-ed3f11534bdf tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Lock "2c812714-dccc-4d1e-bdb5-c11e446949c3-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1045.647900] env[65726]: INFO nova.compute.manager [None req-fe2eb30a-001a-49d1-91b1-ed3f11534bdf tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] [instance: 2c812714-dccc-4d1e-bdb5-c11e446949c3] Terminating instance [ 1045.653303] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-031ff252-21d5-4d9e-b884-2b2939b62196 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: a56be6fa-6027-46ee-9aa8-332fe4222d68] Volume attach. Driver type: vmdk {{(pid=65726) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1045.653524] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-031ff252-21d5-4d9e-b884-2b2939b62196 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: a56be6fa-6027-46ee-9aa8-332fe4222d68] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-995275', 'volume_id': '30181d53-b2b9-4253-a8f5-28a3abf812f9', 'name': 'volume-30181d53-b2b9-4253-a8f5-28a3abf812f9', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'a56be6fa-6027-46ee-9aa8-332fe4222d68', 'attached_at': '', 'detached_at': '', 'volume_id': '30181d53-b2b9-4253-a8f5-28a3abf812f9', 'serial': '30181d53-b2b9-4253-a8f5-28a3abf812f9'} {{(pid=65726) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1045.654768] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afcef58a-00cf-4f1c-87ab-fa928b8db622 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.678094] env[65726]: DEBUG nova.network.neutron [None req-534e859d-9bab-4a75-8dfe-7304a9dfecfd tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: f947529e-c930-4420-91a7-1e677b5ac2f1] Successfully created port: 4521ce71-46b3-44a7-a0d8-6172da592893 {{(pid=65726) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1045.682242] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cade3479-e0a3-46f0-9ea5-231969a91a87 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.685942] env[65726]: DEBUG nova.compute.manager [None req-534e859d-9bab-4a75-8dfe-7304a9dfecfd tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: f947529e-c930-4420-91a7-1e677b5ac2f1] Start building block device mappings for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1045.718518] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-031ff252-21d5-4d9e-b884-2b2939b62196 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: a56be6fa-6027-46ee-9aa8-332fe4222d68] Reconfiguring VM instance instance-00000053 to attach disk [datastore2] volume-30181d53-b2b9-4253-a8f5-28a3abf812f9/volume-30181d53-b2b9-4253-a8f5-28a3abf812f9.vmdk or device None with type thin {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1045.719269] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7110b6e3-757b-4bd3-a2e5-32a4f9769f9c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.745712] env[65726]: DEBUG oslo_vmware.api [None req-031ff252-21d5-4d9e-b884-2b2939b62196 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Waiting for the task: (returnval){ [ 1045.745712] env[65726]: value = "task-5116577" [ 1045.745712] env[65726]: _type = "Task" [ 1045.745712] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1045.766570] env[65726]: DEBUG oslo_vmware.api [None req-031ff252-21d5-4d9e-b884-2b2939b62196 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5116577, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.809044] env[65726]: DEBUG oslo_concurrency.lockutils [None req-cc47cd13-9595-40ad-a0c8-afcfcfe7f019 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Acquiring lock "48923e3d-b00d-4034-bacf-7be82009fb08" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1045.809199] env[65726]: DEBUG oslo_concurrency.lockutils [None req-cc47cd13-9595-40ad-a0c8-afcfcfe7f019 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Lock "48923e3d-b00d-4034-bacf-7be82009fb08" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1045.810420] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116574, 'name': CreateVM_Task, 'duration_secs': 0.675133} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1045.813510] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ad4f00b5-3765-494f-8132-8b17b29d219a] Created VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1045.815136] env[65726]: WARNING openstack [None req-0c1711f1-74cb-4ba7-a859-716c68812e37 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1045.815513] env[65726]: WARNING openstack [None req-0c1711f1-74cb-4ba7-a859-716c68812e37 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1045.820597] env[65726]: DEBUG oslo_concurrency.lockutils [None req-0c1711f1-74cb-4ba7-a859-716c68812e37 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1045.820762] env[65726]: DEBUG oslo_concurrency.lockutils [None req-0c1711f1-74cb-4ba7-a859-716c68812e37 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1045.821082] env[65726]: DEBUG oslo_concurrency.lockutils [None req-0c1711f1-74cb-4ba7-a859-716c68812e37 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1045.822203] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-96baf438-3597-4d43-a510-2245b606520f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.829248] env[65726]: DEBUG oslo_vmware.api [None req-0c1711f1-74cb-4ba7-a859-716c68812e37 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Waiting for the task: (returnval){ [ 1045.829248] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]522d3877-8207-1aba-41a1-bd8029df72c1" [ 1045.829248] env[65726]: _type = "Task" [ 1045.829248] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1045.842708] env[65726]: DEBUG oslo_vmware.api [None req-0c1711f1-74cb-4ba7-a859-716c68812e37 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]522d3877-8207-1aba-41a1-bd8029df72c1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.862116] env[65726]: DEBUG oslo_vmware.api [None req-10701ce3-ca7d-4146-b0df-5677afb5eba0 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] Task: {'id': task-5116576, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.090168} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1045.862402] env[65726]: DEBUG oslo_vmware.api [None req-03f877a9-8771-4f37-b985-f35105d89922 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Task: {'id': task-5116575, 'name': PowerOffVM_Task, 'duration_secs': 0.469094} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1045.865471] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-10701ce3-ca7d-4146-b0df-5677afb5eba0 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] [instance: df08d166-7811-49b5-9c16-a3434229d623] Extended root virtual disk {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1045.865777] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-03f877a9-8771-4f37-b985-f35105d89922 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: 9fc0c9ee-9dc7-4059-b8d2-a5db965e00c2] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1045.865939] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-03f877a9-8771-4f37-b985-f35105d89922 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: 9fc0c9ee-9dc7-4059-b8d2-a5db965e00c2] Unregistering the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1045.867015] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e699c28-0da2-4d23-a40a-82ae94f83f5b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.869629] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d926f538-5afb-4748-b805-9cfd343dffe1 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.897205] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-10701ce3-ca7d-4146-b0df-5677afb5eba0 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] [instance: df08d166-7811-49b5-9c16-a3434229d623] Reconfiguring VM instance instance-0000005a to attach disk [datastore2] df08d166-7811-49b5-9c16-a3434229d623/df08d166-7811-49b5-9c16-a3434229d623.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1045.900625] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-69f4d8b2-479b-4685-a874-e06fc3d02700 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.921622] env[65726]: DEBUG oslo_vmware.api [None req-10701ce3-ca7d-4146-b0df-5677afb5eba0 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] Waiting for the task: (returnval){ [ 1045.921622] env[65726]: value = "task-5116579" [ 1045.921622] env[65726]: _type = "Task" [ 1045.921622] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1045.933937] env[65726]: DEBUG oslo_vmware.api [None req-10701ce3-ca7d-4146-b0df-5677afb5eba0 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] Task: {'id': task-5116579, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.947190] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-03f877a9-8771-4f37-b985-f35105d89922 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: 9fc0c9ee-9dc7-4059-b8d2-a5db965e00c2] Unregistered the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1045.947190] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-03f877a9-8771-4f37-b985-f35105d89922 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: 9fc0c9ee-9dc7-4059-b8d2-a5db965e00c2] Deleting contents of the VM from datastore datastore2 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1045.947190] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-03f877a9-8771-4f37-b985-f35105d89922 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Deleting the datastore file [datastore2] 9fc0c9ee-9dc7-4059-b8d2-a5db965e00c2 {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1045.947190] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-fc05249a-b851-411e-84e9-cbe4067c0518 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.957953] env[65726]: DEBUG oslo_vmware.api [None req-03f877a9-8771-4f37-b985-f35105d89922 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Waiting for the task: (returnval){ [ 1045.957953] env[65726]: value = "task-5116580" [ 1045.957953] env[65726]: _type = "Task" [ 1045.957953] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1045.959464] env[65726]: DEBUG oslo_concurrency.lockutils [None req-4e9ea6a2-e7a5-42fb-bc17-3cf3a7f02eb9 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] Acquiring lock "refresh_cache-2b8eb60b-dc9b-47a4-9ddc-205f2af69888" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1045.959691] env[65726]: DEBUG oslo_concurrency.lockutils [None req-4e9ea6a2-e7a5-42fb-bc17-3cf3a7f02eb9 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] Acquired lock "refresh_cache-2b8eb60b-dc9b-47a4-9ddc-205f2af69888" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1045.959952] env[65726]: DEBUG nova.network.neutron [None req-4e9ea6a2-e7a5-42fb-bc17-3cf3a7f02eb9 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] [instance: 2b8eb60b-dc9b-47a4-9ddc-205f2af69888] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1045.974824] env[65726]: DEBUG oslo_vmware.api [None req-03f877a9-8771-4f37-b985-f35105d89922 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Task: {'id': task-5116580, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.047942] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-735cb05a-64c7-4291-9341-290c6ae58ec1 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.058783] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54d3a345-667a-4794-9aa2-b7b6826ecb7c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.095160] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a1fcf65-a30e-43c3-8638-97d05c4c59e2 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.104796] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ecc8402-d180-4a02-9adf-79d2d41428bd {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.125549] env[65726]: DEBUG nova.compute.provider_tree [None req-5c22cc7c-7dcc-417f-8f16-1f19636dd1f0 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1046.151039] env[65726]: DEBUG nova.compute.manager [None req-fe2eb30a-001a-49d1-91b1-ed3f11534bdf tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] [instance: 2c812714-dccc-4d1e-bdb5-c11e446949c3] Start destroying the instance on the hypervisor. {{(pid=65726) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1046.155072] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-fe2eb30a-001a-49d1-91b1-ed3f11534bdf tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] [instance: 2c812714-dccc-4d1e-bdb5-c11e446949c3] Destroying instance {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1046.155072] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0706b587-46e6-4aef-b7c6-300d325ac144 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.163485] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-fe2eb30a-001a-49d1-91b1-ed3f11534bdf tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] [instance: 2c812714-dccc-4d1e-bdb5-c11e446949c3] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1046.163485] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e681fca5-8237-47da-8c36-cc2fe9223a47 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.171197] env[65726]: DEBUG oslo_vmware.api [None req-fe2eb30a-001a-49d1-91b1-ed3f11534bdf tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Waiting for the task: (returnval){ [ 1046.171197] env[65726]: value = "task-5116581" [ 1046.171197] env[65726]: _type = "Task" [ 1046.171197] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1046.182491] env[65726]: DEBUG oslo_vmware.api [None req-fe2eb30a-001a-49d1-91b1-ed3f11534bdf tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Task: {'id': task-5116581, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.260355] env[65726]: DEBUG oslo_vmware.api [None req-031ff252-21d5-4d9e-b884-2b2939b62196 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5116577, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.315648] env[65726]: DEBUG nova.compute.manager [None req-cc47cd13-9595-40ad-a0c8-afcfcfe7f019 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] [instance: 48923e3d-b00d-4034-bacf-7be82009fb08] Starting instance... {{(pid=65726) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 1046.351457] env[65726]: DEBUG oslo_vmware.api [None req-0c1711f1-74cb-4ba7-a859-716c68812e37 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]522d3877-8207-1aba-41a1-bd8029df72c1, 'name': SearchDatastore_Task, 'duration_secs': 0.018011} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1046.352153] env[65726]: DEBUG oslo_concurrency.lockutils [None req-0c1711f1-74cb-4ba7-a859-716c68812e37 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1046.352653] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-0c1711f1-74cb-4ba7-a859-716c68812e37 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: ad4f00b5-3765-494f-8132-8b17b29d219a] Processing image b52362a3-ee8a-4cbf-b06f-513b0cc8f95c {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1046.353153] env[65726]: DEBUG oslo_concurrency.lockutils [None req-0c1711f1-74cb-4ba7-a859-716c68812e37 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1046.356603] env[65726]: DEBUG oslo_concurrency.lockutils [None req-0c1711f1-74cb-4ba7-a859-716c68812e37 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1046.356603] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-0c1711f1-74cb-4ba7-a859-716c68812e37 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1046.356603] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-864c4a8b-48eb-4d57-a4bd-3999c44cea5a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.367755] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-0c1711f1-74cb-4ba7-a859-716c68812e37 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1046.368452] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-0c1711f1-74cb-4ba7-a859-716c68812e37 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65726) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1046.369542] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bb72f30c-f668-4a01-b485-c8eac1670155 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.378239] env[65726]: DEBUG oslo_vmware.api [None req-0c1711f1-74cb-4ba7-a859-716c68812e37 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Waiting for the task: (returnval){ [ 1046.378239] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]525217af-3f1a-32fe-fb8c-abef407b832f" [ 1046.378239] env[65726]: _type = "Task" [ 1046.378239] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1046.390782] env[65726]: DEBUG oslo_vmware.api [None req-0c1711f1-74cb-4ba7-a859-716c68812e37 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]525217af-3f1a-32fe-fb8c-abef407b832f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.435134] env[65726]: DEBUG oslo_vmware.api [None req-10701ce3-ca7d-4146-b0df-5677afb5eba0 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] Task: {'id': task-5116579, 'name': ReconfigVM_Task, 'duration_secs': 0.343713} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1046.435463] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-10701ce3-ca7d-4146-b0df-5677afb5eba0 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] [instance: df08d166-7811-49b5-9c16-a3434229d623] Reconfigured VM instance instance-0000005a to attach disk [datastore2] df08d166-7811-49b5-9c16-a3434229d623/df08d166-7811-49b5-9c16-a3434229d623.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1046.436297] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f31b4cd5-71c6-4042-aad7-ca0a347dd14b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.444511] env[65726]: DEBUG oslo_vmware.api [None req-10701ce3-ca7d-4146-b0df-5677afb5eba0 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] Waiting for the task: (returnval){ [ 1046.444511] env[65726]: value = "task-5116582" [ 1046.444511] env[65726]: _type = "Task" [ 1046.444511] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1046.454640] env[65726]: DEBUG oslo_vmware.api [None req-10701ce3-ca7d-4146-b0df-5677afb5eba0 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] Task: {'id': task-5116582, 'name': Rename_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.471867] env[65726]: WARNING openstack [None req-4e9ea6a2-e7a5-42fb-bc17-3cf3a7f02eb9 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1046.472409] env[65726]: WARNING openstack [None req-4e9ea6a2-e7a5-42fb-bc17-3cf3a7f02eb9 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1046.479561] env[65726]: DEBUG oslo_vmware.api [None req-03f877a9-8771-4f37-b985-f35105d89922 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Task: {'id': task-5116580, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.252699} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1046.480210] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-03f877a9-8771-4f37-b985-f35105d89922 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Deleted the datastore file {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1046.480390] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-03f877a9-8771-4f37-b985-f35105d89922 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: 9fc0c9ee-9dc7-4059-b8d2-a5db965e00c2] Deleted contents of the VM from datastore datastore2 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1046.480603] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-03f877a9-8771-4f37-b985-f35105d89922 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: 9fc0c9ee-9dc7-4059-b8d2-a5db965e00c2] Instance destroyed {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1046.480784] env[65726]: INFO nova.compute.manager [None req-03f877a9-8771-4f37-b985-f35105d89922 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: 9fc0c9ee-9dc7-4059-b8d2-a5db965e00c2] Took 1.19 seconds to destroy the instance on the hypervisor. [ 1046.481082] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-03f877a9-8771-4f37-b985-f35105d89922 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1046.481305] env[65726]: DEBUG nova.compute.manager [-] [instance: 9fc0c9ee-9dc7-4059-b8d2-a5db965e00c2] Deallocating network for instance {{(pid=65726) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1046.481434] env[65726]: DEBUG nova.network.neutron [-] [instance: 9fc0c9ee-9dc7-4059-b8d2-a5db965e00c2] deallocate_for_instance() {{(pid=65726) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 1046.482078] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1046.482495] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1046.629112] env[65726]: DEBUG nova.scheduler.client.report [None req-5c22cc7c-7dcc-417f-8f16-1f19636dd1f0 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1046.684672] env[65726]: DEBUG oslo_vmware.api [None req-fe2eb30a-001a-49d1-91b1-ed3f11534bdf tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Task: {'id': task-5116581, 'name': PowerOffVM_Task, 'duration_secs': 0.379093} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1046.684918] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-fe2eb30a-001a-49d1-91b1-ed3f11534bdf tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] [instance: 2c812714-dccc-4d1e-bdb5-c11e446949c3] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1046.685097] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-fe2eb30a-001a-49d1-91b1-ed3f11534bdf tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] [instance: 2c812714-dccc-4d1e-bdb5-c11e446949c3] Unregistering the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1046.685386] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1011e120-c542-4c69-aec5-1233357e05dc {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.697448] env[65726]: DEBUG nova.compute.manager [None req-534e859d-9bab-4a75-8dfe-7304a9dfecfd tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: f947529e-c930-4420-91a7-1e677b5ac2f1] Start spawning the instance on the hypervisor. {{(pid=65726) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1046.733222] env[65726]: DEBUG nova.virt.hardware [None req-534e859d-9bab-4a75-8dfe-7304a9dfecfd tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T19:28:34Z,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1539f06fa3534e90acbc3a60c4b8bd3f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T19:28:35Z,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1046.733493] env[65726]: DEBUG nova.virt.hardware [None req-534e859d-9bab-4a75-8dfe-7304a9dfecfd tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1046.733642] env[65726]: DEBUG nova.virt.hardware [None req-534e859d-9bab-4a75-8dfe-7304a9dfecfd tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1046.733816] env[65726]: DEBUG nova.virt.hardware [None req-534e859d-9bab-4a75-8dfe-7304a9dfecfd tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1046.733954] env[65726]: DEBUG nova.virt.hardware [None req-534e859d-9bab-4a75-8dfe-7304a9dfecfd tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1046.734112] env[65726]: DEBUG nova.virt.hardware [None req-534e859d-9bab-4a75-8dfe-7304a9dfecfd tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1046.734341] env[65726]: DEBUG nova.virt.hardware [None req-534e859d-9bab-4a75-8dfe-7304a9dfecfd tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1046.734526] env[65726]: DEBUG nova.virt.hardware [None req-534e859d-9bab-4a75-8dfe-7304a9dfecfd tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1046.734694] env[65726]: DEBUG nova.virt.hardware [None req-534e859d-9bab-4a75-8dfe-7304a9dfecfd tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1046.734854] env[65726]: DEBUG nova.virt.hardware [None req-534e859d-9bab-4a75-8dfe-7304a9dfecfd tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1046.735047] env[65726]: DEBUG nova.virt.hardware [None req-534e859d-9bab-4a75-8dfe-7304a9dfecfd tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1046.735976] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7592bc7f-1893-461e-977d-8a04691c62c4 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.746422] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-daf3dd77-97ec-4022-b09d-e3760bc62ed2 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.760936] env[65726]: DEBUG oslo_vmware.api [None req-031ff252-21d5-4d9e-b884-2b2939b62196 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5116577, 'name': ReconfigVM_Task, 'duration_secs': 0.625921} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1046.772045] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-031ff252-21d5-4d9e-b884-2b2939b62196 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: a56be6fa-6027-46ee-9aa8-332fe4222d68] Reconfigured VM instance instance-00000053 to attach disk [datastore2] volume-30181d53-b2b9-4253-a8f5-28a3abf812f9/volume-30181d53-b2b9-4253-a8f5-28a3abf812f9.vmdk or device None with type thin {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1046.777638] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e2304621-1b7e-44f3-9e52-277bb3e0f56c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.795069] env[65726]: DEBUG oslo_vmware.api [None req-031ff252-21d5-4d9e-b884-2b2939b62196 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Waiting for the task: (returnval){ [ 1046.795069] env[65726]: value = "task-5116584" [ 1046.795069] env[65726]: _type = "Task" [ 1046.795069] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1046.805737] env[65726]: DEBUG oslo_vmware.api [None req-031ff252-21d5-4d9e-b884-2b2939b62196 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5116584, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.836526] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-fe2eb30a-001a-49d1-91b1-ed3f11534bdf tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] [instance: 2c812714-dccc-4d1e-bdb5-c11e446949c3] Unregistered the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1046.838104] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-fe2eb30a-001a-49d1-91b1-ed3f11534bdf tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] [instance: 2c812714-dccc-4d1e-bdb5-c11e446949c3] Deleting contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1046.838104] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-fe2eb30a-001a-49d1-91b1-ed3f11534bdf tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Deleting the datastore file [datastore1] 2c812714-dccc-4d1e-bdb5-c11e446949c3 {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1046.838104] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-407af28a-dbc5-4370-b329-c74ec2752adf {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.846158] env[65726]: DEBUG oslo_concurrency.lockutils [None req-cc47cd13-9595-40ad-a0c8-afcfcfe7f019 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1046.848137] env[65726]: DEBUG oslo_vmware.api [None req-fe2eb30a-001a-49d1-91b1-ed3f11534bdf tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Waiting for the task: (returnval){ [ 1046.848137] env[65726]: value = "task-5116585" [ 1046.848137] env[65726]: _type = "Task" [ 1046.848137] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1046.858337] env[65726]: DEBUG oslo_vmware.api [None req-fe2eb30a-001a-49d1-91b1-ed3f11534bdf tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Task: {'id': task-5116585, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.887886] env[65726]: DEBUG oslo_vmware.api [None req-0c1711f1-74cb-4ba7-a859-716c68812e37 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]525217af-3f1a-32fe-fb8c-abef407b832f, 'name': SearchDatastore_Task, 'duration_secs': 0.014131} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1046.888718] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3c3adf7c-3af7-4b6a-9251-8fc36960951d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.894861] env[65726]: DEBUG oslo_vmware.api [None req-0c1711f1-74cb-4ba7-a859-716c68812e37 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Waiting for the task: (returnval){ [ 1046.894861] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]522db038-1956-6750-b50d-a6faca0f5dbf" [ 1046.894861] env[65726]: _type = "Task" [ 1046.894861] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1046.909417] env[65726]: DEBUG oslo_vmware.api [None req-0c1711f1-74cb-4ba7-a859-716c68812e37 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]522db038-1956-6750-b50d-a6faca0f5dbf, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.926135] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1046.926135] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1046.955331] env[65726]: DEBUG oslo_vmware.api [None req-10701ce3-ca7d-4146-b0df-5677afb5eba0 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] Task: {'id': task-5116582, 'name': Rename_Task, 'duration_secs': 0.155632} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1046.955704] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-10701ce3-ca7d-4146-b0df-5677afb5eba0 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] [instance: df08d166-7811-49b5-9c16-a3434229d623] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1046.956020] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f83610e9-3062-4893-b5a6-03b34f0ce9f1 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.968673] env[65726]: DEBUG oslo_vmware.api [None req-10701ce3-ca7d-4146-b0df-5677afb5eba0 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] Waiting for the task: (returnval){ [ 1046.968673] env[65726]: value = "task-5116586" [ 1046.968673] env[65726]: _type = "Task" [ 1046.968673] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1046.981029] env[65726]: DEBUG oslo_vmware.api [None req-10701ce3-ca7d-4146-b0df-5677afb5eba0 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] Task: {'id': task-5116586, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.137158] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5c22cc7c-7dcc-417f-8f16-1f19636dd1f0 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.471s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1047.137717] env[65726]: DEBUG nova.compute.manager [None req-5c22cc7c-7dcc-417f-8f16-1f19636dd1f0 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: 51f5337d-2e23-4b7c-b590-76be6bd82411] Start building networks asynchronously for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1047.142649] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a3acab1e-0b06-47ad-b154-32b257cd5af7 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 2.979s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1047.142649] env[65726]: DEBUG nova.objects.instance [None req-a3acab1e-0b06-47ad-b154-32b257cd5af7 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: fb8eac12-aedd-426f-9dd3-6c52f9d4c71d] Trying to apply a migration context that does not seem to be set for this instance {{(pid=65726) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 1047.308406] env[65726]: DEBUG oslo_vmware.api [None req-031ff252-21d5-4d9e-b884-2b2939b62196 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5116584, 'name': ReconfigVM_Task, 'duration_secs': 0.249072} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1047.309508] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-031ff252-21d5-4d9e-b884-2b2939b62196 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: a56be6fa-6027-46ee-9aa8-332fe4222d68] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-995275', 'volume_id': '30181d53-b2b9-4253-a8f5-28a3abf812f9', 'name': 'volume-30181d53-b2b9-4253-a8f5-28a3abf812f9', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'a56be6fa-6027-46ee-9aa8-332fe4222d68', 'attached_at': '', 'detached_at': '', 'volume_id': '30181d53-b2b9-4253-a8f5-28a3abf812f9', 'serial': '30181d53-b2b9-4253-a8f5-28a3abf812f9'} {{(pid=65726) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1047.356508] env[65726]: DEBUG nova.network.neutron [None req-534e859d-9bab-4a75-8dfe-7304a9dfecfd tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: f947529e-c930-4420-91a7-1e677b5ac2f1] Successfully updated port: 4521ce71-46b3-44a7-a0d8-6172da592893 {{(pid=65726) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 1047.366862] env[65726]: DEBUG oslo_vmware.api [None req-fe2eb30a-001a-49d1-91b1-ed3f11534bdf tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Task: {'id': task-5116585, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.444063} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1047.366994] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-fe2eb30a-001a-49d1-91b1-ed3f11534bdf tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Deleted the datastore file {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1047.367241] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-fe2eb30a-001a-49d1-91b1-ed3f11534bdf tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] [instance: 2c812714-dccc-4d1e-bdb5-c11e446949c3] Deleted contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1047.367951] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-fe2eb30a-001a-49d1-91b1-ed3f11534bdf tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] [instance: 2c812714-dccc-4d1e-bdb5-c11e446949c3] Instance destroyed {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1047.368239] env[65726]: INFO nova.compute.manager [None req-fe2eb30a-001a-49d1-91b1-ed3f11534bdf tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] [instance: 2c812714-dccc-4d1e-bdb5-c11e446949c3] Took 1.22 seconds to destroy the instance on the hypervisor. [ 1047.368477] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-fe2eb30a-001a-49d1-91b1-ed3f11534bdf tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1047.369370] env[65726]: DEBUG nova.compute.manager [-] [instance: 2c812714-dccc-4d1e-bdb5-c11e446949c3] Deallocating network for instance {{(pid=65726) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1047.369498] env[65726]: DEBUG nova.network.neutron [-] [instance: 2c812714-dccc-4d1e-bdb5-c11e446949c3] deallocate_for_instance() {{(pid=65726) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 1047.370694] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1047.371357] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1047.420309] env[65726]: DEBUG oslo_vmware.api [None req-0c1711f1-74cb-4ba7-a859-716c68812e37 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]522db038-1956-6750-b50d-a6faca0f5dbf, 'name': SearchDatastore_Task, 'duration_secs': 0.019276} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1047.420309] env[65726]: DEBUG oslo_concurrency.lockutils [None req-0c1711f1-74cb-4ba7-a859-716c68812e37 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1047.420309] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-0c1711f1-74cb-4ba7-a859-716c68812e37 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore2] ad4f00b5-3765-494f-8132-8b17b29d219a/ad4f00b5-3765-494f-8132-8b17b29d219a.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1047.420309] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-68a9ceb7-905f-4293-998e-0cdad588f72e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.431757] env[65726]: DEBUG oslo_vmware.api [None req-0c1711f1-74cb-4ba7-a859-716c68812e37 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Waiting for the task: (returnval){ [ 1047.431757] env[65726]: value = "task-5116587" [ 1047.431757] env[65726]: _type = "Task" [ 1047.431757] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1047.446597] env[65726]: DEBUG oslo_vmware.api [None req-0c1711f1-74cb-4ba7-a859-716c68812e37 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Task: {'id': task-5116587, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.464386] env[65726]: DEBUG oslo_concurrency.lockutils [None req-03150333-3109-4819-9fe0-28ed6751ecba tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Acquiring lock "fb8eac12-aedd-426f-9dd3-6c52f9d4c71d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1047.464745] env[65726]: DEBUG oslo_concurrency.lockutils [None req-03150333-3109-4819-9fe0-28ed6751ecba tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Lock "fb8eac12-aedd-426f-9dd3-6c52f9d4c71d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1047.465052] env[65726]: DEBUG oslo_concurrency.lockutils [None req-03150333-3109-4819-9fe0-28ed6751ecba tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Acquiring lock "fb8eac12-aedd-426f-9dd3-6c52f9d4c71d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1047.465331] env[65726]: DEBUG oslo_concurrency.lockutils [None req-03150333-3109-4819-9fe0-28ed6751ecba tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Lock "fb8eac12-aedd-426f-9dd3-6c52f9d4c71d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1047.465599] env[65726]: DEBUG oslo_concurrency.lockutils [None req-03150333-3109-4819-9fe0-28ed6751ecba tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Lock "fb8eac12-aedd-426f-9dd3-6c52f9d4c71d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1047.469015] env[65726]: INFO nova.compute.manager [None req-03150333-3109-4819-9fe0-28ed6751ecba tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: fb8eac12-aedd-426f-9dd3-6c52f9d4c71d] Terminating instance [ 1047.482232] env[65726]: DEBUG oslo_vmware.api [None req-10701ce3-ca7d-4146-b0df-5677afb5eba0 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] Task: {'id': task-5116586, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.648953] env[65726]: DEBUG nova.compute.utils [None req-5c22cc7c-7dcc-417f-8f16-1f19636dd1f0 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Using /dev/sd instead of None {{(pid=65726) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1047.655456] env[65726]: DEBUG nova.compute.manager [None req-5c22cc7c-7dcc-417f-8f16-1f19636dd1f0 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: 51f5337d-2e23-4b7c-b590-76be6bd82411] Allocating IP information in the background. {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1047.655456] env[65726]: DEBUG nova.network.neutron [None req-5c22cc7c-7dcc-417f-8f16-1f19636dd1f0 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: 51f5337d-2e23-4b7c-b590-76be6bd82411] allocate_for_instance() {{(pid=65726) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 1047.655456] env[65726]: WARNING neutronclient.v2_0.client [None req-5c22cc7c-7dcc-417f-8f16-1f19636dd1f0 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1047.655724] env[65726]: WARNING neutronclient.v2_0.client [None req-5c22cc7c-7dcc-417f-8f16-1f19636dd1f0 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1047.656237] env[65726]: WARNING openstack [None req-5c22cc7c-7dcc-417f-8f16-1f19636dd1f0 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1047.656645] env[65726]: WARNING openstack [None req-5c22cc7c-7dcc-417f-8f16-1f19636dd1f0 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1047.664845] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a3acab1e-0b06-47ad-b154-32b257cd5af7 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 0.523s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1047.668122] env[65726]: DEBUG oslo_concurrency.lockutils [None req-17e357f7-df51-4538-bd2b-ea061ce92818 tempest-ServerMetadataTestJSON-637854735 tempest-ServerMetadataTestJSON-637854735-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.979s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1047.668355] env[65726]: DEBUG nova.objects.instance [None req-17e357f7-df51-4538-bd2b-ea061ce92818 tempest-ServerMetadataTestJSON-637854735 tempest-ServerMetadataTestJSON-637854735-project-member] Lazy-loading 'resources' on Instance uuid 699127a9-f40f-4c1c-ba4a-625af097350b {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1047.714248] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1047.714936] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1047.729393] env[65726]: WARNING openstack [None req-4e9ea6a2-e7a5-42fb-bc17-3cf3a7f02eb9 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1047.729977] env[65726]: WARNING openstack [None req-4e9ea6a2-e7a5-42fb-bc17-3cf3a7f02eb9 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1047.798891] env[65726]: DEBUG nova.policy [None req-5c22cc7c-7dcc-417f-8f16-1f19636dd1f0 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9b65b9194933469c908c7cc04478444b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd46ccb00794f458b85da4a93879139ee', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65726) authorize /opt/stack/nova/nova/policy.py:192}} [ 1047.860381] env[65726]: DEBUG oslo_concurrency.lockutils [None req-534e859d-9bab-4a75-8dfe-7304a9dfecfd tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Acquiring lock "refresh_cache-f947529e-c930-4420-91a7-1e677b5ac2f1" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1047.860558] env[65726]: DEBUG oslo_concurrency.lockutils [None req-534e859d-9bab-4a75-8dfe-7304a9dfecfd tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Acquired lock "refresh_cache-f947529e-c930-4420-91a7-1e677b5ac2f1" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1047.860729] env[65726]: DEBUG nova.network.neutron [None req-534e859d-9bab-4a75-8dfe-7304a9dfecfd tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: f947529e-c930-4420-91a7-1e677b5ac2f1] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1047.948347] env[65726]: DEBUG oslo_vmware.api [None req-0c1711f1-74cb-4ba7-a859-716c68812e37 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Task: {'id': task-5116587, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.982068] env[65726]: DEBUG nova.compute.manager [None req-03150333-3109-4819-9fe0-28ed6751ecba tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: fb8eac12-aedd-426f-9dd3-6c52f9d4c71d] Start destroying the instance on the hypervisor. {{(pid=65726) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1047.982406] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-03150333-3109-4819-9fe0-28ed6751ecba tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: fb8eac12-aedd-426f-9dd3-6c52f9d4c71d] Destroying instance {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1047.982698] env[65726]: DEBUG oslo_vmware.api [None req-10701ce3-ca7d-4146-b0df-5677afb5eba0 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] Task: {'id': task-5116586, 'name': PowerOnVM_Task, 'duration_secs': 0.572271} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1047.986029] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e6f785e-49c8-4610-958d-375ee2e97538 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.987139] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-10701ce3-ca7d-4146-b0df-5677afb5eba0 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] [instance: df08d166-7811-49b5-9c16-a3434229d623] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1047.987346] env[65726]: INFO nova.compute.manager [None req-10701ce3-ca7d-4146-b0df-5677afb5eba0 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] [instance: df08d166-7811-49b5-9c16-a3434229d623] Took 8.04 seconds to spawn the instance on the hypervisor. [ 1047.987694] env[65726]: DEBUG nova.compute.manager [None req-10701ce3-ca7d-4146-b0df-5677afb5eba0 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] [instance: df08d166-7811-49b5-9c16-a3434229d623] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1047.988610] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0eb145f-c39a-4f49-a8a0-b207c0b31e9c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.006883] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-03150333-3109-4819-9fe0-28ed6751ecba tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: fb8eac12-aedd-426f-9dd3-6c52f9d4c71d] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1048.007694] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-34ba61cd-a23b-4a64-aaaa-5b876ab8a8f4 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.020546] env[65726]: DEBUG oslo_vmware.api [None req-03150333-3109-4819-9fe0-28ed6751ecba tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Waiting for the task: (returnval){ [ 1048.020546] env[65726]: value = "task-5116588" [ 1048.020546] env[65726]: _type = "Task" [ 1048.020546] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1048.036043] env[65726]: DEBUG oslo_vmware.api [None req-03150333-3109-4819-9fe0-28ed6751ecba tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': task-5116588, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.046158] env[65726]: WARNING openstack [None req-4e9ea6a2-e7a5-42fb-bc17-3cf3a7f02eb9 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1048.046658] env[65726]: WARNING openstack [None req-4e9ea6a2-e7a5-42fb-bc17-3cf3a7f02eb9 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1048.158163] env[65726]: DEBUG nova.compute.manager [None req-5c22cc7c-7dcc-417f-8f16-1f19636dd1f0 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: 51f5337d-2e23-4b7c-b590-76be6bd82411] Start building block device mappings for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1048.172211] env[65726]: DEBUG nova.network.neutron [-] [instance: 9fc0c9ee-9dc7-4059-b8d2-a5db965e00c2] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1048.366307] env[65726]: DEBUG nova.objects.instance [None req-031ff252-21d5-4d9e-b884-2b2939b62196 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Lazy-loading 'flavor' on Instance uuid a56be6fa-6027-46ee-9aa8-332fe4222d68 {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1048.367168] env[65726]: WARNING openstack [None req-534e859d-9bab-4a75-8dfe-7304a9dfecfd tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1048.367554] env[65726]: WARNING openstack [None req-534e859d-9bab-4a75-8dfe-7304a9dfecfd tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1048.372855] env[65726]: DEBUG nova.network.neutron [None req-534e859d-9bab-4a75-8dfe-7304a9dfecfd tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: f947529e-c930-4420-91a7-1e677b5ac2f1] Instance cache missing network info. {{(pid=65726) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 1048.381084] env[65726]: DEBUG nova.network.neutron [None req-5c22cc7c-7dcc-417f-8f16-1f19636dd1f0 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: 51f5337d-2e23-4b7c-b590-76be6bd82411] Successfully created port: 7407bcd9-3202-4798-9eaa-252fb3fdb21f {{(pid=65726) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1048.431724] env[65726]: DEBUG nova.compute.manager [req-149379c7-cbd6-469b-b734-64715acd8c43 req-9f435d4a-fca3-4185-a2ae-e00a2fd06321 service nova] [instance: ad4f00b5-3765-494f-8132-8b17b29d219a] Received event network-changed-508e0342-f4cf-409d-96e4-742e92925ba6 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1048.431918] env[65726]: DEBUG nova.compute.manager [req-149379c7-cbd6-469b-b734-64715acd8c43 req-9f435d4a-fca3-4185-a2ae-e00a2fd06321 service nova] [instance: ad4f00b5-3765-494f-8132-8b17b29d219a] Refreshing instance network info cache due to event network-changed-508e0342-f4cf-409d-96e4-742e92925ba6. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 1048.432220] env[65726]: DEBUG oslo_concurrency.lockutils [req-149379c7-cbd6-469b-b734-64715acd8c43 req-9f435d4a-fca3-4185-a2ae-e00a2fd06321 service nova] Acquiring lock "refresh_cache-ad4f00b5-3765-494f-8132-8b17b29d219a" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1048.432323] env[65726]: DEBUG oslo_concurrency.lockutils [req-149379c7-cbd6-469b-b734-64715acd8c43 req-9f435d4a-fca3-4185-a2ae-e00a2fd06321 service nova] Acquired lock "refresh_cache-ad4f00b5-3765-494f-8132-8b17b29d219a" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1048.432484] env[65726]: DEBUG nova.network.neutron [req-149379c7-cbd6-469b-b734-64715acd8c43 req-9f435d4a-fca3-4185-a2ae-e00a2fd06321 service nova] [instance: ad4f00b5-3765-494f-8132-8b17b29d219a] Refreshing network info cache for port 508e0342-f4cf-409d-96e4-742e92925ba6 {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 1048.435786] env[65726]: DEBUG nova.network.neutron [None req-4e9ea6a2-e7a5-42fb-bc17-3cf3a7f02eb9 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] [instance: 2b8eb60b-dc9b-47a4-9ddc-205f2af69888] Updating instance_info_cache with network_info: [{"id": "f210c8b9-8838-4774-838c-aeb3860a63ca", "address": "fa:16:3e:5a:3b:53", "network": {"id": "b99bfe5a-6e05-45dc-81ff-61ad5e98453a", "bridge": "br-int", "label": "tempest-InstanceActionsTestJSON-314314203-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "878eef0927ed47dbb370522e93a5453f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "65e4a2b4-fd64-4ac9-b2ec-bac768b501c5", "external-id": "nsx-vlan-transportzone-449", "segmentation_id": 449, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf210c8b9-88", "ovs_interfaceid": "f210c8b9-8838-4774-838c-aeb3860a63ca", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1048.452940] env[65726]: DEBUG oslo_vmware.api [None req-0c1711f1-74cb-4ba7-a859-716c68812e37 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Task: {'id': task-5116587, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.012573} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1048.456149] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-0c1711f1-74cb-4ba7-a859-716c68812e37 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore2] ad4f00b5-3765-494f-8132-8b17b29d219a/ad4f00b5-3765-494f-8132-8b17b29d219a.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1048.457207] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-0c1711f1-74cb-4ba7-a859-716c68812e37 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: ad4f00b5-3765-494f-8132-8b17b29d219a] Extending root virtual disk to 1048576 {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1048.459343] env[65726]: WARNING openstack [None req-534e859d-9bab-4a75-8dfe-7304a9dfecfd tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1048.459819] env[65726]: WARNING openstack [None req-534e859d-9bab-4a75-8dfe-7304a9dfecfd tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1048.468342] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-dd3aecea-aab7-406c-932e-bd30021bcd00 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.480485] env[65726]: DEBUG oslo_vmware.api [None req-0c1711f1-74cb-4ba7-a859-716c68812e37 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Waiting for the task: (returnval){ [ 1048.480485] env[65726]: value = "task-5116589" [ 1048.480485] env[65726]: _type = "Task" [ 1048.480485] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1048.493630] env[65726]: DEBUG oslo_vmware.api [None req-0c1711f1-74cb-4ba7-a859-716c68812e37 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Task: {'id': task-5116589, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.522192] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea64f510-7b97-4e79-a828-a99a2723d548 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.528030] env[65726]: INFO nova.compute.manager [None req-10701ce3-ca7d-4146-b0df-5677afb5eba0 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] [instance: df08d166-7811-49b5-9c16-a3434229d623] Took 16.42 seconds to build instance. [ 1048.541302] env[65726]: DEBUG oslo_vmware.api [None req-03150333-3109-4819-9fe0-28ed6751ecba tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': task-5116588, 'name': PowerOffVM_Task, 'duration_secs': 0.415038} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1048.541716] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-03150333-3109-4819-9fe0-28ed6751ecba tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: fb8eac12-aedd-426f-9dd3-6c52f9d4c71d] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1048.541886] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-03150333-3109-4819-9fe0-28ed6751ecba tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: fb8eac12-aedd-426f-9dd3-6c52f9d4c71d] Unregistering the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1048.543616] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12530a62-268a-4347-b974-4a83763915d5 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.546710] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-869c21c9-4fe6-4f97-9da6-6bddfb94242a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.588737] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47a7d8cb-62ae-4c51-93ba-236694540c2f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.599891] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e429cd7-c553-4e2b-8cc5-cbc5185443d8 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.610887] env[65726]: WARNING openstack [None req-534e859d-9bab-4a75-8dfe-7304a9dfecfd tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1048.611040] env[65726]: WARNING openstack [None req-534e859d-9bab-4a75-8dfe-7304a9dfecfd tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1048.633334] env[65726]: DEBUG nova.compute.provider_tree [None req-17e357f7-df51-4538-bd2b-ea061ce92818 tempest-ServerMetadataTestJSON-637854735 tempest-ServerMetadataTestJSON-637854735-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1048.670085] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-03150333-3109-4819-9fe0-28ed6751ecba tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: fb8eac12-aedd-426f-9dd3-6c52f9d4c71d] Unregistered the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1048.670377] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-03150333-3109-4819-9fe0-28ed6751ecba tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: fb8eac12-aedd-426f-9dd3-6c52f9d4c71d] Deleting contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1048.671141] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-03150333-3109-4819-9fe0-28ed6751ecba tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Deleting the datastore file [datastore1] fb8eac12-aedd-426f-9dd3-6c52f9d4c71d {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1048.676294] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-92cf907b-b3ff-48f4-b55e-9297e3257a19 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.680414] env[65726]: INFO nova.compute.manager [-] [instance: 9fc0c9ee-9dc7-4059-b8d2-a5db965e00c2] Took 2.20 seconds to deallocate network for instance. [ 1048.681849] env[65726]: DEBUG nova.compute.manager [req-19e95e49-31c3-43fb-9a2f-7222f2da85ff req-a71baafb-7c70-4153-98a3-6fd10a301949 service nova] [instance: f947529e-c930-4420-91a7-1e677b5ac2f1] Received event network-vif-plugged-4521ce71-46b3-44a7-a0d8-6172da592893 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1048.682142] env[65726]: DEBUG oslo_concurrency.lockutils [req-19e95e49-31c3-43fb-9a2f-7222f2da85ff req-a71baafb-7c70-4153-98a3-6fd10a301949 service nova] Acquiring lock "f947529e-c930-4420-91a7-1e677b5ac2f1-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1048.682402] env[65726]: DEBUG oslo_concurrency.lockutils [req-19e95e49-31c3-43fb-9a2f-7222f2da85ff req-a71baafb-7c70-4153-98a3-6fd10a301949 service nova] Lock "f947529e-c930-4420-91a7-1e677b5ac2f1-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1048.682602] env[65726]: DEBUG oslo_concurrency.lockutils [req-19e95e49-31c3-43fb-9a2f-7222f2da85ff req-a71baafb-7c70-4153-98a3-6fd10a301949 service nova] Lock "f947529e-c930-4420-91a7-1e677b5ac2f1-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1048.682798] env[65726]: DEBUG nova.compute.manager [req-19e95e49-31c3-43fb-9a2f-7222f2da85ff req-a71baafb-7c70-4153-98a3-6fd10a301949 service nova] [instance: f947529e-c930-4420-91a7-1e677b5ac2f1] No waiting events found dispatching network-vif-plugged-4521ce71-46b3-44a7-a0d8-6172da592893 {{(pid=65726) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1048.683067] env[65726]: WARNING nova.compute.manager [req-19e95e49-31c3-43fb-9a2f-7222f2da85ff req-a71baafb-7c70-4153-98a3-6fd10a301949 service nova] [instance: f947529e-c930-4420-91a7-1e677b5ac2f1] Received unexpected event network-vif-plugged-4521ce71-46b3-44a7-a0d8-6172da592893 for instance with vm_state building and task_state spawning. [ 1048.692660] env[65726]: DEBUG oslo_vmware.api [None req-03150333-3109-4819-9fe0-28ed6751ecba tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Waiting for the task: (returnval){ [ 1048.692660] env[65726]: value = "task-5116591" [ 1048.692660] env[65726]: _type = "Task" [ 1048.692660] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1048.702305] env[65726]: DEBUG oslo_vmware.api [None req-03150333-3109-4819-9fe0-28ed6751ecba tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': task-5116591, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.796628] env[65726]: DEBUG nova.network.neutron [-] [instance: 2c812714-dccc-4d1e-bdb5-c11e446949c3] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1048.879418] env[65726]: DEBUG oslo_concurrency.lockutils [None req-031ff252-21d5-4d9e-b884-2b2939b62196 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Lock "a56be6fa-6027-46ee-9aa8-332fe4222d68" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 8.370s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1048.939086] env[65726]: WARNING openstack [req-149379c7-cbd6-469b-b734-64715acd8c43 req-9f435d4a-fca3-4185-a2ae-e00a2fd06321 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1048.939509] env[65726]: WARNING openstack [req-149379c7-cbd6-469b-b734-64715acd8c43 req-9f435d4a-fca3-4185-a2ae-e00a2fd06321 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1048.946614] env[65726]: DEBUG oslo_concurrency.lockutils [None req-4e9ea6a2-e7a5-42fb-bc17-3cf3a7f02eb9 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] Releasing lock "refresh_cache-2b8eb60b-dc9b-47a4-9ddc-205f2af69888" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1048.984488] env[65726]: DEBUG nova.network.neutron [None req-534e859d-9bab-4a75-8dfe-7304a9dfecfd tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: f947529e-c930-4420-91a7-1e677b5ac2f1] Updating instance_info_cache with network_info: [{"id": "4521ce71-46b3-44a7-a0d8-6172da592893", "address": "fa:16:3e:b2:98:d2", "network": {"id": "4c972d9d-db3b-4779-ac39-39d9a2d141dd", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1418913861-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cbb17a09b35c4c22ade5c4082b10bda9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "50cf0a70-948d-4611-af05-94c1483064ed", "external-id": "nsx-vlan-transportzone-536", "segmentation_id": 536, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4521ce71-46", "ovs_interfaceid": "4521ce71-46b3-44a7-a0d8-6172da592893", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1048.999593] env[65726]: DEBUG oslo_vmware.api [None req-0c1711f1-74cb-4ba7-a859-716c68812e37 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Task: {'id': task-5116589, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.073422} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1048.999593] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-0c1711f1-74cb-4ba7-a859-716c68812e37 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: ad4f00b5-3765-494f-8132-8b17b29d219a] Extended root virtual disk {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1049.000370] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06100dc7-2956-449e-8a31-f8db8c329282 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.026348] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-0c1711f1-74cb-4ba7-a859-716c68812e37 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: ad4f00b5-3765-494f-8132-8b17b29d219a] Reconfiguring VM instance instance-0000005b to attach disk [datastore2] ad4f00b5-3765-494f-8132-8b17b29d219a/ad4f00b5-3765-494f-8132-8b17b29d219a.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1049.027256] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5f93cfc8-8927-49ff-baef-a88e778b8dcc {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.060673] env[65726]: DEBUG oslo_concurrency.lockutils [None req-10701ce3-ca7d-4146-b0df-5677afb5eba0 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] Lock "df08d166-7811-49b5-9c16-a3434229d623" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.961s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1049.068540] env[65726]: DEBUG oslo_vmware.api [None req-0c1711f1-74cb-4ba7-a859-716c68812e37 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Waiting for the task: (returnval){ [ 1049.068540] env[65726]: value = "task-5116592" [ 1049.068540] env[65726]: _type = "Task" [ 1049.068540] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1049.082510] env[65726]: DEBUG oslo_vmware.api [None req-0c1711f1-74cb-4ba7-a859-716c68812e37 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Task: {'id': task-5116592, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.139749] env[65726]: DEBUG nova.scheduler.client.report [None req-17e357f7-df51-4538-bd2b-ea061ce92818 tempest-ServerMetadataTestJSON-637854735 tempest-ServerMetadataTestJSON-637854735-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1049.150779] env[65726]: WARNING openstack [req-149379c7-cbd6-469b-b734-64715acd8c43 req-9f435d4a-fca3-4185-a2ae-e00a2fd06321 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1049.150779] env[65726]: WARNING openstack [req-149379c7-cbd6-469b-b734-64715acd8c43 req-9f435d4a-fca3-4185-a2ae-e00a2fd06321 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1049.173571] env[65726]: DEBUG nova.compute.manager [None req-5c22cc7c-7dcc-417f-8f16-1f19636dd1f0 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: 51f5337d-2e23-4b7c-b590-76be6bd82411] Start spawning the instance on the hypervisor. {{(pid=65726) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1049.191759] env[65726]: DEBUG oslo_concurrency.lockutils [None req-03f877a9-8771-4f37-b985-f35105d89922 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1049.200506] env[65726]: DEBUG nova.virt.hardware [None req-5c22cc7c-7dcc-417f-8f16-1f19636dd1f0 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T19:28:34Z,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1539f06fa3534e90acbc3a60c4b8bd3f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T19:28:35Z,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1049.200928] env[65726]: DEBUG nova.virt.hardware [None req-5c22cc7c-7dcc-417f-8f16-1f19636dd1f0 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1049.201266] env[65726]: DEBUG nova.virt.hardware [None req-5c22cc7c-7dcc-417f-8f16-1f19636dd1f0 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1049.201525] env[65726]: DEBUG nova.virt.hardware [None req-5c22cc7c-7dcc-417f-8f16-1f19636dd1f0 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1049.201717] env[65726]: DEBUG nova.virt.hardware [None req-5c22cc7c-7dcc-417f-8f16-1f19636dd1f0 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1049.201909] env[65726]: DEBUG nova.virt.hardware [None req-5c22cc7c-7dcc-417f-8f16-1f19636dd1f0 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1049.202223] env[65726]: DEBUG nova.virt.hardware [None req-5c22cc7c-7dcc-417f-8f16-1f19636dd1f0 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1049.202462] env[65726]: DEBUG nova.virt.hardware [None req-5c22cc7c-7dcc-417f-8f16-1f19636dd1f0 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1049.202694] env[65726]: DEBUG nova.virt.hardware [None req-5c22cc7c-7dcc-417f-8f16-1f19636dd1f0 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1049.202864] env[65726]: DEBUG nova.virt.hardware [None req-5c22cc7c-7dcc-417f-8f16-1f19636dd1f0 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1049.203131] env[65726]: DEBUG nova.virt.hardware [None req-5c22cc7c-7dcc-417f-8f16-1f19636dd1f0 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1049.204450] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1a9552f-dcce-4e6a-b269-f7fcf7102693 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.213047] env[65726]: DEBUG oslo_vmware.api [None req-03150333-3109-4819-9fe0-28ed6751ecba tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': task-5116591, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.219057] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6349cbbf-e5fa-4751-811e-7b40b5d7a725 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.298855] env[65726]: WARNING openstack [req-149379c7-cbd6-469b-b734-64715acd8c43 req-9f435d4a-fca3-4185-a2ae-e00a2fd06321 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1049.299609] env[65726]: WARNING openstack [req-149379c7-cbd6-469b-b734-64715acd8c43 req-9f435d4a-fca3-4185-a2ae-e00a2fd06321 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1049.310021] env[65726]: INFO nova.compute.manager [-] [instance: 2c812714-dccc-4d1e-bdb5-c11e446949c3] Took 1.94 seconds to deallocate network for instance. [ 1049.453781] env[65726]: DEBUG nova.compute.manager [None req-4e9ea6a2-e7a5-42fb-bc17-3cf3a7f02eb9 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] [instance: 2b8eb60b-dc9b-47a4-9ddc-205f2af69888] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1049.456292] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-571037c9-965d-4b10-a435-754a876df619 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.495372] env[65726]: DEBUG nova.network.neutron [req-149379c7-cbd6-469b-b734-64715acd8c43 req-9f435d4a-fca3-4185-a2ae-e00a2fd06321 service nova] [instance: ad4f00b5-3765-494f-8132-8b17b29d219a] Updated VIF entry in instance network info cache for port 508e0342-f4cf-409d-96e4-742e92925ba6. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 1049.495775] env[65726]: DEBUG nova.network.neutron [req-149379c7-cbd6-469b-b734-64715acd8c43 req-9f435d4a-fca3-4185-a2ae-e00a2fd06321 service nova] [instance: ad4f00b5-3765-494f-8132-8b17b29d219a] Updating instance_info_cache with network_info: [{"id": "508e0342-f4cf-409d-96e4-742e92925ba6", "address": "fa:16:3e:89:e2:97", "network": {"id": "ad1373b5-e0ed-4ee6-a97c-af931afbab35", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-909418417-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a090d553766847e58e3231b966c92565", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5f60c972-a72d-4c5f-a250-faadfd6eafbe", "external-id": "nsx-vlan-transportzone-932", "segmentation_id": 932, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap508e0342-f4", "ovs_interfaceid": "508e0342-f4cf-409d-96e4-742e92925ba6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1049.498238] env[65726]: DEBUG oslo_concurrency.lockutils [None req-534e859d-9bab-4a75-8dfe-7304a9dfecfd tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Releasing lock "refresh_cache-f947529e-c930-4420-91a7-1e677b5ac2f1" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1049.498238] env[65726]: DEBUG nova.compute.manager [None req-534e859d-9bab-4a75-8dfe-7304a9dfecfd tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: f947529e-c930-4420-91a7-1e677b5ac2f1] Instance network_info: |[{"id": "4521ce71-46b3-44a7-a0d8-6172da592893", "address": "fa:16:3e:b2:98:d2", "network": {"id": "4c972d9d-db3b-4779-ac39-39d9a2d141dd", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1418913861-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cbb17a09b35c4c22ade5c4082b10bda9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "50cf0a70-948d-4611-af05-94c1483064ed", "external-id": "nsx-vlan-transportzone-536", "segmentation_id": 536, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4521ce71-46", "ovs_interfaceid": "4521ce71-46b3-44a7-a0d8-6172da592893", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1049.498432] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-534e859d-9bab-4a75-8dfe-7304a9dfecfd tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: f947529e-c930-4420-91a7-1e677b5ac2f1] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b2:98:d2', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '50cf0a70-948d-4611-af05-94c1483064ed', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4521ce71-46b3-44a7-a0d8-6172da592893', 'vif_model': 'vmxnet3'}] {{(pid=65726) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1049.508637] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-534e859d-9bab-4a75-8dfe-7304a9dfecfd tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1049.509620] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f947529e-c930-4420-91a7-1e677b5ac2f1] Creating VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1049.509870] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-979ebfab-4259-4b49-b18d-63582cd33f1b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.531986] env[65726]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1049.531986] env[65726]: value = "task-5116593" [ 1049.531986] env[65726]: _type = "Task" [ 1049.531986] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1049.547753] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116593, 'name': CreateVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.581630] env[65726]: DEBUG oslo_vmware.api [None req-0c1711f1-74cb-4ba7-a859-716c68812e37 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Task: {'id': task-5116592, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.645598] env[65726]: DEBUG oslo_concurrency.lockutils [None req-17e357f7-df51-4538-bd2b-ea061ce92818 tempest-ServerMetadataTestJSON-637854735 tempest-ServerMetadataTestJSON-637854735-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.977s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1049.649398] env[65726]: DEBUG oslo_concurrency.lockutils [None req-cc47cd13-9595-40ad-a0c8-afcfcfe7f019 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.802s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1049.649648] env[65726]: INFO nova.compute.claims [None req-cc47cd13-9595-40ad-a0c8-afcfcfe7f019 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] [instance: 48923e3d-b00d-4034-bacf-7be82009fb08] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1049.683689] env[65726]: INFO nova.scheduler.client.report [None req-17e357f7-df51-4538-bd2b-ea061ce92818 tempest-ServerMetadataTestJSON-637854735 tempest-ServerMetadataTestJSON-637854735-project-member] Deleted allocations for instance 699127a9-f40f-4c1c-ba4a-625af097350b [ 1049.706410] env[65726]: DEBUG oslo_vmware.api [None req-03150333-3109-4819-9fe0-28ed6751ecba tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': task-5116591, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.59367} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1049.706751] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-03150333-3109-4819-9fe0-28ed6751ecba tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Deleted the datastore file {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1049.706982] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-03150333-3109-4819-9fe0-28ed6751ecba tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: fb8eac12-aedd-426f-9dd3-6c52f9d4c71d] Deleted contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1049.707678] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-03150333-3109-4819-9fe0-28ed6751ecba tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: fb8eac12-aedd-426f-9dd3-6c52f9d4c71d] Instance destroyed {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1049.707928] env[65726]: INFO nova.compute.manager [None req-03150333-3109-4819-9fe0-28ed6751ecba tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: fb8eac12-aedd-426f-9dd3-6c52f9d4c71d] Took 1.73 seconds to destroy the instance on the hypervisor. [ 1049.708182] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-03150333-3109-4819-9fe0-28ed6751ecba tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1049.709205] env[65726]: DEBUG nova.compute.manager [-] [instance: fb8eac12-aedd-426f-9dd3-6c52f9d4c71d] Deallocating network for instance {{(pid=65726) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1049.709319] env[65726]: DEBUG nova.network.neutron [-] [instance: fb8eac12-aedd-426f-9dd3-6c52f9d4c71d] deallocate_for_instance() {{(pid=65726) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 1049.710284] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1049.710613] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1049.819264] env[65726]: DEBUG oslo_concurrency.lockutils [None req-fe2eb30a-001a-49d1-91b1-ed3f11534bdf tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1049.902866] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1049.902866] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1049.942029] env[65726]: DEBUG oslo_concurrency.lockutils [None req-42911402-8543-4c4b-aab9-86db149010d2 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Acquiring lock "5cdf1d61-b817-4986-a48c-804a2374497a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1049.942319] env[65726]: DEBUG oslo_concurrency.lockutils [None req-42911402-8543-4c4b-aab9-86db149010d2 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Lock "5cdf1d61-b817-4986-a48c-804a2374497a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.002s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1049.999564] env[65726]: DEBUG oslo_concurrency.lockutils [req-149379c7-cbd6-469b-b734-64715acd8c43 req-9f435d4a-fca3-4185-a2ae-e00a2fd06321 service nova] Releasing lock "refresh_cache-ad4f00b5-3765-494f-8132-8b17b29d219a" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1050.048594] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116593, 'name': CreateVM_Task} progress is 25%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.079531] env[65726]: DEBUG nova.network.neutron [None req-5c22cc7c-7dcc-417f-8f16-1f19636dd1f0 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: 51f5337d-2e23-4b7c-b590-76be6bd82411] Successfully updated port: 7407bcd9-3202-4798-9eaa-252fb3fdb21f {{(pid=65726) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 1050.091185] env[65726]: DEBUG oslo_vmware.api [None req-0c1711f1-74cb-4ba7-a859-716c68812e37 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Task: {'id': task-5116592, 'name': ReconfigVM_Task, 'duration_secs': 0.725752} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1050.091658] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-0c1711f1-74cb-4ba7-a859-716c68812e37 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: ad4f00b5-3765-494f-8132-8b17b29d219a] Reconfigured VM instance instance-0000005b to attach disk [datastore2] ad4f00b5-3765-494f-8132-8b17b29d219a/ad4f00b5-3765-494f-8132-8b17b29d219a.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1050.092487] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f01419f2-36a3-4407-b4d6-bf45abec41f3 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.101832] env[65726]: DEBUG oslo_vmware.api [None req-0c1711f1-74cb-4ba7-a859-716c68812e37 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Waiting for the task: (returnval){ [ 1050.101832] env[65726]: value = "task-5116594" [ 1050.101832] env[65726]: _type = "Task" [ 1050.101832] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1050.116675] env[65726]: DEBUG oslo_vmware.api [None req-0c1711f1-74cb-4ba7-a859-716c68812e37 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Task: {'id': task-5116594, 'name': Rename_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.194043] env[65726]: DEBUG oslo_concurrency.lockutils [None req-17e357f7-df51-4538-bd2b-ea061ce92818 tempest-ServerMetadataTestJSON-637854735 tempest-ServerMetadataTestJSON-637854735-project-member] Lock "699127a9-f40f-4c1c-ba4a-625af097350b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.601s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1050.447787] env[65726]: DEBUG nova.compute.manager [None req-42911402-8543-4c4b-aab9-86db149010d2 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 5cdf1d61-b817-4986-a48c-804a2374497a] Starting instance... {{(pid=65726) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 1050.483155] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94a2f578-62f8-4620-8d91-b1ea7a296497 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.494233] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-4e9ea6a2-e7a5-42fb-bc17-3cf3a7f02eb9 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] [instance: 2b8eb60b-dc9b-47a4-9ddc-205f2af69888] Doing hard reboot of VM {{(pid=65726) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1063}} [ 1050.494233] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ResetVM_Task with opID=oslo.vmware-23176ac1-86b7-490b-9651-f79387f31f79 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.502385] env[65726]: DEBUG oslo_vmware.api [None req-4e9ea6a2-e7a5-42fb-bc17-3cf3a7f02eb9 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] Waiting for the task: (returnval){ [ 1050.502385] env[65726]: value = "task-5116595" [ 1050.502385] env[65726]: _type = "Task" [ 1050.502385] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1050.513507] env[65726]: DEBUG oslo_vmware.api [None req-4e9ea6a2-e7a5-42fb-bc17-3cf3a7f02eb9 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] Task: {'id': task-5116595, 'name': ResetVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.546810] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116593, 'name': CreateVM_Task, 'duration_secs': 0.809421} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1050.547060] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f947529e-c930-4420-91a7-1e677b5ac2f1] Created VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1050.547840] env[65726]: WARNING openstack [None req-534e859d-9bab-4a75-8dfe-7304a9dfecfd tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1050.548214] env[65726]: WARNING openstack [None req-534e859d-9bab-4a75-8dfe-7304a9dfecfd tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1050.553500] env[65726]: DEBUG oslo_concurrency.lockutils [None req-534e859d-9bab-4a75-8dfe-7304a9dfecfd tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1050.553589] env[65726]: DEBUG oslo_concurrency.lockutils [None req-534e859d-9bab-4a75-8dfe-7304a9dfecfd tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1050.553936] env[65726]: DEBUG oslo_concurrency.lockutils [None req-534e859d-9bab-4a75-8dfe-7304a9dfecfd tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1050.554244] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-792c49c4-8868-438e-8f57-db8584c7e496 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.559891] env[65726]: DEBUG oslo_vmware.api [None req-534e859d-9bab-4a75-8dfe-7304a9dfecfd tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Waiting for the task: (returnval){ [ 1050.559891] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5298accc-55c9-c375-5be6-109ca317fba2" [ 1050.559891] env[65726]: _type = "Task" [ 1050.559891] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1050.570118] env[65726]: DEBUG oslo_vmware.api [None req-534e859d-9bab-4a75-8dfe-7304a9dfecfd tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5298accc-55c9-c375-5be6-109ca317fba2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.584062] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5c22cc7c-7dcc-417f-8f16-1f19636dd1f0 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Acquiring lock "refresh_cache-51f5337d-2e23-4b7c-b590-76be6bd82411" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1050.584339] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5c22cc7c-7dcc-417f-8f16-1f19636dd1f0 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Acquired lock "refresh_cache-51f5337d-2e23-4b7c-b590-76be6bd82411" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1050.586014] env[65726]: DEBUG nova.network.neutron [None req-5c22cc7c-7dcc-417f-8f16-1f19636dd1f0 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: 51f5337d-2e23-4b7c-b590-76be6bd82411] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1050.613897] env[65726]: DEBUG oslo_vmware.api [None req-0c1711f1-74cb-4ba7-a859-716c68812e37 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Task: {'id': task-5116594, 'name': Rename_Task, 'duration_secs': 0.208795} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1050.614221] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-0c1711f1-74cb-4ba7-a859-716c68812e37 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: ad4f00b5-3765-494f-8132-8b17b29d219a] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1050.614486] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-72d79c8b-1a2c-4ae3-8714-da52ab10a596 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.622395] env[65726]: DEBUG oslo_vmware.api [None req-0c1711f1-74cb-4ba7-a859-716c68812e37 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Waiting for the task: (returnval){ [ 1050.622395] env[65726]: value = "task-5116596" [ 1050.622395] env[65726]: _type = "Task" [ 1050.622395] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1050.631842] env[65726]: DEBUG oslo_vmware.api [None req-0c1711f1-74cb-4ba7-a859-716c68812e37 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Task: {'id': task-5116596, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.682701] env[65726]: DEBUG nova.network.neutron [-] [instance: fb8eac12-aedd-426f-9dd3-6c52f9d4c71d] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1050.948442] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53d17bdc-e04d-4528-9ba8-46a8908e7d5e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.961059] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a0daaad-f748-4ed9-8935-f7580714be1e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.002027] env[65726]: DEBUG oslo_concurrency.lockutils [None req-42911402-8543-4c4b-aab9-86db149010d2 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1051.002942] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5298d3fe-0196-496a-bfce-1f9dad8eeae3 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.006127] env[65726]: DEBUG oslo_concurrency.lockutils [None req-80fb9493-d54c-4f0a-a5fd-0197feeffb7b tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Acquiring lock "3b4d9461-747a-461c-b231-a0de02f0cb83" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1051.006367] env[65726]: DEBUG oslo_concurrency.lockutils [None req-80fb9493-d54c-4f0a-a5fd-0197feeffb7b tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Lock "3b4d9461-747a-461c-b231-a0de02f0cb83" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1051.022381] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bebba0ad-bbdb-4b84-82e4-059aa9182904 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.026843] env[65726]: DEBUG oslo_vmware.api [None req-4e9ea6a2-e7a5-42fb-bc17-3cf3a7f02eb9 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] Task: {'id': task-5116595, 'name': ResetVM_Task, 'duration_secs': 0.109785} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1051.028645] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-4e9ea6a2-e7a5-42fb-bc17-3cf3a7f02eb9 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] [instance: 2b8eb60b-dc9b-47a4-9ddc-205f2af69888] Did hard reboot of VM {{(pid=65726) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1067}} [ 1051.028645] env[65726]: DEBUG nova.compute.manager [None req-4e9ea6a2-e7a5-42fb-bc17-3cf3a7f02eb9 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] [instance: 2b8eb60b-dc9b-47a4-9ddc-205f2af69888] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1051.030971] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88740b1f-a723-49c6-a1ec-fd12e95937ea {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.044852] env[65726]: DEBUG nova.compute.provider_tree [None req-cc47cd13-9595-40ad-a0c8-afcfcfe7f019 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1051.080728] env[65726]: DEBUG oslo_vmware.api [None req-534e859d-9bab-4a75-8dfe-7304a9dfecfd tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5298accc-55c9-c375-5be6-109ca317fba2, 'name': SearchDatastore_Task, 'duration_secs': 0.011835} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1051.081104] env[65726]: DEBUG oslo_concurrency.lockutils [None req-534e859d-9bab-4a75-8dfe-7304a9dfecfd tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1051.081335] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-534e859d-9bab-4a75-8dfe-7304a9dfecfd tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: f947529e-c930-4420-91a7-1e677b5ac2f1] Processing image b52362a3-ee8a-4cbf-b06f-513b0cc8f95c {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1051.081566] env[65726]: DEBUG oslo_concurrency.lockutils [None req-534e859d-9bab-4a75-8dfe-7304a9dfecfd tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1051.081698] env[65726]: DEBUG oslo_concurrency.lockutils [None req-534e859d-9bab-4a75-8dfe-7304a9dfecfd tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1051.081865] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-534e859d-9bab-4a75-8dfe-7304a9dfecfd tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1051.082220] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-26ae89f4-8df1-4178-93f1-016481be8b4f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.089419] env[65726]: WARNING openstack [None req-5c22cc7c-7dcc-417f-8f16-1f19636dd1f0 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1051.089419] env[65726]: WARNING openstack [None req-5c22cc7c-7dcc-417f-8f16-1f19636dd1f0 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1051.094805] env[65726]: DEBUG nova.network.neutron [None req-5c22cc7c-7dcc-417f-8f16-1f19636dd1f0 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: 51f5337d-2e23-4b7c-b590-76be6bd82411] Instance cache missing network info. {{(pid=65726) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 1051.101032] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-534e859d-9bab-4a75-8dfe-7304a9dfecfd tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1051.101032] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-534e859d-9bab-4a75-8dfe-7304a9dfecfd tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65726) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1051.101032] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f6bfbe92-ee75-41fc-9a80-40222bac4f16 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.107018] env[65726]: DEBUG oslo_vmware.api [None req-534e859d-9bab-4a75-8dfe-7304a9dfecfd tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Waiting for the task: (returnval){ [ 1051.107018] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]529a5175-eb6c-43d6-9e59-686b777e421f" [ 1051.107018] env[65726]: _type = "Task" [ 1051.107018] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1051.118592] env[65726]: DEBUG oslo_vmware.api [None req-534e859d-9bab-4a75-8dfe-7304a9dfecfd tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]529a5175-eb6c-43d6-9e59-686b777e421f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.133594] env[65726]: DEBUG oslo_vmware.api [None req-0c1711f1-74cb-4ba7-a859-716c68812e37 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Task: {'id': task-5116596, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.157708] env[65726]: DEBUG nova.compute.manager [None req-dc5e36ff-c9dd-4078-bec5-b01919099186 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: a56be6fa-6027-46ee-9aa8-332fe4222d68] Stashing vm_state: active {{(pid=65726) _prep_resize /opt/stack/nova/nova/compute/manager.py:6193}} [ 1051.186238] env[65726]: INFO nova.compute.manager [-] [instance: fb8eac12-aedd-426f-9dd3-6c52f9d4c71d] Took 1.48 seconds to deallocate network for instance. [ 1051.202144] env[65726]: WARNING openstack [None req-5c22cc7c-7dcc-417f-8f16-1f19636dd1f0 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1051.202580] env[65726]: WARNING openstack [None req-5c22cc7c-7dcc-417f-8f16-1f19636dd1f0 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1051.350203] env[65726]: WARNING openstack [None req-5c22cc7c-7dcc-417f-8f16-1f19636dd1f0 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1051.350203] env[65726]: WARNING openstack [None req-5c22cc7c-7dcc-417f-8f16-1f19636dd1f0 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1051.502277] env[65726]: DEBUG nova.network.neutron [None req-5c22cc7c-7dcc-417f-8f16-1f19636dd1f0 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: 51f5337d-2e23-4b7c-b590-76be6bd82411] Updating instance_info_cache with network_info: [{"id": "7407bcd9-3202-4798-9eaa-252fb3fdb21f", "address": "fa:16:3e:79:d2:9d", "network": {"id": "51a66448-0f72-4266-8868-856b48be839f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-155658961-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d46ccb00794f458b85da4a93879139ee", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b1f3e6c3-5584-4852-9017-476ab8ac4946", "external-id": "nsx-vlan-transportzone-304", "segmentation_id": 304, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7407bcd9-32", "ovs_interfaceid": "7407bcd9-3202-4798-9eaa-252fb3fdb21f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1051.512627] env[65726]: DEBUG nova.compute.manager [None req-80fb9493-d54c-4f0a-a5fd-0197feeffb7b tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] [instance: 3b4d9461-747a-461c-b231-a0de02f0cb83] Starting instance... {{(pid=65726) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 1051.547137] env[65726]: DEBUG nova.scheduler.client.report [None req-cc47cd13-9595-40ad-a0c8-afcfcfe7f019 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1051.555083] env[65726]: DEBUG nova.compute.manager [req-0a4be146-49f1-48e7-8aed-7b2432d6edd4 req-51fc7ac6-17cd-479e-9cc1-77a1a4f10a3b service nova] [instance: fb8eac12-aedd-426f-9dd3-6c52f9d4c71d] Received event network-vif-deleted-839bf297-99c8-450e-9cd5-546932e1bea1 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1051.557038] env[65726]: DEBUG oslo_concurrency.lockutils [None req-4e9ea6a2-e7a5-42fb-bc17-3cf3a7f02eb9 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] Lock "2b8eb60b-dc9b-47a4-9ddc-205f2af69888" "released" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: held 6.134s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1051.597256] env[65726]: DEBUG nova.compute.manager [req-b9bc6299-3cef-4157-aa38-4a04c25c4da2 req-7ae20637-21bb-4a61-9347-b466f63404f4 service nova] [instance: f947529e-c930-4420-91a7-1e677b5ac2f1] Received event network-changed-4521ce71-46b3-44a7-a0d8-6172da592893 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1051.597448] env[65726]: DEBUG nova.compute.manager [req-b9bc6299-3cef-4157-aa38-4a04c25c4da2 req-7ae20637-21bb-4a61-9347-b466f63404f4 service nova] [instance: f947529e-c930-4420-91a7-1e677b5ac2f1] Refreshing instance network info cache due to event network-changed-4521ce71-46b3-44a7-a0d8-6172da592893. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 1051.597706] env[65726]: DEBUG oslo_concurrency.lockutils [req-b9bc6299-3cef-4157-aa38-4a04c25c4da2 req-7ae20637-21bb-4a61-9347-b466f63404f4 service nova] Acquiring lock "refresh_cache-f947529e-c930-4420-91a7-1e677b5ac2f1" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1051.597792] env[65726]: DEBUG oslo_concurrency.lockutils [req-b9bc6299-3cef-4157-aa38-4a04c25c4da2 req-7ae20637-21bb-4a61-9347-b466f63404f4 service nova] Acquired lock "refresh_cache-f947529e-c930-4420-91a7-1e677b5ac2f1" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1051.597944] env[65726]: DEBUG nova.network.neutron [req-b9bc6299-3cef-4157-aa38-4a04c25c4da2 req-7ae20637-21bb-4a61-9347-b466f63404f4 service nova] [instance: f947529e-c930-4420-91a7-1e677b5ac2f1] Refreshing network info cache for port 4521ce71-46b3-44a7-a0d8-6172da592893 {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 1051.618492] env[65726]: DEBUG oslo_vmware.api [None req-534e859d-9bab-4a75-8dfe-7304a9dfecfd tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]529a5175-eb6c-43d6-9e59-686b777e421f, 'name': SearchDatastore_Task, 'duration_secs': 0.030705} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1051.623021] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a507f288-3ece-421e-af09-d0135ddabe32 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.629950] env[65726]: DEBUG oslo_vmware.api [None req-534e859d-9bab-4a75-8dfe-7304a9dfecfd tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Waiting for the task: (returnval){ [ 1051.629950] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52e1b935-6460-933c-2299-314c200145fd" [ 1051.629950] env[65726]: _type = "Task" [ 1051.629950] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1051.640199] env[65726]: DEBUG oslo_vmware.api [None req-0c1711f1-74cb-4ba7-a859-716c68812e37 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Task: {'id': task-5116596, 'name': PowerOnVM_Task, 'duration_secs': 0.559671} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1051.640199] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-0c1711f1-74cb-4ba7-a859-716c68812e37 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: ad4f00b5-3765-494f-8132-8b17b29d219a] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1051.640199] env[65726]: INFO nova.compute.manager [None req-0c1711f1-74cb-4ba7-a859-716c68812e37 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: ad4f00b5-3765-494f-8132-8b17b29d219a] Took 9.32 seconds to spawn the instance on the hypervisor. [ 1051.642700] env[65726]: DEBUG nova.compute.manager [None req-0c1711f1-74cb-4ba7-a859-716c68812e37 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: ad4f00b5-3765-494f-8132-8b17b29d219a] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1051.642700] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a291c6d6-217a-4dae-90be-d8cf5d19ca03 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.648724] env[65726]: DEBUG oslo_vmware.api [None req-534e859d-9bab-4a75-8dfe-7304a9dfecfd tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52e1b935-6460-933c-2299-314c200145fd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.691495] env[65726]: DEBUG oslo_concurrency.lockutils [None req-dc5e36ff-c9dd-4078-bec5-b01919099186 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1051.697146] env[65726]: DEBUG oslo_concurrency.lockutils [None req-03150333-3109-4819-9fe0-28ed6751ecba tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1052.006925] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5c22cc7c-7dcc-417f-8f16-1f19636dd1f0 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Releasing lock "refresh_cache-51f5337d-2e23-4b7c-b590-76be6bd82411" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1052.007345] env[65726]: DEBUG nova.compute.manager [None req-5c22cc7c-7dcc-417f-8f16-1f19636dd1f0 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: 51f5337d-2e23-4b7c-b590-76be6bd82411] Instance network_info: |[{"id": "7407bcd9-3202-4798-9eaa-252fb3fdb21f", "address": "fa:16:3e:79:d2:9d", "network": {"id": "51a66448-0f72-4266-8868-856b48be839f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-155658961-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d46ccb00794f458b85da4a93879139ee", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b1f3e6c3-5584-4852-9017-476ab8ac4946", "external-id": "nsx-vlan-transportzone-304", "segmentation_id": 304, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7407bcd9-32", "ovs_interfaceid": "7407bcd9-3202-4798-9eaa-252fb3fdb21f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1052.007817] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-5c22cc7c-7dcc-417f-8f16-1f19636dd1f0 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: 51f5337d-2e23-4b7c-b590-76be6bd82411] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:79:d2:9d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'b1f3e6c3-5584-4852-9017-476ab8ac4946', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7407bcd9-3202-4798-9eaa-252fb3fdb21f', 'vif_model': 'vmxnet3'}] {{(pid=65726) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1052.017959] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-5c22cc7c-7dcc-417f-8f16-1f19636dd1f0 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1052.020550] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 51f5337d-2e23-4b7c-b590-76be6bd82411] Creating VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1052.022864] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-904f93b6-36d7-4532-9f68-909c0e0addc8 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.045869] env[65726]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1052.045869] env[65726]: value = "task-5116597" [ 1052.045869] env[65726]: _type = "Task" [ 1052.045869] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1052.056870] env[65726]: DEBUG oslo_concurrency.lockutils [None req-cc47cd13-9595-40ad-a0c8-afcfcfe7f019 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.408s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1052.056870] env[65726]: DEBUG nova.compute.manager [None req-cc47cd13-9595-40ad-a0c8-afcfcfe7f019 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] [instance: 48923e3d-b00d-4034-bacf-7be82009fb08] Start building networks asynchronously for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1052.059935] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116597, 'name': CreateVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1052.061112] env[65726]: DEBUG oslo_concurrency.lockutils [None req-80fb9493-d54c-4f0a-a5fd-0197feeffb7b tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1052.061415] env[65726]: DEBUG oslo_concurrency.lockutils [None req-03f877a9-8771-4f37-b985-f35105d89922 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.870s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1052.061658] env[65726]: DEBUG nova.objects.instance [None req-03f877a9-8771-4f37-b985-f35105d89922 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Lazy-loading 'resources' on Instance uuid 9fc0c9ee-9dc7-4059-b8d2-a5db965e00c2 {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1052.102038] env[65726]: WARNING openstack [req-b9bc6299-3cef-4157-aa38-4a04c25c4da2 req-7ae20637-21bb-4a61-9347-b466f63404f4 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1052.102513] env[65726]: WARNING openstack [req-b9bc6299-3cef-4157-aa38-4a04c25c4da2 req-7ae20637-21bb-4a61-9347-b466f63404f4 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1052.149767] env[65726]: DEBUG oslo_vmware.api [None req-534e859d-9bab-4a75-8dfe-7304a9dfecfd tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52e1b935-6460-933c-2299-314c200145fd, 'name': SearchDatastore_Task, 'duration_secs': 0.019074} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1052.150075] env[65726]: DEBUG oslo_concurrency.lockutils [None req-534e859d-9bab-4a75-8dfe-7304a9dfecfd tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1052.150374] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-534e859d-9bab-4a75-8dfe-7304a9dfecfd tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore2] f947529e-c930-4420-91a7-1e677b5ac2f1/f947529e-c930-4420-91a7-1e677b5ac2f1.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1052.150725] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-194f2739-caf2-4ff7-8c26-68ccfea2fd9b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.172037] env[65726]: DEBUG oslo_vmware.api [None req-534e859d-9bab-4a75-8dfe-7304a9dfecfd tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Waiting for the task: (returnval){ [ 1052.172037] env[65726]: value = "task-5116598" [ 1052.172037] env[65726]: _type = "Task" [ 1052.172037] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1052.182730] env[65726]: INFO nova.compute.manager [None req-0c1711f1-74cb-4ba7-a859-716c68812e37 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: ad4f00b5-3765-494f-8132-8b17b29d219a] Took 17.75 seconds to build instance. [ 1052.191266] env[65726]: DEBUG oslo_vmware.api [None req-534e859d-9bab-4a75-8dfe-7304a9dfecfd tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': task-5116598, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1052.291681] env[65726]: WARNING openstack [req-b9bc6299-3cef-4157-aa38-4a04c25c4da2 req-7ae20637-21bb-4a61-9347-b466f63404f4 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1052.291681] env[65726]: WARNING openstack [req-b9bc6299-3cef-4157-aa38-4a04c25c4da2 req-7ae20637-21bb-4a61-9347-b466f63404f4 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1052.431876] env[65726]: WARNING openstack [req-b9bc6299-3cef-4157-aa38-4a04c25c4da2 req-7ae20637-21bb-4a61-9347-b466f63404f4 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1052.432434] env[65726]: WARNING openstack [req-b9bc6299-3cef-4157-aa38-4a04c25c4da2 req-7ae20637-21bb-4a61-9347-b466f63404f4 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1052.471056] env[65726]: DEBUG oslo_concurrency.lockutils [None req-f3c5543b-5b17-4006-8d18-e03fa5503534 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] Acquiring lock "2b8eb60b-dc9b-47a4-9ddc-205f2af69888" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1052.471420] env[65726]: DEBUG oslo_concurrency.lockutils [None req-f3c5543b-5b17-4006-8d18-e03fa5503534 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] Lock "2b8eb60b-dc9b-47a4-9ddc-205f2af69888" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1052.471675] env[65726]: DEBUG oslo_concurrency.lockutils [None req-f3c5543b-5b17-4006-8d18-e03fa5503534 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] Acquiring lock "2b8eb60b-dc9b-47a4-9ddc-205f2af69888-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1052.471907] env[65726]: DEBUG oslo_concurrency.lockutils [None req-f3c5543b-5b17-4006-8d18-e03fa5503534 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] Lock "2b8eb60b-dc9b-47a4-9ddc-205f2af69888-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1052.472209] env[65726]: DEBUG oslo_concurrency.lockutils [None req-f3c5543b-5b17-4006-8d18-e03fa5503534 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] Lock "2b8eb60b-dc9b-47a4-9ddc-205f2af69888-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1052.475661] env[65726]: INFO nova.compute.manager [None req-f3c5543b-5b17-4006-8d18-e03fa5503534 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] [instance: 2b8eb60b-dc9b-47a4-9ddc-205f2af69888] Terminating instance [ 1052.557778] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116597, 'name': CreateVM_Task, 'duration_secs': 0.452904} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1052.558155] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 51f5337d-2e23-4b7c-b590-76be6bd82411] Created VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1052.558876] env[65726]: WARNING openstack [None req-5c22cc7c-7dcc-417f-8f16-1f19636dd1f0 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1052.559372] env[65726]: WARNING openstack [None req-5c22cc7c-7dcc-417f-8f16-1f19636dd1f0 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1052.565303] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5c22cc7c-7dcc-417f-8f16-1f19636dd1f0 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1052.565494] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5c22cc7c-7dcc-417f-8f16-1f19636dd1f0 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1052.565817] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5c22cc7c-7dcc-417f-8f16-1f19636dd1f0 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1052.571688] env[65726]: DEBUG nova.compute.utils [None req-cc47cd13-9595-40ad-a0c8-afcfcfe7f019 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Using /dev/sd instead of None {{(pid=65726) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1052.573568] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c993e565-b968-4661-a18e-bab8ac93fc73 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.576689] env[65726]: DEBUG nova.compute.manager [None req-cc47cd13-9595-40ad-a0c8-afcfcfe7f019 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] [instance: 48923e3d-b00d-4034-bacf-7be82009fb08] Allocating IP information in the background. {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1052.577080] env[65726]: DEBUG nova.network.neutron [None req-cc47cd13-9595-40ad-a0c8-afcfcfe7f019 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] [instance: 48923e3d-b00d-4034-bacf-7be82009fb08] allocate_for_instance() {{(pid=65726) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 1052.577553] env[65726]: WARNING neutronclient.v2_0.client [None req-cc47cd13-9595-40ad-a0c8-afcfcfe7f019 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1052.577974] env[65726]: WARNING neutronclient.v2_0.client [None req-cc47cd13-9595-40ad-a0c8-afcfcfe7f019 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1052.578869] env[65726]: WARNING openstack [None req-cc47cd13-9595-40ad-a0c8-afcfcfe7f019 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1052.580751] env[65726]: WARNING openstack [None req-cc47cd13-9595-40ad-a0c8-afcfcfe7f019 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1052.593317] env[65726]: DEBUG oslo_vmware.api [None req-5c22cc7c-7dcc-417f-8f16-1f19636dd1f0 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Waiting for the task: (returnval){ [ 1052.593317] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]523cbfbe-00b0-ca0b-531d-f8256ecd54ef" [ 1052.593317] env[65726]: _type = "Task" [ 1052.593317] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1052.607398] env[65726]: DEBUG oslo_vmware.api [None req-5c22cc7c-7dcc-417f-8f16-1f19636dd1f0 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]523cbfbe-00b0-ca0b-531d-f8256ecd54ef, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1052.609771] env[65726]: DEBUG nova.network.neutron [req-b9bc6299-3cef-4157-aa38-4a04c25c4da2 req-7ae20637-21bb-4a61-9347-b466f63404f4 service nova] [instance: f947529e-c930-4420-91a7-1e677b5ac2f1] Updated VIF entry in instance network info cache for port 4521ce71-46b3-44a7-a0d8-6172da592893. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 1052.610179] env[65726]: DEBUG nova.network.neutron [req-b9bc6299-3cef-4157-aa38-4a04c25c4da2 req-7ae20637-21bb-4a61-9347-b466f63404f4 service nova] [instance: f947529e-c930-4420-91a7-1e677b5ac2f1] Updating instance_info_cache with network_info: [{"id": "4521ce71-46b3-44a7-a0d8-6172da592893", "address": "fa:16:3e:b2:98:d2", "network": {"id": "4c972d9d-db3b-4779-ac39-39d9a2d141dd", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1418913861-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cbb17a09b35c4c22ade5c4082b10bda9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "50cf0a70-948d-4611-af05-94c1483064ed", "external-id": "nsx-vlan-transportzone-536", "segmentation_id": 536, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4521ce71-46", "ovs_interfaceid": "4521ce71-46b3-44a7-a0d8-6172da592893", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1052.632239] env[65726]: DEBUG nova.policy [None req-cc47cd13-9595-40ad-a0c8-afcfcfe7f019 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2614400ef47b4d20ae684c79641ec2a6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1d5cb47905144da49aaf07ddb7d263af', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65726) authorize /opt/stack/nova/nova/policy.py:192}} [ 1052.687130] env[65726]: DEBUG oslo_concurrency.lockutils [None req-0c1711f1-74cb-4ba7-a859-716c68812e37 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Lock "ad4f00b5-3765-494f-8132-8b17b29d219a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 19.268s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1052.692448] env[65726]: DEBUG oslo_vmware.api [None req-534e859d-9bab-4a75-8dfe-7304a9dfecfd tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': task-5116598, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1052.905950] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f40bbe9c-9078-4fb4-91f0-c7d4531c6e32 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.917674] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d12df024-a655-45f6-9108-3e9537514f51 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.965456] env[65726]: DEBUG nova.network.neutron [None req-cc47cd13-9595-40ad-a0c8-afcfcfe7f019 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] [instance: 48923e3d-b00d-4034-bacf-7be82009fb08] Successfully created port: 03191f4c-fd30-4ddc-abfd-75a70e8ab0d6 {{(pid=65726) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1052.969401] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-587b5fbf-f2d5-4471-a16a-b7e5c5b7f131 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.978350] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab6a7fd9-202b-43a0-9d16-9d0870ecf0a3 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.983495] env[65726]: DEBUG nova.compute.manager [None req-f3c5543b-5b17-4006-8d18-e03fa5503534 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] [instance: 2b8eb60b-dc9b-47a4-9ddc-205f2af69888] Start destroying the instance on the hypervisor. {{(pid=65726) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1052.983724] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-f3c5543b-5b17-4006-8d18-e03fa5503534 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] [instance: 2b8eb60b-dc9b-47a4-9ddc-205f2af69888] Destroying instance {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1052.984591] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ff60c92-edb1-441c-b5dd-03bc999f9107 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.003033] env[65726]: DEBUG nova.compute.provider_tree [None req-03f877a9-8771-4f37-b985-f35105d89922 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1053.005602] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-f3c5543b-5b17-4006-8d18-e03fa5503534 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] [instance: 2b8eb60b-dc9b-47a4-9ddc-205f2af69888] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1053.009027] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-cbe2abf7-50c9-4d5a-9319-87de44218ea5 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.014711] env[65726]: DEBUG oslo_vmware.api [None req-f3c5543b-5b17-4006-8d18-e03fa5503534 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] Waiting for the task: (returnval){ [ 1053.014711] env[65726]: value = "task-5116599" [ 1053.014711] env[65726]: _type = "Task" [ 1053.014711] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1053.025428] env[65726]: DEBUG oslo_vmware.api [None req-f3c5543b-5b17-4006-8d18-e03fa5503534 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] Task: {'id': task-5116599, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1053.087815] env[65726]: DEBUG nova.compute.manager [None req-cc47cd13-9595-40ad-a0c8-afcfcfe7f019 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] [instance: 48923e3d-b00d-4034-bacf-7be82009fb08] Start building block device mappings for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1053.106957] env[65726]: DEBUG oslo_vmware.api [None req-5c22cc7c-7dcc-417f-8f16-1f19636dd1f0 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]523cbfbe-00b0-ca0b-531d-f8256ecd54ef, 'name': SearchDatastore_Task, 'duration_secs': 0.06351} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1053.107377] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5c22cc7c-7dcc-417f-8f16-1f19636dd1f0 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1053.108307] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-5c22cc7c-7dcc-417f-8f16-1f19636dd1f0 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: 51f5337d-2e23-4b7c-b590-76be6bd82411] Processing image b52362a3-ee8a-4cbf-b06f-513b0cc8f95c {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1053.108307] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5c22cc7c-7dcc-417f-8f16-1f19636dd1f0 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1053.108307] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5c22cc7c-7dcc-417f-8f16-1f19636dd1f0 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1053.108539] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-5c22cc7c-7dcc-417f-8f16-1f19636dd1f0 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1053.108746] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ee7ebdc5-eb5d-4e5a-86c8-a158b02504d4 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.117180] env[65726]: DEBUG oslo_concurrency.lockutils [req-b9bc6299-3cef-4157-aa38-4a04c25c4da2 req-7ae20637-21bb-4a61-9347-b466f63404f4 service nova] Releasing lock "refresh_cache-f947529e-c930-4420-91a7-1e677b5ac2f1" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1053.117180] env[65726]: DEBUG nova.compute.manager [req-b9bc6299-3cef-4157-aa38-4a04c25c4da2 req-7ae20637-21bb-4a61-9347-b466f63404f4 service nova] [instance: 9fc0c9ee-9dc7-4059-b8d2-a5db965e00c2] Received event network-vif-deleted-e4315658-0732-463a-b063-b1c4cf7cdafd {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1053.117180] env[65726]: DEBUG nova.compute.manager [req-b9bc6299-3cef-4157-aa38-4a04c25c4da2 req-7ae20637-21bb-4a61-9347-b466f63404f4 service nova] [instance: 2c812714-dccc-4d1e-bdb5-c11e446949c3] Received event network-vif-deleted-8addcb43-85bf-481e-92bc-5bbdf48c0ac4 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1053.117180] env[65726]: DEBUG nova.compute.manager [req-b9bc6299-3cef-4157-aa38-4a04c25c4da2 req-7ae20637-21bb-4a61-9347-b466f63404f4 service nova] [instance: 2c812714-dccc-4d1e-bdb5-c11e446949c3] Received event network-vif-deleted-203fa726-d08f-47a8-a7cd-389ed31da940 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1053.117180] env[65726]: DEBUG nova.compute.manager [req-b9bc6299-3cef-4157-aa38-4a04c25c4da2 req-7ae20637-21bb-4a61-9347-b466f63404f4 service nova] [instance: 51f5337d-2e23-4b7c-b590-76be6bd82411] Received event network-vif-plugged-7407bcd9-3202-4798-9eaa-252fb3fdb21f {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1053.117180] env[65726]: DEBUG oslo_concurrency.lockutils [req-b9bc6299-3cef-4157-aa38-4a04c25c4da2 req-7ae20637-21bb-4a61-9347-b466f63404f4 service nova] Acquiring lock "51f5337d-2e23-4b7c-b590-76be6bd82411-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1053.117180] env[65726]: DEBUG oslo_concurrency.lockutils [req-b9bc6299-3cef-4157-aa38-4a04c25c4da2 req-7ae20637-21bb-4a61-9347-b466f63404f4 service nova] Lock "51f5337d-2e23-4b7c-b590-76be6bd82411-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1053.117383] env[65726]: DEBUG oslo_concurrency.lockutils [req-b9bc6299-3cef-4157-aa38-4a04c25c4da2 req-7ae20637-21bb-4a61-9347-b466f63404f4 service nova] Lock "51f5337d-2e23-4b7c-b590-76be6bd82411-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1053.118209] env[65726]: DEBUG nova.compute.manager [req-b9bc6299-3cef-4157-aa38-4a04c25c4da2 req-7ae20637-21bb-4a61-9347-b466f63404f4 service nova] [instance: 51f5337d-2e23-4b7c-b590-76be6bd82411] No waiting events found dispatching network-vif-plugged-7407bcd9-3202-4798-9eaa-252fb3fdb21f {{(pid=65726) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1053.118209] env[65726]: WARNING nova.compute.manager [req-b9bc6299-3cef-4157-aa38-4a04c25c4da2 req-7ae20637-21bb-4a61-9347-b466f63404f4 service nova] [instance: 51f5337d-2e23-4b7c-b590-76be6bd82411] Received unexpected event network-vif-plugged-7407bcd9-3202-4798-9eaa-252fb3fdb21f for instance with vm_state building and task_state spawning. [ 1053.118209] env[65726]: DEBUG nova.compute.manager [req-b9bc6299-3cef-4157-aa38-4a04c25c4da2 req-7ae20637-21bb-4a61-9347-b466f63404f4 service nova] [instance: 51f5337d-2e23-4b7c-b590-76be6bd82411] Received event network-changed-7407bcd9-3202-4798-9eaa-252fb3fdb21f {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1053.118209] env[65726]: DEBUG nova.compute.manager [req-b9bc6299-3cef-4157-aa38-4a04c25c4da2 req-7ae20637-21bb-4a61-9347-b466f63404f4 service nova] [instance: 51f5337d-2e23-4b7c-b590-76be6bd82411] Refreshing instance network info cache due to event network-changed-7407bcd9-3202-4798-9eaa-252fb3fdb21f. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 1053.118209] env[65726]: DEBUG oslo_concurrency.lockutils [req-b9bc6299-3cef-4157-aa38-4a04c25c4da2 req-7ae20637-21bb-4a61-9347-b466f63404f4 service nova] Acquiring lock "refresh_cache-51f5337d-2e23-4b7c-b590-76be6bd82411" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1053.118209] env[65726]: DEBUG oslo_concurrency.lockutils [req-b9bc6299-3cef-4157-aa38-4a04c25c4da2 req-7ae20637-21bb-4a61-9347-b466f63404f4 service nova] Acquired lock "refresh_cache-51f5337d-2e23-4b7c-b590-76be6bd82411" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1053.118419] env[65726]: DEBUG nova.network.neutron [req-b9bc6299-3cef-4157-aa38-4a04c25c4da2 req-7ae20637-21bb-4a61-9347-b466f63404f4 service nova] [instance: 51f5337d-2e23-4b7c-b590-76be6bd82411] Refreshing network info cache for port 7407bcd9-3202-4798-9eaa-252fb3fdb21f {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 1053.125672] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-5c22cc7c-7dcc-417f-8f16-1f19636dd1f0 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1053.126668] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-5c22cc7c-7dcc-417f-8f16-1f19636dd1f0 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65726) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1053.126735] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-573dc150-521c-4a37-b41f-f1d4f1cd5c14 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.134876] env[65726]: DEBUG oslo_vmware.api [None req-5c22cc7c-7dcc-417f-8f16-1f19636dd1f0 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Waiting for the task: (returnval){ [ 1053.134876] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52db8b3f-d322-afa5-5c39-8d513a486eb9" [ 1053.134876] env[65726]: _type = "Task" [ 1053.134876] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1053.144117] env[65726]: DEBUG oslo_vmware.api [None req-5c22cc7c-7dcc-417f-8f16-1f19636dd1f0 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52db8b3f-d322-afa5-5c39-8d513a486eb9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1053.189362] env[65726]: DEBUG oslo_vmware.api [None req-534e859d-9bab-4a75-8dfe-7304a9dfecfd tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': task-5116598, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.56838} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1053.189648] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-534e859d-9bab-4a75-8dfe-7304a9dfecfd tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore2] f947529e-c930-4420-91a7-1e677b5ac2f1/f947529e-c930-4420-91a7-1e677b5ac2f1.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1053.189858] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-534e859d-9bab-4a75-8dfe-7304a9dfecfd tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: f947529e-c930-4420-91a7-1e677b5ac2f1] Extending root virtual disk to 1048576 {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1053.190136] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-9b054939-d6c4-4c45-9c12-4fb92405094b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.198889] env[65726]: DEBUG oslo_vmware.api [None req-534e859d-9bab-4a75-8dfe-7304a9dfecfd tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Waiting for the task: (returnval){ [ 1053.198889] env[65726]: value = "task-5116600" [ 1053.198889] env[65726]: _type = "Task" [ 1053.198889] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1053.207651] env[65726]: DEBUG oslo_vmware.api [None req-534e859d-9bab-4a75-8dfe-7304a9dfecfd tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': task-5116600, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1053.507588] env[65726]: DEBUG nova.scheduler.client.report [None req-03f877a9-8771-4f37-b985-f35105d89922 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1053.525737] env[65726]: DEBUG oslo_vmware.api [None req-f3c5543b-5b17-4006-8d18-e03fa5503534 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] Task: {'id': task-5116599, 'name': PowerOffVM_Task, 'duration_secs': 0.397888} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1053.526631] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-f3c5543b-5b17-4006-8d18-e03fa5503534 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] [instance: 2b8eb60b-dc9b-47a4-9ddc-205f2af69888] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1053.526804] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-f3c5543b-5b17-4006-8d18-e03fa5503534 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] [instance: 2b8eb60b-dc9b-47a4-9ddc-205f2af69888] Unregistering the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1053.527080] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-037a08a9-0ea8-42e5-9812-66a268e674e1 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.590630] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-f3c5543b-5b17-4006-8d18-e03fa5503534 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] [instance: 2b8eb60b-dc9b-47a4-9ddc-205f2af69888] Unregistered the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1053.591116] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-f3c5543b-5b17-4006-8d18-e03fa5503534 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] [instance: 2b8eb60b-dc9b-47a4-9ddc-205f2af69888] Deleting contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1053.591223] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-f3c5543b-5b17-4006-8d18-e03fa5503534 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] Deleting the datastore file [datastore1] 2b8eb60b-dc9b-47a4-9ddc-205f2af69888 {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1053.591635] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-50528fdc-60bc-430f-8a66-99a020f9e70c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.604752] env[65726]: DEBUG oslo_vmware.api [None req-f3c5543b-5b17-4006-8d18-e03fa5503534 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] Waiting for the task: (returnval){ [ 1053.604752] env[65726]: value = "task-5116602" [ 1053.604752] env[65726]: _type = "Task" [ 1053.604752] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1053.613539] env[65726]: DEBUG oslo_vmware.api [None req-f3c5543b-5b17-4006-8d18-e03fa5503534 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] Task: {'id': task-5116602, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1053.623596] env[65726]: WARNING openstack [req-b9bc6299-3cef-4157-aa38-4a04c25c4da2 req-7ae20637-21bb-4a61-9347-b466f63404f4 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1053.624075] env[65726]: WARNING openstack [req-b9bc6299-3cef-4157-aa38-4a04c25c4da2 req-7ae20637-21bb-4a61-9347-b466f63404f4 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1053.646250] env[65726]: DEBUG oslo_vmware.api [None req-5c22cc7c-7dcc-417f-8f16-1f19636dd1f0 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52db8b3f-d322-afa5-5c39-8d513a486eb9, 'name': SearchDatastore_Task, 'duration_secs': 0.05996} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1053.647076] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-17a3ef07-4499-4519-98d3-1fd707398461 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.653987] env[65726]: DEBUG oslo_vmware.api [None req-5c22cc7c-7dcc-417f-8f16-1f19636dd1f0 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Waiting for the task: (returnval){ [ 1053.653987] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5299b863-b914-b3cf-21e3-5f23b48fa413" [ 1053.653987] env[65726]: _type = "Task" [ 1053.653987] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1053.664023] env[65726]: DEBUG oslo_vmware.api [None req-5c22cc7c-7dcc-417f-8f16-1f19636dd1f0 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5299b863-b914-b3cf-21e3-5f23b48fa413, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1053.710550] env[65726]: DEBUG oslo_vmware.api [None req-534e859d-9bab-4a75-8dfe-7304a9dfecfd tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': task-5116600, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.073309} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1053.710951] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-534e859d-9bab-4a75-8dfe-7304a9dfecfd tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: f947529e-c930-4420-91a7-1e677b5ac2f1] Extended root virtual disk {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1053.711856] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6102134c-40b3-4e26-bb17-6751c26e9a59 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.737892] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-534e859d-9bab-4a75-8dfe-7304a9dfecfd tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: f947529e-c930-4420-91a7-1e677b5ac2f1] Reconfiguring VM instance instance-0000005c to attach disk [datastore2] f947529e-c930-4420-91a7-1e677b5ac2f1/f947529e-c930-4420-91a7-1e677b5ac2f1.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1053.738305] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f962d137-80d9-4070-9e7d-969d68c8f43f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.760513] env[65726]: DEBUG oslo_vmware.api [None req-534e859d-9bab-4a75-8dfe-7304a9dfecfd tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Waiting for the task: (returnval){ [ 1053.760513] env[65726]: value = "task-5116603" [ 1053.760513] env[65726]: _type = "Task" [ 1053.760513] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1053.769611] env[65726]: DEBUG oslo_vmware.api [None req-534e859d-9bab-4a75-8dfe-7304a9dfecfd tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': task-5116603, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1053.837351] env[65726]: DEBUG nova.compute.manager [req-e33d75da-e766-4a75-8533-638c91c379d6 req-2c28f547-af7b-4318-9aa2-687434625b9d service nova] [instance: df08d166-7811-49b5-9c16-a3434229d623] Received event network-changed-70efdc7d-7775-4990-b851-d60195bd1504 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1053.837561] env[65726]: DEBUG nova.compute.manager [req-e33d75da-e766-4a75-8533-638c91c379d6 req-2c28f547-af7b-4318-9aa2-687434625b9d service nova] [instance: df08d166-7811-49b5-9c16-a3434229d623] Refreshing instance network info cache due to event network-changed-70efdc7d-7775-4990-b851-d60195bd1504. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 1053.837789] env[65726]: DEBUG oslo_concurrency.lockutils [req-e33d75da-e766-4a75-8533-638c91c379d6 req-2c28f547-af7b-4318-9aa2-687434625b9d service nova] Acquiring lock "refresh_cache-df08d166-7811-49b5-9c16-a3434229d623" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1053.837937] env[65726]: DEBUG oslo_concurrency.lockutils [req-e33d75da-e766-4a75-8533-638c91c379d6 req-2c28f547-af7b-4318-9aa2-687434625b9d service nova] Acquired lock "refresh_cache-df08d166-7811-49b5-9c16-a3434229d623" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1053.838145] env[65726]: DEBUG nova.network.neutron [req-e33d75da-e766-4a75-8533-638c91c379d6 req-2c28f547-af7b-4318-9aa2-687434625b9d service nova] [instance: df08d166-7811-49b5-9c16-a3434229d623] Refreshing network info cache for port 70efdc7d-7775-4990-b851-d60195bd1504 {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 1053.881149] env[65726]: WARNING openstack [req-b9bc6299-3cef-4157-aa38-4a04c25c4da2 req-7ae20637-21bb-4a61-9347-b466f63404f4 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1053.881573] env[65726]: WARNING openstack [req-b9bc6299-3cef-4157-aa38-4a04c25c4da2 req-7ae20637-21bb-4a61-9347-b466f63404f4 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1054.011059] env[65726]: WARNING openstack [req-b9bc6299-3cef-4157-aa38-4a04c25c4da2 req-7ae20637-21bb-4a61-9347-b466f63404f4 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1054.011477] env[65726]: WARNING openstack [req-b9bc6299-3cef-4157-aa38-4a04c25c4da2 req-7ae20637-21bb-4a61-9347-b466f63404f4 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1054.026670] env[65726]: DEBUG oslo_concurrency.lockutils [None req-03f877a9-8771-4f37-b985-f35105d89922 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.965s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1054.031457] env[65726]: DEBUG oslo_concurrency.lockutils [None req-fe2eb30a-001a-49d1-91b1-ed3f11534bdf tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.212s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1054.031739] env[65726]: DEBUG nova.objects.instance [None req-fe2eb30a-001a-49d1-91b1-ed3f11534bdf tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Lazy-loading 'resources' on Instance uuid 2c812714-dccc-4d1e-bdb5-c11e446949c3 {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1054.058849] env[65726]: INFO nova.scheduler.client.report [None req-03f877a9-8771-4f37-b985-f35105d89922 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Deleted allocations for instance 9fc0c9ee-9dc7-4059-b8d2-a5db965e00c2 [ 1054.101332] env[65726]: DEBUG nova.compute.manager [None req-cc47cd13-9595-40ad-a0c8-afcfcfe7f019 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] [instance: 48923e3d-b00d-4034-bacf-7be82009fb08] Start spawning the instance on the hypervisor. {{(pid=65726) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1054.117844] env[65726]: DEBUG oslo_vmware.api [None req-f3c5543b-5b17-4006-8d18-e03fa5503534 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] Task: {'id': task-5116602, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.154768} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1054.123507] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-f3c5543b-5b17-4006-8d18-e03fa5503534 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] Deleted the datastore file {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1054.123704] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-f3c5543b-5b17-4006-8d18-e03fa5503534 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] [instance: 2b8eb60b-dc9b-47a4-9ddc-205f2af69888] Deleted contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1054.123878] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-f3c5543b-5b17-4006-8d18-e03fa5503534 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] [instance: 2b8eb60b-dc9b-47a4-9ddc-205f2af69888] Instance destroyed {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1054.124059] env[65726]: INFO nova.compute.manager [None req-f3c5543b-5b17-4006-8d18-e03fa5503534 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] [instance: 2b8eb60b-dc9b-47a4-9ddc-205f2af69888] Took 1.14 seconds to destroy the instance on the hypervisor. [ 1054.124307] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-f3c5543b-5b17-4006-8d18-e03fa5503534 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1054.126880] env[65726]: DEBUG nova.compute.manager [-] [instance: 2b8eb60b-dc9b-47a4-9ddc-205f2af69888] Deallocating network for instance {{(pid=65726) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1054.127042] env[65726]: DEBUG nova.network.neutron [-] [instance: 2b8eb60b-dc9b-47a4-9ddc-205f2af69888] deallocate_for_instance() {{(pid=65726) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 1054.127685] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1054.127943] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1054.144150] env[65726]: DEBUG nova.virt.hardware [None req-cc47cd13-9595-40ad-a0c8-afcfcfe7f019 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T19:28:34Z,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1539f06fa3534e90acbc3a60c4b8bd3f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T19:28:35Z,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1054.144477] env[65726]: DEBUG nova.virt.hardware [None req-cc47cd13-9595-40ad-a0c8-afcfcfe7f019 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1054.144566] env[65726]: DEBUG nova.virt.hardware [None req-cc47cd13-9595-40ad-a0c8-afcfcfe7f019 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1054.144743] env[65726]: DEBUG nova.virt.hardware [None req-cc47cd13-9595-40ad-a0c8-afcfcfe7f019 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1054.144922] env[65726]: DEBUG nova.virt.hardware [None req-cc47cd13-9595-40ad-a0c8-afcfcfe7f019 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1054.145107] env[65726]: DEBUG nova.virt.hardware [None req-cc47cd13-9595-40ad-a0c8-afcfcfe7f019 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1054.145323] env[65726]: DEBUG nova.virt.hardware [None req-cc47cd13-9595-40ad-a0c8-afcfcfe7f019 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1054.145480] env[65726]: DEBUG nova.virt.hardware [None req-cc47cd13-9595-40ad-a0c8-afcfcfe7f019 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1054.145714] env[65726]: DEBUG nova.virt.hardware [None req-cc47cd13-9595-40ad-a0c8-afcfcfe7f019 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1054.145898] env[65726]: DEBUG nova.virt.hardware [None req-cc47cd13-9595-40ad-a0c8-afcfcfe7f019 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1054.146092] env[65726]: DEBUG nova.virt.hardware [None req-cc47cd13-9595-40ad-a0c8-afcfcfe7f019 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1054.147613] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54d87f16-0b4c-4cc9-9f78-cf81052c7279 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.162153] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-754aca18-1679-4b27-aac6-ea8002345707 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.185077] env[65726]: DEBUG oslo_vmware.api [None req-5c22cc7c-7dcc-417f-8f16-1f19636dd1f0 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5299b863-b914-b3cf-21e3-5f23b48fa413, 'name': SearchDatastore_Task, 'duration_secs': 0.010748} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1054.185077] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5c22cc7c-7dcc-417f-8f16-1f19636dd1f0 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1054.185077] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-5c22cc7c-7dcc-417f-8f16-1f19636dd1f0 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore2] 51f5337d-2e23-4b7c-b590-76be6bd82411/51f5337d-2e23-4b7c-b590-76be6bd82411.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1054.185368] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f51dae3f-08de-47d4-9c31-c2f25e41d337 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.195031] env[65726]: DEBUG oslo_vmware.api [None req-5c22cc7c-7dcc-417f-8f16-1f19636dd1f0 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Waiting for the task: (returnval){ [ 1054.195031] env[65726]: value = "task-5116604" [ 1054.195031] env[65726]: _type = "Task" [ 1054.195031] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1054.200769] env[65726]: DEBUG nova.network.neutron [req-b9bc6299-3cef-4157-aa38-4a04c25c4da2 req-7ae20637-21bb-4a61-9347-b466f63404f4 service nova] [instance: 51f5337d-2e23-4b7c-b590-76be6bd82411] Updated VIF entry in instance network info cache for port 7407bcd9-3202-4798-9eaa-252fb3fdb21f. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 1054.201132] env[65726]: DEBUG nova.network.neutron [req-b9bc6299-3cef-4157-aa38-4a04c25c4da2 req-7ae20637-21bb-4a61-9347-b466f63404f4 service nova] [instance: 51f5337d-2e23-4b7c-b590-76be6bd82411] Updating instance_info_cache with network_info: [{"id": "7407bcd9-3202-4798-9eaa-252fb3fdb21f", "address": "fa:16:3e:79:d2:9d", "network": {"id": "51a66448-0f72-4266-8868-856b48be839f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-155658961-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d46ccb00794f458b85da4a93879139ee", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b1f3e6c3-5584-4852-9017-476ab8ac4946", "external-id": "nsx-vlan-transportzone-304", "segmentation_id": 304, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7407bcd9-32", "ovs_interfaceid": "7407bcd9-3202-4798-9eaa-252fb3fdb21f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1054.210343] env[65726]: DEBUG oslo_vmware.api [None req-5c22cc7c-7dcc-417f-8f16-1f19636dd1f0 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': task-5116604, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1054.225331] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1054.225670] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1054.272723] env[65726]: DEBUG oslo_vmware.api [None req-534e859d-9bab-4a75-8dfe-7304a9dfecfd tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': task-5116603, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1054.344043] env[65726]: WARNING openstack [req-e33d75da-e766-4a75-8533-638c91c379d6 req-2c28f547-af7b-4318-9aa2-687434625b9d service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1054.344043] env[65726]: WARNING openstack [req-e33d75da-e766-4a75-8533-638c91c379d6 req-2c28f547-af7b-4318-9aa2-687434625b9d service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1054.567070] env[65726]: DEBUG nova.network.neutron [None req-cc47cd13-9595-40ad-a0c8-afcfcfe7f019 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] [instance: 48923e3d-b00d-4034-bacf-7be82009fb08] Successfully updated port: 03191f4c-fd30-4ddc-abfd-75a70e8ab0d6 {{(pid=65726) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 1054.579872] env[65726]: DEBUG oslo_concurrency.lockutils [None req-03f877a9-8771-4f37-b985-f35105d89922 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Lock "9fc0c9ee-9dc7-4059-b8d2-a5db965e00c2" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.801s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1054.703848] env[65726]: DEBUG oslo_concurrency.lockutils [req-b9bc6299-3cef-4157-aa38-4a04c25c4da2 req-7ae20637-21bb-4a61-9347-b466f63404f4 service nova] Releasing lock "refresh_cache-51f5337d-2e23-4b7c-b590-76be6bd82411" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1054.710765] env[65726]: DEBUG oslo_vmware.api [None req-5c22cc7c-7dcc-417f-8f16-1f19636dd1f0 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': task-5116604, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1054.716725] env[65726]: DEBUG nova.compute.manager [req-14e3a7eb-6b88-4929-9e3d-16e37cd5335b req-93bc39ac-d31a-4d65-852b-c3b651ae620c service nova] [instance: 48923e3d-b00d-4034-bacf-7be82009fb08] Received event network-vif-plugged-03191f4c-fd30-4ddc-abfd-75a70e8ab0d6 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1054.716994] env[65726]: DEBUG oslo_concurrency.lockutils [req-14e3a7eb-6b88-4929-9e3d-16e37cd5335b req-93bc39ac-d31a-4d65-852b-c3b651ae620c service nova] Acquiring lock "48923e3d-b00d-4034-bacf-7be82009fb08-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1054.719901] env[65726]: DEBUG oslo_concurrency.lockutils [req-14e3a7eb-6b88-4929-9e3d-16e37cd5335b req-93bc39ac-d31a-4d65-852b-c3b651ae620c service nova] Lock "48923e3d-b00d-4034-bacf-7be82009fb08-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1054.720327] env[65726]: DEBUG oslo_concurrency.lockutils [req-14e3a7eb-6b88-4929-9e3d-16e37cd5335b req-93bc39ac-d31a-4d65-852b-c3b651ae620c service nova] Lock "48923e3d-b00d-4034-bacf-7be82009fb08-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.003s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1054.720541] env[65726]: DEBUG nova.compute.manager [req-14e3a7eb-6b88-4929-9e3d-16e37cd5335b req-93bc39ac-d31a-4d65-852b-c3b651ae620c service nova] [instance: 48923e3d-b00d-4034-bacf-7be82009fb08] No waiting events found dispatching network-vif-plugged-03191f4c-fd30-4ddc-abfd-75a70e8ab0d6 {{(pid=65726) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1054.720716] env[65726]: WARNING nova.compute.manager [req-14e3a7eb-6b88-4929-9e3d-16e37cd5335b req-93bc39ac-d31a-4d65-852b-c3b651ae620c service nova] [instance: 48923e3d-b00d-4034-bacf-7be82009fb08] Received unexpected event network-vif-plugged-03191f4c-fd30-4ddc-abfd-75a70e8ab0d6 for instance with vm_state building and task_state spawning. [ 1054.735824] env[65726]: WARNING openstack [req-e33d75da-e766-4a75-8533-638c91c379d6 req-2c28f547-af7b-4318-9aa2-687434625b9d service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1054.737130] env[65726]: WARNING openstack [req-e33d75da-e766-4a75-8533-638c91c379d6 req-2c28f547-af7b-4318-9aa2-687434625b9d service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1054.781526] env[65726]: DEBUG oslo_vmware.api [None req-534e859d-9bab-4a75-8dfe-7304a9dfecfd tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': task-5116603, 'name': ReconfigVM_Task, 'duration_secs': 0.531701} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1054.781682] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-534e859d-9bab-4a75-8dfe-7304a9dfecfd tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: f947529e-c930-4420-91a7-1e677b5ac2f1] Reconfigured VM instance instance-0000005c to attach disk [datastore2] f947529e-c930-4420-91a7-1e677b5ac2f1/f947529e-c930-4420-91a7-1e677b5ac2f1.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1054.784902] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0743bad4-5173-450f-80f7-09b66dc01a74 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.793143] env[65726]: DEBUG oslo_vmware.api [None req-534e859d-9bab-4a75-8dfe-7304a9dfecfd tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Waiting for the task: (returnval){ [ 1054.793143] env[65726]: value = "task-5116605" [ 1054.793143] env[65726]: _type = "Task" [ 1054.793143] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1054.804584] env[65726]: DEBUG oslo_vmware.api [None req-534e859d-9bab-4a75-8dfe-7304a9dfecfd tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': task-5116605, 'name': Rename_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1054.818302] env[65726]: WARNING openstack [req-e33d75da-e766-4a75-8533-638c91c379d6 req-2c28f547-af7b-4318-9aa2-687434625b9d service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1054.818828] env[65726]: WARNING openstack [req-e33d75da-e766-4a75-8533-638c91c379d6 req-2c28f547-af7b-4318-9aa2-687434625b9d service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1054.885065] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c16f72cb-ab67-44a0-b666-c7fd1cd12ac1 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.893858] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66a2a7a9-e7da-4c7b-840e-825cc6cb9438 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.928599] env[65726]: DEBUG nova.network.neutron [req-e33d75da-e766-4a75-8533-638c91c379d6 req-2c28f547-af7b-4318-9aa2-687434625b9d service nova] [instance: df08d166-7811-49b5-9c16-a3434229d623] Updated VIF entry in instance network info cache for port 70efdc7d-7775-4990-b851-d60195bd1504. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 1054.928956] env[65726]: DEBUG nova.network.neutron [req-e33d75da-e766-4a75-8533-638c91c379d6 req-2c28f547-af7b-4318-9aa2-687434625b9d service nova] [instance: df08d166-7811-49b5-9c16-a3434229d623] Updating instance_info_cache with network_info: [{"id": "70efdc7d-7775-4990-b851-d60195bd1504", "address": "fa:16:3e:be:5b:19", "network": {"id": "835ec128-d214-4f54-ab53-88f8378b7d74", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-1935476539-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.151", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "305ecace479440ad9a616e546ccd38ce", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3b107fab-ee71-47db-ad4d-3c6f05546843", "external-id": "cl2-zone-554", "segmentation_id": 554, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap70efdc7d-77", "ovs_interfaceid": "70efdc7d-7775-4990-b851-d60195bd1504", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1054.930705] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-501e6251-c753-4bab-a01c-ee1283848a29 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.939884] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01581e43-fcc1-4ced-b2e1-3dcc1752477e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.955572] env[65726]: DEBUG nova.compute.provider_tree [None req-fe2eb30a-001a-49d1-91b1-ed3f11534bdf tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1055.007197] env[65726]: DEBUG nova.network.neutron [-] [instance: 2b8eb60b-dc9b-47a4-9ddc-205f2af69888] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1055.072428] env[65726]: DEBUG oslo_concurrency.lockutils [None req-cc47cd13-9595-40ad-a0c8-afcfcfe7f019 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Acquiring lock "refresh_cache-48923e3d-b00d-4034-bacf-7be82009fb08" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1055.072428] env[65726]: DEBUG oslo_concurrency.lockutils [None req-cc47cd13-9595-40ad-a0c8-afcfcfe7f019 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Acquired lock "refresh_cache-48923e3d-b00d-4034-bacf-7be82009fb08" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1055.072428] env[65726]: DEBUG nova.network.neutron [None req-cc47cd13-9595-40ad-a0c8-afcfcfe7f019 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] [instance: 48923e3d-b00d-4034-bacf-7be82009fb08] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1055.204220] env[65726]: DEBUG oslo_vmware.api [None req-5c22cc7c-7dcc-417f-8f16-1f19636dd1f0 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': task-5116604, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1055.304256] env[65726]: DEBUG oslo_vmware.api [None req-534e859d-9bab-4a75-8dfe-7304a9dfecfd tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': task-5116605, 'name': Rename_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1055.435135] env[65726]: DEBUG oslo_concurrency.lockutils [req-e33d75da-e766-4a75-8533-638c91c379d6 req-2c28f547-af7b-4318-9aa2-687434625b9d service nova] Releasing lock "refresh_cache-df08d166-7811-49b5-9c16-a3434229d623" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1055.458552] env[65726]: DEBUG nova.scheduler.client.report [None req-fe2eb30a-001a-49d1-91b1-ed3f11534bdf tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1055.477633] env[65726]: DEBUG oslo_concurrency.lockutils [None req-4c7a0966-9c18-4500-91fc-ea14494d2511 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Acquiring lock "6d2b86c2-0a44-433f-8b3f-2e9c7f693baa" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1055.482259] env[65726]: DEBUG oslo_concurrency.lockutils [None req-4c7a0966-9c18-4500-91fc-ea14494d2511 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Lock "6d2b86c2-0a44-433f-8b3f-2e9c7f693baa" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1055.482259] env[65726]: DEBUG oslo_concurrency.lockutils [None req-4c7a0966-9c18-4500-91fc-ea14494d2511 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Acquiring lock "6d2b86c2-0a44-433f-8b3f-2e9c7f693baa-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1055.482259] env[65726]: DEBUG oslo_concurrency.lockutils [None req-4c7a0966-9c18-4500-91fc-ea14494d2511 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Lock "6d2b86c2-0a44-433f-8b3f-2e9c7f693baa-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1055.482259] env[65726]: DEBUG oslo_concurrency.lockutils [None req-4c7a0966-9c18-4500-91fc-ea14494d2511 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Lock "6d2b86c2-0a44-433f-8b3f-2e9c7f693baa-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1055.482259] env[65726]: INFO nova.compute.manager [None req-4c7a0966-9c18-4500-91fc-ea14494d2511 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: 6d2b86c2-0a44-433f-8b3f-2e9c7f693baa] Terminating instance [ 1055.509965] env[65726]: INFO nova.compute.manager [-] [instance: 2b8eb60b-dc9b-47a4-9ddc-205f2af69888] Took 1.38 seconds to deallocate network for instance. [ 1055.575577] env[65726]: WARNING openstack [None req-cc47cd13-9595-40ad-a0c8-afcfcfe7f019 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1055.575986] env[65726]: WARNING openstack [None req-cc47cd13-9595-40ad-a0c8-afcfcfe7f019 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1055.580985] env[65726]: DEBUG nova.network.neutron [None req-cc47cd13-9595-40ad-a0c8-afcfcfe7f019 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] [instance: 48923e3d-b00d-4034-bacf-7be82009fb08] Instance cache missing network info. {{(pid=65726) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 1055.649057] env[65726]: WARNING openstack [None req-cc47cd13-9595-40ad-a0c8-afcfcfe7f019 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1055.649453] env[65726]: WARNING openstack [None req-cc47cd13-9595-40ad-a0c8-afcfcfe7f019 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1055.704357] env[65726]: DEBUG oslo_vmware.api [None req-5c22cc7c-7dcc-417f-8f16-1f19636dd1f0 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': task-5116604, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1055.721831] env[65726]: WARNING openstack [None req-cc47cd13-9595-40ad-a0c8-afcfcfe7f019 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1055.722258] env[65726]: WARNING openstack [None req-cc47cd13-9595-40ad-a0c8-afcfcfe7f019 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1055.804322] env[65726]: DEBUG oslo_vmware.api [None req-534e859d-9bab-4a75-8dfe-7304a9dfecfd tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': task-5116605, 'name': Rename_Task, 'duration_secs': 0.579986} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1055.804687] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-534e859d-9bab-4a75-8dfe-7304a9dfecfd tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: f947529e-c930-4420-91a7-1e677b5ac2f1] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1055.804938] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ba052e3a-0ac8-4dd0-97ad-daf56a5f0479 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.814152] env[65726]: DEBUG oslo_vmware.api [None req-534e859d-9bab-4a75-8dfe-7304a9dfecfd tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Waiting for the task: (returnval){ [ 1055.814152] env[65726]: value = "task-5116606" [ 1055.814152] env[65726]: _type = "Task" [ 1055.814152] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1055.825488] env[65726]: DEBUG nova.network.neutron [None req-cc47cd13-9595-40ad-a0c8-afcfcfe7f019 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] [instance: 48923e3d-b00d-4034-bacf-7be82009fb08] Updating instance_info_cache with network_info: [{"id": "03191f4c-fd30-4ddc-abfd-75a70e8ab0d6", "address": "fa:16:3e:de:9d:4a", "network": {"id": "a2c8f0c5-7fad-4b72-8e1a-376e03d3c0ff", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-2135463722-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1d5cb47905144da49aaf07ddb7d263af", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "98f447de-d71e-41ef-bc37-ed97b4a1f58f", "external-id": "nsx-vlan-transportzone-904", "segmentation_id": 904, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap03191f4c-fd", "ovs_interfaceid": "03191f4c-fd30-4ddc-abfd-75a70e8ab0d6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1055.826783] env[65726]: DEBUG oslo_vmware.api [None req-534e859d-9bab-4a75-8dfe-7304a9dfecfd tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': task-5116606, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1055.869435] env[65726]: DEBUG nova.compute.manager [req-e3219cf8-b0e2-46c3-b1cb-eacecd8c8311 req-9fbbdd5f-5266-4bbc-a2e1-d88f95e1db5a service nova] [instance: ad4f00b5-3765-494f-8132-8b17b29d219a] Received event network-changed-508e0342-f4cf-409d-96e4-742e92925ba6 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1055.869720] env[65726]: DEBUG nova.compute.manager [req-e3219cf8-b0e2-46c3-b1cb-eacecd8c8311 req-9fbbdd5f-5266-4bbc-a2e1-d88f95e1db5a service nova] [instance: ad4f00b5-3765-494f-8132-8b17b29d219a] Refreshing instance network info cache due to event network-changed-508e0342-f4cf-409d-96e4-742e92925ba6. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 1055.869947] env[65726]: DEBUG oslo_concurrency.lockutils [req-e3219cf8-b0e2-46c3-b1cb-eacecd8c8311 req-9fbbdd5f-5266-4bbc-a2e1-d88f95e1db5a service nova] Acquiring lock "refresh_cache-ad4f00b5-3765-494f-8132-8b17b29d219a" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1055.870121] env[65726]: DEBUG oslo_concurrency.lockutils [req-e3219cf8-b0e2-46c3-b1cb-eacecd8c8311 req-9fbbdd5f-5266-4bbc-a2e1-d88f95e1db5a service nova] Acquired lock "refresh_cache-ad4f00b5-3765-494f-8132-8b17b29d219a" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1055.871069] env[65726]: DEBUG nova.network.neutron [req-e3219cf8-b0e2-46c3-b1cb-eacecd8c8311 req-9fbbdd5f-5266-4bbc-a2e1-d88f95e1db5a service nova] [instance: ad4f00b5-3765-494f-8132-8b17b29d219a] Refreshing network info cache for port 508e0342-f4cf-409d-96e4-742e92925ba6 {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 1055.964669] env[65726]: DEBUG oslo_concurrency.lockutils [None req-fe2eb30a-001a-49d1-91b1-ed3f11534bdf tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.933s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1055.968043] env[65726]: DEBUG oslo_concurrency.lockutils [None req-42911402-8543-4c4b-aab9-86db149010d2 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.966s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1055.970187] env[65726]: INFO nova.compute.claims [None req-42911402-8543-4c4b-aab9-86db149010d2 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 5cdf1d61-b817-4986-a48c-804a2374497a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1055.984823] env[65726]: DEBUG nova.compute.manager [None req-4c7a0966-9c18-4500-91fc-ea14494d2511 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: 6d2b86c2-0a44-433f-8b3f-2e9c7f693baa] Start destroying the instance on the hypervisor. {{(pid=65726) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1055.985465] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-4c7a0966-9c18-4500-91fc-ea14494d2511 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: 6d2b86c2-0a44-433f-8b3f-2e9c7f693baa] Destroying instance {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1055.987296] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e3c68b3-e89f-4040-93bb-3033d3eb4406 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.998297] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-4c7a0966-9c18-4500-91fc-ea14494d2511 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: 6d2b86c2-0a44-433f-8b3f-2e9c7f693baa] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1055.998630] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4980111e-5554-494b-b1a3-5f6b440cf83f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.005557] env[65726]: INFO nova.scheduler.client.report [None req-fe2eb30a-001a-49d1-91b1-ed3f11534bdf tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Deleted allocations for instance 2c812714-dccc-4d1e-bdb5-c11e446949c3 [ 1056.007884] env[65726]: DEBUG oslo_vmware.api [None req-4c7a0966-9c18-4500-91fc-ea14494d2511 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Waiting for the task: (returnval){ [ 1056.007884] env[65726]: value = "task-5116607" [ 1056.007884] env[65726]: _type = "Task" [ 1056.007884] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1056.018090] env[65726]: DEBUG oslo_concurrency.lockutils [None req-f3c5543b-5b17-4006-8d18-e03fa5503534 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1056.020971] env[65726]: DEBUG oslo_vmware.api [None req-4c7a0966-9c18-4500-91fc-ea14494d2511 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Task: {'id': task-5116607, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1056.209879] env[65726]: DEBUG oslo_vmware.api [None req-5c22cc7c-7dcc-417f-8f16-1f19636dd1f0 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': task-5116604, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.537267} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1056.210990] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-5c22cc7c-7dcc-417f-8f16-1f19636dd1f0 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore2] 51f5337d-2e23-4b7c-b590-76be6bd82411/51f5337d-2e23-4b7c-b590-76be6bd82411.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1056.210990] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-5c22cc7c-7dcc-417f-8f16-1f19636dd1f0 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: 51f5337d-2e23-4b7c-b590-76be6bd82411] Extending root virtual disk to 1048576 {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1056.211277] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-dc9d9ce3-02ea-44de-a857-07ab34befdaa {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.219618] env[65726]: DEBUG oslo_vmware.api [None req-5c22cc7c-7dcc-417f-8f16-1f19636dd1f0 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Waiting for the task: (returnval){ [ 1056.219618] env[65726]: value = "task-5116608" [ 1056.219618] env[65726]: _type = "Task" [ 1056.219618] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1056.231994] env[65726]: DEBUG oslo_vmware.api [None req-5c22cc7c-7dcc-417f-8f16-1f19636dd1f0 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': task-5116608, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1056.325728] env[65726]: DEBUG oslo_vmware.api [None req-534e859d-9bab-4a75-8dfe-7304a9dfecfd tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': task-5116606, 'name': PowerOnVM_Task} progress is 88%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1056.328512] env[65726]: DEBUG oslo_concurrency.lockutils [None req-cc47cd13-9595-40ad-a0c8-afcfcfe7f019 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Releasing lock "refresh_cache-48923e3d-b00d-4034-bacf-7be82009fb08" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1056.328859] env[65726]: DEBUG nova.compute.manager [None req-cc47cd13-9595-40ad-a0c8-afcfcfe7f019 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] [instance: 48923e3d-b00d-4034-bacf-7be82009fb08] Instance network_info: |[{"id": "03191f4c-fd30-4ddc-abfd-75a70e8ab0d6", "address": "fa:16:3e:de:9d:4a", "network": {"id": "a2c8f0c5-7fad-4b72-8e1a-376e03d3c0ff", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-2135463722-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1d5cb47905144da49aaf07ddb7d263af", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "98f447de-d71e-41ef-bc37-ed97b4a1f58f", "external-id": "nsx-vlan-transportzone-904", "segmentation_id": 904, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap03191f4c-fd", "ovs_interfaceid": "03191f4c-fd30-4ddc-abfd-75a70e8ab0d6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1056.329371] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-cc47cd13-9595-40ad-a0c8-afcfcfe7f019 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] [instance: 48923e3d-b00d-4034-bacf-7be82009fb08] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:de:9d:4a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '98f447de-d71e-41ef-bc37-ed97b4a1f58f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '03191f4c-fd30-4ddc-abfd-75a70e8ab0d6', 'vif_model': 'vmxnet3'}] {{(pid=65726) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1056.338683] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-cc47cd13-9595-40ad-a0c8-afcfcfe7f019 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1056.338972] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 48923e3d-b00d-4034-bacf-7be82009fb08] Creating VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1056.339242] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a164b273-8b32-4b4e-8b83-8b9c1fce2d08 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.361734] env[65726]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1056.361734] env[65726]: value = "task-5116609" [ 1056.361734] env[65726]: _type = "Task" [ 1056.361734] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1056.374732] env[65726]: WARNING openstack [req-e3219cf8-b0e2-46c3-b1cb-eacecd8c8311 req-9fbbdd5f-5266-4bbc-a2e1-d88f95e1db5a service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1056.375139] env[65726]: WARNING openstack [req-e3219cf8-b0e2-46c3-b1cb-eacecd8c8311 req-9fbbdd5f-5266-4bbc-a2e1-d88f95e1db5a service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1056.382767] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116609, 'name': CreateVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1056.521882] env[65726]: DEBUG oslo_concurrency.lockutils [None req-fe2eb30a-001a-49d1-91b1-ed3f11534bdf tempest-ServersTestMultiNic-1552153896 tempest-ServersTestMultiNic-1552153896-project-member] Lock "2c812714-dccc-4d1e-bdb5-c11e446949c3" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 10.879s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1056.526943] env[65726]: DEBUG oslo_vmware.api [None req-4c7a0966-9c18-4500-91fc-ea14494d2511 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Task: {'id': task-5116607, 'name': PowerOffVM_Task, 'duration_secs': 0.215833} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1056.527466] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-4c7a0966-9c18-4500-91fc-ea14494d2511 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: 6d2b86c2-0a44-433f-8b3f-2e9c7f693baa] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1056.527643] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-4c7a0966-9c18-4500-91fc-ea14494d2511 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: 6d2b86c2-0a44-433f-8b3f-2e9c7f693baa] Unregistering the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1056.527922] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ea83e745-21bf-4829-ae60-6441d362557d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.538707] env[65726]: WARNING openstack [req-e3219cf8-b0e2-46c3-b1cb-eacecd8c8311 req-9fbbdd5f-5266-4bbc-a2e1-d88f95e1db5a service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1056.539255] env[65726]: WARNING openstack [req-e3219cf8-b0e2-46c3-b1cb-eacecd8c8311 req-9fbbdd5f-5266-4bbc-a2e1-d88f95e1db5a service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1056.600731] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-4c7a0966-9c18-4500-91fc-ea14494d2511 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: 6d2b86c2-0a44-433f-8b3f-2e9c7f693baa] Unregistered the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1056.600988] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-4c7a0966-9c18-4500-91fc-ea14494d2511 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: 6d2b86c2-0a44-433f-8b3f-2e9c7f693baa] Deleting contents of the VM from datastore datastore2 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1056.601198] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-4c7a0966-9c18-4500-91fc-ea14494d2511 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Deleting the datastore file [datastore2] 6d2b86c2-0a44-433f-8b3f-2e9c7f693baa {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1056.601577] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-cf29ad0e-16bd-409d-85eb-530a19bcfac5 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.610876] env[65726]: DEBUG oslo_vmware.api [None req-4c7a0966-9c18-4500-91fc-ea14494d2511 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Waiting for the task: (returnval){ [ 1056.610876] env[65726]: value = "task-5116611" [ 1056.610876] env[65726]: _type = "Task" [ 1056.610876] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1056.621162] env[65726]: WARNING openstack [req-e3219cf8-b0e2-46c3-b1cb-eacecd8c8311 req-9fbbdd5f-5266-4bbc-a2e1-d88f95e1db5a service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1056.621726] env[65726]: WARNING openstack [req-e3219cf8-b0e2-46c3-b1cb-eacecd8c8311 req-9fbbdd5f-5266-4bbc-a2e1-d88f95e1db5a service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1056.634044] env[65726]: DEBUG oslo_vmware.api [None req-4c7a0966-9c18-4500-91fc-ea14494d2511 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Task: {'id': task-5116611, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1056.720339] env[65726]: DEBUG nova.network.neutron [req-e3219cf8-b0e2-46c3-b1cb-eacecd8c8311 req-9fbbdd5f-5266-4bbc-a2e1-d88f95e1db5a service nova] [instance: ad4f00b5-3765-494f-8132-8b17b29d219a] Updated VIF entry in instance network info cache for port 508e0342-f4cf-409d-96e4-742e92925ba6. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 1056.721015] env[65726]: DEBUG nova.network.neutron [req-e3219cf8-b0e2-46c3-b1cb-eacecd8c8311 req-9fbbdd5f-5266-4bbc-a2e1-d88f95e1db5a service nova] [instance: ad4f00b5-3765-494f-8132-8b17b29d219a] Updating instance_info_cache with network_info: [{"id": "508e0342-f4cf-409d-96e4-742e92925ba6", "address": "fa:16:3e:89:e2:97", "network": {"id": "ad1373b5-e0ed-4ee6-a97c-af931afbab35", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-909418417-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.247", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a090d553766847e58e3231b966c92565", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5f60c972-a72d-4c5f-a250-faadfd6eafbe", "external-id": "nsx-vlan-transportzone-932", "segmentation_id": 932, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap508e0342-f4", "ovs_interfaceid": "508e0342-f4cf-409d-96e4-742e92925ba6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1056.731515] env[65726]: DEBUG oslo_vmware.api [None req-5c22cc7c-7dcc-417f-8f16-1f19636dd1f0 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': task-5116608, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.097345} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1056.731833] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-5c22cc7c-7dcc-417f-8f16-1f19636dd1f0 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: 51f5337d-2e23-4b7c-b590-76be6bd82411] Extended root virtual disk {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1056.732817] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a926e247-3299-4052-9fb5-ccd09cc31eb4 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.758768] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-5c22cc7c-7dcc-417f-8f16-1f19636dd1f0 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: 51f5337d-2e23-4b7c-b590-76be6bd82411] Reconfiguring VM instance instance-0000005d to attach disk [datastore2] 51f5337d-2e23-4b7c-b590-76be6bd82411/51f5337d-2e23-4b7c-b590-76be6bd82411.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1056.760054] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-06b22b11-cf44-4eec-8dcc-053ee63ffb1a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.782259] env[65726]: DEBUG oslo_vmware.api [None req-5c22cc7c-7dcc-417f-8f16-1f19636dd1f0 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Waiting for the task: (returnval){ [ 1056.782259] env[65726]: value = "task-5116612" [ 1056.782259] env[65726]: _type = "Task" [ 1056.782259] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1056.792662] env[65726]: DEBUG oslo_vmware.api [None req-5c22cc7c-7dcc-417f-8f16-1f19636dd1f0 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': task-5116612, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1056.826709] env[65726]: DEBUG oslo_vmware.api [None req-534e859d-9bab-4a75-8dfe-7304a9dfecfd tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': task-5116606, 'name': PowerOnVM_Task, 'duration_secs': 0.761179} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1056.828007] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-534e859d-9bab-4a75-8dfe-7304a9dfecfd tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: f947529e-c930-4420-91a7-1e677b5ac2f1] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1056.828229] env[65726]: INFO nova.compute.manager [None req-534e859d-9bab-4a75-8dfe-7304a9dfecfd tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: f947529e-c930-4420-91a7-1e677b5ac2f1] Took 10.13 seconds to spawn the instance on the hypervisor. [ 1056.828384] env[65726]: DEBUG nova.compute.manager [None req-534e859d-9bab-4a75-8dfe-7304a9dfecfd tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: f947529e-c930-4420-91a7-1e677b5ac2f1] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1056.830044] env[65726]: DEBUG nova.compute.manager [req-e8cd49c9-99d6-46ad-b9cd-f1c981faa41f req-d5c9f35f-5b96-46db-a3b0-d06626465a20 service nova] [instance: 48923e3d-b00d-4034-bacf-7be82009fb08] Received event network-changed-03191f4c-fd30-4ddc-abfd-75a70e8ab0d6 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1056.830044] env[65726]: DEBUG nova.compute.manager [req-e8cd49c9-99d6-46ad-b9cd-f1c981faa41f req-d5c9f35f-5b96-46db-a3b0-d06626465a20 service nova] [instance: 48923e3d-b00d-4034-bacf-7be82009fb08] Refreshing instance network info cache due to event network-changed-03191f4c-fd30-4ddc-abfd-75a70e8ab0d6. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 1056.830044] env[65726]: DEBUG oslo_concurrency.lockutils [req-e8cd49c9-99d6-46ad-b9cd-f1c981faa41f req-d5c9f35f-5b96-46db-a3b0-d06626465a20 service nova] Acquiring lock "refresh_cache-48923e3d-b00d-4034-bacf-7be82009fb08" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1056.830044] env[65726]: DEBUG oslo_concurrency.lockutils [req-e8cd49c9-99d6-46ad-b9cd-f1c981faa41f req-d5c9f35f-5b96-46db-a3b0-d06626465a20 service nova] Acquired lock "refresh_cache-48923e3d-b00d-4034-bacf-7be82009fb08" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1056.830524] env[65726]: DEBUG nova.network.neutron [req-e8cd49c9-99d6-46ad-b9cd-f1c981faa41f req-d5c9f35f-5b96-46db-a3b0-d06626465a20 service nova] [instance: 48923e3d-b00d-4034-bacf-7be82009fb08] Refreshing network info cache for port 03191f4c-fd30-4ddc-abfd-75a70e8ab0d6 {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 1056.831949] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c05ce3f2-c953-49b5-998a-b398440c37ae {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.874322] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116609, 'name': CreateVM_Task, 'duration_secs': 0.388605} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1056.874582] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 48923e3d-b00d-4034-bacf-7be82009fb08] Created VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1056.875433] env[65726]: WARNING openstack [None req-cc47cd13-9595-40ad-a0c8-afcfcfe7f019 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1056.875723] env[65726]: WARNING openstack [None req-cc47cd13-9595-40ad-a0c8-afcfcfe7f019 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1056.880988] env[65726]: DEBUG oslo_concurrency.lockutils [None req-cc47cd13-9595-40ad-a0c8-afcfcfe7f019 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1056.881235] env[65726]: DEBUG oslo_concurrency.lockutils [None req-cc47cd13-9595-40ad-a0c8-afcfcfe7f019 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1056.881575] env[65726]: DEBUG oslo_concurrency.lockutils [None req-cc47cd13-9595-40ad-a0c8-afcfcfe7f019 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1056.882219] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4b834ce4-1d84-4a64-946c-24738f0cb6fd {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.888106] env[65726]: DEBUG oslo_vmware.api [None req-cc47cd13-9595-40ad-a0c8-afcfcfe7f019 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Waiting for the task: (returnval){ [ 1056.888106] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52da846d-2a50-245f-5acd-940459d31abc" [ 1056.888106] env[65726]: _type = "Task" [ 1056.888106] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1056.897852] env[65726]: DEBUG oslo_vmware.api [None req-cc47cd13-9595-40ad-a0c8-afcfcfe7f019 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52da846d-2a50-245f-5acd-940459d31abc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1057.123890] env[65726]: DEBUG oslo_vmware.api [None req-4c7a0966-9c18-4500-91fc-ea14494d2511 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Task: {'id': task-5116611, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.210313} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1057.124209] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-4c7a0966-9c18-4500-91fc-ea14494d2511 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Deleted the datastore file {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1057.124441] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-4c7a0966-9c18-4500-91fc-ea14494d2511 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: 6d2b86c2-0a44-433f-8b3f-2e9c7f693baa] Deleted contents of the VM from datastore datastore2 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1057.124659] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-4c7a0966-9c18-4500-91fc-ea14494d2511 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: 6d2b86c2-0a44-433f-8b3f-2e9c7f693baa] Instance destroyed {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1057.124962] env[65726]: INFO nova.compute.manager [None req-4c7a0966-9c18-4500-91fc-ea14494d2511 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] [instance: 6d2b86c2-0a44-433f-8b3f-2e9c7f693baa] Took 1.14 seconds to destroy the instance on the hypervisor. [ 1057.125501] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-4c7a0966-9c18-4500-91fc-ea14494d2511 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1057.125501] env[65726]: DEBUG nova.compute.manager [-] [instance: 6d2b86c2-0a44-433f-8b3f-2e9c7f693baa] Deallocating network for instance {{(pid=65726) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1057.125615] env[65726]: DEBUG nova.network.neutron [-] [instance: 6d2b86c2-0a44-433f-8b3f-2e9c7f693baa] deallocate_for_instance() {{(pid=65726) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 1057.126308] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1057.126498] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1057.167996] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1057.168317] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1057.227527] env[65726]: DEBUG oslo_concurrency.lockutils [req-e3219cf8-b0e2-46c3-b1cb-eacecd8c8311 req-9fbbdd5f-5266-4bbc-a2e1-d88f95e1db5a service nova] Releasing lock "refresh_cache-ad4f00b5-3765-494f-8132-8b17b29d219a" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1057.228410] env[65726]: DEBUG nova.compute.manager [req-e3219cf8-b0e2-46c3-b1cb-eacecd8c8311 req-9fbbdd5f-5266-4bbc-a2e1-d88f95e1db5a service nova] [instance: 2b8eb60b-dc9b-47a4-9ddc-205f2af69888] Received event network-vif-deleted-f210c8b9-8838-4774-838c-aeb3860a63ca {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1057.241705] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff47d523-ccea-42ba-b16f-903c464b2171 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.251887] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9540cdc0-f5be-404a-af65-f1d5bdcbd02e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.290086] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c48a974-f0b8-4f70-9701-1216d2b0984e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.299164] env[65726]: DEBUG oslo_vmware.api [None req-5c22cc7c-7dcc-417f-8f16-1f19636dd1f0 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': task-5116612, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1057.303959] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-239e3ac0-2382-4dea-b194-f93b4d985340 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.329583] env[65726]: DEBUG nova.compute.provider_tree [None req-42911402-8543-4c4b-aab9-86db149010d2 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1057.337751] env[65726]: WARNING openstack [req-e8cd49c9-99d6-46ad-b9cd-f1c981faa41f req-d5c9f35f-5b96-46db-a3b0-d06626465a20 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1057.338461] env[65726]: WARNING openstack [req-e8cd49c9-99d6-46ad-b9cd-f1c981faa41f req-d5c9f35f-5b96-46db-a3b0-d06626465a20 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1057.356442] env[65726]: INFO nova.compute.manager [None req-534e859d-9bab-4a75-8dfe-7304a9dfecfd tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: f947529e-c930-4420-91a7-1e677b5ac2f1] Took 18.63 seconds to build instance. [ 1057.399677] env[65726]: DEBUG oslo_vmware.api [None req-cc47cd13-9595-40ad-a0c8-afcfcfe7f019 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52da846d-2a50-245f-5acd-940459d31abc, 'name': SearchDatastore_Task, 'duration_secs': 0.011058} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1057.400191] env[65726]: DEBUG oslo_concurrency.lockutils [None req-cc47cd13-9595-40ad-a0c8-afcfcfe7f019 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1057.400294] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-cc47cd13-9595-40ad-a0c8-afcfcfe7f019 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] [instance: 48923e3d-b00d-4034-bacf-7be82009fb08] Processing image b52362a3-ee8a-4cbf-b06f-513b0cc8f95c {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1057.400517] env[65726]: DEBUG oslo_concurrency.lockutils [None req-cc47cd13-9595-40ad-a0c8-afcfcfe7f019 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1057.400669] env[65726]: DEBUG oslo_concurrency.lockutils [None req-cc47cd13-9595-40ad-a0c8-afcfcfe7f019 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1057.400849] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-cc47cd13-9595-40ad-a0c8-afcfcfe7f019 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1057.401349] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0d8771dd-9c29-4f57-92c6-e4e8df012d47 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.411658] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-cc47cd13-9595-40ad-a0c8-afcfcfe7f019 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1057.411842] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-cc47cd13-9595-40ad-a0c8-afcfcfe7f019 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65726) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1057.412673] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ae9998d4-b4db-4d63-b2b4-dc8b87cc74d3 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.419719] env[65726]: DEBUG oslo_vmware.api [None req-cc47cd13-9595-40ad-a0c8-afcfcfe7f019 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Waiting for the task: (returnval){ [ 1057.419719] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]527f77c9-56d2-9b96-e741-59e6030f4b69" [ 1057.419719] env[65726]: _type = "Task" [ 1057.419719] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1057.429085] env[65726]: DEBUG oslo_vmware.api [None req-cc47cd13-9595-40ad-a0c8-afcfcfe7f019 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]527f77c9-56d2-9b96-e741-59e6030f4b69, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1057.622831] env[65726]: WARNING openstack [req-e8cd49c9-99d6-46ad-b9cd-f1c981faa41f req-d5c9f35f-5b96-46db-a3b0-d06626465a20 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1057.623232] env[65726]: WARNING openstack [req-e8cd49c9-99d6-46ad-b9cd-f1c981faa41f req-d5c9f35f-5b96-46db-a3b0-d06626465a20 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1057.768370] env[65726]: WARNING openstack [req-e8cd49c9-99d6-46ad-b9cd-f1c981faa41f req-d5c9f35f-5b96-46db-a3b0-d06626465a20 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1057.768790] env[65726]: WARNING openstack [req-e8cd49c9-99d6-46ad-b9cd-f1c981faa41f req-d5c9f35f-5b96-46db-a3b0-d06626465a20 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1057.798452] env[65726]: DEBUG oslo_vmware.api [None req-5c22cc7c-7dcc-417f-8f16-1f19636dd1f0 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': task-5116612, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1057.833033] env[65726]: DEBUG nova.scheduler.client.report [None req-42911402-8543-4c4b-aab9-86db149010d2 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1057.858024] env[65726]: DEBUG nova.network.neutron [req-e8cd49c9-99d6-46ad-b9cd-f1c981faa41f req-d5c9f35f-5b96-46db-a3b0-d06626465a20 service nova] [instance: 48923e3d-b00d-4034-bacf-7be82009fb08] Updated VIF entry in instance network info cache for port 03191f4c-fd30-4ddc-abfd-75a70e8ab0d6. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 1057.858627] env[65726]: DEBUG nova.network.neutron [req-e8cd49c9-99d6-46ad-b9cd-f1c981faa41f req-d5c9f35f-5b96-46db-a3b0-d06626465a20 service nova] [instance: 48923e3d-b00d-4034-bacf-7be82009fb08] Updating instance_info_cache with network_info: [{"id": "03191f4c-fd30-4ddc-abfd-75a70e8ab0d6", "address": "fa:16:3e:de:9d:4a", "network": {"id": "a2c8f0c5-7fad-4b72-8e1a-376e03d3c0ff", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-2135463722-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1d5cb47905144da49aaf07ddb7d263af", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "98f447de-d71e-41ef-bc37-ed97b4a1f58f", "external-id": "nsx-vlan-transportzone-904", "segmentation_id": 904, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap03191f4c-fd", "ovs_interfaceid": "03191f4c-fd30-4ddc-abfd-75a70e8ab0d6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1057.860308] env[65726]: DEBUG oslo_concurrency.lockutils [None req-534e859d-9bab-4a75-8dfe-7304a9dfecfd tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Lock "f947529e-c930-4420-91a7-1e677b5ac2f1" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 20.147s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1057.932224] env[65726]: DEBUG oslo_vmware.api [None req-cc47cd13-9595-40ad-a0c8-afcfcfe7f019 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]527f77c9-56d2-9b96-e741-59e6030f4b69, 'name': SearchDatastore_Task, 'duration_secs': 0.010121} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1057.933325] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ed98170b-e21a-433b-8d8b-c882c96d6c83 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.940056] env[65726]: DEBUG oslo_vmware.api [None req-cc47cd13-9595-40ad-a0c8-afcfcfe7f019 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Waiting for the task: (returnval){ [ 1057.940056] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5281b972-266c-8752-1de3-6279d2964052" [ 1057.940056] env[65726]: _type = "Task" [ 1057.940056] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1057.949751] env[65726]: DEBUG oslo_vmware.api [None req-cc47cd13-9595-40ad-a0c8-afcfcfe7f019 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5281b972-266c-8752-1de3-6279d2964052, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1057.971826] env[65726]: DEBUG nova.network.neutron [-] [instance: 6d2b86c2-0a44-433f-8b3f-2e9c7f693baa] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1058.284884] env[65726]: DEBUG oslo_concurrency.lockutils [None req-690d3db3-4b44-4e7c-bc5b-a409b814ad64 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Acquiring lock "f947529e-c930-4420-91a7-1e677b5ac2f1" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1058.285105] env[65726]: DEBUG oslo_concurrency.lockutils [None req-690d3db3-4b44-4e7c-bc5b-a409b814ad64 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Lock "f947529e-c930-4420-91a7-1e677b5ac2f1" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1058.285297] env[65726]: DEBUG nova.compute.manager [None req-690d3db3-4b44-4e7c-bc5b-a409b814ad64 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: f947529e-c930-4420-91a7-1e677b5ac2f1] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1058.286644] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20dbc122-9061-4bdd-a590-e25fa1ac6a06 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.312052] env[65726]: DEBUG nova.compute.manager [None req-690d3db3-4b44-4e7c-bc5b-a409b814ad64 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: f947529e-c930-4420-91a7-1e677b5ac2f1] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=65726) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3459}} [ 1058.312826] env[65726]: DEBUG nova.objects.instance [None req-690d3db3-4b44-4e7c-bc5b-a409b814ad64 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Lazy-loading 'flavor' on Instance uuid f947529e-c930-4420-91a7-1e677b5ac2f1 {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1058.315267] env[65726]: DEBUG oslo_vmware.api [None req-5c22cc7c-7dcc-417f-8f16-1f19636dd1f0 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': task-5116612, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1058.338815] env[65726]: DEBUG oslo_concurrency.lockutils [None req-42911402-8543-4c4b-aab9-86db149010d2 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.371s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1058.339560] env[65726]: DEBUG nova.compute.manager [None req-42911402-8543-4c4b-aab9-86db149010d2 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 5cdf1d61-b817-4986-a48c-804a2374497a] Start building networks asynchronously for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1058.342634] env[65726]: DEBUG oslo_concurrency.lockutils [None req-dc5e36ff-c9dd-4078-bec5-b01919099186 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 6.652s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1058.362288] env[65726]: DEBUG oslo_concurrency.lockutils [req-e8cd49c9-99d6-46ad-b9cd-f1c981faa41f req-d5c9f35f-5b96-46db-a3b0-d06626465a20 service nova] Releasing lock "refresh_cache-48923e3d-b00d-4034-bacf-7be82009fb08" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1058.451866] env[65726]: DEBUG oslo_vmware.api [None req-cc47cd13-9595-40ad-a0c8-afcfcfe7f019 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5281b972-266c-8752-1de3-6279d2964052, 'name': SearchDatastore_Task, 'duration_secs': 0.011045} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1058.452241] env[65726]: DEBUG oslo_concurrency.lockutils [None req-cc47cd13-9595-40ad-a0c8-afcfcfe7f019 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1058.452537] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-cc47cd13-9595-40ad-a0c8-afcfcfe7f019 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore2] 48923e3d-b00d-4034-bacf-7be82009fb08/48923e3d-b00d-4034-bacf-7be82009fb08.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1058.452808] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8b0fd0e9-9080-4157-a8b0-6203b559ebcc {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.460546] env[65726]: DEBUG oslo_vmware.api [None req-cc47cd13-9595-40ad-a0c8-afcfcfe7f019 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Waiting for the task: (returnval){ [ 1058.460546] env[65726]: value = "task-5116613" [ 1058.460546] env[65726]: _type = "Task" [ 1058.460546] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1058.471767] env[65726]: DEBUG oslo_vmware.api [None req-cc47cd13-9595-40ad-a0c8-afcfcfe7f019 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Task: {'id': task-5116613, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1058.474377] env[65726]: INFO nova.compute.manager [-] [instance: 6d2b86c2-0a44-433f-8b3f-2e9c7f693baa] Took 1.35 seconds to deallocate network for instance. [ 1058.801351] env[65726]: DEBUG oslo_vmware.api [None req-5c22cc7c-7dcc-417f-8f16-1f19636dd1f0 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': task-5116612, 'name': ReconfigVM_Task, 'duration_secs': 1.648674} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1058.801351] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-5c22cc7c-7dcc-417f-8f16-1f19636dd1f0 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: 51f5337d-2e23-4b7c-b590-76be6bd82411] Reconfigured VM instance instance-0000005d to attach disk [datastore2] 51f5337d-2e23-4b7c-b590-76be6bd82411/51f5337d-2e23-4b7c-b590-76be6bd82411.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1058.801351] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-bb8f66cf-2d2e-4ab2-ace6-4a1a48309296 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.812994] env[65726]: DEBUG oslo_vmware.api [None req-5c22cc7c-7dcc-417f-8f16-1f19636dd1f0 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Waiting for the task: (returnval){ [ 1058.812994] env[65726]: value = "task-5116614" [ 1058.812994] env[65726]: _type = "Task" [ 1058.812994] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1058.828355] env[65726]: DEBUG oslo_vmware.api [None req-5c22cc7c-7dcc-417f-8f16-1f19636dd1f0 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': task-5116614, 'name': Rename_Task} progress is 6%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1058.850399] env[65726]: DEBUG nova.compute.utils [None req-42911402-8543-4c4b-aab9-86db149010d2 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Using /dev/sd instead of None {{(pid=65726) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1058.853855] env[65726]: INFO nova.compute.claims [None req-dc5e36ff-c9dd-4078-bec5-b01919099186 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: a56be6fa-6027-46ee-9aa8-332fe4222d68] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1058.863959] env[65726]: DEBUG nova.compute.manager [None req-42911402-8543-4c4b-aab9-86db149010d2 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 5cdf1d61-b817-4986-a48c-804a2374497a] Allocating IP information in the background. {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1058.864369] env[65726]: DEBUG nova.network.neutron [None req-42911402-8543-4c4b-aab9-86db149010d2 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 5cdf1d61-b817-4986-a48c-804a2374497a] allocate_for_instance() {{(pid=65726) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 1058.864538] env[65726]: WARNING neutronclient.v2_0.client [None req-42911402-8543-4c4b-aab9-86db149010d2 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1058.864935] env[65726]: WARNING neutronclient.v2_0.client [None req-42911402-8543-4c4b-aab9-86db149010d2 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1058.865630] env[65726]: WARNING openstack [None req-42911402-8543-4c4b-aab9-86db149010d2 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1058.866114] env[65726]: WARNING openstack [None req-42911402-8543-4c4b-aab9-86db149010d2 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1058.894031] env[65726]: DEBUG nova.compute.manager [req-f089cc86-1c2c-46b6-859c-c46edfefc5a5 req-e9697989-1bf7-462b-be3a-68ebfae42676 service nova] [instance: 6d2b86c2-0a44-433f-8b3f-2e9c7f693baa] Received event network-vif-deleted-b487d53a-9bda-4e3c-be61-f26acbc8ae0f {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1058.935916] env[65726]: DEBUG nova.policy [None req-42911402-8543-4c4b-aab9-86db149010d2 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a82797497789484a8fbf88a23a449c95', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f93d852e2c904f42981cbfff882558e9', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65726) authorize /opt/stack/nova/nova/policy.py:192}} [ 1058.973699] env[65726]: DEBUG oslo_vmware.api [None req-cc47cd13-9595-40ad-a0c8-afcfcfe7f019 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Task: {'id': task-5116613, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1058.983551] env[65726]: DEBUG oslo_concurrency.lockutils [None req-4c7a0966-9c18-4500-91fc-ea14494d2511 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1059.298260] env[65726]: DEBUG nova.network.neutron [None req-42911402-8543-4c4b-aab9-86db149010d2 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 5cdf1d61-b817-4986-a48c-804a2374497a] Successfully created port: 94113d1c-2016-4a98-b42a-6ae90eee0b24 {{(pid=65726) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1059.324018] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-690d3db3-4b44-4e7c-bc5b-a409b814ad64 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: f947529e-c930-4420-91a7-1e677b5ac2f1] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1059.328993] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6e27d22b-0635-4d50-8e92-b12acfcd6913 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.332028] env[65726]: DEBUG oslo_vmware.api [None req-5c22cc7c-7dcc-417f-8f16-1f19636dd1f0 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': task-5116614, 'name': Rename_Task, 'duration_secs': 0.177187} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1059.332028] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-5c22cc7c-7dcc-417f-8f16-1f19636dd1f0 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: 51f5337d-2e23-4b7c-b590-76be6bd82411] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1059.332615] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e36a1ad9-14c6-4cc4-9175-926e311c799f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.337400] env[65726]: DEBUG oslo_vmware.api [None req-690d3db3-4b44-4e7c-bc5b-a409b814ad64 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Waiting for the task: (returnval){ [ 1059.337400] env[65726]: value = "task-5116615" [ 1059.337400] env[65726]: _type = "Task" [ 1059.337400] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1059.343081] env[65726]: DEBUG oslo_vmware.api [None req-5c22cc7c-7dcc-417f-8f16-1f19636dd1f0 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Waiting for the task: (returnval){ [ 1059.343081] env[65726]: value = "task-5116616" [ 1059.343081] env[65726]: _type = "Task" [ 1059.343081] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1059.351341] env[65726]: DEBUG oslo_vmware.api [None req-690d3db3-4b44-4e7c-bc5b-a409b814ad64 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': task-5116615, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.360738] env[65726]: DEBUG oslo_vmware.api [None req-5c22cc7c-7dcc-417f-8f16-1f19636dd1f0 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': task-5116616, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.366071] env[65726]: INFO nova.compute.resource_tracker [None req-dc5e36ff-c9dd-4078-bec5-b01919099186 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: a56be6fa-6027-46ee-9aa8-332fe4222d68] Updating resource usage from migration 0ae40754-d17a-4cf6-a2d0-e1747b172ad4 [ 1059.369966] env[65726]: DEBUG nova.compute.manager [None req-42911402-8543-4c4b-aab9-86db149010d2 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 5cdf1d61-b817-4986-a48c-804a2374497a] Start building block device mappings for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1059.477116] env[65726]: DEBUG oslo_vmware.api [None req-cc47cd13-9595-40ad-a0c8-afcfcfe7f019 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Task: {'id': task-5116613, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.777895} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1059.481499] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-cc47cd13-9595-40ad-a0c8-afcfcfe7f019 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore2] 48923e3d-b00d-4034-bacf-7be82009fb08/48923e3d-b00d-4034-bacf-7be82009fb08.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1059.481499] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-cc47cd13-9595-40ad-a0c8-afcfcfe7f019 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] [instance: 48923e3d-b00d-4034-bacf-7be82009fb08] Extending root virtual disk to 1048576 {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1059.484865] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-93aa498e-eeca-475c-ab3d-9a599ce6b889 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.498292] env[65726]: DEBUG oslo_vmware.api [None req-cc47cd13-9595-40ad-a0c8-afcfcfe7f019 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Waiting for the task: (returnval){ [ 1059.498292] env[65726]: value = "task-5116617" [ 1059.498292] env[65726]: _type = "Task" [ 1059.498292] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1059.513362] env[65726]: DEBUG oslo_vmware.api [None req-cc47cd13-9595-40ad-a0c8-afcfcfe7f019 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Task: {'id': task-5116617, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.667060] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce24762a-a451-4923-aded-067e431f5fda {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.676732] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d83fd41-5a91-4075-9d9a-ec4ca93237b8 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.708743] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb4974f7-3030-4d29-b8b0-92660839fc44 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.717638] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18d7c43a-4a62-4163-877d-eef4f1279ba7 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.733359] env[65726]: DEBUG nova.compute.provider_tree [None req-dc5e36ff-c9dd-4078-bec5-b01919099186 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1059.854837] env[65726]: DEBUG oslo_vmware.api [None req-690d3db3-4b44-4e7c-bc5b-a409b814ad64 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': task-5116615, 'name': PowerOffVM_Task, 'duration_secs': 0.2021} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1059.855689] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-690d3db3-4b44-4e7c-bc5b-a409b814ad64 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: f947529e-c930-4420-91a7-1e677b5ac2f1] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1059.855958] env[65726]: DEBUG nova.compute.manager [None req-690d3db3-4b44-4e7c-bc5b-a409b814ad64 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: f947529e-c930-4420-91a7-1e677b5ac2f1] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1059.857029] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7dda80f-37cc-478c-a056-b96021a49916 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.865725] env[65726]: DEBUG oslo_vmware.api [None req-5c22cc7c-7dcc-417f-8f16-1f19636dd1f0 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': task-5116616, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.986036] env[65726]: DEBUG oslo_concurrency.lockutils [None req-f263eae6-96b2-45f2-b03a-d97fe5d0f1d8 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Acquiring lock "6fd70b4a-63e4-4258-9cff-f3b582500b9f" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1059.987547] env[65726]: DEBUG oslo_concurrency.lockutils [None req-f263eae6-96b2-45f2-b03a-d97fe5d0f1d8 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Lock "6fd70b4a-63e4-4258-9cff-f3b582500b9f" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.001s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1059.987547] env[65726]: INFO nova.compute.manager [None req-f263eae6-96b2-45f2-b03a-d97fe5d0f1d8 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] [instance: 6fd70b4a-63e4-4258-9cff-f3b582500b9f] Shelving [ 1060.010992] env[65726]: DEBUG oslo_vmware.api [None req-cc47cd13-9595-40ad-a0c8-afcfcfe7f019 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Task: {'id': task-5116617, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.097693} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1060.011340] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-cc47cd13-9595-40ad-a0c8-afcfcfe7f019 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] [instance: 48923e3d-b00d-4034-bacf-7be82009fb08] Extended root virtual disk {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1060.012228] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0999abf5-fbe9-4f51-93a5-ee675f5cd48c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.035904] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-cc47cd13-9595-40ad-a0c8-afcfcfe7f019 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] [instance: 48923e3d-b00d-4034-bacf-7be82009fb08] Reconfiguring VM instance instance-0000005e to attach disk [datastore2] 48923e3d-b00d-4034-bacf-7be82009fb08/48923e3d-b00d-4034-bacf-7be82009fb08.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1060.036698] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-24801761-599a-4ba4-ae21-783943254bb1 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.058181] env[65726]: DEBUG oslo_vmware.api [None req-cc47cd13-9595-40ad-a0c8-afcfcfe7f019 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Waiting for the task: (returnval){ [ 1060.058181] env[65726]: value = "task-5116618" [ 1060.058181] env[65726]: _type = "Task" [ 1060.058181] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1060.067958] env[65726]: DEBUG oslo_vmware.api [None req-cc47cd13-9595-40ad-a0c8-afcfcfe7f019 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Task: {'id': task-5116618, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1060.237976] env[65726]: DEBUG nova.scheduler.client.report [None req-dc5e36ff-c9dd-4078-bec5-b01919099186 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1060.357278] env[65726]: DEBUG oslo_vmware.api [None req-5c22cc7c-7dcc-417f-8f16-1f19636dd1f0 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': task-5116616, 'name': PowerOnVM_Task, 'duration_secs': 0.642793} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1060.357656] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-5c22cc7c-7dcc-417f-8f16-1f19636dd1f0 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: 51f5337d-2e23-4b7c-b590-76be6bd82411] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1060.357907] env[65726]: INFO nova.compute.manager [None req-5c22cc7c-7dcc-417f-8f16-1f19636dd1f0 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: 51f5337d-2e23-4b7c-b590-76be6bd82411] Took 11.18 seconds to spawn the instance on the hypervisor. [ 1060.358157] env[65726]: DEBUG nova.compute.manager [None req-5c22cc7c-7dcc-417f-8f16-1f19636dd1f0 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: 51f5337d-2e23-4b7c-b590-76be6bd82411] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1060.359140] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89bb8720-22e2-4993-adb2-3a0995d07e36 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.381086] env[65726]: DEBUG nova.compute.manager [None req-42911402-8543-4c4b-aab9-86db149010d2 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 5cdf1d61-b817-4986-a48c-804a2374497a] Start spawning the instance on the hypervisor. {{(pid=65726) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1060.384672] env[65726]: DEBUG oslo_concurrency.lockutils [None req-690d3db3-4b44-4e7c-bc5b-a409b814ad64 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Lock "f947529e-c930-4420-91a7-1e677b5ac2f1" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.100s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1060.414177] env[65726]: DEBUG nova.virt.hardware [None req-42911402-8543-4c4b-aab9-86db149010d2 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T19:28:34Z,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1539f06fa3534e90acbc3a60c4b8bd3f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T19:28:35Z,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1060.414177] env[65726]: DEBUG nova.virt.hardware [None req-42911402-8543-4c4b-aab9-86db149010d2 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1060.414177] env[65726]: DEBUG nova.virt.hardware [None req-42911402-8543-4c4b-aab9-86db149010d2 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1060.414738] env[65726]: DEBUG nova.virt.hardware [None req-42911402-8543-4c4b-aab9-86db149010d2 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1060.415068] env[65726]: DEBUG nova.virt.hardware [None req-42911402-8543-4c4b-aab9-86db149010d2 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1060.415450] env[65726]: DEBUG nova.virt.hardware [None req-42911402-8543-4c4b-aab9-86db149010d2 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1060.417231] env[65726]: DEBUG nova.virt.hardware [None req-42911402-8543-4c4b-aab9-86db149010d2 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1060.417231] env[65726]: DEBUG nova.virt.hardware [None req-42911402-8543-4c4b-aab9-86db149010d2 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1060.417231] env[65726]: DEBUG nova.virt.hardware [None req-42911402-8543-4c4b-aab9-86db149010d2 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1060.417231] env[65726]: DEBUG nova.virt.hardware [None req-42911402-8543-4c4b-aab9-86db149010d2 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1060.417231] env[65726]: DEBUG nova.virt.hardware [None req-42911402-8543-4c4b-aab9-86db149010d2 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1060.417897] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dce72692-040b-4a19-b166-ec50028cb312 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.430062] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-516f4709-daf6-4d37-811d-5d1c3983db0f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.568895] env[65726]: DEBUG oslo_vmware.api [None req-cc47cd13-9595-40ad-a0c8-afcfcfe7f019 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Task: {'id': task-5116618, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1060.744075] env[65726]: DEBUG oslo_concurrency.lockutils [None req-dc5e36ff-c9dd-4078-bec5-b01919099186 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.401s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1060.744275] env[65726]: INFO nova.compute.manager [None req-dc5e36ff-c9dd-4078-bec5-b01919099186 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: a56be6fa-6027-46ee-9aa8-332fe4222d68] Migrating [ 1060.751086] env[65726]: DEBUG oslo_concurrency.lockutils [None req-03150333-3109-4819-9fe0-28ed6751ecba tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 9.054s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1060.751328] env[65726]: DEBUG nova.objects.instance [None req-03150333-3109-4819-9fe0-28ed6751ecba tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Lazy-loading 'resources' on Instance uuid fb8eac12-aedd-426f-9dd3-6c52f9d4c71d {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1060.882102] env[65726]: INFO nova.compute.manager [None req-5c22cc7c-7dcc-417f-8f16-1f19636dd1f0 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: 51f5337d-2e23-4b7c-b590-76be6bd82411] Took 21.23 seconds to build instance. [ 1060.975757] env[65726]: DEBUG nova.network.neutron [None req-42911402-8543-4c4b-aab9-86db149010d2 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 5cdf1d61-b817-4986-a48c-804a2374497a] Successfully updated port: 94113d1c-2016-4a98-b42a-6ae90eee0b24 {{(pid=65726) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 1061.000121] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-f263eae6-96b2-45f2-b03a-d97fe5d0f1d8 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] [instance: 6fd70b4a-63e4-4258-9cff-f3b582500b9f] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1061.000121] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8fa7fc49-2693-4c57-a269-76e0676f794c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.007185] env[65726]: DEBUG oslo_vmware.api [None req-f263eae6-96b2-45f2-b03a-d97fe5d0f1d8 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Waiting for the task: (returnval){ [ 1061.007185] env[65726]: value = "task-5116619" [ 1061.007185] env[65726]: _type = "Task" [ 1061.007185] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1061.019183] env[65726]: DEBUG oslo_vmware.api [None req-f263eae6-96b2-45f2-b03a-d97fe5d0f1d8 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Task: {'id': task-5116619, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1061.072696] env[65726]: DEBUG oslo_vmware.api [None req-cc47cd13-9595-40ad-a0c8-afcfcfe7f019 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Task: {'id': task-5116618, 'name': ReconfigVM_Task, 'duration_secs': 0.746964} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1061.073120] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-cc47cd13-9595-40ad-a0c8-afcfcfe7f019 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] [instance: 48923e3d-b00d-4034-bacf-7be82009fb08] Reconfigured VM instance instance-0000005e to attach disk [datastore2] 48923e3d-b00d-4034-bacf-7be82009fb08/48923e3d-b00d-4034-bacf-7be82009fb08.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1061.073933] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2a6ed0f3-a37e-4998-8ab3-b7d68dce343b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.083219] env[65726]: DEBUG oslo_vmware.api [None req-cc47cd13-9595-40ad-a0c8-afcfcfe7f019 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Waiting for the task: (returnval){ [ 1061.083219] env[65726]: value = "task-5116620" [ 1061.083219] env[65726]: _type = "Task" [ 1061.083219] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1061.093859] env[65726]: DEBUG oslo_vmware.api [None req-cc47cd13-9595-40ad-a0c8-afcfcfe7f019 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Task: {'id': task-5116620, 'name': Rename_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1061.096554] env[65726]: DEBUG nova.compute.manager [req-8a38f678-b1eb-44ce-a0b6-27fccef96a8a req-f9e78b09-c6ec-435f-bc68-ec22e156a752 service nova] [instance: 5cdf1d61-b817-4986-a48c-804a2374497a] Received event network-vif-plugged-94113d1c-2016-4a98-b42a-6ae90eee0b24 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1061.097031] env[65726]: DEBUG oslo_concurrency.lockutils [req-8a38f678-b1eb-44ce-a0b6-27fccef96a8a req-f9e78b09-c6ec-435f-bc68-ec22e156a752 service nova] Acquiring lock "5cdf1d61-b817-4986-a48c-804a2374497a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1061.097887] env[65726]: DEBUG oslo_concurrency.lockutils [req-8a38f678-b1eb-44ce-a0b6-27fccef96a8a req-f9e78b09-c6ec-435f-bc68-ec22e156a752 service nova] Lock "5cdf1d61-b817-4986-a48c-804a2374497a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1061.097887] env[65726]: DEBUG oslo_concurrency.lockutils [req-8a38f678-b1eb-44ce-a0b6-27fccef96a8a req-f9e78b09-c6ec-435f-bc68-ec22e156a752 service nova] Lock "5cdf1d61-b817-4986-a48c-804a2374497a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1061.097887] env[65726]: DEBUG nova.compute.manager [req-8a38f678-b1eb-44ce-a0b6-27fccef96a8a req-f9e78b09-c6ec-435f-bc68-ec22e156a752 service nova] [instance: 5cdf1d61-b817-4986-a48c-804a2374497a] No waiting events found dispatching network-vif-plugged-94113d1c-2016-4a98-b42a-6ae90eee0b24 {{(pid=65726) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1061.098050] env[65726]: WARNING nova.compute.manager [req-8a38f678-b1eb-44ce-a0b6-27fccef96a8a req-f9e78b09-c6ec-435f-bc68-ec22e156a752 service nova] [instance: 5cdf1d61-b817-4986-a48c-804a2374497a] Received unexpected event network-vif-plugged-94113d1c-2016-4a98-b42a-6ae90eee0b24 for instance with vm_state building and task_state spawning. [ 1061.118106] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5ecca65c-f696-44b9-8b91-de86779bab97 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Acquiring lock "f947529e-c930-4420-91a7-1e677b5ac2f1" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1061.118106] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5ecca65c-f696-44b9-8b91-de86779bab97 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Lock "f947529e-c930-4420-91a7-1e677b5ac2f1" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1061.118106] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5ecca65c-f696-44b9-8b91-de86779bab97 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Acquiring lock "f947529e-c930-4420-91a7-1e677b5ac2f1-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1061.118304] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5ecca65c-f696-44b9-8b91-de86779bab97 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Lock "f947529e-c930-4420-91a7-1e677b5ac2f1-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1061.118427] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5ecca65c-f696-44b9-8b91-de86779bab97 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Lock "f947529e-c930-4420-91a7-1e677b5ac2f1-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1061.121082] env[65726]: INFO nova.compute.manager [None req-5ecca65c-f696-44b9-8b91-de86779bab97 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: f947529e-c930-4420-91a7-1e677b5ac2f1] Terminating instance [ 1061.265645] env[65726]: DEBUG oslo_concurrency.lockutils [None req-dc5e36ff-c9dd-4078-bec5-b01919099186 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Acquiring lock "refresh_cache-a56be6fa-6027-46ee-9aa8-332fe4222d68" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1061.265857] env[65726]: DEBUG oslo_concurrency.lockutils [None req-dc5e36ff-c9dd-4078-bec5-b01919099186 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Acquired lock "refresh_cache-a56be6fa-6027-46ee-9aa8-332fe4222d68" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1061.266074] env[65726]: DEBUG nova.network.neutron [None req-dc5e36ff-c9dd-4078-bec5-b01919099186 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: a56be6fa-6027-46ee-9aa8-332fe4222d68] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1061.386380] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5c22cc7c-7dcc-417f-8f16-1f19636dd1f0 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Lock "51f5337d-2e23-4b7c-b590-76be6bd82411" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 22.747s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1061.479495] env[65726]: DEBUG oslo_concurrency.lockutils [None req-42911402-8543-4c4b-aab9-86db149010d2 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Acquiring lock "refresh_cache-5cdf1d61-b817-4986-a48c-804a2374497a" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1061.479684] env[65726]: DEBUG oslo_concurrency.lockutils [None req-42911402-8543-4c4b-aab9-86db149010d2 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Acquired lock "refresh_cache-5cdf1d61-b817-4986-a48c-804a2374497a" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1061.479870] env[65726]: DEBUG nova.network.neutron [None req-42911402-8543-4c4b-aab9-86db149010d2 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 5cdf1d61-b817-4986-a48c-804a2374497a] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1061.498670] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71cdf421-41be-4fbf-8a20-eeecf1b8c7d1 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.507305] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5887d449-3ec2-4275-aadc-d134dae7fa7f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.521582] env[65726]: DEBUG oslo_vmware.api [None req-f263eae6-96b2-45f2-b03a-d97fe5d0f1d8 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Task: {'id': task-5116619, 'name': PowerOffVM_Task, 'duration_secs': 0.18829} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1061.522425] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-f263eae6-96b2-45f2-b03a-d97fe5d0f1d8 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] [instance: 6fd70b4a-63e4-4258-9cff-f3b582500b9f] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1061.547799] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8330c70-ea5e-41de-b539-e6c19f160005 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.551996] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1482212-42c1-4732-9b28-1d08b3b605a9 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.574397] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b0c2486-2ea9-4f1c-b460-b0adf482028c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.578978] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1003aec4-2813-4ed4-9124-f149b710afea {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.597369] env[65726]: DEBUG nova.compute.provider_tree [None req-03150333-3109-4819-9fe0-28ed6751ecba tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1061.607436] env[65726]: DEBUG oslo_vmware.api [None req-cc47cd13-9595-40ad-a0c8-afcfcfe7f019 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Task: {'id': task-5116620, 'name': Rename_Task, 'duration_secs': 0.16804} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1061.607692] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-cc47cd13-9595-40ad-a0c8-afcfcfe7f019 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] [instance: 48923e3d-b00d-4034-bacf-7be82009fb08] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1061.607944] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6669bf75-99e8-4d44-b17a-bc75cbd7da62 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.614850] env[65726]: DEBUG oslo_vmware.api [None req-cc47cd13-9595-40ad-a0c8-afcfcfe7f019 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Waiting for the task: (returnval){ [ 1061.614850] env[65726]: value = "task-5116621" [ 1061.614850] env[65726]: _type = "Task" [ 1061.614850] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1061.625125] env[65726]: DEBUG nova.compute.manager [None req-5ecca65c-f696-44b9-8b91-de86779bab97 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: f947529e-c930-4420-91a7-1e677b5ac2f1] Start destroying the instance on the hypervisor. {{(pid=65726) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1061.625346] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-5ecca65c-f696-44b9-8b91-de86779bab97 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: f947529e-c930-4420-91a7-1e677b5ac2f1] Destroying instance {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1061.625629] env[65726]: DEBUG oslo_vmware.api [None req-cc47cd13-9595-40ad-a0c8-afcfcfe7f019 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Task: {'id': task-5116621, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1061.626378] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5698dad7-e294-43fc-ae66-90ac44447198 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.634700] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-5ecca65c-f696-44b9-8b91-de86779bab97 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: f947529e-c930-4420-91a7-1e677b5ac2f1] Unregistering the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1061.634913] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-99179a60-9979-46e3-b73d-3bf2354ea847 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.728850] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-5ecca65c-f696-44b9-8b91-de86779bab97 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: f947529e-c930-4420-91a7-1e677b5ac2f1] Unregistered the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1061.731035] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-5ecca65c-f696-44b9-8b91-de86779bab97 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: f947529e-c930-4420-91a7-1e677b5ac2f1] Deleting contents of the VM from datastore datastore2 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1061.731397] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-5ecca65c-f696-44b9-8b91-de86779bab97 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Deleting the datastore file [datastore2] f947529e-c930-4420-91a7-1e677b5ac2f1 {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1061.731983] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5c330239-719e-42cb-8858-3602c693c77c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.742873] env[65726]: DEBUG oslo_vmware.api [None req-5ecca65c-f696-44b9-8b91-de86779bab97 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Waiting for the task: (returnval){ [ 1061.742873] env[65726]: value = "task-5116623" [ 1061.742873] env[65726]: _type = "Task" [ 1061.742873] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1061.754270] env[65726]: DEBUG oslo_vmware.api [None req-5ecca65c-f696-44b9-8b91-de86779bab97 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': task-5116623, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1061.769611] env[65726]: WARNING openstack [None req-dc5e36ff-c9dd-4078-bec5-b01919099186 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1061.770054] env[65726]: WARNING openstack [None req-dc5e36ff-c9dd-4078-bec5-b01919099186 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1061.987020] env[65726]: WARNING openstack [None req-42911402-8543-4c4b-aab9-86db149010d2 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1061.987020] env[65726]: WARNING openstack [None req-42911402-8543-4c4b-aab9-86db149010d2 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1061.990907] env[65726]: DEBUG nova.network.neutron [None req-42911402-8543-4c4b-aab9-86db149010d2 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 5cdf1d61-b817-4986-a48c-804a2374497a] Instance cache missing network info. {{(pid=65726) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 1062.085485] env[65726]: WARNING openstack [None req-dc5e36ff-c9dd-4078-bec5-b01919099186 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1062.086085] env[65726]: WARNING openstack [None req-dc5e36ff-c9dd-4078-bec5-b01919099186 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1062.102490] env[65726]: DEBUG nova.scheduler.client.report [None req-03150333-3109-4819-9fe0-28ed6751ecba tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1062.106739] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-f263eae6-96b2-45f2-b03a-d97fe5d0f1d8 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] [instance: 6fd70b4a-63e4-4258-9cff-f3b582500b9f] Creating Snapshot of the VM instance {{(pid=65726) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1062.107876] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-b0bbab1e-8651-4b97-8edb-c7f3ca172de0 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.118158] env[65726]: DEBUG oslo_vmware.api [None req-f263eae6-96b2-45f2-b03a-d97fe5d0f1d8 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Waiting for the task: (returnval){ [ 1062.118158] env[65726]: value = "task-5116624" [ 1062.118158] env[65726]: _type = "Task" [ 1062.118158] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1062.134643] env[65726]: DEBUG oslo_vmware.api [None req-f263eae6-96b2-45f2-b03a-d97fe5d0f1d8 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Task: {'id': task-5116624, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1062.139895] env[65726]: DEBUG oslo_vmware.api [None req-cc47cd13-9595-40ad-a0c8-afcfcfe7f019 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Task: {'id': task-5116621, 'name': PowerOnVM_Task} progress is 96%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1062.142351] env[65726]: WARNING openstack [None req-42911402-8543-4c4b-aab9-86db149010d2 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1062.142904] env[65726]: WARNING openstack [None req-42911402-8543-4c4b-aab9-86db149010d2 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1062.256210] env[65726]: DEBUG oslo_vmware.api [None req-5ecca65c-f696-44b9-8b91-de86779bab97 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': task-5116623, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.193745} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1062.256210] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-5ecca65c-f696-44b9-8b91-de86779bab97 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Deleted the datastore file {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1062.256210] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-5ecca65c-f696-44b9-8b91-de86779bab97 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: f947529e-c930-4420-91a7-1e677b5ac2f1] Deleted contents of the VM from datastore datastore2 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1062.256210] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-5ecca65c-f696-44b9-8b91-de86779bab97 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: f947529e-c930-4420-91a7-1e677b5ac2f1] Instance destroyed {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1062.256210] env[65726]: INFO nova.compute.manager [None req-5ecca65c-f696-44b9-8b91-de86779bab97 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: f947529e-c930-4420-91a7-1e677b5ac2f1] Took 0.63 seconds to destroy the instance on the hypervisor. [ 1062.256210] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-5ecca65c-f696-44b9-8b91-de86779bab97 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1062.256210] env[65726]: DEBUG nova.compute.manager [-] [instance: f947529e-c930-4420-91a7-1e677b5ac2f1] Deallocating network for instance {{(pid=65726) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1062.256210] env[65726]: DEBUG nova.network.neutron [-] [instance: f947529e-c930-4420-91a7-1e677b5ac2f1] deallocate_for_instance() {{(pid=65726) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 1062.256210] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1062.256210] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1062.290976] env[65726]: WARNING openstack [None req-dc5e36ff-c9dd-4078-bec5-b01919099186 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1062.291827] env[65726]: WARNING openstack [None req-dc5e36ff-c9dd-4078-bec5-b01919099186 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1062.599800] env[65726]: WARNING openstack [None req-42911402-8543-4c4b-aab9-86db149010d2 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1062.600547] env[65726]: WARNING openstack [None req-42911402-8543-4c4b-aab9-86db149010d2 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1062.611185] env[65726]: DEBUG oslo_concurrency.lockutils [None req-03150333-3109-4819-9fe0-28ed6751ecba tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.860s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1062.614229] env[65726]: DEBUG oslo_concurrency.lockutils [None req-80fb9493-d54c-4f0a-a5fd-0197feeffb7b tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 10.553s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1062.616406] env[65726]: INFO nova.compute.claims [None req-80fb9493-d54c-4f0a-a5fd-0197feeffb7b tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] [instance: 3b4d9461-747a-461c-b231-a0de02f0cb83] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1062.633387] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1062.633686] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1062.648374] env[65726]: INFO nova.scheduler.client.report [None req-03150333-3109-4819-9fe0-28ed6751ecba tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Deleted allocations for instance fb8eac12-aedd-426f-9dd3-6c52f9d4c71d [ 1062.654967] env[65726]: DEBUG oslo_vmware.api [None req-cc47cd13-9595-40ad-a0c8-afcfcfe7f019 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Task: {'id': task-5116621, 'name': PowerOnVM_Task, 'duration_secs': 0.878191} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1062.658540] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-cc47cd13-9595-40ad-a0c8-afcfcfe7f019 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] [instance: 48923e3d-b00d-4034-bacf-7be82009fb08] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1062.659053] env[65726]: INFO nova.compute.manager [None req-cc47cd13-9595-40ad-a0c8-afcfcfe7f019 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] [instance: 48923e3d-b00d-4034-bacf-7be82009fb08] Took 8.56 seconds to spawn the instance on the hypervisor. [ 1062.659153] env[65726]: DEBUG nova.compute.manager [None req-cc47cd13-9595-40ad-a0c8-afcfcfe7f019 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] [instance: 48923e3d-b00d-4034-bacf-7be82009fb08] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1062.663759] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d09780f0-361a-4581-a4b4-80d29a2c0281 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.667747] env[65726]: DEBUG oslo_vmware.api [None req-f263eae6-96b2-45f2-b03a-d97fe5d0f1d8 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Task: {'id': task-5116624, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1062.823839] env[65726]: DEBUG nova.network.neutron [None req-dc5e36ff-c9dd-4078-bec5-b01919099186 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: a56be6fa-6027-46ee-9aa8-332fe4222d68] Updating instance_info_cache with network_info: [{"id": "4e2dfbcd-41c3-4aaa-a2ec-50c855c12307", "address": "fa:16:3e:1d:82:59", "network": {"id": "0d0f561d-ac41-4ef5-9926-c707409799f7", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1134321217-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.160", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4eece77569624f90bf64e5c51974173f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2ce62383-8e84-4e26-955b-74c11392f4c9", "external-id": "nsx-vlan-transportzone-215", "segmentation_id": 215, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4e2dfbcd-41", "ovs_interfaceid": "4e2dfbcd-41c3-4aaa-a2ec-50c855c12307", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1062.945121] env[65726]: DEBUG nova.network.neutron [None req-42911402-8543-4c4b-aab9-86db149010d2 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 5cdf1d61-b817-4986-a48c-804a2374497a] Updating instance_info_cache with network_info: [{"id": "94113d1c-2016-4a98-b42a-6ae90eee0b24", "address": "fa:16:3e:4d:82:49", "network": {"id": "9dd55fb6-0dd2-484d-b94d-fbdccb958c79", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1005598846-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f93d852e2c904f42981cbfff882558e9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "74f30339-6421-4654-bddb-81d7f34db9d7", "external-id": "nsx-vlan-transportzone-899", "segmentation_id": 899, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap94113d1c-20", "ovs_interfaceid": "94113d1c-2016-4a98-b42a-6ae90eee0b24", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1063.140319] env[65726]: DEBUG oslo_vmware.api [None req-f263eae6-96b2-45f2-b03a-d97fe5d0f1d8 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Task: {'id': task-5116624, 'name': CreateSnapshot_Task, 'duration_secs': 0.814917} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1063.140570] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-f263eae6-96b2-45f2-b03a-d97fe5d0f1d8 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] [instance: 6fd70b4a-63e4-4258-9cff-f3b582500b9f] Created Snapshot of the VM instance {{(pid=65726) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1063.141293] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-381e36c0-fa0c-4d3b-939c-c7e2bdb4b9f3 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.164738] env[65726]: DEBUG oslo_concurrency.lockutils [None req-03150333-3109-4819-9fe0-28ed6751ecba tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Lock "fb8eac12-aedd-426f-9dd3-6c52f9d4c71d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 15.697s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1063.188380] env[65726]: INFO nova.compute.manager [None req-cc47cd13-9595-40ad-a0c8-afcfcfe7f019 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] [instance: 48923e3d-b00d-4034-bacf-7be82009fb08] Took 16.37 seconds to build instance. [ 1063.328408] env[65726]: DEBUG oslo_concurrency.lockutils [None req-dc5e36ff-c9dd-4078-bec5-b01919099186 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Releasing lock "refresh_cache-a56be6fa-6027-46ee-9aa8-332fe4222d68" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1063.450644] env[65726]: DEBUG oslo_concurrency.lockutils [None req-42911402-8543-4c4b-aab9-86db149010d2 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Releasing lock "refresh_cache-5cdf1d61-b817-4986-a48c-804a2374497a" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1063.450644] env[65726]: DEBUG nova.compute.manager [None req-42911402-8543-4c4b-aab9-86db149010d2 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 5cdf1d61-b817-4986-a48c-804a2374497a] Instance network_info: |[{"id": "94113d1c-2016-4a98-b42a-6ae90eee0b24", "address": "fa:16:3e:4d:82:49", "network": {"id": "9dd55fb6-0dd2-484d-b94d-fbdccb958c79", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1005598846-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f93d852e2c904f42981cbfff882558e9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "74f30339-6421-4654-bddb-81d7f34db9d7", "external-id": "nsx-vlan-transportzone-899", "segmentation_id": 899, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap94113d1c-20", "ovs_interfaceid": "94113d1c-2016-4a98-b42a-6ae90eee0b24", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1063.450644] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-42911402-8543-4c4b-aab9-86db149010d2 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 5cdf1d61-b817-4986-a48c-804a2374497a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:4d:82:49', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '74f30339-6421-4654-bddb-81d7f34db9d7', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '94113d1c-2016-4a98-b42a-6ae90eee0b24', 'vif_model': 'vmxnet3'}] {{(pid=65726) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1063.459686] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-42911402-8543-4c4b-aab9-86db149010d2 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1063.463022] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5cdf1d61-b817-4986-a48c-804a2374497a] Creating VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1063.463022] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ac5d70fa-2a12-4615-9569-d2e6af48e655 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.482180] env[65726]: DEBUG nova.compute.manager [req-5fae9768-f36a-41c8-b244-0788a001723c req-ff91bb2f-6d27-4dd1-9655-9b949d9181da service nova] [instance: f947529e-c930-4420-91a7-1e677b5ac2f1] Received event network-vif-deleted-4521ce71-46b3-44a7-a0d8-6172da592893 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1063.482442] env[65726]: INFO nova.compute.manager [req-5fae9768-f36a-41c8-b244-0788a001723c req-ff91bb2f-6d27-4dd1-9655-9b949d9181da service nova] [instance: f947529e-c930-4420-91a7-1e677b5ac2f1] Neutron deleted interface 4521ce71-46b3-44a7-a0d8-6172da592893; detaching it from the instance and deleting it from the info cache [ 1063.482649] env[65726]: DEBUG nova.network.neutron [req-5fae9768-f36a-41c8-b244-0788a001723c req-ff91bb2f-6d27-4dd1-9655-9b949d9181da service nova] [instance: f947529e-c930-4420-91a7-1e677b5ac2f1] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1063.492341] env[65726]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1063.492341] env[65726]: value = "task-5116625" [ 1063.492341] env[65726]: _type = "Task" [ 1063.492341] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1063.498102] env[65726]: DEBUG nova.network.neutron [-] [instance: f947529e-c930-4420-91a7-1e677b5ac2f1] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1063.506679] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116625, 'name': CreateVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1063.557917] env[65726]: DEBUG nova.compute.manager [req-b6bd140f-5338-4661-948f-a8376c22ff3f req-250ce9de-3d1f-482a-94b8-337b92322431 service nova] [instance: 5cdf1d61-b817-4986-a48c-804a2374497a] Received event network-changed-94113d1c-2016-4a98-b42a-6ae90eee0b24 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1063.558214] env[65726]: DEBUG nova.compute.manager [req-b6bd140f-5338-4661-948f-a8376c22ff3f req-250ce9de-3d1f-482a-94b8-337b92322431 service nova] [instance: 5cdf1d61-b817-4986-a48c-804a2374497a] Refreshing instance network info cache due to event network-changed-94113d1c-2016-4a98-b42a-6ae90eee0b24. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 1063.559026] env[65726]: DEBUG oslo_concurrency.lockutils [req-b6bd140f-5338-4661-948f-a8376c22ff3f req-250ce9de-3d1f-482a-94b8-337b92322431 service nova] Acquiring lock "refresh_cache-5cdf1d61-b817-4986-a48c-804a2374497a" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1063.559026] env[65726]: DEBUG oslo_concurrency.lockutils [req-b6bd140f-5338-4661-948f-a8376c22ff3f req-250ce9de-3d1f-482a-94b8-337b92322431 service nova] Acquired lock "refresh_cache-5cdf1d61-b817-4986-a48c-804a2374497a" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1063.559026] env[65726]: DEBUG nova.network.neutron [req-b6bd140f-5338-4661-948f-a8376c22ff3f req-250ce9de-3d1f-482a-94b8-337b92322431 service nova] [instance: 5cdf1d61-b817-4986-a48c-804a2374497a] Refreshing network info cache for port 94113d1c-2016-4a98-b42a-6ae90eee0b24 {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 1063.665015] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-f263eae6-96b2-45f2-b03a-d97fe5d0f1d8 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] [instance: 6fd70b4a-63e4-4258-9cff-f3b582500b9f] Creating linked-clone VM from snapshot {{(pid=65726) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1063.668361] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-b37c217f-075d-437a-9a3e-47b1f52325c0 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.679247] env[65726]: DEBUG oslo_vmware.api [None req-f263eae6-96b2-45f2-b03a-d97fe5d0f1d8 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Waiting for the task: (returnval){ [ 1063.679247] env[65726]: value = "task-5116626" [ 1063.679247] env[65726]: _type = "Task" [ 1063.679247] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1063.690812] env[65726]: DEBUG oslo_concurrency.lockutils [None req-cc47cd13-9595-40ad-a0c8-afcfcfe7f019 tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Lock "48923e3d-b00d-4034-bacf-7be82009fb08" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.880s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1063.695638] env[65726]: DEBUG oslo_vmware.api [None req-f263eae6-96b2-45f2-b03a-d97fe5d0f1d8 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Task: {'id': task-5116626, 'name': CloneVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1063.768716] env[65726]: DEBUG oslo_concurrency.lockutils [None req-91ff63ad-fe79-48d0-901e-fc0ed403574a tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Acquiring lock "0d42fa75-20e9-4646-9b08-17015b7f068c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1063.768939] env[65726]: DEBUG oslo_concurrency.lockutils [None req-91ff63ad-fe79-48d0-901e-fc0ed403574a tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Lock "0d42fa75-20e9-4646-9b08-17015b7f068c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1063.941280] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03791f67-9105-4162-8b1b-4d06329235e8 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.952557] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b393d83c-058e-4025-97e3-bddfbfba52c6 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.992598] env[65726]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-35f5f411-04ab-49e8-b4bc-d1fbbc3a57b9 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.995512] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d6f17ed-dc96-4c19-85f7-3f1a048ad329 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.003455] env[65726]: INFO nova.compute.manager [-] [instance: f947529e-c930-4420-91a7-1e677b5ac2f1] Took 1.75 seconds to deallocate network for instance. [ 1064.020690] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33648b8a-69be-406b-ba37-f9b093e8663f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.032702] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb2168ad-6686-4db8-8376-6b87aa6038a8 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.037667] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116625, 'name': CreateVM_Task} progress is 99%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.052440] env[65726]: DEBUG nova.compute.provider_tree [None req-80fb9493-d54c-4f0a-a5fd-0197feeffb7b tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1064.068887] env[65726]: WARNING openstack [req-b6bd140f-5338-4661-948f-a8376c22ff3f req-250ce9de-3d1f-482a-94b8-337b92322431 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1064.069157] env[65726]: WARNING openstack [req-b6bd140f-5338-4661-948f-a8376c22ff3f req-250ce9de-3d1f-482a-94b8-337b92322431 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1064.076681] env[65726]: DEBUG nova.compute.manager [req-5fae9768-f36a-41c8-b244-0788a001723c req-ff91bb2f-6d27-4dd1-9655-9b949d9181da service nova] [instance: f947529e-c930-4420-91a7-1e677b5ac2f1] Detach interface failed, port_id=4521ce71-46b3-44a7-a0d8-6172da592893, reason: Instance f947529e-c930-4420-91a7-1e677b5ac2f1 could not be found. {{(pid=65726) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11669}} [ 1064.079745] env[65726]: DEBUG nova.scheduler.client.report [None req-80fb9493-d54c-4f0a-a5fd-0197feeffb7b tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1064.193935] env[65726]: DEBUG oslo_vmware.api [None req-f263eae6-96b2-45f2-b03a-d97fe5d0f1d8 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Task: {'id': task-5116626, 'name': CloneVM_Task} progress is 94%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.202450] env[65726]: WARNING openstack [req-b6bd140f-5338-4661-948f-a8376c22ff3f req-250ce9de-3d1f-482a-94b8-337b92322431 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1064.202949] env[65726]: WARNING openstack [req-b6bd140f-5338-4661-948f-a8376c22ff3f req-250ce9de-3d1f-482a-94b8-337b92322431 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1064.226571] env[65726]: DEBUG oslo_concurrency.lockutils [None req-ab8fab95-1b19-4328-a74a-7a9a0971be1c tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Acquiring lock "48923e3d-b00d-4034-bacf-7be82009fb08" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1064.226880] env[65726]: DEBUG oslo_concurrency.lockutils [None req-ab8fab95-1b19-4328-a74a-7a9a0971be1c tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Lock "48923e3d-b00d-4034-bacf-7be82009fb08" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1064.227105] env[65726]: DEBUG oslo_concurrency.lockutils [None req-ab8fab95-1b19-4328-a74a-7a9a0971be1c tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Acquiring lock "48923e3d-b00d-4034-bacf-7be82009fb08-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1064.227289] env[65726]: DEBUG oslo_concurrency.lockutils [None req-ab8fab95-1b19-4328-a74a-7a9a0971be1c tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Lock "48923e3d-b00d-4034-bacf-7be82009fb08-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1064.227449] env[65726]: DEBUG oslo_concurrency.lockutils [None req-ab8fab95-1b19-4328-a74a-7a9a0971be1c tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Lock "48923e3d-b00d-4034-bacf-7be82009fb08-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1064.230049] env[65726]: INFO nova.compute.manager [None req-ab8fab95-1b19-4328-a74a-7a9a0971be1c tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] [instance: 48923e3d-b00d-4034-bacf-7be82009fb08] Terminating instance [ 1064.270120] env[65726]: WARNING openstack [req-b6bd140f-5338-4661-948f-a8376c22ff3f req-250ce9de-3d1f-482a-94b8-337b92322431 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1064.270590] env[65726]: WARNING openstack [req-b6bd140f-5338-4661-948f-a8376c22ff3f req-250ce9de-3d1f-482a-94b8-337b92322431 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1064.278852] env[65726]: DEBUG nova.compute.manager [None req-91ff63ad-fe79-48d0-901e-fc0ed403574a tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: 0d42fa75-20e9-4646-9b08-17015b7f068c] Starting instance... {{(pid=65726) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 1064.418503] env[65726]: DEBUG nova.network.neutron [req-b6bd140f-5338-4661-948f-a8376c22ff3f req-250ce9de-3d1f-482a-94b8-337b92322431 service nova] [instance: 5cdf1d61-b817-4986-a48c-804a2374497a] Updated VIF entry in instance network info cache for port 94113d1c-2016-4a98-b42a-6ae90eee0b24. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 1064.418898] env[65726]: DEBUG nova.network.neutron [req-b6bd140f-5338-4661-948f-a8376c22ff3f req-250ce9de-3d1f-482a-94b8-337b92322431 service nova] [instance: 5cdf1d61-b817-4986-a48c-804a2374497a] Updating instance_info_cache with network_info: [{"id": "94113d1c-2016-4a98-b42a-6ae90eee0b24", "address": "fa:16:3e:4d:82:49", "network": {"id": "9dd55fb6-0dd2-484d-b94d-fbdccb958c79", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1005598846-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f93d852e2c904f42981cbfff882558e9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "74f30339-6421-4654-bddb-81d7f34db9d7", "external-id": "nsx-vlan-transportzone-899", "segmentation_id": 899, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap94113d1c-20", "ovs_interfaceid": "94113d1c-2016-4a98-b42a-6ae90eee0b24", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1064.506806] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116625, 'name': CreateVM_Task, 'duration_secs': 0.572481} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1064.507059] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5cdf1d61-b817-4986-a48c-804a2374497a] Created VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1064.507880] env[65726]: WARNING openstack [None req-42911402-8543-4c4b-aab9-86db149010d2 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1064.508476] env[65726]: WARNING openstack [None req-42911402-8543-4c4b-aab9-86db149010d2 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1064.514154] env[65726]: DEBUG oslo_concurrency.lockutils [None req-42911402-8543-4c4b-aab9-86db149010d2 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1064.514445] env[65726]: DEBUG oslo_concurrency.lockutils [None req-42911402-8543-4c4b-aab9-86db149010d2 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1064.514759] env[65726]: DEBUG oslo_concurrency.lockutils [None req-42911402-8543-4c4b-aab9-86db149010d2 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1064.515054] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-47da3204-50f3-4a64-a2dc-0ed9d7075b19 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.521227] env[65726]: DEBUG oslo_vmware.api [None req-42911402-8543-4c4b-aab9-86db149010d2 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Waiting for the task: (returnval){ [ 1064.521227] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52f131e8-4381-c71a-f2c3-dbe7a810b7a3" [ 1064.521227] env[65726]: _type = "Task" [ 1064.521227] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1064.530305] env[65726]: DEBUG oslo_vmware.api [None req-42911402-8543-4c4b-aab9-86db149010d2 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52f131e8-4381-c71a-f2c3-dbe7a810b7a3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.539936] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5ecca65c-f696-44b9-8b91-de86779bab97 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1064.584081] env[65726]: DEBUG oslo_concurrency.lockutils [None req-80fb9493-d54c-4f0a-a5fd-0197feeffb7b tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 1.970s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1064.584906] env[65726]: DEBUG nova.compute.manager [None req-80fb9493-d54c-4f0a-a5fd-0197feeffb7b tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] [instance: 3b4d9461-747a-461c-b231-a0de02f0cb83] Start building networks asynchronously for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1064.587636] env[65726]: DEBUG oslo_concurrency.lockutils [None req-f3c5543b-5b17-4006-8d18-e03fa5503534 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 8.570s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1064.587857] env[65726]: DEBUG nova.objects.instance [None req-f3c5543b-5b17-4006-8d18-e03fa5503534 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] Lazy-loading 'resources' on Instance uuid 2b8eb60b-dc9b-47a4-9ddc-205f2af69888 {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1064.691783] env[65726]: DEBUG oslo_vmware.api [None req-f263eae6-96b2-45f2-b03a-d97fe5d0f1d8 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Task: {'id': task-5116626, 'name': CloneVM_Task} progress is 94%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.734750] env[65726]: DEBUG nova.compute.manager [None req-ab8fab95-1b19-4328-a74a-7a9a0971be1c tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] [instance: 48923e3d-b00d-4034-bacf-7be82009fb08] Start destroying the instance on the hypervisor. {{(pid=65726) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1064.735420] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-ab8fab95-1b19-4328-a74a-7a9a0971be1c tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] [instance: 48923e3d-b00d-4034-bacf-7be82009fb08] Destroying instance {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1064.736773] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e61eef92-4cac-497c-87f6-a4fb437a3da1 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.749543] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-ab8fab95-1b19-4328-a74a-7a9a0971be1c tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] [instance: 48923e3d-b00d-4034-bacf-7be82009fb08] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1064.749865] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-95c87f58-1472-4fa6-bb35-c6d6de83360d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.764735] env[65726]: DEBUG oslo_concurrency.lockutils [None req-8c26e8da-5d87-42a2-a8c5-15e9c65a8b9f tempest-InstanceActionsV221TestJSON-647428531 tempest-InstanceActionsV221TestJSON-647428531-project-member] Acquiring lock "b0fec5ff-8c1e-41f9-b30a-6253a34baeec" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1064.764735] env[65726]: DEBUG oslo_concurrency.lockutils [None req-8c26e8da-5d87-42a2-a8c5-15e9c65a8b9f tempest-InstanceActionsV221TestJSON-647428531 tempest-InstanceActionsV221TestJSON-647428531-project-member] Lock "b0fec5ff-8c1e-41f9-b30a-6253a34baeec" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1064.766078] env[65726]: DEBUG oslo_vmware.api [None req-ab8fab95-1b19-4328-a74a-7a9a0971be1c tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Waiting for the task: (returnval){ [ 1064.766078] env[65726]: value = "task-5116627" [ 1064.766078] env[65726]: _type = "Task" [ 1064.766078] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1064.776340] env[65726]: DEBUG oslo_vmware.api [None req-ab8fab95-1b19-4328-a74a-7a9a0971be1c tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Task: {'id': task-5116627, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.803535] env[65726]: DEBUG oslo_concurrency.lockutils [None req-91ff63ad-fe79-48d0-901e-fc0ed403574a tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1064.850997] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48c8fa59-6fd5-4861-b967-c1010070fe18 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.885232] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-dc5e36ff-c9dd-4078-bec5-b01919099186 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: a56be6fa-6027-46ee-9aa8-332fe4222d68] Updating instance 'a56be6fa-6027-46ee-9aa8-332fe4222d68' progress to 0 {{(pid=65726) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1064.923536] env[65726]: DEBUG oslo_concurrency.lockutils [req-b6bd140f-5338-4661-948f-a8376c22ff3f req-250ce9de-3d1f-482a-94b8-337b92322431 service nova] Releasing lock "refresh_cache-5cdf1d61-b817-4986-a48c-804a2374497a" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1064.924029] env[65726]: DEBUG nova.compute.manager [req-b6bd140f-5338-4661-948f-a8376c22ff3f req-250ce9de-3d1f-482a-94b8-337b92322431 service nova] [instance: 51f5337d-2e23-4b7c-b590-76be6bd82411] Received event network-changed-7407bcd9-3202-4798-9eaa-252fb3fdb21f {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1064.924124] env[65726]: DEBUG nova.compute.manager [req-b6bd140f-5338-4661-948f-a8376c22ff3f req-250ce9de-3d1f-482a-94b8-337b92322431 service nova] [instance: 51f5337d-2e23-4b7c-b590-76be6bd82411] Refreshing instance network info cache due to event network-changed-7407bcd9-3202-4798-9eaa-252fb3fdb21f. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 1064.924272] env[65726]: DEBUG oslo_concurrency.lockutils [req-b6bd140f-5338-4661-948f-a8376c22ff3f req-250ce9de-3d1f-482a-94b8-337b92322431 service nova] Acquiring lock "refresh_cache-51f5337d-2e23-4b7c-b590-76be6bd82411" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1064.924408] env[65726]: DEBUG oslo_concurrency.lockutils [req-b6bd140f-5338-4661-948f-a8376c22ff3f req-250ce9de-3d1f-482a-94b8-337b92322431 service nova] Acquired lock "refresh_cache-51f5337d-2e23-4b7c-b590-76be6bd82411" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1064.924582] env[65726]: DEBUG nova.network.neutron [req-b6bd140f-5338-4661-948f-a8376c22ff3f req-250ce9de-3d1f-482a-94b8-337b92322431 service nova] [instance: 51f5337d-2e23-4b7c-b590-76be6bd82411] Refreshing network info cache for port 7407bcd9-3202-4798-9eaa-252fb3fdb21f {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 1065.034300] env[65726]: DEBUG oslo_vmware.api [None req-42911402-8543-4c4b-aab9-86db149010d2 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52f131e8-4381-c71a-f2c3-dbe7a810b7a3, 'name': SearchDatastore_Task, 'duration_secs': 0.021869} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1065.034723] env[65726]: DEBUG oslo_concurrency.lockutils [None req-42911402-8543-4c4b-aab9-86db149010d2 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1065.034993] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-42911402-8543-4c4b-aab9-86db149010d2 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 5cdf1d61-b817-4986-a48c-804a2374497a] Processing image b52362a3-ee8a-4cbf-b06f-513b0cc8f95c {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1065.035316] env[65726]: DEBUG oslo_concurrency.lockutils [None req-42911402-8543-4c4b-aab9-86db149010d2 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1065.035522] env[65726]: DEBUG oslo_concurrency.lockutils [None req-42911402-8543-4c4b-aab9-86db149010d2 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1065.035744] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-42911402-8543-4c4b-aab9-86db149010d2 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1065.036078] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0d533c2a-ae07-416f-95f3-4f6732dee8f9 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.060538] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-42911402-8543-4c4b-aab9-86db149010d2 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1065.060801] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-42911402-8543-4c4b-aab9-86db149010d2 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65726) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1065.061665] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f5a08131-7cd6-4639-a3dd-05f18931521a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.069157] env[65726]: DEBUG oslo_vmware.api [None req-42911402-8543-4c4b-aab9-86db149010d2 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Waiting for the task: (returnval){ [ 1065.069157] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]522651f6-2415-56f7-6238-26a0ad94b8ce" [ 1065.069157] env[65726]: _type = "Task" [ 1065.069157] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1065.078899] env[65726]: DEBUG oslo_vmware.api [None req-42911402-8543-4c4b-aab9-86db149010d2 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]522651f6-2415-56f7-6238-26a0ad94b8ce, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.090779] env[65726]: DEBUG nova.compute.utils [None req-80fb9493-d54c-4f0a-a5fd-0197feeffb7b tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Using /dev/sd instead of None {{(pid=65726) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1065.092422] env[65726]: DEBUG nova.compute.manager [None req-80fb9493-d54c-4f0a-a5fd-0197feeffb7b tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] [instance: 3b4d9461-747a-461c-b231-a0de02f0cb83] Allocating IP information in the background. {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1065.092651] env[65726]: DEBUG nova.network.neutron [None req-80fb9493-d54c-4f0a-a5fd-0197feeffb7b tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] [instance: 3b4d9461-747a-461c-b231-a0de02f0cb83] allocate_for_instance() {{(pid=65726) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 1065.093278] env[65726]: WARNING neutronclient.v2_0.client [None req-80fb9493-d54c-4f0a-a5fd-0197feeffb7b tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1065.093409] env[65726]: WARNING neutronclient.v2_0.client [None req-80fb9493-d54c-4f0a-a5fd-0197feeffb7b tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1065.093993] env[65726]: WARNING openstack [None req-80fb9493-d54c-4f0a-a5fd-0197feeffb7b tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1065.095033] env[65726]: WARNING openstack [None req-80fb9493-d54c-4f0a-a5fd-0197feeffb7b tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1065.149854] env[65726]: DEBUG nova.policy [None req-80fb9493-d54c-4f0a-a5fd-0197feeffb7b tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b776acdc25b648d2ba0575d7a5d5fec0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'eb811e38506b4297bf01bc5f2e098370', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65726) authorize /opt/stack/nova/nova/policy.py:192}} [ 1065.195222] env[65726]: DEBUG oslo_vmware.api [None req-f263eae6-96b2-45f2-b03a-d97fe5d0f1d8 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Task: {'id': task-5116626, 'name': CloneVM_Task} progress is 95%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.267759] env[65726]: DEBUG nova.compute.manager [None req-8c26e8da-5d87-42a2-a8c5-15e9c65a8b9f tempest-InstanceActionsV221TestJSON-647428531 tempest-InstanceActionsV221TestJSON-647428531-project-member] [instance: b0fec5ff-8c1e-41f9-b30a-6253a34baeec] Starting instance... {{(pid=65726) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 1065.285663] env[65726]: DEBUG oslo_vmware.api [None req-ab8fab95-1b19-4328-a74a-7a9a0971be1c tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Task: {'id': task-5116627, 'name': PowerOffVM_Task, 'duration_secs': 0.46418} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1065.286093] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-ab8fab95-1b19-4328-a74a-7a9a0971be1c tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] [instance: 48923e3d-b00d-4034-bacf-7be82009fb08] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1065.286572] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-ab8fab95-1b19-4328-a74a-7a9a0971be1c tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] [instance: 48923e3d-b00d-4034-bacf-7be82009fb08] Unregistering the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1065.286665] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-bb0e9a78-20f0-4e2e-8642-6ee4827346f7 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.363349] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-ab8fab95-1b19-4328-a74a-7a9a0971be1c tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] [instance: 48923e3d-b00d-4034-bacf-7be82009fb08] Unregistered the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1065.363667] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-ab8fab95-1b19-4328-a74a-7a9a0971be1c tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] [instance: 48923e3d-b00d-4034-bacf-7be82009fb08] Deleting contents of the VM from datastore datastore2 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1065.363918] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-ab8fab95-1b19-4328-a74a-7a9a0971be1c tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Deleting the datastore file [datastore2] 48923e3d-b00d-4034-bacf-7be82009fb08 {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1065.364241] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0ed919cb-d9e3-4f57-b59c-fb41f3703329 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.371665] env[65726]: DEBUG oslo_vmware.api [None req-ab8fab95-1b19-4328-a74a-7a9a0971be1c tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Waiting for the task: (returnval){ [ 1065.371665] env[65726]: value = "task-5116629" [ 1065.371665] env[65726]: _type = "Task" [ 1065.371665] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1065.384828] env[65726]: DEBUG oslo_vmware.api [None req-ab8fab95-1b19-4328-a74a-7a9a0971be1c tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Task: {'id': task-5116629, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.386980] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c71f5ce-1ac7-4ab6-9b91-7092bd1460bf {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.392602] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-dc5e36ff-c9dd-4078-bec5-b01919099186 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: a56be6fa-6027-46ee-9aa8-332fe4222d68] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1065.394898] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c18e20b2-76fa-4e53-9b1a-50e982fdb48e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.397631] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0fe3656-c3a4-4332-8925-523ca31797af {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.406828] env[65726]: DEBUG oslo_vmware.api [None req-dc5e36ff-c9dd-4078-bec5-b01919099186 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Waiting for the task: (returnval){ [ 1065.406828] env[65726]: value = "task-5116630" [ 1065.406828] env[65726]: _type = "Task" [ 1065.406828] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1065.437179] env[65726]: WARNING openstack [req-b6bd140f-5338-4661-948f-a8376c22ff3f req-250ce9de-3d1f-482a-94b8-337b92322431 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1065.438180] env[65726]: WARNING openstack [req-b6bd140f-5338-4661-948f-a8376c22ff3f req-250ce9de-3d1f-482a-94b8-337b92322431 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1065.452227] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe29cbd7-9a83-4c10-b608-ebda92231de1 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.462339] env[65726]: DEBUG oslo_vmware.api [None req-dc5e36ff-c9dd-4078-bec5-b01919099186 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5116630, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.466040] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bdf38e7b-ef59-4ba7-8d53-31ea5ff2e576 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.485040] env[65726]: DEBUG nova.compute.provider_tree [None req-f3c5543b-5b17-4006-8d18-e03fa5503534 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1065.491679] env[65726]: DEBUG nova.network.neutron [None req-80fb9493-d54c-4f0a-a5fd-0197feeffb7b tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] [instance: 3b4d9461-747a-461c-b231-a0de02f0cb83] Successfully created port: 20c6b53e-2740-46bc-a94f-59a15f8b2c61 {{(pid=65726) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1065.500420] env[65726]: DEBUG nova.scheduler.client.report [None req-f3c5543b-5b17-4006-8d18-e03fa5503534 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1065.580533] env[65726]: DEBUG oslo_vmware.api [None req-42911402-8543-4c4b-aab9-86db149010d2 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]522651f6-2415-56f7-6238-26a0ad94b8ce, 'name': SearchDatastore_Task, 'duration_secs': 0.030134} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1065.581027] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d167170b-33e3-440e-a667-f0e4490b296a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.587690] env[65726]: DEBUG oslo_vmware.api [None req-42911402-8543-4c4b-aab9-86db149010d2 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Waiting for the task: (returnval){ [ 1065.587690] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52b4d5f1-68c3-c501-36e9-0f59d3817375" [ 1065.587690] env[65726]: _type = "Task" [ 1065.587690] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1065.595691] env[65726]: DEBUG oslo_vmware.api [None req-42911402-8543-4c4b-aab9-86db149010d2 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52b4d5f1-68c3-c501-36e9-0f59d3817375, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.604425] env[65726]: DEBUG nova.compute.manager [None req-80fb9493-d54c-4f0a-a5fd-0197feeffb7b tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] [instance: 3b4d9461-747a-461c-b231-a0de02f0cb83] Start building block device mappings for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1065.640333] env[65726]: WARNING openstack [req-b6bd140f-5338-4661-948f-a8376c22ff3f req-250ce9de-3d1f-482a-94b8-337b92322431 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1065.640825] env[65726]: WARNING openstack [req-b6bd140f-5338-4661-948f-a8376c22ff3f req-250ce9de-3d1f-482a-94b8-337b92322431 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1065.694382] env[65726]: DEBUG oslo_vmware.api [None req-f263eae6-96b2-45f2-b03a-d97fe5d0f1d8 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Task: {'id': task-5116626, 'name': CloneVM_Task, 'duration_secs': 1.739318} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1065.694679] env[65726]: INFO nova.virt.vmwareapi.vmops [None req-f263eae6-96b2-45f2-b03a-d97fe5d0f1d8 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] [instance: 6fd70b4a-63e4-4258-9cff-f3b582500b9f] Created linked-clone VM from snapshot [ 1065.695486] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f693be52-1b3e-4ed4-927e-a860b02ce225 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.705240] env[65726]: DEBUG nova.virt.vmwareapi.images [None req-f263eae6-96b2-45f2-b03a-d97fe5d0f1d8 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] [instance: 6fd70b4a-63e4-4258-9cff-f3b582500b9f] Uploading image eb421501-6112-4a68-9c48-c82505e774d2 {{(pid=65726) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 1065.745189] env[65726]: DEBUG oslo_vmware.rw_handles [None req-f263eae6-96b2-45f2-b03a-d97fe5d0f1d8 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1065.745189] env[65726]: value = "vm-995285" [ 1065.745189] env[65726]: _type = "VirtualMachine" [ 1065.745189] env[65726]: }. {{(pid=65726) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1065.745482] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-e8e8247a-bf87-418f-b518-05e929e9c56c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.754140] env[65726]: DEBUG oslo_vmware.rw_handles [None req-f263eae6-96b2-45f2-b03a-d97fe5d0f1d8 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Lease: (returnval){ [ 1065.754140] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]529c5d2f-0b3d-e619-cbc2-9e3c38541098" [ 1065.754140] env[65726]: _type = "HttpNfcLease" [ 1065.754140] env[65726]: } obtained for exporting VM: (result){ [ 1065.754140] env[65726]: value = "vm-995285" [ 1065.754140] env[65726]: _type = "VirtualMachine" [ 1065.754140] env[65726]: }. {{(pid=65726) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1065.754395] env[65726]: DEBUG oslo_vmware.api [None req-f263eae6-96b2-45f2-b03a-d97fe5d0f1d8 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Waiting for the lease: (returnval){ [ 1065.754395] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]529c5d2f-0b3d-e619-cbc2-9e3c38541098" [ 1065.754395] env[65726]: _type = "HttpNfcLease" [ 1065.754395] env[65726]: } to be ready. {{(pid=65726) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1065.766698] env[65726]: WARNING openstack [req-b6bd140f-5338-4661-948f-a8376c22ff3f req-250ce9de-3d1f-482a-94b8-337b92322431 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1065.767123] env[65726]: WARNING openstack [req-b6bd140f-5338-4661-948f-a8376c22ff3f req-250ce9de-3d1f-482a-94b8-337b92322431 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1065.779440] env[65726]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1065.779440] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]529c5d2f-0b3d-e619-cbc2-9e3c38541098" [ 1065.779440] env[65726]: _type = "HttpNfcLease" [ 1065.779440] env[65726]: } is initializing. {{(pid=65726) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1065.800639] env[65726]: DEBUG oslo_concurrency.lockutils [None req-8c26e8da-5d87-42a2-a8c5-15e9c65a8b9f tempest-InstanceActionsV221TestJSON-647428531 tempest-InstanceActionsV221TestJSON-647428531-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1065.863564] env[65726]: DEBUG nova.network.neutron [req-b6bd140f-5338-4661-948f-a8376c22ff3f req-250ce9de-3d1f-482a-94b8-337b92322431 service nova] [instance: 51f5337d-2e23-4b7c-b590-76be6bd82411] Updated VIF entry in instance network info cache for port 7407bcd9-3202-4798-9eaa-252fb3fdb21f. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 1065.863927] env[65726]: DEBUG nova.network.neutron [req-b6bd140f-5338-4661-948f-a8376c22ff3f req-250ce9de-3d1f-482a-94b8-337b92322431 service nova] [instance: 51f5337d-2e23-4b7c-b590-76be6bd82411] Updating instance_info_cache with network_info: [{"id": "7407bcd9-3202-4798-9eaa-252fb3fdb21f", "address": "fa:16:3e:79:d2:9d", "network": {"id": "51a66448-0f72-4266-8868-856b48be839f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-155658961-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.213", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d46ccb00794f458b85da4a93879139ee", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b1f3e6c3-5584-4852-9017-476ab8ac4946", "external-id": "nsx-vlan-transportzone-304", "segmentation_id": 304, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7407bcd9-32", "ovs_interfaceid": "7407bcd9-3202-4798-9eaa-252fb3fdb21f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1065.888131] env[65726]: DEBUG oslo_vmware.api [None req-ab8fab95-1b19-4328-a74a-7a9a0971be1c tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Task: {'id': task-5116629, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.177067} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1065.888342] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-ab8fab95-1b19-4328-a74a-7a9a0971be1c tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Deleted the datastore file {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1065.888514] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-ab8fab95-1b19-4328-a74a-7a9a0971be1c tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] [instance: 48923e3d-b00d-4034-bacf-7be82009fb08] Deleted contents of the VM from datastore datastore2 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1065.888682] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-ab8fab95-1b19-4328-a74a-7a9a0971be1c tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] [instance: 48923e3d-b00d-4034-bacf-7be82009fb08] Instance destroyed {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1065.888845] env[65726]: INFO nova.compute.manager [None req-ab8fab95-1b19-4328-a74a-7a9a0971be1c tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] [instance: 48923e3d-b00d-4034-bacf-7be82009fb08] Took 1.15 seconds to destroy the instance on the hypervisor. [ 1065.889096] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-ab8fab95-1b19-4328-a74a-7a9a0971be1c tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1065.889288] env[65726]: DEBUG nova.compute.manager [-] [instance: 48923e3d-b00d-4034-bacf-7be82009fb08] Deallocating network for instance {{(pid=65726) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1065.889384] env[65726]: DEBUG nova.network.neutron [-] [instance: 48923e3d-b00d-4034-bacf-7be82009fb08] deallocate_for_instance() {{(pid=65726) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 1065.889877] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1065.890175] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1065.925718] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1065.926010] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1065.946256] env[65726]: DEBUG oslo_vmware.api [None req-dc5e36ff-c9dd-4078-bec5-b01919099186 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5116630, 'name': PowerOffVM_Task, 'duration_secs': 0.293668} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1065.946393] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-dc5e36ff-c9dd-4078-bec5-b01919099186 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: a56be6fa-6027-46ee-9aa8-332fe4222d68] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1065.946648] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-dc5e36ff-c9dd-4078-bec5-b01919099186 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: a56be6fa-6027-46ee-9aa8-332fe4222d68] Updating instance 'a56be6fa-6027-46ee-9aa8-332fe4222d68' progress to 17 {{(pid=65726) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1066.005578] env[65726]: DEBUG oslo_concurrency.lockutils [None req-f3c5543b-5b17-4006-8d18-e03fa5503534 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.418s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1066.007992] env[65726]: DEBUG oslo_concurrency.lockutils [None req-4c7a0966-9c18-4500-91fc-ea14494d2511 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 7.025s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1066.008239] env[65726]: DEBUG nova.objects.instance [None req-4c7a0966-9c18-4500-91fc-ea14494d2511 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Lazy-loading 'resources' on Instance uuid 6d2b86c2-0a44-433f-8b3f-2e9c7f693baa {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1066.024857] env[65726]: INFO nova.scheduler.client.report [None req-f3c5543b-5b17-4006-8d18-e03fa5503534 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] Deleted allocations for instance 2b8eb60b-dc9b-47a4-9ddc-205f2af69888 [ 1066.098463] env[65726]: DEBUG oslo_vmware.api [None req-42911402-8543-4c4b-aab9-86db149010d2 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52b4d5f1-68c3-c501-36e9-0f59d3817375, 'name': SearchDatastore_Task, 'duration_secs': 0.011462} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1066.098745] env[65726]: DEBUG oslo_concurrency.lockutils [None req-42911402-8543-4c4b-aab9-86db149010d2 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1066.099016] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-42911402-8543-4c4b-aab9-86db149010d2 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] 5cdf1d61-b817-4986-a48c-804a2374497a/5cdf1d61-b817-4986-a48c-804a2374497a.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1066.099301] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d869ff09-f40a-404b-924d-736d4a65584a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.107355] env[65726]: DEBUG oslo_vmware.api [None req-42911402-8543-4c4b-aab9-86db149010d2 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Waiting for the task: (returnval){ [ 1066.107355] env[65726]: value = "task-5116632" [ 1066.107355] env[65726]: _type = "Task" [ 1066.107355] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1066.120206] env[65726]: DEBUG oslo_vmware.api [None req-42911402-8543-4c4b-aab9-86db149010d2 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': task-5116632, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.150096] env[65726]: DEBUG nova.compute.manager [req-f8c9dae6-5ed5-4000-b7e5-fc13945a7f74 req-c5486d3e-5689-482f-acd9-1e037a25d06e service nova] [instance: 48923e3d-b00d-4034-bacf-7be82009fb08] Received event network-vif-deleted-03191f4c-fd30-4ddc-abfd-75a70e8ab0d6 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1066.150381] env[65726]: INFO nova.compute.manager [req-f8c9dae6-5ed5-4000-b7e5-fc13945a7f74 req-c5486d3e-5689-482f-acd9-1e037a25d06e service nova] [instance: 48923e3d-b00d-4034-bacf-7be82009fb08] Neutron deleted interface 03191f4c-fd30-4ddc-abfd-75a70e8ab0d6; detaching it from the instance and deleting it from the info cache [ 1066.150645] env[65726]: DEBUG nova.network.neutron [req-f8c9dae6-5ed5-4000-b7e5-fc13945a7f74 req-c5486d3e-5689-482f-acd9-1e037a25d06e service nova] [instance: 48923e3d-b00d-4034-bacf-7be82009fb08] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1066.263592] env[65726]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1066.263592] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]529c5d2f-0b3d-e619-cbc2-9e3c38541098" [ 1066.263592] env[65726]: _type = "HttpNfcLease" [ 1066.263592] env[65726]: } is ready. {{(pid=65726) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1066.263956] env[65726]: DEBUG oslo_vmware.rw_handles [None req-f263eae6-96b2-45f2-b03a-d97fe5d0f1d8 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1066.263956] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]529c5d2f-0b3d-e619-cbc2-9e3c38541098" [ 1066.263956] env[65726]: _type = "HttpNfcLease" [ 1066.263956] env[65726]: }. {{(pid=65726) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1066.264740] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ceb0a3e-0efb-438f-9f7e-5f1330a1d681 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.273519] env[65726]: DEBUG oslo_vmware.rw_handles [None req-f263eae6-96b2-45f2-b03a-d97fe5d0f1d8 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52fbf44b-bcce-284e-0923-049a7f39cb78/disk-0.vmdk from lease info. {{(pid=65726) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1066.273709] env[65726]: DEBUG oslo_vmware.rw_handles [None req-f263eae6-96b2-45f2-b03a-d97fe5d0f1d8 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52fbf44b-bcce-284e-0923-049a7f39cb78/disk-0.vmdk for reading. {{(pid=65726) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1066.366891] env[65726]: DEBUG oslo_concurrency.lockutils [req-b6bd140f-5338-4661-948f-a8376c22ff3f req-250ce9de-3d1f-482a-94b8-337b92322431 service nova] Releasing lock "refresh_cache-51f5337d-2e23-4b7c-b590-76be6bd82411" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1066.454477] env[65726]: DEBUG nova.virt.hardware [None req-dc5e36ff-c9dd-4078-bec5-b01919099186 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1066.454737] env[65726]: DEBUG nova.virt.hardware [None req-dc5e36ff-c9dd-4078-bec5-b01919099186 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1066.454915] env[65726]: DEBUG nova.virt.hardware [None req-dc5e36ff-c9dd-4078-bec5-b01919099186 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1066.455146] env[65726]: DEBUG nova.virt.hardware [None req-dc5e36ff-c9dd-4078-bec5-b01919099186 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1066.455320] env[65726]: DEBUG nova.virt.hardware [None req-dc5e36ff-c9dd-4078-bec5-b01919099186 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1066.455501] env[65726]: DEBUG nova.virt.hardware [None req-dc5e36ff-c9dd-4078-bec5-b01919099186 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1066.459028] env[65726]: DEBUG nova.virt.hardware [None req-dc5e36ff-c9dd-4078-bec5-b01919099186 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1066.459028] env[65726]: DEBUG nova.virt.hardware [None req-dc5e36ff-c9dd-4078-bec5-b01919099186 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1066.459028] env[65726]: DEBUG nova.virt.hardware [None req-dc5e36ff-c9dd-4078-bec5-b01919099186 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1066.459028] env[65726]: DEBUG nova.virt.hardware [None req-dc5e36ff-c9dd-4078-bec5-b01919099186 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1066.459028] env[65726]: DEBUG nova.virt.hardware [None req-dc5e36ff-c9dd-4078-bec5-b01919099186 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1066.463395] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-297f1285-7199-4c64-bbb1-982bf70ef18b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.483545] env[65726]: DEBUG oslo_vmware.api [None req-dc5e36ff-c9dd-4078-bec5-b01919099186 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Waiting for the task: (returnval){ [ 1066.483545] env[65726]: value = "task-5116633" [ 1066.483545] env[65726]: _type = "Task" [ 1066.483545] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1066.494097] env[65726]: DEBUG oslo_vmware.api [None req-dc5e36ff-c9dd-4078-bec5-b01919099186 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5116633, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.534119] env[65726]: DEBUG oslo_concurrency.lockutils [None req-f3c5543b-5b17-4006-8d18-e03fa5503534 tempest-InstanceActionsTestJSON-181224484 tempest-InstanceActionsTestJSON-181224484-project-member] Lock "2b8eb60b-dc9b-47a4-9ddc-205f2af69888" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 14.063s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1066.593447] env[65726]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-9b27393d-96d9-4ba9-b9f4-0f4d59195db2 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.616804] env[65726]: DEBUG nova.compute.manager [None req-80fb9493-d54c-4f0a-a5fd-0197feeffb7b tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] [instance: 3b4d9461-747a-461c-b231-a0de02f0cb83] Start spawning the instance on the hypervisor. {{(pid=65726) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1066.626436] env[65726]: DEBUG nova.network.neutron [-] [instance: 48923e3d-b00d-4034-bacf-7be82009fb08] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1066.628021] env[65726]: DEBUG oslo_vmware.api [None req-42911402-8543-4c4b-aab9-86db149010d2 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': task-5116632, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.645830] env[65726]: DEBUG nova.virt.hardware [None req-80fb9493-d54c-4f0a-a5fd-0197feeffb7b tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T19:28:34Z,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1539f06fa3534e90acbc3a60c4b8bd3f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T19:28:35Z,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1066.646043] env[65726]: DEBUG nova.virt.hardware [None req-80fb9493-d54c-4f0a-a5fd-0197feeffb7b tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1066.646207] env[65726]: DEBUG nova.virt.hardware [None req-80fb9493-d54c-4f0a-a5fd-0197feeffb7b tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1066.646388] env[65726]: DEBUG nova.virt.hardware [None req-80fb9493-d54c-4f0a-a5fd-0197feeffb7b tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1066.646745] env[65726]: DEBUG nova.virt.hardware [None req-80fb9493-d54c-4f0a-a5fd-0197feeffb7b tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1066.646745] env[65726]: DEBUG nova.virt.hardware [None req-80fb9493-d54c-4f0a-a5fd-0197feeffb7b tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1066.646860] env[65726]: DEBUG nova.virt.hardware [None req-80fb9493-d54c-4f0a-a5fd-0197feeffb7b tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1066.647018] env[65726]: DEBUG nova.virt.hardware [None req-80fb9493-d54c-4f0a-a5fd-0197feeffb7b tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1066.647201] env[65726]: DEBUG nova.virt.hardware [None req-80fb9493-d54c-4f0a-a5fd-0197feeffb7b tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1066.647353] env[65726]: DEBUG nova.virt.hardware [None req-80fb9493-d54c-4f0a-a5fd-0197feeffb7b tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1066.647545] env[65726]: DEBUG nova.virt.hardware [None req-80fb9493-d54c-4f0a-a5fd-0197feeffb7b tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1066.650043] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d97081f-467b-4af7-9f69-5f03fb12b7f3 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.659315] env[65726]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1c95803c-de5f-435c-90f7-40fa5481149b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.676735] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cdb54fc8-04a6-4a98-8684-c147a0c3da75 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.688304] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9667b2a-28a3-421c-939e-ca73b362cb89 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.721181] env[65726]: DEBUG nova.compute.manager [req-f8c9dae6-5ed5-4000-b7e5-fc13945a7f74 req-c5486d3e-5689-482f-acd9-1e037a25d06e service nova] [instance: 48923e3d-b00d-4034-bacf-7be82009fb08] Detach interface failed, port_id=03191f4c-fd30-4ddc-abfd-75a70e8ab0d6, reason: Instance 48923e3d-b00d-4034-bacf-7be82009fb08 could not be found. {{(pid=65726) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11669}} [ 1066.807343] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a8f6003-3340-4822-9c31-c6b3cb8e9b00 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.815548] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99a6d85f-6226-4d90-8041-0ac2bbcb3cf3 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.849218] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be2c711b-eed6-4525-bf68-90c4aa7b035a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.858919] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8206d05-b2ae-4c13-b2fe-5d01de3d1677 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.876327] env[65726]: DEBUG nova.compute.provider_tree [None req-4c7a0966-9c18-4500-91fc-ea14494d2511 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1066.961536] env[65726]: DEBUG oslo_service.periodic_task [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=65726) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1066.961785] env[65726]: DEBUG oslo_service.periodic_task [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=65726) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1066.995113] env[65726]: DEBUG oslo_vmware.api [None req-dc5e36ff-c9dd-4078-bec5-b01919099186 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5116633, 'name': ReconfigVM_Task, 'duration_secs': 0.294162} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1066.996788] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-dc5e36ff-c9dd-4078-bec5-b01919099186 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: a56be6fa-6027-46ee-9aa8-332fe4222d68] Updating instance 'a56be6fa-6027-46ee-9aa8-332fe4222d68' progress to 33 {{(pid=65726) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1067.086463] env[65726]: DEBUG nova.network.neutron [None req-80fb9493-d54c-4f0a-a5fd-0197feeffb7b tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] [instance: 3b4d9461-747a-461c-b231-a0de02f0cb83] Successfully updated port: 20c6b53e-2740-46bc-a94f-59a15f8b2c61 {{(pid=65726) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 1067.119769] env[65726]: DEBUG oslo_vmware.api [None req-42911402-8543-4c4b-aab9-86db149010d2 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': task-5116632, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.548169} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1067.120311] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-42911402-8543-4c4b-aab9-86db149010d2 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] 5cdf1d61-b817-4986-a48c-804a2374497a/5cdf1d61-b817-4986-a48c-804a2374497a.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1067.120659] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-42911402-8543-4c4b-aab9-86db149010d2 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 5cdf1d61-b817-4986-a48c-804a2374497a] Extending root virtual disk to 1048576 {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1067.121033] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c66f31e6-ec05-4618-be9a-bc2d35942cef {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.129460] env[65726]: INFO nova.compute.manager [-] [instance: 48923e3d-b00d-4034-bacf-7be82009fb08] Took 1.24 seconds to deallocate network for instance. [ 1067.132636] env[65726]: DEBUG oslo_vmware.api [None req-42911402-8543-4c4b-aab9-86db149010d2 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Waiting for the task: (returnval){ [ 1067.132636] env[65726]: value = "task-5116634" [ 1067.132636] env[65726]: _type = "Task" [ 1067.132636] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1067.149413] env[65726]: DEBUG oslo_vmware.api [None req-42911402-8543-4c4b-aab9-86db149010d2 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': task-5116634, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.379904] env[65726]: DEBUG nova.scheduler.client.report [None req-4c7a0966-9c18-4500-91fc-ea14494d2511 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1067.468865] env[65726]: DEBUG oslo_service.periodic_task [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=65726) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1067.471060] env[65726]: DEBUG oslo_service.periodic_task [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=65726) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1067.471060] env[65726]: DEBUG oslo_service.periodic_task [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=65726) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1067.471060] env[65726]: DEBUG oslo_service.periodic_task [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=65726) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1067.471060] env[65726]: DEBUG oslo_service.periodic_task [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=65726) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1067.471060] env[65726]: DEBUG oslo_service.periodic_task [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=65726) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1067.471060] env[65726]: DEBUG nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=65726) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11251}} [ 1067.471060] env[65726]: DEBUG oslo_service.periodic_task [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Running periodic task ComputeManager.update_available_resource {{(pid=65726) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1067.503431] env[65726]: DEBUG nova.virt.hardware [None req-dc5e36ff-c9dd-4078-bec5-b01919099186 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1067.503643] env[65726]: DEBUG nova.virt.hardware [None req-dc5e36ff-c9dd-4078-bec5-b01919099186 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1067.503676] env[65726]: DEBUG nova.virt.hardware [None req-dc5e36ff-c9dd-4078-bec5-b01919099186 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1067.504718] env[65726]: DEBUG nova.virt.hardware [None req-dc5e36ff-c9dd-4078-bec5-b01919099186 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1067.505059] env[65726]: DEBUG nova.virt.hardware [None req-dc5e36ff-c9dd-4078-bec5-b01919099186 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1067.505325] env[65726]: DEBUG nova.virt.hardware [None req-dc5e36ff-c9dd-4078-bec5-b01919099186 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1067.505908] env[65726]: DEBUG nova.virt.hardware [None req-dc5e36ff-c9dd-4078-bec5-b01919099186 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1067.505908] env[65726]: DEBUG nova.virt.hardware [None req-dc5e36ff-c9dd-4078-bec5-b01919099186 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1067.505908] env[65726]: DEBUG nova.virt.hardware [None req-dc5e36ff-c9dd-4078-bec5-b01919099186 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1067.506249] env[65726]: DEBUG nova.virt.hardware [None req-dc5e36ff-c9dd-4078-bec5-b01919099186 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1067.508975] env[65726]: DEBUG nova.virt.hardware [None req-dc5e36ff-c9dd-4078-bec5-b01919099186 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1067.512553] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-dc5e36ff-c9dd-4078-bec5-b01919099186 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: a56be6fa-6027-46ee-9aa8-332fe4222d68] Reconfiguring VM instance instance-00000053 to detach disk 2000 {{(pid=65726) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1067.513667] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2a022f77-fc6b-4160-a9bf-36ad739965a1 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.541398] env[65726]: DEBUG nova.compute.manager [req-7ee8789d-79ad-4cee-a50e-7ed356ad18ef req-e8befe32-3533-4233-8ef2-302e904c14de service nova] [instance: 3b4d9461-747a-461c-b231-a0de02f0cb83] Received event network-vif-plugged-20c6b53e-2740-46bc-a94f-59a15f8b2c61 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1067.541700] env[65726]: DEBUG oslo_concurrency.lockutils [req-7ee8789d-79ad-4cee-a50e-7ed356ad18ef req-e8befe32-3533-4233-8ef2-302e904c14de service nova] Acquiring lock "3b4d9461-747a-461c-b231-a0de02f0cb83-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1067.543199] env[65726]: DEBUG oslo_concurrency.lockutils [req-7ee8789d-79ad-4cee-a50e-7ed356ad18ef req-e8befe32-3533-4233-8ef2-302e904c14de service nova] Lock "3b4d9461-747a-461c-b231-a0de02f0cb83-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1067.543199] env[65726]: DEBUG oslo_concurrency.lockutils [req-7ee8789d-79ad-4cee-a50e-7ed356ad18ef req-e8befe32-3533-4233-8ef2-302e904c14de service nova] Lock "3b4d9461-747a-461c-b231-a0de02f0cb83-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1067.543199] env[65726]: DEBUG nova.compute.manager [req-7ee8789d-79ad-4cee-a50e-7ed356ad18ef req-e8befe32-3533-4233-8ef2-302e904c14de service nova] [instance: 3b4d9461-747a-461c-b231-a0de02f0cb83] No waiting events found dispatching network-vif-plugged-20c6b53e-2740-46bc-a94f-59a15f8b2c61 {{(pid=65726) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1067.543199] env[65726]: WARNING nova.compute.manager [req-7ee8789d-79ad-4cee-a50e-7ed356ad18ef req-e8befe32-3533-4233-8ef2-302e904c14de service nova] [instance: 3b4d9461-747a-461c-b231-a0de02f0cb83] Received unexpected event network-vif-plugged-20c6b53e-2740-46bc-a94f-59a15f8b2c61 for instance with vm_state building and task_state spawning. [ 1067.551040] env[65726]: DEBUG oslo_vmware.api [None req-dc5e36ff-c9dd-4078-bec5-b01919099186 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Waiting for the task: (returnval){ [ 1067.551040] env[65726]: value = "task-5116635" [ 1067.551040] env[65726]: _type = "Task" [ 1067.551040] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1067.561814] env[65726]: DEBUG oslo_vmware.api [None req-dc5e36ff-c9dd-4078-bec5-b01919099186 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5116635, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.591138] env[65726]: DEBUG oslo_concurrency.lockutils [None req-80fb9493-d54c-4f0a-a5fd-0197feeffb7b tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Acquiring lock "refresh_cache-3b4d9461-747a-461c-b231-a0de02f0cb83" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1067.591138] env[65726]: DEBUG oslo_concurrency.lockutils [None req-80fb9493-d54c-4f0a-a5fd-0197feeffb7b tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Acquired lock "refresh_cache-3b4d9461-747a-461c-b231-a0de02f0cb83" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1067.591138] env[65726]: DEBUG nova.network.neutron [None req-80fb9493-d54c-4f0a-a5fd-0197feeffb7b tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] [instance: 3b4d9461-747a-461c-b231-a0de02f0cb83] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1067.642386] env[65726]: DEBUG oslo_concurrency.lockutils [None req-ab8fab95-1b19-4328-a74a-7a9a0971be1c tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1067.649523] env[65726]: DEBUG oslo_vmware.api [None req-42911402-8543-4c4b-aab9-86db149010d2 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': task-5116634, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.205069} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1067.649928] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-42911402-8543-4c4b-aab9-86db149010d2 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 5cdf1d61-b817-4986-a48c-804a2374497a] Extended root virtual disk {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1067.650908] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d226301-07df-4021-8236-3da486b904b3 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.679537] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-42911402-8543-4c4b-aab9-86db149010d2 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 5cdf1d61-b817-4986-a48c-804a2374497a] Reconfiguring VM instance instance-0000005f to attach disk [datastore1] 5cdf1d61-b817-4986-a48c-804a2374497a/5cdf1d61-b817-4986-a48c-804a2374497a.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1067.680075] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f9052d51-54df-4587-92a2-630e55ea564c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.702587] env[65726]: DEBUG oslo_vmware.api [None req-42911402-8543-4c4b-aab9-86db149010d2 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Waiting for the task: (returnval){ [ 1067.702587] env[65726]: value = "task-5116636" [ 1067.702587] env[65726]: _type = "Task" [ 1067.702587] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1067.713140] env[65726]: DEBUG oslo_vmware.api [None req-42911402-8543-4c4b-aab9-86db149010d2 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': task-5116636, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.886083] env[65726]: DEBUG oslo_concurrency.lockutils [None req-4c7a0966-9c18-4500-91fc-ea14494d2511 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.878s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1067.891186] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5ecca65c-f696-44b9-8b91-de86779bab97 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.350s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1067.891186] env[65726]: DEBUG nova.objects.instance [None req-5ecca65c-f696-44b9-8b91-de86779bab97 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Lazy-loading 'resources' on Instance uuid f947529e-c930-4420-91a7-1e677b5ac2f1 {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1067.920434] env[65726]: INFO nova.scheduler.client.report [None req-4c7a0966-9c18-4500-91fc-ea14494d2511 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Deleted allocations for instance 6d2b86c2-0a44-433f-8b3f-2e9c7f693baa [ 1067.975097] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1068.064469] env[65726]: DEBUG oslo_vmware.api [None req-dc5e36ff-c9dd-4078-bec5-b01919099186 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5116635, 'name': ReconfigVM_Task, 'duration_secs': 0.220282} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1068.065637] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-dc5e36ff-c9dd-4078-bec5-b01919099186 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: a56be6fa-6027-46ee-9aa8-332fe4222d68] Reconfigured VM instance instance-00000053 to detach disk 2000 {{(pid=65726) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1068.067159] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48c22d77-282f-492f-8428-4cf71af521c9 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.096947] env[65726]: WARNING openstack [None req-80fb9493-d54c-4f0a-a5fd-0197feeffb7b tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1068.097774] env[65726]: WARNING openstack [None req-80fb9493-d54c-4f0a-a5fd-0197feeffb7b tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1068.103369] env[65726]: DEBUG nova.network.neutron [None req-80fb9493-d54c-4f0a-a5fd-0197feeffb7b tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] [instance: 3b4d9461-747a-461c-b231-a0de02f0cb83] Instance cache missing network info. {{(pid=65726) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 1068.113326] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-dc5e36ff-c9dd-4078-bec5-b01919099186 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: a56be6fa-6027-46ee-9aa8-332fe4222d68] Reconfiguring VM instance instance-00000053 to attach disk [datastore1] a56be6fa-6027-46ee-9aa8-332fe4222d68/a56be6fa-6027-46ee-9aa8-332fe4222d68.vmdk or device None with type thin {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1068.114840] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4557aa95-efd8-4126-84db-526305e8ef41 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.136033] env[65726]: DEBUG oslo_vmware.api [None req-dc5e36ff-c9dd-4078-bec5-b01919099186 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Waiting for the task: (returnval){ [ 1068.136033] env[65726]: value = "task-5116637" [ 1068.136033] env[65726]: _type = "Task" [ 1068.136033] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1068.145853] env[65726]: DEBUG oslo_vmware.api [None req-dc5e36ff-c9dd-4078-bec5-b01919099186 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5116637, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.174821] env[65726]: WARNING openstack [None req-80fb9493-d54c-4f0a-a5fd-0197feeffb7b tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1068.175346] env[65726]: WARNING openstack [None req-80fb9493-d54c-4f0a-a5fd-0197feeffb7b tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1068.221066] env[65726]: DEBUG oslo_vmware.api [None req-42911402-8543-4c4b-aab9-86db149010d2 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': task-5116636, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.254218] env[65726]: WARNING openstack [None req-80fb9493-d54c-4f0a-a5fd-0197feeffb7b tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1068.254689] env[65726]: WARNING openstack [None req-80fb9493-d54c-4f0a-a5fd-0197feeffb7b tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1068.417432] env[65726]: DEBUG nova.network.neutron [None req-80fb9493-d54c-4f0a-a5fd-0197feeffb7b tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] [instance: 3b4d9461-747a-461c-b231-a0de02f0cb83] Updating instance_info_cache with network_info: [{"id": "20c6b53e-2740-46bc-a94f-59a15f8b2c61", "address": "fa:16:3e:09:2a:fb", "network": {"id": "53facc8b-af7e-44de-8c6f-6bea798b3f0b", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-2084216438-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "eb811e38506b4297bf01bc5f2e098370", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "19671de9-8b5b-4710-adc3-7419f3c0f171", "external-id": "nsx-vlan-transportzone-421", "segmentation_id": 421, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap20c6b53e-27", "ovs_interfaceid": "20c6b53e-2740-46bc-a94f-59a15f8b2c61", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1068.432280] env[65726]: DEBUG oslo_concurrency.lockutils [None req-4c7a0966-9c18-4500-91fc-ea14494d2511 tempest-ServerRescueNegativeTestJSON-1470543753 tempest-ServerRescueNegativeTestJSON-1470543753-project-member] Lock "6d2b86c2-0a44-433f-8b3f-2e9c7f693baa" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 12.954s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1068.648212] env[65726]: DEBUG oslo_vmware.api [None req-dc5e36ff-c9dd-4078-bec5-b01919099186 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5116637, 'name': ReconfigVM_Task, 'duration_secs': 0.346118} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1068.648898] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-dc5e36ff-c9dd-4078-bec5-b01919099186 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: a56be6fa-6027-46ee-9aa8-332fe4222d68] Reconfigured VM instance instance-00000053 to attach disk [datastore1] a56be6fa-6027-46ee-9aa8-332fe4222d68/a56be6fa-6027-46ee-9aa8-332fe4222d68.vmdk or device None with type thin {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1068.649096] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-dc5e36ff-c9dd-4078-bec5-b01919099186 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: a56be6fa-6027-46ee-9aa8-332fe4222d68] Updating instance 'a56be6fa-6027-46ee-9aa8-332fe4222d68' progress to 50 {{(pid=65726) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1068.667332] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fdb9aa41-8d43-4ad4-adab-5540df6c576b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.676087] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f261eb51-1878-4b7e-bc6c-e07a037f3f22 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.713574] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b387970-de7c-4b43-a6e7-fb56370ca91a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.728620] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40e5bdd2-34dd-4a59-8fdd-b1940b537b65 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.734977] env[65726]: DEBUG oslo_vmware.api [None req-42911402-8543-4c4b-aab9-86db149010d2 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': task-5116636, 'name': ReconfigVM_Task, 'duration_secs': 0.586331} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1068.735932] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-42911402-8543-4c4b-aab9-86db149010d2 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 5cdf1d61-b817-4986-a48c-804a2374497a] Reconfigured VM instance instance-0000005f to attach disk [datastore1] 5cdf1d61-b817-4986-a48c-804a2374497a/5cdf1d61-b817-4986-a48c-804a2374497a.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1068.736925] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3247693d-fdd7-46e3-bfb1-35016563bca9 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.750585] env[65726]: DEBUG nova.compute.provider_tree [None req-5ecca65c-f696-44b9-8b91-de86779bab97 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1068.754920] env[65726]: DEBUG oslo_vmware.api [None req-42911402-8543-4c4b-aab9-86db149010d2 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Waiting for the task: (returnval){ [ 1068.754920] env[65726]: value = "task-5116638" [ 1068.754920] env[65726]: _type = "Task" [ 1068.754920] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1068.764997] env[65726]: DEBUG oslo_vmware.api [None req-42911402-8543-4c4b-aab9-86db149010d2 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': task-5116638, 'name': Rename_Task} progress is 10%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.920616] env[65726]: DEBUG oslo_concurrency.lockutils [None req-80fb9493-d54c-4f0a-a5fd-0197feeffb7b tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Releasing lock "refresh_cache-3b4d9461-747a-461c-b231-a0de02f0cb83" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1068.921036] env[65726]: DEBUG nova.compute.manager [None req-80fb9493-d54c-4f0a-a5fd-0197feeffb7b tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] [instance: 3b4d9461-747a-461c-b231-a0de02f0cb83] Instance network_info: |[{"id": "20c6b53e-2740-46bc-a94f-59a15f8b2c61", "address": "fa:16:3e:09:2a:fb", "network": {"id": "53facc8b-af7e-44de-8c6f-6bea798b3f0b", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-2084216438-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "eb811e38506b4297bf01bc5f2e098370", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "19671de9-8b5b-4710-adc3-7419f3c0f171", "external-id": "nsx-vlan-transportzone-421", "segmentation_id": 421, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap20c6b53e-27", "ovs_interfaceid": "20c6b53e-2740-46bc-a94f-59a15f8b2c61", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1068.921543] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-80fb9493-d54c-4f0a-a5fd-0197feeffb7b tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] [instance: 3b4d9461-747a-461c-b231-a0de02f0cb83] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:09:2a:fb', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '19671de9-8b5b-4710-adc3-7419f3c0f171', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '20c6b53e-2740-46bc-a94f-59a15f8b2c61', 'vif_model': 'vmxnet3'}] {{(pid=65726) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1068.930274] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-80fb9493-d54c-4f0a-a5fd-0197feeffb7b tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Creating folder: Project (eb811e38506b4297bf01bc5f2e098370). Parent ref: group-v995008. {{(pid=65726) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1068.930742] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c9e39810-16b2-441f-bb10-1efe5bbff73e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.943647] env[65726]: INFO nova.virt.vmwareapi.vm_util [None req-80fb9493-d54c-4f0a-a5fd-0197feeffb7b tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Created folder: Project (eb811e38506b4297bf01bc5f2e098370) in parent group-v995008. [ 1068.943840] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-80fb9493-d54c-4f0a-a5fd-0197feeffb7b tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Creating folder: Instances. Parent ref: group-v995286. {{(pid=65726) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1068.944028] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-6d8d11ec-966b-47e3-a338-0fb196d30325 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.955583] env[65726]: INFO nova.virt.vmwareapi.vm_util [None req-80fb9493-d54c-4f0a-a5fd-0197feeffb7b tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Created folder: Instances in parent group-v995286. [ 1068.955875] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-80fb9493-d54c-4f0a-a5fd-0197feeffb7b tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1068.956107] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3b4d9461-747a-461c-b231-a0de02f0cb83] Creating VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1068.956334] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-70aa8be2-d4d5-4175-bac3-449c2bc96e56 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.977859] env[65726]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1068.977859] env[65726]: value = "task-5116641" [ 1068.977859] env[65726]: _type = "Task" [ 1068.977859] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1068.987156] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116641, 'name': CreateVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1069.159668] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae3ef6a4-fd2d-4863-bf17-503cbb270c74 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.183839] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7295da12-9e96-4ede-836c-5093b3f73bad {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.209531] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-dc5e36ff-c9dd-4078-bec5-b01919099186 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: a56be6fa-6027-46ee-9aa8-332fe4222d68] Updating instance 'a56be6fa-6027-46ee-9aa8-332fe4222d68' progress to 67 {{(pid=65726) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1069.258648] env[65726]: DEBUG nova.scheduler.client.report [None req-5ecca65c-f696-44b9-8b91-de86779bab97 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1069.272495] env[65726]: DEBUG oslo_vmware.api [None req-42911402-8543-4c4b-aab9-86db149010d2 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': task-5116638, 'name': Rename_Task, 'duration_secs': 0.166572} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1069.272789] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-42911402-8543-4c4b-aab9-86db149010d2 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 5cdf1d61-b817-4986-a48c-804a2374497a] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1069.273248] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0501e202-a8cf-47b3-b7a9-8d3058be8ca4 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.282599] env[65726]: DEBUG oslo_vmware.api [None req-42911402-8543-4c4b-aab9-86db149010d2 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Waiting for the task: (returnval){ [ 1069.282599] env[65726]: value = "task-5116642" [ 1069.282599] env[65726]: _type = "Task" [ 1069.282599] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1069.295067] env[65726]: DEBUG oslo_vmware.api [None req-42911402-8543-4c4b-aab9-86db149010d2 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': task-5116642, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1069.488794] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116641, 'name': CreateVM_Task, 'duration_secs': 0.38076} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1069.489173] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3b4d9461-747a-461c-b231-a0de02f0cb83] Created VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1069.489910] env[65726]: WARNING openstack [None req-80fb9493-d54c-4f0a-a5fd-0197feeffb7b tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1069.490319] env[65726]: WARNING openstack [None req-80fb9493-d54c-4f0a-a5fd-0197feeffb7b tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1069.496286] env[65726]: DEBUG oslo_concurrency.lockutils [None req-80fb9493-d54c-4f0a-a5fd-0197feeffb7b tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1069.496410] env[65726]: DEBUG oslo_concurrency.lockutils [None req-80fb9493-d54c-4f0a-a5fd-0197feeffb7b tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1069.496697] env[65726]: DEBUG oslo_concurrency.lockutils [None req-80fb9493-d54c-4f0a-a5fd-0197feeffb7b tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1069.497032] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fc19c816-d7fc-4808-82f0-ff8bb8aa5660 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.503540] env[65726]: DEBUG oslo_vmware.api [None req-80fb9493-d54c-4f0a-a5fd-0197feeffb7b tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Waiting for the task: (returnval){ [ 1069.503540] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]520d6621-43f3-a05f-256b-7a0b354377a1" [ 1069.503540] env[65726]: _type = "Task" [ 1069.503540] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1069.513520] env[65726]: DEBUG oslo_vmware.api [None req-80fb9493-d54c-4f0a-a5fd-0197feeffb7b tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]520d6621-43f3-a05f-256b-7a0b354377a1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1069.581033] env[65726]: DEBUG nova.compute.manager [req-91caea6d-a041-4e9f-9ca7-19377c7191c0 req-3a008e24-62ff-4c75-9b36-0a4b62cbc341 service nova] [instance: 3b4d9461-747a-461c-b231-a0de02f0cb83] Received event network-changed-20c6b53e-2740-46bc-a94f-59a15f8b2c61 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1069.581033] env[65726]: DEBUG nova.compute.manager [req-91caea6d-a041-4e9f-9ca7-19377c7191c0 req-3a008e24-62ff-4c75-9b36-0a4b62cbc341 service nova] [instance: 3b4d9461-747a-461c-b231-a0de02f0cb83] Refreshing instance network info cache due to event network-changed-20c6b53e-2740-46bc-a94f-59a15f8b2c61. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 1069.581889] env[65726]: DEBUG oslo_concurrency.lockutils [req-91caea6d-a041-4e9f-9ca7-19377c7191c0 req-3a008e24-62ff-4c75-9b36-0a4b62cbc341 service nova] Acquiring lock "refresh_cache-3b4d9461-747a-461c-b231-a0de02f0cb83" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1069.581889] env[65726]: DEBUG oslo_concurrency.lockutils [req-91caea6d-a041-4e9f-9ca7-19377c7191c0 req-3a008e24-62ff-4c75-9b36-0a4b62cbc341 service nova] Acquired lock "refresh_cache-3b4d9461-747a-461c-b231-a0de02f0cb83" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1069.581889] env[65726]: DEBUG nova.network.neutron [req-91caea6d-a041-4e9f-9ca7-19377c7191c0 req-3a008e24-62ff-4c75-9b36-0a4b62cbc341 service nova] [instance: 3b4d9461-747a-461c-b231-a0de02f0cb83] Refreshing network info cache for port 20c6b53e-2740-46bc-a94f-59a15f8b2c61 {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 1069.765592] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5ecca65c-f696-44b9-8b91-de86779bab97 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.876s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1069.769196] env[65726]: DEBUG oslo_concurrency.lockutils [None req-91ff63ad-fe79-48d0-901e-fc0ed403574a tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.965s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1069.771027] env[65726]: INFO nova.compute.claims [None req-91ff63ad-fe79-48d0-901e-fc0ed403574a tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: 0d42fa75-20e9-4646-9b08-17015b7f068c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1069.794986] env[65726]: DEBUG oslo_vmware.api [None req-42911402-8543-4c4b-aab9-86db149010d2 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': task-5116642, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1069.810869] env[65726]: INFO nova.scheduler.client.report [None req-5ecca65c-f696-44b9-8b91-de86779bab97 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Deleted allocations for instance f947529e-c930-4420-91a7-1e677b5ac2f1 [ 1070.018617] env[65726]: DEBUG oslo_vmware.api [None req-80fb9493-d54c-4f0a-a5fd-0197feeffb7b tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]520d6621-43f3-a05f-256b-7a0b354377a1, 'name': SearchDatastore_Task, 'duration_secs': 0.016501} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1070.019180] env[65726]: DEBUG oslo_concurrency.lockutils [None req-80fb9493-d54c-4f0a-a5fd-0197feeffb7b tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1070.019553] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-80fb9493-d54c-4f0a-a5fd-0197feeffb7b tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] [instance: 3b4d9461-747a-461c-b231-a0de02f0cb83] Processing image b52362a3-ee8a-4cbf-b06f-513b0cc8f95c {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1070.019893] env[65726]: DEBUG oslo_concurrency.lockutils [None req-80fb9493-d54c-4f0a-a5fd-0197feeffb7b tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1070.020155] env[65726]: DEBUG oslo_concurrency.lockutils [None req-80fb9493-d54c-4f0a-a5fd-0197feeffb7b tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1070.020515] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-80fb9493-d54c-4f0a-a5fd-0197feeffb7b tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1070.021310] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9f41af09-e781-46ec-b6c6-d7975f1d5b1e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.033037] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-80fb9493-d54c-4f0a-a5fd-0197feeffb7b tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1070.033205] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-80fb9493-d54c-4f0a-a5fd-0197feeffb7b tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65726) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1070.034129] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4eb47156-678e-45d2-9dbc-9385063e4e6f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.041231] env[65726]: DEBUG oslo_vmware.api [None req-80fb9493-d54c-4f0a-a5fd-0197feeffb7b tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Waiting for the task: (returnval){ [ 1070.041231] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]522c7e01-8904-f057-8091-be01a282a929" [ 1070.041231] env[65726]: _type = "Task" [ 1070.041231] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1070.052046] env[65726]: DEBUG oslo_vmware.api [None req-80fb9493-d54c-4f0a-a5fd-0197feeffb7b tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]522c7e01-8904-f057-8091-be01a282a929, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.085621] env[65726]: WARNING openstack [req-91caea6d-a041-4e9f-9ca7-19377c7191c0 req-3a008e24-62ff-4c75-9b36-0a4b62cbc341 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1070.085621] env[65726]: WARNING openstack [req-91caea6d-a041-4e9f-9ca7-19377c7191c0 req-3a008e24-62ff-4c75-9b36-0a4b62cbc341 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1070.231830] env[65726]: WARNING openstack [req-91caea6d-a041-4e9f-9ca7-19377c7191c0 req-3a008e24-62ff-4c75-9b36-0a4b62cbc341 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1070.232240] env[65726]: WARNING openstack [req-91caea6d-a041-4e9f-9ca7-19377c7191c0 req-3a008e24-62ff-4c75-9b36-0a4b62cbc341 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1070.293380] env[65726]: DEBUG oslo_vmware.api [None req-42911402-8543-4c4b-aab9-86db149010d2 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': task-5116642, 'name': PowerOnVM_Task, 'duration_secs': 0.68655} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1070.294737] env[65726]: WARNING openstack [req-91caea6d-a041-4e9f-9ca7-19377c7191c0 req-3a008e24-62ff-4c75-9b36-0a4b62cbc341 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1070.295042] env[65726]: WARNING openstack [req-91caea6d-a041-4e9f-9ca7-19377c7191c0 req-3a008e24-62ff-4c75-9b36-0a4b62cbc341 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1070.305110] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-42911402-8543-4c4b-aab9-86db149010d2 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 5cdf1d61-b817-4986-a48c-804a2374497a] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1070.305359] env[65726]: INFO nova.compute.manager [None req-42911402-8543-4c4b-aab9-86db149010d2 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 5cdf1d61-b817-4986-a48c-804a2374497a] Took 9.92 seconds to spawn the instance on the hypervisor. [ 1070.305546] env[65726]: DEBUG nova.compute.manager [None req-42911402-8543-4c4b-aab9-86db149010d2 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 5cdf1d61-b817-4986-a48c-804a2374497a] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1070.306868] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4270253b-e8df-4e99-90e2-c8640fa1ee71 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.325944] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5ecca65c-f696-44b9-8b91-de86779bab97 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Lock "f947529e-c930-4420-91a7-1e677b5ac2f1" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.208s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1070.434421] env[65726]: DEBUG nova.network.neutron [req-91caea6d-a041-4e9f-9ca7-19377c7191c0 req-3a008e24-62ff-4c75-9b36-0a4b62cbc341 service nova] [instance: 3b4d9461-747a-461c-b231-a0de02f0cb83] Updated VIF entry in instance network info cache for port 20c6b53e-2740-46bc-a94f-59a15f8b2c61. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 1070.434833] env[65726]: DEBUG nova.network.neutron [req-91caea6d-a041-4e9f-9ca7-19377c7191c0 req-3a008e24-62ff-4c75-9b36-0a4b62cbc341 service nova] [instance: 3b4d9461-747a-461c-b231-a0de02f0cb83] Updating instance_info_cache with network_info: [{"id": "20c6b53e-2740-46bc-a94f-59a15f8b2c61", "address": "fa:16:3e:09:2a:fb", "network": {"id": "53facc8b-af7e-44de-8c6f-6bea798b3f0b", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-2084216438-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "eb811e38506b4297bf01bc5f2e098370", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "19671de9-8b5b-4710-adc3-7419f3c0f171", "external-id": "nsx-vlan-transportzone-421", "segmentation_id": 421, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap20c6b53e-27", "ovs_interfaceid": "20c6b53e-2740-46bc-a94f-59a15f8b2c61", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1070.553013] env[65726]: DEBUG oslo_vmware.api [None req-80fb9493-d54c-4f0a-a5fd-0197feeffb7b tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]522c7e01-8904-f057-8091-be01a282a929, 'name': SearchDatastore_Task, 'duration_secs': 0.019973} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1070.553875] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e7e78d40-20a2-440d-97b0-123a2b672f3a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.560269] env[65726]: DEBUG oslo_vmware.api [None req-80fb9493-d54c-4f0a-a5fd-0197feeffb7b tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Waiting for the task: (returnval){ [ 1070.560269] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]523c37e9-0daa-b5fb-663e-8aa4394e9ba8" [ 1070.560269] env[65726]: _type = "Task" [ 1070.560269] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1070.571306] env[65726]: DEBUG oslo_vmware.api [None req-80fb9493-d54c-4f0a-a5fd-0197feeffb7b tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]523c37e9-0daa-b5fb-663e-8aa4394e9ba8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.833611] env[65726]: INFO nova.compute.manager [None req-42911402-8543-4c4b-aab9-86db149010d2 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 5cdf1d61-b817-4986-a48c-804a2374497a] Took 19.88 seconds to build instance. [ 1070.894251] env[65726]: WARNING neutronclient.v2_0.client [None req-dc5e36ff-c9dd-4078-bec5-b01919099186 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1070.937607] env[65726]: DEBUG oslo_concurrency.lockutils [req-91caea6d-a041-4e9f-9ca7-19377c7191c0 req-3a008e24-62ff-4c75-9b36-0a4b62cbc341 service nova] Releasing lock "refresh_cache-3b4d9461-747a-461c-b231-a0de02f0cb83" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1070.943834] env[65726]: DEBUG nova.network.neutron [None req-dc5e36ff-c9dd-4078-bec5-b01919099186 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: a56be6fa-6027-46ee-9aa8-332fe4222d68] Port 4e2dfbcd-41c3-4aaa-a2ec-50c855c12307 binding to destination host cpu-1 is already ACTIVE {{(pid=65726) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3236}} [ 1071.017036] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff6d6376-8956-4bd5-9a3e-01ab3219a136 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.026712] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48095b1d-07c8-43ea-ba00-558d7ea27c72 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.065281] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a31746d2-2cd6-458b-9b31-5e52f20f6093 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.076883] env[65726]: DEBUG oslo_vmware.api [None req-80fb9493-d54c-4f0a-a5fd-0197feeffb7b tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]523c37e9-0daa-b5fb-663e-8aa4394e9ba8, 'name': SearchDatastore_Task, 'duration_secs': 0.015822} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1071.077271] env[65726]: DEBUG oslo_concurrency.lockutils [None req-80fb9493-d54c-4f0a-a5fd-0197feeffb7b tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1071.077529] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-80fb9493-d54c-4f0a-a5fd-0197feeffb7b tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore2] 3b4d9461-747a-461c-b231-a0de02f0cb83/3b4d9461-747a-461c-b231-a0de02f0cb83.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1071.078871] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64f70a72-677f-4fee-9cce-d24522ace3bc {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.082824] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b2498273-ef74-4601-9e01-7f7a4821c1f1 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.096682] env[65726]: DEBUG nova.compute.provider_tree [None req-91ff63ad-fe79-48d0-901e-fc0ed403574a tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1071.099815] env[65726]: DEBUG oslo_vmware.api [None req-80fb9493-d54c-4f0a-a5fd-0197feeffb7b tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Waiting for the task: (returnval){ [ 1071.099815] env[65726]: value = "task-5116643" [ 1071.099815] env[65726]: _type = "Task" [ 1071.099815] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1071.108371] env[65726]: DEBUG oslo_vmware.api [None req-80fb9493-d54c-4f0a-a5fd-0197feeffb7b tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Task: {'id': task-5116643, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1071.337025] env[65726]: DEBUG oslo_concurrency.lockutils [None req-42911402-8543-4c4b-aab9-86db149010d2 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Lock "5cdf1d61-b817-4986-a48c-804a2374497a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 21.394s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1071.456019] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9ca95ca0-14c4-4870-aa74-cd2a795fbbc9 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Acquiring lock "d8801049-fc34-4b2f-8d5e-b9459c3e03a6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1071.456019] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9ca95ca0-14c4-4870-aa74-cd2a795fbbc9 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Lock "d8801049-fc34-4b2f-8d5e-b9459c3e03a6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1071.600752] env[65726]: DEBUG nova.scheduler.client.report [None req-91ff63ad-fe79-48d0-901e-fc0ed403574a tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1071.619401] env[65726]: DEBUG oslo_vmware.api [None req-80fb9493-d54c-4f0a-a5fd-0197feeffb7b tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Task: {'id': task-5116643, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1071.964907] env[65726]: DEBUG nova.compute.manager [None req-9ca95ca0-14c4-4870-aa74-cd2a795fbbc9 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: d8801049-fc34-4b2f-8d5e-b9459c3e03a6] Starting instance... {{(pid=65726) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 1071.973788] env[65726]: DEBUG oslo_concurrency.lockutils [None req-dc5e36ff-c9dd-4078-bec5-b01919099186 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Acquiring lock "a56be6fa-6027-46ee-9aa8-332fe4222d68-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1071.974061] env[65726]: DEBUG oslo_concurrency.lockutils [None req-dc5e36ff-c9dd-4078-bec5-b01919099186 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Lock "a56be6fa-6027-46ee-9aa8-332fe4222d68-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1071.974267] env[65726]: DEBUG oslo_concurrency.lockutils [None req-dc5e36ff-c9dd-4078-bec5-b01919099186 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Lock "a56be6fa-6027-46ee-9aa8-332fe4222d68-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1072.032649] env[65726]: INFO nova.compute.manager [None req-e7763bc8-6024-44cb-aee3-214195b1fd06 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 5cdf1d61-b817-4986-a48c-804a2374497a] Rebuilding instance [ 1072.091887] env[65726]: DEBUG nova.compute.manager [None req-e7763bc8-6024-44cb-aee3-214195b1fd06 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 5cdf1d61-b817-4986-a48c-804a2374497a] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1072.093525] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c41cb91-bdf4-408b-a6ba-67e323185c41 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.111196] env[65726]: DEBUG oslo_concurrency.lockutils [None req-91ff63ad-fe79-48d0-901e-fc0ed403574a tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.342s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1072.111277] env[65726]: DEBUG nova.compute.manager [None req-91ff63ad-fe79-48d0-901e-fc0ed403574a tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: 0d42fa75-20e9-4646-9b08-17015b7f068c] Start building networks asynchronously for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1072.114385] env[65726]: DEBUG oslo_concurrency.lockutils [None req-8c26e8da-5d87-42a2-a8c5-15e9c65a8b9f tempest-InstanceActionsV221TestJSON-647428531 tempest-InstanceActionsV221TestJSON-647428531-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.314s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1072.115725] env[65726]: INFO nova.compute.claims [None req-8c26e8da-5d87-42a2-a8c5-15e9c65a8b9f tempest-InstanceActionsV221TestJSON-647428531 tempest-InstanceActionsV221TestJSON-647428531-project-member] [instance: b0fec5ff-8c1e-41f9-b30a-6253a34baeec] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1072.125884] env[65726]: DEBUG oslo_vmware.api [None req-80fb9493-d54c-4f0a-a5fd-0197feeffb7b tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Task: {'id': task-5116643, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.670962} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1072.126127] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-80fb9493-d54c-4f0a-a5fd-0197feeffb7b tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore2] 3b4d9461-747a-461c-b231-a0de02f0cb83/3b4d9461-747a-461c-b231-a0de02f0cb83.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1072.126364] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-80fb9493-d54c-4f0a-a5fd-0197feeffb7b tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] [instance: 3b4d9461-747a-461c-b231-a0de02f0cb83] Extending root virtual disk to 1048576 {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1072.126645] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-de9495cf-a324-4a3f-9ad6-30bd4197096c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.136164] env[65726]: DEBUG oslo_vmware.api [None req-80fb9493-d54c-4f0a-a5fd-0197feeffb7b tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Waiting for the task: (returnval){ [ 1072.136164] env[65726]: value = "task-5116644" [ 1072.136164] env[65726]: _type = "Task" [ 1072.136164] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1072.153022] env[65726]: DEBUG oslo_vmware.api [None req-80fb9493-d54c-4f0a-a5fd-0197feeffb7b tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Task: {'id': task-5116644, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1072.496381] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9ca95ca0-14c4-4870-aa74-cd2a795fbbc9 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1072.624308] env[65726]: DEBUG nova.compute.utils [None req-91ff63ad-fe79-48d0-901e-fc0ed403574a tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Using /dev/sd instead of None {{(pid=65726) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1072.626724] env[65726]: DEBUG nova.compute.manager [None req-91ff63ad-fe79-48d0-901e-fc0ed403574a tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: 0d42fa75-20e9-4646-9b08-17015b7f068c] Allocating IP information in the background. {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1072.626874] env[65726]: DEBUG nova.network.neutron [None req-91ff63ad-fe79-48d0-901e-fc0ed403574a tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: 0d42fa75-20e9-4646-9b08-17015b7f068c] allocate_for_instance() {{(pid=65726) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 1072.627325] env[65726]: WARNING neutronclient.v2_0.client [None req-91ff63ad-fe79-48d0-901e-fc0ed403574a tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1072.627741] env[65726]: WARNING neutronclient.v2_0.client [None req-91ff63ad-fe79-48d0-901e-fc0ed403574a tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1072.628441] env[65726]: WARNING openstack [None req-91ff63ad-fe79-48d0-901e-fc0ed403574a tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1072.628893] env[65726]: WARNING openstack [None req-91ff63ad-fe79-48d0-901e-fc0ed403574a tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1072.637359] env[65726]: DEBUG nova.compute.manager [None req-91ff63ad-fe79-48d0-901e-fc0ed403574a tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: 0d42fa75-20e9-4646-9b08-17015b7f068c] Start building block device mappings for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1072.651053] env[65726]: DEBUG oslo_vmware.api [None req-80fb9493-d54c-4f0a-a5fd-0197feeffb7b tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Task: {'id': task-5116644, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.145509} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1072.651153] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-80fb9493-d54c-4f0a-a5fd-0197feeffb7b tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] [instance: 3b4d9461-747a-461c-b231-a0de02f0cb83] Extended root virtual disk {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1072.652033] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dff52231-6e44-4a87-aa87-805f209a4f64 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.676103] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-80fb9493-d54c-4f0a-a5fd-0197feeffb7b tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] [instance: 3b4d9461-747a-461c-b231-a0de02f0cb83] Reconfiguring VM instance instance-00000060 to attach disk [datastore2] 3b4d9461-747a-461c-b231-a0de02f0cb83/3b4d9461-747a-461c-b231-a0de02f0cb83.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1072.676463] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ffb7449c-9571-4179-b835-c03b780dae52 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.694692] env[65726]: DEBUG nova.policy [None req-91ff63ad-fe79-48d0-901e-fc0ed403574a tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9b65b9194933469c908c7cc04478444b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd46ccb00794f458b85da4a93879139ee', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65726) authorize /opt/stack/nova/nova/policy.py:192}} [ 1072.703934] env[65726]: DEBUG oslo_vmware.api [None req-80fb9493-d54c-4f0a-a5fd-0197feeffb7b tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Waiting for the task: (returnval){ [ 1072.703934] env[65726]: value = "task-5116645" [ 1072.703934] env[65726]: _type = "Task" [ 1072.703934] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1072.714533] env[65726]: DEBUG oslo_vmware.api [None req-80fb9493-d54c-4f0a-a5fd-0197feeffb7b tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Task: {'id': task-5116645, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1072.983394] env[65726]: WARNING openstack [None req-dc5e36ff-c9dd-4078-bec5-b01919099186 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1072.983878] env[65726]: WARNING openstack [None req-dc5e36ff-c9dd-4078-bec5-b01919099186 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1073.011970] env[65726]: DEBUG nova.network.neutron [None req-91ff63ad-fe79-48d0-901e-fc0ed403574a tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: 0d42fa75-20e9-4646-9b08-17015b7f068c] Successfully created port: 7a95cb45-ebfc-4ac0-ad58-4bb57170e47a {{(pid=65726) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1073.045191] env[65726]: DEBUG oslo_concurrency.lockutils [None req-dc5e36ff-c9dd-4078-bec5-b01919099186 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Acquiring lock "refresh_cache-a56be6fa-6027-46ee-9aa8-332fe4222d68" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1073.045456] env[65726]: DEBUG oslo_concurrency.lockutils [None req-dc5e36ff-c9dd-4078-bec5-b01919099186 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Acquired lock "refresh_cache-a56be6fa-6027-46ee-9aa8-332fe4222d68" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1073.045583] env[65726]: DEBUG nova.network.neutron [None req-dc5e36ff-c9dd-4078-bec5-b01919099186 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: a56be6fa-6027-46ee-9aa8-332fe4222d68] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1073.110066] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-e7763bc8-6024-44cb-aee3-214195b1fd06 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 5cdf1d61-b817-4986-a48c-804a2374497a] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1073.110684] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f3184e7c-9178-4334-9135-4c87332f93ab {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.120299] env[65726]: DEBUG oslo_vmware.api [None req-e7763bc8-6024-44cb-aee3-214195b1fd06 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Waiting for the task: (returnval){ [ 1073.120299] env[65726]: value = "task-5116646" [ 1073.120299] env[65726]: _type = "Task" [ 1073.120299] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1073.134501] env[65726]: DEBUG oslo_vmware.api [None req-e7763bc8-6024-44cb-aee3-214195b1fd06 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': task-5116646, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1073.219741] env[65726]: DEBUG oslo_vmware.api [None req-80fb9493-d54c-4f0a-a5fd-0197feeffb7b tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Task: {'id': task-5116645, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1073.366458] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31c50e4b-19bb-4d95-8cd3-4d37de3a7df5 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.375302] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-274a85a4-67a3-4b35-b1a8-ed5389143216 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.410242] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c104191-5117-4a55-91e5-7a4bf33471aa {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.419269] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4415008-8fec-40f7-89f1-6445f116a91a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.435870] env[65726]: DEBUG nova.compute.provider_tree [None req-8c26e8da-5d87-42a2-a8c5-15e9c65a8b9f tempest-InstanceActionsV221TestJSON-647428531 tempest-InstanceActionsV221TestJSON-647428531-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1073.550637] env[65726]: WARNING openstack [None req-dc5e36ff-c9dd-4078-bec5-b01919099186 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1073.551097] env[65726]: WARNING openstack [None req-dc5e36ff-c9dd-4078-bec5-b01919099186 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1073.630790] env[65726]: DEBUG oslo_vmware.api [None req-e7763bc8-6024-44cb-aee3-214195b1fd06 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': task-5116646, 'name': PowerOffVM_Task, 'duration_secs': 0.215184} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1073.635165] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-e7763bc8-6024-44cb-aee3-214195b1fd06 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 5cdf1d61-b817-4986-a48c-804a2374497a] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1073.635419] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-e7763bc8-6024-44cb-aee3-214195b1fd06 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 5cdf1d61-b817-4986-a48c-804a2374497a] Destroying instance {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1073.636491] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78e57998-f6da-48d3-8bad-52517fc752d2 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.645689] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-e7763bc8-6024-44cb-aee3-214195b1fd06 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 5cdf1d61-b817-4986-a48c-804a2374497a] Unregistering the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1073.645822] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9a50c5cb-69f2-4751-bd94-61d3259786ec {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.648239] env[65726]: DEBUG nova.compute.manager [None req-91ff63ad-fe79-48d0-901e-fc0ed403574a tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: 0d42fa75-20e9-4646-9b08-17015b7f068c] Start spawning the instance on the hypervisor. {{(pid=65726) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1073.679573] env[65726]: WARNING openstack [None req-dc5e36ff-c9dd-4078-bec5-b01919099186 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1073.680071] env[65726]: WARNING openstack [None req-dc5e36ff-c9dd-4078-bec5-b01919099186 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1073.691240] env[65726]: DEBUG nova.virt.hardware [None req-91ff63ad-fe79-48d0-901e-fc0ed403574a tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T19:28:34Z,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1539f06fa3534e90acbc3a60c4b8bd3f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T19:28:35Z,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1073.691533] env[65726]: DEBUG nova.virt.hardware [None req-91ff63ad-fe79-48d0-901e-fc0ed403574a tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1073.691710] env[65726]: DEBUG nova.virt.hardware [None req-91ff63ad-fe79-48d0-901e-fc0ed403574a tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1073.691909] env[65726]: DEBUG nova.virt.hardware [None req-91ff63ad-fe79-48d0-901e-fc0ed403574a tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1073.692132] env[65726]: DEBUG nova.virt.hardware [None req-91ff63ad-fe79-48d0-901e-fc0ed403574a tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1073.692294] env[65726]: DEBUG nova.virt.hardware [None req-91ff63ad-fe79-48d0-901e-fc0ed403574a tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1073.692581] env[65726]: DEBUG nova.virt.hardware [None req-91ff63ad-fe79-48d0-901e-fc0ed403574a tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1073.692798] env[65726]: DEBUG nova.virt.hardware [None req-91ff63ad-fe79-48d0-901e-fc0ed403574a tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1073.692977] env[65726]: DEBUG nova.virt.hardware [None req-91ff63ad-fe79-48d0-901e-fc0ed403574a tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1073.693164] env[65726]: DEBUG nova.virt.hardware [None req-91ff63ad-fe79-48d0-901e-fc0ed403574a tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1073.693341] env[65726]: DEBUG nova.virt.hardware [None req-91ff63ad-fe79-48d0-901e-fc0ed403574a tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1073.694625] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51e8e883-01aa-4dea-995d-f307de46a915 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.704613] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80aa7c47-6142-4521-bf0b-aa2527d9b581 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.719308] env[65726]: DEBUG oslo_vmware.api [None req-80fb9493-d54c-4f0a-a5fd-0197feeffb7b tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Task: {'id': task-5116645, 'name': ReconfigVM_Task, 'duration_secs': 0.51975} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1073.733044] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-80fb9493-d54c-4f0a-a5fd-0197feeffb7b tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] [instance: 3b4d9461-747a-461c-b231-a0de02f0cb83] Reconfigured VM instance instance-00000060 to attach disk [datastore2] 3b4d9461-747a-461c-b231-a0de02f0cb83/3b4d9461-747a-461c-b231-a0de02f0cb83.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1073.735614] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9e1d1b03-dedb-4530-841d-b5876e8c0a79 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.737689] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-e7763bc8-6024-44cb-aee3-214195b1fd06 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 5cdf1d61-b817-4986-a48c-804a2374497a] Unregistered the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1073.737883] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-e7763bc8-6024-44cb-aee3-214195b1fd06 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 5cdf1d61-b817-4986-a48c-804a2374497a] Deleting contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1073.738081] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-e7763bc8-6024-44cb-aee3-214195b1fd06 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Deleting the datastore file [datastore1] 5cdf1d61-b817-4986-a48c-804a2374497a {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1073.738363] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-137a3761-2a24-4b63-8a15-91574f20836b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.749235] env[65726]: DEBUG oslo_vmware.api [None req-80fb9493-d54c-4f0a-a5fd-0197feeffb7b tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Waiting for the task: (returnval){ [ 1073.749235] env[65726]: value = "task-5116648" [ 1073.749235] env[65726]: _type = "Task" [ 1073.749235] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1073.749543] env[65726]: DEBUG oslo_vmware.api [None req-e7763bc8-6024-44cb-aee3-214195b1fd06 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Waiting for the task: (returnval){ [ 1073.749543] env[65726]: value = "task-5116649" [ 1073.749543] env[65726]: _type = "Task" [ 1073.749543] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1073.767357] env[65726]: DEBUG oslo_vmware.api [None req-80fb9493-d54c-4f0a-a5fd-0197feeffb7b tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Task: {'id': task-5116648, 'name': Rename_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1073.767966] env[65726]: DEBUG oslo_vmware.api [None req-e7763bc8-6024-44cb-aee3-214195b1fd06 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': task-5116649, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1073.771887] env[65726]: WARNING openstack [None req-dc5e36ff-c9dd-4078-bec5-b01919099186 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1073.772434] env[65726]: WARNING openstack [None req-dc5e36ff-c9dd-4078-bec5-b01919099186 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1073.890926] env[65726]: DEBUG nova.network.neutron [None req-dc5e36ff-c9dd-4078-bec5-b01919099186 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: a56be6fa-6027-46ee-9aa8-332fe4222d68] Updating instance_info_cache with network_info: [{"id": "4e2dfbcd-41c3-4aaa-a2ec-50c855c12307", "address": "fa:16:3e:1d:82:59", "network": {"id": "0d0f561d-ac41-4ef5-9926-c707409799f7", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1134321217-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.160", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4eece77569624f90bf64e5c51974173f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2ce62383-8e84-4e26-955b-74c11392f4c9", "external-id": "nsx-vlan-transportzone-215", "segmentation_id": 215, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4e2dfbcd-41", "ovs_interfaceid": "4e2dfbcd-41c3-4aaa-a2ec-50c855c12307", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1073.940615] env[65726]: DEBUG nova.scheduler.client.report [None req-8c26e8da-5d87-42a2-a8c5-15e9c65a8b9f tempest-InstanceActionsV221TestJSON-647428531 tempest-InstanceActionsV221TestJSON-647428531-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1074.266152] env[65726]: DEBUG oslo_vmware.api [None req-e7763bc8-6024-44cb-aee3-214195b1fd06 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': task-5116649, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.468871} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1074.266450] env[65726]: DEBUG oslo_vmware.api [None req-80fb9493-d54c-4f0a-a5fd-0197feeffb7b tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Task: {'id': task-5116648, 'name': Rename_Task, 'duration_secs': 0.335509} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1074.267020] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-e7763bc8-6024-44cb-aee3-214195b1fd06 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Deleted the datastore file {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1074.267020] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-e7763bc8-6024-44cb-aee3-214195b1fd06 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 5cdf1d61-b817-4986-a48c-804a2374497a] Deleted contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1074.267020] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-e7763bc8-6024-44cb-aee3-214195b1fd06 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 5cdf1d61-b817-4986-a48c-804a2374497a] Instance destroyed {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1074.270029] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-80fb9493-d54c-4f0a-a5fd-0197feeffb7b tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] [instance: 3b4d9461-747a-461c-b231-a0de02f0cb83] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1074.270187] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0d7ea14f-bc69-4d1e-97c0-16596975ac59 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.278099] env[65726]: DEBUG oslo_vmware.api [None req-80fb9493-d54c-4f0a-a5fd-0197feeffb7b tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Waiting for the task: (returnval){ [ 1074.278099] env[65726]: value = "task-5116650" [ 1074.278099] env[65726]: _type = "Task" [ 1074.278099] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1074.287713] env[65726]: DEBUG oslo_vmware.api [None req-80fb9493-d54c-4f0a-a5fd-0197feeffb7b tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Task: {'id': task-5116650, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1074.394318] env[65726]: DEBUG oslo_concurrency.lockutils [None req-dc5e36ff-c9dd-4078-bec5-b01919099186 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Releasing lock "refresh_cache-a56be6fa-6027-46ee-9aa8-332fe4222d68" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1074.448905] env[65726]: DEBUG oslo_concurrency.lockutils [None req-8c26e8da-5d87-42a2-a8c5-15e9c65a8b9f tempest-InstanceActionsV221TestJSON-647428531 tempest-InstanceActionsV221TestJSON-647428531-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.333s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1074.448905] env[65726]: DEBUG nova.compute.manager [None req-8c26e8da-5d87-42a2-a8c5-15e9c65a8b9f tempest-InstanceActionsV221TestJSON-647428531 tempest-InstanceActionsV221TestJSON-647428531-project-member] [instance: b0fec5ff-8c1e-41f9-b30a-6253a34baeec] Start building networks asynchronously for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1074.451298] env[65726]: DEBUG oslo_concurrency.lockutils [None req-ab8fab95-1b19-4328-a74a-7a9a0971be1c tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 6.809s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1074.451476] env[65726]: DEBUG nova.objects.instance [None req-ab8fab95-1b19-4328-a74a-7a9a0971be1c tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Lazy-loading 'resources' on Instance uuid 48923e3d-b00d-4034-bacf-7be82009fb08 {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1074.567309] env[65726]: DEBUG nova.compute.manager [req-d92190f1-2df0-4812-bc4c-24fe98f68426 req-b6507eef-4dea-4592-a40b-e8a99e3fb47e service nova] [instance: 0d42fa75-20e9-4646-9b08-17015b7f068c] Received event network-vif-plugged-7a95cb45-ebfc-4ac0-ad58-4bb57170e47a {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1074.567309] env[65726]: DEBUG oslo_concurrency.lockutils [req-d92190f1-2df0-4812-bc4c-24fe98f68426 req-b6507eef-4dea-4592-a40b-e8a99e3fb47e service nova] Acquiring lock "0d42fa75-20e9-4646-9b08-17015b7f068c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1074.567309] env[65726]: DEBUG oslo_concurrency.lockutils [req-d92190f1-2df0-4812-bc4c-24fe98f68426 req-b6507eef-4dea-4592-a40b-e8a99e3fb47e service nova] Lock "0d42fa75-20e9-4646-9b08-17015b7f068c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1074.567309] env[65726]: DEBUG oslo_concurrency.lockutils [req-d92190f1-2df0-4812-bc4c-24fe98f68426 req-b6507eef-4dea-4592-a40b-e8a99e3fb47e service nova] Lock "0d42fa75-20e9-4646-9b08-17015b7f068c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1074.567309] env[65726]: DEBUG nova.compute.manager [req-d92190f1-2df0-4812-bc4c-24fe98f68426 req-b6507eef-4dea-4592-a40b-e8a99e3fb47e service nova] [instance: 0d42fa75-20e9-4646-9b08-17015b7f068c] No waiting events found dispatching network-vif-plugged-7a95cb45-ebfc-4ac0-ad58-4bb57170e47a {{(pid=65726) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1074.567309] env[65726]: WARNING nova.compute.manager [req-d92190f1-2df0-4812-bc4c-24fe98f68426 req-b6507eef-4dea-4592-a40b-e8a99e3fb47e service nova] [instance: 0d42fa75-20e9-4646-9b08-17015b7f068c] Received unexpected event network-vif-plugged-7a95cb45-ebfc-4ac0-ad58-4bb57170e47a for instance with vm_state building and task_state spawning. [ 1074.680064] env[65726]: DEBUG nova.network.neutron [None req-91ff63ad-fe79-48d0-901e-fc0ed403574a tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: 0d42fa75-20e9-4646-9b08-17015b7f068c] Successfully updated port: 7a95cb45-ebfc-4ac0-ad58-4bb57170e47a {{(pid=65726) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 1074.791362] env[65726]: DEBUG oslo_vmware.api [None req-80fb9493-d54c-4f0a-a5fd-0197feeffb7b tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Task: {'id': task-5116650, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1074.906332] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b7af4ab-f6ee-42f9-af0d-6cee988de243 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.918444] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df3ff4db-db05-4a9e-8c6e-d71156ff8a8e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.955787] env[65726]: DEBUG nova.compute.utils [None req-8c26e8da-5d87-42a2-a8c5-15e9c65a8b9f tempest-InstanceActionsV221TestJSON-647428531 tempest-InstanceActionsV221TestJSON-647428531-project-member] Using /dev/sd instead of None {{(pid=65726) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1074.956977] env[65726]: DEBUG nova.compute.manager [None req-8c26e8da-5d87-42a2-a8c5-15e9c65a8b9f tempest-InstanceActionsV221TestJSON-647428531 tempest-InstanceActionsV221TestJSON-647428531-project-member] [instance: b0fec5ff-8c1e-41f9-b30a-6253a34baeec] Allocating IP information in the background. {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1074.957244] env[65726]: DEBUG nova.network.neutron [None req-8c26e8da-5d87-42a2-a8c5-15e9c65a8b9f tempest-InstanceActionsV221TestJSON-647428531 tempest-InstanceActionsV221TestJSON-647428531-project-member] [instance: b0fec5ff-8c1e-41f9-b30a-6253a34baeec] allocate_for_instance() {{(pid=65726) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 1074.957685] env[65726]: WARNING neutronclient.v2_0.client [None req-8c26e8da-5d87-42a2-a8c5-15e9c65a8b9f tempest-InstanceActionsV221TestJSON-647428531 tempest-InstanceActionsV221TestJSON-647428531-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1074.958062] env[65726]: WARNING neutronclient.v2_0.client [None req-8c26e8da-5d87-42a2-a8c5-15e9c65a8b9f tempest-InstanceActionsV221TestJSON-647428531 tempest-InstanceActionsV221TestJSON-647428531-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1074.959152] env[65726]: WARNING openstack [None req-8c26e8da-5d87-42a2-a8c5-15e9c65a8b9f tempest-InstanceActionsV221TestJSON-647428531 tempest-InstanceActionsV221TestJSON-647428531-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1074.960347] env[65726]: WARNING openstack [None req-8c26e8da-5d87-42a2-a8c5-15e9c65a8b9f tempest-InstanceActionsV221TestJSON-647428531 tempest-InstanceActionsV221TestJSON-647428531-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1075.018698] env[65726]: DEBUG nova.policy [None req-8c26e8da-5d87-42a2-a8c5-15e9c65a8b9f tempest-InstanceActionsV221TestJSON-647428531 tempest-InstanceActionsV221TestJSON-647428531-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '82bfa5a365314f1aac92cc89c043fa5a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a6424f20b3b849e5ad82327218765b0c', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65726) authorize /opt/stack/nova/nova/policy.py:192}} [ 1075.183062] env[65726]: DEBUG oslo_concurrency.lockutils [None req-91ff63ad-fe79-48d0-901e-fc0ed403574a tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Acquiring lock "refresh_cache-0d42fa75-20e9-4646-9b08-17015b7f068c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1075.183267] env[65726]: DEBUG oslo_concurrency.lockutils [None req-91ff63ad-fe79-48d0-901e-fc0ed403574a tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Acquired lock "refresh_cache-0d42fa75-20e9-4646-9b08-17015b7f068c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1075.183499] env[65726]: DEBUG nova.network.neutron [None req-91ff63ad-fe79-48d0-901e-fc0ed403574a tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: 0d42fa75-20e9-4646-9b08-17015b7f068c] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1075.191725] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e27e931-8dc8-4eaa-8bd5-35e339ef68ef {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.203132] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2e8c835-5ed8-4bfa-838b-c61eeab6e4f8 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.241387] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92da5529-23c1-4e1a-a190-cbfcacdb6f9c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.247862] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34ba9d78-07c1-412c-a7fb-cd6305bc15db {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.265247] env[65726]: DEBUG nova.compute.provider_tree [None req-ab8fab95-1b19-4328-a74a-7a9a0971be1c tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1075.271158] env[65726]: DEBUG nova.network.neutron [None req-8c26e8da-5d87-42a2-a8c5-15e9c65a8b9f tempest-InstanceActionsV221TestJSON-647428531 tempest-InstanceActionsV221TestJSON-647428531-project-member] [instance: b0fec5ff-8c1e-41f9-b30a-6253a34baeec] Successfully created port: f11556ce-cfc0-4593-ac46-59266a693222 {{(pid=65726) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1075.293673] env[65726]: DEBUG oslo_vmware.api [None req-80fb9493-d54c-4f0a-a5fd-0197feeffb7b tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Task: {'id': task-5116650, 'name': PowerOnVM_Task, 'duration_secs': 0.720562} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1075.294113] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-80fb9493-d54c-4f0a-a5fd-0197feeffb7b tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] [instance: 3b4d9461-747a-461c-b231-a0de02f0cb83] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1075.294619] env[65726]: INFO nova.compute.manager [None req-80fb9493-d54c-4f0a-a5fd-0197feeffb7b tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] [instance: 3b4d9461-747a-461c-b231-a0de02f0cb83] Took 8.68 seconds to spawn the instance on the hypervisor. [ 1075.294817] env[65726]: DEBUG nova.compute.manager [None req-80fb9493-d54c-4f0a-a5fd-0197feeffb7b tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] [instance: 3b4d9461-747a-461c-b231-a0de02f0cb83] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1075.295664] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-248c7675-87d5-43b2-bead-0def5713bce8 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.314210] env[65726]: DEBUG nova.virt.hardware [None req-e7763bc8-6024-44cb-aee3-214195b1fd06 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T19:28:34Z,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1539f06fa3534e90acbc3a60c4b8bd3f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T19:28:35Z,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1075.314657] env[65726]: DEBUG nova.virt.hardware [None req-e7763bc8-6024-44cb-aee3-214195b1fd06 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1075.314846] env[65726]: DEBUG nova.virt.hardware [None req-e7763bc8-6024-44cb-aee3-214195b1fd06 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1075.315045] env[65726]: DEBUG nova.virt.hardware [None req-e7763bc8-6024-44cb-aee3-214195b1fd06 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1075.315191] env[65726]: DEBUG nova.virt.hardware [None req-e7763bc8-6024-44cb-aee3-214195b1fd06 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1075.315322] env[65726]: DEBUG nova.virt.hardware [None req-e7763bc8-6024-44cb-aee3-214195b1fd06 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1075.315534] env[65726]: DEBUG nova.virt.hardware [None req-e7763bc8-6024-44cb-aee3-214195b1fd06 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1075.315724] env[65726]: DEBUG nova.virt.hardware [None req-e7763bc8-6024-44cb-aee3-214195b1fd06 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1075.315984] env[65726]: DEBUG nova.virt.hardware [None req-e7763bc8-6024-44cb-aee3-214195b1fd06 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1075.316074] env[65726]: DEBUG nova.virt.hardware [None req-e7763bc8-6024-44cb-aee3-214195b1fd06 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1075.316812] env[65726]: DEBUG nova.virt.hardware [None req-e7763bc8-6024-44cb-aee3-214195b1fd06 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1075.317506] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d658ffa9-9208-4789-b80a-7e7ec52f363b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.329646] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92ca4b55-9aee-4efa-ba1d-2a3f310b05f4 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.346396] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-e7763bc8-6024-44cb-aee3-214195b1fd06 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 5cdf1d61-b817-4986-a48c-804a2374497a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:4d:82:49', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '74f30339-6421-4654-bddb-81d7f34db9d7', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '94113d1c-2016-4a98-b42a-6ae90eee0b24', 'vif_model': 'vmxnet3'}] {{(pid=65726) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1075.353921] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-e7763bc8-6024-44cb-aee3-214195b1fd06 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1075.354666] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5cdf1d61-b817-4986-a48c-804a2374497a] Creating VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1075.354866] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-fb84f7fb-e23d-47ed-a8c1-b3614cb0aa7a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.377028] env[65726]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1075.377028] env[65726]: value = "task-5116651" [ 1075.377028] env[65726]: _type = "Task" [ 1075.377028] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1075.387183] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116651, 'name': CreateVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1075.470901] env[65726]: DEBUG nova.compute.manager [None req-8c26e8da-5d87-42a2-a8c5-15e9c65a8b9f tempest-InstanceActionsV221TestJSON-647428531 tempest-InstanceActionsV221TestJSON-647428531-project-member] [instance: b0fec5ff-8c1e-41f9-b30a-6253a34baeec] Start building block device mappings for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1075.687763] env[65726]: WARNING openstack [None req-91ff63ad-fe79-48d0-901e-fc0ed403574a tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1075.688190] env[65726]: WARNING openstack [None req-91ff63ad-fe79-48d0-901e-fc0ed403574a tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1075.693459] env[65726]: DEBUG nova.network.neutron [None req-91ff63ad-fe79-48d0-901e-fc0ed403574a tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: 0d42fa75-20e9-4646-9b08-17015b7f068c] Instance cache missing network info. {{(pid=65726) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 1075.770223] env[65726]: WARNING openstack [None req-91ff63ad-fe79-48d0-901e-fc0ed403574a tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1075.770223] env[65726]: WARNING openstack [None req-91ff63ad-fe79-48d0-901e-fc0ed403574a tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1075.777533] env[65726]: DEBUG nova.scheduler.client.report [None req-ab8fab95-1b19-4328-a74a-7a9a0971be1c tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1075.816257] env[65726]: INFO nova.compute.manager [None req-80fb9493-d54c-4f0a-a5fd-0197feeffb7b tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] [instance: 3b4d9461-747a-461c-b231-a0de02f0cb83] Took 23.80 seconds to build instance. [ 1075.871101] env[65726]: WARNING openstack [None req-91ff63ad-fe79-48d0-901e-fc0ed403574a tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1075.871505] env[65726]: WARNING openstack [None req-91ff63ad-fe79-48d0-901e-fc0ed403574a tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1075.891307] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116651, 'name': CreateVM_Task, 'duration_secs': 0.504657} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1075.893621] env[65726]: DEBUG oslo_vmware.rw_handles [None req-f263eae6-96b2-45f2-b03a-d97fe5d0f1d8 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52fbf44b-bcce-284e-0923-049a7f39cb78/disk-0.vmdk. {{(pid=65726) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1075.893849] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5cdf1d61-b817-4986-a48c-804a2374497a] Created VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1075.894725] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd30f855-27d0-45d4-8b42-b29c1a10636a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.897935] env[65726]: WARNING openstack [None req-e7763bc8-6024-44cb-aee3-214195b1fd06 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1075.898353] env[65726]: WARNING openstack [None req-e7763bc8-6024-44cb-aee3-214195b1fd06 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1075.903386] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e7763bc8-6024-44cb-aee3-214195b1fd06 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1075.903527] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e7763bc8-6024-44cb-aee3-214195b1fd06 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1075.903803] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e7763bc8-6024-44cb-aee3-214195b1fd06 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1075.904257] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3cd323ae-0ac4-469f-9f53-d2c037517a37 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.910093] env[65726]: DEBUG oslo_vmware.rw_handles [None req-f263eae6-96b2-45f2-b03a-d97fe5d0f1d8 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52fbf44b-bcce-284e-0923-049a7f39cb78/disk-0.vmdk is in state: ready. {{(pid=65726) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1075.910625] env[65726]: ERROR oslo_vmware.rw_handles [None req-f263eae6-96b2-45f2-b03a-d97fe5d0f1d8 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52fbf44b-bcce-284e-0923-049a7f39cb78/disk-0.vmdk due to incomplete transfer. [ 1075.911552] env[65726]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-6bf639ff-569c-4b60-a889-1aea3c3aed1c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.913319] env[65726]: DEBUG oslo_vmware.api [None req-e7763bc8-6024-44cb-aee3-214195b1fd06 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Waiting for the task: (returnval){ [ 1075.913319] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52813651-7bfe-598e-ffda-7ac2dda69ebf" [ 1075.913319] env[65726]: _type = "Task" [ 1075.913319] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1075.923062] env[65726]: DEBUG oslo_vmware.api [None req-e7763bc8-6024-44cb-aee3-214195b1fd06 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52813651-7bfe-598e-ffda-7ac2dda69ebf, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1075.928432] env[65726]: DEBUG oslo_vmware.rw_handles [None req-f263eae6-96b2-45f2-b03a-d97fe5d0f1d8 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52fbf44b-bcce-284e-0923-049a7f39cb78/disk-0.vmdk. {{(pid=65726) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1075.928633] env[65726]: DEBUG nova.virt.vmwareapi.images [None req-f263eae6-96b2-45f2-b03a-d97fe5d0f1d8 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] [instance: 6fd70b4a-63e4-4258-9cff-f3b582500b9f] Uploaded image eb421501-6112-4a68-9c48-c82505e774d2 to the Glance image server {{(pid=65726) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 1075.931871] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-f263eae6-96b2-45f2-b03a-d97fe5d0f1d8 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] [instance: 6fd70b4a-63e4-4258-9cff-f3b582500b9f] Destroying the VM {{(pid=65726) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 1075.932168] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-0280eca0-d943-413a-a7b9-5ccbb935c036 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.940412] env[65726]: DEBUG oslo_vmware.api [None req-f263eae6-96b2-45f2-b03a-d97fe5d0f1d8 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Waiting for the task: (returnval){ [ 1075.940412] env[65726]: value = "task-5116652" [ 1075.940412] env[65726]: _type = "Task" [ 1075.940412] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1075.949908] env[65726]: DEBUG oslo_vmware.api [None req-f263eae6-96b2-45f2-b03a-d97fe5d0f1d8 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Task: {'id': task-5116652, 'name': Destroy_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1075.972581] env[65726]: DEBUG nova.network.neutron [None req-91ff63ad-fe79-48d0-901e-fc0ed403574a tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: 0d42fa75-20e9-4646-9b08-17015b7f068c] Updating instance_info_cache with network_info: [{"id": "7a95cb45-ebfc-4ac0-ad58-4bb57170e47a", "address": "fa:16:3e:2b:ea:81", "network": {"id": "51a66448-0f72-4266-8868-856b48be839f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-155658961-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d46ccb00794f458b85da4a93879139ee", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b1f3e6c3-5584-4852-9017-476ab8ac4946", "external-id": "nsx-vlan-transportzone-304", "segmentation_id": 304, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7a95cb45-eb", "ovs_interfaceid": "7a95cb45-ebfc-4ac0-ad58-4bb57170e47a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1076.047784] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b85fe0f1-d7fb-4e47-a9de-2b131698ef7c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.922524] env[65726]: DEBUG oslo_concurrency.lockutils [None req-ab8fab95-1b19-4328-a74a-7a9a0971be1c tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.471s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1076.924741] env[65726]: DEBUG oslo_concurrency.lockutils [None req-80fb9493-d54c-4f0a-a5fd-0197feeffb7b tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Lock "3b4d9461-747a-461c-b231-a0de02f0cb83" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 25.918s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1076.925481] env[65726]: DEBUG oslo_concurrency.lockutils [None req-91ff63ad-fe79-48d0-901e-fc0ed403574a tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Releasing lock "refresh_cache-0d42fa75-20e9-4646-9b08-17015b7f068c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1076.925869] env[65726]: DEBUG nova.compute.manager [None req-91ff63ad-fe79-48d0-901e-fc0ed403574a tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: 0d42fa75-20e9-4646-9b08-17015b7f068c] Instance network_info: |[{"id": "7a95cb45-ebfc-4ac0-ad58-4bb57170e47a", "address": "fa:16:3e:2b:ea:81", "network": {"id": "51a66448-0f72-4266-8868-856b48be839f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-155658961-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d46ccb00794f458b85da4a93879139ee", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b1f3e6c3-5584-4852-9017-476ab8ac4946", "external-id": "nsx-vlan-transportzone-304", "segmentation_id": 304, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7a95cb45-eb", "ovs_interfaceid": "7a95cb45-ebfc-4ac0-ad58-4bb57170e47a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1076.926885] env[65726]: DEBUG nova.compute.manager [None req-8c26e8da-5d87-42a2-a8c5-15e9c65a8b9f tempest-InstanceActionsV221TestJSON-647428531 tempest-InstanceActionsV221TestJSON-647428531-project-member] [instance: b0fec5ff-8c1e-41f9-b30a-6253a34baeec] Start spawning the instance on the hypervisor. {{(pid=65726) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1076.933451] env[65726]: DEBUG nova.network.neutron [None req-8c26e8da-5d87-42a2-a8c5-15e9c65a8b9f tempest-InstanceActionsV221TestJSON-647428531 tempest-InstanceActionsV221TestJSON-647428531-project-member] [instance: b0fec5ff-8c1e-41f9-b30a-6253a34baeec] Successfully updated port: f11556ce-cfc0-4593-ac46-59266a693222 {{(pid=65726) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 1076.935235] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 8.960s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1076.935409] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1076.935770] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=65726) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:937}} [ 1076.935957] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9ca95ca0-14c4-4870-aa74-cd2a795fbbc9 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.440s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1076.937353] env[65726]: INFO nova.compute.claims [None req-9ca95ca0-14c4-4870-aa74-cd2a795fbbc9 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: d8801049-fc34-4b2f-8d5e-b9459c3e03a6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1076.947329] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-91ff63ad-fe79-48d0-901e-fc0ed403574a tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: 0d42fa75-20e9-4646-9b08-17015b7f068c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:2b:ea:81', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'b1f3e6c3-5584-4852-9017-476ab8ac4946', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7a95cb45-ebfc-4ac0-ad58-4bb57170e47a', 'vif_model': 'vmxnet3'}] {{(pid=65726) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1076.955246] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-91ff63ad-fe79-48d0-901e-fc0ed403574a tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1076.958975] env[65726]: DEBUG nova.compute.manager [req-9776b961-4d23-419a-aa3e-30ffbff82604 req-f8156270-90df-4e5b-8b3f-3e915dfebcfc service nova] [instance: 0d42fa75-20e9-4646-9b08-17015b7f068c] Received event network-changed-7a95cb45-ebfc-4ac0-ad58-4bb57170e47a {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1076.959198] env[65726]: DEBUG nova.compute.manager [req-9776b961-4d23-419a-aa3e-30ffbff82604 req-f8156270-90df-4e5b-8b3f-3e915dfebcfc service nova] [instance: 0d42fa75-20e9-4646-9b08-17015b7f068c] Refreshing instance network info cache due to event network-changed-7a95cb45-ebfc-4ac0-ad58-4bb57170e47a. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 1076.959364] env[65726]: DEBUG oslo_concurrency.lockutils [req-9776b961-4d23-419a-aa3e-30ffbff82604 req-f8156270-90df-4e5b-8b3f-3e915dfebcfc service nova] Acquiring lock "refresh_cache-0d42fa75-20e9-4646-9b08-17015b7f068c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1076.959506] env[65726]: DEBUG oslo_concurrency.lockutils [req-9776b961-4d23-419a-aa3e-30ffbff82604 req-f8156270-90df-4e5b-8b3f-3e915dfebcfc service nova] Acquired lock "refresh_cache-0d42fa75-20e9-4646-9b08-17015b7f068c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1076.959656] env[65726]: DEBUG nova.network.neutron [req-9776b961-4d23-419a-aa3e-30ffbff82604 req-f8156270-90df-4e5b-8b3f-3e915dfebcfc service nova] [instance: 0d42fa75-20e9-4646-9b08-17015b7f068c] Refreshing network info cache for port 7a95cb45-ebfc-4ac0-ad58-4bb57170e47a {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 1076.961812] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d609a04e-90f5-45f2-9c67-51ca2ccdacad {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.968497] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cccadc2d-bc92-4827-9f43-732c01baa477 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.971045] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0d42fa75-20e9-4646-9b08-17015b7f068c] Creating VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1076.972663] env[65726]: INFO nova.scheduler.client.report [None req-ab8fab95-1b19-4328-a74a-7a9a0971be1c tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Deleted allocations for instance 48923e3d-b00d-4034-bacf-7be82009fb08 [ 1076.982449] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ef506b5f-aef3-41ed-a463-409088625b77 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.006596] env[65726]: DEBUG nova.virt.hardware [None req-8c26e8da-5d87-42a2-a8c5-15e9c65a8b9f tempest-InstanceActionsV221TestJSON-647428531 tempest-InstanceActionsV221TestJSON-647428531-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T19:28:34Z,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1539f06fa3534e90acbc3a60c4b8bd3f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T19:28:35Z,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1077.007440] env[65726]: DEBUG nova.virt.hardware [None req-8c26e8da-5d87-42a2-a8c5-15e9c65a8b9f tempest-InstanceActionsV221TestJSON-647428531 tempest-InstanceActionsV221TestJSON-647428531-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1077.007750] env[65726]: DEBUG nova.virt.hardware [None req-8c26e8da-5d87-42a2-a8c5-15e9c65a8b9f tempest-InstanceActionsV221TestJSON-647428531 tempest-InstanceActionsV221TestJSON-647428531-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1077.008077] env[65726]: DEBUG nova.virt.hardware [None req-8c26e8da-5d87-42a2-a8c5-15e9c65a8b9f tempest-InstanceActionsV221TestJSON-647428531 tempest-InstanceActionsV221TestJSON-647428531-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1077.008325] env[65726]: DEBUG nova.virt.hardware [None req-8c26e8da-5d87-42a2-a8c5-15e9c65a8b9f tempest-InstanceActionsV221TestJSON-647428531 tempest-InstanceActionsV221TestJSON-647428531-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1077.008568] env[65726]: DEBUG nova.virt.hardware [None req-8c26e8da-5d87-42a2-a8c5-15e9c65a8b9f tempest-InstanceActionsV221TestJSON-647428531 tempest-InstanceActionsV221TestJSON-647428531-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1077.008897] env[65726]: DEBUG nova.virt.hardware [None req-8c26e8da-5d87-42a2-a8c5-15e9c65a8b9f tempest-InstanceActionsV221TestJSON-647428531 tempest-InstanceActionsV221TestJSON-647428531-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1077.009191] env[65726]: DEBUG nova.virt.hardware [None req-8c26e8da-5d87-42a2-a8c5-15e9c65a8b9f tempest-InstanceActionsV221TestJSON-647428531 tempest-InstanceActionsV221TestJSON-647428531-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1077.009481] env[65726]: DEBUG nova.virt.hardware [None req-8c26e8da-5d87-42a2-a8c5-15e9c65a8b9f tempest-InstanceActionsV221TestJSON-647428531 tempest-InstanceActionsV221TestJSON-647428531-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1077.009770] env[65726]: DEBUG nova.virt.hardware [None req-8c26e8da-5d87-42a2-a8c5-15e9c65a8b9f tempest-InstanceActionsV221TestJSON-647428531 tempest-InstanceActionsV221TestJSON-647428531-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1077.010079] env[65726]: DEBUG nova.virt.hardware [None req-8c26e8da-5d87-42a2-a8c5-15e9c65a8b9f tempest-InstanceActionsV221TestJSON-647428531 tempest-InstanceActionsV221TestJSON-647428531-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1077.011041] env[65726]: DEBUG oslo_vmware.api [None req-f263eae6-96b2-45f2-b03a-d97fe5d0f1d8 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Task: {'id': task-5116652, 'name': Destroy_Task, 'duration_secs': 0.426115} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1077.011429] env[65726]: DEBUG oslo_vmware.api [None req-e7763bc8-6024-44cb-aee3-214195b1fd06 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52813651-7bfe-598e-ffda-7ac2dda69ebf, 'name': SearchDatastore_Task, 'duration_secs': 0.012858} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1077.012797] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-070aaf10-6947-49bb-95c9-1bea43da8919 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.021654] env[65726]: INFO nova.virt.vmwareapi.vm_util [None req-f263eae6-96b2-45f2-b03a-d97fe5d0f1d8 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] [instance: 6fd70b4a-63e4-4258-9cff-f3b582500b9f] Destroyed the VM [ 1077.021654] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-f263eae6-96b2-45f2-b03a-d97fe5d0f1d8 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] [instance: 6fd70b4a-63e4-4258-9cff-f3b582500b9f] Deleting Snapshot of the VM instance {{(pid=65726) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1077.021972] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e7763bc8-6024-44cb-aee3-214195b1fd06 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1077.022246] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-e7763bc8-6024-44cb-aee3-214195b1fd06 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 5cdf1d61-b817-4986-a48c-804a2374497a] Processing image b52362a3-ee8a-4cbf-b06f-513b0cc8f95c {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1077.022526] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e7763bc8-6024-44cb-aee3-214195b1fd06 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1077.022703] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e7763bc8-6024-44cb-aee3-214195b1fd06 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1077.022868] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-e7763bc8-6024-44cb-aee3-214195b1fd06 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1077.023694] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-dc5e36ff-c9dd-4078-bec5-b01919099186 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: a56be6fa-6027-46ee-9aa8-332fe4222d68] Updating instance 'a56be6fa-6027-46ee-9aa8-332fe4222d68' progress to 83 {{(pid=65726) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1077.029514] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-76d60462-b3c8-4636-8fdf-9acc0863dd08 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.031419] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f00d0d8d-cc80-4326-a4fd-308c3e229b0c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.038301] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-feabe0f5-7ee1-477f-ac77-ef1fd621d2ef {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.039101] env[65726]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1077.039101] env[65726]: value = "task-5116653" [ 1077.039101] env[65726]: _type = "Task" [ 1077.039101] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1077.048425] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-053661ff-4b0a-4b60-9a0c-174fca6f2a3a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.065484] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-e7763bc8-6024-44cb-aee3-214195b1fd06 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1077.065484] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-e7763bc8-6024-44cb-aee3-214195b1fd06 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65726) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1077.066168] env[65726]: DEBUG oslo_vmware.api [None req-f263eae6-96b2-45f2-b03a-d97fe5d0f1d8 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Waiting for the task: (returnval){ [ 1077.066168] env[65726]: value = "task-5116654" [ 1077.066168] env[65726]: _type = "Task" [ 1077.066168] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1077.068446] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-69a78095-0f16-4586-b115-1ad7495085ef {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.071772] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e0672ae-1647-4eb6-a63d-af50f92ae5de {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.089885] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116653, 'name': CreateVM_Task} progress is 15%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1077.095533] env[65726]: DEBUG oslo_vmware.api [None req-e7763bc8-6024-44cb-aee3-214195b1fd06 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Waiting for the task: (returnval){ [ 1077.095533] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52d1e99c-fe20-cfa7-748b-4fbcb3136ea1" [ 1077.095533] env[65726]: _type = "Task" [ 1077.095533] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1077.100217] env[65726]: DEBUG oslo_vmware.api [None req-f263eae6-96b2-45f2-b03a-d97fe5d0f1d8 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Task: {'id': task-5116654, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1077.104624] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d8f026c-7078-45ef-8a58-e0ecf4fd3e20 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.116506] env[65726]: DEBUG oslo_vmware.api [None req-e7763bc8-6024-44cb-aee3-214195b1fd06 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52d1e99c-fe20-cfa7-748b-4fbcb3136ea1, 'name': SearchDatastore_Task, 'duration_secs': 0.013558} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1077.117137] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-96a4655f-3f8c-4037-8d19-8464e3f05246 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.144660] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=178789MB free_disk=95GB free_vcpus=48 pci_devices=None {{(pid=65726) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1136}} [ 1077.144810] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1077.149260] env[65726]: DEBUG oslo_vmware.api [None req-e7763bc8-6024-44cb-aee3-214195b1fd06 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Waiting for the task: (returnval){ [ 1077.149260] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52ee6e20-5417-5fe0-dd48-02ba134e3792" [ 1077.149260] env[65726]: _type = "Task" [ 1077.149260] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1077.158780] env[65726]: DEBUG oslo_vmware.api [None req-e7763bc8-6024-44cb-aee3-214195b1fd06 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52ee6e20-5417-5fe0-dd48-02ba134e3792, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1077.435606] env[65726]: DEBUG oslo_concurrency.lockutils [None req-8c26e8da-5d87-42a2-a8c5-15e9c65a8b9f tempest-InstanceActionsV221TestJSON-647428531 tempest-InstanceActionsV221TestJSON-647428531-project-member] Acquiring lock "refresh_cache-b0fec5ff-8c1e-41f9-b30a-6253a34baeec" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1077.435818] env[65726]: DEBUG oslo_concurrency.lockutils [None req-8c26e8da-5d87-42a2-a8c5-15e9c65a8b9f tempest-InstanceActionsV221TestJSON-647428531 tempest-InstanceActionsV221TestJSON-647428531-project-member] Acquired lock "refresh_cache-b0fec5ff-8c1e-41f9-b30a-6253a34baeec" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1077.435986] env[65726]: DEBUG nova.network.neutron [None req-8c26e8da-5d87-42a2-a8c5-15e9c65a8b9f tempest-InstanceActionsV221TestJSON-647428531 tempest-InstanceActionsV221TestJSON-647428531-project-member] [instance: b0fec5ff-8c1e-41f9-b30a-6253a34baeec] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1077.462987] env[65726]: WARNING openstack [req-9776b961-4d23-419a-aa3e-30ffbff82604 req-f8156270-90df-4e5b-8b3f-3e915dfebcfc service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1077.463380] env[65726]: WARNING openstack [req-9776b961-4d23-419a-aa3e-30ffbff82604 req-f8156270-90df-4e5b-8b3f-3e915dfebcfc service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1077.470309] env[65726]: INFO nova.compute.manager [None req-a4ccf9b3-2389-45a5-bcf4-8de43acbac4c tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] [instance: 3b4d9461-747a-461c-b231-a0de02f0cb83] Rescuing [ 1077.470559] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a4ccf9b3-2389-45a5-bcf4-8de43acbac4c tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Acquiring lock "refresh_cache-3b4d9461-747a-461c-b231-a0de02f0cb83" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1077.470708] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a4ccf9b3-2389-45a5-bcf4-8de43acbac4c tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Acquired lock "refresh_cache-3b4d9461-747a-461c-b231-a0de02f0cb83" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1077.470867] env[65726]: DEBUG nova.network.neutron [None req-a4ccf9b3-2389-45a5-bcf4-8de43acbac4c tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] [instance: 3b4d9461-747a-461c-b231-a0de02f0cb83] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1077.503595] env[65726]: DEBUG oslo_concurrency.lockutils [None req-ab8fab95-1b19-4328-a74a-7a9a0971be1c tempest-ImagesOneServerNegativeTestJSON-1125193517 tempest-ImagesOneServerNegativeTestJSON-1125193517-project-member] Lock "48923e3d-b00d-4034-bacf-7be82009fb08" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 13.277s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1077.543837] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-dc5e36ff-c9dd-4078-bec5-b01919099186 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: a56be6fa-6027-46ee-9aa8-332fe4222d68] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1077.544268] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9dcd163e-3701-4527-8c05-527f08cc3fab {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.558232] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116653, 'name': CreateVM_Task, 'duration_secs': 0.339888} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1077.559985] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0d42fa75-20e9-4646-9b08-17015b7f068c] Created VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1077.560445] env[65726]: DEBUG oslo_vmware.api [None req-dc5e36ff-c9dd-4078-bec5-b01919099186 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Waiting for the task: (returnval){ [ 1077.560445] env[65726]: value = "task-5116655" [ 1077.560445] env[65726]: _type = "Task" [ 1077.560445] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1077.565669] env[65726]: WARNING openstack [None req-91ff63ad-fe79-48d0-901e-fc0ed403574a tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1077.566141] env[65726]: WARNING openstack [None req-91ff63ad-fe79-48d0-901e-fc0ed403574a tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1077.571794] env[65726]: DEBUG oslo_concurrency.lockutils [None req-91ff63ad-fe79-48d0-901e-fc0ed403574a tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1077.572018] env[65726]: DEBUG oslo_concurrency.lockutils [None req-91ff63ad-fe79-48d0-901e-fc0ed403574a tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1077.572424] env[65726]: DEBUG oslo_concurrency.lockutils [None req-91ff63ad-fe79-48d0-901e-fc0ed403574a tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1077.573087] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2c70886d-e717-404e-96a9-b690bb68b0f4 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.589462] env[65726]: DEBUG oslo_vmware.api [None req-f263eae6-96b2-45f2-b03a-d97fe5d0f1d8 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Task: {'id': task-5116654, 'name': RemoveSnapshot_Task, 'duration_secs': 0.391473} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1077.593962] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-f263eae6-96b2-45f2-b03a-d97fe5d0f1d8 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] [instance: 6fd70b4a-63e4-4258-9cff-f3b582500b9f] Deleted Snapshot of the VM instance {{(pid=65726) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1077.594455] env[65726]: DEBUG nova.compute.manager [None req-f263eae6-96b2-45f2-b03a-d97fe5d0f1d8 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] [instance: 6fd70b4a-63e4-4258-9cff-f3b582500b9f] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1077.594784] env[65726]: DEBUG oslo_vmware.api [None req-dc5e36ff-c9dd-4078-bec5-b01919099186 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5116655, 'name': PowerOnVM_Task} progress is 33%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1077.595071] env[65726]: DEBUG oslo_vmware.api [None req-91ff63ad-fe79-48d0-901e-fc0ed403574a tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Waiting for the task: (returnval){ [ 1077.595071] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5275f91e-7ba1-ced4-538e-2a0f3a246652" [ 1077.595071] env[65726]: _type = "Task" [ 1077.595071] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1077.596574] env[65726]: WARNING openstack [req-9776b961-4d23-419a-aa3e-30ffbff82604 req-f8156270-90df-4e5b-8b3f-3e915dfebcfc service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1077.596936] env[65726]: WARNING openstack [req-9776b961-4d23-419a-aa3e-30ffbff82604 req-f8156270-90df-4e5b-8b3f-3e915dfebcfc service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1077.604915] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3166067-dbb0-47c1-b09c-d9007746e405 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.619015] env[65726]: DEBUG oslo_vmware.api [None req-91ff63ad-fe79-48d0-901e-fc0ed403574a tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5275f91e-7ba1-ced4-538e-2a0f3a246652, 'name': SearchDatastore_Task, 'duration_secs': 0.012196} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1077.621013] env[65726]: DEBUG oslo_concurrency.lockutils [None req-91ff63ad-fe79-48d0-901e-fc0ed403574a tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1077.621298] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-91ff63ad-fe79-48d0-901e-fc0ed403574a tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: 0d42fa75-20e9-4646-9b08-17015b7f068c] Processing image b52362a3-ee8a-4cbf-b06f-513b0cc8f95c {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1077.621504] env[65726]: DEBUG oslo_concurrency.lockutils [None req-91ff63ad-fe79-48d0-901e-fc0ed403574a tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1077.660623] env[65726]: DEBUG oslo_vmware.api [None req-e7763bc8-6024-44cb-aee3-214195b1fd06 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52ee6e20-5417-5fe0-dd48-02ba134e3792, 'name': SearchDatastore_Task, 'duration_secs': 0.015222} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1077.660929] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e7763bc8-6024-44cb-aee3-214195b1fd06 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1077.661220] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-e7763bc8-6024-44cb-aee3-214195b1fd06 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore2] 5cdf1d61-b817-4986-a48c-804a2374497a/5cdf1d61-b817-4986-a48c-804a2374497a.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1077.661512] env[65726]: DEBUG oslo_concurrency.lockutils [None req-91ff63ad-fe79-48d0-901e-fc0ed403574a tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1077.661730] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-91ff63ad-fe79-48d0-901e-fc0ed403574a tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1077.661970] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6b0f1619-0cf3-47c1-a5e4-0f6b2b71fba0 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.664132] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4fa27538-335d-41c9-b307-78d2e99d9148 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.669958] env[65726]: WARNING openstack [req-9776b961-4d23-419a-aa3e-30ffbff82604 req-f8156270-90df-4e5b-8b3f-3e915dfebcfc service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1077.670375] env[65726]: WARNING openstack [req-9776b961-4d23-419a-aa3e-30ffbff82604 req-f8156270-90df-4e5b-8b3f-3e915dfebcfc service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1077.685017] env[65726]: DEBUG oslo_vmware.api [None req-e7763bc8-6024-44cb-aee3-214195b1fd06 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Waiting for the task: (returnval){ [ 1077.685017] env[65726]: value = "task-5116656" [ 1077.685017] env[65726]: _type = "Task" [ 1077.685017] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1077.686869] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-91ff63ad-fe79-48d0-901e-fc0ed403574a tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1077.687148] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-91ff63ad-fe79-48d0-901e-fc0ed403574a tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65726) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1077.691652] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-86af6fea-8e3e-42da-929b-834fdda6376c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.700097] env[65726]: DEBUG oslo_vmware.api [None req-e7763bc8-6024-44cb-aee3-214195b1fd06 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': task-5116656, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1077.701615] env[65726]: DEBUG oslo_vmware.api [None req-91ff63ad-fe79-48d0-901e-fc0ed403574a tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Waiting for the task: (returnval){ [ 1077.701615] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]528cbf9f-b3ab-16c2-df86-06a46b68dfe3" [ 1077.701615] env[65726]: _type = "Task" [ 1077.701615] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1077.711423] env[65726]: DEBUG oslo_vmware.api [None req-91ff63ad-fe79-48d0-901e-fc0ed403574a tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]528cbf9f-b3ab-16c2-df86-06a46b68dfe3, 'name': SearchDatastore_Task, 'duration_secs': 0.010339} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1077.712634] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-75ce27a2-0624-4ffb-abbb-884a8ab946ea {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.722740] env[65726]: DEBUG oslo_vmware.api [None req-91ff63ad-fe79-48d0-901e-fc0ed403574a tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Waiting for the task: (returnval){ [ 1077.722740] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52392a34-7a3c-e29b-01bd-84fe395c9e9c" [ 1077.722740] env[65726]: _type = "Task" [ 1077.722740] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1077.732214] env[65726]: DEBUG oslo_vmware.api [None req-91ff63ad-fe79-48d0-901e-fc0ed403574a tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52392a34-7a3c-e29b-01bd-84fe395c9e9c, 'name': SearchDatastore_Task, 'duration_secs': 0.009549} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1077.732520] env[65726]: DEBUG oslo_concurrency.lockutils [None req-91ff63ad-fe79-48d0-901e-fc0ed403574a tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1077.732785] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-91ff63ad-fe79-48d0-901e-fc0ed403574a tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore2] 0d42fa75-20e9-4646-9b08-17015b7f068c/0d42fa75-20e9-4646-9b08-17015b7f068c.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1077.733085] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4b89bed0-f0e1-441b-bba2-0fde10f093fe {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.740316] env[65726]: DEBUG oslo_vmware.api [None req-91ff63ad-fe79-48d0-901e-fc0ed403574a tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Waiting for the task: (returnval){ [ 1077.740316] env[65726]: value = "task-5116657" [ 1077.740316] env[65726]: _type = "Task" [ 1077.740316] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1077.749112] env[65726]: DEBUG oslo_vmware.api [None req-91ff63ad-fe79-48d0-901e-fc0ed403574a tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': task-5116657, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1077.760180] env[65726]: DEBUG nova.network.neutron [req-9776b961-4d23-419a-aa3e-30ffbff82604 req-f8156270-90df-4e5b-8b3f-3e915dfebcfc service nova] [instance: 0d42fa75-20e9-4646-9b08-17015b7f068c] Updated VIF entry in instance network info cache for port 7a95cb45-ebfc-4ac0-ad58-4bb57170e47a. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 1077.760662] env[65726]: DEBUG nova.network.neutron [req-9776b961-4d23-419a-aa3e-30ffbff82604 req-f8156270-90df-4e5b-8b3f-3e915dfebcfc service nova] [instance: 0d42fa75-20e9-4646-9b08-17015b7f068c] Updating instance_info_cache with network_info: [{"id": "7a95cb45-ebfc-4ac0-ad58-4bb57170e47a", "address": "fa:16:3e:2b:ea:81", "network": {"id": "51a66448-0f72-4266-8868-856b48be839f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-155658961-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d46ccb00794f458b85da4a93879139ee", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b1f3e6c3-5584-4852-9017-476ab8ac4946", "external-id": "nsx-vlan-transportzone-304", "segmentation_id": 304, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7a95cb45-eb", "ovs_interfaceid": "7a95cb45-ebfc-4ac0-ad58-4bb57170e47a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1077.939263] env[65726]: WARNING openstack [None req-8c26e8da-5d87-42a2-a8c5-15e9c65a8b9f tempest-InstanceActionsV221TestJSON-647428531 tempest-InstanceActionsV221TestJSON-647428531-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1077.939701] env[65726]: WARNING openstack [None req-8c26e8da-5d87-42a2-a8c5-15e9c65a8b9f tempest-InstanceActionsV221TestJSON-647428531 tempest-InstanceActionsV221TestJSON-647428531-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1077.947052] env[65726]: DEBUG nova.network.neutron [None req-8c26e8da-5d87-42a2-a8c5-15e9c65a8b9f tempest-InstanceActionsV221TestJSON-647428531 tempest-InstanceActionsV221TestJSON-647428531-project-member] [instance: b0fec5ff-8c1e-41f9-b30a-6253a34baeec] Instance cache missing network info. {{(pid=65726) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 1077.980793] env[65726]: WARNING openstack [None req-a4ccf9b3-2389-45a5-bcf4-8de43acbac4c tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1077.981340] env[65726]: WARNING openstack [None req-a4ccf9b3-2389-45a5-bcf4-8de43acbac4c tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1078.044314] env[65726]: WARNING openstack [None req-8c26e8da-5d87-42a2-a8c5-15e9c65a8b9f tempest-InstanceActionsV221TestJSON-647428531 tempest-InstanceActionsV221TestJSON-647428531-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1078.044707] env[65726]: WARNING openstack [None req-8c26e8da-5d87-42a2-a8c5-15e9c65a8b9f tempest-InstanceActionsV221TestJSON-647428531 tempest-InstanceActionsV221TestJSON-647428531-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1078.089215] env[65726]: DEBUG oslo_vmware.api [None req-dc5e36ff-c9dd-4078-bec5-b01919099186 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5116655, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1078.124324] env[65726]: WARNING openstack [None req-a4ccf9b3-2389-45a5-bcf4-8de43acbac4c tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1078.124811] env[65726]: WARNING openstack [None req-a4ccf9b3-2389-45a5-bcf4-8de43acbac4c tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1078.137568] env[65726]: INFO nova.compute.manager [None req-f263eae6-96b2-45f2-b03a-d97fe5d0f1d8 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] [instance: 6fd70b4a-63e4-4258-9cff-f3b582500b9f] Shelve offloading [ 1078.151624] env[65726]: WARNING openstack [None req-8c26e8da-5d87-42a2-a8c5-15e9c65a8b9f tempest-InstanceActionsV221TestJSON-647428531 tempest-InstanceActionsV221TestJSON-647428531-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1078.152105] env[65726]: WARNING openstack [None req-8c26e8da-5d87-42a2-a8c5-15e9c65a8b9f tempest-InstanceActionsV221TestJSON-647428531 tempest-InstanceActionsV221TestJSON-647428531-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1078.198800] env[65726]: DEBUG oslo_vmware.api [None req-e7763bc8-6024-44cb-aee3-214195b1fd06 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': task-5116656, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1078.204465] env[65726]: WARNING openstack [None req-a4ccf9b3-2389-45a5-bcf4-8de43acbac4c tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1078.204928] env[65726]: WARNING openstack [None req-a4ccf9b3-2389-45a5-bcf4-8de43acbac4c tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1078.250188] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4234ca5d-f66f-4b96-bf77-4c313e80b044 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.257656] env[65726]: DEBUG oslo_vmware.api [None req-91ff63ad-fe79-48d0-901e-fc0ed403574a tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': task-5116657, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1078.263577] env[65726]: DEBUG oslo_concurrency.lockutils [req-9776b961-4d23-419a-aa3e-30ffbff82604 req-f8156270-90df-4e5b-8b3f-3e915dfebcfc service nova] Releasing lock "refresh_cache-0d42fa75-20e9-4646-9b08-17015b7f068c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1078.265212] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a56ec172-e46a-4087-b466-597d77f5218b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.308281] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3bb81eb8-0100-4fcb-978f-7551cad37763 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.319632] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb6e02d6-a1d5-400c-93dc-07bb6e65b7d4 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.342115] env[65726]: DEBUG nova.compute.provider_tree [None req-9ca95ca0-14c4-4870-aa74-cd2a795fbbc9 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1078.353535] env[65726]: DEBUG nova.network.neutron [None req-8c26e8da-5d87-42a2-a8c5-15e9c65a8b9f tempest-InstanceActionsV221TestJSON-647428531 tempest-InstanceActionsV221TestJSON-647428531-project-member] [instance: b0fec5ff-8c1e-41f9-b30a-6253a34baeec] Updating instance_info_cache with network_info: [{"id": "f11556ce-cfc0-4593-ac46-59266a693222", "address": "fa:16:3e:b0:ac:64", "network": {"id": "a1dbed2e-e67f-415c-807a-2296f7957462", "bridge": "br-int", "label": "tempest-InstanceActionsV221TestJSON-1168406842-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a6424f20b3b849e5ad82327218765b0c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5fb99c57-eaa0-447b-bb33-baced85d9c00", "external-id": "nsx-vlan-transportzone-253", "segmentation_id": 253, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf11556ce-cf", "ovs_interfaceid": "f11556ce-cfc0-4593-ac46-59266a693222", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1078.435658] env[65726]: DEBUG nova.network.neutron [None req-a4ccf9b3-2389-45a5-bcf4-8de43acbac4c tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] [instance: 3b4d9461-747a-461c-b231-a0de02f0cb83] Updating instance_info_cache with network_info: [{"id": "20c6b53e-2740-46bc-a94f-59a15f8b2c61", "address": "fa:16:3e:09:2a:fb", "network": {"id": "53facc8b-af7e-44de-8c6f-6bea798b3f0b", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-2084216438-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "eb811e38506b4297bf01bc5f2e098370", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "19671de9-8b5b-4710-adc3-7419f3c0f171", "external-id": "nsx-vlan-transportzone-421", "segmentation_id": 421, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap20c6b53e-27", "ovs_interfaceid": "20c6b53e-2740-46bc-a94f-59a15f8b2c61", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1078.590257] env[65726]: DEBUG oslo_vmware.api [None req-dc5e36ff-c9dd-4078-bec5-b01919099186 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5116655, 'name': PowerOnVM_Task, 'duration_secs': 0.805135} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1078.590669] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-dc5e36ff-c9dd-4078-bec5-b01919099186 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: a56be6fa-6027-46ee-9aa8-332fe4222d68] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1078.590938] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-dc5e36ff-c9dd-4078-bec5-b01919099186 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: a56be6fa-6027-46ee-9aa8-332fe4222d68] Updating instance 'a56be6fa-6027-46ee-9aa8-332fe4222d68' progress to 100 {{(pid=65726) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1078.641454] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-f263eae6-96b2-45f2-b03a-d97fe5d0f1d8 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] [instance: 6fd70b4a-63e4-4258-9cff-f3b582500b9f] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1078.641831] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a7a97a15-a07c-4497-ac3f-df0333075a57 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.647603] env[65726]: DEBUG nova.compute.manager [req-8f135682-c4a1-48a4-bfec-bf68e16816af req-a86d94c4-1b49-497f-9232-802976bb30d0 service nova] [instance: b0fec5ff-8c1e-41f9-b30a-6253a34baeec] Received event network-vif-plugged-f11556ce-cfc0-4593-ac46-59266a693222 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1078.647803] env[65726]: DEBUG oslo_concurrency.lockutils [req-8f135682-c4a1-48a4-bfec-bf68e16816af req-a86d94c4-1b49-497f-9232-802976bb30d0 service nova] Acquiring lock "b0fec5ff-8c1e-41f9-b30a-6253a34baeec-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1078.648021] env[65726]: DEBUG oslo_concurrency.lockutils [req-8f135682-c4a1-48a4-bfec-bf68e16816af req-a86d94c4-1b49-497f-9232-802976bb30d0 service nova] Lock "b0fec5ff-8c1e-41f9-b30a-6253a34baeec-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1078.648188] env[65726]: DEBUG oslo_concurrency.lockutils [req-8f135682-c4a1-48a4-bfec-bf68e16816af req-a86d94c4-1b49-497f-9232-802976bb30d0 service nova] Lock "b0fec5ff-8c1e-41f9-b30a-6253a34baeec-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1078.648355] env[65726]: DEBUG nova.compute.manager [req-8f135682-c4a1-48a4-bfec-bf68e16816af req-a86d94c4-1b49-497f-9232-802976bb30d0 service nova] [instance: b0fec5ff-8c1e-41f9-b30a-6253a34baeec] No waiting events found dispatching network-vif-plugged-f11556ce-cfc0-4593-ac46-59266a693222 {{(pid=65726) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1078.648514] env[65726]: WARNING nova.compute.manager [req-8f135682-c4a1-48a4-bfec-bf68e16816af req-a86d94c4-1b49-497f-9232-802976bb30d0 service nova] [instance: b0fec5ff-8c1e-41f9-b30a-6253a34baeec] Received unexpected event network-vif-plugged-f11556ce-cfc0-4593-ac46-59266a693222 for instance with vm_state building and task_state spawning. [ 1078.648668] env[65726]: DEBUG nova.compute.manager [req-8f135682-c4a1-48a4-bfec-bf68e16816af req-a86d94c4-1b49-497f-9232-802976bb30d0 service nova] [instance: b0fec5ff-8c1e-41f9-b30a-6253a34baeec] Received event network-changed-f11556ce-cfc0-4593-ac46-59266a693222 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1078.648815] env[65726]: DEBUG nova.compute.manager [req-8f135682-c4a1-48a4-bfec-bf68e16816af req-a86d94c4-1b49-497f-9232-802976bb30d0 service nova] [instance: b0fec5ff-8c1e-41f9-b30a-6253a34baeec] Refreshing instance network info cache due to event network-changed-f11556ce-cfc0-4593-ac46-59266a693222. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 1078.648976] env[65726]: DEBUG oslo_concurrency.lockutils [req-8f135682-c4a1-48a4-bfec-bf68e16816af req-a86d94c4-1b49-497f-9232-802976bb30d0 service nova] Acquiring lock "refresh_cache-b0fec5ff-8c1e-41f9-b30a-6253a34baeec" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1078.654076] env[65726]: DEBUG oslo_vmware.api [None req-f263eae6-96b2-45f2-b03a-d97fe5d0f1d8 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Waiting for the task: (returnval){ [ 1078.654076] env[65726]: value = "task-5116658" [ 1078.654076] env[65726]: _type = "Task" [ 1078.654076] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1078.663936] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-f263eae6-96b2-45f2-b03a-d97fe5d0f1d8 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] [instance: 6fd70b4a-63e4-4258-9cff-f3b582500b9f] VM already powered off {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1078.664199] env[65726]: DEBUG nova.compute.manager [None req-f263eae6-96b2-45f2-b03a-d97fe5d0f1d8 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] [instance: 6fd70b4a-63e4-4258-9cff-f3b582500b9f] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1078.664979] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb2dfbea-ed85-40e7-8314-5956ab35b710 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.672767] env[65726]: DEBUG oslo_concurrency.lockutils [None req-f263eae6-96b2-45f2-b03a-d97fe5d0f1d8 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Acquiring lock "refresh_cache-6fd70b4a-63e4-4258-9cff-f3b582500b9f" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1078.672934] env[65726]: DEBUG oslo_concurrency.lockutils [None req-f263eae6-96b2-45f2-b03a-d97fe5d0f1d8 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Acquired lock "refresh_cache-6fd70b4a-63e4-4258-9cff-f3b582500b9f" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1078.673114] env[65726]: DEBUG nova.network.neutron [None req-f263eae6-96b2-45f2-b03a-d97fe5d0f1d8 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] [instance: 6fd70b4a-63e4-4258-9cff-f3b582500b9f] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1078.696348] env[65726]: DEBUG oslo_vmware.api [None req-e7763bc8-6024-44cb-aee3-214195b1fd06 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': task-5116656, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.521897} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1078.696618] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-e7763bc8-6024-44cb-aee3-214195b1fd06 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore2] 5cdf1d61-b817-4986-a48c-804a2374497a/5cdf1d61-b817-4986-a48c-804a2374497a.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1078.696830] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-e7763bc8-6024-44cb-aee3-214195b1fd06 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 5cdf1d61-b817-4986-a48c-804a2374497a] Extending root virtual disk to 1048576 {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1078.697191] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-14843d47-3edf-4020-88bc-e7adb5f96347 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.704764] env[65726]: DEBUG oslo_vmware.api [None req-e7763bc8-6024-44cb-aee3-214195b1fd06 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Waiting for the task: (returnval){ [ 1078.704764] env[65726]: value = "task-5116659" [ 1078.704764] env[65726]: _type = "Task" [ 1078.704764] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1078.713271] env[65726]: DEBUG oslo_vmware.api [None req-e7763bc8-6024-44cb-aee3-214195b1fd06 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': task-5116659, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1078.753554] env[65726]: DEBUG oslo_vmware.api [None req-91ff63ad-fe79-48d0-901e-fc0ed403574a tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': task-5116657, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.81016} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1078.753822] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-91ff63ad-fe79-48d0-901e-fc0ed403574a tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore2] 0d42fa75-20e9-4646-9b08-17015b7f068c/0d42fa75-20e9-4646-9b08-17015b7f068c.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1078.754049] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-91ff63ad-fe79-48d0-901e-fc0ed403574a tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: 0d42fa75-20e9-4646-9b08-17015b7f068c] Extending root virtual disk to 1048576 {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1078.754316] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-dea6d1ac-1c53-4ae1-9805-a7f47b123710 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.761556] env[65726]: DEBUG oslo_vmware.api [None req-91ff63ad-fe79-48d0-901e-fc0ed403574a tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Waiting for the task: (returnval){ [ 1078.761556] env[65726]: value = "task-5116660" [ 1078.761556] env[65726]: _type = "Task" [ 1078.761556] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1078.772398] env[65726]: DEBUG oslo_vmware.api [None req-91ff63ad-fe79-48d0-901e-fc0ed403574a tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': task-5116660, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1078.847083] env[65726]: DEBUG nova.scheduler.client.report [None req-9ca95ca0-14c4-4870-aa74-cd2a795fbbc9 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1078.856603] env[65726]: DEBUG oslo_concurrency.lockutils [None req-8c26e8da-5d87-42a2-a8c5-15e9c65a8b9f tempest-InstanceActionsV221TestJSON-647428531 tempest-InstanceActionsV221TestJSON-647428531-project-member] Releasing lock "refresh_cache-b0fec5ff-8c1e-41f9-b30a-6253a34baeec" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1078.856603] env[65726]: DEBUG nova.compute.manager [None req-8c26e8da-5d87-42a2-a8c5-15e9c65a8b9f tempest-InstanceActionsV221TestJSON-647428531 tempest-InstanceActionsV221TestJSON-647428531-project-member] [instance: b0fec5ff-8c1e-41f9-b30a-6253a34baeec] Instance network_info: |[{"id": "f11556ce-cfc0-4593-ac46-59266a693222", "address": "fa:16:3e:b0:ac:64", "network": {"id": "a1dbed2e-e67f-415c-807a-2296f7957462", "bridge": "br-int", "label": "tempest-InstanceActionsV221TestJSON-1168406842-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a6424f20b3b849e5ad82327218765b0c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5fb99c57-eaa0-447b-bb33-baced85d9c00", "external-id": "nsx-vlan-transportzone-253", "segmentation_id": 253, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf11556ce-cf", "ovs_interfaceid": "f11556ce-cfc0-4593-ac46-59266a693222", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1078.857098] env[65726]: DEBUG oslo_concurrency.lockutils [req-8f135682-c4a1-48a4-bfec-bf68e16816af req-a86d94c4-1b49-497f-9232-802976bb30d0 service nova] Acquired lock "refresh_cache-b0fec5ff-8c1e-41f9-b30a-6253a34baeec" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1078.857826] env[65726]: DEBUG nova.network.neutron [req-8f135682-c4a1-48a4-bfec-bf68e16816af req-a86d94c4-1b49-497f-9232-802976bb30d0 service nova] [instance: b0fec5ff-8c1e-41f9-b30a-6253a34baeec] Refreshing network info cache for port f11556ce-cfc0-4593-ac46-59266a693222 {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 1078.859016] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-8c26e8da-5d87-42a2-a8c5-15e9c65a8b9f tempest-InstanceActionsV221TestJSON-647428531 tempest-InstanceActionsV221TestJSON-647428531-project-member] [instance: b0fec5ff-8c1e-41f9-b30a-6253a34baeec] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b0:ac:64', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '5fb99c57-eaa0-447b-bb33-baced85d9c00', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f11556ce-cfc0-4593-ac46-59266a693222', 'vif_model': 'vmxnet3'}] {{(pid=65726) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1078.867018] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-8c26e8da-5d87-42a2-a8c5-15e9c65a8b9f tempest-InstanceActionsV221TestJSON-647428531 tempest-InstanceActionsV221TestJSON-647428531-project-member] Creating folder: Project (a6424f20b3b849e5ad82327218765b0c). Parent ref: group-v995008. {{(pid=65726) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1078.868163] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-30d7ab8a-ce10-4b06-9137-7440790eff18 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.881140] env[65726]: INFO nova.virt.vmwareapi.vm_util [None req-8c26e8da-5d87-42a2-a8c5-15e9c65a8b9f tempest-InstanceActionsV221TestJSON-647428531 tempest-InstanceActionsV221TestJSON-647428531-project-member] Created folder: Project (a6424f20b3b849e5ad82327218765b0c) in parent group-v995008. [ 1078.881359] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-8c26e8da-5d87-42a2-a8c5-15e9c65a8b9f tempest-InstanceActionsV221TestJSON-647428531 tempest-InstanceActionsV221TestJSON-647428531-project-member] Creating folder: Instances. Parent ref: group-v995291. {{(pid=65726) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1078.881610] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-3cb8bc4a-5c72-46cc-b0f9-296fe2621ca9 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.892656] env[65726]: INFO nova.virt.vmwareapi.vm_util [None req-8c26e8da-5d87-42a2-a8c5-15e9c65a8b9f tempest-InstanceActionsV221TestJSON-647428531 tempest-InstanceActionsV221TestJSON-647428531-project-member] Created folder: Instances in parent group-v995291. [ 1078.893062] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-8c26e8da-5d87-42a2-a8c5-15e9c65a8b9f tempest-InstanceActionsV221TestJSON-647428531 tempest-InstanceActionsV221TestJSON-647428531-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1078.893302] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b0fec5ff-8c1e-41f9-b30a-6253a34baeec] Creating VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1078.893653] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-72eee5ea-f4a7-449b-8500-e43173ff4799 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.915466] env[65726]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1078.915466] env[65726]: value = "task-5116663" [ 1078.915466] env[65726]: _type = "Task" [ 1078.915466] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1078.928368] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116663, 'name': CreateVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1078.939458] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a4ccf9b3-2389-45a5-bcf4-8de43acbac4c tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Releasing lock "refresh_cache-3b4d9461-747a-461c-b231-a0de02f0cb83" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1079.177624] env[65726]: WARNING openstack [None req-f263eae6-96b2-45f2-b03a-d97fe5d0f1d8 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1079.177624] env[65726]: WARNING openstack [None req-f263eae6-96b2-45f2-b03a-d97fe5d0f1d8 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1079.215618] env[65726]: DEBUG oslo_vmware.api [None req-e7763bc8-6024-44cb-aee3-214195b1fd06 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': task-5116659, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.260271} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1079.215980] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-e7763bc8-6024-44cb-aee3-214195b1fd06 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 5cdf1d61-b817-4986-a48c-804a2374497a] Extended root virtual disk {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1079.216964] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34622688-f546-4a57-bea0-0444f5ca3e67 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.244503] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-e7763bc8-6024-44cb-aee3-214195b1fd06 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 5cdf1d61-b817-4986-a48c-804a2374497a] Reconfiguring VM instance instance-0000005f to attach disk [datastore2] 5cdf1d61-b817-4986-a48c-804a2374497a/5cdf1d61-b817-4986-a48c-804a2374497a.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1079.249465] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1188ffba-96ad-4b67-b40c-558f42c952f5 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.274307] env[65726]: DEBUG oslo_vmware.api [None req-91ff63ad-fe79-48d0-901e-fc0ed403574a tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': task-5116660, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.095591} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1079.275807] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-91ff63ad-fe79-48d0-901e-fc0ed403574a tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: 0d42fa75-20e9-4646-9b08-17015b7f068c] Extended root virtual disk {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1079.276201] env[65726]: DEBUG oslo_vmware.api [None req-e7763bc8-6024-44cb-aee3-214195b1fd06 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Waiting for the task: (returnval){ [ 1079.276201] env[65726]: value = "task-5116664" [ 1079.276201] env[65726]: _type = "Task" [ 1079.276201] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1079.276901] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eaab4352-5439-45a0-ab1d-d64b993754ce {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.304125] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-91ff63ad-fe79-48d0-901e-fc0ed403574a tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: 0d42fa75-20e9-4646-9b08-17015b7f068c] Reconfiguring VM instance instance-00000061 to attach disk [datastore2] 0d42fa75-20e9-4646-9b08-17015b7f068c/0d42fa75-20e9-4646-9b08-17015b7f068c.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1079.312496] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1a5f607f-c85f-4f97-938d-2c9a6c12709d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.328587] env[65726]: DEBUG oslo_vmware.api [None req-e7763bc8-6024-44cb-aee3-214195b1fd06 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': task-5116664, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1079.335979] env[65726]: DEBUG oslo_vmware.api [None req-91ff63ad-fe79-48d0-901e-fc0ed403574a tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Waiting for the task: (returnval){ [ 1079.335979] env[65726]: value = "task-5116665" [ 1079.335979] env[65726]: _type = "Task" [ 1079.335979] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1079.348466] env[65726]: DEBUG oslo_vmware.api [None req-91ff63ad-fe79-48d0-901e-fc0ed403574a tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': task-5116665, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1079.350230] env[65726]: WARNING openstack [None req-f263eae6-96b2-45f2-b03a-d97fe5d0f1d8 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1079.350629] env[65726]: WARNING openstack [None req-f263eae6-96b2-45f2-b03a-d97fe5d0f1d8 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1079.358871] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9ca95ca0-14c4-4870-aa74-cd2a795fbbc9 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.423s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1079.359478] env[65726]: DEBUG nova.compute.manager [None req-9ca95ca0-14c4-4870-aa74-cd2a795fbbc9 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: d8801049-fc34-4b2f-8d5e-b9459c3e03a6] Start building networks asynchronously for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1079.363336] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 2.218s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1079.370731] env[65726]: WARNING openstack [req-8f135682-c4a1-48a4-bfec-bf68e16816af req-a86d94c4-1b49-497f-9232-802976bb30d0 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1079.371586] env[65726]: WARNING openstack [req-8f135682-c4a1-48a4-bfec-bf68e16816af req-a86d94c4-1b49-497f-9232-802976bb30d0 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1079.439097] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116663, 'name': CreateVM_Task, 'duration_secs': 0.373977} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1079.439097] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b0fec5ff-8c1e-41f9-b30a-6253a34baeec] Created VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1079.439097] env[65726]: WARNING openstack [None req-8c26e8da-5d87-42a2-a8c5-15e9c65a8b9f tempest-InstanceActionsV221TestJSON-647428531 tempest-InstanceActionsV221TestJSON-647428531-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1079.439097] env[65726]: WARNING openstack [None req-8c26e8da-5d87-42a2-a8c5-15e9c65a8b9f tempest-InstanceActionsV221TestJSON-647428531 tempest-InstanceActionsV221TestJSON-647428531-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1079.444960] env[65726]: DEBUG oslo_concurrency.lockutils [None req-8c26e8da-5d87-42a2-a8c5-15e9c65a8b9f tempest-InstanceActionsV221TestJSON-647428531 tempest-InstanceActionsV221TestJSON-647428531-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1079.444960] env[65726]: DEBUG oslo_concurrency.lockutils [None req-8c26e8da-5d87-42a2-a8c5-15e9c65a8b9f tempest-InstanceActionsV221TestJSON-647428531 tempest-InstanceActionsV221TestJSON-647428531-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1079.444960] env[65726]: DEBUG oslo_concurrency.lockutils [None req-8c26e8da-5d87-42a2-a8c5-15e9c65a8b9f tempest-InstanceActionsV221TestJSON-647428531 tempest-InstanceActionsV221TestJSON-647428531-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1079.446505] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-93f284cf-21e0-4f6f-9490-a261a135f9ea {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.455179] env[65726]: DEBUG oslo_vmware.api [None req-8c26e8da-5d87-42a2-a8c5-15e9c65a8b9f tempest-InstanceActionsV221TestJSON-647428531 tempest-InstanceActionsV221TestJSON-647428531-project-member] Waiting for the task: (returnval){ [ 1079.455179] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52bbbd7d-092f-e649-309a-14505804d7b5" [ 1079.455179] env[65726]: _type = "Task" [ 1079.455179] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1079.464882] env[65726]: DEBUG oslo_vmware.api [None req-8c26e8da-5d87-42a2-a8c5-15e9c65a8b9f tempest-InstanceActionsV221TestJSON-647428531 tempest-InstanceActionsV221TestJSON-647428531-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52bbbd7d-092f-e649-309a-14505804d7b5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1079.479790] env[65726]: WARNING openstack [None req-f263eae6-96b2-45f2-b03a-d97fe5d0f1d8 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1079.480347] env[65726]: WARNING openstack [None req-f263eae6-96b2-45f2-b03a-d97fe5d0f1d8 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1079.654862] env[65726]: WARNING openstack [req-8f135682-c4a1-48a4-bfec-bf68e16816af req-a86d94c4-1b49-497f-9232-802976bb30d0 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1079.654862] env[65726]: WARNING openstack [req-8f135682-c4a1-48a4-bfec-bf68e16816af req-a86d94c4-1b49-497f-9232-802976bb30d0 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1079.692137] env[65726]: DEBUG nova.network.neutron [None req-f263eae6-96b2-45f2-b03a-d97fe5d0f1d8 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] [instance: 6fd70b4a-63e4-4258-9cff-f3b582500b9f] Updating instance_info_cache with network_info: [{"id": "153b4c9d-d01d-4254-9aa6-040705be347a", "address": "fa:16:3e:17:18:d9", "network": {"id": "93e38d7e-b4d7-4d4a-8393-64ab2f6769de", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1445539922-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "06e221dc693640929cac7bc5af45948f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "399f3826-705c-45f7-9fe0-3a08a945151a", "external-id": "nsx-vlan-transportzone-936", "segmentation_id": 936, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap153b4c9d-d0", "ovs_interfaceid": "153b4c9d-d01d-4254-9aa6-040705be347a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1079.790918] env[65726]: DEBUG oslo_vmware.api [None req-e7763bc8-6024-44cb-aee3-214195b1fd06 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': task-5116664, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1079.801656] env[65726]: WARNING openstack [req-8f135682-c4a1-48a4-bfec-bf68e16816af req-a86d94c4-1b49-497f-9232-802976bb30d0 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1079.803255] env[65726]: WARNING openstack [req-8f135682-c4a1-48a4-bfec-bf68e16816af req-a86d94c4-1b49-497f-9232-802976bb30d0 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1079.846089] env[65726]: DEBUG oslo_vmware.api [None req-91ff63ad-fe79-48d0-901e-fc0ed403574a tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': task-5116665, 'name': ReconfigVM_Task, 'duration_secs': 0.299895} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1079.846424] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-91ff63ad-fe79-48d0-901e-fc0ed403574a tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: 0d42fa75-20e9-4646-9b08-17015b7f068c] Reconfigured VM instance instance-00000061 to attach disk [datastore2] 0d42fa75-20e9-4646-9b08-17015b7f068c/0d42fa75-20e9-4646-9b08-17015b7f068c.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1079.847118] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-81402533-04ef-46e3-810f-1891064489b4 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.855266] env[65726]: DEBUG oslo_vmware.api [None req-91ff63ad-fe79-48d0-901e-fc0ed403574a tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Waiting for the task: (returnval){ [ 1079.855266] env[65726]: value = "task-5116666" [ 1079.855266] env[65726]: _type = "Task" [ 1079.855266] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1079.864787] env[65726]: DEBUG oslo_vmware.api [None req-91ff63ad-fe79-48d0-901e-fc0ed403574a tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': task-5116666, 'name': Rename_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1079.867493] env[65726]: DEBUG nova.compute.utils [None req-9ca95ca0-14c4-4870-aa74-cd2a795fbbc9 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Using /dev/sd instead of None {{(pid=65726) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1079.880020] env[65726]: DEBUG nova.compute.manager [None req-9ca95ca0-14c4-4870-aa74-cd2a795fbbc9 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: d8801049-fc34-4b2f-8d5e-b9459c3e03a6] Allocating IP information in the background. {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1079.880020] env[65726]: DEBUG nova.network.neutron [None req-9ca95ca0-14c4-4870-aa74-cd2a795fbbc9 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: d8801049-fc34-4b2f-8d5e-b9459c3e03a6] allocate_for_instance() {{(pid=65726) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 1079.880020] env[65726]: WARNING neutronclient.v2_0.client [None req-9ca95ca0-14c4-4870-aa74-cd2a795fbbc9 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1079.880020] env[65726]: WARNING neutronclient.v2_0.client [None req-9ca95ca0-14c4-4870-aa74-cd2a795fbbc9 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1079.880020] env[65726]: WARNING openstack [None req-9ca95ca0-14c4-4870-aa74-cd2a795fbbc9 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1079.880287] env[65726]: WARNING openstack [None req-9ca95ca0-14c4-4870-aa74-cd2a795fbbc9 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1079.947690] env[65726]: DEBUG nova.policy [None req-9ca95ca0-14c4-4870-aa74-cd2a795fbbc9 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4cd1ed63db174f21be78f86554fada0a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'cbb17a09b35c4c22ade5c4082b10bda9', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65726) authorize /opt/stack/nova/nova/policy.py:192}} [ 1079.965611] env[65726]: DEBUG oslo_vmware.api [None req-8c26e8da-5d87-42a2-a8c5-15e9c65a8b9f tempest-InstanceActionsV221TestJSON-647428531 tempest-InstanceActionsV221TestJSON-647428531-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52bbbd7d-092f-e649-309a-14505804d7b5, 'name': SearchDatastore_Task, 'duration_secs': 0.011356} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1079.966041] env[65726]: DEBUG oslo_concurrency.lockutils [None req-8c26e8da-5d87-42a2-a8c5-15e9c65a8b9f tempest-InstanceActionsV221TestJSON-647428531 tempest-InstanceActionsV221TestJSON-647428531-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1079.966312] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-8c26e8da-5d87-42a2-a8c5-15e9c65a8b9f tempest-InstanceActionsV221TestJSON-647428531 tempest-InstanceActionsV221TestJSON-647428531-project-member] [instance: b0fec5ff-8c1e-41f9-b30a-6253a34baeec] Processing image b52362a3-ee8a-4cbf-b06f-513b0cc8f95c {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1079.966798] env[65726]: DEBUG oslo_concurrency.lockutils [None req-8c26e8da-5d87-42a2-a8c5-15e9c65a8b9f tempest-InstanceActionsV221TestJSON-647428531 tempest-InstanceActionsV221TestJSON-647428531-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1079.966999] env[65726]: DEBUG oslo_concurrency.lockutils [None req-8c26e8da-5d87-42a2-a8c5-15e9c65a8b9f tempest-InstanceActionsV221TestJSON-647428531 tempest-InstanceActionsV221TestJSON-647428531-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1079.967785] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-8c26e8da-5d87-42a2-a8c5-15e9c65a8b9f tempest-InstanceActionsV221TestJSON-647428531 tempest-InstanceActionsV221TestJSON-647428531-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1079.968148] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b9bdd080-b0ef-4d85-8bd4-bac9472df7eb {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.980930] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-8c26e8da-5d87-42a2-a8c5-15e9c65a8b9f tempest-InstanceActionsV221TestJSON-647428531 tempest-InstanceActionsV221TestJSON-647428531-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1079.981150] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-8c26e8da-5d87-42a2-a8c5-15e9c65a8b9f tempest-InstanceActionsV221TestJSON-647428531 tempest-InstanceActionsV221TestJSON-647428531-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65726) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1079.982124] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ad92758e-c788-4c67-83f3-0afa7bb3e0a3 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.987335] env[65726]: DEBUG nova.network.neutron [req-8f135682-c4a1-48a4-bfec-bf68e16816af req-a86d94c4-1b49-497f-9232-802976bb30d0 service nova] [instance: b0fec5ff-8c1e-41f9-b30a-6253a34baeec] Updated VIF entry in instance network info cache for port f11556ce-cfc0-4593-ac46-59266a693222. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 1079.987920] env[65726]: DEBUG nova.network.neutron [req-8f135682-c4a1-48a4-bfec-bf68e16816af req-a86d94c4-1b49-497f-9232-802976bb30d0 service nova] [instance: b0fec5ff-8c1e-41f9-b30a-6253a34baeec] Updating instance_info_cache with network_info: [{"id": "f11556ce-cfc0-4593-ac46-59266a693222", "address": "fa:16:3e:b0:ac:64", "network": {"id": "a1dbed2e-e67f-415c-807a-2296f7957462", "bridge": "br-int", "label": "tempest-InstanceActionsV221TestJSON-1168406842-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a6424f20b3b849e5ad82327218765b0c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5fb99c57-eaa0-447b-bb33-baced85d9c00", "external-id": "nsx-vlan-transportzone-253", "segmentation_id": 253, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf11556ce-cf", "ovs_interfaceid": "f11556ce-cfc0-4593-ac46-59266a693222", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1079.998174] env[65726]: DEBUG oslo_vmware.api [None req-8c26e8da-5d87-42a2-a8c5-15e9c65a8b9f tempest-InstanceActionsV221TestJSON-647428531 tempest-InstanceActionsV221TestJSON-647428531-project-member] Waiting for the task: (returnval){ [ 1079.998174] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5203d0b6-39b5-c8a8-9e20-c7eabf925e30" [ 1079.998174] env[65726]: _type = "Task" [ 1079.998174] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1080.010283] env[65726]: DEBUG oslo_vmware.api [None req-8c26e8da-5d87-42a2-a8c5-15e9c65a8b9f tempest-InstanceActionsV221TestJSON-647428531 tempest-InstanceActionsV221TestJSON-647428531-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5203d0b6-39b5-c8a8-9e20-c7eabf925e30, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1080.194519] env[65726]: DEBUG oslo_concurrency.lockutils [None req-f263eae6-96b2-45f2-b03a-d97fe5d0f1d8 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Releasing lock "refresh_cache-6fd70b4a-63e4-4258-9cff-f3b582500b9f" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1080.195256] env[65726]: WARNING openstack [None req-f263eae6-96b2-45f2-b03a-d97fe5d0f1d8 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1080.195594] env[65726]: WARNING openstack [None req-f263eae6-96b2-45f2-b03a-d97fe5d0f1d8 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1080.201737] env[65726]: WARNING openstack [None req-f263eae6-96b2-45f2-b03a-d97fe5d0f1d8 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1080.202064] env[65726]: WARNING openstack [None req-f263eae6-96b2-45f2-b03a-d97fe5d0f1d8 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1080.242112] env[65726]: DEBUG nova.network.neutron [None req-9ca95ca0-14c4-4870-aa74-cd2a795fbbc9 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: d8801049-fc34-4b2f-8d5e-b9459c3e03a6] Successfully created port: b333a678-e6d6-494e-bc4d-7e498e370067 {{(pid=65726) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1080.291294] env[65726]: DEBUG oslo_vmware.api [None req-e7763bc8-6024-44cb-aee3-214195b1fd06 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': task-5116664, 'name': ReconfigVM_Task, 'duration_secs': 0.999662} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1080.291496] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-e7763bc8-6024-44cb-aee3-214195b1fd06 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 5cdf1d61-b817-4986-a48c-804a2374497a] Reconfigured VM instance instance-0000005f to attach disk [datastore2] 5cdf1d61-b817-4986-a48c-804a2374497a/5cdf1d61-b817-4986-a48c-804a2374497a.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1080.292742] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e89bf1c4-f817-4b3c-92b1-2087389760ec {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.308543] env[65726]: DEBUG oslo_vmware.api [None req-e7763bc8-6024-44cb-aee3-214195b1fd06 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Waiting for the task: (returnval){ [ 1080.308543] env[65726]: value = "task-5116667" [ 1080.308543] env[65726]: _type = "Task" [ 1080.308543] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1080.318852] env[65726]: DEBUG oslo_vmware.api [None req-e7763bc8-6024-44cb-aee3-214195b1fd06 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': task-5116667, 'name': Rename_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1080.367158] env[65726]: DEBUG oslo_vmware.api [None req-91ff63ad-fe79-48d0-901e-fc0ed403574a tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': task-5116666, 'name': Rename_Task, 'duration_secs': 0.161897} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1080.367492] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-91ff63ad-fe79-48d0-901e-fc0ed403574a tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: 0d42fa75-20e9-4646-9b08-17015b7f068c] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1080.368010] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7e0b1c06-5cc4-45a1-996e-e0ca49c68164 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.376068] env[65726]: DEBUG oslo_vmware.api [None req-91ff63ad-fe79-48d0-901e-fc0ed403574a tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Waiting for the task: (returnval){ [ 1080.376068] env[65726]: value = "task-5116668" [ 1080.376068] env[65726]: _type = "Task" [ 1080.376068] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1080.381330] env[65726]: DEBUG nova.compute.manager [None req-9ca95ca0-14c4-4870-aa74-cd2a795fbbc9 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: d8801049-fc34-4b2f-8d5e-b9459c3e03a6] Start building block device mappings for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1080.390112] env[65726]: DEBUG oslo_vmware.api [None req-91ff63ad-fe79-48d0-901e-fc0ed403574a tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': task-5116668, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1080.395982] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Applying migration context for instance a56be6fa-6027-46ee-9aa8-332fe4222d68 as it has an incoming, in-progress migration 0ae40754-d17a-4cf6-a2d0-e1747b172ad4. Migration status is finished {{(pid=65726) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1046}} [ 1080.397715] env[65726]: INFO nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] [instance: a56be6fa-6027-46ee-9aa8-332fe4222d68] Updating resource usage from migration 0ae40754-d17a-4cf6-a2d0-e1747b172ad4 [ 1080.444058] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Instance 6fd70b4a-63e4-4258-9cff-f3b582500b9f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65726) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1080.444307] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Instance aa6ce489-c62f-4481-87b7-e74242aeb8ca actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65726) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1080.444437] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Instance 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65726) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1080.444557] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Instance df08d166-7811-49b5-9c16-a3434229d623 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65726) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1080.444669] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Instance ad4f00b5-3765-494f-8132-8b17b29d219a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65726) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1080.444788] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Instance 51f5337d-2e23-4b7c-b590-76be6bd82411 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65726) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1080.444913] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Instance 5cdf1d61-b817-4986-a48c-804a2374497a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65726) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1080.445303] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Migration 0ae40754-d17a-4cf6-a2d0-e1747b172ad4 is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65726) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1745}} [ 1080.445605] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Instance a56be6fa-6027-46ee-9aa8-332fe4222d68 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=65726) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1080.445810] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Instance 3b4d9461-747a-461c-b231-a0de02f0cb83 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65726) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1080.445974] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Instance 0d42fa75-20e9-4646-9b08-17015b7f068c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65726) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1080.446157] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Instance b0fec5ff-8c1e-41f9-b30a-6253a34baeec actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65726) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1080.446706] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Instance d8801049-fc34-4b2f-8d5e-b9459c3e03a6 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65726) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1080.446841] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Total usable vcpus: 48, total allocated vcpus: 13 {{(pid=65726) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1159}} [ 1080.447113] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=3072MB phys_disk=100GB used_disk=13GB total_vcpus=48 used_vcpus=13 pci_stats=[] stats={'failed_builds': '0', 'num_instances': '12', 'num_vm_active': '8', 'num_task_None': '6', 'num_os_type_None': '12', 'num_proj_96149159e18e44f9bf3453e67681f224': '1', 'io_workload': '4', 'num_vm_shelved': '1', 'num_task_shelving_offloading': '1', 'num_proj_06e221dc693640929cac7bc5af45948f': '1', 'num_proj_4ffd45f4a7a041199a4fc7f69f5e7e9b': '1', 'num_task_resize_finish': '1', 'num_proj_4eece77569624f90bf64e5c51974173f': '1', 'num_proj_305ecace479440ad9a616e546ccd38ce': '1', 'num_proj_a090d553766847e58e3231b966c92565': '1', 'num_proj_d46ccb00794f458b85da4a93879139ee': '2', 'num_task_rebuild_spawning': '1', 'num_proj_f93d852e2c904f42981cbfff882558e9': '1', 'num_task_rescuing': '1', 'num_proj_eb811e38506b4297bf01bc5f2e098370': '1', 'num_vm_building': '3', 'num_task_spawning': '2', 'num_proj_a6424f20b3b849e5ad82327218765b0c': '1', 'num_proj_cbb17a09b35c4c22ade5c4082b10bda9': '1'} {{(pid=65726) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1168}} [ 1080.494195] env[65726]: DEBUG oslo_concurrency.lockutils [req-8f135682-c4a1-48a4-bfec-bf68e16816af req-a86d94c4-1b49-497f-9232-802976bb30d0 service nova] Releasing lock "refresh_cache-b0fec5ff-8c1e-41f9-b30a-6253a34baeec" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1080.495016] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-a4ccf9b3-2389-45a5-bcf4-8de43acbac4c tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] [instance: 3b4d9461-747a-461c-b231-a0de02f0cb83] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1080.495380] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8980cfc9-6878-4927-a6df-f353007be8ff {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.511951] env[65726]: DEBUG oslo_vmware.api [None req-8c26e8da-5d87-42a2-a8c5-15e9c65a8b9f tempest-InstanceActionsV221TestJSON-647428531 tempest-InstanceActionsV221TestJSON-647428531-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5203d0b6-39b5-c8a8-9e20-c7eabf925e30, 'name': SearchDatastore_Task, 'duration_secs': 0.013589} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1080.514814] env[65726]: DEBUG oslo_vmware.api [None req-a4ccf9b3-2389-45a5-bcf4-8de43acbac4c tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Waiting for the task: (returnval){ [ 1080.514814] env[65726]: value = "task-5116669" [ 1080.514814] env[65726]: _type = "Task" [ 1080.514814] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1080.519801] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c258381e-0864-4bae-bc02-da44fbde73a2 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.538494] env[65726]: DEBUG oslo_vmware.api [None req-8c26e8da-5d87-42a2-a8c5-15e9c65a8b9f tempest-InstanceActionsV221TestJSON-647428531 tempest-InstanceActionsV221TestJSON-647428531-project-member] Waiting for the task: (returnval){ [ 1080.538494] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52fa864d-4b05-5389-9be3-a7a1a026feda" [ 1080.538494] env[65726]: _type = "Task" [ 1080.538494] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1080.542993] env[65726]: DEBUG oslo_vmware.api [None req-a4ccf9b3-2389-45a5-bcf4-8de43acbac4c tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Task: {'id': task-5116669, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1080.556570] env[65726]: DEBUG oslo_vmware.api [None req-8c26e8da-5d87-42a2-a8c5-15e9c65a8b9f tempest-InstanceActionsV221TestJSON-647428531 tempest-InstanceActionsV221TestJSON-647428531-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52fa864d-4b05-5389-9be3-a7a1a026feda, 'name': SearchDatastore_Task, 'duration_secs': 0.013226} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1080.556932] env[65726]: DEBUG oslo_concurrency.lockutils [None req-8c26e8da-5d87-42a2-a8c5-15e9c65a8b9f tempest-InstanceActionsV221TestJSON-647428531 tempest-InstanceActionsV221TestJSON-647428531-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1080.557251] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-8c26e8da-5d87-42a2-a8c5-15e9c65a8b9f tempest-InstanceActionsV221TestJSON-647428531 tempest-InstanceActionsV221TestJSON-647428531-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore2] b0fec5ff-8c1e-41f9-b30a-6253a34baeec/b0fec5ff-8c1e-41f9-b30a-6253a34baeec.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1080.557551] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-64c4f05d-c531-497a-8af5-65e55ddb2863 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.567980] env[65726]: DEBUG oslo_vmware.api [None req-8c26e8da-5d87-42a2-a8c5-15e9c65a8b9f tempest-InstanceActionsV221TestJSON-647428531 tempest-InstanceActionsV221TestJSON-647428531-project-member] Waiting for the task: (returnval){ [ 1080.567980] env[65726]: value = "task-5116670" [ 1080.567980] env[65726]: _type = "Task" [ 1080.567980] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1080.580906] env[65726]: DEBUG oslo_vmware.api [None req-8c26e8da-5d87-42a2-a8c5-15e9c65a8b9f tempest-InstanceActionsV221TestJSON-647428531 tempest-InstanceActionsV221TestJSON-647428531-project-member] Task: {'id': task-5116670, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1080.696108] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-457a2ff8-382f-4a1a-b06e-78893ab98f61 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.705885] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e42125a-c4b3-48a5-b0fb-472ad45c1f4f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.741745] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ec71a59-24de-4dda-a602-46bf7267b6ed {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.752433] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9be97b2-9c8c-48c8-b392-d4c3222bd4b2 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.772615] env[65726]: DEBUG nova.compute.provider_tree [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1080.820946] env[65726]: DEBUG oslo_vmware.api [None req-e7763bc8-6024-44cb-aee3-214195b1fd06 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': task-5116667, 'name': Rename_Task, 'duration_secs': 0.161756} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1080.821247] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-e7763bc8-6024-44cb-aee3-214195b1fd06 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 5cdf1d61-b817-4986-a48c-804a2374497a] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1080.821582] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-06a27369-80c6-44ae-89d3-ddb1da75d635 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.829445] env[65726]: DEBUG oslo_vmware.api [None req-e7763bc8-6024-44cb-aee3-214195b1fd06 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Waiting for the task: (returnval){ [ 1080.829445] env[65726]: value = "task-5116671" [ 1080.829445] env[65726]: _type = "Task" [ 1080.829445] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1080.840708] env[65726]: DEBUG oslo_vmware.api [None req-e7763bc8-6024-44cb-aee3-214195b1fd06 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': task-5116671, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1080.891813] env[65726]: DEBUG oslo_vmware.api [None req-91ff63ad-fe79-48d0-901e-fc0ed403574a tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': task-5116668, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1081.040394] env[65726]: DEBUG oslo_vmware.api [None req-a4ccf9b3-2389-45a5-bcf4-8de43acbac4c tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Task: {'id': task-5116669, 'name': PowerOffVM_Task, 'duration_secs': 0.221928} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1081.042177] env[65726]: WARNING openstack [None req-14a9b6f0-9c85-4271-a1df-1d81c0fc4a98 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1081.042666] env[65726]: WARNING openstack [None req-14a9b6f0-9c85-4271-a1df-1d81c0fc4a98 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1081.051214] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-a4ccf9b3-2389-45a5-bcf4-8de43acbac4c tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] [instance: 3b4d9461-747a-461c-b231-a0de02f0cb83] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1081.051374] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34209494-ffdb-4a36-92dd-35b9296615b4 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.078260] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4309025-b70f-4f13-9841-7eaf042f515e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.087726] env[65726]: DEBUG oslo_vmware.api [None req-8c26e8da-5d87-42a2-a8c5-15e9c65a8b9f tempest-InstanceActionsV221TestJSON-647428531 tempest-InstanceActionsV221TestJSON-647428531-project-member] Task: {'id': task-5116670, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1081.119915] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-a4ccf9b3-2389-45a5-bcf4-8de43acbac4c tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] [instance: 3b4d9461-747a-461c-b231-a0de02f0cb83] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1081.120445] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9c35e545-4c37-4e61-a626-d374d3be7e56 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.124630] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-f263eae6-96b2-45f2-b03a-d97fe5d0f1d8 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] [instance: 6fd70b4a-63e4-4258-9cff-f3b582500b9f] Destroying instance {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1081.125616] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da6cefa7-9052-4922-97b8-027949f100a6 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.134561] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-f263eae6-96b2-45f2-b03a-d97fe5d0f1d8 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] [instance: 6fd70b4a-63e4-4258-9cff-f3b582500b9f] Unregistering the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1081.136105] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-bf9283bb-b883-4a34-8854-12d1a58ac91e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.138498] env[65726]: DEBUG oslo_vmware.api [None req-a4ccf9b3-2389-45a5-bcf4-8de43acbac4c tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Waiting for the task: (returnval){ [ 1081.138498] env[65726]: value = "task-5116672" [ 1081.138498] env[65726]: _type = "Task" [ 1081.138498] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1081.244033] env[65726]: WARNING openstack [None req-14a9b6f0-9c85-4271-a1df-1d81c0fc4a98 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1081.245344] env[65726]: WARNING openstack [None req-14a9b6f0-9c85-4271-a1df-1d81c0fc4a98 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1081.251484] env[65726]: WARNING neutronclient.v2_0.client [None req-14a9b6f0-9c85-4271-a1df-1d81c0fc4a98 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1081.255335] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-f263eae6-96b2-45f2-b03a-d97fe5d0f1d8 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] [instance: 6fd70b4a-63e4-4258-9cff-f3b582500b9f] Unregistered the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1081.255535] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-f263eae6-96b2-45f2-b03a-d97fe5d0f1d8 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] [instance: 6fd70b4a-63e4-4258-9cff-f3b582500b9f] Deleting contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1081.255715] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-f263eae6-96b2-45f2-b03a-d97fe5d0f1d8 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Deleting the datastore file [datastore1] 6fd70b4a-63e4-4258-9cff-f3b582500b9f {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1081.256020] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1f0a1af5-7666-4258-8d37-901000c103ea {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.266875] env[65726]: DEBUG oslo_vmware.api [None req-f263eae6-96b2-45f2-b03a-d97fe5d0f1d8 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Waiting for the task: (returnval){ [ 1081.266875] env[65726]: value = "task-5116674" [ 1081.266875] env[65726]: _type = "Task" [ 1081.266875] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1081.278727] env[65726]: DEBUG nova.scheduler.client.report [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1081.282253] env[65726]: DEBUG oslo_vmware.api [None req-f263eae6-96b2-45f2-b03a-d97fe5d0f1d8 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Task: {'id': task-5116674, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1081.344058] env[65726]: DEBUG oslo_vmware.api [None req-e7763bc8-6024-44cb-aee3-214195b1fd06 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': task-5116671, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1081.392412] env[65726]: DEBUG oslo_vmware.api [None req-91ff63ad-fe79-48d0-901e-fc0ed403574a tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': task-5116668, 'name': PowerOnVM_Task, 'duration_secs': 0.858693} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1081.392412] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-91ff63ad-fe79-48d0-901e-fc0ed403574a tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: 0d42fa75-20e9-4646-9b08-17015b7f068c] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1081.392412] env[65726]: INFO nova.compute.manager [None req-91ff63ad-fe79-48d0-901e-fc0ed403574a tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: 0d42fa75-20e9-4646-9b08-17015b7f068c] Took 7.74 seconds to spawn the instance on the hypervisor. [ 1081.392412] env[65726]: DEBUG nova.compute.manager [None req-91ff63ad-fe79-48d0-901e-fc0ed403574a tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: 0d42fa75-20e9-4646-9b08-17015b7f068c] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1081.394173] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df6bfc43-f6ea-4f7f-9088-ad7a1abc9936 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.398980] env[65726]: DEBUG nova.compute.manager [None req-9ca95ca0-14c4-4870-aa74-cd2a795fbbc9 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: d8801049-fc34-4b2f-8d5e-b9459c3e03a6] Start spawning the instance on the hypervisor. {{(pid=65726) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1081.437026] env[65726]: DEBUG nova.virt.hardware [None req-9ca95ca0-14c4-4870-aa74-cd2a795fbbc9 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T19:28:34Z,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1539f06fa3534e90acbc3a60c4b8bd3f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T19:28:35Z,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1081.437026] env[65726]: DEBUG nova.virt.hardware [None req-9ca95ca0-14c4-4870-aa74-cd2a795fbbc9 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1081.437026] env[65726]: DEBUG nova.virt.hardware [None req-9ca95ca0-14c4-4870-aa74-cd2a795fbbc9 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1081.437026] env[65726]: DEBUG nova.virt.hardware [None req-9ca95ca0-14c4-4870-aa74-cd2a795fbbc9 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1081.437026] env[65726]: DEBUG nova.virt.hardware [None req-9ca95ca0-14c4-4870-aa74-cd2a795fbbc9 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1081.437026] env[65726]: DEBUG nova.virt.hardware [None req-9ca95ca0-14c4-4870-aa74-cd2a795fbbc9 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1081.437026] env[65726]: DEBUG nova.virt.hardware [None req-9ca95ca0-14c4-4870-aa74-cd2a795fbbc9 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1081.437026] env[65726]: DEBUG nova.virt.hardware [None req-9ca95ca0-14c4-4870-aa74-cd2a795fbbc9 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1081.437495] env[65726]: DEBUG nova.virt.hardware [None req-9ca95ca0-14c4-4870-aa74-cd2a795fbbc9 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1081.437820] env[65726]: DEBUG nova.virt.hardware [None req-9ca95ca0-14c4-4870-aa74-cd2a795fbbc9 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1081.438138] env[65726]: DEBUG nova.virt.hardware [None req-9ca95ca0-14c4-4870-aa74-cd2a795fbbc9 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1081.439207] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94700f74-43c2-4f71-a457-2e57cd770ac2 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.450833] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93a4f32f-cf12-4d99-8a15-b40a3c126212 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.479622] env[65726]: DEBUG nova.network.neutron [None req-14a9b6f0-9c85-4271-a1df-1d81c0fc4a98 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: a56be6fa-6027-46ee-9aa8-332fe4222d68] Port 4e2dfbcd-41c3-4aaa-a2ec-50c855c12307 binding to destination host cpu-1 is already ACTIVE {{(pid=65726) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3236}} [ 1081.479913] env[65726]: DEBUG oslo_concurrency.lockutils [None req-14a9b6f0-9c85-4271-a1df-1d81c0fc4a98 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Acquiring lock "refresh_cache-a56be6fa-6027-46ee-9aa8-332fe4222d68" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1081.480149] env[65726]: DEBUG oslo_concurrency.lockutils [None req-14a9b6f0-9c85-4271-a1df-1d81c0fc4a98 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Acquired lock "refresh_cache-a56be6fa-6027-46ee-9aa8-332fe4222d68" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1081.480384] env[65726]: DEBUG nova.network.neutron [None req-14a9b6f0-9c85-4271-a1df-1d81c0fc4a98 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: a56be6fa-6027-46ee-9aa8-332fe4222d68] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1081.583925] env[65726]: DEBUG oslo_vmware.api [None req-8c26e8da-5d87-42a2-a8c5-15e9c65a8b9f tempest-InstanceActionsV221TestJSON-647428531 tempest-InstanceActionsV221TestJSON-647428531-project-member] Task: {'id': task-5116670, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.549541} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1081.584230] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-8c26e8da-5d87-42a2-a8c5-15e9c65a8b9f tempest-InstanceActionsV221TestJSON-647428531 tempest-InstanceActionsV221TestJSON-647428531-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore2] b0fec5ff-8c1e-41f9-b30a-6253a34baeec/b0fec5ff-8c1e-41f9-b30a-6253a34baeec.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1081.584442] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-8c26e8da-5d87-42a2-a8c5-15e9c65a8b9f tempest-InstanceActionsV221TestJSON-647428531 tempest-InstanceActionsV221TestJSON-647428531-project-member] [instance: b0fec5ff-8c1e-41f9-b30a-6253a34baeec] Extending root virtual disk to 1048576 {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1081.584704] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5421e3df-b643-464d-b51b-4932053f2c2a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.593412] env[65726]: DEBUG oslo_vmware.api [None req-8c26e8da-5d87-42a2-a8c5-15e9c65a8b9f tempest-InstanceActionsV221TestJSON-647428531 tempest-InstanceActionsV221TestJSON-647428531-project-member] Waiting for the task: (returnval){ [ 1081.593412] env[65726]: value = "task-5116675" [ 1081.593412] env[65726]: _type = "Task" [ 1081.593412] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1081.607033] env[65726]: DEBUG oslo_vmware.api [None req-8c26e8da-5d87-42a2-a8c5-15e9c65a8b9f tempest-InstanceActionsV221TestJSON-647428531 tempest-InstanceActionsV221TestJSON-647428531-project-member] Task: {'id': task-5116675, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1081.652231] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-a4ccf9b3-2389-45a5-bcf4-8de43acbac4c tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] [instance: 3b4d9461-747a-461c-b231-a0de02f0cb83] VM already powered off {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1081.652231] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-a4ccf9b3-2389-45a5-bcf4-8de43acbac4c tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] [instance: 3b4d9461-747a-461c-b231-a0de02f0cb83] Processing image b52362a3-ee8a-4cbf-b06f-513b0cc8f95c {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1081.652231] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a4ccf9b3-2389-45a5-bcf4-8de43acbac4c tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1081.652231] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a4ccf9b3-2389-45a5-bcf4-8de43acbac4c tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1081.652231] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-a4ccf9b3-2389-45a5-bcf4-8de43acbac4c tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1081.652231] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-db56ef0a-25ca-4570-895c-91339d7ea7ed {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.665693] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-a4ccf9b3-2389-45a5-bcf4-8de43acbac4c tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1081.665693] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-a4ccf9b3-2389-45a5-bcf4-8de43acbac4c tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65726) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1081.666489] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-77d0575b-be64-4ae0-8caa-45d981b7a242 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.674455] env[65726]: DEBUG oslo_vmware.api [None req-a4ccf9b3-2389-45a5-bcf4-8de43acbac4c tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Waiting for the task: (returnval){ [ 1081.674455] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52897834-e1a9-534a-131b-a715bdb0aa83" [ 1081.674455] env[65726]: _type = "Task" [ 1081.674455] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1081.683677] env[65726]: DEBUG oslo_vmware.api [None req-a4ccf9b3-2389-45a5-bcf4-8de43acbac4c tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52897834-e1a9-534a-131b-a715bdb0aa83, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1081.736834] env[65726]: DEBUG nova.compute.manager [req-875287d4-2129-4110-b590-50836062fb91 req-8e3010db-9e03-4e18-80de-4894b6c0abd4 service nova] [instance: 6fd70b4a-63e4-4258-9cff-f3b582500b9f] Received event network-vif-unplugged-153b4c9d-d01d-4254-9aa6-040705be347a {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1081.737114] env[65726]: DEBUG oslo_concurrency.lockutils [req-875287d4-2129-4110-b590-50836062fb91 req-8e3010db-9e03-4e18-80de-4894b6c0abd4 service nova] Acquiring lock "6fd70b4a-63e4-4258-9cff-f3b582500b9f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1081.737310] env[65726]: DEBUG oslo_concurrency.lockutils [req-875287d4-2129-4110-b590-50836062fb91 req-8e3010db-9e03-4e18-80de-4894b6c0abd4 service nova] Lock "6fd70b4a-63e4-4258-9cff-f3b582500b9f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1081.737469] env[65726]: DEBUG oslo_concurrency.lockutils [req-875287d4-2129-4110-b590-50836062fb91 req-8e3010db-9e03-4e18-80de-4894b6c0abd4 service nova] Lock "6fd70b4a-63e4-4258-9cff-f3b582500b9f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1081.738829] env[65726]: DEBUG nova.compute.manager [req-875287d4-2129-4110-b590-50836062fb91 req-8e3010db-9e03-4e18-80de-4894b6c0abd4 service nova] [instance: 6fd70b4a-63e4-4258-9cff-f3b582500b9f] No waiting events found dispatching network-vif-unplugged-153b4c9d-d01d-4254-9aa6-040705be347a {{(pid=65726) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1081.738829] env[65726]: WARNING nova.compute.manager [req-875287d4-2129-4110-b590-50836062fb91 req-8e3010db-9e03-4e18-80de-4894b6c0abd4 service nova] [instance: 6fd70b4a-63e4-4258-9cff-f3b582500b9f] Received unexpected event network-vif-unplugged-153b4c9d-d01d-4254-9aa6-040705be347a for instance with vm_state shelved and task_state shelving_offloading. [ 1081.776643] env[65726]: DEBUG oslo_vmware.api [None req-f263eae6-96b2-45f2-b03a-d97fe5d0f1d8 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Task: {'id': task-5116674, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1081.783769] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=65726) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1097}} [ 1081.783994] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.421s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1081.827338] env[65726]: DEBUG nova.compute.manager [req-91f8ab41-606a-46b5-abe3-5d1935f66539 req-93d5f4f1-c911-4400-9655-f783958b1f87 service nova] [instance: d8801049-fc34-4b2f-8d5e-b9459c3e03a6] Received event network-vif-plugged-b333a678-e6d6-494e-bc4d-7e498e370067 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1081.827562] env[65726]: DEBUG oslo_concurrency.lockutils [req-91f8ab41-606a-46b5-abe3-5d1935f66539 req-93d5f4f1-c911-4400-9655-f783958b1f87 service nova] Acquiring lock "d8801049-fc34-4b2f-8d5e-b9459c3e03a6-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1081.827764] env[65726]: DEBUG oslo_concurrency.lockutils [req-91f8ab41-606a-46b5-abe3-5d1935f66539 req-93d5f4f1-c911-4400-9655-f783958b1f87 service nova] Lock "d8801049-fc34-4b2f-8d5e-b9459c3e03a6-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1081.827928] env[65726]: DEBUG oslo_concurrency.lockutils [req-91f8ab41-606a-46b5-abe3-5d1935f66539 req-93d5f4f1-c911-4400-9655-f783958b1f87 service nova] Lock "d8801049-fc34-4b2f-8d5e-b9459c3e03a6-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1081.828106] env[65726]: DEBUG nova.compute.manager [req-91f8ab41-606a-46b5-abe3-5d1935f66539 req-93d5f4f1-c911-4400-9655-f783958b1f87 service nova] [instance: d8801049-fc34-4b2f-8d5e-b9459c3e03a6] No waiting events found dispatching network-vif-plugged-b333a678-e6d6-494e-bc4d-7e498e370067 {{(pid=65726) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1081.828266] env[65726]: WARNING nova.compute.manager [req-91f8ab41-606a-46b5-abe3-5d1935f66539 req-93d5f4f1-c911-4400-9655-f783958b1f87 service nova] [instance: d8801049-fc34-4b2f-8d5e-b9459c3e03a6] Received unexpected event network-vif-plugged-b333a678-e6d6-494e-bc4d-7e498e370067 for instance with vm_state building and task_state spawning. [ 1081.840470] env[65726]: DEBUG oslo_vmware.api [None req-e7763bc8-6024-44cb-aee3-214195b1fd06 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': task-5116671, 'name': PowerOnVM_Task, 'duration_secs': 0.522242} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1081.840470] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-e7763bc8-6024-44cb-aee3-214195b1fd06 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 5cdf1d61-b817-4986-a48c-804a2374497a] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1081.840673] env[65726]: DEBUG nova.compute.manager [None req-e7763bc8-6024-44cb-aee3-214195b1fd06 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 5cdf1d61-b817-4986-a48c-804a2374497a] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1081.841487] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a88efaf7-c64e-4172-85b8-7fda72544054 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.909702] env[65726]: DEBUG nova.network.neutron [None req-9ca95ca0-14c4-4870-aa74-cd2a795fbbc9 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: d8801049-fc34-4b2f-8d5e-b9459c3e03a6] Successfully updated port: b333a678-e6d6-494e-bc4d-7e498e370067 {{(pid=65726) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 1081.922962] env[65726]: INFO nova.compute.manager [None req-91ff63ad-fe79-48d0-901e-fc0ed403574a tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: 0d42fa75-20e9-4646-9b08-17015b7f068c] Took 17.14 seconds to build instance. [ 1081.983271] env[65726]: WARNING openstack [None req-14a9b6f0-9c85-4271-a1df-1d81c0fc4a98 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1081.983699] env[65726]: WARNING openstack [None req-14a9b6f0-9c85-4271-a1df-1d81c0fc4a98 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1082.104295] env[65726]: DEBUG oslo_vmware.api [None req-8c26e8da-5d87-42a2-a8c5-15e9c65a8b9f tempest-InstanceActionsV221TestJSON-647428531 tempest-InstanceActionsV221TestJSON-647428531-project-member] Task: {'id': task-5116675, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.075095} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1082.104666] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-8c26e8da-5d87-42a2-a8c5-15e9c65a8b9f tempest-InstanceActionsV221TestJSON-647428531 tempest-InstanceActionsV221TestJSON-647428531-project-member] [instance: b0fec5ff-8c1e-41f9-b30a-6253a34baeec] Extended root virtual disk {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1082.105387] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b07888fc-779b-4acc-a652-f4c359063a7d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.111311] env[65726]: WARNING openstack [None req-14a9b6f0-9c85-4271-a1df-1d81c0fc4a98 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1082.111667] env[65726]: WARNING openstack [None req-14a9b6f0-9c85-4271-a1df-1d81c0fc4a98 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1082.136342] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-8c26e8da-5d87-42a2-a8c5-15e9c65a8b9f tempest-InstanceActionsV221TestJSON-647428531 tempest-InstanceActionsV221TestJSON-647428531-project-member] [instance: b0fec5ff-8c1e-41f9-b30a-6253a34baeec] Reconfiguring VM instance instance-00000062 to attach disk [datastore2] b0fec5ff-8c1e-41f9-b30a-6253a34baeec/b0fec5ff-8c1e-41f9-b30a-6253a34baeec.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1082.136669] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a8637374-5879-4477-a6d8-0da60cb5be88 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.162069] env[65726]: DEBUG oslo_vmware.api [None req-8c26e8da-5d87-42a2-a8c5-15e9c65a8b9f tempest-InstanceActionsV221TestJSON-647428531 tempest-InstanceActionsV221TestJSON-647428531-project-member] Waiting for the task: (returnval){ [ 1082.162069] env[65726]: value = "task-5116676" [ 1082.162069] env[65726]: _type = "Task" [ 1082.162069] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1082.172632] env[65726]: DEBUG oslo_vmware.api [None req-8c26e8da-5d87-42a2-a8c5-15e9c65a8b9f tempest-InstanceActionsV221TestJSON-647428531 tempest-InstanceActionsV221TestJSON-647428531-project-member] Task: {'id': task-5116676, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1082.183675] env[65726]: DEBUG oslo_vmware.api [None req-a4ccf9b3-2389-45a5-bcf4-8de43acbac4c tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52897834-e1a9-534a-131b-a715bdb0aa83, 'name': SearchDatastore_Task, 'duration_secs': 0.083639} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1082.185056] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9f1d9250-ff17-4231-b5d5-005431f5b4c3 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.191418] env[65726]: WARNING openstack [None req-14a9b6f0-9c85-4271-a1df-1d81c0fc4a98 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1082.191808] env[65726]: WARNING openstack [None req-14a9b6f0-9c85-4271-a1df-1d81c0fc4a98 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1082.201920] env[65726]: DEBUG oslo_vmware.api [None req-a4ccf9b3-2389-45a5-bcf4-8de43acbac4c tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Waiting for the task: (returnval){ [ 1082.201920] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5209ea29-d882-7943-e2df-d9aa0654860f" [ 1082.201920] env[65726]: _type = "Task" [ 1082.201920] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1082.212032] env[65726]: DEBUG oslo_vmware.api [None req-a4ccf9b3-2389-45a5-bcf4-8de43acbac4c tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5209ea29-d882-7943-e2df-d9aa0654860f, 'name': SearchDatastore_Task, 'duration_secs': 0.0117} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1082.212317] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a4ccf9b3-2389-45a5-bcf4-8de43acbac4c tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1082.212628] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-a4ccf9b3-2389-45a5-bcf4-8de43acbac4c tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Copying virtual disk from [datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore2] 3b4d9461-747a-461c-b231-a0de02f0cb83/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c-rescue.vmdk. {{(pid=65726) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 1082.212915] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e6549a3f-e95b-44d1-aab5-0392ef5c485b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.222435] env[65726]: DEBUG oslo_vmware.api [None req-a4ccf9b3-2389-45a5-bcf4-8de43acbac4c tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Waiting for the task: (returnval){ [ 1082.222435] env[65726]: value = "task-5116677" [ 1082.222435] env[65726]: _type = "Task" [ 1082.222435] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1082.231479] env[65726]: DEBUG oslo_vmware.api [None req-a4ccf9b3-2389-45a5-bcf4-8de43acbac4c tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Task: {'id': task-5116677, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1082.277934] env[65726]: DEBUG oslo_vmware.api [None req-f263eae6-96b2-45f2-b03a-d97fe5d0f1d8 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Task: {'id': task-5116674, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.704062} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1082.278227] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-f263eae6-96b2-45f2-b03a-d97fe5d0f1d8 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Deleted the datastore file {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1082.278408] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-f263eae6-96b2-45f2-b03a-d97fe5d0f1d8 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] [instance: 6fd70b4a-63e4-4258-9cff-f3b582500b9f] Deleted contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1082.278714] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-f263eae6-96b2-45f2-b03a-d97fe5d0f1d8 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] [instance: 6fd70b4a-63e4-4258-9cff-f3b582500b9f] Instance destroyed {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1082.286069] env[65726]: DEBUG nova.network.neutron [None req-14a9b6f0-9c85-4271-a1df-1d81c0fc4a98 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: a56be6fa-6027-46ee-9aa8-332fe4222d68] Updating instance_info_cache with network_info: [{"id": "4e2dfbcd-41c3-4aaa-a2ec-50c855c12307", "address": "fa:16:3e:1d:82:59", "network": {"id": "0d0f561d-ac41-4ef5-9926-c707409799f7", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1134321217-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.160", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4eece77569624f90bf64e5c51974173f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2ce62383-8e84-4e26-955b-74c11392f4c9", "external-id": "nsx-vlan-transportzone-215", "segmentation_id": 215, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4e2dfbcd-41", "ovs_interfaceid": "4e2dfbcd-41c3-4aaa-a2ec-50c855c12307", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1082.306160] env[65726]: INFO nova.scheduler.client.report [None req-f263eae6-96b2-45f2-b03a-d97fe5d0f1d8 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Deleted allocations for instance 6fd70b4a-63e4-4258-9cff-f3b582500b9f [ 1082.360880] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e7763bc8-6024-44cb-aee3-214195b1fd06 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1082.361409] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e7763bc8-6024-44cb-aee3-214195b1fd06 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 0.001s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1082.361847] env[65726]: DEBUG nova.objects.instance [None req-e7763bc8-6024-44cb-aee3-214195b1fd06 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 5cdf1d61-b817-4986-a48c-804a2374497a] Trying to apply a migration context that does not seem to be set for this instance {{(pid=65726) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 1082.413133] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9ca95ca0-14c4-4870-aa74-cd2a795fbbc9 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Acquiring lock "refresh_cache-d8801049-fc34-4b2f-8d5e-b9459c3e03a6" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1082.413294] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9ca95ca0-14c4-4870-aa74-cd2a795fbbc9 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Acquired lock "refresh_cache-d8801049-fc34-4b2f-8d5e-b9459c3e03a6" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1082.413479] env[65726]: DEBUG nova.network.neutron [None req-9ca95ca0-14c4-4870-aa74-cd2a795fbbc9 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: d8801049-fc34-4b2f-8d5e-b9459c3e03a6] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1082.425609] env[65726]: DEBUG oslo_concurrency.lockutils [None req-91ff63ad-fe79-48d0-901e-fc0ed403574a tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Lock "0d42fa75-20e9-4646-9b08-17015b7f068c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 18.656s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1082.675997] env[65726]: DEBUG oslo_vmware.api [None req-8c26e8da-5d87-42a2-a8c5-15e9c65a8b9f tempest-InstanceActionsV221TestJSON-647428531 tempest-InstanceActionsV221TestJSON-647428531-project-member] Task: {'id': task-5116676, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1082.733891] env[65726]: DEBUG oslo_vmware.api [None req-a4ccf9b3-2389-45a5-bcf4-8de43acbac4c tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Task: {'id': task-5116677, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1082.788705] env[65726]: DEBUG oslo_concurrency.lockutils [None req-14a9b6f0-9c85-4271-a1df-1d81c0fc4a98 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Releasing lock "refresh_cache-a56be6fa-6027-46ee-9aa8-332fe4222d68" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1082.811860] env[65726]: DEBUG oslo_concurrency.lockutils [None req-f263eae6-96b2-45f2-b03a-d97fe5d0f1d8 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1082.920073] env[65726]: WARNING openstack [None req-9ca95ca0-14c4-4870-aa74-cd2a795fbbc9 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1082.920073] env[65726]: WARNING openstack [None req-9ca95ca0-14c4-4870-aa74-cd2a795fbbc9 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1082.924663] env[65726]: DEBUG nova.network.neutron [None req-9ca95ca0-14c4-4870-aa74-cd2a795fbbc9 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: d8801049-fc34-4b2f-8d5e-b9459c3e03a6] Instance cache missing network info. {{(pid=65726) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 1083.017805] env[65726]: WARNING openstack [None req-9ca95ca0-14c4-4870-aa74-cd2a795fbbc9 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1083.018463] env[65726]: WARNING openstack [None req-9ca95ca0-14c4-4870-aa74-cd2a795fbbc9 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1083.092884] env[65726]: WARNING openstack [None req-9ca95ca0-14c4-4870-aa74-cd2a795fbbc9 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1083.093312] env[65726]: WARNING openstack [None req-9ca95ca0-14c4-4870-aa74-cd2a795fbbc9 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1083.176177] env[65726]: DEBUG oslo_vmware.api [None req-8c26e8da-5d87-42a2-a8c5-15e9c65a8b9f tempest-InstanceActionsV221TestJSON-647428531 tempest-InstanceActionsV221TestJSON-647428531-project-member] Task: {'id': task-5116676, 'name': ReconfigVM_Task, 'duration_secs': 0.642458} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1083.176540] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-8c26e8da-5d87-42a2-a8c5-15e9c65a8b9f tempest-InstanceActionsV221TestJSON-647428531 tempest-InstanceActionsV221TestJSON-647428531-project-member] [instance: b0fec5ff-8c1e-41f9-b30a-6253a34baeec] Reconfigured VM instance instance-00000062 to attach disk [datastore2] b0fec5ff-8c1e-41f9-b30a-6253a34baeec/b0fec5ff-8c1e-41f9-b30a-6253a34baeec.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1083.177220] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-88f3ca80-bcec-4248-8304-c26137965d67 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.186166] env[65726]: DEBUG oslo_vmware.api [None req-8c26e8da-5d87-42a2-a8c5-15e9c65a8b9f tempest-InstanceActionsV221TestJSON-647428531 tempest-InstanceActionsV221TestJSON-647428531-project-member] Waiting for the task: (returnval){ [ 1083.186166] env[65726]: value = "task-5116678" [ 1083.186166] env[65726]: _type = "Task" [ 1083.186166] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1083.203295] env[65726]: DEBUG oslo_vmware.api [None req-8c26e8da-5d87-42a2-a8c5-15e9c65a8b9f tempest-InstanceActionsV221TestJSON-647428531 tempest-InstanceActionsV221TestJSON-647428531-project-member] Task: {'id': task-5116678, 'name': Rename_Task} progress is 6%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1083.221557] env[65726]: DEBUG nova.network.neutron [None req-9ca95ca0-14c4-4870-aa74-cd2a795fbbc9 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: d8801049-fc34-4b2f-8d5e-b9459c3e03a6] Updating instance_info_cache with network_info: [{"id": "b333a678-e6d6-494e-bc4d-7e498e370067", "address": "fa:16:3e:52:4a:a1", "network": {"id": "4c972d9d-db3b-4779-ac39-39d9a2d141dd", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1418913861-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cbb17a09b35c4c22ade5c4082b10bda9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "50cf0a70-948d-4611-af05-94c1483064ed", "external-id": "nsx-vlan-transportzone-536", "segmentation_id": 536, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb333a678-e6", "ovs_interfaceid": "b333a678-e6d6-494e-bc4d-7e498e370067", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1083.235019] env[65726]: DEBUG oslo_vmware.api [None req-a4ccf9b3-2389-45a5-bcf4-8de43acbac4c tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Task: {'id': task-5116677, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.524258} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1083.235296] env[65726]: INFO nova.virt.vmwareapi.ds_util [None req-a4ccf9b3-2389-45a5-bcf4-8de43acbac4c tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Copied virtual disk from [datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore2] 3b4d9461-747a-461c-b231-a0de02f0cb83/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c-rescue.vmdk. [ 1083.236116] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d83fcc1c-7955-444b-9546-3023d68a11f6 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.264442] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-a4ccf9b3-2389-45a5-bcf4-8de43acbac4c tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] [instance: 3b4d9461-747a-461c-b231-a0de02f0cb83] Reconfiguring VM instance instance-00000060 to attach disk [datastore2] 3b4d9461-747a-461c-b231-a0de02f0cb83/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c-rescue.vmdk or device None with type thin {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1083.265093] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5e5855e5-f2ca-4b51-ad83-46459c3e23bc {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.285876] env[65726]: DEBUG oslo_vmware.api [None req-a4ccf9b3-2389-45a5-bcf4-8de43acbac4c tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Waiting for the task: (returnval){ [ 1083.285876] env[65726]: value = "task-5116679" [ 1083.285876] env[65726]: _type = "Task" [ 1083.285876] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1083.295724] env[65726]: DEBUG nova.compute.manager [None req-14a9b6f0-9c85-4271-a1df-1d81c0fc4a98 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: a56be6fa-6027-46ee-9aa8-332fe4222d68] Hypervisor driver does not support instance shared storage check, assuming it's not on shared storage {{(pid=65726) _is_instance_storage_shared /opt/stack/nova/nova/compute/manager.py:924}} [ 1083.297850] env[65726]: DEBUG oslo_vmware.api [None req-a4ccf9b3-2389-45a5-bcf4-8de43acbac4c tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Task: {'id': task-5116679, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1083.374478] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e7763bc8-6024-44cb-aee3-214195b1fd06 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.013s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1083.375884] env[65726]: DEBUG oslo_concurrency.lockutils [None req-f263eae6-96b2-45f2-b03a-d97fe5d0f1d8 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.565s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1083.380279] env[65726]: DEBUG nova.objects.instance [None req-f263eae6-96b2-45f2-b03a-d97fe5d0f1d8 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Lazy-loading 'resources' on Instance uuid 6fd70b4a-63e4-4258-9cff-f3b582500b9f {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1083.536637] env[65726]: DEBUG oslo_concurrency.lockutils [None req-4f5aff8f-c14b-4955-a84c-9d59624f9d81 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Acquiring lock "5cdf1d61-b817-4986-a48c-804a2374497a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1083.536922] env[65726]: DEBUG oslo_concurrency.lockutils [None req-4f5aff8f-c14b-4955-a84c-9d59624f9d81 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Lock "5cdf1d61-b817-4986-a48c-804a2374497a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1083.537196] env[65726]: DEBUG oslo_concurrency.lockutils [None req-4f5aff8f-c14b-4955-a84c-9d59624f9d81 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Acquiring lock "5cdf1d61-b817-4986-a48c-804a2374497a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1083.537389] env[65726]: DEBUG oslo_concurrency.lockutils [None req-4f5aff8f-c14b-4955-a84c-9d59624f9d81 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Lock "5cdf1d61-b817-4986-a48c-804a2374497a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1083.537558] env[65726]: DEBUG oslo_concurrency.lockutils [None req-4f5aff8f-c14b-4955-a84c-9d59624f9d81 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Lock "5cdf1d61-b817-4986-a48c-804a2374497a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1083.539870] env[65726]: INFO nova.compute.manager [None req-4f5aff8f-c14b-4955-a84c-9d59624f9d81 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 5cdf1d61-b817-4986-a48c-804a2374497a] Terminating instance [ 1083.672262] env[65726]: DEBUG oslo_concurrency.lockutils [None req-dd8aa948-a330-4efc-a6ba-843d94f1837a tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Acquiring lock "aa6ce489-c62f-4481-87b7-e74242aeb8ca" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1083.672552] env[65726]: DEBUG oslo_concurrency.lockutils [None req-dd8aa948-a330-4efc-a6ba-843d94f1837a tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Lock "aa6ce489-c62f-4481-87b7-e74242aeb8ca" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1083.699359] env[65726]: DEBUG oslo_vmware.api [None req-8c26e8da-5d87-42a2-a8c5-15e9c65a8b9f tempest-InstanceActionsV221TestJSON-647428531 tempest-InstanceActionsV221TestJSON-647428531-project-member] Task: {'id': task-5116678, 'name': Rename_Task, 'duration_secs': 0.351595} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1083.699653] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-8c26e8da-5d87-42a2-a8c5-15e9c65a8b9f tempest-InstanceActionsV221TestJSON-647428531 tempest-InstanceActionsV221TestJSON-647428531-project-member] [instance: b0fec5ff-8c1e-41f9-b30a-6253a34baeec] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1083.699904] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d9f15aae-c3a8-46b8-b1d6-9cf45e06653d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.709016] env[65726]: DEBUG oslo_vmware.api [None req-8c26e8da-5d87-42a2-a8c5-15e9c65a8b9f tempest-InstanceActionsV221TestJSON-647428531 tempest-InstanceActionsV221TestJSON-647428531-project-member] Waiting for the task: (returnval){ [ 1083.709016] env[65726]: value = "task-5116680" [ 1083.709016] env[65726]: _type = "Task" [ 1083.709016] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1083.724168] env[65726]: DEBUG oslo_vmware.api [None req-8c26e8da-5d87-42a2-a8c5-15e9c65a8b9f tempest-InstanceActionsV221TestJSON-647428531 tempest-InstanceActionsV221TestJSON-647428531-project-member] Task: {'id': task-5116680, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1083.724372] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9ca95ca0-14c4-4870-aa74-cd2a795fbbc9 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Releasing lock "refresh_cache-d8801049-fc34-4b2f-8d5e-b9459c3e03a6" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1083.724729] env[65726]: DEBUG nova.compute.manager [None req-9ca95ca0-14c4-4870-aa74-cd2a795fbbc9 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: d8801049-fc34-4b2f-8d5e-b9459c3e03a6] Instance network_info: |[{"id": "b333a678-e6d6-494e-bc4d-7e498e370067", "address": "fa:16:3e:52:4a:a1", "network": {"id": "4c972d9d-db3b-4779-ac39-39d9a2d141dd", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1418913861-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cbb17a09b35c4c22ade5c4082b10bda9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "50cf0a70-948d-4611-af05-94c1483064ed", "external-id": "nsx-vlan-transportzone-536", "segmentation_id": 536, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb333a678-e6", "ovs_interfaceid": "b333a678-e6d6-494e-bc4d-7e498e370067", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1083.725668] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-9ca95ca0-14c4-4870-aa74-cd2a795fbbc9 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: d8801049-fc34-4b2f-8d5e-b9459c3e03a6] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:52:4a:a1', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '50cf0a70-948d-4611-af05-94c1483064ed', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b333a678-e6d6-494e-bc4d-7e498e370067', 'vif_model': 'vmxnet3'}] {{(pid=65726) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1083.736498] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-9ca95ca0-14c4-4870-aa74-cd2a795fbbc9 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1083.736498] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d8801049-fc34-4b2f-8d5e-b9459c3e03a6] Creating VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1083.736498] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-bcdcbb12-7311-4029-ae93-5616415cdb22 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.762024] env[65726]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1083.762024] env[65726]: value = "task-5116681" [ 1083.762024] env[65726]: _type = "Task" [ 1083.762024] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1083.773285] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116681, 'name': CreateVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1083.798396] env[65726]: DEBUG oslo_vmware.api [None req-a4ccf9b3-2389-45a5-bcf4-8de43acbac4c tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Task: {'id': task-5116679, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1083.884041] env[65726]: DEBUG nova.objects.instance [None req-f263eae6-96b2-45f2-b03a-d97fe5d0f1d8 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Lazy-loading 'numa_topology' on Instance uuid 6fd70b4a-63e4-4258-9cff-f3b582500b9f {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1084.046137] env[65726]: DEBUG nova.compute.manager [None req-4f5aff8f-c14b-4955-a84c-9d59624f9d81 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 5cdf1d61-b817-4986-a48c-804a2374497a] Start destroying the instance on the hypervisor. {{(pid=65726) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1084.046137] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-4f5aff8f-c14b-4955-a84c-9d59624f9d81 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 5cdf1d61-b817-4986-a48c-804a2374497a] Destroying instance {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1084.046418] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8971171-d756-46a8-a01e-22a7001fd9fb {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.055710] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-4f5aff8f-c14b-4955-a84c-9d59624f9d81 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 5cdf1d61-b817-4986-a48c-804a2374497a] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1084.056037] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0b775524-d2ba-4711-8f5c-56b30ad072fb {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.069230] env[65726]: DEBUG oslo_vmware.api [None req-4f5aff8f-c14b-4955-a84c-9d59624f9d81 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Waiting for the task: (returnval){ [ 1084.069230] env[65726]: value = "task-5116682" [ 1084.069230] env[65726]: _type = "Task" [ 1084.069230] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1084.085572] env[65726]: DEBUG oslo_vmware.api [None req-4f5aff8f-c14b-4955-a84c-9d59624f9d81 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': task-5116682, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1084.096051] env[65726]: DEBUG nova.compute.manager [req-708292a0-fd27-4a3f-8519-16ad5ba287b3 req-6f58ab55-2aa3-4257-b0cf-7b2339f092a6 service nova] [instance: 6fd70b4a-63e4-4258-9cff-f3b582500b9f] Received event network-changed-153b4c9d-d01d-4254-9aa6-040705be347a {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1084.096394] env[65726]: DEBUG nova.compute.manager [req-708292a0-fd27-4a3f-8519-16ad5ba287b3 req-6f58ab55-2aa3-4257-b0cf-7b2339f092a6 service nova] [instance: 6fd70b4a-63e4-4258-9cff-f3b582500b9f] Refreshing instance network info cache due to event network-changed-153b4c9d-d01d-4254-9aa6-040705be347a. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 1084.096658] env[65726]: DEBUG oslo_concurrency.lockutils [req-708292a0-fd27-4a3f-8519-16ad5ba287b3 req-6f58ab55-2aa3-4257-b0cf-7b2339f092a6 service nova] Acquiring lock "refresh_cache-6fd70b4a-63e4-4258-9cff-f3b582500b9f" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1084.096862] env[65726]: DEBUG oslo_concurrency.lockutils [req-708292a0-fd27-4a3f-8519-16ad5ba287b3 req-6f58ab55-2aa3-4257-b0cf-7b2339f092a6 service nova] Acquired lock "refresh_cache-6fd70b4a-63e4-4258-9cff-f3b582500b9f" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1084.097154] env[65726]: DEBUG nova.network.neutron [req-708292a0-fd27-4a3f-8519-16ad5ba287b3 req-6f58ab55-2aa3-4257-b0cf-7b2339f092a6 service nova] [instance: 6fd70b4a-63e4-4258-9cff-f3b582500b9f] Refreshing network info cache for port 153b4c9d-d01d-4254-9aa6-040705be347a {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 1084.119350] env[65726]: DEBUG nova.compute.manager [req-b0ae93f4-f230-4542-aaac-0ce403689f36 req-530b0252-afb1-4ea1-9240-a7b795baeea4 service nova] [instance: d8801049-fc34-4b2f-8d5e-b9459c3e03a6] Received event network-changed-b333a678-e6d6-494e-bc4d-7e498e370067 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1084.119350] env[65726]: DEBUG nova.compute.manager [req-b0ae93f4-f230-4542-aaac-0ce403689f36 req-530b0252-afb1-4ea1-9240-a7b795baeea4 service nova] [instance: d8801049-fc34-4b2f-8d5e-b9459c3e03a6] Refreshing instance network info cache due to event network-changed-b333a678-e6d6-494e-bc4d-7e498e370067. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 1084.119350] env[65726]: DEBUG oslo_concurrency.lockutils [req-b0ae93f4-f230-4542-aaac-0ce403689f36 req-530b0252-afb1-4ea1-9240-a7b795baeea4 service nova] Acquiring lock "refresh_cache-d8801049-fc34-4b2f-8d5e-b9459c3e03a6" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1084.119350] env[65726]: DEBUG oslo_concurrency.lockutils [req-b0ae93f4-f230-4542-aaac-0ce403689f36 req-530b0252-afb1-4ea1-9240-a7b795baeea4 service nova] Acquired lock "refresh_cache-d8801049-fc34-4b2f-8d5e-b9459c3e03a6" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1084.119777] env[65726]: DEBUG nova.network.neutron [req-b0ae93f4-f230-4542-aaac-0ce403689f36 req-530b0252-afb1-4ea1-9240-a7b795baeea4 service nova] [instance: d8801049-fc34-4b2f-8d5e-b9459c3e03a6] Refreshing network info cache for port b333a678-e6d6-494e-bc4d-7e498e370067 {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 1084.177047] env[65726]: INFO nova.compute.manager [None req-dd8aa948-a330-4efc-a6ba-843d94f1837a tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] [instance: aa6ce489-c62f-4481-87b7-e74242aeb8ca] Detaching volume 3547dfa9-73e6-4baa-b777-e22891db153f [ 1084.217091] env[65726]: INFO nova.virt.block_device [None req-dd8aa948-a330-4efc-a6ba-843d94f1837a tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] [instance: aa6ce489-c62f-4481-87b7-e74242aeb8ca] Attempting to driver detach volume 3547dfa9-73e6-4baa-b777-e22891db153f from mountpoint /dev/sdb [ 1084.217409] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-dd8aa948-a330-4efc-a6ba-843d94f1837a tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] [instance: aa6ce489-c62f-4481-87b7-e74242aeb8ca] Volume detach. Driver type: vmdk {{(pid=65726) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1084.218163] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-dd8aa948-a330-4efc-a6ba-843d94f1837a tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] [instance: aa6ce489-c62f-4481-87b7-e74242aeb8ca] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-995264', 'volume_id': '3547dfa9-73e6-4baa-b777-e22891db153f', 'name': 'volume-3547dfa9-73e6-4baa-b777-e22891db153f', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'aa6ce489-c62f-4481-87b7-e74242aeb8ca', 'attached_at': '', 'detached_at': '', 'volume_id': '3547dfa9-73e6-4baa-b777-e22891db153f', 'serial': '3547dfa9-73e6-4baa-b777-e22891db153f'} {{(pid=65726) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1084.219405] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7601abd5-b5c1-4ddb-a8d4-febb82c99ff4 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.227204] env[65726]: DEBUG oslo_vmware.api [None req-8c26e8da-5d87-42a2-a8c5-15e9c65a8b9f tempest-InstanceActionsV221TestJSON-647428531 tempest-InstanceActionsV221TestJSON-647428531-project-member] Task: {'id': task-5116680, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1084.249870] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e950d94-569e-461f-859e-9bbedf008f17 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.259148] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a040e449-705b-4954-a7b5-651112459b22 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.273269] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116681, 'name': CreateVM_Task, 'duration_secs': 0.384011} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1084.289957] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d8801049-fc34-4b2f-8d5e-b9459c3e03a6] Created VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1084.291562] env[65726]: WARNING openstack [None req-9ca95ca0-14c4-4870-aa74-cd2a795fbbc9 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1084.292228] env[65726]: WARNING openstack [None req-9ca95ca0-14c4-4870-aa74-cd2a795fbbc9 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1084.298517] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9ca95ca0-14c4-4870-aa74-cd2a795fbbc9 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1084.298778] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9ca95ca0-14c4-4870-aa74-cd2a795fbbc9 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1084.299049] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9ca95ca0-14c4-4870-aa74-cd2a795fbbc9 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1084.303210] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-626523c4-8c8b-4aa3-95ae-5a94b2f036c4 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.309538] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2e89609d-fa07-499b-ad53-cf263e0370a9 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.331874] env[65726]: DEBUG oslo_vmware.api [None req-9ca95ca0-14c4-4870-aa74-cd2a795fbbc9 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Waiting for the task: (returnval){ [ 1084.331874] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52788563-f17f-2127-b4e9-8b35d9de91f8" [ 1084.331874] env[65726]: _type = "Task" [ 1084.331874] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1084.332289] env[65726]: DEBUG oslo_vmware.api [None req-a4ccf9b3-2389-45a5-bcf4-8de43acbac4c tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Task: {'id': task-5116679, 'name': ReconfigVM_Task, 'duration_secs': 0.746845} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1084.332628] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-dd8aa948-a330-4efc-a6ba-843d94f1837a tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] The volume has not been displaced from its original location: [datastore1] volume-3547dfa9-73e6-4baa-b777-e22891db153f/volume-3547dfa9-73e6-4baa-b777-e22891db153f.vmdk. No consolidation needed. {{(pid=65726) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1084.338614] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-dd8aa948-a330-4efc-a6ba-843d94f1837a tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] [instance: aa6ce489-c62f-4481-87b7-e74242aeb8ca] Reconfiguring VM instance instance-0000004e to detach disk 2001 {{(pid=65726) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1084.339227] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-a4ccf9b3-2389-45a5-bcf4-8de43acbac4c tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] [instance: 3b4d9461-747a-461c-b231-a0de02f0cb83] Reconfigured VM instance instance-00000060 to attach disk [datastore2] 3b4d9461-747a-461c-b231-a0de02f0cb83/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c-rescue.vmdk or device None with type thin {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1084.339609] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bc42f158-2a3d-4a32-a184-3421e85b074c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.358194] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b17c737-7bf1-47be-9648-970a29343c53 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.391945] env[65726]: DEBUG nova.objects.base [None req-f263eae6-96b2-45f2-b03a-d97fe5d0f1d8 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Object Instance<6fd70b4a-63e4-4258-9cff-f3b582500b9f> lazy-loaded attributes: resources,numa_topology {{(pid=65726) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1084.394759] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8b805a0f-15e5-4dcb-9f8c-08345f7da515 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.405994] env[65726]: DEBUG oslo_vmware.api [None req-dd8aa948-a330-4efc-a6ba-843d94f1837a tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Waiting for the task: (returnval){ [ 1084.405994] env[65726]: value = "task-5116683" [ 1084.405994] env[65726]: _type = "Task" [ 1084.405994] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1084.406283] env[65726]: DEBUG oslo_vmware.api [None req-9ca95ca0-14c4-4870-aa74-cd2a795fbbc9 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52788563-f17f-2127-b4e9-8b35d9de91f8, 'name': SearchDatastore_Task, 'duration_secs': 0.015339} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1084.408151] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9ca95ca0-14c4-4870-aa74-cd2a795fbbc9 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1084.408151] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-9ca95ca0-14c4-4870-aa74-cd2a795fbbc9 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: d8801049-fc34-4b2f-8d5e-b9459c3e03a6] Processing image b52362a3-ee8a-4cbf-b06f-513b0cc8f95c {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1084.408151] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9ca95ca0-14c4-4870-aa74-cd2a795fbbc9 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1084.408151] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9ca95ca0-14c4-4870-aa74-cd2a795fbbc9 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1084.408151] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-9ca95ca0-14c4-4870-aa74-cd2a795fbbc9 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1084.413807] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-21ec84c0-8ed1-4b43-a17b-1558ed62e946 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.416546] env[65726]: DEBUG oslo_vmware.api [None req-a4ccf9b3-2389-45a5-bcf4-8de43acbac4c tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Waiting for the task: (returnval){ [ 1084.416546] env[65726]: value = "task-5116684" [ 1084.416546] env[65726]: _type = "Task" [ 1084.416546] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1084.425682] env[65726]: DEBUG oslo_vmware.api [None req-dd8aa948-a330-4efc-a6ba-843d94f1837a tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Task: {'id': task-5116683, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1084.428044] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-9ca95ca0-14c4-4870-aa74-cd2a795fbbc9 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1084.428044] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-9ca95ca0-14c4-4870-aa74-cd2a795fbbc9 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65726) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1084.428703] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-769e5c12-dcf5-48ef-9897-c116aec8bf9c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.436301] env[65726]: DEBUG oslo_concurrency.lockutils [None req-14a9b6f0-9c85-4271-a1df-1d81c0fc4a98 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1084.436301] env[65726]: DEBUG oslo_vmware.api [None req-a4ccf9b3-2389-45a5-bcf4-8de43acbac4c tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Task: {'id': task-5116684, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1084.439859] env[65726]: DEBUG oslo_vmware.api [None req-9ca95ca0-14c4-4870-aa74-cd2a795fbbc9 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Waiting for the task: (returnval){ [ 1084.439859] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52cabf33-082b-f666-a9b3-962da9d98f29" [ 1084.439859] env[65726]: _type = "Task" [ 1084.439859] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1084.463009] env[65726]: DEBUG oslo_vmware.api [None req-9ca95ca0-14c4-4870-aa74-cd2a795fbbc9 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52cabf33-082b-f666-a9b3-962da9d98f29, 'name': SearchDatastore_Task, 'duration_secs': 0.012567} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1084.464420] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cefe2127-a711-40c5-b3b4-df5382e1b009 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.474559] env[65726]: DEBUG oslo_vmware.api [None req-9ca95ca0-14c4-4870-aa74-cd2a795fbbc9 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Waiting for the task: (returnval){ [ 1084.474559] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52c2d97d-5c3b-41e1-8429-c6807191e179" [ 1084.474559] env[65726]: _type = "Task" [ 1084.474559] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1084.491147] env[65726]: DEBUG oslo_vmware.api [None req-9ca95ca0-14c4-4870-aa74-cd2a795fbbc9 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52c2d97d-5c3b-41e1-8429-c6807191e179, 'name': SearchDatastore_Task, 'duration_secs': 0.014955} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1084.491496] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9ca95ca0-14c4-4870-aa74-cd2a795fbbc9 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1084.491712] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-9ca95ca0-14c4-4870-aa74-cd2a795fbbc9 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] d8801049-fc34-4b2f-8d5e-b9459c3e03a6/d8801049-fc34-4b2f-8d5e-b9459c3e03a6.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1084.492372] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5edf0345-ee4e-4919-992c-d2a23427a942 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.504402] env[65726]: DEBUG oslo_vmware.api [None req-9ca95ca0-14c4-4870-aa74-cd2a795fbbc9 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Waiting for the task: (returnval){ [ 1084.504402] env[65726]: value = "task-5116685" [ 1084.504402] env[65726]: _type = "Task" [ 1084.504402] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1084.518289] env[65726]: DEBUG oslo_vmware.api [None req-9ca95ca0-14c4-4870-aa74-cd2a795fbbc9 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': task-5116685, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1084.553037] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5101d24a-5b00-4587-8b37-5a45f8c4ea31 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Acquiring lock "0361ae8f-634d-4161-a594-7f93dbb1b7cd" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1084.554073] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5101d24a-5b00-4587-8b37-5a45f8c4ea31 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Lock "0361ae8f-634d-4161-a594-7f93dbb1b7cd" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1084.587797] env[65726]: DEBUG oslo_vmware.api [None req-4f5aff8f-c14b-4955-a84c-9d59624f9d81 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': task-5116682, 'name': PowerOffVM_Task, 'duration_secs': 0.210374} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1084.589159] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-4f5aff8f-c14b-4955-a84c-9d59624f9d81 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 5cdf1d61-b817-4986-a48c-804a2374497a] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1084.589159] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-4f5aff8f-c14b-4955-a84c-9d59624f9d81 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 5cdf1d61-b817-4986-a48c-804a2374497a] Unregistering the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1084.589159] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ab3fd6e6-223a-4764-a3a9-b053ad3dde32 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.600353] env[65726]: WARNING openstack [req-708292a0-fd27-4a3f-8519-16ad5ba287b3 req-6f58ab55-2aa3-4257-b0cf-7b2339f092a6 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1084.600862] env[65726]: WARNING openstack [req-708292a0-fd27-4a3f-8519-16ad5ba287b3 req-6f58ab55-2aa3-4257-b0cf-7b2339f092a6 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1084.624862] env[65726]: WARNING openstack [req-b0ae93f4-f230-4542-aaac-0ce403689f36 req-530b0252-afb1-4ea1-9240-a7b795baeea4 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1084.625290] env[65726]: WARNING openstack [req-b0ae93f4-f230-4542-aaac-0ce403689f36 req-530b0252-afb1-4ea1-9240-a7b795baeea4 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1084.679383] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-4f5aff8f-c14b-4955-a84c-9d59624f9d81 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 5cdf1d61-b817-4986-a48c-804a2374497a] Unregistered the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1084.680190] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-4f5aff8f-c14b-4955-a84c-9d59624f9d81 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 5cdf1d61-b817-4986-a48c-804a2374497a] Deleting contents of the VM from datastore datastore2 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1084.680190] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-4f5aff8f-c14b-4955-a84c-9d59624f9d81 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Deleting the datastore file [datastore2] 5cdf1d61-b817-4986-a48c-804a2374497a {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1084.680190] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1da61c4f-06bd-4689-a873-922b7389c0be {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.687764] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a14e722-ce6d-46d2-a064-14d6a61d06ca {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.693105] env[65726]: DEBUG oslo_vmware.api [None req-4f5aff8f-c14b-4955-a84c-9d59624f9d81 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Waiting for the task: (returnval){ [ 1084.693105] env[65726]: value = "task-5116687" [ 1084.693105] env[65726]: _type = "Task" [ 1084.693105] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1084.701645] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b576af1-fbf1-44df-b96c-0054abd197ee {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.713794] env[65726]: DEBUG oslo_vmware.api [None req-4f5aff8f-c14b-4955-a84c-9d59624f9d81 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': task-5116687, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1084.750060] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56a3d276-9f59-4b04-bb10-de2edad16538 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.756397] env[65726]: DEBUG oslo_vmware.api [None req-8c26e8da-5d87-42a2-a8c5-15e9c65a8b9f tempest-InstanceActionsV221TestJSON-647428531 tempest-InstanceActionsV221TestJSON-647428531-project-member] Task: {'id': task-5116680, 'name': PowerOnVM_Task, 'duration_secs': 0.637903} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1084.757080] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-8c26e8da-5d87-42a2-a8c5-15e9c65a8b9f tempest-InstanceActionsV221TestJSON-647428531 tempest-InstanceActionsV221TestJSON-647428531-project-member] [instance: b0fec5ff-8c1e-41f9-b30a-6253a34baeec] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1084.757349] env[65726]: INFO nova.compute.manager [None req-8c26e8da-5d87-42a2-a8c5-15e9c65a8b9f tempest-InstanceActionsV221TestJSON-647428531 tempest-InstanceActionsV221TestJSON-647428531-project-member] [instance: b0fec5ff-8c1e-41f9-b30a-6253a34baeec] Took 7.83 seconds to spawn the instance on the hypervisor. [ 1084.757559] env[65726]: DEBUG nova.compute.manager [None req-8c26e8da-5d87-42a2-a8c5-15e9c65a8b9f tempest-InstanceActionsV221TestJSON-647428531 tempest-InstanceActionsV221TestJSON-647428531-project-member] [instance: b0fec5ff-8c1e-41f9-b30a-6253a34baeec] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1084.758355] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d06c2ec7-43ac-4bbc-bade-d88224b403e6 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.764969] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-076131eb-33ae-4a66-8340-9666b695fedc {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.792009] env[65726]: DEBUG nova.compute.provider_tree [None req-f263eae6-96b2-45f2-b03a-d97fe5d0f1d8 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1084.794823] env[65726]: WARNING openstack [req-708292a0-fd27-4a3f-8519-16ad5ba287b3 req-6f58ab55-2aa3-4257-b0cf-7b2339f092a6 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1084.795251] env[65726]: WARNING openstack [req-708292a0-fd27-4a3f-8519-16ad5ba287b3 req-6f58ab55-2aa3-4257-b0cf-7b2339f092a6 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1084.803351] env[65726]: WARNING openstack [req-b0ae93f4-f230-4542-aaac-0ce403689f36 req-530b0252-afb1-4ea1-9240-a7b795baeea4 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1084.803461] env[65726]: WARNING openstack [req-b0ae93f4-f230-4542-aaac-0ce403689f36 req-530b0252-afb1-4ea1-9240-a7b795baeea4 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1084.876293] env[65726]: WARNING openstack [req-708292a0-fd27-4a3f-8519-16ad5ba287b3 req-6f58ab55-2aa3-4257-b0cf-7b2339f092a6 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1084.876662] env[65726]: WARNING openstack [req-708292a0-fd27-4a3f-8519-16ad5ba287b3 req-6f58ab55-2aa3-4257-b0cf-7b2339f092a6 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1084.903116] env[65726]: WARNING openstack [req-b0ae93f4-f230-4542-aaac-0ce403689f36 req-530b0252-afb1-4ea1-9240-a7b795baeea4 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1084.903522] env[65726]: WARNING openstack [req-b0ae93f4-f230-4542-aaac-0ce403689f36 req-530b0252-afb1-4ea1-9240-a7b795baeea4 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1084.923914] env[65726]: DEBUG oslo_vmware.api [None req-dd8aa948-a330-4efc-a6ba-843d94f1837a tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Task: {'id': task-5116683, 'name': ReconfigVM_Task, 'duration_secs': 0.290051} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1084.924686] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-dd8aa948-a330-4efc-a6ba-843d94f1837a tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] [instance: aa6ce489-c62f-4481-87b7-e74242aeb8ca] Reconfigured VM instance instance-0000004e to detach disk 2001 {{(pid=65726) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1084.932332] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-104f992a-8201-4a38-8c0e-122e4e6e4038 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.942373] env[65726]: DEBUG oslo_vmware.api [None req-a4ccf9b3-2389-45a5-bcf4-8de43acbac4c tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Task: {'id': task-5116684, 'name': ReconfigVM_Task, 'duration_secs': 0.190476} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1084.946787] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-a4ccf9b3-2389-45a5-bcf4-8de43acbac4c tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] [instance: 3b4d9461-747a-461c-b231-a0de02f0cb83] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1084.947649] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5af98892-3ba5-40a9-ab14-bd2e5207dc7f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.953777] env[65726]: DEBUG oslo_vmware.api [None req-dd8aa948-a330-4efc-a6ba-843d94f1837a tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Waiting for the task: (returnval){ [ 1084.953777] env[65726]: value = "task-5116688" [ 1084.953777] env[65726]: _type = "Task" [ 1084.953777] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1084.958241] env[65726]: DEBUG oslo_concurrency.lockutils [None req-1c19f444-aee2-49ca-b67e-d3cef83d3664 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Acquiring lock "6fd70b4a-63e4-4258-9cff-f3b582500b9f" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1084.958567] env[65726]: DEBUG oslo_vmware.api [None req-a4ccf9b3-2389-45a5-bcf4-8de43acbac4c tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Waiting for the task: (returnval){ [ 1084.958567] env[65726]: value = "task-5116689" [ 1084.958567] env[65726]: _type = "Task" [ 1084.958567] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1084.967018] env[65726]: DEBUG oslo_vmware.api [None req-dd8aa948-a330-4efc-a6ba-843d94f1837a tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Task: {'id': task-5116688, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1084.972849] env[65726]: DEBUG oslo_vmware.api [None req-a4ccf9b3-2389-45a5-bcf4-8de43acbac4c tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Task: {'id': task-5116689, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1085.016311] env[65726]: DEBUG oslo_vmware.api [None req-9ca95ca0-14c4-4870-aa74-cd2a795fbbc9 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': task-5116685, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1085.022669] env[65726]: DEBUG nova.network.neutron [req-708292a0-fd27-4a3f-8519-16ad5ba287b3 req-6f58ab55-2aa3-4257-b0cf-7b2339f092a6 service nova] [instance: 6fd70b4a-63e4-4258-9cff-f3b582500b9f] Updated VIF entry in instance network info cache for port 153b4c9d-d01d-4254-9aa6-040705be347a. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 1085.023114] env[65726]: DEBUG nova.network.neutron [req-708292a0-fd27-4a3f-8519-16ad5ba287b3 req-6f58ab55-2aa3-4257-b0cf-7b2339f092a6 service nova] [instance: 6fd70b4a-63e4-4258-9cff-f3b582500b9f] Updating instance_info_cache with network_info: [{"id": "153b4c9d-d01d-4254-9aa6-040705be347a", "address": "fa:16:3e:17:18:d9", "network": {"id": "93e38d7e-b4d7-4d4a-8393-64ab2f6769de", "bridge": null, "label": "tempest-ServersNegativeTestJSON-1445539922-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "06e221dc693640929cac7bc5af45948f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap153b4c9d-d0", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1085.059082] env[65726]: DEBUG nova.compute.manager [None req-5101d24a-5b00-4587-8b37-5a45f8c4ea31 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 0361ae8f-634d-4161-a594-7f93dbb1b7cd] Starting instance... {{(pid=65726) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 1085.099991] env[65726]: DEBUG nova.network.neutron [req-b0ae93f4-f230-4542-aaac-0ce403689f36 req-530b0252-afb1-4ea1-9240-a7b795baeea4 service nova] [instance: d8801049-fc34-4b2f-8d5e-b9459c3e03a6] Updated VIF entry in instance network info cache for port b333a678-e6d6-494e-bc4d-7e498e370067. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 1085.100425] env[65726]: DEBUG nova.network.neutron [req-b0ae93f4-f230-4542-aaac-0ce403689f36 req-530b0252-afb1-4ea1-9240-a7b795baeea4 service nova] [instance: d8801049-fc34-4b2f-8d5e-b9459c3e03a6] Updating instance_info_cache with network_info: [{"id": "b333a678-e6d6-494e-bc4d-7e498e370067", "address": "fa:16:3e:52:4a:a1", "network": {"id": "4c972d9d-db3b-4779-ac39-39d9a2d141dd", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1418913861-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cbb17a09b35c4c22ade5c4082b10bda9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "50cf0a70-948d-4611-af05-94c1483064ed", "external-id": "nsx-vlan-transportzone-536", "segmentation_id": 536, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb333a678-e6", "ovs_interfaceid": "b333a678-e6d6-494e-bc4d-7e498e370067", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1085.208361] env[65726]: DEBUG oslo_vmware.api [None req-4f5aff8f-c14b-4955-a84c-9d59624f9d81 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': task-5116687, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1085.291679] env[65726]: INFO nova.compute.manager [None req-8c26e8da-5d87-42a2-a8c5-15e9c65a8b9f tempest-InstanceActionsV221TestJSON-647428531 tempest-InstanceActionsV221TestJSON-647428531-project-member] [instance: b0fec5ff-8c1e-41f9-b30a-6253a34baeec] Took 19.52 seconds to build instance. [ 1085.311593] env[65726]: DEBUG nova.scheduler.client.report [None req-f263eae6-96b2-45f2-b03a-d97fe5d0f1d8 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1085.469018] env[65726]: DEBUG oslo_vmware.api [None req-dd8aa948-a330-4efc-a6ba-843d94f1837a tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Task: {'id': task-5116688, 'name': ReconfigVM_Task, 'duration_secs': 0.182656} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1085.469099] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-dd8aa948-a330-4efc-a6ba-843d94f1837a tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] [instance: aa6ce489-c62f-4481-87b7-e74242aeb8ca] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-995264', 'volume_id': '3547dfa9-73e6-4baa-b777-e22891db153f', 'name': 'volume-3547dfa9-73e6-4baa-b777-e22891db153f', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'aa6ce489-c62f-4481-87b7-e74242aeb8ca', 'attached_at': '', 'detached_at': '', 'volume_id': '3547dfa9-73e6-4baa-b777-e22891db153f', 'serial': '3547dfa9-73e6-4baa-b777-e22891db153f'} {{(pid=65726) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1085.476429] env[65726]: DEBUG oslo_vmware.api [None req-a4ccf9b3-2389-45a5-bcf4-8de43acbac4c tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Task: {'id': task-5116689, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1085.522034] env[65726]: DEBUG oslo_vmware.api [None req-9ca95ca0-14c4-4870-aa74-cd2a795fbbc9 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': task-5116685, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.858904} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1085.522034] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-9ca95ca0-14c4-4870-aa74-cd2a795fbbc9 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] d8801049-fc34-4b2f-8d5e-b9459c3e03a6/d8801049-fc34-4b2f-8d5e-b9459c3e03a6.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1085.522034] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-9ca95ca0-14c4-4870-aa74-cd2a795fbbc9 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: d8801049-fc34-4b2f-8d5e-b9459c3e03a6] Extending root virtual disk to 1048576 {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1085.522034] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-791f0b0b-c0ce-4dc1-b35d-d7e4dd4bed44 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.525794] env[65726]: DEBUG oslo_concurrency.lockutils [req-708292a0-fd27-4a3f-8519-16ad5ba287b3 req-6f58ab55-2aa3-4257-b0cf-7b2339f092a6 service nova] Releasing lock "refresh_cache-6fd70b4a-63e4-4258-9cff-f3b582500b9f" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1085.525881] env[65726]: DEBUG nova.compute.manager [req-708292a0-fd27-4a3f-8519-16ad5ba287b3 req-6f58ab55-2aa3-4257-b0cf-7b2339f092a6 service nova] [instance: 51f5337d-2e23-4b7c-b590-76be6bd82411] Received event network-changed-7407bcd9-3202-4798-9eaa-252fb3fdb21f {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1085.526074] env[65726]: DEBUG nova.compute.manager [req-708292a0-fd27-4a3f-8519-16ad5ba287b3 req-6f58ab55-2aa3-4257-b0cf-7b2339f092a6 service nova] [instance: 51f5337d-2e23-4b7c-b590-76be6bd82411] Refreshing instance network info cache due to event network-changed-7407bcd9-3202-4798-9eaa-252fb3fdb21f. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 1085.526334] env[65726]: DEBUG oslo_concurrency.lockutils [req-708292a0-fd27-4a3f-8519-16ad5ba287b3 req-6f58ab55-2aa3-4257-b0cf-7b2339f092a6 service nova] Acquiring lock "refresh_cache-51f5337d-2e23-4b7c-b590-76be6bd82411" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1085.526959] env[65726]: DEBUG oslo_concurrency.lockutils [req-708292a0-fd27-4a3f-8519-16ad5ba287b3 req-6f58ab55-2aa3-4257-b0cf-7b2339f092a6 service nova] Acquired lock "refresh_cache-51f5337d-2e23-4b7c-b590-76be6bd82411" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1085.526959] env[65726]: DEBUG nova.network.neutron [req-708292a0-fd27-4a3f-8519-16ad5ba287b3 req-6f58ab55-2aa3-4257-b0cf-7b2339f092a6 service nova] [instance: 51f5337d-2e23-4b7c-b590-76be6bd82411] Refreshing network info cache for port 7407bcd9-3202-4798-9eaa-252fb3fdb21f {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 1085.529455] env[65726]: DEBUG oslo_vmware.api [None req-9ca95ca0-14c4-4870-aa74-cd2a795fbbc9 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Waiting for the task: (returnval){ [ 1085.529455] env[65726]: value = "task-5116690" [ 1085.529455] env[65726]: _type = "Task" [ 1085.529455] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1085.544389] env[65726]: DEBUG oslo_vmware.api [None req-9ca95ca0-14c4-4870-aa74-cd2a795fbbc9 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': task-5116690, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1085.583502] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5101d24a-5b00-4587-8b37-5a45f8c4ea31 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1085.603395] env[65726]: DEBUG oslo_concurrency.lockutils [req-b0ae93f4-f230-4542-aaac-0ce403689f36 req-530b0252-afb1-4ea1-9240-a7b795baeea4 service nova] Releasing lock "refresh_cache-d8801049-fc34-4b2f-8d5e-b9459c3e03a6" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1085.707378] env[65726]: DEBUG oslo_vmware.api [None req-4f5aff8f-c14b-4955-a84c-9d59624f9d81 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': task-5116687, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.603688} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1085.707609] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-4f5aff8f-c14b-4955-a84c-9d59624f9d81 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Deleted the datastore file {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1085.707797] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-4f5aff8f-c14b-4955-a84c-9d59624f9d81 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 5cdf1d61-b817-4986-a48c-804a2374497a] Deleted contents of the VM from datastore datastore2 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1085.707966] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-4f5aff8f-c14b-4955-a84c-9d59624f9d81 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 5cdf1d61-b817-4986-a48c-804a2374497a] Instance destroyed {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1085.708365] env[65726]: INFO nova.compute.manager [None req-4f5aff8f-c14b-4955-a84c-9d59624f9d81 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 5cdf1d61-b817-4986-a48c-804a2374497a] Took 1.66 seconds to destroy the instance on the hypervisor. [ 1085.708452] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-4f5aff8f-c14b-4955-a84c-9d59624f9d81 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1085.708651] env[65726]: DEBUG nova.compute.manager [-] [instance: 5cdf1d61-b817-4986-a48c-804a2374497a] Deallocating network for instance {{(pid=65726) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1085.708751] env[65726]: DEBUG nova.network.neutron [-] [instance: 5cdf1d61-b817-4986-a48c-804a2374497a] deallocate_for_instance() {{(pid=65726) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 1085.709324] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1085.709605] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1085.749440] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1085.749750] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1085.794432] env[65726]: DEBUG oslo_concurrency.lockutils [None req-8c26e8da-5d87-42a2-a8c5-15e9c65a8b9f tempest-InstanceActionsV221TestJSON-647428531 tempest-InstanceActionsV221TestJSON-647428531-project-member] Lock "b0fec5ff-8c1e-41f9-b30a-6253a34baeec" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 21.030s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1085.817231] env[65726]: DEBUG oslo_concurrency.lockutils [None req-f263eae6-96b2-45f2-b03a-d97fe5d0f1d8 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.441s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1085.819948] env[65726]: DEBUG oslo_concurrency.lockutils [None req-14a9b6f0-9c85-4271-a1df-1d81c0fc4a98 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: waited 1.385s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1085.931606] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a6f4e31f-bf92-46ff-ad83-0dbcdb2a78cf tempest-InstanceActionsV221TestJSON-647428531 tempest-InstanceActionsV221TestJSON-647428531-project-member] Acquiring lock "b0fec5ff-8c1e-41f9-b30a-6253a34baeec" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1085.931877] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a6f4e31f-bf92-46ff-ad83-0dbcdb2a78cf tempest-InstanceActionsV221TestJSON-647428531 tempest-InstanceActionsV221TestJSON-647428531-project-member] Lock "b0fec5ff-8c1e-41f9-b30a-6253a34baeec" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1085.932101] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a6f4e31f-bf92-46ff-ad83-0dbcdb2a78cf tempest-InstanceActionsV221TestJSON-647428531 tempest-InstanceActionsV221TestJSON-647428531-project-member] Acquiring lock "b0fec5ff-8c1e-41f9-b30a-6253a34baeec-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1085.932288] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a6f4e31f-bf92-46ff-ad83-0dbcdb2a78cf tempest-InstanceActionsV221TestJSON-647428531 tempest-InstanceActionsV221TestJSON-647428531-project-member] Lock "b0fec5ff-8c1e-41f9-b30a-6253a34baeec-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1085.932457] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a6f4e31f-bf92-46ff-ad83-0dbcdb2a78cf tempest-InstanceActionsV221TestJSON-647428531 tempest-InstanceActionsV221TestJSON-647428531-project-member] Lock "b0fec5ff-8c1e-41f9-b30a-6253a34baeec-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1085.934584] env[65726]: INFO nova.compute.manager [None req-a6f4e31f-bf92-46ff-ad83-0dbcdb2a78cf tempest-InstanceActionsV221TestJSON-647428531 tempest-InstanceActionsV221TestJSON-647428531-project-member] [instance: b0fec5ff-8c1e-41f9-b30a-6253a34baeec] Terminating instance [ 1085.972318] env[65726]: DEBUG oslo_vmware.api [None req-a4ccf9b3-2389-45a5-bcf4-8de43acbac4c tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Task: {'id': task-5116689, 'name': PowerOnVM_Task, 'duration_secs': 0.603731} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1085.972781] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-a4ccf9b3-2389-45a5-bcf4-8de43acbac4c tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] [instance: 3b4d9461-747a-461c-b231-a0de02f0cb83] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1085.977053] env[65726]: DEBUG nova.compute.manager [None req-a4ccf9b3-2389-45a5-bcf4-8de43acbac4c tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] [instance: 3b4d9461-747a-461c-b231-a0de02f0cb83] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1085.978502] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75951598-e45b-4f0f-9583-ead40da04f66 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.020556] env[65726]: DEBUG nova.objects.instance [None req-dd8aa948-a330-4efc-a6ba-843d94f1837a tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Lazy-loading 'flavor' on Instance uuid aa6ce489-c62f-4481-87b7-e74242aeb8ca {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1086.031544] env[65726]: WARNING openstack [req-708292a0-fd27-4a3f-8519-16ad5ba287b3 req-6f58ab55-2aa3-4257-b0cf-7b2339f092a6 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1086.032816] env[65726]: WARNING openstack [req-708292a0-fd27-4a3f-8519-16ad5ba287b3 req-6f58ab55-2aa3-4257-b0cf-7b2339f092a6 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1086.049402] env[65726]: DEBUG oslo_vmware.api [None req-9ca95ca0-14c4-4870-aa74-cd2a795fbbc9 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': task-5116690, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.074542} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1086.049661] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-9ca95ca0-14c4-4870-aa74-cd2a795fbbc9 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: d8801049-fc34-4b2f-8d5e-b9459c3e03a6] Extended root virtual disk {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1086.050515] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f14e2509-be8a-4c41-b036-2f740d7aff54 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.073642] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-9ca95ca0-14c4-4870-aa74-cd2a795fbbc9 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: d8801049-fc34-4b2f-8d5e-b9459c3e03a6] Reconfiguring VM instance instance-00000063 to attach disk [datastore1] d8801049-fc34-4b2f-8d5e-b9459c3e03a6/d8801049-fc34-4b2f-8d5e-b9459c3e03a6.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1086.077964] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f9a7661b-9ef6-4b4a-b482-a11b6d8b2357 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.099863] env[65726]: DEBUG oslo_vmware.api [None req-9ca95ca0-14c4-4870-aa74-cd2a795fbbc9 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Waiting for the task: (returnval){ [ 1086.099863] env[65726]: value = "task-5116691" [ 1086.099863] env[65726]: _type = "Task" [ 1086.099863] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1086.108885] env[65726]: DEBUG oslo_vmware.api [None req-9ca95ca0-14c4-4870-aa74-cd2a795fbbc9 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': task-5116691, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1086.140169] env[65726]: DEBUG nova.compute.manager [req-9668386f-7100-4ee1-84f3-695c7b14a3e0 req-8b18a951-fd66-4737-bd92-c99e88c7f6d4 service nova] [instance: 0d42fa75-20e9-4646-9b08-17015b7f068c] Received event network-changed-7a95cb45-ebfc-4ac0-ad58-4bb57170e47a {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1086.140312] env[65726]: DEBUG nova.compute.manager [req-9668386f-7100-4ee1-84f3-695c7b14a3e0 req-8b18a951-fd66-4737-bd92-c99e88c7f6d4 service nova] [instance: 0d42fa75-20e9-4646-9b08-17015b7f068c] Refreshing instance network info cache due to event network-changed-7a95cb45-ebfc-4ac0-ad58-4bb57170e47a. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 1086.140665] env[65726]: DEBUG oslo_concurrency.lockutils [req-9668386f-7100-4ee1-84f3-695c7b14a3e0 req-8b18a951-fd66-4737-bd92-c99e88c7f6d4 service nova] Acquiring lock "refresh_cache-0d42fa75-20e9-4646-9b08-17015b7f068c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1086.140855] env[65726]: DEBUG oslo_concurrency.lockutils [req-9668386f-7100-4ee1-84f3-695c7b14a3e0 req-8b18a951-fd66-4737-bd92-c99e88c7f6d4 service nova] Acquired lock "refresh_cache-0d42fa75-20e9-4646-9b08-17015b7f068c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1086.141034] env[65726]: DEBUG nova.network.neutron [req-9668386f-7100-4ee1-84f3-695c7b14a3e0 req-8b18a951-fd66-4737-bd92-c99e88c7f6d4 service nova] [instance: 0d42fa75-20e9-4646-9b08-17015b7f068c] Refreshing network info cache for port 7a95cb45-ebfc-4ac0-ad58-4bb57170e47a {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 1086.172834] env[65726]: WARNING openstack [req-708292a0-fd27-4a3f-8519-16ad5ba287b3 req-6f58ab55-2aa3-4257-b0cf-7b2339f092a6 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1086.173261] env[65726]: WARNING openstack [req-708292a0-fd27-4a3f-8519-16ad5ba287b3 req-6f58ab55-2aa3-4257-b0cf-7b2339f092a6 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1086.240368] env[65726]: WARNING openstack [req-708292a0-fd27-4a3f-8519-16ad5ba287b3 req-6f58ab55-2aa3-4257-b0cf-7b2339f092a6 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1086.240945] env[65726]: WARNING openstack [req-708292a0-fd27-4a3f-8519-16ad5ba287b3 req-6f58ab55-2aa3-4257-b0cf-7b2339f092a6 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1086.326640] env[65726]: DEBUG nova.objects.instance [None req-14a9b6f0-9c85-4271-a1df-1d81c0fc4a98 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Lazy-loading 'migration_context' on Instance uuid a56be6fa-6027-46ee-9aa8-332fe4222d68 {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1086.333038] env[65726]: DEBUG oslo_concurrency.lockutils [None req-f263eae6-96b2-45f2-b03a-d97fe5d0f1d8 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Lock "6fd70b4a-63e4-4258-9cff-f3b582500b9f" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 26.346s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1086.333269] env[65726]: DEBUG oslo_concurrency.lockutils [None req-1c19f444-aee2-49ca-b67e-d3cef83d3664 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Lock "6fd70b4a-63e4-4258-9cff-f3b582500b9f" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 1.375s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1086.334603] env[65726]: INFO nova.compute.manager [None req-1c19f444-aee2-49ca-b67e-d3cef83d3664 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] [instance: 6fd70b4a-63e4-4258-9cff-f3b582500b9f] Unshelving [ 1086.343839] env[65726]: DEBUG nova.network.neutron [req-708292a0-fd27-4a3f-8519-16ad5ba287b3 req-6f58ab55-2aa3-4257-b0cf-7b2339f092a6 service nova] [instance: 51f5337d-2e23-4b7c-b590-76be6bd82411] Updated VIF entry in instance network info cache for port 7407bcd9-3202-4798-9eaa-252fb3fdb21f. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 1086.344222] env[65726]: DEBUG nova.network.neutron [req-708292a0-fd27-4a3f-8519-16ad5ba287b3 req-6f58ab55-2aa3-4257-b0cf-7b2339f092a6 service nova] [instance: 51f5337d-2e23-4b7c-b590-76be6bd82411] Updating instance_info_cache with network_info: [{"id": "7407bcd9-3202-4798-9eaa-252fb3fdb21f", "address": "fa:16:3e:79:d2:9d", "network": {"id": "51a66448-0f72-4266-8868-856b48be839f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-155658961-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.213", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d46ccb00794f458b85da4a93879139ee", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b1f3e6c3-5584-4852-9017-476ab8ac4946", "external-id": "nsx-vlan-transportzone-304", "segmentation_id": 304, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7407bcd9-32", "ovs_interfaceid": "7407bcd9-3202-4798-9eaa-252fb3fdb21f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1086.439551] env[65726]: DEBUG nova.compute.manager [None req-a6f4e31f-bf92-46ff-ad83-0dbcdb2a78cf tempest-InstanceActionsV221TestJSON-647428531 tempest-InstanceActionsV221TestJSON-647428531-project-member] [instance: b0fec5ff-8c1e-41f9-b30a-6253a34baeec] Start destroying the instance on the hypervisor. {{(pid=65726) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1086.439885] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-a6f4e31f-bf92-46ff-ad83-0dbcdb2a78cf tempest-InstanceActionsV221TestJSON-647428531 tempest-InstanceActionsV221TestJSON-647428531-project-member] [instance: b0fec5ff-8c1e-41f9-b30a-6253a34baeec] Destroying instance {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1086.441202] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ee51ab4-3f72-421d-9614-0681d1b0dc90 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.452678] env[65726]: DEBUG nova.network.neutron [-] [instance: 5cdf1d61-b817-4986-a48c-804a2374497a] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1086.454091] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-a6f4e31f-bf92-46ff-ad83-0dbcdb2a78cf tempest-InstanceActionsV221TestJSON-647428531 tempest-InstanceActionsV221TestJSON-647428531-project-member] [instance: b0fec5ff-8c1e-41f9-b30a-6253a34baeec] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1086.454405] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-da5a2299-dd3e-4895-933c-ff734a1bf85f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.463295] env[65726]: DEBUG oslo_vmware.api [None req-a6f4e31f-bf92-46ff-ad83-0dbcdb2a78cf tempest-InstanceActionsV221TestJSON-647428531 tempest-InstanceActionsV221TestJSON-647428531-project-member] Waiting for the task: (returnval){ [ 1086.463295] env[65726]: value = "task-5116692" [ 1086.463295] env[65726]: _type = "Task" [ 1086.463295] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1086.477088] env[65726]: DEBUG oslo_vmware.api [None req-a6f4e31f-bf92-46ff-ad83-0dbcdb2a78cf tempest-InstanceActionsV221TestJSON-647428531 tempest-InstanceActionsV221TestJSON-647428531-project-member] Task: {'id': task-5116692, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1086.610886] env[65726]: DEBUG oslo_vmware.api [None req-9ca95ca0-14c4-4870-aa74-cd2a795fbbc9 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': task-5116691, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1086.645566] env[65726]: WARNING openstack [req-9668386f-7100-4ee1-84f3-695c7b14a3e0 req-8b18a951-fd66-4737-bd92-c99e88c7f6d4 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1086.645970] env[65726]: WARNING openstack [req-9668386f-7100-4ee1-84f3-695c7b14a3e0 req-8b18a951-fd66-4737-bd92-c99e88c7f6d4 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1086.781923] env[65726]: WARNING openstack [req-9668386f-7100-4ee1-84f3-695c7b14a3e0 req-8b18a951-fd66-4737-bd92-c99e88c7f6d4 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1086.782317] env[65726]: WARNING openstack [req-9668386f-7100-4ee1-84f3-695c7b14a3e0 req-8b18a951-fd66-4737-bd92-c99e88c7f6d4 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1086.852420] env[65726]: DEBUG oslo_concurrency.lockutils [req-708292a0-fd27-4a3f-8519-16ad5ba287b3 req-6f58ab55-2aa3-4257-b0cf-7b2339f092a6 service nova] Releasing lock "refresh_cache-51f5337d-2e23-4b7c-b590-76be6bd82411" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1086.852420] env[65726]: DEBUG nova.compute.manager [req-708292a0-fd27-4a3f-8519-16ad5ba287b3 req-6f58ab55-2aa3-4257-b0cf-7b2339f092a6 service nova] [instance: 0d42fa75-20e9-4646-9b08-17015b7f068c] Received event network-changed-7a95cb45-ebfc-4ac0-ad58-4bb57170e47a {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1086.852420] env[65726]: DEBUG nova.compute.manager [req-708292a0-fd27-4a3f-8519-16ad5ba287b3 req-6f58ab55-2aa3-4257-b0cf-7b2339f092a6 service nova] [instance: 0d42fa75-20e9-4646-9b08-17015b7f068c] Refreshing instance network info cache due to event network-changed-7a95cb45-ebfc-4ac0-ad58-4bb57170e47a. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 1086.852420] env[65726]: DEBUG oslo_concurrency.lockutils [req-708292a0-fd27-4a3f-8519-16ad5ba287b3 req-6f58ab55-2aa3-4257-b0cf-7b2339f092a6 service nova] Acquiring lock "refresh_cache-0d42fa75-20e9-4646-9b08-17015b7f068c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1086.903365] env[65726]: WARNING openstack [req-9668386f-7100-4ee1-84f3-695c7b14a3e0 req-8b18a951-fd66-4737-bd92-c99e88c7f6d4 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1086.903774] env[65726]: WARNING openstack [req-9668386f-7100-4ee1-84f3-695c7b14a3e0 req-8b18a951-fd66-4737-bd92-c99e88c7f6d4 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1086.966846] env[65726]: INFO nova.compute.manager [-] [instance: 5cdf1d61-b817-4986-a48c-804a2374497a] Took 1.26 seconds to deallocate network for instance. [ 1086.985108] env[65726]: DEBUG oslo_vmware.api [None req-a6f4e31f-bf92-46ff-ad83-0dbcdb2a78cf tempest-InstanceActionsV221TestJSON-647428531 tempest-InstanceActionsV221TestJSON-647428531-project-member] Task: {'id': task-5116692, 'name': PowerOffVM_Task, 'duration_secs': 0.394567} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1086.985432] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-a6f4e31f-bf92-46ff-ad83-0dbcdb2a78cf tempest-InstanceActionsV221TestJSON-647428531 tempest-InstanceActionsV221TestJSON-647428531-project-member] [instance: b0fec5ff-8c1e-41f9-b30a-6253a34baeec] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1086.985596] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-a6f4e31f-bf92-46ff-ad83-0dbcdb2a78cf tempest-InstanceActionsV221TestJSON-647428531 tempest-InstanceActionsV221TestJSON-647428531-project-member] [instance: b0fec5ff-8c1e-41f9-b30a-6253a34baeec] Unregistering the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1086.985863] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-384f7712-e3e8-4ce5-93b7-c672c1677821 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.005744] env[65726]: DEBUG nova.network.neutron [req-9668386f-7100-4ee1-84f3-695c7b14a3e0 req-8b18a951-fd66-4737-bd92-c99e88c7f6d4 service nova] [instance: 0d42fa75-20e9-4646-9b08-17015b7f068c] Updated VIF entry in instance network info cache for port 7a95cb45-ebfc-4ac0-ad58-4bb57170e47a. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 1087.006319] env[65726]: DEBUG nova.network.neutron [req-9668386f-7100-4ee1-84f3-695c7b14a3e0 req-8b18a951-fd66-4737-bd92-c99e88c7f6d4 service nova] [instance: 0d42fa75-20e9-4646-9b08-17015b7f068c] Updating instance_info_cache with network_info: [{"id": "7a95cb45-ebfc-4ac0-ad58-4bb57170e47a", "address": "fa:16:3e:2b:ea:81", "network": {"id": "51a66448-0f72-4266-8868-856b48be839f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-155658961-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d46ccb00794f458b85da4a93879139ee", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b1f3e6c3-5584-4852-9017-476ab8ac4946", "external-id": "nsx-vlan-transportzone-304", "segmentation_id": 304, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7a95cb45-eb", "ovs_interfaceid": "7a95cb45-ebfc-4ac0-ad58-4bb57170e47a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1087.011083] env[65726]: DEBUG oslo_concurrency.lockutils [None req-33cd1beb-aa3c-4bd3-a35f-a33f8bcc8760 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Acquiring lock "aa6ce489-c62f-4481-87b7-e74242aeb8ca" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1087.033892] env[65726]: DEBUG oslo_concurrency.lockutils [None req-dd8aa948-a330-4efc-a6ba-843d94f1837a tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Lock "aa6ce489-c62f-4481-87b7-e74242aeb8ca" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.358s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1087.036809] env[65726]: DEBUG oslo_concurrency.lockutils [None req-33cd1beb-aa3c-4bd3-a35f-a33f8bcc8760 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Lock "aa6ce489-c62f-4481-87b7-e74242aeb8ca" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.025s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1087.037357] env[65726]: DEBUG nova.compute.manager [None req-33cd1beb-aa3c-4bd3-a35f-a33f8bcc8760 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] [instance: aa6ce489-c62f-4481-87b7-e74242aeb8ca] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1087.038730] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2bb3200e-e3c7-4dfd-b5b8-d0a837ca0b6f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.048516] env[65726]: DEBUG nova.compute.manager [None req-33cd1beb-aa3c-4bd3-a35f-a33f8bcc8760 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] [instance: aa6ce489-c62f-4481-87b7-e74242aeb8ca] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=65726) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3459}} [ 1087.049150] env[65726]: DEBUG nova.objects.instance [None req-33cd1beb-aa3c-4bd3-a35f-a33f8bcc8760 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Lazy-loading 'flavor' on Instance uuid aa6ce489-c62f-4481-87b7-e74242aeb8ca {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1087.064475] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-a6f4e31f-bf92-46ff-ad83-0dbcdb2a78cf tempest-InstanceActionsV221TestJSON-647428531 tempest-InstanceActionsV221TestJSON-647428531-project-member] [instance: b0fec5ff-8c1e-41f9-b30a-6253a34baeec] Unregistered the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1087.064701] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-a6f4e31f-bf92-46ff-ad83-0dbcdb2a78cf tempest-InstanceActionsV221TestJSON-647428531 tempest-InstanceActionsV221TestJSON-647428531-project-member] [instance: b0fec5ff-8c1e-41f9-b30a-6253a34baeec] Deleting contents of the VM from datastore datastore2 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1087.064833] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-a6f4e31f-bf92-46ff-ad83-0dbcdb2a78cf tempest-InstanceActionsV221TestJSON-647428531 tempest-InstanceActionsV221TestJSON-647428531-project-member] Deleting the datastore file [datastore2] b0fec5ff-8c1e-41f9-b30a-6253a34baeec {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1087.067717] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-06f0af94-1ea0-4528-8c43-c4d58bcb2d19 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.078347] env[65726]: DEBUG oslo_vmware.api [None req-a6f4e31f-bf92-46ff-ad83-0dbcdb2a78cf tempest-InstanceActionsV221TestJSON-647428531 tempest-InstanceActionsV221TestJSON-647428531-project-member] Waiting for the task: (returnval){ [ 1087.078347] env[65726]: value = "task-5116694" [ 1087.078347] env[65726]: _type = "Task" [ 1087.078347] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1087.090373] env[65726]: DEBUG oslo_vmware.api [None req-a6f4e31f-bf92-46ff-ad83-0dbcdb2a78cf tempest-InstanceActionsV221TestJSON-647428531 tempest-InstanceActionsV221TestJSON-647428531-project-member] Task: {'id': task-5116694, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1087.100019] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1dff705-9dff-48ea-b140-45d187d69802 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.113149] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-641722d7-63ae-4b5a-a2d1-f49c877181d0 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.117088] env[65726]: DEBUG oslo_vmware.api [None req-9ca95ca0-14c4-4870-aa74-cd2a795fbbc9 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': task-5116691, 'name': ReconfigVM_Task, 'duration_secs': 0.787802} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1087.117478] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-9ca95ca0-14c4-4870-aa74-cd2a795fbbc9 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: d8801049-fc34-4b2f-8d5e-b9459c3e03a6] Reconfigured VM instance instance-00000063 to attach disk [datastore1] d8801049-fc34-4b2f-8d5e-b9459c3e03a6/d8801049-fc34-4b2f-8d5e-b9459c3e03a6.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1087.118608] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-16057ba2-72bf-4567-9f06-bf0d7464b998 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.152606] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5edb0374-7d4b-4806-88c3-7f274d6bb2f9 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.156775] env[65726]: DEBUG oslo_vmware.api [None req-9ca95ca0-14c4-4870-aa74-cd2a795fbbc9 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Waiting for the task: (returnval){ [ 1087.156775] env[65726]: value = "task-5116695" [ 1087.156775] env[65726]: _type = "Task" [ 1087.156775] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1087.168567] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1097c72e-2b5a-4d08-a271-6b9226246dff {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.180652] env[65726]: DEBUG oslo_vmware.api [None req-9ca95ca0-14c4-4870-aa74-cd2a795fbbc9 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': task-5116695, 'name': Rename_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1087.196675] env[65726]: DEBUG nova.compute.provider_tree [None req-14a9b6f0-9c85-4271-a1df-1d81c0fc4a98 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1087.364961] env[65726]: DEBUG oslo_concurrency.lockutils [None req-1c19f444-aee2-49ca-b67e-d3cef83d3664 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1087.481418] env[65726]: DEBUG oslo_concurrency.lockutils [None req-4f5aff8f-c14b-4955-a84c-9d59624f9d81 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1087.513425] env[65726]: DEBUG oslo_concurrency.lockutils [req-9668386f-7100-4ee1-84f3-695c7b14a3e0 req-8b18a951-fd66-4737-bd92-c99e88c7f6d4 service nova] Releasing lock "refresh_cache-0d42fa75-20e9-4646-9b08-17015b7f068c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1087.513789] env[65726]: DEBUG nova.compute.manager [req-9668386f-7100-4ee1-84f3-695c7b14a3e0 req-8b18a951-fd66-4737-bd92-c99e88c7f6d4 service nova] [instance: 51f5337d-2e23-4b7c-b590-76be6bd82411] Received event network-changed-7407bcd9-3202-4798-9eaa-252fb3fdb21f {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1087.513997] env[65726]: DEBUG nova.compute.manager [req-9668386f-7100-4ee1-84f3-695c7b14a3e0 req-8b18a951-fd66-4737-bd92-c99e88c7f6d4 service nova] [instance: 51f5337d-2e23-4b7c-b590-76be6bd82411] Refreshing instance network info cache due to event network-changed-7407bcd9-3202-4798-9eaa-252fb3fdb21f. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 1087.514300] env[65726]: DEBUG oslo_concurrency.lockutils [req-9668386f-7100-4ee1-84f3-695c7b14a3e0 req-8b18a951-fd66-4737-bd92-c99e88c7f6d4 service nova] Acquiring lock "refresh_cache-51f5337d-2e23-4b7c-b590-76be6bd82411" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1087.514590] env[65726]: DEBUG oslo_concurrency.lockutils [req-9668386f-7100-4ee1-84f3-695c7b14a3e0 req-8b18a951-fd66-4737-bd92-c99e88c7f6d4 service nova] Acquired lock "refresh_cache-51f5337d-2e23-4b7c-b590-76be6bd82411" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1087.514679] env[65726]: DEBUG nova.network.neutron [req-9668386f-7100-4ee1-84f3-695c7b14a3e0 req-8b18a951-fd66-4737-bd92-c99e88c7f6d4 service nova] [instance: 51f5337d-2e23-4b7c-b590-76be6bd82411] Refreshing network info cache for port 7407bcd9-3202-4798-9eaa-252fb3fdb21f {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 1087.516006] env[65726]: DEBUG oslo_concurrency.lockutils [req-708292a0-fd27-4a3f-8519-16ad5ba287b3 req-6f58ab55-2aa3-4257-b0cf-7b2339f092a6 service nova] Acquired lock "refresh_cache-0d42fa75-20e9-4646-9b08-17015b7f068c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1087.516481] env[65726]: DEBUG nova.network.neutron [req-708292a0-fd27-4a3f-8519-16ad5ba287b3 req-6f58ab55-2aa3-4257-b0cf-7b2339f092a6 service nova] [instance: 0d42fa75-20e9-4646-9b08-17015b7f068c] Refreshing network info cache for port 7a95cb45-ebfc-4ac0-ad58-4bb57170e47a {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 1087.588931] env[65726]: DEBUG oslo_vmware.api [None req-a6f4e31f-bf92-46ff-ad83-0dbcdb2a78cf tempest-InstanceActionsV221TestJSON-647428531 tempest-InstanceActionsV221TestJSON-647428531-project-member] Task: {'id': task-5116694, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.279526} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1087.589214] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-a6f4e31f-bf92-46ff-ad83-0dbcdb2a78cf tempest-InstanceActionsV221TestJSON-647428531 tempest-InstanceActionsV221TestJSON-647428531-project-member] Deleted the datastore file {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1087.589394] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-a6f4e31f-bf92-46ff-ad83-0dbcdb2a78cf tempest-InstanceActionsV221TestJSON-647428531 tempest-InstanceActionsV221TestJSON-647428531-project-member] [instance: b0fec5ff-8c1e-41f9-b30a-6253a34baeec] Deleted contents of the VM from datastore datastore2 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1087.589565] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-a6f4e31f-bf92-46ff-ad83-0dbcdb2a78cf tempest-InstanceActionsV221TestJSON-647428531 tempest-InstanceActionsV221TestJSON-647428531-project-member] [instance: b0fec5ff-8c1e-41f9-b30a-6253a34baeec] Instance destroyed {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1087.589733] env[65726]: INFO nova.compute.manager [None req-a6f4e31f-bf92-46ff-ad83-0dbcdb2a78cf tempest-InstanceActionsV221TestJSON-647428531 tempest-InstanceActionsV221TestJSON-647428531-project-member] [instance: b0fec5ff-8c1e-41f9-b30a-6253a34baeec] Took 1.15 seconds to destroy the instance on the hypervisor. [ 1087.589984] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-a6f4e31f-bf92-46ff-ad83-0dbcdb2a78cf tempest-InstanceActionsV221TestJSON-647428531 tempest-InstanceActionsV221TestJSON-647428531-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1087.590204] env[65726]: DEBUG nova.compute.manager [-] [instance: b0fec5ff-8c1e-41f9-b30a-6253a34baeec] Deallocating network for instance {{(pid=65726) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1087.590302] env[65726]: DEBUG nova.network.neutron [-] [instance: b0fec5ff-8c1e-41f9-b30a-6253a34baeec] deallocate_for_instance() {{(pid=65726) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 1087.590855] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1087.591093] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1087.631058] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1087.631355] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1087.667638] env[65726]: DEBUG oslo_vmware.api [None req-9ca95ca0-14c4-4870-aa74-cd2a795fbbc9 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': task-5116695, 'name': Rename_Task, 'duration_secs': 0.170563} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1087.667918] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-9ca95ca0-14c4-4870-aa74-cd2a795fbbc9 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: d8801049-fc34-4b2f-8d5e-b9459c3e03a6] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1087.668187] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f66f5eaa-a396-4e1e-b785-275b602c0f35 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.675897] env[65726]: DEBUG oslo_vmware.api [None req-9ca95ca0-14c4-4870-aa74-cd2a795fbbc9 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Waiting for the task: (returnval){ [ 1087.675897] env[65726]: value = "task-5116696" [ 1087.675897] env[65726]: _type = "Task" [ 1087.675897] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1087.685046] env[65726]: DEBUG oslo_vmware.api [None req-9ca95ca0-14c4-4870-aa74-cd2a795fbbc9 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': task-5116696, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1087.700289] env[65726]: DEBUG nova.scheduler.client.report [None req-14a9b6f0-9c85-4271-a1df-1d81c0fc4a98 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1087.785720] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5eb7fa6b-cbd6-4a47-b227-e9c16f779f4a tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Acquiring lock "4809fcf6-59d4-409c-bd8d-981dc85a686f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1087.786049] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5eb7fa6b-cbd6-4a47-b227-e9c16f779f4a tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Lock "4809fcf6-59d4-409c-bd8d-981dc85a686f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1087.902400] env[65726]: DEBUG nova.compute.manager [req-7a7d0c4a-2252-4189-beae-327b6277ec7b req-0dce6a46-ae7e-4905-8283-e66961abd2b2 service nova] [instance: b0fec5ff-8c1e-41f9-b30a-6253a34baeec] Received event network-vif-deleted-f11556ce-cfc0-4593-ac46-59266a693222 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1087.902705] env[65726]: INFO nova.compute.manager [req-7a7d0c4a-2252-4189-beae-327b6277ec7b req-0dce6a46-ae7e-4905-8283-e66961abd2b2 service nova] [instance: b0fec5ff-8c1e-41f9-b30a-6253a34baeec] Neutron deleted interface f11556ce-cfc0-4593-ac46-59266a693222; detaching it from the instance and deleting it from the info cache [ 1087.902916] env[65726]: DEBUG nova.network.neutron [req-7a7d0c4a-2252-4189-beae-327b6277ec7b req-0dce6a46-ae7e-4905-8283-e66961abd2b2 service nova] [instance: b0fec5ff-8c1e-41f9-b30a-6253a34baeec] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1088.018774] env[65726]: WARNING openstack [req-9668386f-7100-4ee1-84f3-695c7b14a3e0 req-8b18a951-fd66-4737-bd92-c99e88c7f6d4 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1088.019207] env[65726]: WARNING openstack [req-9668386f-7100-4ee1-84f3-695c7b14a3e0 req-8b18a951-fd66-4737-bd92-c99e88c7f6d4 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1088.027038] env[65726]: WARNING openstack [req-708292a0-fd27-4a3f-8519-16ad5ba287b3 req-6f58ab55-2aa3-4257-b0cf-7b2339f092a6 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1088.027407] env[65726]: WARNING openstack [req-708292a0-fd27-4a3f-8519-16ad5ba287b3 req-6f58ab55-2aa3-4257-b0cf-7b2339f092a6 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1088.060911] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-33cd1beb-aa3c-4bd3-a35f-a33f8bcc8760 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] [instance: aa6ce489-c62f-4481-87b7-e74242aeb8ca] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1088.061591] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d7d37004-0986-47d8-8e5e-bfd8e4c3ebb6 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.071445] env[65726]: DEBUG oslo_vmware.api [None req-33cd1beb-aa3c-4bd3-a35f-a33f8bcc8760 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Waiting for the task: (returnval){ [ 1088.071445] env[65726]: value = "task-5116697" [ 1088.071445] env[65726]: _type = "Task" [ 1088.071445] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1088.085497] env[65726]: DEBUG oslo_vmware.api [None req-33cd1beb-aa3c-4bd3-a35f-a33f8bcc8760 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Task: {'id': task-5116697, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1088.161731] env[65726]: WARNING openstack [req-708292a0-fd27-4a3f-8519-16ad5ba287b3 req-6f58ab55-2aa3-4257-b0cf-7b2339f092a6 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1088.162158] env[65726]: WARNING openstack [req-708292a0-fd27-4a3f-8519-16ad5ba287b3 req-6f58ab55-2aa3-4257-b0cf-7b2339f092a6 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1088.188791] env[65726]: DEBUG oslo_vmware.api [None req-9ca95ca0-14c4-4870-aa74-cd2a795fbbc9 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': task-5116696, 'name': PowerOnVM_Task} progress is 88%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1088.192463] env[65726]: WARNING openstack [req-9668386f-7100-4ee1-84f3-695c7b14a3e0 req-8b18a951-fd66-4737-bd92-c99e88c7f6d4 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1088.192987] env[65726]: WARNING openstack [req-9668386f-7100-4ee1-84f3-695c7b14a3e0 req-8b18a951-fd66-4737-bd92-c99e88c7f6d4 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1088.238030] env[65726]: WARNING openstack [req-708292a0-fd27-4a3f-8519-16ad5ba287b3 req-6f58ab55-2aa3-4257-b0cf-7b2339f092a6 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1088.238030] env[65726]: WARNING openstack [req-708292a0-fd27-4a3f-8519-16ad5ba287b3 req-6f58ab55-2aa3-4257-b0cf-7b2339f092a6 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1088.275151] env[65726]: WARNING openstack [req-9668386f-7100-4ee1-84f3-695c7b14a3e0 req-8b18a951-fd66-4737-bd92-c99e88c7f6d4 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1088.275151] env[65726]: WARNING openstack [req-9668386f-7100-4ee1-84f3-695c7b14a3e0 req-8b18a951-fd66-4737-bd92-c99e88c7f6d4 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1088.294292] env[65726]: DEBUG nova.compute.manager [None req-5eb7fa6b-cbd6-4a47-b227-e9c16f779f4a tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] [instance: 4809fcf6-59d4-409c-bd8d-981dc85a686f] Starting instance... {{(pid=65726) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 1088.342213] env[65726]: DEBUG nova.network.neutron [req-708292a0-fd27-4a3f-8519-16ad5ba287b3 req-6f58ab55-2aa3-4257-b0cf-7b2339f092a6 service nova] [instance: 0d42fa75-20e9-4646-9b08-17015b7f068c] Updated VIF entry in instance network info cache for port 7a95cb45-ebfc-4ac0-ad58-4bb57170e47a. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 1088.342665] env[65726]: DEBUG nova.network.neutron [req-708292a0-fd27-4a3f-8519-16ad5ba287b3 req-6f58ab55-2aa3-4257-b0cf-7b2339f092a6 service nova] [instance: 0d42fa75-20e9-4646-9b08-17015b7f068c] Updating instance_info_cache with network_info: [{"id": "7a95cb45-ebfc-4ac0-ad58-4bb57170e47a", "address": "fa:16:3e:2b:ea:81", "network": {"id": "51a66448-0f72-4266-8868-856b48be839f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-155658961-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d46ccb00794f458b85da4a93879139ee", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b1f3e6c3-5584-4852-9017-476ab8ac4946", "external-id": "nsx-vlan-transportzone-304", "segmentation_id": 304, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7a95cb45-eb", "ovs_interfaceid": "7a95cb45-ebfc-4ac0-ad58-4bb57170e47a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1088.360394] env[65726]: DEBUG nova.network.neutron [-] [instance: b0fec5ff-8c1e-41f9-b30a-6253a34baeec] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1088.369345] env[65726]: DEBUG nova.network.neutron [req-9668386f-7100-4ee1-84f3-695c7b14a3e0 req-8b18a951-fd66-4737-bd92-c99e88c7f6d4 service nova] [instance: 51f5337d-2e23-4b7c-b590-76be6bd82411] Updated VIF entry in instance network info cache for port 7407bcd9-3202-4798-9eaa-252fb3fdb21f. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 1088.370095] env[65726]: DEBUG nova.network.neutron [req-9668386f-7100-4ee1-84f3-695c7b14a3e0 req-8b18a951-fd66-4737-bd92-c99e88c7f6d4 service nova] [instance: 51f5337d-2e23-4b7c-b590-76be6bd82411] Updating instance_info_cache with network_info: [{"id": "7407bcd9-3202-4798-9eaa-252fb3fdb21f", "address": "fa:16:3e:79:d2:9d", "network": {"id": "51a66448-0f72-4266-8868-856b48be839f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-155658961-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.213", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d46ccb00794f458b85da4a93879139ee", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b1f3e6c3-5584-4852-9017-476ab8ac4946", "external-id": "nsx-vlan-transportzone-304", "segmentation_id": 304, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7407bcd9-32", "ovs_interfaceid": "7407bcd9-3202-4798-9eaa-252fb3fdb21f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1088.405660] env[65726]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-020a4f2c-9e10-4081-86e7-47a87fb11c85 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.415733] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-222aad0e-1564-43cf-975a-982f4f912083 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.453890] env[65726]: DEBUG nova.compute.manager [req-7a7d0c4a-2252-4189-beae-327b6277ec7b req-0dce6a46-ae7e-4905-8283-e66961abd2b2 service nova] [instance: b0fec5ff-8c1e-41f9-b30a-6253a34baeec] Detach interface failed, port_id=f11556ce-cfc0-4593-ac46-59266a693222, reason: Instance b0fec5ff-8c1e-41f9-b30a-6253a34baeec could not be found. {{(pid=65726) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11669}} [ 1088.539040] env[65726]: DEBUG oslo_concurrency.lockutils [None req-6efd9d49-6747-4178-8149-554fa3ea5be4 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Acquiring lock "2fe6fc5a-f5c7-4f8e-96df-4e621a252f04" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1088.539330] env[65726]: DEBUG oslo_concurrency.lockutils [None req-6efd9d49-6747-4178-8149-554fa3ea5be4 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Lock "2fe6fc5a-f5c7-4f8e-96df-4e621a252f04" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1088.582993] env[65726]: DEBUG oslo_vmware.api [None req-33cd1beb-aa3c-4bd3-a35f-a33f8bcc8760 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Task: {'id': task-5116697, 'name': PowerOffVM_Task, 'duration_secs': 0.272622} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1088.583296] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-33cd1beb-aa3c-4bd3-a35f-a33f8bcc8760 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] [instance: aa6ce489-c62f-4481-87b7-e74242aeb8ca] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1088.583489] env[65726]: DEBUG nova.compute.manager [None req-33cd1beb-aa3c-4bd3-a35f-a33f8bcc8760 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] [instance: aa6ce489-c62f-4481-87b7-e74242aeb8ca] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1088.584330] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6951bce3-86d3-41ff-a504-0b19dd3dff1e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.686834] env[65726]: DEBUG oslo_vmware.api [None req-9ca95ca0-14c4-4870-aa74-cd2a795fbbc9 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': task-5116696, 'name': PowerOnVM_Task, 'duration_secs': 0.869622} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1088.687130] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-9ca95ca0-14c4-4870-aa74-cd2a795fbbc9 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: d8801049-fc34-4b2f-8d5e-b9459c3e03a6] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1088.687332] env[65726]: INFO nova.compute.manager [None req-9ca95ca0-14c4-4870-aa74-cd2a795fbbc9 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: d8801049-fc34-4b2f-8d5e-b9459c3e03a6] Took 7.29 seconds to spawn the instance on the hypervisor. [ 1088.687505] env[65726]: DEBUG nova.compute.manager [None req-9ca95ca0-14c4-4870-aa74-cd2a795fbbc9 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: d8801049-fc34-4b2f-8d5e-b9459c3e03a6] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1088.688593] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1d6ba25-0e26-41d9-8e8b-4ebacf12c718 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.712593] env[65726]: DEBUG oslo_concurrency.lockutils [None req-14a9b6f0-9c85-4271-a1df-1d81c0fc4a98 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: held 2.893s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1088.718389] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5101d24a-5b00-4587-8b37-5a45f8c4ea31 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.135s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1088.719807] env[65726]: INFO nova.compute.claims [None req-5101d24a-5b00-4587-8b37-5a45f8c4ea31 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 0361ae8f-634d-4161-a594-7f93dbb1b7cd] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1088.845850] env[65726]: DEBUG oslo_concurrency.lockutils [req-708292a0-fd27-4a3f-8519-16ad5ba287b3 req-6f58ab55-2aa3-4257-b0cf-7b2339f092a6 service nova] Releasing lock "refresh_cache-0d42fa75-20e9-4646-9b08-17015b7f068c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1088.853823] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5eb7fa6b-cbd6-4a47-b227-e9c16f779f4a tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1088.863028] env[65726]: INFO nova.compute.manager [-] [instance: b0fec5ff-8c1e-41f9-b30a-6253a34baeec] Took 1.27 seconds to deallocate network for instance. [ 1088.874019] env[65726]: DEBUG oslo_concurrency.lockutils [req-9668386f-7100-4ee1-84f3-695c7b14a3e0 req-8b18a951-fd66-4737-bd92-c99e88c7f6d4 service nova] Releasing lock "refresh_cache-51f5337d-2e23-4b7c-b590-76be6bd82411" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1088.874163] env[65726]: DEBUG nova.compute.manager [req-9668386f-7100-4ee1-84f3-695c7b14a3e0 req-8b18a951-fd66-4737-bd92-c99e88c7f6d4 service nova] [instance: 5cdf1d61-b817-4986-a48c-804a2374497a] Received event network-vif-deleted-94113d1c-2016-4a98-b42a-6ae90eee0b24 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1088.875127] env[65726]: INFO nova.compute.manager [req-9668386f-7100-4ee1-84f3-695c7b14a3e0 req-8b18a951-fd66-4737-bd92-c99e88c7f6d4 service nova] [instance: 5cdf1d61-b817-4986-a48c-804a2374497a] Neutron deleted interface 94113d1c-2016-4a98-b42a-6ae90eee0b24; detaching it from the instance and deleting it from the info cache [ 1088.875127] env[65726]: DEBUG nova.network.neutron [req-9668386f-7100-4ee1-84f3-695c7b14a3e0 req-8b18a951-fd66-4737-bd92-c99e88c7f6d4 service nova] [instance: 5cdf1d61-b817-4986-a48c-804a2374497a] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1089.043180] env[65726]: INFO nova.compute.manager [None req-6efd9d49-6747-4178-8149-554fa3ea5be4 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04] Detaching volume 98eb7918-3cfa-4296-9855-bce47b982be1 [ 1089.082905] env[65726]: INFO nova.virt.block_device [None req-6efd9d49-6747-4178-8149-554fa3ea5be4 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04] Attempting to driver detach volume 98eb7918-3cfa-4296-9855-bce47b982be1 from mountpoint /dev/sdb [ 1089.083241] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-6efd9d49-6747-4178-8149-554fa3ea5be4 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04] Volume detach. Driver type: vmdk {{(pid=65726) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1089.083447] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-6efd9d49-6747-4178-8149-554fa3ea5be4 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-995255', 'volume_id': '98eb7918-3cfa-4296-9855-bce47b982be1', 'name': 'volume-98eb7918-3cfa-4296-9855-bce47b982be1', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'attached', 'instance': '2fe6fc5a-f5c7-4f8e-96df-4e621a252f04', 'attached_at': '', 'detached_at': '', 'volume_id': '98eb7918-3cfa-4296-9855-bce47b982be1', 'serial': '98eb7918-3cfa-4296-9855-bce47b982be1'} {{(pid=65726) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1089.084511] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd8fb798-d9b9-4bd0-9d47-c5567c97b7d2 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.111238] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c3464ba-9283-4832-a2a1-79a28519619e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.114414] env[65726]: DEBUG oslo_concurrency.lockutils [None req-33cd1beb-aa3c-4bd3-a35f-a33f8bcc8760 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Lock "aa6ce489-c62f-4481-87b7-e74242aeb8ca" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.078s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1089.121282] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0bc10dcc-13b9-45b4-bf0e-8a92a926f1cd {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.145821] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a5d3c30-baf2-49f0-8f66-87361df33e7a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.167108] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-6efd9d49-6747-4178-8149-554fa3ea5be4 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] The volume has not been displaced from its original location: [datastore2] volume-98eb7918-3cfa-4296-9855-bce47b982be1/volume-98eb7918-3cfa-4296-9855-bce47b982be1.vmdk. No consolidation needed. {{(pid=65726) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1089.173236] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-6efd9d49-6747-4178-8149-554fa3ea5be4 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04] Reconfiguring VM instance instance-0000003f to detach disk 2001 {{(pid=65726) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1089.173996] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5d37c07c-8220-434f-8de1-6eb593e0a4d7 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.194331] env[65726]: DEBUG oslo_vmware.api [None req-6efd9d49-6747-4178-8149-554fa3ea5be4 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Waiting for the task: (returnval){ [ 1089.194331] env[65726]: value = "task-5116698" [ 1089.194331] env[65726]: _type = "Task" [ 1089.194331] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1089.206549] env[65726]: DEBUG oslo_vmware.api [None req-6efd9d49-6747-4178-8149-554fa3ea5be4 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Task: {'id': task-5116698, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1089.208880] env[65726]: INFO nova.compute.manager [None req-9ca95ca0-14c4-4870-aa74-cd2a795fbbc9 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: d8801049-fc34-4b2f-8d5e-b9459c3e03a6] Took 16.73 seconds to build instance. [ 1089.371837] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a6f4e31f-bf92-46ff-ad83-0dbcdb2a78cf tempest-InstanceActionsV221TestJSON-647428531 tempest-InstanceActionsV221TestJSON-647428531-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1089.377054] env[65726]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7480c890-8c76-428c-8175-4f31c7c07557 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.391066] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2ae44a7-e667-481e-8b10-b1f69244bce3 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.438053] env[65726]: DEBUG nova.compute.manager [req-9668386f-7100-4ee1-84f3-695c7b14a3e0 req-8b18a951-fd66-4737-bd92-c99e88c7f6d4 service nova] [instance: 5cdf1d61-b817-4986-a48c-804a2374497a] Detach interface failed, port_id=94113d1c-2016-4a98-b42a-6ae90eee0b24, reason: Instance 5cdf1d61-b817-4986-a48c-804a2374497a could not be found. {{(pid=65726) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11669}} [ 1089.513444] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-363013ba-5788-4fdc-a143-f7c5355f9a25 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.520916] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-0857f350-c753-4ac4-a5de-c6f94d195846 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: d8801049-fc34-4b2f-8d5e-b9459c3e03a6] Suspending the VM {{(pid=65726) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1162}} [ 1089.521158] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-05e77af8-02b3-4acc-bbf5-2b696dc827a2 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.528500] env[65726]: DEBUG oslo_vmware.api [None req-0857f350-c753-4ac4-a5de-c6f94d195846 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Waiting for the task: (returnval){ [ 1089.528500] env[65726]: value = "task-5116699" [ 1089.528500] env[65726]: _type = "Task" [ 1089.528500] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1089.542651] env[65726]: DEBUG oslo_vmware.api [None req-0857f350-c753-4ac4-a5de-c6f94d195846 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': task-5116699, 'name': SuspendVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1089.654478] env[65726]: DEBUG nova.objects.instance [None req-6a8181d2-ce0d-4923-b0a1-3e2a66a63f15 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Lazy-loading 'flavor' on Instance uuid aa6ce489-c62f-4481-87b7-e74242aeb8ca {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1089.706961] env[65726]: DEBUG oslo_vmware.api [None req-6efd9d49-6747-4178-8149-554fa3ea5be4 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Task: {'id': task-5116698, 'name': ReconfigVM_Task, 'duration_secs': 0.383378} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1089.707360] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-6efd9d49-6747-4178-8149-554fa3ea5be4 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04] Reconfigured VM instance instance-0000003f to detach disk 2001 {{(pid=65726) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1089.713188] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9ca95ca0-14c4-4870-aa74-cd2a795fbbc9 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Lock "d8801049-fc34-4b2f-8d5e-b9459c3e03a6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 18.257s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1089.713616] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5bf2fb93-718e-4e59-be75-ad0bb527f26f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.738753] env[65726]: DEBUG oslo_vmware.api [None req-6efd9d49-6747-4178-8149-554fa3ea5be4 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Waiting for the task: (returnval){ [ 1089.738753] env[65726]: value = "task-5116700" [ 1089.738753] env[65726]: _type = "Task" [ 1089.738753] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1089.753597] env[65726]: DEBUG oslo_vmware.api [None req-6efd9d49-6747-4178-8149-554fa3ea5be4 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Task: {'id': task-5116700, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1089.968453] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0107dc09-5f59-4156-a174-e0bc5b9070e7 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.977290] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bac0e922-895f-4900-a11a-88a13dd2a595 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.010118] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-998b11b5-6ffa-4e78-aa75-c32257b8cde5 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.018801] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c94ca9ae-cb2f-4e8f-a071-c385d0913e89 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.037973] env[65726]: DEBUG nova.compute.provider_tree [None req-5101d24a-5b00-4587-8b37-5a45f8c4ea31 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1090.048821] env[65726]: DEBUG oslo_vmware.api [None req-0857f350-c753-4ac4-a5de-c6f94d195846 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': task-5116699, 'name': SuspendVM_Task} progress is 50%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1090.158635] env[65726]: DEBUG oslo_concurrency.lockutils [None req-6a8181d2-ce0d-4923-b0a1-3e2a66a63f15 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Acquiring lock "refresh_cache-aa6ce489-c62f-4481-87b7-e74242aeb8ca" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1090.158810] env[65726]: DEBUG oslo_concurrency.lockutils [None req-6a8181d2-ce0d-4923-b0a1-3e2a66a63f15 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Acquired lock "refresh_cache-aa6ce489-c62f-4481-87b7-e74242aeb8ca" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1090.159036] env[65726]: DEBUG nova.network.neutron [None req-6a8181d2-ce0d-4923-b0a1-3e2a66a63f15 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] [instance: aa6ce489-c62f-4481-87b7-e74242aeb8ca] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1090.159178] env[65726]: DEBUG nova.objects.instance [None req-6a8181d2-ce0d-4923-b0a1-3e2a66a63f15 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Lazy-loading 'info_cache' on Instance uuid aa6ce489-c62f-4481-87b7-e74242aeb8ca {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1090.256769] env[65726]: DEBUG oslo_vmware.api [None req-6efd9d49-6747-4178-8149-554fa3ea5be4 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Task: {'id': task-5116700, 'name': ReconfigVM_Task, 'duration_secs': 0.310091} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1090.257164] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-6efd9d49-6747-4178-8149-554fa3ea5be4 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-995255', 'volume_id': '98eb7918-3cfa-4296-9855-bce47b982be1', 'name': 'volume-98eb7918-3cfa-4296-9855-bce47b982be1', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'attached', 'instance': '2fe6fc5a-f5c7-4f8e-96df-4e621a252f04', 'attached_at': '', 'detached_at': '', 'volume_id': '98eb7918-3cfa-4296-9855-bce47b982be1', 'serial': '98eb7918-3cfa-4296-9855-bce47b982be1'} {{(pid=65726) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1090.262073] env[65726]: INFO nova.compute.manager [None req-14a9b6f0-9c85-4271-a1df-1d81c0fc4a98 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: a56be6fa-6027-46ee-9aa8-332fe4222d68] Swapping old allocation on dict_keys(['07c4692f-bdb4-4058-9173-ff9664830295']) held by migration 0ae40754-d17a-4cf6-a2d0-e1747b172ad4 for instance [ 1090.284710] env[65726]: DEBUG nova.scheduler.client.report [None req-14a9b6f0-9c85-4271-a1df-1d81c0fc4a98 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Overwriting current allocation {'allocations': {'07c4692f-bdb4-4058-9173-ff9664830295': {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}, 'generation': 140}}, 'project_id': '4eece77569624f90bf64e5c51974173f', 'user_id': 'b251fcbd04044fcda192bc3febac1a90', 'consumer_generation': 1} on consumer a56be6fa-6027-46ee-9aa8-332fe4222d68 {{(pid=65726) move_allocations /opt/stack/nova/nova/scheduler/client/report.py:2036}} [ 1090.344204] env[65726]: WARNING openstack [None req-14a9b6f0-9c85-4271-a1df-1d81c0fc4a98 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1090.344756] env[65726]: WARNING openstack [None req-14a9b6f0-9c85-4271-a1df-1d81c0fc4a98 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1090.395987] env[65726]: DEBUG oslo_concurrency.lockutils [None req-14a9b6f0-9c85-4271-a1df-1d81c0fc4a98 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Acquiring lock "refresh_cache-a56be6fa-6027-46ee-9aa8-332fe4222d68" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1090.396295] env[65726]: DEBUG oslo_concurrency.lockutils [None req-14a9b6f0-9c85-4271-a1df-1d81c0fc4a98 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Acquired lock "refresh_cache-a56be6fa-6027-46ee-9aa8-332fe4222d68" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1090.396373] env[65726]: DEBUG nova.network.neutron [None req-14a9b6f0-9c85-4271-a1df-1d81c0fc4a98 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: a56be6fa-6027-46ee-9aa8-332fe4222d68] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1090.545504] env[65726]: DEBUG nova.scheduler.client.report [None req-5101d24a-5b00-4587-8b37-5a45f8c4ea31 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1090.558795] env[65726]: DEBUG oslo_vmware.api [None req-0857f350-c753-4ac4-a5de-c6f94d195846 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': task-5116699, 'name': SuspendVM_Task, 'duration_secs': 1.014281} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1090.559212] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-0857f350-c753-4ac4-a5de-c6f94d195846 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: d8801049-fc34-4b2f-8d5e-b9459c3e03a6] Suspended the VM {{(pid=65726) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1166}} [ 1090.559459] env[65726]: DEBUG nova.compute.manager [None req-0857f350-c753-4ac4-a5de-c6f94d195846 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: d8801049-fc34-4b2f-8d5e-b9459c3e03a6] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1090.560696] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a8cf09d-f9c8-409d-b386-49f9e08f5e9e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.663803] env[65726]: DEBUG nova.objects.base [None req-6a8181d2-ce0d-4923-b0a1-3e2a66a63f15 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Object Instance lazy-loaded attributes: flavor,info_cache {{(pid=65726) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1090.806550] env[65726]: DEBUG nova.objects.instance [None req-6efd9d49-6747-4178-8149-554fa3ea5be4 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Lazy-loading 'flavor' on Instance uuid 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04 {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1090.899408] env[65726]: WARNING openstack [None req-14a9b6f0-9c85-4271-a1df-1d81c0fc4a98 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1090.899798] env[65726]: WARNING openstack [None req-14a9b6f0-9c85-4271-a1df-1d81c0fc4a98 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1091.052168] env[65726]: WARNING openstack [None req-14a9b6f0-9c85-4271-a1df-1d81c0fc4a98 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1091.052617] env[65726]: WARNING openstack [None req-14a9b6f0-9c85-4271-a1df-1d81c0fc4a98 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1091.059575] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5101d24a-5b00-4587-8b37-5a45f8c4ea31 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.341s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1091.060059] env[65726]: DEBUG nova.compute.manager [None req-5101d24a-5b00-4587-8b37-5a45f8c4ea31 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 0361ae8f-634d-4161-a594-7f93dbb1b7cd] Start building networks asynchronously for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1091.062877] env[65726]: DEBUG oslo_concurrency.lockutils [None req-1c19f444-aee2-49ca-b67e-d3cef83d3664 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.698s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1091.063101] env[65726]: DEBUG nova.objects.instance [None req-1c19f444-aee2-49ca-b67e-d3cef83d3664 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Lazy-loading 'pci_requests' on Instance uuid 6fd70b4a-63e4-4258-9cff-f3b582500b9f {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1091.121152] env[65726]: WARNING openstack [None req-14a9b6f0-9c85-4271-a1df-1d81c0fc4a98 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1091.121529] env[65726]: WARNING openstack [None req-14a9b6f0-9c85-4271-a1df-1d81c0fc4a98 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1091.168907] env[65726]: WARNING openstack [None req-6a8181d2-ce0d-4923-b0a1-3e2a66a63f15 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1091.169322] env[65726]: WARNING openstack [None req-6a8181d2-ce0d-4923-b0a1-3e2a66a63f15 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1091.214933] env[65726]: DEBUG nova.network.neutron [None req-14a9b6f0-9c85-4271-a1df-1d81c0fc4a98 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: a56be6fa-6027-46ee-9aa8-332fe4222d68] Updating instance_info_cache with network_info: [{"id": "4e2dfbcd-41c3-4aaa-a2ec-50c855c12307", "address": "fa:16:3e:1d:82:59", "network": {"id": "0d0f561d-ac41-4ef5-9926-c707409799f7", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1134321217-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.160", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4eece77569624f90bf64e5c51974173f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2ce62383-8e84-4e26-955b-74c11392f4c9", "external-id": "nsx-vlan-transportzone-215", "segmentation_id": 215, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4e2dfbcd-41", "ovs_interfaceid": "4e2dfbcd-41c3-4aaa-a2ec-50c855c12307", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1091.287571] env[65726]: WARNING openstack [None req-6a8181d2-ce0d-4923-b0a1-3e2a66a63f15 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1091.288393] env[65726]: WARNING openstack [None req-6a8181d2-ce0d-4923-b0a1-3e2a66a63f15 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1091.354691] env[65726]: WARNING openstack [None req-6a8181d2-ce0d-4923-b0a1-3e2a66a63f15 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1091.355228] env[65726]: WARNING openstack [None req-6a8181d2-ce0d-4923-b0a1-3e2a66a63f15 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1091.468204] env[65726]: DEBUG nova.network.neutron [None req-6a8181d2-ce0d-4923-b0a1-3e2a66a63f15 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] [instance: aa6ce489-c62f-4481-87b7-e74242aeb8ca] Updating instance_info_cache with network_info: [{"id": "d5a805f5-b6a8-42ec-92a0-4a652b51c1e6", "address": "fa:16:3e:f5:d5:d7", "network": {"id": "547d7b16-4591-4d80-9e9b-c8a9a3a78dab", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-2077874536-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.242", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4ffd45f4a7a041199a4fc7f69f5e7e9b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2ed91b7b-b4ec-486d-ab34-af0afb7ec691", "external-id": "nsx-vlan-transportzone-75", "segmentation_id": 75, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd5a805f5-b6", "ovs_interfaceid": "d5a805f5-b6a8-42ec-92a0-4a652b51c1e6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1091.566649] env[65726]: DEBUG nova.compute.utils [None req-5101d24a-5b00-4587-8b37-5a45f8c4ea31 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Using /dev/sd instead of None {{(pid=65726) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1091.570375] env[65726]: DEBUG nova.objects.instance [None req-1c19f444-aee2-49ca-b67e-d3cef83d3664 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Lazy-loading 'numa_topology' on Instance uuid 6fd70b4a-63e4-4258-9cff-f3b582500b9f {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1091.571865] env[65726]: DEBUG nova.compute.manager [None req-5101d24a-5b00-4587-8b37-5a45f8c4ea31 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 0361ae8f-634d-4161-a594-7f93dbb1b7cd] Allocating IP information in the background. {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1091.571865] env[65726]: DEBUG nova.network.neutron [None req-5101d24a-5b00-4587-8b37-5a45f8c4ea31 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 0361ae8f-634d-4161-a594-7f93dbb1b7cd] allocate_for_instance() {{(pid=65726) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 1091.572013] env[65726]: WARNING neutronclient.v2_0.client [None req-5101d24a-5b00-4587-8b37-5a45f8c4ea31 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1091.572443] env[65726]: WARNING neutronclient.v2_0.client [None req-5101d24a-5b00-4587-8b37-5a45f8c4ea31 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1091.573151] env[65726]: WARNING openstack [None req-5101d24a-5b00-4587-8b37-5a45f8c4ea31 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1091.573572] env[65726]: WARNING openstack [None req-5101d24a-5b00-4587-8b37-5a45f8c4ea31 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1091.629160] env[65726]: DEBUG nova.policy [None req-5101d24a-5b00-4587-8b37-5a45f8c4ea31 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a82797497789484a8fbf88a23a449c95', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f93d852e2c904f42981cbfff882558e9', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65726) authorize /opt/stack/nova/nova/policy.py:192}} [ 1091.717431] env[65726]: DEBUG oslo_concurrency.lockutils [None req-14a9b6f0-9c85-4271-a1df-1d81c0fc4a98 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Releasing lock "refresh_cache-a56be6fa-6027-46ee-9aa8-332fe4222d68" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1091.718495] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-156fb567-aa46-46ca-9572-6b4ec06cdf01 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.726360] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14606434-65e6-432c-896b-45d3c4ca0f5d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.815682] env[65726]: DEBUG oslo_concurrency.lockutils [None req-6efd9d49-6747-4178-8149-554fa3ea5be4 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Lock "2fe6fc5a-f5c7-4f8e-96df-4e621a252f04" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.276s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1091.866510] env[65726]: DEBUG nova.network.neutron [None req-5101d24a-5b00-4587-8b37-5a45f8c4ea31 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 0361ae8f-634d-4161-a594-7f93dbb1b7cd] Successfully created port: 2bfd7147-87fd-4a19-b28b-eb95f51f4d00 {{(pid=65726) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1091.948904] env[65726]: DEBUG oslo_concurrency.lockutils [None req-f67ad2d6-6757-4118-93c4-4106809d054d tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Acquiring lock "d8801049-fc34-4b2f-8d5e-b9459c3e03a6" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1091.949427] env[65726]: DEBUG oslo_concurrency.lockutils [None req-f67ad2d6-6757-4118-93c4-4106809d054d tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Lock "d8801049-fc34-4b2f-8d5e-b9459c3e03a6" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1091.949674] env[65726]: DEBUG oslo_concurrency.lockutils [None req-f67ad2d6-6757-4118-93c4-4106809d054d tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Acquiring lock "d8801049-fc34-4b2f-8d5e-b9459c3e03a6-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1091.949890] env[65726]: DEBUG oslo_concurrency.lockutils [None req-f67ad2d6-6757-4118-93c4-4106809d054d tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Lock "d8801049-fc34-4b2f-8d5e-b9459c3e03a6-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1091.950087] env[65726]: DEBUG oslo_concurrency.lockutils [None req-f67ad2d6-6757-4118-93c4-4106809d054d tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Lock "d8801049-fc34-4b2f-8d5e-b9459c3e03a6-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1091.952997] env[65726]: INFO nova.compute.manager [None req-f67ad2d6-6757-4118-93c4-4106809d054d tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: d8801049-fc34-4b2f-8d5e-b9459c3e03a6] Terminating instance [ 1091.972616] env[65726]: DEBUG oslo_concurrency.lockutils [None req-6a8181d2-ce0d-4923-b0a1-3e2a66a63f15 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Releasing lock "refresh_cache-aa6ce489-c62f-4481-87b7-e74242aeb8ca" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1092.069937] env[65726]: DEBUG nova.compute.manager [None req-5101d24a-5b00-4587-8b37-5a45f8c4ea31 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 0361ae8f-634d-4161-a594-7f93dbb1b7cd] Start building block device mappings for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1092.081837] env[65726]: INFO nova.compute.claims [None req-1c19f444-aee2-49ca-b67e-d3cef83d3664 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] [instance: 6fd70b4a-63e4-4258-9cff-f3b582500b9f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1092.457486] env[65726]: DEBUG nova.compute.manager [None req-f67ad2d6-6757-4118-93c4-4106809d054d tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: d8801049-fc34-4b2f-8d5e-b9459c3e03a6] Start destroying the instance on the hypervisor. {{(pid=65726) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1092.457724] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-f67ad2d6-6757-4118-93c4-4106809d054d tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: d8801049-fc34-4b2f-8d5e-b9459c3e03a6] Destroying instance {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1092.458701] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0f2e1c7-3393-4510-bcf9-0c9917d2d5f3 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.467891] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-f67ad2d6-6757-4118-93c4-4106809d054d tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: d8801049-fc34-4b2f-8d5e-b9459c3e03a6] Unregistering the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1092.468177] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d162d0fe-4133-47b0-ae23-abee90d0a2f9 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.546954] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-f67ad2d6-6757-4118-93c4-4106809d054d tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: d8801049-fc34-4b2f-8d5e-b9459c3e03a6] Unregistered the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1092.547333] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-f67ad2d6-6757-4118-93c4-4106809d054d tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: d8801049-fc34-4b2f-8d5e-b9459c3e03a6] Deleting contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1092.547388] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-f67ad2d6-6757-4118-93c4-4106809d054d tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Deleting the datastore file [datastore1] d8801049-fc34-4b2f-8d5e-b9459c3e03a6 {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1092.547671] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-021103c1-b573-49b5-9df1-716cc7983fef {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.555821] env[65726]: DEBUG oslo_vmware.api [None req-f67ad2d6-6757-4118-93c4-4106809d054d tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Waiting for the task: (returnval){ [ 1092.555821] env[65726]: value = "task-5116702" [ 1092.555821] env[65726]: _type = "Task" [ 1092.555821] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1092.565593] env[65726]: DEBUG oslo_vmware.api [None req-f67ad2d6-6757-4118-93c4-4106809d054d tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': task-5116702, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1092.818854] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-14a9b6f0-9c85-4271-a1df-1d81c0fc4a98 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: a56be6fa-6027-46ee-9aa8-332fe4222d68] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1092.819292] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-180c0550-48e4-42f1-8091-f44f9759c3f2 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.827850] env[65726]: DEBUG oslo_vmware.api [None req-14a9b6f0-9c85-4271-a1df-1d81c0fc4a98 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Waiting for the task: (returnval){ [ 1092.827850] env[65726]: value = "task-5116703" [ 1092.827850] env[65726]: _type = "Task" [ 1092.827850] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1092.840766] env[65726]: DEBUG oslo_vmware.api [None req-14a9b6f0-9c85-4271-a1df-1d81c0fc4a98 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5116703, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1092.896538] env[65726]: DEBUG oslo_concurrency.lockutils [None req-eab5602f-7673-4d76-aa16-7ebf0d2c87e3 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Acquiring lock "2fe6fc5a-f5c7-4f8e-96df-4e621a252f04" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1092.896964] env[65726]: DEBUG oslo_concurrency.lockutils [None req-eab5602f-7673-4d76-aa16-7ebf0d2c87e3 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Lock "2fe6fc5a-f5c7-4f8e-96df-4e621a252f04" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1092.897289] env[65726]: DEBUG oslo_concurrency.lockutils [None req-eab5602f-7673-4d76-aa16-7ebf0d2c87e3 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Acquiring lock "2fe6fc5a-f5c7-4f8e-96df-4e621a252f04-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1092.897596] env[65726]: DEBUG oslo_concurrency.lockutils [None req-eab5602f-7673-4d76-aa16-7ebf0d2c87e3 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Lock "2fe6fc5a-f5c7-4f8e-96df-4e621a252f04-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1092.897853] env[65726]: DEBUG oslo_concurrency.lockutils [None req-eab5602f-7673-4d76-aa16-7ebf0d2c87e3 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Lock "2fe6fc5a-f5c7-4f8e-96df-4e621a252f04-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1092.900362] env[65726]: INFO nova.compute.manager [None req-eab5602f-7673-4d76-aa16-7ebf0d2c87e3 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04] Terminating instance [ 1092.978560] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-6a8181d2-ce0d-4923-b0a1-3e2a66a63f15 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] [instance: aa6ce489-c62f-4481-87b7-e74242aeb8ca] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1092.979329] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a0431655-c38f-4f98-9a42-836863ef573c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.987510] env[65726]: DEBUG oslo_vmware.api [None req-6a8181d2-ce0d-4923-b0a1-3e2a66a63f15 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Waiting for the task: (returnval){ [ 1092.987510] env[65726]: value = "task-5116704" [ 1092.987510] env[65726]: _type = "Task" [ 1092.987510] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1092.997142] env[65726]: DEBUG oslo_vmware.api [None req-6a8181d2-ce0d-4923-b0a1-3e2a66a63f15 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Task: {'id': task-5116704, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1093.068104] env[65726]: DEBUG oslo_vmware.api [None req-f67ad2d6-6757-4118-93c4-4106809d054d tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': task-5116702, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.157106} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1093.068420] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-f67ad2d6-6757-4118-93c4-4106809d054d tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Deleted the datastore file {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1093.068626] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-f67ad2d6-6757-4118-93c4-4106809d054d tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: d8801049-fc34-4b2f-8d5e-b9459c3e03a6] Deleted contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1093.068847] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-f67ad2d6-6757-4118-93c4-4106809d054d tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: d8801049-fc34-4b2f-8d5e-b9459c3e03a6] Instance destroyed {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1093.069045] env[65726]: INFO nova.compute.manager [None req-f67ad2d6-6757-4118-93c4-4106809d054d tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: d8801049-fc34-4b2f-8d5e-b9459c3e03a6] Took 0.61 seconds to destroy the instance on the hypervisor. [ 1093.069351] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-f67ad2d6-6757-4118-93c4-4106809d054d tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1093.069592] env[65726]: DEBUG nova.compute.manager [-] [instance: d8801049-fc34-4b2f-8d5e-b9459c3e03a6] Deallocating network for instance {{(pid=65726) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1093.069695] env[65726]: DEBUG nova.network.neutron [-] [instance: d8801049-fc34-4b2f-8d5e-b9459c3e03a6] deallocate_for_instance() {{(pid=65726) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 1093.070234] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1093.070517] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1093.082679] env[65726]: DEBUG nova.compute.manager [None req-5101d24a-5b00-4587-8b37-5a45f8c4ea31 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 0361ae8f-634d-4161-a594-7f93dbb1b7cd] Start spawning the instance on the hypervisor. {{(pid=65726) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1093.116390] env[65726]: DEBUG nova.virt.hardware [None req-5101d24a-5b00-4587-8b37-5a45f8c4ea31 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T19:28:34Z,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1539f06fa3534e90acbc3a60c4b8bd3f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T19:28:35Z,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1093.116749] env[65726]: DEBUG nova.virt.hardware [None req-5101d24a-5b00-4587-8b37-5a45f8c4ea31 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1093.116981] env[65726]: DEBUG nova.virt.hardware [None req-5101d24a-5b00-4587-8b37-5a45f8c4ea31 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1093.117272] env[65726]: DEBUG nova.virt.hardware [None req-5101d24a-5b00-4587-8b37-5a45f8c4ea31 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1093.117483] env[65726]: DEBUG nova.virt.hardware [None req-5101d24a-5b00-4587-8b37-5a45f8c4ea31 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1093.117731] env[65726]: DEBUG nova.virt.hardware [None req-5101d24a-5b00-4587-8b37-5a45f8c4ea31 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1093.118036] env[65726]: DEBUG nova.virt.hardware [None req-5101d24a-5b00-4587-8b37-5a45f8c4ea31 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1093.118325] env[65726]: DEBUG nova.virt.hardware [None req-5101d24a-5b00-4587-8b37-5a45f8c4ea31 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1093.118603] env[65726]: DEBUG nova.virt.hardware [None req-5101d24a-5b00-4587-8b37-5a45f8c4ea31 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1093.118835] env[65726]: DEBUG nova.virt.hardware [None req-5101d24a-5b00-4587-8b37-5a45f8c4ea31 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1093.119072] env[65726]: DEBUG nova.virt.hardware [None req-5101d24a-5b00-4587-8b37-5a45f8c4ea31 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1093.120086] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04bd8a10-a86f-47f8-986d-8ca24d850a22 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.127435] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1093.127891] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1093.144668] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2ddfb43-84a5-4910-890c-796e1af9832d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.296682] env[65726]: DEBUG nova.compute.manager [req-266b5865-bb46-4dbd-8a83-a498b13c53d2 req-8d64c442-a077-4732-8086-82aec8acc41f service nova] [instance: 0361ae8f-634d-4161-a594-7f93dbb1b7cd] Received event network-vif-plugged-2bfd7147-87fd-4a19-b28b-eb95f51f4d00 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1093.296896] env[65726]: DEBUG oslo_concurrency.lockutils [req-266b5865-bb46-4dbd-8a83-a498b13c53d2 req-8d64c442-a077-4732-8086-82aec8acc41f service nova] Acquiring lock "0361ae8f-634d-4161-a594-7f93dbb1b7cd-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1093.297178] env[65726]: DEBUG oslo_concurrency.lockutils [req-266b5865-bb46-4dbd-8a83-a498b13c53d2 req-8d64c442-a077-4732-8086-82aec8acc41f service nova] Lock "0361ae8f-634d-4161-a594-7f93dbb1b7cd-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1093.297355] env[65726]: DEBUG oslo_concurrency.lockutils [req-266b5865-bb46-4dbd-8a83-a498b13c53d2 req-8d64c442-a077-4732-8086-82aec8acc41f service nova] Lock "0361ae8f-634d-4161-a594-7f93dbb1b7cd-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1093.297524] env[65726]: DEBUG nova.compute.manager [req-266b5865-bb46-4dbd-8a83-a498b13c53d2 req-8d64c442-a077-4732-8086-82aec8acc41f service nova] [instance: 0361ae8f-634d-4161-a594-7f93dbb1b7cd] No waiting events found dispatching network-vif-plugged-2bfd7147-87fd-4a19-b28b-eb95f51f4d00 {{(pid=65726) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1093.297685] env[65726]: WARNING nova.compute.manager [req-266b5865-bb46-4dbd-8a83-a498b13c53d2 req-8d64c442-a077-4732-8086-82aec8acc41f service nova] [instance: 0361ae8f-634d-4161-a594-7f93dbb1b7cd] Received unexpected event network-vif-plugged-2bfd7147-87fd-4a19-b28b-eb95f51f4d00 for instance with vm_state building and task_state spawning. [ 1093.325241] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af67476a-643e-414e-9dc7-c0038ce11022 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.339729] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-563d88e1-095d-45d0-9254-1350bed2094b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.347812] env[65726]: DEBUG oslo_vmware.api [None req-14a9b6f0-9c85-4271-a1df-1d81c0fc4a98 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5116703, 'name': PowerOffVM_Task, 'duration_secs': 0.225853} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1093.349323] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-14a9b6f0-9c85-4271-a1df-1d81c0fc4a98 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: a56be6fa-6027-46ee-9aa8-332fe4222d68] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1093.349927] env[65726]: DEBUG nova.virt.hardware [None req-14a9b6f0-9c85-4271-a1df-1d81c0fc4a98 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1093.350369] env[65726]: DEBUG nova.virt.hardware [None req-14a9b6f0-9c85-4271-a1df-1d81c0fc4a98 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1093.350369] env[65726]: DEBUG nova.virt.hardware [None req-14a9b6f0-9c85-4271-a1df-1d81c0fc4a98 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1093.350509] env[65726]: DEBUG nova.virt.hardware [None req-14a9b6f0-9c85-4271-a1df-1d81c0fc4a98 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1093.350766] env[65726]: DEBUG nova.virt.hardware [None req-14a9b6f0-9c85-4271-a1df-1d81c0fc4a98 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1093.350851] env[65726]: DEBUG nova.virt.hardware [None req-14a9b6f0-9c85-4271-a1df-1d81c0fc4a98 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1093.351113] env[65726]: DEBUG nova.virt.hardware [None req-14a9b6f0-9c85-4271-a1df-1d81c0fc4a98 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1093.351216] env[65726]: DEBUG nova.virt.hardware [None req-14a9b6f0-9c85-4271-a1df-1d81c0fc4a98 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1093.351400] env[65726]: DEBUG nova.virt.hardware [None req-14a9b6f0-9c85-4271-a1df-1d81c0fc4a98 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1093.351568] env[65726]: DEBUG nova.virt.hardware [None req-14a9b6f0-9c85-4271-a1df-1d81c0fc4a98 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1093.351876] env[65726]: DEBUG nova.virt.hardware [None req-14a9b6f0-9c85-4271-a1df-1d81c0fc4a98 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1093.391527] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7f6577f2-54d9-418c-9090-48b40ffb7539 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.410743] env[65726]: DEBUG nova.network.neutron [None req-5101d24a-5b00-4587-8b37-5a45f8c4ea31 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 0361ae8f-634d-4161-a594-7f93dbb1b7cd] Successfully updated port: 2bfd7147-87fd-4a19-b28b-eb95f51f4d00 {{(pid=65726) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 1093.412587] env[65726]: DEBUG nova.compute.manager [None req-eab5602f-7673-4d76-aa16-7ebf0d2c87e3 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04] Start destroying the instance on the hypervisor. {{(pid=65726) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1093.412587] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-eab5602f-7673-4d76-aa16-7ebf0d2c87e3 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04] Destroying instance {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1093.413340] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43b73ea6-c949-41b2-9002-9d2a1bfdb4a3 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.416934] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2cdef1a2-158a-4843-91bd-9b1cdd2ce5a3 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.433584] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d17d3c97-1376-4843-b4b5-7d3247e48617 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.438268] env[65726]: DEBUG oslo_vmware.api [None req-14a9b6f0-9c85-4271-a1df-1d81c0fc4a98 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Waiting for the task: (returnval){ [ 1093.438268] env[65726]: value = "task-5116705" [ 1093.438268] env[65726]: _type = "Task" [ 1093.438268] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1093.438558] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-eab5602f-7673-4d76-aa16-7ebf0d2c87e3 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1093.439767] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2fabc7be-f2fd-4caa-8987-d161e19aa814 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.455261] env[65726]: DEBUG nova.compute.provider_tree [None req-1c19f444-aee2-49ca-b67e-d3cef83d3664 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1093.460258] env[65726]: DEBUG oslo_vmware.api [None req-eab5602f-7673-4d76-aa16-7ebf0d2c87e3 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Waiting for the task: (returnval){ [ 1093.460258] env[65726]: value = "task-5116706" [ 1093.460258] env[65726]: _type = "Task" [ 1093.460258] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1093.464339] env[65726]: DEBUG oslo_vmware.api [None req-14a9b6f0-9c85-4271-a1df-1d81c0fc4a98 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5116705, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1093.470790] env[65726]: DEBUG nova.compute.manager [req-696f5876-23d0-4e25-b2c1-903e0c0df38a req-9b47c72a-75ab-4765-a3c5-3fa5eb88d848 service nova] [instance: d8801049-fc34-4b2f-8d5e-b9459c3e03a6] Received event network-vif-deleted-b333a678-e6d6-494e-bc4d-7e498e370067 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1093.471099] env[65726]: INFO nova.compute.manager [req-696f5876-23d0-4e25-b2c1-903e0c0df38a req-9b47c72a-75ab-4765-a3c5-3fa5eb88d848 service nova] [instance: d8801049-fc34-4b2f-8d5e-b9459c3e03a6] Neutron deleted interface b333a678-e6d6-494e-bc4d-7e498e370067; detaching it from the instance and deleting it from the info cache [ 1093.471319] env[65726]: DEBUG nova.network.neutron [req-696f5876-23d0-4e25-b2c1-903e0c0df38a req-9b47c72a-75ab-4765-a3c5-3fa5eb88d848 service nova] [instance: d8801049-fc34-4b2f-8d5e-b9459c3e03a6] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1093.480272] env[65726]: DEBUG oslo_vmware.api [None req-eab5602f-7673-4d76-aa16-7ebf0d2c87e3 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Task: {'id': task-5116706, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1093.498208] env[65726]: DEBUG oslo_vmware.api [None req-6a8181d2-ce0d-4923-b0a1-3e2a66a63f15 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Task: {'id': task-5116704, 'name': PowerOnVM_Task, 'duration_secs': 0.392023} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1093.498496] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-6a8181d2-ce0d-4923-b0a1-3e2a66a63f15 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] [instance: aa6ce489-c62f-4481-87b7-e74242aeb8ca] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1093.498708] env[65726]: DEBUG nova.compute.manager [None req-6a8181d2-ce0d-4923-b0a1-3e2a66a63f15 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] [instance: aa6ce489-c62f-4481-87b7-e74242aeb8ca] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1093.499608] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c00f10d5-fba3-4665-9cb8-69495980378d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.917156] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5101d24a-5b00-4587-8b37-5a45f8c4ea31 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Acquiring lock "refresh_cache-0361ae8f-634d-4161-a594-7f93dbb1b7cd" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1093.917455] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5101d24a-5b00-4587-8b37-5a45f8c4ea31 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Acquired lock "refresh_cache-0361ae8f-634d-4161-a594-7f93dbb1b7cd" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1093.917706] env[65726]: DEBUG nova.network.neutron [None req-5101d24a-5b00-4587-8b37-5a45f8c4ea31 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 0361ae8f-634d-4161-a594-7f93dbb1b7cd] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1093.919071] env[65726]: DEBUG nova.network.neutron [-] [instance: d8801049-fc34-4b2f-8d5e-b9459c3e03a6] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1093.949662] env[65726]: DEBUG oslo_vmware.api [None req-14a9b6f0-9c85-4271-a1df-1d81c0fc4a98 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5116705, 'name': ReconfigVM_Task, 'duration_secs': 0.171615} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1093.950562] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d130fe7a-7736-4493-a3b6-d77baf01fba8 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.972820] env[65726]: DEBUG nova.scheduler.client.report [None req-1c19f444-aee2-49ca-b67e-d3cef83d3664 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1093.977277] env[65726]: DEBUG nova.virt.hardware [None req-14a9b6f0-9c85-4271-a1df-1d81c0fc4a98 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1093.977542] env[65726]: DEBUG nova.virt.hardware [None req-14a9b6f0-9c85-4271-a1df-1d81c0fc4a98 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1093.977732] env[65726]: DEBUG nova.virt.hardware [None req-14a9b6f0-9c85-4271-a1df-1d81c0fc4a98 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1093.977929] env[65726]: DEBUG nova.virt.hardware [None req-14a9b6f0-9c85-4271-a1df-1d81c0fc4a98 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1093.978086] env[65726]: DEBUG nova.virt.hardware [None req-14a9b6f0-9c85-4271-a1df-1d81c0fc4a98 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1093.978248] env[65726]: DEBUG nova.virt.hardware [None req-14a9b6f0-9c85-4271-a1df-1d81c0fc4a98 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1093.978475] env[65726]: DEBUG nova.virt.hardware [None req-14a9b6f0-9c85-4271-a1df-1d81c0fc4a98 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1093.978637] env[65726]: DEBUG nova.virt.hardware [None req-14a9b6f0-9c85-4271-a1df-1d81c0fc4a98 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1093.978805] env[65726]: DEBUG nova.virt.hardware [None req-14a9b6f0-9c85-4271-a1df-1d81c0fc4a98 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1093.978967] env[65726]: DEBUG nova.virt.hardware [None req-14a9b6f0-9c85-4271-a1df-1d81c0fc4a98 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1093.979157] env[65726]: DEBUG nova.virt.hardware [None req-14a9b6f0-9c85-4271-a1df-1d81c0fc4a98 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1093.983635] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9b7851c3-d45a-413c-8966-09626ac75193 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.986324] env[65726]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e8d281d4-afc6-4470-b1ed-c701af259e12 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.996072] env[65726]: DEBUG oslo_vmware.api [None req-14a9b6f0-9c85-4271-a1df-1d81c0fc4a98 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Waiting for the task: (returnval){ [ 1093.996072] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52eadecb-48e9-f5e6-7ae2-88b48d924538" [ 1093.996072] env[65726]: _type = "Task" [ 1093.996072] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1094.002255] env[65726]: DEBUG oslo_vmware.api [None req-eab5602f-7673-4d76-aa16-7ebf0d2c87e3 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Task: {'id': task-5116706, 'name': PowerOffVM_Task, 'duration_secs': 0.261771} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1094.008976] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-296796f4-d942-4561-b902-79cc31428ec3 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.019929] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-eab5602f-7673-4d76-aa16-7ebf0d2c87e3 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1094.020135] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-eab5602f-7673-4d76-aa16-7ebf0d2c87e3 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04] Unregistering the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1094.023154] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7236b7e8-81b2-4741-943c-0f34fde73919 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.033948] env[65726]: DEBUG oslo_vmware.api [None req-14a9b6f0-9c85-4271-a1df-1d81c0fc4a98 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52eadecb-48e9-f5e6-7ae2-88b48d924538, 'name': SearchDatastore_Task, 'duration_secs': 0.008479} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1094.041541] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-14a9b6f0-9c85-4271-a1df-1d81c0fc4a98 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: a56be6fa-6027-46ee-9aa8-332fe4222d68] Reconfiguring VM instance instance-00000053 to detach disk 2000 {{(pid=65726) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1094.055755] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-093b44d8-f865-4883-8699-0da84e2c06b6 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.067122] env[65726]: DEBUG nova.compute.manager [req-696f5876-23d0-4e25-b2c1-903e0c0df38a req-9b47c72a-75ab-4765-a3c5-3fa5eb88d848 service nova] [instance: d8801049-fc34-4b2f-8d5e-b9459c3e03a6] Detach interface failed, port_id=b333a678-e6d6-494e-bc4d-7e498e370067, reason: Instance d8801049-fc34-4b2f-8d5e-b9459c3e03a6 could not be found. {{(pid=65726) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11669}} [ 1094.073759] env[65726]: DEBUG oslo_vmware.api [None req-14a9b6f0-9c85-4271-a1df-1d81c0fc4a98 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Waiting for the task: (returnval){ [ 1094.073759] env[65726]: value = "task-5116708" [ 1094.073759] env[65726]: _type = "Task" [ 1094.073759] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1094.087451] env[65726]: DEBUG oslo_vmware.api [None req-14a9b6f0-9c85-4271-a1df-1d81c0fc4a98 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5116708, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1094.094700] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-eab5602f-7673-4d76-aa16-7ebf0d2c87e3 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04] Unregistered the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1094.094971] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-eab5602f-7673-4d76-aa16-7ebf0d2c87e3 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04] Deleting contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1094.095246] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-eab5602f-7673-4d76-aa16-7ebf0d2c87e3 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Deleting the datastore file [datastore1] 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04 {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1094.095553] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b728b521-2854-44e8-b29b-bd4e101aeaf2 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.104089] env[65726]: DEBUG oslo_vmware.api [None req-eab5602f-7673-4d76-aa16-7ebf0d2c87e3 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Waiting for the task: (returnval){ [ 1094.104089] env[65726]: value = "task-5116709" [ 1094.104089] env[65726]: _type = "Task" [ 1094.104089] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1094.115474] env[65726]: DEBUG oslo_vmware.api [None req-eab5602f-7673-4d76-aa16-7ebf0d2c87e3 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Task: {'id': task-5116709, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1094.421645] env[65726]: WARNING openstack [None req-5101d24a-5b00-4587-8b37-5a45f8c4ea31 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1094.422055] env[65726]: WARNING openstack [None req-5101d24a-5b00-4587-8b37-5a45f8c4ea31 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1094.427396] env[65726]: DEBUG nova.network.neutron [None req-5101d24a-5b00-4587-8b37-5a45f8c4ea31 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 0361ae8f-634d-4161-a594-7f93dbb1b7cd] Instance cache missing network info. {{(pid=65726) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 1094.429307] env[65726]: INFO nova.compute.manager [-] [instance: d8801049-fc34-4b2f-8d5e-b9459c3e03a6] Took 1.36 seconds to deallocate network for instance. [ 1094.481286] env[65726]: DEBUG oslo_concurrency.lockutils [None req-1c19f444-aee2-49ca-b67e-d3cef83d3664 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.418s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1094.481927] env[65726]: WARNING openstack [None req-1c19f444-aee2-49ca-b67e-d3cef83d3664 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1094.482292] env[65726]: WARNING openstack [None req-1c19f444-aee2-49ca-b67e-d3cef83d3664 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1094.491958] env[65726]: WARNING openstack [None req-5101d24a-5b00-4587-8b37-5a45f8c4ea31 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1094.492358] env[65726]: WARNING openstack [None req-5101d24a-5b00-4587-8b37-5a45f8c4ea31 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1094.498864] env[65726]: DEBUG oslo_concurrency.lockutils [None req-4f5aff8f-c14b-4955-a84c-9d59624f9d81 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 7.018s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1094.499126] env[65726]: DEBUG nova.objects.instance [None req-4f5aff8f-c14b-4955-a84c-9d59624f9d81 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Lazy-loading 'resources' on Instance uuid 5cdf1d61-b817-4986-a48c-804a2374497a {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1094.526914] env[65726]: INFO nova.network.neutron [None req-1c19f444-aee2-49ca-b67e-d3cef83d3664 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] [instance: 6fd70b4a-63e4-4258-9cff-f3b582500b9f] Updating port 153b4c9d-d01d-4254-9aa6-040705be347a with attributes {'binding_host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 1094.583375] env[65726]: WARNING openstack [None req-5101d24a-5b00-4587-8b37-5a45f8c4ea31 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1094.583756] env[65726]: WARNING openstack [None req-5101d24a-5b00-4587-8b37-5a45f8c4ea31 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1094.597260] env[65726]: DEBUG oslo_vmware.api [None req-14a9b6f0-9c85-4271-a1df-1d81c0fc4a98 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5116708, 'name': ReconfigVM_Task, 'duration_secs': 0.231707} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1094.597429] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-14a9b6f0-9c85-4271-a1df-1d81c0fc4a98 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: a56be6fa-6027-46ee-9aa8-332fe4222d68] Reconfigured VM instance instance-00000053 to detach disk 2000 {{(pid=65726) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1094.598353] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-381e1c1e-e5c7-4b63-a1ca-487dd3c9ce0a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.625770] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-14a9b6f0-9c85-4271-a1df-1d81c0fc4a98 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: a56be6fa-6027-46ee-9aa8-332fe4222d68] Reconfiguring VM instance instance-00000053 to attach disk [datastore1] a56be6fa-6027-46ee-9aa8-332fe4222d68/a56be6fa-6027-46ee-9aa8-332fe4222d68.vmdk or device None with type thin {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1094.631466] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f23409f4-7d75-43fc-ab9b-387dca83317c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.657657] env[65726]: DEBUG oslo_vmware.api [None req-eab5602f-7673-4d76-aa16-7ebf0d2c87e3 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Task: {'id': task-5116709, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.149451} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1094.659295] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-eab5602f-7673-4d76-aa16-7ebf0d2c87e3 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Deleted the datastore file {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1094.659483] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-eab5602f-7673-4d76-aa16-7ebf0d2c87e3 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04] Deleted contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1094.659670] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-eab5602f-7673-4d76-aa16-7ebf0d2c87e3 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04] Instance destroyed {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1094.659842] env[65726]: INFO nova.compute.manager [None req-eab5602f-7673-4d76-aa16-7ebf0d2c87e3 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04] Took 1.25 seconds to destroy the instance on the hypervisor. [ 1094.660112] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-eab5602f-7673-4d76-aa16-7ebf0d2c87e3 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1094.660395] env[65726]: DEBUG oslo_vmware.api [None req-14a9b6f0-9c85-4271-a1df-1d81c0fc4a98 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Waiting for the task: (returnval){ [ 1094.660395] env[65726]: value = "task-5116710" [ 1094.660395] env[65726]: _type = "Task" [ 1094.660395] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1094.660629] env[65726]: DEBUG nova.compute.manager [-] [instance: 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04] Deallocating network for instance {{(pid=65726) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1094.661362] env[65726]: DEBUG nova.network.neutron [-] [instance: 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04] deallocate_for_instance() {{(pid=65726) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 1094.661362] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1094.661517] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1094.679175] env[65726]: DEBUG oslo_vmware.api [None req-14a9b6f0-9c85-4271-a1df-1d81c0fc4a98 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5116710, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1094.695830] env[65726]: DEBUG nova.network.neutron [None req-5101d24a-5b00-4587-8b37-5a45f8c4ea31 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 0361ae8f-634d-4161-a594-7f93dbb1b7cd] Updating instance_info_cache with network_info: [{"id": "2bfd7147-87fd-4a19-b28b-eb95f51f4d00", "address": "fa:16:3e:d6:07:64", "network": {"id": "9dd55fb6-0dd2-484d-b94d-fbdccb958c79", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1005598846-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f93d852e2c904f42981cbfff882558e9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "74f30339-6421-4654-bddb-81d7f34db9d7", "external-id": "nsx-vlan-transportzone-899", "segmentation_id": 899, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2bfd7147-87", "ovs_interfaceid": "2bfd7147-87fd-4a19-b28b-eb95f51f4d00", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1094.727617] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1094.727905] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1094.938715] env[65726]: DEBUG oslo_concurrency.lockutils [None req-f67ad2d6-6757-4118-93c4-4106809d054d tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1095.185156] env[65726]: DEBUG oslo_vmware.api [None req-14a9b6f0-9c85-4271-a1df-1d81c0fc4a98 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5116710, 'name': ReconfigVM_Task, 'duration_secs': 0.357854} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1095.185423] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-14a9b6f0-9c85-4271-a1df-1d81c0fc4a98 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: a56be6fa-6027-46ee-9aa8-332fe4222d68] Reconfigured VM instance instance-00000053 to attach disk [datastore1] a56be6fa-6027-46ee-9aa8-332fe4222d68/a56be6fa-6027-46ee-9aa8-332fe4222d68.vmdk or device None with type thin {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1095.186393] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c57c962-6ceb-4de7-be5d-eb53cd888d40 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.209062] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5101d24a-5b00-4587-8b37-5a45f8c4ea31 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Releasing lock "refresh_cache-0361ae8f-634d-4161-a594-7f93dbb1b7cd" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1095.209447] env[65726]: DEBUG nova.compute.manager [None req-5101d24a-5b00-4587-8b37-5a45f8c4ea31 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 0361ae8f-634d-4161-a594-7f93dbb1b7cd] Instance network_info: |[{"id": "2bfd7147-87fd-4a19-b28b-eb95f51f4d00", "address": "fa:16:3e:d6:07:64", "network": {"id": "9dd55fb6-0dd2-484d-b94d-fbdccb958c79", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1005598846-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f93d852e2c904f42981cbfff882558e9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "74f30339-6421-4654-bddb-81d7f34db9d7", "external-id": "nsx-vlan-transportzone-899", "segmentation_id": 899, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2bfd7147-87", "ovs_interfaceid": "2bfd7147-87fd-4a19-b28b-eb95f51f4d00", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1095.210460] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-5101d24a-5b00-4587-8b37-5a45f8c4ea31 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 0361ae8f-634d-4161-a594-7f93dbb1b7cd] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d6:07:64', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '74f30339-6421-4654-bddb-81d7f34db9d7', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2bfd7147-87fd-4a19-b28b-eb95f51f4d00', 'vif_model': 'vmxnet3'}] {{(pid=65726) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1095.218425] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-5101d24a-5b00-4587-8b37-5a45f8c4ea31 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1095.219179] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1361a918-8944-4a91-bb20-2acca55dc4a4 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.222964] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0361ae8f-634d-4161-a594-7f93dbb1b7cd] Creating VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1095.226025] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0382390-db86-4f40-8c3b-a462aa92ccc7 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.226529] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ff9f4bec-44a4-4783-b9c9-5e20ea9eebdf {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.269878] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f78cd43-e461-47bc-b9f8-6f0301926b70 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.273542] env[65726]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1095.273542] env[65726]: value = "task-5116711" [ 1095.273542] env[65726]: _type = "Task" [ 1095.273542] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1095.274511] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-373bda1d-2da3-4d04-8d41-44bf8e2eea75 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.311065] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54ac93e1-c44f-4191-a671-ae84099e608b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.334290] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116711, 'name': CreateVM_Task} progress is 15%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1095.336779] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e6faf97-5282-490a-91ea-aa00f86fa370 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.340979] env[65726]: DEBUG nova.compute.manager [req-25423325-6f11-4f34-8f71-68313bb76b36 req-6d4804bd-89fe-4118-ac39-3cab4b6e1edb service nova] [instance: 0361ae8f-634d-4161-a594-7f93dbb1b7cd] Received event network-changed-2bfd7147-87fd-4a19-b28b-eb95f51f4d00 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1095.341360] env[65726]: DEBUG nova.compute.manager [req-25423325-6f11-4f34-8f71-68313bb76b36 req-6d4804bd-89fe-4118-ac39-3cab4b6e1edb service nova] [instance: 0361ae8f-634d-4161-a594-7f93dbb1b7cd] Refreshing instance network info cache due to event network-changed-2bfd7147-87fd-4a19-b28b-eb95f51f4d00. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 1095.341486] env[65726]: DEBUG oslo_concurrency.lockutils [req-25423325-6f11-4f34-8f71-68313bb76b36 req-6d4804bd-89fe-4118-ac39-3cab4b6e1edb service nova] Acquiring lock "refresh_cache-0361ae8f-634d-4161-a594-7f93dbb1b7cd" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1095.341679] env[65726]: DEBUG oslo_concurrency.lockutils [req-25423325-6f11-4f34-8f71-68313bb76b36 req-6d4804bd-89fe-4118-ac39-3cab4b6e1edb service nova] Acquired lock "refresh_cache-0361ae8f-634d-4161-a594-7f93dbb1b7cd" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1095.341679] env[65726]: DEBUG nova.network.neutron [req-25423325-6f11-4f34-8f71-68313bb76b36 req-6d4804bd-89fe-4118-ac39-3cab4b6e1edb service nova] [instance: 0361ae8f-634d-4161-a594-7f93dbb1b7cd] Refreshing network info cache for port 2bfd7147-87fd-4a19-b28b-eb95f51f4d00 {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 1095.348118] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc97bcad-db5d-4538-bee5-9f7cca991213 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.356806] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-14a9b6f0-9c85-4271-a1df-1d81c0fc4a98 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: a56be6fa-6027-46ee-9aa8-332fe4222d68] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1095.357662] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2e4a9675-258b-4ae7-ad4c-25555b0aaca4 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.370322] env[65726]: DEBUG nova.compute.provider_tree [None req-4f5aff8f-c14b-4955-a84c-9d59624f9d81 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1095.375159] env[65726]: DEBUG oslo_vmware.api [None req-14a9b6f0-9c85-4271-a1df-1d81c0fc4a98 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Waiting for the task: (returnval){ [ 1095.375159] env[65726]: value = "task-5116712" [ 1095.375159] env[65726]: _type = "Task" [ 1095.375159] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1095.385374] env[65726]: DEBUG oslo_vmware.api [None req-14a9b6f0-9c85-4271-a1df-1d81c0fc4a98 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5116712, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1095.510633] env[65726]: DEBUG nova.compute.manager [req-5ddca665-8ddd-45d9-8d4a-2490e4385a74 req-0e89414c-6eb7-429d-a8bb-684ef9d9cf44 service nova] [instance: 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04] Received event network-vif-deleted-25b9d108-0b43-4459-b9db-7bd90a495bb3 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1095.510879] env[65726]: INFO nova.compute.manager [req-5ddca665-8ddd-45d9-8d4a-2490e4385a74 req-0e89414c-6eb7-429d-a8bb-684ef9d9cf44 service nova] [instance: 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04] Neutron deleted interface 25b9d108-0b43-4459-b9db-7bd90a495bb3; detaching it from the instance and deleting it from the info cache [ 1095.511061] env[65726]: DEBUG nova.network.neutron [req-5ddca665-8ddd-45d9-8d4a-2490e4385a74 req-0e89414c-6eb7-429d-a8bb-684ef9d9cf44 service nova] [instance: 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1095.635887] env[65726]: DEBUG nova.network.neutron [-] [instance: 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1095.785355] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116711, 'name': CreateVM_Task, 'duration_secs': 0.385252} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1095.785541] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0361ae8f-634d-4161-a594-7f93dbb1b7cd] Created VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1095.786354] env[65726]: WARNING openstack [None req-5101d24a-5b00-4587-8b37-5a45f8c4ea31 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1095.786734] env[65726]: WARNING openstack [None req-5101d24a-5b00-4587-8b37-5a45f8c4ea31 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1095.791899] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5101d24a-5b00-4587-8b37-5a45f8c4ea31 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1095.792164] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5101d24a-5b00-4587-8b37-5a45f8c4ea31 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1095.792422] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5101d24a-5b00-4587-8b37-5a45f8c4ea31 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1095.792899] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a5be4f0d-5a3a-482a-ad92-2bac25bf5325 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.798689] env[65726]: DEBUG oslo_vmware.api [None req-5101d24a-5b00-4587-8b37-5a45f8c4ea31 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Waiting for the task: (returnval){ [ 1095.798689] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52600c42-3c2d-6079-8c65-ee894b7f7b42" [ 1095.798689] env[65726]: _type = "Task" [ 1095.798689] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1095.808700] env[65726]: DEBUG oslo_vmware.api [None req-5101d24a-5b00-4587-8b37-5a45f8c4ea31 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52600c42-3c2d-6079-8c65-ee894b7f7b42, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1095.847889] env[65726]: WARNING openstack [req-25423325-6f11-4f34-8f71-68313bb76b36 req-6d4804bd-89fe-4118-ac39-3cab4b6e1edb service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1095.848437] env[65726]: WARNING openstack [req-25423325-6f11-4f34-8f71-68313bb76b36 req-6d4804bd-89fe-4118-ac39-3cab4b6e1edb service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1095.874070] env[65726]: DEBUG nova.scheduler.client.report [None req-4f5aff8f-c14b-4955-a84c-9d59624f9d81 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1095.886845] env[65726]: DEBUG oslo_vmware.api [None req-14a9b6f0-9c85-4271-a1df-1d81c0fc4a98 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5116712, 'name': PowerOnVM_Task, 'duration_secs': 0.47554} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1095.891193] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-14a9b6f0-9c85-4271-a1df-1d81c0fc4a98 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: a56be6fa-6027-46ee-9aa8-332fe4222d68] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1095.956420] env[65726]: WARNING openstack [req-25423325-6f11-4f34-8f71-68313bb76b36 req-6d4804bd-89fe-4118-ac39-3cab4b6e1edb service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1095.956893] env[65726]: WARNING openstack [req-25423325-6f11-4f34-8f71-68313bb76b36 req-6d4804bd-89fe-4118-ac39-3cab4b6e1edb service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1096.015942] env[65726]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-73e156f2-1f55-47c4-af78-cf994c115957 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.025999] env[65726]: WARNING openstack [req-25423325-6f11-4f34-8f71-68313bb76b36 req-6d4804bd-89fe-4118-ac39-3cab4b6e1edb service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1096.026505] env[65726]: WARNING openstack [req-25423325-6f11-4f34-8f71-68313bb76b36 req-6d4804bd-89fe-4118-ac39-3cab4b6e1edb service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1096.040906] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a60b292b-76fd-4201-b26b-dff7f78e0b34 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.062414] env[65726]: DEBUG oslo_concurrency.lockutils [None req-1c19f444-aee2-49ca-b67e-d3cef83d3664 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Acquiring lock "refresh_cache-6fd70b4a-63e4-4258-9cff-f3b582500b9f" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1096.062561] env[65726]: DEBUG oslo_concurrency.lockutils [None req-1c19f444-aee2-49ca-b67e-d3cef83d3664 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Acquired lock "refresh_cache-6fd70b4a-63e4-4258-9cff-f3b582500b9f" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1096.062898] env[65726]: DEBUG nova.network.neutron [None req-1c19f444-aee2-49ca-b67e-d3cef83d3664 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] [instance: 6fd70b4a-63e4-4258-9cff-f3b582500b9f] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1096.076847] env[65726]: DEBUG nova.compute.manager [req-5ddca665-8ddd-45d9-8d4a-2490e4385a74 req-0e89414c-6eb7-429d-a8bb-684ef9d9cf44 service nova] [instance: 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04] Detach interface failed, port_id=25b9d108-0b43-4459-b9db-7bd90a495bb3, reason: Instance 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04 could not be found. {{(pid=65726) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11669}} [ 1096.079412] env[65726]: WARNING openstack [None req-1c19f444-aee2-49ca-b67e-d3cef83d3664 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1096.079570] env[65726]: WARNING openstack [None req-1c19f444-aee2-49ca-b67e-d3cef83d3664 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1096.128197] env[65726]: DEBUG nova.network.neutron [req-25423325-6f11-4f34-8f71-68313bb76b36 req-6d4804bd-89fe-4118-ac39-3cab4b6e1edb service nova] [instance: 0361ae8f-634d-4161-a594-7f93dbb1b7cd] Updated VIF entry in instance network info cache for port 2bfd7147-87fd-4a19-b28b-eb95f51f4d00. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 1096.128829] env[65726]: DEBUG nova.network.neutron [req-25423325-6f11-4f34-8f71-68313bb76b36 req-6d4804bd-89fe-4118-ac39-3cab4b6e1edb service nova] [instance: 0361ae8f-634d-4161-a594-7f93dbb1b7cd] Updating instance_info_cache with network_info: [{"id": "2bfd7147-87fd-4a19-b28b-eb95f51f4d00", "address": "fa:16:3e:d6:07:64", "network": {"id": "9dd55fb6-0dd2-484d-b94d-fbdccb958c79", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1005598846-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f93d852e2c904f42981cbfff882558e9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "74f30339-6421-4654-bddb-81d7f34db9d7", "external-id": "nsx-vlan-transportzone-899", "segmentation_id": 899, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2bfd7147-87", "ovs_interfaceid": "2bfd7147-87fd-4a19-b28b-eb95f51f4d00", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1096.139655] env[65726]: INFO nova.compute.manager [-] [instance: 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04] Took 1.48 seconds to deallocate network for instance. [ 1096.177554] env[65726]: WARNING openstack [None req-1c19f444-aee2-49ca-b67e-d3cef83d3664 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1096.177928] env[65726]: WARNING openstack [None req-1c19f444-aee2-49ca-b67e-d3cef83d3664 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1096.248163] env[65726]: WARNING openstack [None req-1c19f444-aee2-49ca-b67e-d3cef83d3664 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1096.248163] env[65726]: WARNING openstack [None req-1c19f444-aee2-49ca-b67e-d3cef83d3664 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1096.309831] env[65726]: DEBUG oslo_vmware.api [None req-5101d24a-5b00-4587-8b37-5a45f8c4ea31 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52600c42-3c2d-6079-8c65-ee894b7f7b42, 'name': SearchDatastore_Task, 'duration_secs': 0.011265} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1096.310271] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5101d24a-5b00-4587-8b37-5a45f8c4ea31 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1096.310552] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-5101d24a-5b00-4587-8b37-5a45f8c4ea31 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 0361ae8f-634d-4161-a594-7f93dbb1b7cd] Processing image b52362a3-ee8a-4cbf-b06f-513b0cc8f95c {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1096.310761] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5101d24a-5b00-4587-8b37-5a45f8c4ea31 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1096.310903] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5101d24a-5b00-4587-8b37-5a45f8c4ea31 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1096.311094] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-5101d24a-5b00-4587-8b37-5a45f8c4ea31 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1096.311397] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-afa69f02-9f36-4f7a-8a14-f49410f49a95 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.325186] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-5101d24a-5b00-4587-8b37-5a45f8c4ea31 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1096.325426] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-5101d24a-5b00-4587-8b37-5a45f8c4ea31 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65726) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1096.326216] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-96ad7c0d-c3d4-4ded-82b4-a4f694d2a0dc {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.331954] env[65726]: DEBUG oslo_vmware.api [None req-5101d24a-5b00-4587-8b37-5a45f8c4ea31 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Waiting for the task: (returnval){ [ 1096.331954] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52331b26-5fd8-17d1-2c27-084ff7456871" [ 1096.331954] env[65726]: _type = "Task" [ 1096.331954] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1096.334362] env[65726]: DEBUG nova.network.neutron [None req-1c19f444-aee2-49ca-b67e-d3cef83d3664 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] [instance: 6fd70b4a-63e4-4258-9cff-f3b582500b9f] Updating instance_info_cache with network_info: [{"id": "153b4c9d-d01d-4254-9aa6-040705be347a", "address": "fa:16:3e:17:18:d9", "network": {"id": "93e38d7e-b4d7-4d4a-8393-64ab2f6769de", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1445539922-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "06e221dc693640929cac7bc5af45948f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "399f3826-705c-45f7-9fe0-3a08a945151a", "external-id": "nsx-vlan-transportzone-936", "segmentation_id": 936, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap153b4c9d-d0", "ovs_interfaceid": "153b4c9d-d01d-4254-9aa6-040705be347a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1096.347665] env[65726]: DEBUG oslo_vmware.api [None req-5101d24a-5b00-4587-8b37-5a45f8c4ea31 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52331b26-5fd8-17d1-2c27-084ff7456871, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1096.382359] env[65726]: DEBUG oslo_concurrency.lockutils [None req-4f5aff8f-c14b-4955-a84c-9d59624f9d81 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.883s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1096.384983] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5eb7fa6b-cbd6-4a47-b227-e9c16f779f4a tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.531s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1096.386529] env[65726]: INFO nova.compute.claims [None req-5eb7fa6b-cbd6-4a47-b227-e9c16f779f4a tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] [instance: 4809fcf6-59d4-409c-bd8d-981dc85a686f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1096.411717] env[65726]: INFO nova.scheduler.client.report [None req-4f5aff8f-c14b-4955-a84c-9d59624f9d81 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Deleted allocations for instance 5cdf1d61-b817-4986-a48c-804a2374497a [ 1096.631901] env[65726]: DEBUG oslo_concurrency.lockutils [req-25423325-6f11-4f34-8f71-68313bb76b36 req-6d4804bd-89fe-4118-ac39-3cab4b6e1edb service nova] Releasing lock "refresh_cache-0361ae8f-634d-4161-a594-7f93dbb1b7cd" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1096.647806] env[65726]: DEBUG oslo_concurrency.lockutils [None req-eab5602f-7673-4d76-aa16-7ebf0d2c87e3 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1096.839751] env[65726]: DEBUG oslo_concurrency.lockutils [None req-1c19f444-aee2-49ca-b67e-d3cef83d3664 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Releasing lock "refresh_cache-6fd70b4a-63e4-4258-9cff-f3b582500b9f" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1096.848388] env[65726]: DEBUG oslo_vmware.api [None req-5101d24a-5b00-4587-8b37-5a45f8c4ea31 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52331b26-5fd8-17d1-2c27-084ff7456871, 'name': SearchDatastore_Task, 'duration_secs': 0.017599} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1096.849212] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-764d52f5-c018-4658-8c54-0d4b6b0ae883 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.854720] env[65726]: DEBUG oslo_vmware.api [None req-5101d24a-5b00-4587-8b37-5a45f8c4ea31 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Waiting for the task: (returnval){ [ 1096.854720] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]526544ef-82e4-4743-61ab-e3f177b9434f" [ 1096.854720] env[65726]: _type = "Task" [ 1096.854720] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1096.862697] env[65726]: DEBUG oslo_vmware.api [None req-5101d24a-5b00-4587-8b37-5a45f8c4ea31 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]526544ef-82e4-4743-61ab-e3f177b9434f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1096.871021] env[65726]: DEBUG nova.virt.hardware [None req-1c19f444-aee2-49ca-b67e-d3cef83d3664 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='ab1fe94413c8fc331a65bd267a012f1d',container_format='bare',created_at=2025-12-12T19:38:43Z,direct_url=,disk_format='vmdk',id=eb421501-6112-4a68-9c48-c82505e774d2,min_disk=1,min_ram=0,name='tempest-ServersNegativeTestJSON-server-2007878152-shelved',owner='06e221dc693640929cac7bc5af45948f',properties=ImageMetaProps,protected=,size=31667200,status='active',tags=,updated_at=2025-12-12T19:38:59Z,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1096.871272] env[65726]: DEBUG nova.virt.hardware [None req-1c19f444-aee2-49ca-b67e-d3cef83d3664 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1096.871429] env[65726]: DEBUG nova.virt.hardware [None req-1c19f444-aee2-49ca-b67e-d3cef83d3664 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1096.871603] env[65726]: DEBUG nova.virt.hardware [None req-1c19f444-aee2-49ca-b67e-d3cef83d3664 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1096.871745] env[65726]: DEBUG nova.virt.hardware [None req-1c19f444-aee2-49ca-b67e-d3cef83d3664 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1096.871888] env[65726]: DEBUG nova.virt.hardware [None req-1c19f444-aee2-49ca-b67e-d3cef83d3664 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1096.872106] env[65726]: DEBUG nova.virt.hardware [None req-1c19f444-aee2-49ca-b67e-d3cef83d3664 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1096.872263] env[65726]: DEBUG nova.virt.hardware [None req-1c19f444-aee2-49ca-b67e-d3cef83d3664 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1096.872435] env[65726]: DEBUG nova.virt.hardware [None req-1c19f444-aee2-49ca-b67e-d3cef83d3664 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1096.872677] env[65726]: DEBUG nova.virt.hardware [None req-1c19f444-aee2-49ca-b67e-d3cef83d3664 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1096.872855] env[65726]: DEBUG nova.virt.hardware [None req-1c19f444-aee2-49ca-b67e-d3cef83d3664 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1096.873684] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a2b8238-7e79-4d9c-aba7-c90ec3cf6e01 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.881456] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aead4be3-0582-4b1a-9f87-53fdc477ce28 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.897587] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-1c19f444-aee2-49ca-b67e-d3cef83d3664 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] [instance: 6fd70b4a-63e4-4258-9cff-f3b582500b9f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:17:18:d9', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '399f3826-705c-45f7-9fe0-3a08a945151a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '153b4c9d-d01d-4254-9aa6-040705be347a', 'vif_model': 'vmxnet3'}] {{(pid=65726) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1096.904913] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-1c19f444-aee2-49ca-b67e-d3cef83d3664 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1096.908804] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6fd70b4a-63e4-4258-9cff-f3b582500b9f] Creating VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1096.910479] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9d4ed1d6-69ed-447d-84c4-d6631a67c15e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.930630] env[65726]: DEBUG oslo_concurrency.lockutils [None req-4f5aff8f-c14b-4955-a84c-9d59624f9d81 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Lock "5cdf1d61-b817-4986-a48c-804a2374497a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 13.394s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1096.936661] env[65726]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1096.936661] env[65726]: value = "task-5116713" [ 1096.936661] env[65726]: _type = "Task" [ 1096.936661] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1096.946455] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116713, 'name': CreateVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1096.947486] env[65726]: INFO nova.compute.manager [None req-14a9b6f0-9c85-4271-a1df-1d81c0fc4a98 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: a56be6fa-6027-46ee-9aa8-332fe4222d68] Updating instance to original state: 'active' [ 1097.364301] env[65726]: DEBUG nova.compute.manager [req-36e9a068-90ac-4bf7-baef-ed0fe02cb856 req-1741adb2-124c-4a22-9886-36c52f579b78 service nova] [instance: 6fd70b4a-63e4-4258-9cff-f3b582500b9f] Received event network-vif-plugged-153b4c9d-d01d-4254-9aa6-040705be347a {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1097.364728] env[65726]: DEBUG oslo_concurrency.lockutils [req-36e9a068-90ac-4bf7-baef-ed0fe02cb856 req-1741adb2-124c-4a22-9886-36c52f579b78 service nova] Acquiring lock "6fd70b4a-63e4-4258-9cff-f3b582500b9f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1097.364728] env[65726]: DEBUG oslo_concurrency.lockutils [req-36e9a068-90ac-4bf7-baef-ed0fe02cb856 req-1741adb2-124c-4a22-9886-36c52f579b78 service nova] Lock "6fd70b4a-63e4-4258-9cff-f3b582500b9f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1097.364957] env[65726]: DEBUG oslo_concurrency.lockutils [req-36e9a068-90ac-4bf7-baef-ed0fe02cb856 req-1741adb2-124c-4a22-9886-36c52f579b78 service nova] Lock "6fd70b4a-63e4-4258-9cff-f3b582500b9f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1097.365090] env[65726]: DEBUG nova.compute.manager [req-36e9a068-90ac-4bf7-baef-ed0fe02cb856 req-1741adb2-124c-4a22-9886-36c52f579b78 service nova] [instance: 6fd70b4a-63e4-4258-9cff-f3b582500b9f] No waiting events found dispatching network-vif-plugged-153b4c9d-d01d-4254-9aa6-040705be347a {{(pid=65726) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1097.365282] env[65726]: WARNING nova.compute.manager [req-36e9a068-90ac-4bf7-baef-ed0fe02cb856 req-1741adb2-124c-4a22-9886-36c52f579b78 service nova] [instance: 6fd70b4a-63e4-4258-9cff-f3b582500b9f] Received unexpected event network-vif-plugged-153b4c9d-d01d-4254-9aa6-040705be347a for instance with vm_state shelved_offloaded and task_state spawning. [ 1097.365441] env[65726]: DEBUG nova.compute.manager [req-36e9a068-90ac-4bf7-baef-ed0fe02cb856 req-1741adb2-124c-4a22-9886-36c52f579b78 service nova] [instance: 6fd70b4a-63e4-4258-9cff-f3b582500b9f] Received event network-changed-153b4c9d-d01d-4254-9aa6-040705be347a {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1097.365592] env[65726]: DEBUG nova.compute.manager [req-36e9a068-90ac-4bf7-baef-ed0fe02cb856 req-1741adb2-124c-4a22-9886-36c52f579b78 service nova] [instance: 6fd70b4a-63e4-4258-9cff-f3b582500b9f] Refreshing instance network info cache due to event network-changed-153b4c9d-d01d-4254-9aa6-040705be347a. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 1097.365834] env[65726]: DEBUG oslo_concurrency.lockutils [req-36e9a068-90ac-4bf7-baef-ed0fe02cb856 req-1741adb2-124c-4a22-9886-36c52f579b78 service nova] Acquiring lock "refresh_cache-6fd70b4a-63e4-4258-9cff-f3b582500b9f" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1097.365988] env[65726]: DEBUG oslo_concurrency.lockutils [req-36e9a068-90ac-4bf7-baef-ed0fe02cb856 req-1741adb2-124c-4a22-9886-36c52f579b78 service nova] Acquired lock "refresh_cache-6fd70b4a-63e4-4258-9cff-f3b582500b9f" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1097.366194] env[65726]: DEBUG nova.network.neutron [req-36e9a068-90ac-4bf7-baef-ed0fe02cb856 req-1741adb2-124c-4a22-9886-36c52f579b78 service nova] [instance: 6fd70b4a-63e4-4258-9cff-f3b582500b9f] Refreshing network info cache for port 153b4c9d-d01d-4254-9aa6-040705be347a {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 1097.371257] env[65726]: DEBUG oslo_vmware.api [None req-5101d24a-5b00-4587-8b37-5a45f8c4ea31 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]526544ef-82e4-4743-61ab-e3f177b9434f, 'name': SearchDatastore_Task, 'duration_secs': 0.042249} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1097.372046] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5101d24a-5b00-4587-8b37-5a45f8c4ea31 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1097.372046] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-5101d24a-5b00-4587-8b37-5a45f8c4ea31 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] 0361ae8f-634d-4161-a594-7f93dbb1b7cd/0361ae8f-634d-4161-a594-7f93dbb1b7cd.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1097.372271] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f7ef18cb-73ae-4595-86f7-a246c34255f3 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.381116] env[65726]: DEBUG oslo_vmware.api [None req-5101d24a-5b00-4587-8b37-5a45f8c4ea31 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Waiting for the task: (returnval){ [ 1097.381116] env[65726]: value = "task-5116714" [ 1097.381116] env[65726]: _type = "Task" [ 1097.381116] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1097.392334] env[65726]: DEBUG oslo_vmware.api [None req-5101d24a-5b00-4587-8b37-5a45f8c4ea31 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': task-5116714, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1097.450786] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116713, 'name': CreateVM_Task, 'duration_secs': 0.452476} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1097.450918] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6fd70b4a-63e4-4258-9cff-f3b582500b9f] Created VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1097.451695] env[65726]: WARNING openstack [None req-1c19f444-aee2-49ca-b67e-d3cef83d3664 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1097.452069] env[65726]: WARNING openstack [None req-1c19f444-aee2-49ca-b67e-d3cef83d3664 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1097.457100] env[65726]: DEBUG oslo_concurrency.lockutils [None req-1c19f444-aee2-49ca-b67e-d3cef83d3664 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/eb421501-6112-4a68-9c48-c82505e774d2" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1097.457281] env[65726]: DEBUG oslo_concurrency.lockutils [None req-1c19f444-aee2-49ca-b67e-d3cef83d3664 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Acquired lock "[datastore1] devstack-image-cache_base/eb421501-6112-4a68-9c48-c82505e774d2" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1097.458531] env[65726]: DEBUG oslo_concurrency.lockutils [None req-1c19f444-aee2-49ca-b67e-d3cef83d3664 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/eb421501-6112-4a68-9c48-c82505e774d2" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1097.463766] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8968560f-6f38-4a33-ac98-1828d21bb6d6 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.471918] env[65726]: DEBUG oslo_vmware.api [None req-1c19f444-aee2-49ca-b67e-d3cef83d3664 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Waiting for the task: (returnval){ [ 1097.471918] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52bca6a4-4e1c-efe6-6439-b4b5148ad4a3" [ 1097.471918] env[65726]: _type = "Task" [ 1097.471918] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1097.481408] env[65726]: DEBUG oslo_vmware.api [None req-1c19f444-aee2-49ca-b67e-d3cef83d3664 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52bca6a4-4e1c-efe6-6439-b4b5148ad4a3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1097.631423] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b2f210d-73ba-487f-abd6-92d777c72cda {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.639943] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3fac310-f632-4553-b5a2-ea092721072e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.673450] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40afaf3b-2485-47db-b552-ec10c42c9a84 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.682416] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89e6a715-a70c-4419-8a70-c2a64b610523 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.699373] env[65726]: DEBUG nova.compute.provider_tree [None req-5eb7fa6b-cbd6-4a47-b227-e9c16f779f4a tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1097.873635] env[65726]: WARNING openstack [req-36e9a068-90ac-4bf7-baef-ed0fe02cb856 req-1741adb2-124c-4a22-9886-36c52f579b78 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1097.874180] env[65726]: WARNING openstack [req-36e9a068-90ac-4bf7-baef-ed0fe02cb856 req-1741adb2-124c-4a22-9886-36c52f579b78 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1097.898204] env[65726]: DEBUG oslo_vmware.api [None req-5101d24a-5b00-4587-8b37-5a45f8c4ea31 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': task-5116714, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.503894} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1097.898760] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-5101d24a-5b00-4587-8b37-5a45f8c4ea31 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] 0361ae8f-634d-4161-a594-7f93dbb1b7cd/0361ae8f-634d-4161-a594-7f93dbb1b7cd.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1097.898951] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-5101d24a-5b00-4587-8b37-5a45f8c4ea31 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 0361ae8f-634d-4161-a594-7f93dbb1b7cd] Extending root virtual disk to 1048576 {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1097.899201] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ecf0a97b-1ec7-4f36-abcc-6e144d2410ba {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.911273] env[65726]: DEBUG oslo_vmware.api [None req-5101d24a-5b00-4587-8b37-5a45f8c4ea31 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Waiting for the task: (returnval){ [ 1097.911273] env[65726]: value = "task-5116715" [ 1097.911273] env[65726]: _type = "Task" [ 1097.911273] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1097.926645] env[65726]: DEBUG oslo_vmware.api [None req-5101d24a-5b00-4587-8b37-5a45f8c4ea31 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': task-5116715, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1097.984563] env[65726]: DEBUG oslo_concurrency.lockutils [None req-1c19f444-aee2-49ca-b67e-d3cef83d3664 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Releasing lock "[datastore1] devstack-image-cache_base/eb421501-6112-4a68-9c48-c82505e774d2" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1097.984868] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-1c19f444-aee2-49ca-b67e-d3cef83d3664 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] [instance: 6fd70b4a-63e4-4258-9cff-f3b582500b9f] Processing image eb421501-6112-4a68-9c48-c82505e774d2 {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1097.985156] env[65726]: DEBUG oslo_concurrency.lockutils [None req-1c19f444-aee2-49ca-b67e-d3cef83d3664 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/eb421501-6112-4a68-9c48-c82505e774d2/eb421501-6112-4a68-9c48-c82505e774d2.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1097.985334] env[65726]: DEBUG oslo_concurrency.lockutils [None req-1c19f444-aee2-49ca-b67e-d3cef83d3664 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Acquired lock "[datastore1] devstack-image-cache_base/eb421501-6112-4a68-9c48-c82505e774d2/eb421501-6112-4a68-9c48-c82505e774d2.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1097.985526] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-1c19f444-aee2-49ca-b67e-d3cef83d3664 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1097.986168] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3599fe6b-0771-4fbb-8487-191810555e01 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.989546] env[65726]: WARNING openstack [req-36e9a068-90ac-4bf7-baef-ed0fe02cb856 req-1741adb2-124c-4a22-9886-36c52f579b78 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1097.990086] env[65726]: WARNING openstack [req-36e9a068-90ac-4bf7-baef-ed0fe02cb856 req-1741adb2-124c-4a22-9886-36c52f579b78 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1098.007344] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-1c19f444-aee2-49ca-b67e-d3cef83d3664 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1098.007540] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-1c19f444-aee2-49ca-b67e-d3cef83d3664 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65726) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1098.008781] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-168a28aa-d316-4ba2-9acb-fe8646778012 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.017014] env[65726]: DEBUG oslo_vmware.api [None req-1c19f444-aee2-49ca-b67e-d3cef83d3664 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Waiting for the task: (returnval){ [ 1098.017014] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5209e010-0ad8-75e0-c916-67f47a2e1a24" [ 1098.017014] env[65726]: _type = "Task" [ 1098.017014] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1098.035330] env[65726]: DEBUG oslo_vmware.api [None req-1c19f444-aee2-49ca-b67e-d3cef83d3664 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5209e010-0ad8-75e0-c916-67f47a2e1a24, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1098.067268] env[65726]: WARNING openstack [req-36e9a068-90ac-4bf7-baef-ed0fe02cb856 req-1741adb2-124c-4a22-9886-36c52f579b78 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1098.067731] env[65726]: WARNING openstack [req-36e9a068-90ac-4bf7-baef-ed0fe02cb856 req-1741adb2-124c-4a22-9886-36c52f579b78 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1098.108624] env[65726]: DEBUG nova.objects.instance [None req-b56a2566-f325-47b4-827c-75797fd41834 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] Lazy-loading 'flavor' on Instance uuid df08d166-7811-49b5-9c16-a3434229d623 {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1098.156539] env[65726]: DEBUG nova.network.neutron [req-36e9a068-90ac-4bf7-baef-ed0fe02cb856 req-1741adb2-124c-4a22-9886-36c52f579b78 service nova] [instance: 6fd70b4a-63e4-4258-9cff-f3b582500b9f] Updated VIF entry in instance network info cache for port 153b4c9d-d01d-4254-9aa6-040705be347a. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 1098.158242] env[65726]: DEBUG nova.network.neutron [req-36e9a068-90ac-4bf7-baef-ed0fe02cb856 req-1741adb2-124c-4a22-9886-36c52f579b78 service nova] [instance: 6fd70b4a-63e4-4258-9cff-f3b582500b9f] Updating instance_info_cache with network_info: [{"id": "153b4c9d-d01d-4254-9aa6-040705be347a", "address": "fa:16:3e:17:18:d9", "network": {"id": "93e38d7e-b4d7-4d4a-8393-64ab2f6769de", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1445539922-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "06e221dc693640929cac7bc5af45948f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "399f3826-705c-45f7-9fe0-3a08a945151a", "external-id": "nsx-vlan-transportzone-936", "segmentation_id": 936, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap153b4c9d-d0", "ovs_interfaceid": "153b4c9d-d01d-4254-9aa6-040705be347a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1098.202743] env[65726]: DEBUG nova.scheduler.client.report [None req-5eb7fa6b-cbd6-4a47-b227-e9c16f779f4a tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1098.422632] env[65726]: DEBUG oslo_vmware.api [None req-5101d24a-5b00-4587-8b37-5a45f8c4ea31 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': task-5116715, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.063265} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1098.422632] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-5101d24a-5b00-4587-8b37-5a45f8c4ea31 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 0361ae8f-634d-4161-a594-7f93dbb1b7cd] Extended root virtual disk {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1098.423528] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aead344c-b24e-43bf-878e-c9101d8af533 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.449120] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-5101d24a-5b00-4587-8b37-5a45f8c4ea31 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 0361ae8f-634d-4161-a594-7f93dbb1b7cd] Reconfiguring VM instance instance-00000064 to attach disk [datastore1] 0361ae8f-634d-4161-a594-7f93dbb1b7cd/0361ae8f-634d-4161-a594-7f93dbb1b7cd.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1098.449434] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ff8621af-e3bf-4007-b21e-5ec17d37bb17 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.470534] env[65726]: DEBUG oslo_vmware.api [None req-5101d24a-5b00-4587-8b37-5a45f8c4ea31 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Waiting for the task: (returnval){ [ 1098.470534] env[65726]: value = "task-5116716" [ 1098.470534] env[65726]: _type = "Task" [ 1098.470534] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1098.479416] env[65726]: DEBUG oslo_vmware.api [None req-5101d24a-5b00-4587-8b37-5a45f8c4ea31 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': task-5116716, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1098.528852] env[65726]: DEBUG oslo_concurrency.lockutils [None req-0288a821-ce9c-40cd-86ef-962ac009580c tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Acquiring lock "a56be6fa-6027-46ee-9aa8-332fe4222d68" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1098.529252] env[65726]: DEBUG oslo_concurrency.lockutils [None req-0288a821-ce9c-40cd-86ef-962ac009580c tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Lock "a56be6fa-6027-46ee-9aa8-332fe4222d68" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1098.529802] env[65726]: DEBUG oslo_concurrency.lockutils [None req-0288a821-ce9c-40cd-86ef-962ac009580c tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Acquiring lock "a56be6fa-6027-46ee-9aa8-332fe4222d68-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1098.529919] env[65726]: DEBUG oslo_concurrency.lockutils [None req-0288a821-ce9c-40cd-86ef-962ac009580c tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Lock "a56be6fa-6027-46ee-9aa8-332fe4222d68-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1098.530145] env[65726]: DEBUG oslo_concurrency.lockutils [None req-0288a821-ce9c-40cd-86ef-962ac009580c tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Lock "a56be6fa-6027-46ee-9aa8-332fe4222d68-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1098.532257] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-1c19f444-aee2-49ca-b67e-d3cef83d3664 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] [instance: 6fd70b4a-63e4-4258-9cff-f3b582500b9f] Preparing fetch location {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 1098.532500] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-1c19f444-aee2-49ca-b67e-d3cef83d3664 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] [instance: 6fd70b4a-63e4-4258-9cff-f3b582500b9f] Fetch image to [datastore1] OSTACK_IMG_5f557822-0d62-4142-9f38-0161dfdccbb6/OSTACK_IMG_5f557822-0d62-4142-9f38-0161dfdccbb6.vmdk {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 1098.532733] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-1c19f444-aee2-49ca-b67e-d3cef83d3664 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] [instance: 6fd70b4a-63e4-4258-9cff-f3b582500b9f] Downloading stream optimized image eb421501-6112-4a68-9c48-c82505e774d2 to [datastore1] OSTACK_IMG_5f557822-0d62-4142-9f38-0161dfdccbb6/OSTACK_IMG_5f557822-0d62-4142-9f38-0161dfdccbb6.vmdk on the data store datastore1 as vApp {{(pid=65726) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 1098.532963] env[65726]: DEBUG nova.virt.vmwareapi.images [None req-1c19f444-aee2-49ca-b67e-d3cef83d3664 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] [instance: 6fd70b4a-63e4-4258-9cff-f3b582500b9f] Downloading image file data eb421501-6112-4a68-9c48-c82505e774d2 to the ESX as VM named 'OSTACK_IMG_5f557822-0d62-4142-9f38-0161dfdccbb6' {{(pid=65726) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 1098.537060] env[65726]: INFO nova.compute.manager [None req-0288a821-ce9c-40cd-86ef-962ac009580c tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: a56be6fa-6027-46ee-9aa8-332fe4222d68] Terminating instance [ 1098.615314] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b56a2566-f325-47b4-827c-75797fd41834 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] Acquiring lock "refresh_cache-df08d166-7811-49b5-9c16-a3434229d623" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1098.615492] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b56a2566-f325-47b4-827c-75797fd41834 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] Acquired lock "refresh_cache-df08d166-7811-49b5-9c16-a3434229d623" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1098.616201] env[65726]: WARNING openstack [None req-b56a2566-f325-47b4-827c-75797fd41834 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1098.616564] env[65726]: WARNING openstack [None req-b56a2566-f325-47b4-827c-75797fd41834 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1098.625458] env[65726]: DEBUG oslo_vmware.rw_handles [None req-1c19f444-aee2-49ca-b67e-d3cef83d3664 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 1098.625458] env[65726]: value = "resgroup-9" [ 1098.625458] env[65726]: _type = "ResourcePool" [ 1098.625458] env[65726]: }. {{(pid=65726) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 1098.625984] env[65726]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-c2ebca59-07f8-4276-b2f9-b30931604ff0 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.649542] env[65726]: DEBUG oslo_vmware.rw_handles [None req-1c19f444-aee2-49ca-b67e-d3cef83d3664 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Lease: (returnval){ [ 1098.649542] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52041259-6206-4882-dea0-00fc5d492a1e" [ 1098.649542] env[65726]: _type = "HttpNfcLease" [ 1098.649542] env[65726]: } obtained for vApp import into resource pool (val){ [ 1098.649542] env[65726]: value = "resgroup-9" [ 1098.649542] env[65726]: _type = "ResourcePool" [ 1098.649542] env[65726]: }. {{(pid=65726) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 1098.649850] env[65726]: DEBUG oslo_vmware.api [None req-1c19f444-aee2-49ca-b67e-d3cef83d3664 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Waiting for the lease: (returnval){ [ 1098.649850] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52041259-6206-4882-dea0-00fc5d492a1e" [ 1098.649850] env[65726]: _type = "HttpNfcLease" [ 1098.649850] env[65726]: } to be ready. {{(pid=65726) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1098.657307] env[65726]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1098.657307] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52041259-6206-4882-dea0-00fc5d492a1e" [ 1098.657307] env[65726]: _type = "HttpNfcLease" [ 1098.657307] env[65726]: } is initializing. {{(pid=65726) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1098.661082] env[65726]: DEBUG oslo_concurrency.lockutils [req-36e9a068-90ac-4bf7-baef-ed0fe02cb856 req-1741adb2-124c-4a22-9886-36c52f579b78 service nova] Releasing lock "refresh_cache-6fd70b4a-63e4-4258-9cff-f3b582500b9f" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1098.708857] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5eb7fa6b-cbd6-4a47-b227-e9c16f779f4a tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.324s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1098.709446] env[65726]: DEBUG nova.compute.manager [None req-5eb7fa6b-cbd6-4a47-b227-e9c16f779f4a tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] [instance: 4809fcf6-59d4-409c-bd8d-981dc85a686f] Start building networks asynchronously for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1098.712783] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a6f4e31f-bf92-46ff-ad83-0dbcdb2a78cf tempest-InstanceActionsV221TestJSON-647428531 tempest-InstanceActionsV221TestJSON-647428531-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 9.341s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1098.712783] env[65726]: DEBUG nova.objects.instance [None req-a6f4e31f-bf92-46ff-ad83-0dbcdb2a78cf tempest-InstanceActionsV221TestJSON-647428531 tempest-InstanceActionsV221TestJSON-647428531-project-member] Lazy-loading 'resources' on Instance uuid b0fec5ff-8c1e-41f9-b30a-6253a34baeec {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1098.987467] env[65726]: DEBUG oslo_vmware.api [None req-5101d24a-5b00-4587-8b37-5a45f8c4ea31 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': task-5116716, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1099.002151] env[65726]: DEBUG nova.network.neutron [None req-b56a2566-f325-47b4-827c-75797fd41834 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] [instance: df08d166-7811-49b5-9c16-a3434229d623] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1099.033098] env[65726]: DEBUG nova.compute.manager [req-05b43d10-3696-4fd8-8633-a35327ca1b02 req-0cd452a9-d0fc-4ddd-bb22-423921d86061 service nova] [instance: df08d166-7811-49b5-9c16-a3434229d623] Received event network-changed-70efdc7d-7775-4990-b851-d60195bd1504 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1099.033302] env[65726]: DEBUG nova.compute.manager [req-05b43d10-3696-4fd8-8633-a35327ca1b02 req-0cd452a9-d0fc-4ddd-bb22-423921d86061 service nova] [instance: df08d166-7811-49b5-9c16-a3434229d623] Refreshing instance network info cache due to event network-changed-70efdc7d-7775-4990-b851-d60195bd1504. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 1099.033597] env[65726]: DEBUG oslo_concurrency.lockutils [req-05b43d10-3696-4fd8-8633-a35327ca1b02 req-0cd452a9-d0fc-4ddd-bb22-423921d86061 service nova] Acquiring lock "refresh_cache-df08d166-7811-49b5-9c16-a3434229d623" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1099.046529] env[65726]: DEBUG nova.compute.manager [None req-0288a821-ce9c-40cd-86ef-962ac009580c tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: a56be6fa-6027-46ee-9aa8-332fe4222d68] Start destroying the instance on the hypervisor. {{(pid=65726) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1099.046719] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-0288a821-ce9c-40cd-86ef-962ac009580c tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: a56be6fa-6027-46ee-9aa8-332fe4222d68] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1099.046994] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-cc92977b-eb3e-4b0e-bef3-a1ab47c321f7 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.055259] env[65726]: DEBUG oslo_vmware.api [None req-0288a821-ce9c-40cd-86ef-962ac009580c tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Waiting for the task: (returnval){ [ 1099.055259] env[65726]: value = "task-5116718" [ 1099.055259] env[65726]: _type = "Task" [ 1099.055259] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1099.065596] env[65726]: DEBUG oslo_vmware.api [None req-0288a821-ce9c-40cd-86ef-962ac009580c tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5116718, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1099.159340] env[65726]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1099.159340] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52041259-6206-4882-dea0-00fc5d492a1e" [ 1099.159340] env[65726]: _type = "HttpNfcLease" [ 1099.159340] env[65726]: } is initializing. {{(pid=65726) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1099.216251] env[65726]: DEBUG nova.compute.utils [None req-5eb7fa6b-cbd6-4a47-b227-e9c16f779f4a tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Using /dev/sd instead of None {{(pid=65726) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1099.220069] env[65726]: DEBUG nova.compute.manager [None req-5eb7fa6b-cbd6-4a47-b227-e9c16f779f4a tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] [instance: 4809fcf6-59d4-409c-bd8d-981dc85a686f] Allocating IP information in the background. {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1099.220311] env[65726]: DEBUG nova.network.neutron [None req-5eb7fa6b-cbd6-4a47-b227-e9c16f779f4a tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] [instance: 4809fcf6-59d4-409c-bd8d-981dc85a686f] allocate_for_instance() {{(pid=65726) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 1099.220733] env[65726]: WARNING neutronclient.v2_0.client [None req-5eb7fa6b-cbd6-4a47-b227-e9c16f779f4a tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1099.222121] env[65726]: WARNING neutronclient.v2_0.client [None req-5eb7fa6b-cbd6-4a47-b227-e9c16f779f4a tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1099.222121] env[65726]: WARNING openstack [None req-5eb7fa6b-cbd6-4a47-b227-e9c16f779f4a tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1099.222121] env[65726]: WARNING openstack [None req-5eb7fa6b-cbd6-4a47-b227-e9c16f779f4a tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1099.272558] env[65726]: DEBUG nova.policy [None req-5eb7fa6b-cbd6-4a47-b227-e9c16f779f4a tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b776acdc25b648d2ba0575d7a5d5fec0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'eb811e38506b4297bf01bc5f2e098370', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65726) authorize /opt/stack/nova/nova/policy.py:192}} [ 1099.404785] env[65726]: DEBUG oslo_concurrency.lockutils [None req-be8a3d8a-c73e-4be5-b722-9654488b1331 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Acquiring lock "ad4f00b5-3765-494f-8132-8b17b29d219a" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1099.405061] env[65726]: DEBUG oslo_concurrency.lockutils [None req-be8a3d8a-c73e-4be5-b722-9654488b1331 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Lock "ad4f00b5-3765-494f-8132-8b17b29d219a" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1099.451079] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-242aead6-11df-41b2-9cb1-79ce86d8b2ee {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.461982] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1f4821b-5b8c-41a3-aecc-1bd0409eb935 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.501436] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09a24694-2023-425c-ba21-3298ce5f05fe {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.506413] env[65726]: WARNING openstack [None req-b56a2566-f325-47b4-827c-75797fd41834 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1099.506838] env[65726]: WARNING openstack [None req-b56a2566-f325-47b4-827c-75797fd41834 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1099.519443] env[65726]: DEBUG oslo_vmware.api [None req-5101d24a-5b00-4587-8b37-5a45f8c4ea31 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': task-5116716, 'name': ReconfigVM_Task, 'duration_secs': 0.749237} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1099.519614] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-5101d24a-5b00-4587-8b37-5a45f8c4ea31 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 0361ae8f-634d-4161-a594-7f93dbb1b7cd] Reconfigured VM instance instance-00000064 to attach disk [datastore1] 0361ae8f-634d-4161-a594-7f93dbb1b7cd/0361ae8f-634d-4161-a594-7f93dbb1b7cd.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1099.521321] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0cce5f0b-77ab-4184-b991-5856198a27d3 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.525475] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-096a77b2-d5fb-4c63-b6dd-c2aab832d6cc {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.538088] env[65726]: DEBUG nova.compute.provider_tree [None req-a6f4e31f-bf92-46ff-ad83-0dbcdb2a78cf tempest-InstanceActionsV221TestJSON-647428531 tempest-InstanceActionsV221TestJSON-647428531-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1099.540808] env[65726]: DEBUG oslo_vmware.api [None req-5101d24a-5b00-4587-8b37-5a45f8c4ea31 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Waiting for the task: (returnval){ [ 1099.540808] env[65726]: value = "task-5116719" [ 1099.540808] env[65726]: _type = "Task" [ 1099.540808] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1099.553030] env[65726]: DEBUG oslo_vmware.api [None req-5101d24a-5b00-4587-8b37-5a45f8c4ea31 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': task-5116719, 'name': Rename_Task} progress is 10%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1099.569646] env[65726]: DEBUG nova.network.neutron [None req-5eb7fa6b-cbd6-4a47-b227-e9c16f779f4a tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] [instance: 4809fcf6-59d4-409c-bd8d-981dc85a686f] Successfully created port: 13050a31-7165-4b22-83cc-d8ce9a41eadf {{(pid=65726) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1099.577982] env[65726]: DEBUG oslo_vmware.api [None req-0288a821-ce9c-40cd-86ef-962ac009580c tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5116718, 'name': PowerOffVM_Task, 'duration_secs': 0.347296} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1099.578262] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-0288a821-ce9c-40cd-86ef-962ac009580c tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: a56be6fa-6027-46ee-9aa8-332fe4222d68] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1099.578459] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-0288a821-ce9c-40cd-86ef-962ac009580c tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: a56be6fa-6027-46ee-9aa8-332fe4222d68] Volume detach. Driver type: vmdk {{(pid=65726) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1099.578652] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-0288a821-ce9c-40cd-86ef-962ac009580c tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: a56be6fa-6027-46ee-9aa8-332fe4222d68] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-995275', 'volume_id': '30181d53-b2b9-4253-a8f5-28a3abf812f9', 'name': 'volume-30181d53-b2b9-4253-a8f5-28a3abf812f9', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'attaching', 'instance': 'a56be6fa-6027-46ee-9aa8-332fe4222d68', 'attached_at': '2025-12-12T19:39:15.000000', 'detached_at': '', 'volume_id': '30181d53-b2b9-4253-a8f5-28a3abf812f9', 'serial': '30181d53-b2b9-4253-a8f5-28a3abf812f9'} {{(pid=65726) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1099.579542] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55e74461-b62f-49c9-b8ec-f8ed9d4cc08a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.601211] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbb58d83-d098-43fc-b80f-d9043e9852ac {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.614184] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e307066-413e-4e6d-849c-a17a8baab3c5 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.639740] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41150957-61ad-4748-9634-5d8a9b038d60 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.656877] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-0288a821-ce9c-40cd-86ef-962ac009580c tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] The volume has not been displaced from its original location: [datastore2] volume-30181d53-b2b9-4253-a8f5-28a3abf812f9/volume-30181d53-b2b9-4253-a8f5-28a3abf812f9.vmdk. No consolidation needed. {{(pid=65726) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1099.662266] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-0288a821-ce9c-40cd-86ef-962ac009580c tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: a56be6fa-6027-46ee-9aa8-332fe4222d68] Reconfiguring VM instance instance-00000053 to detach disk 2001 {{(pid=65726) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1099.669542] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1533cec2-dde9-411b-8c5c-3826f737c490 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.692047] env[65726]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1099.692047] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52041259-6206-4882-dea0-00fc5d492a1e" [ 1099.692047] env[65726]: _type = "HttpNfcLease" [ 1099.692047] env[65726]: } is ready. {{(pid=65726) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1099.694251] env[65726]: DEBUG oslo_vmware.rw_handles [None req-1c19f444-aee2-49ca-b67e-d3cef83d3664 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1099.694251] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52041259-6206-4882-dea0-00fc5d492a1e" [ 1099.694251] env[65726]: _type = "HttpNfcLease" [ 1099.694251] env[65726]: }. {{(pid=65726) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 1099.694718] env[65726]: DEBUG oslo_vmware.api [None req-0288a821-ce9c-40cd-86ef-962ac009580c tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Waiting for the task: (returnval){ [ 1099.694718] env[65726]: value = "task-5116720" [ 1099.694718] env[65726]: _type = "Task" [ 1099.694718] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1099.695782] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1880ee5a-a8ea-4586-bcb9-ef48a23d5858 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.713314] env[65726]: DEBUG oslo_vmware.rw_handles [None req-1c19f444-aee2-49ca-b67e-d3cef83d3664 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52610643-3cc8-edf6-4806-155a159ec092/disk-0.vmdk from lease info. {{(pid=65726) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1099.713622] env[65726]: DEBUG oslo_vmware.rw_handles [None req-1c19f444-aee2-49ca-b67e-d3cef83d3664 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Creating HTTP connection to write to file with size = 31667200 and URL = https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52610643-3cc8-edf6-4806-155a159ec092/disk-0.vmdk. {{(pid=65726) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 1099.722481] env[65726]: WARNING openstack [None req-b56a2566-f325-47b4-827c-75797fd41834 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1099.723111] env[65726]: WARNING openstack [None req-b56a2566-f325-47b4-827c-75797fd41834 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1099.733992] env[65726]: DEBUG oslo_vmware.api [None req-0288a821-ce9c-40cd-86ef-962ac009580c tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5116720, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1099.734716] env[65726]: DEBUG nova.compute.manager [None req-5eb7fa6b-cbd6-4a47-b227-e9c16f779f4a tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] [instance: 4809fcf6-59d4-409c-bd8d-981dc85a686f] Start building block device mappings for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1099.802635] env[65726]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-6e332393-eeec-4026-b69b-ac3124fda782 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.873562] env[65726]: WARNING openstack [None req-b56a2566-f325-47b4-827c-75797fd41834 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1099.874046] env[65726]: WARNING openstack [None req-b56a2566-f325-47b4-827c-75797fd41834 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1099.909037] env[65726]: DEBUG nova.compute.utils [None req-be8a3d8a-c73e-4be5-b722-9654488b1331 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Using /dev/sd instead of None {{(pid=65726) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1099.964793] env[65726]: DEBUG nova.network.neutron [None req-b56a2566-f325-47b4-827c-75797fd41834 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] [instance: df08d166-7811-49b5-9c16-a3434229d623] Updating instance_info_cache with network_info: [{"id": "70efdc7d-7775-4990-b851-d60195bd1504", "address": "fa:16:3e:be:5b:19", "network": {"id": "835ec128-d214-4f54-ab53-88f8378b7d74", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-1935476539-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.151", "type": "floating", "version": 4, "meta": {}}]}, {"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "305ecace479440ad9a616e546ccd38ce", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3b107fab-ee71-47db-ad4d-3c6f05546843", "external-id": "cl2-zone-554", "segmentation_id": 554, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap70efdc7d-77", "ovs_interfaceid": "70efdc7d-7775-4990-b851-d60195bd1504", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1100.044582] env[65726]: DEBUG nova.scheduler.client.report [None req-a6f4e31f-bf92-46ff-ad83-0dbcdb2a78cf tempest-InstanceActionsV221TestJSON-647428531 tempest-InstanceActionsV221TestJSON-647428531-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1100.061267] env[65726]: DEBUG oslo_vmware.api [None req-5101d24a-5b00-4587-8b37-5a45f8c4ea31 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': task-5116719, 'name': Rename_Task, 'duration_secs': 0.330659} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1100.063434] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-5101d24a-5b00-4587-8b37-5a45f8c4ea31 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 0361ae8f-634d-4161-a594-7f93dbb1b7cd] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1100.064416] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d4717325-c816-4fb2-8127-8b592119dcaf {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.072733] env[65726]: DEBUG oslo_vmware.api [None req-5101d24a-5b00-4587-8b37-5a45f8c4ea31 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Waiting for the task: (returnval){ [ 1100.072733] env[65726]: value = "task-5116721" [ 1100.072733] env[65726]: _type = "Task" [ 1100.072733] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1100.084087] env[65726]: DEBUG oslo_vmware.api [None req-5101d24a-5b00-4587-8b37-5a45f8c4ea31 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': task-5116721, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1100.211511] env[65726]: DEBUG oslo_vmware.api [None req-0288a821-ce9c-40cd-86ef-962ac009580c tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5116720, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1100.413373] env[65726]: DEBUG oslo_concurrency.lockutils [None req-be8a3d8a-c73e-4be5-b722-9654488b1331 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Lock "ad4f00b5-3765-494f-8132-8b17b29d219a" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.008s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1100.467407] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b56a2566-f325-47b4-827c-75797fd41834 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] Releasing lock "refresh_cache-df08d166-7811-49b5-9c16-a3434229d623" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1100.467833] env[65726]: DEBUG nova.compute.manager [None req-b56a2566-f325-47b4-827c-75797fd41834 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] [instance: df08d166-7811-49b5-9c16-a3434229d623] Inject network info {{(pid=65726) _inject_network_info /opt/stack/nova/nova/compute/manager.py:7794}} [ 1100.468384] env[65726]: DEBUG nova.compute.manager [None req-b56a2566-f325-47b4-827c-75797fd41834 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] [instance: df08d166-7811-49b5-9c16-a3434229d623] network_info to inject: |[{"id": "70efdc7d-7775-4990-b851-d60195bd1504", "address": "fa:16:3e:be:5b:19", "network": {"id": "835ec128-d214-4f54-ab53-88f8378b7d74", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-1935476539-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.151", "type": "floating", "version": 4, "meta": {}}]}, {"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "305ecace479440ad9a616e546ccd38ce", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3b107fab-ee71-47db-ad4d-3c6f05546843", "external-id": "cl2-zone-554", "segmentation_id": 554, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap70efdc7d-77", "ovs_interfaceid": "70efdc7d-7775-4990-b851-d60195bd1504", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65726) _inject_network_info /opt/stack/nova/nova/compute/manager.py:7795}} [ 1100.474648] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-b56a2566-f325-47b4-827c-75797fd41834 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] [instance: df08d166-7811-49b5-9c16-a3434229d623] Reconfiguring VM instance to set the machine id {{(pid=65726) _set_machine_id /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1796}} [ 1100.477070] env[65726]: DEBUG oslo_concurrency.lockutils [req-05b43d10-3696-4fd8-8633-a35327ca1b02 req-0cd452a9-d0fc-4ddd-bb22-423921d86061 service nova] Acquired lock "refresh_cache-df08d166-7811-49b5-9c16-a3434229d623" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1100.477344] env[65726]: DEBUG nova.network.neutron [req-05b43d10-3696-4fd8-8633-a35327ca1b02 req-0cd452a9-d0fc-4ddd-bb22-423921d86061 service nova] [instance: df08d166-7811-49b5-9c16-a3434229d623] Refreshing network info cache for port 70efdc7d-7775-4990-b851-d60195bd1504 {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 1100.479479] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9ca327db-c951-4f2b-a235-8f28a1895948 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.502128] env[65726]: DEBUG oslo_vmware.api [None req-b56a2566-f325-47b4-827c-75797fd41834 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] Waiting for the task: (returnval){ [ 1100.502128] env[65726]: value = "task-5116722" [ 1100.502128] env[65726]: _type = "Task" [ 1100.502128] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1100.513827] env[65726]: DEBUG oslo_vmware.api [None req-b56a2566-f325-47b4-827c-75797fd41834 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] Task: {'id': task-5116722, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1100.555436] env[65726]: DEBUG nova.objects.instance [None req-1dba9ad2-d4be-41a5-9681-c8a1af4e365c tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] Lazy-loading 'flavor' on Instance uuid df08d166-7811-49b5-9c16-a3434229d623 {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1100.560477] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a6f4e31f-bf92-46ff-ad83-0dbcdb2a78cf tempest-InstanceActionsV221TestJSON-647428531 tempest-InstanceActionsV221TestJSON-647428531-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.848s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1100.566041] env[65726]: DEBUG oslo_concurrency.lockutils [None req-f67ad2d6-6757-4118-93c4-4106809d054d tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 5.625s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1100.566041] env[65726]: DEBUG nova.objects.instance [None req-f67ad2d6-6757-4118-93c4-4106809d054d tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Lazy-loading 'resources' on Instance uuid d8801049-fc34-4b2f-8d5e-b9459c3e03a6 {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1100.587696] env[65726]: DEBUG oslo_vmware.api [None req-5101d24a-5b00-4587-8b37-5a45f8c4ea31 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': task-5116721, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1100.597646] env[65726]: INFO nova.scheduler.client.report [None req-a6f4e31f-bf92-46ff-ad83-0dbcdb2a78cf tempest-InstanceActionsV221TestJSON-647428531 tempest-InstanceActionsV221TestJSON-647428531-project-member] Deleted allocations for instance b0fec5ff-8c1e-41f9-b30a-6253a34baeec [ 1100.712359] env[65726]: DEBUG oslo_vmware.api [None req-0288a821-ce9c-40cd-86ef-962ac009580c tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5116720, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1100.801151] env[65726]: DEBUG nova.compute.manager [None req-5eb7fa6b-cbd6-4a47-b227-e9c16f779f4a tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] [instance: 4809fcf6-59d4-409c-bd8d-981dc85a686f] Start spawning the instance on the hypervisor. {{(pid=65726) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1100.830591] env[65726]: DEBUG nova.virt.hardware [None req-5eb7fa6b-cbd6-4a47-b227-e9c16f779f4a tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T19:28:34Z,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1539f06fa3534e90acbc3a60c4b8bd3f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T19:28:35Z,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1100.830866] env[65726]: DEBUG nova.virt.hardware [None req-5eb7fa6b-cbd6-4a47-b227-e9c16f779f4a tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1100.831009] env[65726]: DEBUG nova.virt.hardware [None req-5eb7fa6b-cbd6-4a47-b227-e9c16f779f4a tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1100.831200] env[65726]: DEBUG nova.virt.hardware [None req-5eb7fa6b-cbd6-4a47-b227-e9c16f779f4a tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1100.831344] env[65726]: DEBUG nova.virt.hardware [None req-5eb7fa6b-cbd6-4a47-b227-e9c16f779f4a tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1100.831486] env[65726]: DEBUG nova.virt.hardware [None req-5eb7fa6b-cbd6-4a47-b227-e9c16f779f4a tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1100.831702] env[65726]: DEBUG nova.virt.hardware [None req-5eb7fa6b-cbd6-4a47-b227-e9c16f779f4a tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1100.835028] env[65726]: DEBUG nova.virt.hardware [None req-5eb7fa6b-cbd6-4a47-b227-e9c16f779f4a tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1100.835028] env[65726]: DEBUG nova.virt.hardware [None req-5eb7fa6b-cbd6-4a47-b227-e9c16f779f4a tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1100.835028] env[65726]: DEBUG nova.virt.hardware [None req-5eb7fa6b-cbd6-4a47-b227-e9c16f779f4a tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1100.835028] env[65726]: DEBUG nova.virt.hardware [None req-5eb7fa6b-cbd6-4a47-b227-e9c16f779f4a tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1100.835028] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e5c2c5e-0293-4c42-955e-f062e3934bd7 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.842391] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d22ff633-4404-47cd-bf93-1056c01e6663 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.000027] env[65726]: WARNING openstack [req-05b43d10-3696-4fd8-8633-a35327ca1b02 req-0cd452a9-d0fc-4ddd-bb22-423921d86061 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1101.000027] env[65726]: WARNING openstack [req-05b43d10-3696-4fd8-8633-a35327ca1b02 req-0cd452a9-d0fc-4ddd-bb22-423921d86061 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1101.008045] env[65726]: DEBUG oslo_vmware.rw_handles [None req-1c19f444-aee2-49ca-b67e-d3cef83d3664 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Completed reading data from the image iterator. {{(pid=65726) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 1101.008356] env[65726]: DEBUG oslo_vmware.rw_handles [None req-1c19f444-aee2-49ca-b67e-d3cef83d3664 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52610643-3cc8-edf6-4806-155a159ec092/disk-0.vmdk. {{(pid=65726) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1101.013298] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05825af8-2bea-4aff-af9f-08b0de7be467 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.036098] env[65726]: DEBUG oslo_vmware.rw_handles [None req-1c19f444-aee2-49ca-b67e-d3cef83d3664 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52610643-3cc8-edf6-4806-155a159ec092/disk-0.vmdk is in state: ready. {{(pid=65726) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1101.037086] env[65726]: DEBUG oslo_vmware.rw_handles [None req-1c19f444-aee2-49ca-b67e-d3cef83d3664 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Releasing lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52610643-3cc8-edf6-4806-155a159ec092/disk-0.vmdk. {{(pid=65726) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 1101.037086] env[65726]: DEBUG oslo_vmware.api [None req-b56a2566-f325-47b4-827c-75797fd41834 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] Task: {'id': task-5116722, 'name': ReconfigVM_Task, 'duration_secs': 0.192722} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1101.037086] env[65726]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-9175fd11-51f0-4b30-bf4b-9b2d38237b08 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.038899] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-b56a2566-f325-47b4-827c-75797fd41834 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] [instance: df08d166-7811-49b5-9c16-a3434229d623] Reconfigured VM instance to set the machine id {{(pid=65726) _set_machine_id /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1799}} [ 1101.076577] env[65726]: DEBUG oslo_concurrency.lockutils [None req-1dba9ad2-d4be-41a5-9681-c8a1af4e365c tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] Acquiring lock "refresh_cache-df08d166-7811-49b5-9c16-a3434229d623" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1101.089208] env[65726]: DEBUG oslo_vmware.api [None req-5101d24a-5b00-4587-8b37-5a45f8c4ea31 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': task-5116721, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1101.103588] env[65726]: DEBUG nova.compute.manager [req-6aacd10b-de61-44c1-8dc7-680a1f8abec1 req-123bd5b3-898f-41e0-b2ef-0b5bdcd1395c service nova] [instance: 4809fcf6-59d4-409c-bd8d-981dc85a686f] Received event network-vif-plugged-13050a31-7165-4b22-83cc-d8ce9a41eadf {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1101.103876] env[65726]: DEBUG oslo_concurrency.lockutils [req-6aacd10b-de61-44c1-8dc7-680a1f8abec1 req-123bd5b3-898f-41e0-b2ef-0b5bdcd1395c service nova] Acquiring lock "4809fcf6-59d4-409c-bd8d-981dc85a686f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1101.104245] env[65726]: DEBUG oslo_concurrency.lockutils [req-6aacd10b-de61-44c1-8dc7-680a1f8abec1 req-123bd5b3-898f-41e0-b2ef-0b5bdcd1395c service nova] Lock "4809fcf6-59d4-409c-bd8d-981dc85a686f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1101.104308] env[65726]: DEBUG oslo_concurrency.lockutils [req-6aacd10b-de61-44c1-8dc7-680a1f8abec1 req-123bd5b3-898f-41e0-b2ef-0b5bdcd1395c service nova] Lock "4809fcf6-59d4-409c-bd8d-981dc85a686f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1101.104476] env[65726]: DEBUG nova.compute.manager [req-6aacd10b-de61-44c1-8dc7-680a1f8abec1 req-123bd5b3-898f-41e0-b2ef-0b5bdcd1395c service nova] [instance: 4809fcf6-59d4-409c-bd8d-981dc85a686f] No waiting events found dispatching network-vif-plugged-13050a31-7165-4b22-83cc-d8ce9a41eadf {{(pid=65726) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1101.104652] env[65726]: WARNING nova.compute.manager [req-6aacd10b-de61-44c1-8dc7-680a1f8abec1 req-123bd5b3-898f-41e0-b2ef-0b5bdcd1395c service nova] [instance: 4809fcf6-59d4-409c-bd8d-981dc85a686f] Received unexpected event network-vif-plugged-13050a31-7165-4b22-83cc-d8ce9a41eadf for instance with vm_state building and task_state spawning. [ 1101.108767] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a6f4e31f-bf92-46ff-ad83-0dbcdb2a78cf tempest-InstanceActionsV221TestJSON-647428531 tempest-InstanceActionsV221TestJSON-647428531-project-member] Lock "b0fec5ff-8c1e-41f9-b30a-6253a34baeec" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 15.177s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1101.218489] env[65726]: DEBUG nova.network.neutron [None req-5eb7fa6b-cbd6-4a47-b227-e9c16f779f4a tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] [instance: 4809fcf6-59d4-409c-bd8d-981dc85a686f] Successfully updated port: 13050a31-7165-4b22-83cc-d8ce9a41eadf {{(pid=65726) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 1101.229373] env[65726]: DEBUG oslo_vmware.api [None req-0288a821-ce9c-40cd-86ef-962ac009580c tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5116720, 'name': ReconfigVM_Task, 'duration_secs': 1.323943} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1101.230319] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-0288a821-ce9c-40cd-86ef-962ac009580c tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: a56be6fa-6027-46ee-9aa8-332fe4222d68] Reconfigured VM instance instance-00000053 to detach disk 2001 {{(pid=65726) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1101.241925] env[65726]: WARNING openstack [req-05b43d10-3696-4fd8-8633-a35327ca1b02 req-0cd452a9-d0fc-4ddd-bb22-423921d86061 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1101.242471] env[65726]: WARNING openstack [req-05b43d10-3696-4fd8-8633-a35327ca1b02 req-0cd452a9-d0fc-4ddd-bb22-423921d86061 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1101.251297] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e3dad8c8-6304-41dc-8195-9568e5dfd784 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.268425] env[65726]: DEBUG oslo_vmware.rw_handles [None req-1c19f444-aee2-49ca-b67e-d3cef83d3664 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Closed VMDK write handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52610643-3cc8-edf6-4806-155a159ec092/disk-0.vmdk. {{(pid=65726) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 1101.268712] env[65726]: INFO nova.virt.vmwareapi.images [None req-1c19f444-aee2-49ca-b67e-d3cef83d3664 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] [instance: 6fd70b4a-63e4-4258-9cff-f3b582500b9f] Downloaded image file data eb421501-6112-4a68-9c48-c82505e774d2 [ 1101.274337] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a67a50ae-4459-444f-9cbb-c89c5803d630 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.278077] env[65726]: DEBUG oslo_vmware.api [None req-0288a821-ce9c-40cd-86ef-962ac009580c tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Waiting for the task: (returnval){ [ 1101.278077] env[65726]: value = "task-5116723" [ 1101.278077] env[65726]: _type = "Task" [ 1101.278077] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1101.305443] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0f608b53-03f3-4ef8-a990-511ee5198e78 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.314040] env[65726]: DEBUG oslo_vmware.api [None req-0288a821-ce9c-40cd-86ef-962ac009580c tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5116723, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1101.340954] env[65726]: INFO nova.virt.vmwareapi.images [None req-1c19f444-aee2-49ca-b67e-d3cef83d3664 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] [instance: 6fd70b4a-63e4-4258-9cff-f3b582500b9f] The imported VM was unregistered [ 1101.344704] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-1c19f444-aee2-49ca-b67e-d3cef83d3664 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] [instance: 6fd70b4a-63e4-4258-9cff-f3b582500b9f] Caching image {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 1101.345028] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-1c19f444-aee2-49ca-b67e-d3cef83d3664 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Creating directory with path [datastore1] devstack-image-cache_base/eb421501-6112-4a68-9c48-c82505e774d2 {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1101.345433] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b32ba3a5-073b-4be5-95c0-24d7c15c8904 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.355821] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c34e126a-7282-4178-be11-adc091456212 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.360703] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-1c19f444-aee2-49ca-b67e-d3cef83d3664 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Created directory with path [datastore1] devstack-image-cache_base/eb421501-6112-4a68-9c48-c82505e774d2 {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1101.360937] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-1c19f444-aee2-49ca-b67e-d3cef83d3664 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Moving virtual disk from [datastore1] OSTACK_IMG_5f557822-0d62-4142-9f38-0161dfdccbb6/OSTACK_IMG_5f557822-0d62-4142-9f38-0161dfdccbb6.vmdk to [datastore1] devstack-image-cache_base/eb421501-6112-4a68-9c48-c82505e774d2/eb421501-6112-4a68-9c48-c82505e774d2.vmdk. {{(pid=65726) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 1101.362731] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-8b61154c-9f73-4066-8270-a8f6e78b2bee {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.369645] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e86e55b2-52da-405e-bb01-87c2d94d0d34 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.377739] env[65726]: WARNING openstack [req-05b43d10-3696-4fd8-8633-a35327ca1b02 req-0cd452a9-d0fc-4ddd-bb22-423921d86061 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1101.378224] env[65726]: WARNING openstack [req-05b43d10-3696-4fd8-8633-a35327ca1b02 req-0cd452a9-d0fc-4ddd-bb22-423921d86061 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1101.388057] env[65726]: DEBUG oslo_vmware.api [None req-1c19f444-aee2-49ca-b67e-d3cef83d3664 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Waiting for the task: (returnval){ [ 1101.388057] env[65726]: value = "task-5116725" [ 1101.388057] env[65726]: _type = "Task" [ 1101.388057] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1101.424416] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f372d2ad-0e77-4bc3-beed-0e0cc6a4a0d4 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.432545] env[65726]: DEBUG oslo_vmware.api [None req-1c19f444-aee2-49ca-b67e-d3cef83d3664 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Task: {'id': task-5116725, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1101.440410] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bcded869-f80b-4798-9e0e-2d9653d88213 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.457750] env[65726]: DEBUG nova.compute.provider_tree [None req-f67ad2d6-6757-4118-93c4-4106809d054d tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1101.509449] env[65726]: DEBUG nova.network.neutron [req-05b43d10-3696-4fd8-8633-a35327ca1b02 req-0cd452a9-d0fc-4ddd-bb22-423921d86061 service nova] [instance: df08d166-7811-49b5-9c16-a3434229d623] Updated VIF entry in instance network info cache for port 70efdc7d-7775-4990-b851-d60195bd1504. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 1101.509885] env[65726]: DEBUG nova.network.neutron [req-05b43d10-3696-4fd8-8633-a35327ca1b02 req-0cd452a9-d0fc-4ddd-bb22-423921d86061 service nova] [instance: df08d166-7811-49b5-9c16-a3434229d623] Updating instance_info_cache with network_info: [{"id": "70efdc7d-7775-4990-b851-d60195bd1504", "address": "fa:16:3e:be:5b:19", "network": {"id": "835ec128-d214-4f54-ab53-88f8378b7d74", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-1935476539-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.151", "type": "floating", "version": 4, "meta": {}}]}, {"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "305ecace479440ad9a616e546ccd38ce", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3b107fab-ee71-47db-ad4d-3c6f05546843", "external-id": "cl2-zone-554", "segmentation_id": 554, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap70efdc7d-77", "ovs_interfaceid": "70efdc7d-7775-4990-b851-d60195bd1504", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1101.523604] env[65726]: DEBUG oslo_concurrency.lockutils [None req-be8a3d8a-c73e-4be5-b722-9654488b1331 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Acquiring lock "ad4f00b5-3765-494f-8132-8b17b29d219a" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1101.524045] env[65726]: DEBUG oslo_concurrency.lockutils [None req-be8a3d8a-c73e-4be5-b722-9654488b1331 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Lock "ad4f00b5-3765-494f-8132-8b17b29d219a" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1101.524377] env[65726]: INFO nova.compute.manager [None req-be8a3d8a-c73e-4be5-b722-9654488b1331 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: ad4f00b5-3765-494f-8132-8b17b29d219a] Attaching volume 1c311bc8-792e-43d1-b5e9-6105ef7b377e to /dev/sdb [ 1101.584728] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f7cd011-86a2-4ae8-bb83-cf7cd6ab3164 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.594716] env[65726]: DEBUG oslo_vmware.api [None req-5101d24a-5b00-4587-8b37-5a45f8c4ea31 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': task-5116721, 'name': PowerOnVM_Task, 'duration_secs': 1.376156} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1101.596839] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-5101d24a-5b00-4587-8b37-5a45f8c4ea31 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 0361ae8f-634d-4161-a594-7f93dbb1b7cd] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1101.597414] env[65726]: INFO nova.compute.manager [None req-5101d24a-5b00-4587-8b37-5a45f8c4ea31 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 0361ae8f-634d-4161-a594-7f93dbb1b7cd] Took 8.51 seconds to spawn the instance on the hypervisor. [ 1101.598305] env[65726]: DEBUG nova.compute.manager [None req-5101d24a-5b00-4587-8b37-5a45f8c4ea31 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 0361ae8f-634d-4161-a594-7f93dbb1b7cd] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1101.598934] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d1adc81-5af8-48f0-ad32-dd32918b152a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.603024] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9ede734-47e5-41c0-8265-3773cdfef814 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.626085] env[65726]: DEBUG nova.virt.block_device [None req-be8a3d8a-c73e-4be5-b722-9654488b1331 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: ad4f00b5-3765-494f-8132-8b17b29d219a] Updating existing volume attachment record: 35efd105-6ecf-4f5f-a5b1-6902a6a03a40 {{(pid=65726) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1101.731041] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5eb7fa6b-cbd6-4a47-b227-e9c16f779f4a tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Acquiring lock "refresh_cache-4809fcf6-59d4-409c-bd8d-981dc85a686f" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1101.731261] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5eb7fa6b-cbd6-4a47-b227-e9c16f779f4a tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Acquired lock "refresh_cache-4809fcf6-59d4-409c-bd8d-981dc85a686f" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1101.731454] env[65726]: DEBUG nova.network.neutron [None req-5eb7fa6b-cbd6-4a47-b227-e9c16f779f4a tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] [instance: 4809fcf6-59d4-409c-bd8d-981dc85a686f] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1101.792364] env[65726]: DEBUG oslo_vmware.api [None req-0288a821-ce9c-40cd-86ef-962ac009580c tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5116723, 'name': ReconfigVM_Task, 'duration_secs': 0.203728} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1101.792768] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-0288a821-ce9c-40cd-86ef-962ac009580c tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: a56be6fa-6027-46ee-9aa8-332fe4222d68] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-995275', 'volume_id': '30181d53-b2b9-4253-a8f5-28a3abf812f9', 'name': 'volume-30181d53-b2b9-4253-a8f5-28a3abf812f9', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'attaching', 'instance': 'a56be6fa-6027-46ee-9aa8-332fe4222d68', 'attached_at': '2025-12-12T19:39:15.000000', 'detached_at': '', 'volume_id': '30181d53-b2b9-4253-a8f5-28a3abf812f9', 'serial': '30181d53-b2b9-4253-a8f5-28a3abf812f9'} {{(pid=65726) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1101.793108] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-0288a821-ce9c-40cd-86ef-962ac009580c tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: a56be6fa-6027-46ee-9aa8-332fe4222d68] Destroying instance {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1101.794107] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-737043b1-3466-4d29-95ab-64e1a4c90233 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.807859] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-0288a821-ce9c-40cd-86ef-962ac009580c tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: a56be6fa-6027-46ee-9aa8-332fe4222d68] Unregistering the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1101.807859] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-dd1064f6-f74a-4396-92c1-c094deb87fa0 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.906654] env[65726]: DEBUG oslo_vmware.api [None req-1c19f444-aee2-49ca-b67e-d3cef83d3664 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Task: {'id': task-5116725, 'name': MoveVirtualDisk_Task} progress is 21%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1101.962614] env[65726]: DEBUG nova.scheduler.client.report [None req-f67ad2d6-6757-4118-93c4-4106809d054d tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1102.013039] env[65726]: DEBUG oslo_concurrency.lockutils [req-05b43d10-3696-4fd8-8633-a35327ca1b02 req-0cd452a9-d0fc-4ddd-bb22-423921d86061 service nova] Releasing lock "refresh_cache-df08d166-7811-49b5-9c16-a3434229d623" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1102.013911] env[65726]: DEBUG oslo_concurrency.lockutils [None req-1dba9ad2-d4be-41a5-9681-c8a1af4e365c tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] Acquired lock "refresh_cache-df08d166-7811-49b5-9c16-a3434229d623" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1102.014732] env[65726]: WARNING openstack [None req-1dba9ad2-d4be-41a5-9681-c8a1af4e365c tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1102.015122] env[65726]: WARNING openstack [None req-1dba9ad2-d4be-41a5-9681-c8a1af4e365c tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1102.104224] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-0288a821-ce9c-40cd-86ef-962ac009580c tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: a56be6fa-6027-46ee-9aa8-332fe4222d68] Unregistered the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1102.104387] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-0288a821-ce9c-40cd-86ef-962ac009580c tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: a56be6fa-6027-46ee-9aa8-332fe4222d68] Deleting contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1102.104426] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-0288a821-ce9c-40cd-86ef-962ac009580c tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Deleting the datastore file [datastore1] a56be6fa-6027-46ee-9aa8-332fe4222d68 {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1102.104714] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-04f52b48-ef09-47e6-bf97-8ff673bff8c6 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.116852] env[65726]: DEBUG oslo_vmware.api [None req-0288a821-ce9c-40cd-86ef-962ac009580c tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Waiting for the task: (returnval){ [ 1102.116852] env[65726]: value = "task-5116728" [ 1102.116852] env[65726]: _type = "Task" [ 1102.116852] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1102.134939] env[65726]: DEBUG oslo_vmware.api [None req-0288a821-ce9c-40cd-86ef-962ac009580c tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5116728, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1102.139986] env[65726]: INFO nova.compute.manager [None req-5101d24a-5b00-4587-8b37-5a45f8c4ea31 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 0361ae8f-634d-4161-a594-7f93dbb1b7cd] Took 16.58 seconds to build instance. [ 1102.238252] env[65726]: WARNING openstack [None req-5eb7fa6b-cbd6-4a47-b227-e9c16f779f4a tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1102.238661] env[65726]: WARNING openstack [None req-5eb7fa6b-cbd6-4a47-b227-e9c16f779f4a tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1102.245097] env[65726]: DEBUG nova.network.neutron [None req-5eb7fa6b-cbd6-4a47-b227-e9c16f779f4a tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] [instance: 4809fcf6-59d4-409c-bd8d-981dc85a686f] Instance cache missing network info. {{(pid=65726) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 1102.404798] env[65726]: DEBUG oslo_vmware.api [None req-1c19f444-aee2-49ca-b67e-d3cef83d3664 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Task: {'id': task-5116725, 'name': MoveVirtualDisk_Task} progress is 38%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1102.434253] env[65726]: DEBUG nova.network.neutron [None req-1dba9ad2-d4be-41a5-9681-c8a1af4e365c tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] [instance: df08d166-7811-49b5-9c16-a3434229d623] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1102.468596] env[65726]: DEBUG oslo_concurrency.lockutils [None req-f67ad2d6-6757-4118-93c4-4106809d054d tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.903s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1102.472363] env[65726]: DEBUG oslo_concurrency.lockutils [None req-eab5602f-7673-4d76-aa16-7ebf0d2c87e3 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 5.824s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1102.472680] env[65726]: DEBUG nova.objects.instance [None req-eab5602f-7673-4d76-aa16-7ebf0d2c87e3 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Lazy-loading 'resources' on Instance uuid 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04 {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1102.475118] env[65726]: DEBUG nova.compute.manager [req-86187dd3-4838-4e1f-bf0d-02309cffabe0 req-f30167b4-1658-4552-a4b8-05b102c2bd6a service nova] [instance: df08d166-7811-49b5-9c16-a3434229d623] Received event network-changed-70efdc7d-7775-4990-b851-d60195bd1504 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1102.475582] env[65726]: DEBUG nova.compute.manager [req-86187dd3-4838-4e1f-bf0d-02309cffabe0 req-f30167b4-1658-4552-a4b8-05b102c2bd6a service nova] [instance: df08d166-7811-49b5-9c16-a3434229d623] Refreshing instance network info cache due to event network-changed-70efdc7d-7775-4990-b851-d60195bd1504. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 1102.475582] env[65726]: DEBUG oslo_concurrency.lockutils [req-86187dd3-4838-4e1f-bf0d-02309cffabe0 req-f30167b4-1658-4552-a4b8-05b102c2bd6a service nova] Acquiring lock "refresh_cache-df08d166-7811-49b5-9c16-a3434229d623" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1102.507579] env[65726]: WARNING openstack [None req-5eb7fa6b-cbd6-4a47-b227-e9c16f779f4a tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1102.508028] env[65726]: WARNING openstack [None req-5eb7fa6b-cbd6-4a47-b227-e9c16f779f4a tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1102.520198] env[65726]: INFO nova.scheduler.client.report [None req-f67ad2d6-6757-4118-93c4-4106809d054d tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Deleted allocations for instance d8801049-fc34-4b2f-8d5e-b9459c3e03a6 [ 1102.630718] env[65726]: WARNING openstack [None req-5eb7fa6b-cbd6-4a47-b227-e9c16f779f4a tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1102.630936] env[65726]: WARNING openstack [None req-5eb7fa6b-cbd6-4a47-b227-e9c16f779f4a tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1102.643302] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5101d24a-5b00-4587-8b37-5a45f8c4ea31 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Lock "0361ae8f-634d-4161-a594-7f93dbb1b7cd" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 18.090s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1102.647619] env[65726]: DEBUG oslo_vmware.api [None req-0288a821-ce9c-40cd-86ef-962ac009580c tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5116728, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1102.757069] env[65726]: DEBUG nova.network.neutron [None req-5eb7fa6b-cbd6-4a47-b227-e9c16f779f4a tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] [instance: 4809fcf6-59d4-409c-bd8d-981dc85a686f] Updating instance_info_cache with network_info: [{"id": "13050a31-7165-4b22-83cc-d8ce9a41eadf", "address": "fa:16:3e:02:ed:66", "network": {"id": "53facc8b-af7e-44de-8c6f-6bea798b3f0b", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-2084216438-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "eb811e38506b4297bf01bc5f2e098370", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "19671de9-8b5b-4710-adc3-7419f3c0f171", "external-id": "nsx-vlan-transportzone-421", "segmentation_id": 421, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap13050a31-71", "ovs_interfaceid": "13050a31-7165-4b22-83cc-d8ce9a41eadf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1102.903824] env[65726]: DEBUG oslo_vmware.api [None req-1c19f444-aee2-49ca-b67e-d3cef83d3664 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Task: {'id': task-5116725, 'name': MoveVirtualDisk_Task} progress is 57%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1102.942049] env[65726]: WARNING openstack [None req-1dba9ad2-d4be-41a5-9681-c8a1af4e365c tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1102.942049] env[65726]: WARNING openstack [None req-1dba9ad2-d4be-41a5-9681-c8a1af4e365c tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1103.035894] env[65726]: DEBUG oslo_concurrency.lockutils [None req-f67ad2d6-6757-4118-93c4-4106809d054d tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Lock "d8801049-fc34-4b2f-8d5e-b9459c3e03a6" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 11.086s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1103.132340] env[65726]: DEBUG oslo_vmware.api [None req-0288a821-ce9c-40cd-86ef-962ac009580c tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5116728, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1103.147367] env[65726]: WARNING openstack [None req-1dba9ad2-d4be-41a5-9681-c8a1af4e365c tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1103.147694] env[65726]: WARNING openstack [None req-1dba9ad2-d4be-41a5-9681-c8a1af4e365c tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1103.160974] env[65726]: DEBUG nova.compute.manager [req-8fdb4cdc-0ef8-4cde-8523-b085553a2a7e req-11e1b7dd-4719-4267-b009-34f0360365c6 service nova] [instance: 4809fcf6-59d4-409c-bd8d-981dc85a686f] Received event network-changed-13050a31-7165-4b22-83cc-d8ce9a41eadf {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1103.161195] env[65726]: DEBUG nova.compute.manager [req-8fdb4cdc-0ef8-4cde-8523-b085553a2a7e req-11e1b7dd-4719-4267-b009-34f0360365c6 service nova] [instance: 4809fcf6-59d4-409c-bd8d-981dc85a686f] Refreshing instance network info cache due to event network-changed-13050a31-7165-4b22-83cc-d8ce9a41eadf. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 1103.161469] env[65726]: DEBUG oslo_concurrency.lockutils [req-8fdb4cdc-0ef8-4cde-8523-b085553a2a7e req-11e1b7dd-4719-4267-b009-34f0360365c6 service nova] Acquiring lock "refresh_cache-4809fcf6-59d4-409c-bd8d-981dc85a686f" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1103.248296] env[65726]: WARNING openstack [None req-1dba9ad2-d4be-41a5-9681-c8a1af4e365c tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1103.248821] env[65726]: WARNING openstack [None req-1dba9ad2-d4be-41a5-9681-c8a1af4e365c tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1103.265376] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5eb7fa6b-cbd6-4a47-b227-e9c16f779f4a tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Releasing lock "refresh_cache-4809fcf6-59d4-409c-bd8d-981dc85a686f" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1103.265874] env[65726]: DEBUG nova.compute.manager [None req-5eb7fa6b-cbd6-4a47-b227-e9c16f779f4a tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] [instance: 4809fcf6-59d4-409c-bd8d-981dc85a686f] Instance network_info: |[{"id": "13050a31-7165-4b22-83cc-d8ce9a41eadf", "address": "fa:16:3e:02:ed:66", "network": {"id": "53facc8b-af7e-44de-8c6f-6bea798b3f0b", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-2084216438-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "eb811e38506b4297bf01bc5f2e098370", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "19671de9-8b5b-4710-adc3-7419f3c0f171", "external-id": "nsx-vlan-transportzone-421", "segmentation_id": 421, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap13050a31-71", "ovs_interfaceid": "13050a31-7165-4b22-83cc-d8ce9a41eadf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1103.267019] env[65726]: DEBUG oslo_concurrency.lockutils [req-8fdb4cdc-0ef8-4cde-8523-b085553a2a7e req-11e1b7dd-4719-4267-b009-34f0360365c6 service nova] Acquired lock "refresh_cache-4809fcf6-59d4-409c-bd8d-981dc85a686f" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1103.267157] env[65726]: DEBUG nova.network.neutron [req-8fdb4cdc-0ef8-4cde-8523-b085553a2a7e req-11e1b7dd-4719-4267-b009-34f0360365c6 service nova] [instance: 4809fcf6-59d4-409c-bd8d-981dc85a686f] Refreshing network info cache for port 13050a31-7165-4b22-83cc-d8ce9a41eadf {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 1103.270066] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-5eb7fa6b-cbd6-4a47-b227-e9c16f779f4a tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] [instance: 4809fcf6-59d4-409c-bd8d-981dc85a686f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:02:ed:66', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '19671de9-8b5b-4710-adc3-7419f3c0f171', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '13050a31-7165-4b22-83cc-d8ce9a41eadf', 'vif_model': 'vmxnet3'}] {{(pid=65726) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1103.287769] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-5eb7fa6b-cbd6-4a47-b227-e9c16f779f4a tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1103.292544] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4809fcf6-59d4-409c-bd8d-981dc85a686f] Creating VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1103.295714] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-938d1ac0-6de7-44fb-99fb-a1ee6ab77f77 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.300905] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b7b759b0-fe69-4888-bada-47ae2a291ad5 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.343356] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f91361a-26bc-4ced-81c2-70460382550f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.351376] env[65726]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1103.351376] env[65726]: value = "task-5116729" [ 1103.351376] env[65726]: _type = "Task" [ 1103.351376] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1103.402207] env[65726]: DEBUG nova.network.neutron [None req-1dba9ad2-d4be-41a5-9681-c8a1af4e365c tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] [instance: df08d166-7811-49b5-9c16-a3434229d623] Updating instance_info_cache with network_info: [{"id": "70efdc7d-7775-4990-b851-d60195bd1504", "address": "fa:16:3e:be:5b:19", "network": {"id": "835ec128-d214-4f54-ab53-88f8378b7d74", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-1935476539-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.151", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "305ecace479440ad9a616e546ccd38ce", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3b107fab-ee71-47db-ad4d-3c6f05546843", "external-id": "cl2-zone-554", "segmentation_id": 554, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap70efdc7d-77", "ovs_interfaceid": "70efdc7d-7775-4990-b851-d60195bd1504", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1103.411702] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d30249e3-78c1-4907-896d-f57a4580e480 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.416538] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116729, 'name': CreateVM_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1103.435349] env[65726]: DEBUG oslo_vmware.api [None req-1c19f444-aee2-49ca-b67e-d3cef83d3664 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Task: {'id': task-5116725, 'name': MoveVirtualDisk_Task} progress is 77%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1103.437185] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68f78c72-0c86-45d8-82aa-16901f1f2d5f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.459786] env[65726]: DEBUG nova.compute.provider_tree [None req-eab5602f-7673-4d76-aa16-7ebf0d2c87e3 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1103.632797] env[65726]: DEBUG oslo_vmware.api [None req-0288a821-ce9c-40cd-86ef-962ac009580c tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5116728, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1103.789789] env[65726]: WARNING openstack [req-8fdb4cdc-0ef8-4cde-8523-b085553a2a7e req-11e1b7dd-4719-4267-b009-34f0360365c6 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1103.789950] env[65726]: WARNING openstack [req-8fdb4cdc-0ef8-4cde-8523-b085553a2a7e req-11e1b7dd-4719-4267-b009-34f0360365c6 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1103.870202] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116729, 'name': CreateVM_Task, 'duration_secs': 0.502263} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1103.870396] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4809fcf6-59d4-409c-bd8d-981dc85a686f] Created VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1103.873970] env[65726]: WARNING openstack [None req-5eb7fa6b-cbd6-4a47-b227-e9c16f779f4a tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1103.873970] env[65726]: WARNING openstack [None req-5eb7fa6b-cbd6-4a47-b227-e9c16f779f4a tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1103.878283] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5eb7fa6b-cbd6-4a47-b227-e9c16f779f4a tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1103.878283] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5eb7fa6b-cbd6-4a47-b227-e9c16f779f4a tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1103.878283] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5eb7fa6b-cbd6-4a47-b227-e9c16f779f4a tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1103.878283] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2cb8a930-6852-413f-99e3-5edd7bf0d206 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.888202] env[65726]: DEBUG oslo_vmware.api [None req-5eb7fa6b-cbd6-4a47-b227-e9c16f779f4a tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Waiting for the task: (returnval){ [ 1103.888202] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]522dddb0-3a6a-3fa0-4ac5-a557e118b075" [ 1103.888202] env[65726]: _type = "Task" [ 1103.888202] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1103.908963] env[65726]: DEBUG oslo_vmware.api [None req-5eb7fa6b-cbd6-4a47-b227-e9c16f779f4a tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]522dddb0-3a6a-3fa0-4ac5-a557e118b075, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1103.919988] env[65726]: DEBUG oslo_concurrency.lockutils [None req-1dba9ad2-d4be-41a5-9681-c8a1af4e365c tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] Releasing lock "refresh_cache-df08d166-7811-49b5-9c16-a3434229d623" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1103.920323] env[65726]: DEBUG nova.compute.manager [None req-1dba9ad2-d4be-41a5-9681-c8a1af4e365c tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] [instance: df08d166-7811-49b5-9c16-a3434229d623] Inject network info {{(pid=65726) _inject_network_info /opt/stack/nova/nova/compute/manager.py:7794}} [ 1103.920708] env[65726]: DEBUG nova.compute.manager [None req-1dba9ad2-d4be-41a5-9681-c8a1af4e365c tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] [instance: df08d166-7811-49b5-9c16-a3434229d623] network_info to inject: |[{"id": "70efdc7d-7775-4990-b851-d60195bd1504", "address": "fa:16:3e:be:5b:19", "network": {"id": "835ec128-d214-4f54-ab53-88f8378b7d74", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-1935476539-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.151", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "305ecace479440ad9a616e546ccd38ce", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3b107fab-ee71-47db-ad4d-3c6f05546843", "external-id": "cl2-zone-554", "segmentation_id": 554, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap70efdc7d-77", "ovs_interfaceid": "70efdc7d-7775-4990-b851-d60195bd1504", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65726) _inject_network_info /opt/stack/nova/nova/compute/manager.py:7795}} [ 1103.926374] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-1dba9ad2-d4be-41a5-9681-c8a1af4e365c tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] [instance: df08d166-7811-49b5-9c16-a3434229d623] Reconfiguring VM instance to set the machine id {{(pid=65726) _set_machine_id /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1796}} [ 1103.926510] env[65726]: DEBUG oslo_vmware.api [None req-1c19f444-aee2-49ca-b67e-d3cef83d3664 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Task: {'id': task-5116725, 'name': MoveVirtualDisk_Task} progress is 91%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1103.927518] env[65726]: DEBUG oslo_concurrency.lockutils [req-86187dd3-4838-4e1f-bf0d-02309cffabe0 req-f30167b4-1658-4552-a4b8-05b102c2bd6a service nova] Acquired lock "refresh_cache-df08d166-7811-49b5-9c16-a3434229d623" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1103.927518] env[65726]: DEBUG nova.network.neutron [req-86187dd3-4838-4e1f-bf0d-02309cffabe0 req-f30167b4-1658-4552-a4b8-05b102c2bd6a service nova] [instance: df08d166-7811-49b5-9c16-a3434229d623] Refreshing network info cache for port 70efdc7d-7775-4990-b851-d60195bd1504 {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 1103.928842] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b7c2f8cc-10cc-4101-b245-06749ba193ac {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.941693] env[65726]: WARNING openstack [req-86187dd3-4838-4e1f-bf0d-02309cffabe0 req-f30167b4-1658-4552-a4b8-05b102c2bd6a service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1103.942202] env[65726]: WARNING openstack [req-86187dd3-4838-4e1f-bf0d-02309cffabe0 req-f30167b4-1658-4552-a4b8-05b102c2bd6a service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1103.964025] env[65726]: DEBUG nova.scheduler.client.report [None req-eab5602f-7673-4d76-aa16-7ebf0d2c87e3 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1103.974031] env[65726]: DEBUG oslo_vmware.api [None req-1dba9ad2-d4be-41a5-9681-c8a1af4e365c tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] Waiting for the task: (returnval){ [ 1103.974031] env[65726]: value = "task-5116731" [ 1103.974031] env[65726]: _type = "Task" [ 1103.974031] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1103.991457] env[65726]: DEBUG oslo_vmware.api [None req-1dba9ad2-d4be-41a5-9681-c8a1af4e365c tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] Task: {'id': task-5116731, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1104.019147] env[65726]: WARNING openstack [req-8fdb4cdc-0ef8-4cde-8523-b085553a2a7e req-11e1b7dd-4719-4267-b009-34f0360365c6 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1104.019799] env[65726]: WARNING openstack [req-8fdb4cdc-0ef8-4cde-8523-b085553a2a7e req-11e1b7dd-4719-4267-b009-34f0360365c6 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1104.132173] env[65726]: DEBUG oslo_vmware.api [None req-0288a821-ce9c-40cd-86ef-962ac009580c tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5116728, 'name': DeleteDatastoreFile_Task, 'duration_secs': 2.013218} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1104.135844] env[65726]: WARNING openstack [req-8fdb4cdc-0ef8-4cde-8523-b085553a2a7e req-11e1b7dd-4719-4267-b009-34f0360365c6 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1104.135998] env[65726]: WARNING openstack [req-8fdb4cdc-0ef8-4cde-8523-b085553a2a7e req-11e1b7dd-4719-4267-b009-34f0360365c6 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1104.144267] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-0288a821-ce9c-40cd-86ef-962ac009580c tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Deleted the datastore file {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1104.144480] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-0288a821-ce9c-40cd-86ef-962ac009580c tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: a56be6fa-6027-46ee-9aa8-332fe4222d68] Deleted contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1104.144682] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-0288a821-ce9c-40cd-86ef-962ac009580c tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: a56be6fa-6027-46ee-9aa8-332fe4222d68] Instance destroyed {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1104.144859] env[65726]: INFO nova.compute.manager [None req-0288a821-ce9c-40cd-86ef-962ac009580c tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: a56be6fa-6027-46ee-9aa8-332fe4222d68] Took 5.10 seconds to destroy the instance on the hypervisor. [ 1104.145249] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-0288a821-ce9c-40cd-86ef-962ac009580c tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1104.145739] env[65726]: DEBUG nova.compute.manager [-] [instance: a56be6fa-6027-46ee-9aa8-332fe4222d68] Deallocating network for instance {{(pid=65726) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1104.145838] env[65726]: DEBUG nova.network.neutron [-] [instance: a56be6fa-6027-46ee-9aa8-332fe4222d68] deallocate_for_instance() {{(pid=65726) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 1104.146889] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1104.146889] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1104.176513] env[65726]: WARNING openstack [req-86187dd3-4838-4e1f-bf0d-02309cffabe0 req-f30167b4-1658-4552-a4b8-05b102c2bd6a service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1104.176970] env[65726]: WARNING openstack [req-86187dd3-4838-4e1f-bf0d-02309cffabe0 req-f30167b4-1658-4552-a4b8-05b102c2bd6a service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1104.209347] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1104.209645] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1104.250403] env[65726]: DEBUG nova.network.neutron [req-8fdb4cdc-0ef8-4cde-8523-b085553a2a7e req-11e1b7dd-4719-4267-b009-34f0360365c6 service nova] [instance: 4809fcf6-59d4-409c-bd8d-981dc85a686f] Updated VIF entry in instance network info cache for port 13050a31-7165-4b22-83cc-d8ce9a41eadf. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 1104.251045] env[65726]: DEBUG nova.network.neutron [req-8fdb4cdc-0ef8-4cde-8523-b085553a2a7e req-11e1b7dd-4719-4267-b009-34f0360365c6 service nova] [instance: 4809fcf6-59d4-409c-bd8d-981dc85a686f] Updating instance_info_cache with network_info: [{"id": "13050a31-7165-4b22-83cc-d8ce9a41eadf", "address": "fa:16:3e:02:ed:66", "network": {"id": "53facc8b-af7e-44de-8c6f-6bea798b3f0b", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-2084216438-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "eb811e38506b4297bf01bc5f2e098370", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "19671de9-8b5b-4710-adc3-7419f3c0f171", "external-id": "nsx-vlan-transportzone-421", "segmentation_id": 421, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap13050a31-71", "ovs_interfaceid": "13050a31-7165-4b22-83cc-d8ce9a41eadf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1104.288800] env[65726]: WARNING openstack [req-86187dd3-4838-4e1f-bf0d-02309cffabe0 req-f30167b4-1658-4552-a4b8-05b102c2bd6a service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1104.288800] env[65726]: WARNING openstack [req-86187dd3-4838-4e1f-bf0d-02309cffabe0 req-f30167b4-1658-4552-a4b8-05b102c2bd6a service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1104.399930] env[65726]: DEBUG oslo_vmware.api [None req-5eb7fa6b-cbd6-4a47-b227-e9c16f779f4a tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]522dddb0-3a6a-3fa0-4ac5-a557e118b075, 'name': SearchDatastore_Task, 'duration_secs': 0.10132} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1104.403463] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5eb7fa6b-cbd6-4a47-b227-e9c16f779f4a tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1104.403672] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-5eb7fa6b-cbd6-4a47-b227-e9c16f779f4a tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] [instance: 4809fcf6-59d4-409c-bd8d-981dc85a686f] Processing image b52362a3-ee8a-4cbf-b06f-513b0cc8f95c {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1104.403933] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5eb7fa6b-cbd6-4a47-b227-e9c16f779f4a tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1104.404101] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5eb7fa6b-cbd6-4a47-b227-e9c16f779f4a tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1104.404296] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-5eb7fa6b-cbd6-4a47-b227-e9c16f779f4a tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1104.404587] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-cda93265-b3e0-458e-9a36-4cb92984a0ea {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.413886] env[65726]: DEBUG oslo_vmware.api [None req-1c19f444-aee2-49ca-b67e-d3cef83d3664 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Task: {'id': task-5116725, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.786335} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1104.414364] env[65726]: INFO nova.virt.vmwareapi.ds_util [None req-1c19f444-aee2-49ca-b67e-d3cef83d3664 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Moved virtual disk from [datastore1] OSTACK_IMG_5f557822-0d62-4142-9f38-0161dfdccbb6/OSTACK_IMG_5f557822-0d62-4142-9f38-0161dfdccbb6.vmdk to [datastore1] devstack-image-cache_base/eb421501-6112-4a68-9c48-c82505e774d2/eb421501-6112-4a68-9c48-c82505e774d2.vmdk. [ 1104.414549] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-1c19f444-aee2-49ca-b67e-d3cef83d3664 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] [instance: 6fd70b4a-63e4-4258-9cff-f3b582500b9f] Cleaning up location [datastore1] OSTACK_IMG_5f557822-0d62-4142-9f38-0161dfdccbb6 {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 1104.414708] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-1c19f444-aee2-49ca-b67e-d3cef83d3664 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Deleting the datastore file [datastore1] OSTACK_IMG_5f557822-0d62-4142-9f38-0161dfdccbb6 {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1104.414987] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b6b84aac-31d5-40aa-aecf-862ec08faf46 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.424995] env[65726]: DEBUG oslo_vmware.api [None req-1c19f444-aee2-49ca-b67e-d3cef83d3664 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Waiting for the task: (returnval){ [ 1104.424995] env[65726]: value = "task-5116732" [ 1104.424995] env[65726]: _type = "Task" [ 1104.424995] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1104.433479] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-5eb7fa6b-cbd6-4a47-b227-e9c16f779f4a tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1104.433479] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-5eb7fa6b-cbd6-4a47-b227-e9c16f779f4a tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65726) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1104.433978] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-765ad3c4-2d32-42ed-b39d-268caf118c36 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.440756] env[65726]: DEBUG oslo_vmware.api [None req-1c19f444-aee2-49ca-b67e-d3cef83d3664 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Task: {'id': task-5116732, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1104.445361] env[65726]: DEBUG oslo_vmware.api [None req-5eb7fa6b-cbd6-4a47-b227-e9c16f779f4a tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Waiting for the task: (returnval){ [ 1104.445361] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5266b1c5-4efe-0d0b-0d59-914d8d03431f" [ 1104.445361] env[65726]: _type = "Task" [ 1104.445361] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1104.447688] env[65726]: DEBUG nova.network.neutron [req-86187dd3-4838-4e1f-bf0d-02309cffabe0 req-f30167b4-1658-4552-a4b8-05b102c2bd6a service nova] [instance: df08d166-7811-49b5-9c16-a3434229d623] Updated VIF entry in instance network info cache for port 70efdc7d-7775-4990-b851-d60195bd1504. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 1104.448130] env[65726]: DEBUG nova.network.neutron [req-86187dd3-4838-4e1f-bf0d-02309cffabe0 req-f30167b4-1658-4552-a4b8-05b102c2bd6a service nova] [instance: df08d166-7811-49b5-9c16-a3434229d623] Updating instance_info_cache with network_info: [{"id": "70efdc7d-7775-4990-b851-d60195bd1504", "address": "fa:16:3e:be:5b:19", "network": {"id": "835ec128-d214-4f54-ab53-88f8378b7d74", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-1935476539-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.151", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "305ecace479440ad9a616e546ccd38ce", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3b107fab-ee71-47db-ad4d-3c6f05546843", "external-id": "cl2-zone-554", "segmentation_id": 554, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap70efdc7d-77", "ovs_interfaceid": "70efdc7d-7775-4990-b851-d60195bd1504", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1104.462840] env[65726]: DEBUG oslo_vmware.api [None req-5eb7fa6b-cbd6-4a47-b227-e9c16f779f4a tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5266b1c5-4efe-0d0b-0d59-914d8d03431f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1104.475631] env[65726]: DEBUG oslo_concurrency.lockutils [None req-eab5602f-7673-4d76-aa16-7ebf0d2c87e3 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.003s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1104.489526] env[65726]: DEBUG oslo_vmware.api [None req-1dba9ad2-d4be-41a5-9681-c8a1af4e365c tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] Task: {'id': task-5116731, 'name': ReconfigVM_Task, 'duration_secs': 0.161487} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1104.490102] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-1dba9ad2-d4be-41a5-9681-c8a1af4e365c tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] [instance: df08d166-7811-49b5-9c16-a3434229d623] Reconfigured VM instance to set the machine id {{(pid=65726) _set_machine_id /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1799}} [ 1104.501017] env[65726]: INFO nova.scheduler.client.report [None req-eab5602f-7673-4d76-aa16-7ebf0d2c87e3 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Deleted allocations for instance 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04 [ 1104.740134] env[65726]: DEBUG nova.compute.manager [None req-aea91b5d-face-4f98-b031-6ac6f0d1a83b tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 0361ae8f-634d-4161-a594-7f93dbb1b7cd] Stashing vm_state: active {{(pid=65726) _prep_resize /opt/stack/nova/nova/compute/manager.py:6193}} [ 1104.755625] env[65726]: DEBUG oslo_concurrency.lockutils [req-8fdb4cdc-0ef8-4cde-8523-b085553a2a7e req-11e1b7dd-4719-4267-b009-34f0360365c6 service nova] Releasing lock "refresh_cache-4809fcf6-59d4-409c-bd8d-981dc85a686f" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1104.861430] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e0232c78-b0b0-436b-b3db-dd24c81ff983 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Acquiring lock "e7156a10-631a-4cc5-a544-88ce37763d02" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1104.861660] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e0232c78-b0b0-436b-b3db-dd24c81ff983 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Lock "e7156a10-631a-4cc5-a544-88ce37763d02" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1104.935474] env[65726]: DEBUG oslo_vmware.api [None req-1c19f444-aee2-49ca-b67e-d3cef83d3664 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Task: {'id': task-5116732, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.092232} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1104.935883] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-1c19f444-aee2-49ca-b67e-d3cef83d3664 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Deleted the datastore file {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1104.935985] env[65726]: DEBUG oslo_concurrency.lockutils [None req-1c19f444-aee2-49ca-b67e-d3cef83d3664 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Releasing lock "[datastore1] devstack-image-cache_base/eb421501-6112-4a68-9c48-c82505e774d2/eb421501-6112-4a68-9c48-c82505e774d2.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1104.936226] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-1c19f444-aee2-49ca-b67e-d3cef83d3664 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/eb421501-6112-4a68-9c48-c82505e774d2/eb421501-6112-4a68-9c48-c82505e774d2.vmdk to [datastore1] 6fd70b4a-63e4-4258-9cff-f3b582500b9f/6fd70b4a-63e4-4258-9cff-f3b582500b9f.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1104.936502] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-488eb775-f43f-4534-8d90-1428541fdda3 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.943925] env[65726]: DEBUG oslo_vmware.api [None req-1c19f444-aee2-49ca-b67e-d3cef83d3664 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Waiting for the task: (returnval){ [ 1104.943925] env[65726]: value = "task-5116733" [ 1104.943925] env[65726]: _type = "Task" [ 1104.943925] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1104.952218] env[65726]: DEBUG oslo_vmware.api [None req-1c19f444-aee2-49ca-b67e-d3cef83d3664 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Task: {'id': task-5116733, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1104.957816] env[65726]: DEBUG oslo_concurrency.lockutils [req-86187dd3-4838-4e1f-bf0d-02309cffabe0 req-f30167b4-1658-4552-a4b8-05b102c2bd6a service nova] Releasing lock "refresh_cache-df08d166-7811-49b5-9c16-a3434229d623" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1104.963935] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9575dab1-9756-447a-bbac-d03937ac3ee1 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] Acquiring lock "df08d166-7811-49b5-9c16-a3434229d623" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1104.964189] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9575dab1-9756-447a-bbac-d03937ac3ee1 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] Lock "df08d166-7811-49b5-9c16-a3434229d623" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1104.964384] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9575dab1-9756-447a-bbac-d03937ac3ee1 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] Acquiring lock "df08d166-7811-49b5-9c16-a3434229d623-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1104.964561] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9575dab1-9756-447a-bbac-d03937ac3ee1 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] Lock "df08d166-7811-49b5-9c16-a3434229d623-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1104.964725] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9575dab1-9756-447a-bbac-d03937ac3ee1 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] Lock "df08d166-7811-49b5-9c16-a3434229d623-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1104.966289] env[65726]: DEBUG oslo_vmware.api [None req-5eb7fa6b-cbd6-4a47-b227-e9c16f779f4a tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5266b1c5-4efe-0d0b-0d59-914d8d03431f, 'name': SearchDatastore_Task, 'duration_secs': 0.016777} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1104.966811] env[65726]: INFO nova.compute.manager [None req-9575dab1-9756-447a-bbac-d03937ac3ee1 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] [instance: df08d166-7811-49b5-9c16-a3434229d623] Terminating instance [ 1104.969378] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bf833ba7-e951-45e5-8205-ee0cc11665b3 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.976090] env[65726]: DEBUG oslo_vmware.api [None req-5eb7fa6b-cbd6-4a47-b227-e9c16f779f4a tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Waiting for the task: (returnval){ [ 1104.976090] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52512c5d-69b0-d5a7-31ca-c28adaafc48f" [ 1104.976090] env[65726]: _type = "Task" [ 1104.976090] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1104.987044] env[65726]: DEBUG oslo_vmware.api [None req-5eb7fa6b-cbd6-4a47-b227-e9c16f779f4a tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52512c5d-69b0-d5a7-31ca-c28adaafc48f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1105.006280] env[65726]: DEBUG oslo_concurrency.lockutils [None req-eab5602f-7673-4d76-aa16-7ebf0d2c87e3 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Lock "2fe6fc5a-f5c7-4f8e-96df-4e621a252f04" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 12.109s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1105.188129] env[65726]: DEBUG nova.compute.manager [req-e3cd780a-3977-42b9-9de6-c22eee41d1ce req-dda86d6e-b474-4235-a7e5-b3f5d0436da8 service nova] [instance: a56be6fa-6027-46ee-9aa8-332fe4222d68] Received event network-vif-deleted-4e2dfbcd-41c3-4aaa-a2ec-50c855c12307 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1105.188129] env[65726]: INFO nova.compute.manager [req-e3cd780a-3977-42b9-9de6-c22eee41d1ce req-dda86d6e-b474-4235-a7e5-b3f5d0436da8 service nova] [instance: a56be6fa-6027-46ee-9aa8-332fe4222d68] Neutron deleted interface 4e2dfbcd-41c3-4aaa-a2ec-50c855c12307; detaching it from the instance and deleting it from the info cache [ 1105.188129] env[65726]: DEBUG nova.network.neutron [req-e3cd780a-3977-42b9-9de6-c22eee41d1ce req-dda86d6e-b474-4235-a7e5-b3f5d0436da8 service nova] [instance: a56be6fa-6027-46ee-9aa8-332fe4222d68] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1105.257056] env[65726]: DEBUG nova.network.neutron [-] [instance: a56be6fa-6027-46ee-9aa8-332fe4222d68] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1105.264519] env[65726]: DEBUG oslo_concurrency.lockutils [None req-aea91b5d-face-4f98-b031-6ac6f0d1a83b tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1105.264805] env[65726]: DEBUG oslo_concurrency.lockutils [None req-aea91b5d-face-4f98-b031-6ac6f0d1a83b tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1105.364403] env[65726]: DEBUG nova.compute.manager [None req-e0232c78-b0b0-436b-b3db-dd24c81ff983 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: e7156a10-631a-4cc5-a544-88ce37763d02] Starting instance... {{(pid=65726) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 1105.460317] env[65726]: DEBUG oslo_vmware.api [None req-1c19f444-aee2-49ca-b67e-d3cef83d3664 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Task: {'id': task-5116733, 'name': CopyVirtualDisk_Task} progress is 9%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1105.475039] env[65726]: DEBUG nova.compute.manager [None req-9575dab1-9756-447a-bbac-d03937ac3ee1 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] [instance: df08d166-7811-49b5-9c16-a3434229d623] Start destroying the instance on the hypervisor. {{(pid=65726) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1105.475039] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-9575dab1-9756-447a-bbac-d03937ac3ee1 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] [instance: df08d166-7811-49b5-9c16-a3434229d623] Destroying instance {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1105.475438] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-754c923b-be2e-49e6-b5c6-debd5b6285cf {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.493058] env[65726]: DEBUG oslo_vmware.api [None req-5eb7fa6b-cbd6-4a47-b227-e9c16f779f4a tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52512c5d-69b0-d5a7-31ca-c28adaafc48f, 'name': SearchDatastore_Task, 'duration_secs': 0.017754} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1105.495772] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5eb7fa6b-cbd6-4a47-b227-e9c16f779f4a tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1105.496137] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-5eb7fa6b-cbd6-4a47-b227-e9c16f779f4a tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] 4809fcf6-59d4-409c-bd8d-981dc85a686f/4809fcf6-59d4-409c-bd8d-981dc85a686f.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1105.496525] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-9575dab1-9756-447a-bbac-d03937ac3ee1 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] [instance: df08d166-7811-49b5-9c16-a3434229d623] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1105.496780] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-283de53b-6a76-49ab-b8c9-b92bc2b4ead0 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.499226] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5a058fad-7c40-4f2b-bb55-fc6870197d3b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.510319] env[65726]: DEBUG oslo_vmware.api [None req-9575dab1-9756-447a-bbac-d03937ac3ee1 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] Waiting for the task: (returnval){ [ 1105.510319] env[65726]: value = "task-5116734" [ 1105.510319] env[65726]: _type = "Task" [ 1105.510319] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1105.512582] env[65726]: DEBUG oslo_vmware.api [None req-5eb7fa6b-cbd6-4a47-b227-e9c16f779f4a tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Waiting for the task: (returnval){ [ 1105.512582] env[65726]: value = "task-5116735" [ 1105.512582] env[65726]: _type = "Task" [ 1105.512582] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1105.532960] env[65726]: DEBUG oslo_vmware.api [None req-5eb7fa6b-cbd6-4a47-b227-e9c16f779f4a tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Task: {'id': task-5116735, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1105.533274] env[65726]: DEBUG oslo_vmware.api [None req-9575dab1-9756-447a-bbac-d03937ac3ee1 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] Task: {'id': task-5116734, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1105.691332] env[65726]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2bf37695-a385-4252-90de-a1084e36993f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.706681] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9131cb8f-7f25-4f1e-8f25-77640bb46d05 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.748758] env[65726]: DEBUG nova.compute.manager [req-e3cd780a-3977-42b9-9de6-c22eee41d1ce req-dda86d6e-b474-4235-a7e5-b3f5d0436da8 service nova] [instance: a56be6fa-6027-46ee-9aa8-332fe4222d68] Detach interface failed, port_id=4e2dfbcd-41c3-4aaa-a2ec-50c855c12307, reason: Instance a56be6fa-6027-46ee-9aa8-332fe4222d68 could not be found. {{(pid=65726) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11669}} [ 1105.761086] env[65726]: INFO nova.compute.manager [-] [instance: a56be6fa-6027-46ee-9aa8-332fe4222d68] Took 1.62 seconds to deallocate network for instance. [ 1105.769587] env[65726]: INFO nova.compute.claims [None req-aea91b5d-face-4f98-b031-6ac6f0d1a83b tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 0361ae8f-634d-4161-a594-7f93dbb1b7cd] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1105.890342] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e0232c78-b0b0-436b-b3db-dd24c81ff983 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1105.959786] env[65726]: DEBUG oslo_vmware.api [None req-1c19f444-aee2-49ca-b67e-d3cef83d3664 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Task: {'id': task-5116733, 'name': CopyVirtualDisk_Task} progress is 26%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1106.026643] env[65726]: DEBUG oslo_vmware.api [None req-9575dab1-9756-447a-bbac-d03937ac3ee1 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] Task: {'id': task-5116734, 'name': PowerOffVM_Task, 'duration_secs': 0.216539} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1106.030183] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-9575dab1-9756-447a-bbac-d03937ac3ee1 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] [instance: df08d166-7811-49b5-9c16-a3434229d623] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1106.030409] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-9575dab1-9756-447a-bbac-d03937ac3ee1 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] [instance: df08d166-7811-49b5-9c16-a3434229d623] Unregistering the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1106.030748] env[65726]: DEBUG oslo_vmware.api [None req-5eb7fa6b-cbd6-4a47-b227-e9c16f779f4a tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Task: {'id': task-5116735, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1106.031011] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c10cb6aa-9bde-4e01-a671-3351881c2942 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.116156] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-9575dab1-9756-447a-bbac-d03937ac3ee1 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] [instance: df08d166-7811-49b5-9c16-a3434229d623] Unregistered the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1106.116508] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-9575dab1-9756-447a-bbac-d03937ac3ee1 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] [instance: df08d166-7811-49b5-9c16-a3434229d623] Deleting contents of the VM from datastore datastore2 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1106.116820] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-9575dab1-9756-447a-bbac-d03937ac3ee1 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] Deleting the datastore file [datastore2] df08d166-7811-49b5-9c16-a3434229d623 {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1106.117237] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-539ca84d-9fd8-4a40-a1a7-1174ed9d20d0 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.128201] env[65726]: DEBUG oslo_vmware.api [None req-9575dab1-9756-447a-bbac-d03937ac3ee1 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] Waiting for the task: (returnval){ [ 1106.128201] env[65726]: value = "task-5116737" [ 1106.128201] env[65726]: _type = "Task" [ 1106.128201] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1106.141831] env[65726]: DEBUG oslo_vmware.api [None req-9575dab1-9756-447a-bbac-d03937ac3ee1 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] Task: {'id': task-5116737, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1106.196335] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-be8a3d8a-c73e-4be5-b722-9654488b1331 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: ad4f00b5-3765-494f-8132-8b17b29d219a] Volume attach. Driver type: vmdk {{(pid=65726) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1106.196591] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-be8a3d8a-c73e-4be5-b722-9654488b1331 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: ad4f00b5-3765-494f-8132-8b17b29d219a] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-995298', 'volume_id': '1c311bc8-792e-43d1-b5e9-6105ef7b377e', 'name': 'volume-1c311bc8-792e-43d1-b5e9-6105ef7b377e', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'ad4f00b5-3765-494f-8132-8b17b29d219a', 'attached_at': '', 'detached_at': '', 'volume_id': '1c311bc8-792e-43d1-b5e9-6105ef7b377e', 'serial': '1c311bc8-792e-43d1-b5e9-6105ef7b377e'} {{(pid=65726) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1106.197765] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-154ab146-03fe-4531-90cc-cdaa9707d685 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.220256] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4c7dd47-d064-47eb-bf34-dbfb820da1ae {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.249226] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-be8a3d8a-c73e-4be5-b722-9654488b1331 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: ad4f00b5-3765-494f-8132-8b17b29d219a] Reconfiguring VM instance instance-0000005b to attach disk [datastore1] volume-1c311bc8-792e-43d1-b5e9-6105ef7b377e/volume-1c311bc8-792e-43d1-b5e9-6105ef7b377e.vmdk or device None with type thin {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1106.249686] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8ee0abc7-e607-4d25-8d16-de3fa45463d6 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.271749] env[65726]: DEBUG oslo_vmware.api [None req-be8a3d8a-c73e-4be5-b722-9654488b1331 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Waiting for the task: (returnval){ [ 1106.271749] env[65726]: value = "task-5116738" [ 1106.271749] env[65726]: _type = "Task" [ 1106.271749] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1106.277343] env[65726]: INFO nova.compute.resource_tracker [None req-aea91b5d-face-4f98-b031-6ac6f0d1a83b tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 0361ae8f-634d-4161-a594-7f93dbb1b7cd] Updating resource usage from migration 995b2821-290f-47ee-adb1-3c8a9e262c05 [ 1106.289674] env[65726]: DEBUG oslo_vmware.api [None req-be8a3d8a-c73e-4be5-b722-9654488b1331 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Task: {'id': task-5116738, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1106.320243] env[65726]: INFO nova.compute.manager [None req-0288a821-ce9c-40cd-86ef-962ac009580c tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: a56be6fa-6027-46ee-9aa8-332fe4222d68] Took 0.56 seconds to detach 1 volumes for instance. [ 1106.460597] env[65726]: DEBUG oslo_vmware.api [None req-1c19f444-aee2-49ca-b67e-d3cef83d3664 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Task: {'id': task-5116733, 'name': CopyVirtualDisk_Task} progress is 46%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1106.499779] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97b5587b-a712-42d5-9a05-67ce66dee7d3 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.512070] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ea9b0fd-ef03-471b-9967-1c50c09cfa60 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.559802] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c1acb9a-4c3a-44d3-9ea7-026de566f1d8 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.569541] env[65726]: DEBUG oslo_vmware.api [None req-5eb7fa6b-cbd6-4a47-b227-e9c16f779f4a tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Task: {'id': task-5116735, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1106.574722] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51132e94-aefa-472a-9ea5-8d0d7f160f3f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.594272] env[65726]: DEBUG nova.compute.provider_tree [None req-aea91b5d-face-4f98-b031-6ac6f0d1a83b tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1106.644209] env[65726]: DEBUG oslo_vmware.api [None req-9575dab1-9756-447a-bbac-d03937ac3ee1 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] Task: {'id': task-5116737, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1106.785417] env[65726]: DEBUG oslo_vmware.api [None req-be8a3d8a-c73e-4be5-b722-9654488b1331 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Task: {'id': task-5116738, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1106.831984] env[65726]: DEBUG oslo_concurrency.lockutils [None req-0288a821-ce9c-40cd-86ef-962ac009580c tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1106.960316] env[65726]: DEBUG oslo_vmware.api [None req-1c19f444-aee2-49ca-b67e-d3cef83d3664 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Task: {'id': task-5116733, 'name': CopyVirtualDisk_Task} progress is 66%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1107.038614] env[65726]: DEBUG oslo_vmware.api [None req-5eb7fa6b-cbd6-4a47-b227-e9c16f779f4a tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Task: {'id': task-5116735, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1107.101320] env[65726]: DEBUG nova.scheduler.client.report [None req-aea91b5d-face-4f98-b031-6ac6f0d1a83b tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1107.143945] env[65726]: DEBUG oslo_vmware.api [None req-9575dab1-9756-447a-bbac-d03937ac3ee1 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] Task: {'id': task-5116737, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1107.286512] env[65726]: DEBUG oslo_vmware.api [None req-be8a3d8a-c73e-4be5-b722-9654488b1331 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Task: {'id': task-5116738, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1107.460300] env[65726]: DEBUG oslo_vmware.api [None req-1c19f444-aee2-49ca-b67e-d3cef83d3664 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Task: {'id': task-5116733, 'name': CopyVirtualDisk_Task} progress is 88%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1107.533693] env[65726]: DEBUG oslo_vmware.api [None req-5eb7fa6b-cbd6-4a47-b227-e9c16f779f4a tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Task: {'id': task-5116735, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1107.607069] env[65726]: DEBUG oslo_concurrency.lockutils [None req-aea91b5d-face-4f98-b031-6ac6f0d1a83b tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.342s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1107.607317] env[65726]: INFO nova.compute.manager [None req-aea91b5d-face-4f98-b031-6ac6f0d1a83b tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 0361ae8f-634d-4161-a594-7f93dbb1b7cd] Migrating [ 1107.620356] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e0232c78-b0b0-436b-b3db-dd24c81ff983 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.730s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1107.622283] env[65726]: INFO nova.compute.claims [None req-e0232c78-b0b0-436b-b3db-dd24c81ff983 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: e7156a10-631a-4cc5-a544-88ce37763d02] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1107.653400] env[65726]: DEBUG oslo_vmware.api [None req-9575dab1-9756-447a-bbac-d03937ac3ee1 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] Task: {'id': task-5116737, 'name': DeleteDatastoreFile_Task, 'duration_secs': 1.057955} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1107.653903] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-9575dab1-9756-447a-bbac-d03937ac3ee1 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] Deleted the datastore file {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1107.654252] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-9575dab1-9756-447a-bbac-d03937ac3ee1 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] [instance: df08d166-7811-49b5-9c16-a3434229d623] Deleted contents of the VM from datastore datastore2 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1107.654610] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-9575dab1-9756-447a-bbac-d03937ac3ee1 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] [instance: df08d166-7811-49b5-9c16-a3434229d623] Instance destroyed {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1107.654964] env[65726]: INFO nova.compute.manager [None req-9575dab1-9756-447a-bbac-d03937ac3ee1 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] [instance: df08d166-7811-49b5-9c16-a3434229d623] Took 2.18 seconds to destroy the instance on the hypervisor. [ 1107.656303] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-9575dab1-9756-447a-bbac-d03937ac3ee1 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1107.656303] env[65726]: DEBUG nova.compute.manager [-] [instance: df08d166-7811-49b5-9c16-a3434229d623] Deallocating network for instance {{(pid=65726) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1107.656303] env[65726]: DEBUG nova.network.neutron [-] [instance: df08d166-7811-49b5-9c16-a3434229d623] deallocate_for_instance() {{(pid=65726) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 1107.657104] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1107.657433] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1107.719849] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1107.719849] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1107.788287] env[65726]: DEBUG oslo_vmware.api [None req-be8a3d8a-c73e-4be5-b722-9654488b1331 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Task: {'id': task-5116738, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1107.962274] env[65726]: DEBUG oslo_vmware.api [None req-1c19f444-aee2-49ca-b67e-d3cef83d3664 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Task: {'id': task-5116733, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.991568} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1107.962837] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-1c19f444-aee2-49ca-b67e-d3cef83d3664 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/eb421501-6112-4a68-9c48-c82505e774d2/eb421501-6112-4a68-9c48-c82505e774d2.vmdk to [datastore1] 6fd70b4a-63e4-4258-9cff-f3b582500b9f/6fd70b4a-63e4-4258-9cff-f3b582500b9f.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1107.964292] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67da6aa6-71d4-4d09-8831-c1e1f0183126 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.998678] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-1c19f444-aee2-49ca-b67e-d3cef83d3664 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] [instance: 6fd70b4a-63e4-4258-9cff-f3b582500b9f] Reconfiguring VM instance instance-0000004c to attach disk [datastore1] 6fd70b4a-63e4-4258-9cff-f3b582500b9f/6fd70b4a-63e4-4258-9cff-f3b582500b9f.vmdk or device None with type streamOptimized {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1107.999123] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a20d20e5-181d-4a76-9e97-f61626caaaf2 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.021273] env[65726]: DEBUG oslo_vmware.api [None req-1c19f444-aee2-49ca-b67e-d3cef83d3664 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Waiting for the task: (returnval){ [ 1108.021273] env[65726]: value = "task-5116739" [ 1108.021273] env[65726]: _type = "Task" [ 1108.021273] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1108.033351] env[65726]: DEBUG oslo_concurrency.lockutils [None req-96c35e52-5464-43ed-b438-b393e0533d7f tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Acquiring lock "b353b5e9-500c-42d3-a87f-880a624febec" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1108.033628] env[65726]: DEBUG oslo_concurrency.lockutils [None req-96c35e52-5464-43ed-b438-b393e0533d7f tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Lock "b353b5e9-500c-42d3-a87f-880a624febec" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1108.039996] env[65726]: DEBUG oslo_vmware.api [None req-5eb7fa6b-cbd6-4a47-b227-e9c16f779f4a tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Task: {'id': task-5116735, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.494903} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1108.040588] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-5eb7fa6b-cbd6-4a47-b227-e9c16f779f4a tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] 4809fcf6-59d4-409c-bd8d-981dc85a686f/4809fcf6-59d4-409c-bd8d-981dc85a686f.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1108.040900] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-5eb7fa6b-cbd6-4a47-b227-e9c16f779f4a tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] [instance: 4809fcf6-59d4-409c-bd8d-981dc85a686f] Extending root virtual disk to 1048576 {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1108.045103] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-041f2d9c-0636-49f1-89a3-31f999e914df {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.048222] env[65726]: DEBUG oslo_vmware.api [None req-1c19f444-aee2-49ca-b67e-d3cef83d3664 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Task: {'id': task-5116739, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1108.054745] env[65726]: DEBUG oslo_vmware.api [None req-5eb7fa6b-cbd6-4a47-b227-e9c16f779f4a tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Waiting for the task: (returnval){ [ 1108.054745] env[65726]: value = "task-5116740" [ 1108.054745] env[65726]: _type = "Task" [ 1108.054745] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1108.070564] env[65726]: DEBUG oslo_vmware.api [None req-5eb7fa6b-cbd6-4a47-b227-e9c16f779f4a tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Task: {'id': task-5116740, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1108.135486] env[65726]: DEBUG oslo_concurrency.lockutils [None req-aea91b5d-face-4f98-b031-6ac6f0d1a83b tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Acquiring lock "refresh_cache-0361ae8f-634d-4161-a594-7f93dbb1b7cd" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1108.135486] env[65726]: DEBUG oslo_concurrency.lockutils [None req-aea91b5d-face-4f98-b031-6ac6f0d1a83b tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Acquired lock "refresh_cache-0361ae8f-634d-4161-a594-7f93dbb1b7cd" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1108.135486] env[65726]: DEBUG nova.network.neutron [None req-aea91b5d-face-4f98-b031-6ac6f0d1a83b tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 0361ae8f-634d-4161-a594-7f93dbb1b7cd] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1108.215285] env[65726]: DEBUG nova.compute.manager [req-86e1bfdf-8ac2-4bf8-89d3-1ed95debcb5d req-8f872058-77d3-4e03-b66e-5a8a509daedf service nova] [instance: df08d166-7811-49b5-9c16-a3434229d623] Received event network-vif-deleted-70efdc7d-7775-4990-b851-d60195bd1504 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1108.215528] env[65726]: INFO nova.compute.manager [req-86e1bfdf-8ac2-4bf8-89d3-1ed95debcb5d req-8f872058-77d3-4e03-b66e-5a8a509daedf service nova] [instance: df08d166-7811-49b5-9c16-a3434229d623] Neutron deleted interface 70efdc7d-7775-4990-b851-d60195bd1504; detaching it from the instance and deleting it from the info cache [ 1108.215669] env[65726]: DEBUG nova.network.neutron [req-86e1bfdf-8ac2-4bf8-89d3-1ed95debcb5d req-8f872058-77d3-4e03-b66e-5a8a509daedf service nova] [instance: df08d166-7811-49b5-9c16-a3434229d623] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1108.284488] env[65726]: DEBUG oslo_vmware.api [None req-be8a3d8a-c73e-4be5-b722-9654488b1331 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Task: {'id': task-5116738, 'name': ReconfigVM_Task, 'duration_secs': 1.610216} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1108.284734] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-be8a3d8a-c73e-4be5-b722-9654488b1331 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: ad4f00b5-3765-494f-8132-8b17b29d219a] Reconfigured VM instance instance-0000005b to attach disk [datastore1] volume-1c311bc8-792e-43d1-b5e9-6105ef7b377e/volume-1c311bc8-792e-43d1-b5e9-6105ef7b377e.vmdk or device None with type thin {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1108.289806] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-47eca6da-a517-4868-af43-3e4afda61fd1 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.306898] env[65726]: DEBUG oslo_vmware.api [None req-be8a3d8a-c73e-4be5-b722-9654488b1331 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Waiting for the task: (returnval){ [ 1108.306898] env[65726]: value = "task-5116741" [ 1108.306898] env[65726]: _type = "Task" [ 1108.306898] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1108.318417] env[65726]: DEBUG oslo_vmware.api [None req-be8a3d8a-c73e-4be5-b722-9654488b1331 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Task: {'id': task-5116741, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1108.535964] env[65726]: DEBUG oslo_vmware.api [None req-1c19f444-aee2-49ca-b67e-d3cef83d3664 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Task: {'id': task-5116739, 'name': ReconfigVM_Task, 'duration_secs': 0.362121} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1108.536249] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-1c19f444-aee2-49ca-b67e-d3cef83d3664 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] [instance: 6fd70b4a-63e4-4258-9cff-f3b582500b9f] Reconfigured VM instance instance-0000004c to attach disk [datastore1] 6fd70b4a-63e4-4258-9cff-f3b582500b9f/6fd70b4a-63e4-4258-9cff-f3b582500b9f.vmdk or device None with type streamOptimized {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1108.536918] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-493a2c54-f9ab-49c0-a593-3ecd57432747 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.541442] env[65726]: DEBUG nova.compute.manager [None req-96c35e52-5464-43ed-b438-b393e0533d7f tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: b353b5e9-500c-42d3-a87f-880a624febec] Starting instance... {{(pid=65726) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 1108.545569] env[65726]: DEBUG oslo_vmware.api [None req-1c19f444-aee2-49ca-b67e-d3cef83d3664 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Waiting for the task: (returnval){ [ 1108.545569] env[65726]: value = "task-5116742" [ 1108.545569] env[65726]: _type = "Task" [ 1108.545569] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1108.556078] env[65726]: DEBUG oslo_vmware.api [None req-1c19f444-aee2-49ca-b67e-d3cef83d3664 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Task: {'id': task-5116742, 'name': Rename_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1108.565056] env[65726]: DEBUG oslo_vmware.api [None req-5eb7fa6b-cbd6-4a47-b227-e9c16f779f4a tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Task: {'id': task-5116740, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.070884} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1108.565388] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-5eb7fa6b-cbd6-4a47-b227-e9c16f779f4a tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] [instance: 4809fcf6-59d4-409c-bd8d-981dc85a686f] Extended root virtual disk {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1108.566392] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-572ad75c-7a99-49eb-8656-4cdddecc1936 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.591353] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-5eb7fa6b-cbd6-4a47-b227-e9c16f779f4a tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] [instance: 4809fcf6-59d4-409c-bd8d-981dc85a686f] Reconfiguring VM instance instance-00000065 to attach disk [datastore1] 4809fcf6-59d4-409c-bd8d-981dc85a686f/4809fcf6-59d4-409c-bd8d-981dc85a686f.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1108.592456] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-61ee90b7-23d9-4b27-8ca8-c13467cd536b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.613830] env[65726]: DEBUG oslo_vmware.api [None req-5eb7fa6b-cbd6-4a47-b227-e9c16f779f4a tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Waiting for the task: (returnval){ [ 1108.613830] env[65726]: value = "task-5116743" [ 1108.613830] env[65726]: _type = "Task" [ 1108.613830] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1108.622698] env[65726]: DEBUG oslo_vmware.api [None req-5eb7fa6b-cbd6-4a47-b227-e9c16f779f4a tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Task: {'id': task-5116743, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1108.638971] env[65726]: WARNING openstack [None req-aea91b5d-face-4f98-b031-6ac6f0d1a83b tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1108.639550] env[65726]: WARNING openstack [None req-aea91b5d-face-4f98-b031-6ac6f0d1a83b tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1108.697468] env[65726]: DEBUG nova.network.neutron [-] [instance: df08d166-7811-49b5-9c16-a3434229d623] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1108.720561] env[65726]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-609af6a0-ee70-4f59-8791-33e4c5d9e5c0 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.738625] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3923288-b3db-42d6-940f-f7f33737d63d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.755067] env[65726]: WARNING openstack [None req-aea91b5d-face-4f98-b031-6ac6f0d1a83b tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1108.755495] env[65726]: WARNING openstack [None req-aea91b5d-face-4f98-b031-6ac6f0d1a83b tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1108.789097] env[65726]: DEBUG nova.compute.manager [req-86e1bfdf-8ac2-4bf8-89d3-1ed95debcb5d req-8f872058-77d3-4e03-b66e-5a8a509daedf service nova] [instance: df08d166-7811-49b5-9c16-a3434229d623] Detach interface failed, port_id=70efdc7d-7775-4990-b851-d60195bd1504, reason: Instance df08d166-7811-49b5-9c16-a3434229d623 could not be found. {{(pid=65726) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11669}} [ 1108.817690] env[65726]: DEBUG oslo_vmware.api [None req-be8a3d8a-c73e-4be5-b722-9654488b1331 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Task: {'id': task-5116741, 'name': ReconfigVM_Task, 'duration_secs': 0.203683} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1108.818086] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-be8a3d8a-c73e-4be5-b722-9654488b1331 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: ad4f00b5-3765-494f-8132-8b17b29d219a] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-995298', 'volume_id': '1c311bc8-792e-43d1-b5e9-6105ef7b377e', 'name': 'volume-1c311bc8-792e-43d1-b5e9-6105ef7b377e', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'ad4f00b5-3765-494f-8132-8b17b29d219a', 'attached_at': '', 'detached_at': '', 'volume_id': '1c311bc8-792e-43d1-b5e9-6105ef7b377e', 'serial': '1c311bc8-792e-43d1-b5e9-6105ef7b377e'} {{(pid=65726) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1108.829886] env[65726]: WARNING openstack [None req-aea91b5d-face-4f98-b031-6ac6f0d1a83b tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1108.830285] env[65726]: WARNING openstack [None req-aea91b5d-face-4f98-b031-6ac6f0d1a83b tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1108.880455] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a2c4469-44e0-434b-9200-100c78b86f9d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.888970] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a291a396-f142-4228-b0de-0083569fb5ad {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.922157] env[65726]: DEBUG nova.network.neutron [None req-aea91b5d-face-4f98-b031-6ac6f0d1a83b tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 0361ae8f-634d-4161-a594-7f93dbb1b7cd] Updating instance_info_cache with network_info: [{"id": "2bfd7147-87fd-4a19-b28b-eb95f51f4d00", "address": "fa:16:3e:d6:07:64", "network": {"id": "9dd55fb6-0dd2-484d-b94d-fbdccb958c79", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1005598846-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f93d852e2c904f42981cbfff882558e9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "74f30339-6421-4654-bddb-81d7f34db9d7", "external-id": "nsx-vlan-transportzone-899", "segmentation_id": 899, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2bfd7147-87", "ovs_interfaceid": "2bfd7147-87fd-4a19-b28b-eb95f51f4d00", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1108.923897] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed5e31f5-f488-4281-ad8a-2825a1994316 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.932179] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2cbf2716-49f2-478d-9564-0cf33d9fddd7 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.950962] env[65726]: DEBUG nova.compute.provider_tree [None req-e0232c78-b0b0-436b-b3db-dd24c81ff983 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1109.063247] env[65726]: DEBUG oslo_vmware.api [None req-1c19f444-aee2-49ca-b67e-d3cef83d3664 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Task: {'id': task-5116742, 'name': Rename_Task, 'duration_secs': 0.135858} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1109.063247] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-1c19f444-aee2-49ca-b67e-d3cef83d3664 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] [instance: 6fd70b4a-63e4-4258-9cff-f3b582500b9f] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1109.063247] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a4f8d57b-811b-43bd-b5f7-146aa66f5d08 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.068810] env[65726]: DEBUG oslo_concurrency.lockutils [None req-96c35e52-5464-43ed-b438-b393e0533d7f tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1109.070304] env[65726]: DEBUG oslo_vmware.api [None req-1c19f444-aee2-49ca-b67e-d3cef83d3664 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Waiting for the task: (returnval){ [ 1109.070304] env[65726]: value = "task-5116744" [ 1109.070304] env[65726]: _type = "Task" [ 1109.070304] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1109.079402] env[65726]: DEBUG oslo_vmware.api [None req-1c19f444-aee2-49ca-b67e-d3cef83d3664 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Task: {'id': task-5116744, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1109.128119] env[65726]: DEBUG oslo_vmware.api [None req-5eb7fa6b-cbd6-4a47-b227-e9c16f779f4a tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Task: {'id': task-5116743, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1109.201060] env[65726]: INFO nova.compute.manager [-] [instance: df08d166-7811-49b5-9c16-a3434229d623] Took 1.54 seconds to deallocate network for instance. [ 1109.428091] env[65726]: DEBUG oslo_concurrency.lockutils [None req-aea91b5d-face-4f98-b031-6ac6f0d1a83b tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Releasing lock "refresh_cache-0361ae8f-634d-4161-a594-7f93dbb1b7cd" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1109.453559] env[65726]: DEBUG nova.scheduler.client.report [None req-e0232c78-b0b0-436b-b3db-dd24c81ff983 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1109.582027] env[65726]: DEBUG oslo_vmware.api [None req-1c19f444-aee2-49ca-b67e-d3cef83d3664 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Task: {'id': task-5116744, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1109.625649] env[65726]: DEBUG oslo_vmware.api [None req-5eb7fa6b-cbd6-4a47-b227-e9c16f779f4a tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Task: {'id': task-5116743, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1109.709026] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9575dab1-9756-447a-bbac-d03937ac3ee1 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1109.870163] env[65726]: DEBUG nova.objects.instance [None req-be8a3d8a-c73e-4be5-b722-9654488b1331 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Lazy-loading 'flavor' on Instance uuid ad4f00b5-3765-494f-8132-8b17b29d219a {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1109.959000] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e0232c78-b0b0-436b-b3db-dd24c81ff983 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.339s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1109.959592] env[65726]: DEBUG nova.compute.manager [None req-e0232c78-b0b0-436b-b3db-dd24c81ff983 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: e7156a10-631a-4cc5-a544-88ce37763d02] Start building networks asynchronously for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1109.962248] env[65726]: DEBUG oslo_concurrency.lockutils [None req-0288a821-ce9c-40cd-86ef-962ac009580c tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.131s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1109.962520] env[65726]: DEBUG oslo_concurrency.lockutils [None req-0288a821-ce9c-40cd-86ef-962ac009580c tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1109.965169] env[65726]: DEBUG oslo_concurrency.lockutils [None req-96c35e52-5464-43ed-b438-b393e0533d7f tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.896s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1109.966703] env[65726]: INFO nova.compute.claims [None req-96c35e52-5464-43ed-b438-b393e0533d7f tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: b353b5e9-500c-42d3-a87f-880a624febec] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1109.997519] env[65726]: INFO nova.scheduler.client.report [None req-0288a821-ce9c-40cd-86ef-962ac009580c tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Deleted allocations for instance a56be6fa-6027-46ee-9aa8-332fe4222d68 [ 1110.082739] env[65726]: DEBUG oslo_vmware.api [None req-1c19f444-aee2-49ca-b67e-d3cef83d3664 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Task: {'id': task-5116744, 'name': PowerOnVM_Task, 'duration_secs': 0.535561} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1110.083065] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-1c19f444-aee2-49ca-b67e-d3cef83d3664 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] [instance: 6fd70b4a-63e4-4258-9cff-f3b582500b9f] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1110.127552] env[65726]: DEBUG oslo_vmware.api [None req-5eb7fa6b-cbd6-4a47-b227-e9c16f779f4a tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Task: {'id': task-5116743, 'name': ReconfigVM_Task, 'duration_secs': 1.107751} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1110.127741] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-5eb7fa6b-cbd6-4a47-b227-e9c16f779f4a tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] [instance: 4809fcf6-59d4-409c-bd8d-981dc85a686f] Reconfigured VM instance instance-00000065 to attach disk [datastore1] 4809fcf6-59d4-409c-bd8d-981dc85a686f/4809fcf6-59d4-409c-bd8d-981dc85a686f.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1110.128401] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9ab9344d-0be9-402b-9c6c-c6457a2ad1be {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.135730] env[65726]: DEBUG oslo_vmware.api [None req-5eb7fa6b-cbd6-4a47-b227-e9c16f779f4a tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Waiting for the task: (returnval){ [ 1110.135730] env[65726]: value = "task-5116745" [ 1110.135730] env[65726]: _type = "Task" [ 1110.135730] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1110.144272] env[65726]: DEBUG oslo_vmware.api [None req-5eb7fa6b-cbd6-4a47-b227-e9c16f779f4a tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Task: {'id': task-5116745, 'name': Rename_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1110.186158] env[65726]: DEBUG nova.compute.manager [None req-1c19f444-aee2-49ca-b67e-d3cef83d3664 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] [instance: 6fd70b4a-63e4-4258-9cff-f3b582500b9f] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1110.187425] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22f81631-a8c1-4126-8cd3-84a20de7b47d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.220709] env[65726]: DEBUG oslo_concurrency.lockutils [None req-44fd1bde-b4ab-4b7e-bddb-a60989ac9bd3 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Acquiring lock "ad4f00b5-3765-494f-8132-8b17b29d219a" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1110.374606] env[65726]: DEBUG oslo_concurrency.lockutils [None req-be8a3d8a-c73e-4be5-b722-9654488b1331 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Lock "ad4f00b5-3765-494f-8132-8b17b29d219a" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 8.851s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1110.375576] env[65726]: DEBUG oslo_concurrency.lockutils [None req-44fd1bde-b4ab-4b7e-bddb-a60989ac9bd3 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Lock "ad4f00b5-3765-494f-8132-8b17b29d219a" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.155s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1110.471157] env[65726]: DEBUG nova.compute.utils [None req-e0232c78-b0b0-436b-b3db-dd24c81ff983 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Using /dev/sd instead of None {{(pid=65726) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1110.474792] env[65726]: DEBUG nova.compute.manager [None req-e0232c78-b0b0-436b-b3db-dd24c81ff983 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: e7156a10-631a-4cc5-a544-88ce37763d02] Allocating IP information in the background. {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1110.474792] env[65726]: DEBUG nova.network.neutron [None req-e0232c78-b0b0-436b-b3db-dd24c81ff983 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: e7156a10-631a-4cc5-a544-88ce37763d02] allocate_for_instance() {{(pid=65726) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 1110.475017] env[65726]: WARNING neutronclient.v2_0.client [None req-e0232c78-b0b0-436b-b3db-dd24c81ff983 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1110.475337] env[65726]: WARNING neutronclient.v2_0.client [None req-e0232c78-b0b0-436b-b3db-dd24c81ff983 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1110.475920] env[65726]: WARNING openstack [None req-e0232c78-b0b0-436b-b3db-dd24c81ff983 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1110.476353] env[65726]: WARNING openstack [None req-e0232c78-b0b0-436b-b3db-dd24c81ff983 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1110.505270] env[65726]: DEBUG oslo_concurrency.lockutils [None req-0288a821-ce9c-40cd-86ef-962ac009580c tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Lock "a56be6fa-6027-46ee-9aa8-332fe4222d68" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 11.976s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1110.526623] env[65726]: DEBUG nova.policy [None req-e0232c78-b0b0-436b-b3db-dd24c81ff983 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4cd1ed63db174f21be78f86554fada0a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'cbb17a09b35c4c22ade5c4082b10bda9', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65726) authorize /opt/stack/nova/nova/policy.py:192}} [ 1110.647481] env[65726]: DEBUG oslo_vmware.api [None req-5eb7fa6b-cbd6-4a47-b227-e9c16f779f4a tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Task: {'id': task-5116745, 'name': Rename_Task, 'duration_secs': 0.147578} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1110.647933] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-5eb7fa6b-cbd6-4a47-b227-e9c16f779f4a tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] [instance: 4809fcf6-59d4-409c-bd8d-981dc85a686f] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1110.648332] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3656d02e-f49b-472d-b308-c77e4baf6b4e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.656478] env[65726]: DEBUG oslo_vmware.api [None req-5eb7fa6b-cbd6-4a47-b227-e9c16f779f4a tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Waiting for the task: (returnval){ [ 1110.656478] env[65726]: value = "task-5116746" [ 1110.656478] env[65726]: _type = "Task" [ 1110.656478] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1110.665512] env[65726]: DEBUG oslo_vmware.api [None req-5eb7fa6b-cbd6-4a47-b227-e9c16f779f4a tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Task: {'id': task-5116746, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1110.711532] env[65726]: DEBUG oslo_concurrency.lockutils [None req-1c19f444-aee2-49ca-b67e-d3cef83d3664 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Lock "6fd70b4a-63e4-4258-9cff-f3b582500b9f" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 24.378s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1110.775715] env[65726]: DEBUG nova.network.neutron [None req-e0232c78-b0b0-436b-b3db-dd24c81ff983 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: e7156a10-631a-4cc5-a544-88ce37763d02] Successfully created port: 44a1554d-d380-49b2-ad33-9a251114c93c {{(pid=65726) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1110.882462] env[65726]: INFO nova.compute.manager [None req-44fd1bde-b4ab-4b7e-bddb-a60989ac9bd3 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: ad4f00b5-3765-494f-8132-8b17b29d219a] Detaching volume 1c311bc8-792e-43d1-b5e9-6105ef7b377e [ 1110.920509] env[65726]: INFO nova.virt.block_device [None req-44fd1bde-b4ab-4b7e-bddb-a60989ac9bd3 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: ad4f00b5-3765-494f-8132-8b17b29d219a] Attempting to driver detach volume 1c311bc8-792e-43d1-b5e9-6105ef7b377e from mountpoint /dev/sdb [ 1110.920743] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-44fd1bde-b4ab-4b7e-bddb-a60989ac9bd3 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: ad4f00b5-3765-494f-8132-8b17b29d219a] Volume detach. Driver type: vmdk {{(pid=65726) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1110.920934] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-44fd1bde-b4ab-4b7e-bddb-a60989ac9bd3 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: ad4f00b5-3765-494f-8132-8b17b29d219a] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-995298', 'volume_id': '1c311bc8-792e-43d1-b5e9-6105ef7b377e', 'name': 'volume-1c311bc8-792e-43d1-b5e9-6105ef7b377e', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'ad4f00b5-3765-494f-8132-8b17b29d219a', 'attached_at': '', 'detached_at': '', 'volume_id': '1c311bc8-792e-43d1-b5e9-6105ef7b377e', 'serial': '1c311bc8-792e-43d1-b5e9-6105ef7b377e'} {{(pid=65726) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1110.921902] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78443708-15c8-4456-97c3-005a48661fc8 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.945918] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8c7a452-7135-4f03-8b4a-7bd82fbe7027 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.949630] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1bb01f0-5a95-488a-9f67-9cea3bf47e29 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.969818] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-aea91b5d-face-4f98-b031-6ac6f0d1a83b tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 0361ae8f-634d-4161-a594-7f93dbb1b7cd] Updating instance '0361ae8f-634d-4161-a594-7f93dbb1b7cd' progress to 0 {{(pid=65726) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1110.977057] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-119b91e0-ca17-4ec6-bd65-870f77aa22de {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.999641] env[65726]: DEBUG nova.compute.manager [None req-e0232c78-b0b0-436b-b3db-dd24c81ff983 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: e7156a10-631a-4cc5-a544-88ce37763d02] Start building block device mappings for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1111.007286] env[65726]: DEBUG oslo_concurrency.lockutils [None req-022a70bf-a10f-4483-94dc-011169b6c2d5 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Acquiring lock "interface-51f5337d-2e23-4b7c-b590-76be6bd82411-fb580bad-fd60-471d-95f2-066352417659" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1111.007286] env[65726]: DEBUG oslo_concurrency.lockutils [None req-022a70bf-a10f-4483-94dc-011169b6c2d5 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Lock "interface-51f5337d-2e23-4b7c-b590-76be6bd82411-fb580bad-fd60-471d-95f2-066352417659" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1111.007286] env[65726]: DEBUG nova.objects.instance [None req-022a70bf-a10f-4483-94dc-011169b6c2d5 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Lazy-loading 'flavor' on Instance uuid 51f5337d-2e23-4b7c-b590-76be6bd82411 {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1111.008872] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2042f2b1-7440-4e18-9404-e432e0f022df {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.030209] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-44fd1bde-b4ab-4b7e-bddb-a60989ac9bd3 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] The volume has not been displaced from its original location: [datastore1] volume-1c311bc8-792e-43d1-b5e9-6105ef7b377e/volume-1c311bc8-792e-43d1-b5e9-6105ef7b377e.vmdk. No consolidation needed. {{(pid=65726) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1111.035901] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-44fd1bde-b4ab-4b7e-bddb-a60989ac9bd3 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: ad4f00b5-3765-494f-8132-8b17b29d219a] Reconfiguring VM instance instance-0000005b to detach disk 2001 {{(pid=65726) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1111.036645] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2339200d-fdf3-43f0-9f01-1e699d64f662 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.060473] env[65726]: DEBUG oslo_vmware.api [None req-44fd1bde-b4ab-4b7e-bddb-a60989ac9bd3 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Waiting for the task: (returnval){ [ 1111.060473] env[65726]: value = "task-5116747" [ 1111.060473] env[65726]: _type = "Task" [ 1111.060473] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1111.069535] env[65726]: DEBUG oslo_vmware.api [None req-44fd1bde-b4ab-4b7e-bddb-a60989ac9bd3 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Task: {'id': task-5116747, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1111.172055] env[65726]: DEBUG oslo_vmware.api [None req-5eb7fa6b-cbd6-4a47-b227-e9c16f779f4a tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Task: {'id': task-5116746, 'name': PowerOnVM_Task} progress is 94%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1111.244022] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3416c1c3-d186-43d0-a802-dc1f8ce0ac10 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.252550] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b02aaa70-257d-44a2-909f-b1d982ad73b3 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.288772] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0dcc3242-2e8b-4a06-8511-0be2344f3a9d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.297525] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76d48b4c-087c-4fc4-8738-99f940d020b9 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.313201] env[65726]: DEBUG nova.compute.provider_tree [None req-96c35e52-5464-43ed-b438-b393e0533d7f tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1111.477938] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-aea91b5d-face-4f98-b031-6ac6f0d1a83b tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 0361ae8f-634d-4161-a594-7f93dbb1b7cd] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1111.478293] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d8d624e2-c305-49af-bc9a-aed1d1e82292 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.489449] env[65726]: DEBUG oslo_vmware.api [None req-aea91b5d-face-4f98-b031-6ac6f0d1a83b tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Waiting for the task: (returnval){ [ 1111.489449] env[65726]: value = "task-5116748" [ 1111.489449] env[65726]: _type = "Task" [ 1111.489449] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1111.505017] env[65726]: DEBUG oslo_vmware.api [None req-aea91b5d-face-4f98-b031-6ac6f0d1a83b tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': task-5116748, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1111.514597] env[65726]: WARNING openstack [None req-022a70bf-a10f-4483-94dc-011169b6c2d5 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1111.514954] env[65726]: WARNING openstack [None req-022a70bf-a10f-4483-94dc-011169b6c2d5 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1111.571575] env[65726]: DEBUG oslo_vmware.api [None req-44fd1bde-b4ab-4b7e-bddb-a60989ac9bd3 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Task: {'id': task-5116747, 'name': ReconfigVM_Task, 'duration_secs': 0.268823} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1111.571821] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-44fd1bde-b4ab-4b7e-bddb-a60989ac9bd3 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: ad4f00b5-3765-494f-8132-8b17b29d219a] Reconfigured VM instance instance-0000005b to detach disk 2001 {{(pid=65726) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1111.577078] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b0a0efac-f9b0-44b5-8f8b-2c4bdbc5b4f4 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.602600] env[65726]: DEBUG oslo_vmware.api [None req-44fd1bde-b4ab-4b7e-bddb-a60989ac9bd3 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Waiting for the task: (returnval){ [ 1111.602600] env[65726]: value = "task-5116749" [ 1111.602600] env[65726]: _type = "Task" [ 1111.602600] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1111.612673] env[65726]: DEBUG oslo_vmware.api [None req-44fd1bde-b4ab-4b7e-bddb-a60989ac9bd3 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Task: {'id': task-5116749, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1111.669070] env[65726]: DEBUG oslo_vmware.api [None req-5eb7fa6b-cbd6-4a47-b227-e9c16f779f4a tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Task: {'id': task-5116746, 'name': PowerOnVM_Task, 'duration_secs': 0.567713} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1111.669308] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-5eb7fa6b-cbd6-4a47-b227-e9c16f779f4a tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] [instance: 4809fcf6-59d4-409c-bd8d-981dc85a686f] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1111.669614] env[65726]: INFO nova.compute.manager [None req-5eb7fa6b-cbd6-4a47-b227-e9c16f779f4a tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] [instance: 4809fcf6-59d4-409c-bd8d-981dc85a686f] Took 10.87 seconds to spawn the instance on the hypervisor. [ 1111.669831] env[65726]: DEBUG nova.compute.manager [None req-5eb7fa6b-cbd6-4a47-b227-e9c16f779f4a tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] [instance: 4809fcf6-59d4-409c-bd8d-981dc85a686f] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1111.670816] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f29952d-a4db-4a23-9c4d-66fc93d6bd44 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.816896] env[65726]: DEBUG nova.scheduler.client.report [None req-96c35e52-5464-43ed-b438-b393e0533d7f tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1111.999981] env[65726]: DEBUG oslo_vmware.api [None req-aea91b5d-face-4f98-b031-6ac6f0d1a83b tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': task-5116748, 'name': PowerOffVM_Task, 'duration_secs': 0.25064} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1112.000333] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-aea91b5d-face-4f98-b031-6ac6f0d1a83b tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 0361ae8f-634d-4161-a594-7f93dbb1b7cd] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1112.000468] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-aea91b5d-face-4f98-b031-6ac6f0d1a83b tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 0361ae8f-634d-4161-a594-7f93dbb1b7cd] Updating instance '0361ae8f-634d-4161-a594-7f93dbb1b7cd' progress to 17 {{(pid=65726) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1112.012918] env[65726]: DEBUG nova.compute.manager [None req-e0232c78-b0b0-436b-b3db-dd24c81ff983 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: e7156a10-631a-4cc5-a544-88ce37763d02] Start spawning the instance on the hypervisor. {{(pid=65726) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1112.025851] env[65726]: DEBUG nova.objects.instance [None req-022a70bf-a10f-4483-94dc-011169b6c2d5 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Lazy-loading 'pci_requests' on Instance uuid 51f5337d-2e23-4b7c-b590-76be6bd82411 {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1112.044806] env[65726]: DEBUG nova.virt.hardware [None req-e0232c78-b0b0-436b-b3db-dd24c81ff983 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T19:28:34Z,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1539f06fa3534e90acbc3a60c4b8bd3f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T19:28:35Z,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1112.045309] env[65726]: DEBUG nova.virt.hardware [None req-e0232c78-b0b0-436b-b3db-dd24c81ff983 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1112.045530] env[65726]: DEBUG nova.virt.hardware [None req-e0232c78-b0b0-436b-b3db-dd24c81ff983 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1112.045710] env[65726]: DEBUG nova.virt.hardware [None req-e0232c78-b0b0-436b-b3db-dd24c81ff983 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1112.045853] env[65726]: DEBUG nova.virt.hardware [None req-e0232c78-b0b0-436b-b3db-dd24c81ff983 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1112.045999] env[65726]: DEBUG nova.virt.hardware [None req-e0232c78-b0b0-436b-b3db-dd24c81ff983 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1112.046227] env[65726]: DEBUG nova.virt.hardware [None req-e0232c78-b0b0-436b-b3db-dd24c81ff983 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1112.046380] env[65726]: DEBUG nova.virt.hardware [None req-e0232c78-b0b0-436b-b3db-dd24c81ff983 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1112.046544] env[65726]: DEBUG nova.virt.hardware [None req-e0232c78-b0b0-436b-b3db-dd24c81ff983 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1112.046702] env[65726]: DEBUG nova.virt.hardware [None req-e0232c78-b0b0-436b-b3db-dd24c81ff983 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1112.046869] env[65726]: DEBUG nova.virt.hardware [None req-e0232c78-b0b0-436b-b3db-dd24c81ff983 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1112.048030] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ce60c3a-9121-40c4-80cb-bc80caca465d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.060345] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4fc7d911-45cd-426c-bb62-e5cdcf7bb136 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.113861] env[65726]: DEBUG oslo_vmware.api [None req-44fd1bde-b4ab-4b7e-bddb-a60989ac9bd3 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Task: {'id': task-5116749, 'name': ReconfigVM_Task, 'duration_secs': 0.207254} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1112.114318] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-44fd1bde-b4ab-4b7e-bddb-a60989ac9bd3 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: ad4f00b5-3765-494f-8132-8b17b29d219a] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-995298', 'volume_id': '1c311bc8-792e-43d1-b5e9-6105ef7b377e', 'name': 'volume-1c311bc8-792e-43d1-b5e9-6105ef7b377e', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'ad4f00b5-3765-494f-8132-8b17b29d219a', 'attached_at': '', 'detached_at': '', 'volume_id': '1c311bc8-792e-43d1-b5e9-6105ef7b377e', 'serial': '1c311bc8-792e-43d1-b5e9-6105ef7b377e'} {{(pid=65726) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1112.193041] env[65726]: INFO nova.compute.manager [None req-5eb7fa6b-cbd6-4a47-b227-e9c16f779f4a tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] [instance: 4809fcf6-59d4-409c-bd8d-981dc85a686f] Took 23.39 seconds to build instance. [ 1112.323217] env[65726]: DEBUG oslo_concurrency.lockutils [None req-96c35e52-5464-43ed-b438-b393e0533d7f tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.356s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1112.323217] env[65726]: DEBUG nova.compute.manager [None req-96c35e52-5464-43ed-b438-b393e0533d7f tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: b353b5e9-500c-42d3-a87f-880a624febec] Start building networks asynchronously for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1112.325736] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9575dab1-9756-447a-bbac-d03937ac3ee1 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.617s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1112.325736] env[65726]: DEBUG nova.objects.instance [None req-9575dab1-9756-447a-bbac-d03937ac3ee1 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] Lazy-loading 'resources' on Instance uuid df08d166-7811-49b5-9c16-a3434229d623 {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1112.333516] env[65726]: DEBUG nova.network.neutron [None req-e0232c78-b0b0-436b-b3db-dd24c81ff983 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: e7156a10-631a-4cc5-a544-88ce37763d02] Successfully updated port: 44a1554d-d380-49b2-ad33-9a251114c93c {{(pid=65726) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 1112.508568] env[65726]: DEBUG nova.virt.hardware [None req-aea91b5d-face-4f98-b031-6ac6f0d1a83b tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1112.508817] env[65726]: DEBUG nova.virt.hardware [None req-aea91b5d-face-4f98-b031-6ac6f0d1a83b tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1112.508972] env[65726]: DEBUG nova.virt.hardware [None req-aea91b5d-face-4f98-b031-6ac6f0d1a83b tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1112.509169] env[65726]: DEBUG nova.virt.hardware [None req-aea91b5d-face-4f98-b031-6ac6f0d1a83b tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1112.509315] env[65726]: DEBUG nova.virt.hardware [None req-aea91b5d-face-4f98-b031-6ac6f0d1a83b tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1112.509462] env[65726]: DEBUG nova.virt.hardware [None req-aea91b5d-face-4f98-b031-6ac6f0d1a83b tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1112.509663] env[65726]: DEBUG nova.virt.hardware [None req-aea91b5d-face-4f98-b031-6ac6f0d1a83b tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1112.509832] env[65726]: DEBUG nova.virt.hardware [None req-aea91b5d-face-4f98-b031-6ac6f0d1a83b tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1112.510943] env[65726]: DEBUG nova.virt.hardware [None req-aea91b5d-face-4f98-b031-6ac6f0d1a83b tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1112.510943] env[65726]: DEBUG nova.virt.hardware [None req-aea91b5d-face-4f98-b031-6ac6f0d1a83b tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1112.511151] env[65726]: DEBUG nova.virt.hardware [None req-aea91b5d-face-4f98-b031-6ac6f0d1a83b tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1112.516725] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1af3d719-0e40-44f4-bbac-8a03af5ba46c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.530025] env[65726]: DEBUG nova.objects.base [None req-022a70bf-a10f-4483-94dc-011169b6c2d5 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Object Instance<51f5337d-2e23-4b7c-b590-76be6bd82411> lazy-loaded attributes: flavor,pci_requests {{(pid=65726) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1112.530025] env[65726]: DEBUG nova.network.neutron [None req-022a70bf-a10f-4483-94dc-011169b6c2d5 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: 51f5337d-2e23-4b7c-b590-76be6bd82411] allocate_for_instance() {{(pid=65726) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 1112.530025] env[65726]: WARNING neutronclient.v2_0.client [None req-022a70bf-a10f-4483-94dc-011169b6c2d5 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1112.530025] env[65726]: WARNING neutronclient.v2_0.client [None req-022a70bf-a10f-4483-94dc-011169b6c2d5 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1112.530025] env[65726]: WARNING openstack [None req-022a70bf-a10f-4483-94dc-011169b6c2d5 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1112.530025] env[65726]: WARNING openstack [None req-022a70bf-a10f-4483-94dc-011169b6c2d5 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1112.544897] env[65726]: DEBUG oslo_vmware.api [None req-aea91b5d-face-4f98-b031-6ac6f0d1a83b tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Waiting for the task: (returnval){ [ 1112.544897] env[65726]: value = "task-5116750" [ 1112.544897] env[65726]: _type = "Task" [ 1112.544897] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1112.554108] env[65726]: DEBUG oslo_vmware.api [None req-aea91b5d-face-4f98-b031-6ac6f0d1a83b tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': task-5116750, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1112.618631] env[65726]: DEBUG nova.policy [None req-022a70bf-a10f-4483-94dc-011169b6c2d5 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9b65b9194933469c908c7cc04478444b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd46ccb00794f458b85da4a93879139ee', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65726) authorize /opt/stack/nova/nova/policy.py:192}} [ 1112.666704] env[65726]: DEBUG nova.objects.instance [None req-44fd1bde-b4ab-4b7e-bddb-a60989ac9bd3 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Lazy-loading 'flavor' on Instance uuid ad4f00b5-3765-494f-8132-8b17b29d219a {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1112.696972] env[65726]: DEBUG oslo_concurrency.lockutils [None req-5eb7fa6b-cbd6-4a47-b227-e9c16f779f4a tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Lock "4809fcf6-59d4-409c-bd8d-981dc85a686f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 24.910s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1112.757780] env[65726]: DEBUG nova.compute.manager [req-e963834c-6c4d-40e3-a5b5-fc8d543994f7 req-afd3829c-751b-4eee-a8d8-f35fc0f3eea6 service nova] [instance: e7156a10-631a-4cc5-a544-88ce37763d02] Received event network-vif-plugged-44a1554d-d380-49b2-ad33-9a251114c93c {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1112.758014] env[65726]: DEBUG oslo_concurrency.lockutils [req-e963834c-6c4d-40e3-a5b5-fc8d543994f7 req-afd3829c-751b-4eee-a8d8-f35fc0f3eea6 service nova] Acquiring lock "e7156a10-631a-4cc5-a544-88ce37763d02-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1112.758224] env[65726]: DEBUG oslo_concurrency.lockutils [req-e963834c-6c4d-40e3-a5b5-fc8d543994f7 req-afd3829c-751b-4eee-a8d8-f35fc0f3eea6 service nova] Lock "e7156a10-631a-4cc5-a544-88ce37763d02-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1112.758385] env[65726]: DEBUG oslo_concurrency.lockutils [req-e963834c-6c4d-40e3-a5b5-fc8d543994f7 req-afd3829c-751b-4eee-a8d8-f35fc0f3eea6 service nova] Lock "e7156a10-631a-4cc5-a544-88ce37763d02-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1112.758544] env[65726]: DEBUG nova.compute.manager [req-e963834c-6c4d-40e3-a5b5-fc8d543994f7 req-afd3829c-751b-4eee-a8d8-f35fc0f3eea6 service nova] [instance: e7156a10-631a-4cc5-a544-88ce37763d02] No waiting events found dispatching network-vif-plugged-44a1554d-d380-49b2-ad33-9a251114c93c {{(pid=65726) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1112.758703] env[65726]: WARNING nova.compute.manager [req-e963834c-6c4d-40e3-a5b5-fc8d543994f7 req-afd3829c-751b-4eee-a8d8-f35fc0f3eea6 service nova] [instance: e7156a10-631a-4cc5-a544-88ce37763d02] Received unexpected event network-vif-plugged-44a1554d-d380-49b2-ad33-9a251114c93c for instance with vm_state building and task_state spawning. [ 1112.829154] env[65726]: DEBUG nova.compute.utils [None req-96c35e52-5464-43ed-b438-b393e0533d7f tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Using /dev/sd instead of None {{(pid=65726) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1112.830612] env[65726]: DEBUG nova.compute.manager [None req-96c35e52-5464-43ed-b438-b393e0533d7f tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: b353b5e9-500c-42d3-a87f-880a624febec] Allocating IP information in the background. {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1112.830816] env[65726]: DEBUG nova.network.neutron [None req-96c35e52-5464-43ed-b438-b393e0533d7f tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: b353b5e9-500c-42d3-a87f-880a624febec] allocate_for_instance() {{(pid=65726) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 1112.831160] env[65726]: WARNING neutronclient.v2_0.client [None req-96c35e52-5464-43ed-b438-b393e0533d7f tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1112.831451] env[65726]: WARNING neutronclient.v2_0.client [None req-96c35e52-5464-43ed-b438-b393e0533d7f tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1112.835092] env[65726]: WARNING openstack [None req-96c35e52-5464-43ed-b438-b393e0533d7f tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1112.835092] env[65726]: WARNING openstack [None req-96c35e52-5464-43ed-b438-b393e0533d7f tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1112.844179] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e0232c78-b0b0-436b-b3db-dd24c81ff983 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Acquiring lock "refresh_cache-e7156a10-631a-4cc5-a544-88ce37763d02" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1112.846063] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e0232c78-b0b0-436b-b3db-dd24c81ff983 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Acquired lock "refresh_cache-e7156a10-631a-4cc5-a544-88ce37763d02" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1112.848364] env[65726]: DEBUG nova.network.neutron [None req-e0232c78-b0b0-436b-b3db-dd24c81ff983 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: e7156a10-631a-4cc5-a544-88ce37763d02] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1112.875694] env[65726]: DEBUG oslo_concurrency.lockutils [None req-ad75268a-0b50-4f80-9f3a-c89856a157c4 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Acquiring lock "378c1d18-4bb6-4245-80bf-370fb7af0575" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1112.875778] env[65726]: DEBUG oslo_concurrency.lockutils [None req-ad75268a-0b50-4f80-9f3a-c89856a157c4 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Lock "378c1d18-4bb6-4245-80bf-370fb7af0575" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1112.889771] env[65726]: DEBUG nova.policy [None req-96c35e52-5464-43ed-b438-b393e0533d7f tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e6593c9d94ad4425ab0aede7e76ebb31', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '96149159e18e44f9bf3453e67681f224', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65726) authorize /opt/stack/nova/nova/policy.py:192}} [ 1113.060105] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3fcaace9-99aa-45a5-851a-ba436f212822 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.073135] env[65726]: DEBUG oslo_vmware.api [None req-aea91b5d-face-4f98-b031-6ac6f0d1a83b tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': task-5116750, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1113.075918] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d050792a-b3ec-4a58-a8bd-f0ab10312127 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.108724] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c422cb5-be44-4eb3-82ea-0463b87a3543 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.117354] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2f52299-5263-4a5d-b194-6b28bf699974 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.132717] env[65726]: DEBUG nova.compute.provider_tree [None req-9575dab1-9756-447a-bbac-d03937ac3ee1 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1113.197980] env[65726]: DEBUG nova.network.neutron [None req-96c35e52-5464-43ed-b438-b393e0533d7f tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: b353b5e9-500c-42d3-a87f-880a624febec] Successfully created port: 08b7016b-dc44-4206-a00c-5da943b82a38 {{(pid=65726) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1113.343501] env[65726]: DEBUG nova.compute.manager [None req-96c35e52-5464-43ed-b438-b393e0533d7f tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: b353b5e9-500c-42d3-a87f-880a624febec] Start building block device mappings for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1113.354025] env[65726]: WARNING openstack [None req-e0232c78-b0b0-436b-b3db-dd24c81ff983 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1113.354025] env[65726]: WARNING openstack [None req-e0232c78-b0b0-436b-b3db-dd24c81ff983 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1113.360418] env[65726]: DEBUG nova.network.neutron [None req-e0232c78-b0b0-436b-b3db-dd24c81ff983 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: e7156a10-631a-4cc5-a544-88ce37763d02] Instance cache missing network info. {{(pid=65726) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 1113.380259] env[65726]: DEBUG nova.compute.manager [None req-ad75268a-0b50-4f80-9f3a-c89856a157c4 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 378c1d18-4bb6-4245-80bf-370fb7af0575] Starting instance... {{(pid=65726) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 1113.550887] env[65726]: INFO nova.compute.manager [None req-b72da630-f8d0-435a-9469-cadbe319a3e9 tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] [instance: 4809fcf6-59d4-409c-bd8d-981dc85a686f] Rescuing [ 1113.551193] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b72da630-f8d0-435a-9469-cadbe319a3e9 tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Acquiring lock "refresh_cache-4809fcf6-59d4-409c-bd8d-981dc85a686f" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1113.551338] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b72da630-f8d0-435a-9469-cadbe319a3e9 tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Acquired lock "refresh_cache-4809fcf6-59d4-409c-bd8d-981dc85a686f" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1113.551501] env[65726]: DEBUG nova.network.neutron [None req-b72da630-f8d0-435a-9469-cadbe319a3e9 tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] [instance: 4809fcf6-59d4-409c-bd8d-981dc85a686f] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1113.559929] env[65726]: DEBUG oslo_vmware.api [None req-aea91b5d-face-4f98-b031-6ac6f0d1a83b tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': task-5116750, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1113.635804] env[65726]: DEBUG nova.scheduler.client.report [None req-9575dab1-9756-447a-bbac-d03937ac3ee1 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1113.673309] env[65726]: DEBUG oslo_concurrency.lockutils [None req-44fd1bde-b4ab-4b7e-bddb-a60989ac9bd3 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Lock "ad4f00b5-3765-494f-8132-8b17b29d219a" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.298s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1113.904328] env[65726]: DEBUG oslo_concurrency.lockutils [None req-ad75268a-0b50-4f80-9f3a-c89856a157c4 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1114.054906] env[65726]: WARNING openstack [None req-b72da630-f8d0-435a-9469-cadbe319a3e9 tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1114.055298] env[65726]: WARNING openstack [None req-b72da630-f8d0-435a-9469-cadbe319a3e9 tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1114.065182] env[65726]: DEBUG oslo_vmware.api [None req-aea91b5d-face-4f98-b031-6ac6f0d1a83b tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': task-5116750, 'name': ReconfigVM_Task, 'duration_secs': 1.173664} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1114.065506] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-aea91b5d-face-4f98-b031-6ac6f0d1a83b tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 0361ae8f-634d-4161-a594-7f93dbb1b7cd] Updating instance '0361ae8f-634d-4161-a594-7f93dbb1b7cd' progress to 33 {{(pid=65726) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1114.116020] env[65726]: DEBUG nova.network.neutron [None req-022a70bf-a10f-4483-94dc-011169b6c2d5 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: 51f5337d-2e23-4b7c-b590-76be6bd82411] Successfully updated port: fb580bad-fd60-471d-95f2-066352417659 {{(pid=65726) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 1114.140799] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9575dab1-9756-447a-bbac-d03937ac3ee1 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.815s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1114.144586] env[65726]: DEBUG oslo_concurrency.lockutils [None req-ad75268a-0b50-4f80-9f3a-c89856a157c4 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.240s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1114.147075] env[65726]: INFO nova.compute.claims [None req-ad75268a-0b50-4f80-9f3a-c89856a157c4 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 378c1d18-4bb6-4245-80bf-370fb7af0575] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1114.184310] env[65726]: INFO nova.scheduler.client.report [None req-9575dab1-9756-447a-bbac-d03937ac3ee1 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] Deleted allocations for instance df08d166-7811-49b5-9c16-a3434229d623 [ 1114.264294] env[65726]: WARNING openstack [None req-e0232c78-b0b0-436b-b3db-dd24c81ff983 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1114.264900] env[65726]: WARNING openstack [None req-e0232c78-b0b0-436b-b3db-dd24c81ff983 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1114.354712] env[65726]: DEBUG nova.compute.manager [None req-96c35e52-5464-43ed-b438-b393e0533d7f tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: b353b5e9-500c-42d3-a87f-880a624febec] Start spawning the instance on the hypervisor. {{(pid=65726) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1114.406651] env[65726]: DEBUG nova.virt.hardware [None req-96c35e52-5464-43ed-b438-b393e0533d7f tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T19:28:34Z,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1539f06fa3534e90acbc3a60c4b8bd3f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T19:28:35Z,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1114.406940] env[65726]: DEBUG nova.virt.hardware [None req-96c35e52-5464-43ed-b438-b393e0533d7f tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1114.408406] env[65726]: DEBUG nova.virt.hardware [None req-96c35e52-5464-43ed-b438-b393e0533d7f tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1114.408650] env[65726]: DEBUG nova.virt.hardware [None req-96c35e52-5464-43ed-b438-b393e0533d7f tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1114.408801] env[65726]: DEBUG nova.virt.hardware [None req-96c35e52-5464-43ed-b438-b393e0533d7f tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1114.408955] env[65726]: DEBUG nova.virt.hardware [None req-96c35e52-5464-43ed-b438-b393e0533d7f tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1114.409513] env[65726]: DEBUG nova.virt.hardware [None req-96c35e52-5464-43ed-b438-b393e0533d7f tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1114.409586] env[65726]: DEBUG nova.virt.hardware [None req-96c35e52-5464-43ed-b438-b393e0533d7f tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1114.409879] env[65726]: DEBUG nova.virt.hardware [None req-96c35e52-5464-43ed-b438-b393e0533d7f tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1114.410072] env[65726]: DEBUG nova.virt.hardware [None req-96c35e52-5464-43ed-b438-b393e0533d7f tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1114.410317] env[65726]: DEBUG nova.virt.hardware [None req-96c35e52-5464-43ed-b438-b393e0533d7f tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1114.411232] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-322af997-f442-4a0c-872b-75c1962c9125 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.425412] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2cb6b020-fae3-4071-9f74-a1b64c0fde98 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.439546] env[65726]: WARNING openstack [None req-b72da630-f8d0-435a-9469-cadbe319a3e9 tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1114.439546] env[65726]: WARNING openstack [None req-b72da630-f8d0-435a-9469-cadbe319a3e9 tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1114.504888] env[65726]: WARNING openstack [None req-e0232c78-b0b0-436b-b3db-dd24c81ff983 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1114.505369] env[65726]: WARNING openstack [None req-e0232c78-b0b0-436b-b3db-dd24c81ff983 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1114.572678] env[65726]: DEBUG nova.virt.hardware [None req-aea91b5d-face-4f98-b031-6ac6f0d1a83b tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1114.572972] env[65726]: DEBUG nova.virt.hardware [None req-aea91b5d-face-4f98-b031-6ac6f0d1a83b tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1114.573154] env[65726]: DEBUG nova.virt.hardware [None req-aea91b5d-face-4f98-b031-6ac6f0d1a83b tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1114.573440] env[65726]: DEBUG nova.virt.hardware [None req-aea91b5d-face-4f98-b031-6ac6f0d1a83b tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1114.573511] env[65726]: DEBUG nova.virt.hardware [None req-aea91b5d-face-4f98-b031-6ac6f0d1a83b tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1114.573615] env[65726]: DEBUG nova.virt.hardware [None req-aea91b5d-face-4f98-b031-6ac6f0d1a83b tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1114.573902] env[65726]: DEBUG nova.virt.hardware [None req-aea91b5d-face-4f98-b031-6ac6f0d1a83b tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1114.574101] env[65726]: DEBUG nova.virt.hardware [None req-aea91b5d-face-4f98-b031-6ac6f0d1a83b tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1114.574369] env[65726]: DEBUG nova.virt.hardware [None req-aea91b5d-face-4f98-b031-6ac6f0d1a83b tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1114.574434] env[65726]: DEBUG nova.virt.hardware [None req-aea91b5d-face-4f98-b031-6ac6f0d1a83b tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1114.574601] env[65726]: DEBUG nova.virt.hardware [None req-aea91b5d-face-4f98-b031-6ac6f0d1a83b tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1114.579970] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-aea91b5d-face-4f98-b031-6ac6f0d1a83b tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 0361ae8f-634d-4161-a594-7f93dbb1b7cd] Reconfiguring VM instance instance-00000064 to detach disk 2000 {{(pid=65726) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1114.580352] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-09c6432b-6e47-4df9-a79e-e92551325de5 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.599737] env[65726]: WARNING openstack [None req-b72da630-f8d0-435a-9469-cadbe319a3e9 tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1114.600119] env[65726]: WARNING openstack [None req-b72da630-f8d0-435a-9469-cadbe319a3e9 tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1114.613628] env[65726]: DEBUG oslo_vmware.api [None req-aea91b5d-face-4f98-b031-6ac6f0d1a83b tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Waiting for the task: (returnval){ [ 1114.613628] env[65726]: value = "task-5116751" [ 1114.613628] env[65726]: _type = "Task" [ 1114.613628] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1114.619754] env[65726]: DEBUG oslo_concurrency.lockutils [None req-022a70bf-a10f-4483-94dc-011169b6c2d5 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Acquiring lock "refresh_cache-51f5337d-2e23-4b7c-b590-76be6bd82411" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1114.619953] env[65726]: DEBUG oslo_concurrency.lockutils [None req-022a70bf-a10f-4483-94dc-011169b6c2d5 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Acquired lock "refresh_cache-51f5337d-2e23-4b7c-b590-76be6bd82411" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1114.620162] env[65726]: DEBUG nova.network.neutron [None req-022a70bf-a10f-4483-94dc-011169b6c2d5 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: 51f5337d-2e23-4b7c-b590-76be6bd82411] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1114.629131] env[65726]: DEBUG oslo_vmware.api [None req-aea91b5d-face-4f98-b031-6ac6f0d1a83b tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': task-5116751, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1114.662501] env[65726]: DEBUG nova.network.neutron [None req-e0232c78-b0b0-436b-b3db-dd24c81ff983 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: e7156a10-631a-4cc5-a544-88ce37763d02] Updating instance_info_cache with network_info: [{"id": "44a1554d-d380-49b2-ad33-9a251114c93c", "address": "fa:16:3e:63:0a:a8", "network": {"id": "4c972d9d-db3b-4779-ac39-39d9a2d141dd", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1418913861-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cbb17a09b35c4c22ade5c4082b10bda9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "50cf0a70-948d-4611-af05-94c1483064ed", "external-id": "nsx-vlan-transportzone-536", "segmentation_id": 536, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap44a1554d-d3", "ovs_interfaceid": "44a1554d-d380-49b2-ad33-9a251114c93c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1114.695825] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9575dab1-9756-447a-bbac-d03937ac3ee1 tempest-AttachInterfacesUnderV243Test-711093399 tempest-AttachInterfacesUnderV243Test-711093399-project-member] Lock "df08d166-7811-49b5-9c16-a3434229d623" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.731s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1114.769878] env[65726]: DEBUG nova.network.neutron [None req-b72da630-f8d0-435a-9469-cadbe319a3e9 tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] [instance: 4809fcf6-59d4-409c-bd8d-981dc85a686f] Updating instance_info_cache with network_info: [{"id": "13050a31-7165-4b22-83cc-d8ce9a41eadf", "address": "fa:16:3e:02:ed:66", "network": {"id": "53facc8b-af7e-44de-8c6f-6bea798b3f0b", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-2084216438-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "eb811e38506b4297bf01bc5f2e098370", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "19671de9-8b5b-4710-adc3-7419f3c0f171", "external-id": "nsx-vlan-transportzone-421", "segmentation_id": 421, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap13050a31-71", "ovs_interfaceid": "13050a31-7165-4b22-83cc-d8ce9a41eadf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1114.778338] env[65726]: DEBUG nova.network.neutron [None req-96c35e52-5464-43ed-b438-b393e0533d7f tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: b353b5e9-500c-42d3-a87f-880a624febec] Successfully updated port: 08b7016b-dc44-4206-a00c-5da943b82a38 {{(pid=65726) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 1115.023535] env[65726]: DEBUG nova.compute.manager [req-34f88d20-10e1-45b8-af09-76e3a0989d54 req-8642a19b-9208-4eec-ae30-32e218a9d1e9 service nova] [instance: 51f5337d-2e23-4b7c-b590-76be6bd82411] Received event network-vif-plugged-fb580bad-fd60-471d-95f2-066352417659 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1115.023763] env[65726]: DEBUG oslo_concurrency.lockutils [req-34f88d20-10e1-45b8-af09-76e3a0989d54 req-8642a19b-9208-4eec-ae30-32e218a9d1e9 service nova] Acquiring lock "51f5337d-2e23-4b7c-b590-76be6bd82411-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1115.023978] env[65726]: DEBUG oslo_concurrency.lockutils [req-34f88d20-10e1-45b8-af09-76e3a0989d54 req-8642a19b-9208-4eec-ae30-32e218a9d1e9 service nova] Lock "51f5337d-2e23-4b7c-b590-76be6bd82411-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1115.024163] env[65726]: DEBUG oslo_concurrency.lockutils [req-34f88d20-10e1-45b8-af09-76e3a0989d54 req-8642a19b-9208-4eec-ae30-32e218a9d1e9 service nova] Lock "51f5337d-2e23-4b7c-b590-76be6bd82411-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1115.024331] env[65726]: DEBUG nova.compute.manager [req-34f88d20-10e1-45b8-af09-76e3a0989d54 req-8642a19b-9208-4eec-ae30-32e218a9d1e9 service nova] [instance: 51f5337d-2e23-4b7c-b590-76be6bd82411] No waiting events found dispatching network-vif-plugged-fb580bad-fd60-471d-95f2-066352417659 {{(pid=65726) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1115.024492] env[65726]: WARNING nova.compute.manager [req-34f88d20-10e1-45b8-af09-76e3a0989d54 req-8642a19b-9208-4eec-ae30-32e218a9d1e9 service nova] [instance: 51f5337d-2e23-4b7c-b590-76be6bd82411] Received unexpected event network-vif-plugged-fb580bad-fd60-471d-95f2-066352417659 for instance with vm_state active and task_state None. [ 1115.074920] env[65726]: DEBUG nova.compute.manager [req-ad7c4151-6558-4aaf-ad53-ef58f3e573ff req-fc63acd5-0da7-44ca-be11-d87400e8e8e7 service nova] [instance: e7156a10-631a-4cc5-a544-88ce37763d02] Received event network-changed-44a1554d-d380-49b2-ad33-9a251114c93c {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1115.075278] env[65726]: DEBUG nova.compute.manager [req-ad7c4151-6558-4aaf-ad53-ef58f3e573ff req-fc63acd5-0da7-44ca-be11-d87400e8e8e7 service nova] [instance: e7156a10-631a-4cc5-a544-88ce37763d02] Refreshing instance network info cache due to event network-changed-44a1554d-d380-49b2-ad33-9a251114c93c. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 1115.075609] env[65726]: DEBUG oslo_concurrency.lockutils [req-ad7c4151-6558-4aaf-ad53-ef58f3e573ff req-fc63acd5-0da7-44ca-be11-d87400e8e8e7 service nova] Acquiring lock "refresh_cache-e7156a10-631a-4cc5-a544-88ce37763d02" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1115.124035] env[65726]: DEBUG oslo_vmware.api [None req-aea91b5d-face-4f98-b031-6ac6f0d1a83b tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': task-5116751, 'name': ReconfigVM_Task, 'duration_secs': 0.288704} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1115.124846] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-aea91b5d-face-4f98-b031-6ac6f0d1a83b tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 0361ae8f-634d-4161-a594-7f93dbb1b7cd] Reconfigured VM instance instance-00000064 to detach disk 2000 {{(pid=65726) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1115.125164] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-195b038b-9c5b-41a0-8cb1-57edfcf0ebaf {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.142563] env[65726]: WARNING openstack [None req-022a70bf-a10f-4483-94dc-011169b6c2d5 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1115.142684] env[65726]: WARNING openstack [None req-022a70bf-a10f-4483-94dc-011169b6c2d5 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1115.158207] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-aea91b5d-face-4f98-b031-6ac6f0d1a83b tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 0361ae8f-634d-4161-a594-7f93dbb1b7cd] Reconfiguring VM instance instance-00000064 to attach disk [datastore1] 0361ae8f-634d-4161-a594-7f93dbb1b7cd/0361ae8f-634d-4161-a594-7f93dbb1b7cd.vmdk or device None with type thin {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1115.161608] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-69eed3c8-dc8a-421c-9e81-f7a87bdcffd9 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.178857] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e0232c78-b0b0-436b-b3db-dd24c81ff983 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Releasing lock "refresh_cache-e7156a10-631a-4cc5-a544-88ce37763d02" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1115.179225] env[65726]: DEBUG nova.compute.manager [None req-e0232c78-b0b0-436b-b3db-dd24c81ff983 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: e7156a10-631a-4cc5-a544-88ce37763d02] Instance network_info: |[{"id": "44a1554d-d380-49b2-ad33-9a251114c93c", "address": "fa:16:3e:63:0a:a8", "network": {"id": "4c972d9d-db3b-4779-ac39-39d9a2d141dd", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1418913861-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cbb17a09b35c4c22ade5c4082b10bda9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "50cf0a70-948d-4611-af05-94c1483064ed", "external-id": "nsx-vlan-transportzone-536", "segmentation_id": 536, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap44a1554d-d3", "ovs_interfaceid": "44a1554d-d380-49b2-ad33-9a251114c93c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1115.182020] env[65726]: DEBUG oslo_concurrency.lockutils [req-ad7c4151-6558-4aaf-ad53-ef58f3e573ff req-fc63acd5-0da7-44ca-be11-d87400e8e8e7 service nova] Acquired lock "refresh_cache-e7156a10-631a-4cc5-a544-88ce37763d02" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1115.182020] env[65726]: DEBUG nova.network.neutron [req-ad7c4151-6558-4aaf-ad53-ef58f3e573ff req-fc63acd5-0da7-44ca-be11-d87400e8e8e7 service nova] [instance: e7156a10-631a-4cc5-a544-88ce37763d02] Refreshing network info cache for port 44a1554d-d380-49b2-ad33-9a251114c93c {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 1115.182594] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-e0232c78-b0b0-436b-b3db-dd24c81ff983 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: e7156a10-631a-4cc5-a544-88ce37763d02] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:63:0a:a8', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '50cf0a70-948d-4611-af05-94c1483064ed', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '44a1554d-d380-49b2-ad33-9a251114c93c', 'vif_model': 'vmxnet3'}] {{(pid=65726) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1115.191704] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-e0232c78-b0b0-436b-b3db-dd24c81ff983 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1115.193688] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e7156a10-631a-4cc5-a544-88ce37763d02] Creating VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1115.198510] env[65726]: WARNING nova.network.neutron [None req-022a70bf-a10f-4483-94dc-011169b6c2d5 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: 51f5337d-2e23-4b7c-b590-76be6bd82411] 51a66448-0f72-4266-8868-856b48be839f already exists in list: networks containing: ['51a66448-0f72-4266-8868-856b48be839f']. ignoring it [ 1115.200279] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-87e40028-dd34-4910-bbdf-7bf9a42da52a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.220738] env[65726]: DEBUG oslo_vmware.api [None req-aea91b5d-face-4f98-b031-6ac6f0d1a83b tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Waiting for the task: (returnval){ [ 1115.220738] env[65726]: value = "task-5116752" [ 1115.220738] env[65726]: _type = "Task" [ 1115.220738] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1115.234235] env[65726]: DEBUG oslo_vmware.api [None req-aea91b5d-face-4f98-b031-6ac6f0d1a83b tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': task-5116752, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1115.234721] env[65726]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1115.234721] env[65726]: value = "task-5116753" [ 1115.234721] env[65726]: _type = "Task" [ 1115.234721] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1115.247865] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116753, 'name': CreateVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1115.261027] env[65726]: WARNING openstack [None req-022a70bf-a10f-4483-94dc-011169b6c2d5 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1115.261366] env[65726]: WARNING openstack [None req-022a70bf-a10f-4483-94dc-011169b6c2d5 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1115.274644] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b72da630-f8d0-435a-9469-cadbe319a3e9 tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Releasing lock "refresh_cache-4809fcf6-59d4-409c-bd8d-981dc85a686f" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1115.283271] env[65726]: DEBUG oslo_concurrency.lockutils [None req-96c35e52-5464-43ed-b438-b393e0533d7f tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Acquiring lock "refresh_cache-b353b5e9-500c-42d3-a87f-880a624febec" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1115.283503] env[65726]: DEBUG oslo_concurrency.lockutils [None req-96c35e52-5464-43ed-b438-b393e0533d7f tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Acquired lock "refresh_cache-b353b5e9-500c-42d3-a87f-880a624febec" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1115.284261] env[65726]: DEBUG nova.network.neutron [None req-96c35e52-5464-43ed-b438-b393e0533d7f tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: b353b5e9-500c-42d3-a87f-880a624febec] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1115.391320] env[65726]: WARNING openstack [None req-022a70bf-a10f-4483-94dc-011169b6c2d5 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1115.391849] env[65726]: WARNING openstack [None req-022a70bf-a10f-4483-94dc-011169b6c2d5 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1115.447789] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2d28bee-9e18-4b85-a214-cb382b2aa49e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.458552] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8dee6efb-fac8-4b86-89ae-59bd593b2c7e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.504354] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ccb41f27-6973-4304-b6de-400defc741b1 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.514339] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f4bb21b-ed90-45dc-b33e-6732ca56f4f2 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.538012] env[65726]: DEBUG nova.compute.provider_tree [None req-ad75268a-0b50-4f80-9f3a-c89856a157c4 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1115.545936] env[65726]: WARNING openstack [None req-022a70bf-a10f-4483-94dc-011169b6c2d5 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1115.545936] env[65726]: WARNING openstack [None req-022a70bf-a10f-4483-94dc-011169b6c2d5 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1115.613613] env[65726]: WARNING openstack [None req-022a70bf-a10f-4483-94dc-011169b6c2d5 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1115.614158] env[65726]: WARNING openstack [None req-022a70bf-a10f-4483-94dc-011169b6c2d5 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1115.676653] env[65726]: DEBUG oslo_concurrency.lockutils [None req-04bd2146-ec5d-4e36-832f-8911f884caa3 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Acquiring lock "ad4f00b5-3765-494f-8132-8b17b29d219a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1115.676653] env[65726]: DEBUG oslo_concurrency.lockutils [None req-04bd2146-ec5d-4e36-832f-8911f884caa3 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Lock "ad4f00b5-3765-494f-8132-8b17b29d219a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1115.676840] env[65726]: DEBUG oslo_concurrency.lockutils [None req-04bd2146-ec5d-4e36-832f-8911f884caa3 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Acquiring lock "ad4f00b5-3765-494f-8132-8b17b29d219a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1115.676935] env[65726]: DEBUG oslo_concurrency.lockutils [None req-04bd2146-ec5d-4e36-832f-8911f884caa3 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Lock "ad4f00b5-3765-494f-8132-8b17b29d219a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1115.677072] env[65726]: DEBUG oslo_concurrency.lockutils [None req-04bd2146-ec5d-4e36-832f-8911f884caa3 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Lock "ad4f00b5-3765-494f-8132-8b17b29d219a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1115.682167] env[65726]: INFO nova.compute.manager [None req-04bd2146-ec5d-4e36-832f-8911f884caa3 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: ad4f00b5-3765-494f-8132-8b17b29d219a] Terminating instance [ 1115.693500] env[65726]: WARNING openstack [req-ad7c4151-6558-4aaf-ad53-ef58f3e573ff req-fc63acd5-0da7-44ca-be11-d87400e8e8e7 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1115.693898] env[65726]: WARNING openstack [req-ad7c4151-6558-4aaf-ad53-ef58f3e573ff req-fc63acd5-0da7-44ca-be11-d87400e8e8e7 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1115.709621] env[65726]: DEBUG nova.network.neutron [None req-022a70bf-a10f-4483-94dc-011169b6c2d5 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: 51f5337d-2e23-4b7c-b590-76be6bd82411] Updating instance_info_cache with network_info: [{"id": "7407bcd9-3202-4798-9eaa-252fb3fdb21f", "address": "fa:16:3e:79:d2:9d", "network": {"id": "51a66448-0f72-4266-8868-856b48be839f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-155658961-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.213", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d46ccb00794f458b85da4a93879139ee", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b1f3e6c3-5584-4852-9017-476ab8ac4946", "external-id": "nsx-vlan-transportzone-304", "segmentation_id": 304, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7407bcd9-32", "ovs_interfaceid": "7407bcd9-3202-4798-9eaa-252fb3fdb21f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "fb580bad-fd60-471d-95f2-066352417659", "address": "fa:16:3e:32:0d:d6", "network": {"id": "51a66448-0f72-4266-8868-856b48be839f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-155658961-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d46ccb00794f458b85da4a93879139ee", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b1f3e6c3-5584-4852-9017-476ab8ac4946", "external-id": "nsx-vlan-transportzone-304", "segmentation_id": 304, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfb580bad-fd", "ovs_interfaceid": "fb580bad-fd60-471d-95f2-066352417659", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1115.732979] env[65726]: DEBUG oslo_vmware.api [None req-aea91b5d-face-4f98-b031-6ac6f0d1a83b tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': task-5116752, 'name': ReconfigVM_Task, 'duration_secs': 0.474687} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1115.733314] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-aea91b5d-face-4f98-b031-6ac6f0d1a83b tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 0361ae8f-634d-4161-a594-7f93dbb1b7cd] Reconfigured VM instance instance-00000064 to attach disk [datastore1] 0361ae8f-634d-4161-a594-7f93dbb1b7cd/0361ae8f-634d-4161-a594-7f93dbb1b7cd.vmdk or device None with type thin {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1115.733633] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-aea91b5d-face-4f98-b031-6ac6f0d1a83b tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 0361ae8f-634d-4161-a594-7f93dbb1b7cd] Updating instance '0361ae8f-634d-4161-a594-7f93dbb1b7cd' progress to 50 {{(pid=65726) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1115.751260] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116753, 'name': CreateVM_Task, 'duration_secs': 0.448255} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1115.751433] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e7156a10-631a-4cc5-a544-88ce37763d02] Created VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1115.752210] env[65726]: WARNING openstack [None req-e0232c78-b0b0-436b-b3db-dd24c81ff983 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1115.752548] env[65726]: WARNING openstack [None req-e0232c78-b0b0-436b-b3db-dd24c81ff983 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1115.757582] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e0232c78-b0b0-436b-b3db-dd24c81ff983 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1115.757746] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e0232c78-b0b0-436b-b3db-dd24c81ff983 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1115.758133] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e0232c78-b0b0-436b-b3db-dd24c81ff983 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1115.758421] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1dd8dde5-71f6-4422-96bb-e441b41b55ca {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.764586] env[65726]: DEBUG oslo_vmware.api [None req-e0232c78-b0b0-436b-b3db-dd24c81ff983 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Waiting for the task: (returnval){ [ 1115.764586] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52173b4e-a487-75af-85b7-661587067089" [ 1115.764586] env[65726]: _type = "Task" [ 1115.764586] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1115.777467] env[65726]: DEBUG oslo_vmware.api [None req-e0232c78-b0b0-436b-b3db-dd24c81ff983 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52173b4e-a487-75af-85b7-661587067089, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1115.788886] env[65726]: WARNING openstack [None req-96c35e52-5464-43ed-b438-b393e0533d7f tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1115.789280] env[65726]: WARNING openstack [None req-96c35e52-5464-43ed-b438-b393e0533d7f tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1115.794422] env[65726]: DEBUG nova.network.neutron [None req-96c35e52-5464-43ed-b438-b393e0533d7f tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: b353b5e9-500c-42d3-a87f-880a624febec] Instance cache missing network info. {{(pid=65726) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 1115.813974] env[65726]: WARNING openstack [req-ad7c4151-6558-4aaf-ad53-ef58f3e573ff req-fc63acd5-0da7-44ca-be11-d87400e8e8e7 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1115.814377] env[65726]: WARNING openstack [req-ad7c4151-6558-4aaf-ad53-ef58f3e573ff req-fc63acd5-0da7-44ca-be11-d87400e8e8e7 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1115.861733] env[65726]: WARNING openstack [None req-96c35e52-5464-43ed-b438-b393e0533d7f tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1115.862138] env[65726]: WARNING openstack [None req-96c35e52-5464-43ed-b438-b393e0533d7f tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1115.889971] env[65726]: WARNING openstack [req-ad7c4151-6558-4aaf-ad53-ef58f3e573ff req-fc63acd5-0da7-44ca-be11-d87400e8e8e7 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1115.890577] env[65726]: WARNING openstack [req-ad7c4151-6558-4aaf-ad53-ef58f3e573ff req-fc63acd5-0da7-44ca-be11-d87400e8e8e7 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1116.018101] env[65726]: WARNING openstack [None req-96c35e52-5464-43ed-b438-b393e0533d7f tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1116.018516] env[65726]: WARNING openstack [None req-96c35e52-5464-43ed-b438-b393e0533d7f tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1116.041632] env[65726]: DEBUG nova.scheduler.client.report [None req-ad75268a-0b50-4f80-9f3a-c89856a157c4 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1116.056284] env[65726]: DEBUG nova.network.neutron [req-ad7c4151-6558-4aaf-ad53-ef58f3e573ff req-fc63acd5-0da7-44ca-be11-d87400e8e8e7 service nova] [instance: e7156a10-631a-4cc5-a544-88ce37763d02] Updated VIF entry in instance network info cache for port 44a1554d-d380-49b2-ad33-9a251114c93c. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 1116.056683] env[65726]: DEBUG nova.network.neutron [req-ad7c4151-6558-4aaf-ad53-ef58f3e573ff req-fc63acd5-0da7-44ca-be11-d87400e8e8e7 service nova] [instance: e7156a10-631a-4cc5-a544-88ce37763d02] Updating instance_info_cache with network_info: [{"id": "44a1554d-d380-49b2-ad33-9a251114c93c", "address": "fa:16:3e:63:0a:a8", "network": {"id": "4c972d9d-db3b-4779-ac39-39d9a2d141dd", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1418913861-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cbb17a09b35c4c22ade5c4082b10bda9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "50cf0a70-948d-4611-af05-94c1483064ed", "external-id": "nsx-vlan-transportzone-536", "segmentation_id": 536, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap44a1554d-d3", "ovs_interfaceid": "44a1554d-d380-49b2-ad33-9a251114c93c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1116.143145] env[65726]: DEBUG nova.network.neutron [None req-96c35e52-5464-43ed-b438-b393e0533d7f tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: b353b5e9-500c-42d3-a87f-880a624febec] Updating instance_info_cache with network_info: [{"id": "08b7016b-dc44-4206-a00c-5da943b82a38", "address": "fa:16:3e:c5:e4:14", "network": {"id": "eb3278b7-c190-441b-84d2-305f9b896382", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1754571022-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "96149159e18e44f9bf3453e67681f224", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f77ff7a1-209c-4f3f-b2a0-fd817741e739", "external-id": "nsx-vlan-transportzone-935", "segmentation_id": 935, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap08b7016b-dc", "ovs_interfaceid": "08b7016b-dc44-4206-a00c-5da943b82a38", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1116.186874] env[65726]: DEBUG nova.compute.manager [None req-04bd2146-ec5d-4e36-832f-8911f884caa3 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: ad4f00b5-3765-494f-8132-8b17b29d219a] Start destroying the instance on the hypervisor. {{(pid=65726) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1116.187247] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-04bd2146-ec5d-4e36-832f-8911f884caa3 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: ad4f00b5-3765-494f-8132-8b17b29d219a] Destroying instance {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1116.188168] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72422adb-7eda-47ee-8dde-e2f221ddb781 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.197492] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-04bd2146-ec5d-4e36-832f-8911f884caa3 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: ad4f00b5-3765-494f-8132-8b17b29d219a] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1116.197824] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-42f0a6c0-96a7-4282-a4b2-31b907a45a86 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.205015] env[65726]: DEBUG oslo_vmware.api [None req-04bd2146-ec5d-4e36-832f-8911f884caa3 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Waiting for the task: (returnval){ [ 1116.205015] env[65726]: value = "task-5116754" [ 1116.205015] env[65726]: _type = "Task" [ 1116.205015] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1116.214130] env[65726]: DEBUG oslo_concurrency.lockutils [None req-022a70bf-a10f-4483-94dc-011169b6c2d5 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Releasing lock "refresh_cache-51f5337d-2e23-4b7c-b590-76be6bd82411" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1116.214817] env[65726]: DEBUG oslo_concurrency.lockutils [None req-022a70bf-a10f-4483-94dc-011169b6c2d5 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Acquiring lock "51f5337d-2e23-4b7c-b590-76be6bd82411" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1116.215011] env[65726]: DEBUG oslo_concurrency.lockutils [None req-022a70bf-a10f-4483-94dc-011169b6c2d5 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Acquired lock "51f5337d-2e23-4b7c-b590-76be6bd82411" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1116.215303] env[65726]: DEBUG oslo_vmware.api [None req-04bd2146-ec5d-4e36-832f-8911f884caa3 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Task: {'id': task-5116754, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1116.216454] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c7698ca-cd5c-4e09-baca-a3bacfec9f31 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.235515] env[65726]: DEBUG nova.virt.hardware [None req-022a70bf-a10f-4483-94dc-011169b6c2d5 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1116.235743] env[65726]: DEBUG nova.virt.hardware [None req-022a70bf-a10f-4483-94dc-011169b6c2d5 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1116.235896] env[65726]: DEBUG nova.virt.hardware [None req-022a70bf-a10f-4483-94dc-011169b6c2d5 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1116.236086] env[65726]: DEBUG nova.virt.hardware [None req-022a70bf-a10f-4483-94dc-011169b6c2d5 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1116.236233] env[65726]: DEBUG nova.virt.hardware [None req-022a70bf-a10f-4483-94dc-011169b6c2d5 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1116.236376] env[65726]: DEBUG nova.virt.hardware [None req-022a70bf-a10f-4483-94dc-011169b6c2d5 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1116.236642] env[65726]: DEBUG nova.virt.hardware [None req-022a70bf-a10f-4483-94dc-011169b6c2d5 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1116.236838] env[65726]: DEBUG nova.virt.hardware [None req-022a70bf-a10f-4483-94dc-011169b6c2d5 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1116.237107] env[65726]: DEBUG nova.virt.hardware [None req-022a70bf-a10f-4483-94dc-011169b6c2d5 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1116.237192] env[65726]: DEBUG nova.virt.hardware [None req-022a70bf-a10f-4483-94dc-011169b6c2d5 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1116.237374] env[65726]: DEBUG nova.virt.hardware [None req-022a70bf-a10f-4483-94dc-011169b6c2d5 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1116.244262] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-022a70bf-a10f-4483-94dc-011169b6c2d5 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: 51f5337d-2e23-4b7c-b590-76be6bd82411] Reconfiguring VM to attach interface {{(pid=65726) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 1116.246346] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-010e4c91-802d-4636-93e5-9d3e4022f224 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.259726] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70aeb3c8-51a7-4224-a58e-a1439af9aef0 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.285321] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10840f10-dcb7-4d54-a347-2a65e6fa5807 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.288038] env[65726]: DEBUG oslo_vmware.api [None req-022a70bf-a10f-4483-94dc-011169b6c2d5 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Waiting for the task: (returnval){ [ 1116.288038] env[65726]: value = "task-5116755" [ 1116.288038] env[65726]: _type = "Task" [ 1116.288038] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1116.315973] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-aea91b5d-face-4f98-b031-6ac6f0d1a83b tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 0361ae8f-634d-4161-a594-7f93dbb1b7cd] Updating instance '0361ae8f-634d-4161-a594-7f93dbb1b7cd' progress to 67 {{(pid=65726) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1116.320106] env[65726]: DEBUG oslo_vmware.api [None req-e0232c78-b0b0-436b-b3db-dd24c81ff983 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52173b4e-a487-75af-85b7-661587067089, 'name': SearchDatastore_Task, 'duration_secs': 0.031431} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1116.321425] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e0232c78-b0b0-436b-b3db-dd24c81ff983 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1116.321679] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-e0232c78-b0b0-436b-b3db-dd24c81ff983 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: e7156a10-631a-4cc5-a544-88ce37763d02] Processing image b52362a3-ee8a-4cbf-b06f-513b0cc8f95c {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1116.321977] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e0232c78-b0b0-436b-b3db-dd24c81ff983 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1116.322159] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e0232c78-b0b0-436b-b3db-dd24c81ff983 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1116.322421] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-e0232c78-b0b0-436b-b3db-dd24c81ff983 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1116.323347] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27713fc0-c525-4213-af0f-ee8feb46d4cb {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.330573] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-71794f3f-fd13-46ba-ba7d-b726dab5f13f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.333365] env[65726]: DEBUG oslo_vmware.api [None req-022a70bf-a10f-4483-94dc-011169b6c2d5 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': task-5116755, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1116.339352] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-dcfb1735-1ab3-432a-b25b-08fdbc5ed9b1 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] [instance: 6fd70b4a-63e4-4258-9cff-f3b582500b9f] Suspending the VM {{(pid=65726) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1162}} [ 1116.339665] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-ca07c523-ab5d-4fa4-9ae5-d042d812ea9e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.349429] env[65726]: DEBUG oslo_vmware.api [None req-dcfb1735-1ab3-432a-b25b-08fdbc5ed9b1 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Waiting for the task: (returnval){ [ 1116.349429] env[65726]: value = "task-5116756" [ 1116.349429] env[65726]: _type = "Task" [ 1116.349429] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1116.352356] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-e0232c78-b0b0-436b-b3db-dd24c81ff983 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1116.352610] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-e0232c78-b0b0-436b-b3db-dd24c81ff983 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65726) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1116.356974] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-211a8acb-3b1f-4e8b-a1ca-500e9c6fb7da {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.367395] env[65726]: DEBUG oslo_vmware.api [None req-dcfb1735-1ab3-432a-b25b-08fdbc5ed9b1 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Task: {'id': task-5116756, 'name': SuspendVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1116.370555] env[65726]: DEBUG oslo_vmware.api [None req-e0232c78-b0b0-436b-b3db-dd24c81ff983 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Waiting for the task: (returnval){ [ 1116.370555] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52f2162f-d4f6-e914-2df5-f978d638d0f1" [ 1116.370555] env[65726]: _type = "Task" [ 1116.370555] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1116.382455] env[65726]: DEBUG oslo_vmware.api [None req-e0232c78-b0b0-436b-b3db-dd24c81ff983 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52f2162f-d4f6-e914-2df5-f978d638d0f1, 'name': SearchDatastore_Task, 'duration_secs': 0.01379} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1116.383376] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e7dcb99e-e6ee-4435-b364-dc64cee84ae6 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.389620] env[65726]: DEBUG oslo_vmware.api [None req-e0232c78-b0b0-436b-b3db-dd24c81ff983 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Waiting for the task: (returnval){ [ 1116.389620] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52d35f0f-98c8-ad5b-4e3a-9359e3658f74" [ 1116.389620] env[65726]: _type = "Task" [ 1116.389620] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1116.400198] env[65726]: DEBUG oslo_vmware.api [None req-e0232c78-b0b0-436b-b3db-dd24c81ff983 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52d35f0f-98c8-ad5b-4e3a-9359e3658f74, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1116.546553] env[65726]: DEBUG oslo_concurrency.lockutils [None req-ad75268a-0b50-4f80-9f3a-c89856a157c4 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.402s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1116.547145] env[65726]: DEBUG nova.compute.manager [None req-ad75268a-0b50-4f80-9f3a-c89856a157c4 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 378c1d18-4bb6-4245-80bf-370fb7af0575] Start building networks asynchronously for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1116.561038] env[65726]: DEBUG oslo_concurrency.lockutils [req-ad7c4151-6558-4aaf-ad53-ef58f3e573ff req-fc63acd5-0da7-44ca-be11-d87400e8e8e7 service nova] Releasing lock "refresh_cache-e7156a10-631a-4cc5-a544-88ce37763d02" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1116.561344] env[65726]: DEBUG nova.compute.manager [req-ad7c4151-6558-4aaf-ad53-ef58f3e573ff req-fc63acd5-0da7-44ca-be11-d87400e8e8e7 service nova] [instance: b353b5e9-500c-42d3-a87f-880a624febec] Received event network-vif-plugged-08b7016b-dc44-4206-a00c-5da943b82a38 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1116.561605] env[65726]: DEBUG oslo_concurrency.lockutils [req-ad7c4151-6558-4aaf-ad53-ef58f3e573ff req-fc63acd5-0da7-44ca-be11-d87400e8e8e7 service nova] Acquiring lock "b353b5e9-500c-42d3-a87f-880a624febec-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1116.561825] env[65726]: DEBUG oslo_concurrency.lockutils [req-ad7c4151-6558-4aaf-ad53-ef58f3e573ff req-fc63acd5-0da7-44ca-be11-d87400e8e8e7 service nova] Lock "b353b5e9-500c-42d3-a87f-880a624febec-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1116.561988] env[65726]: DEBUG oslo_concurrency.lockutils [req-ad7c4151-6558-4aaf-ad53-ef58f3e573ff req-fc63acd5-0da7-44ca-be11-d87400e8e8e7 service nova] Lock "b353b5e9-500c-42d3-a87f-880a624febec-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1116.562193] env[65726]: DEBUG nova.compute.manager [req-ad7c4151-6558-4aaf-ad53-ef58f3e573ff req-fc63acd5-0da7-44ca-be11-d87400e8e8e7 service nova] [instance: b353b5e9-500c-42d3-a87f-880a624febec] No waiting events found dispatching network-vif-plugged-08b7016b-dc44-4206-a00c-5da943b82a38 {{(pid=65726) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1116.562366] env[65726]: WARNING nova.compute.manager [req-ad7c4151-6558-4aaf-ad53-ef58f3e573ff req-fc63acd5-0da7-44ca-be11-d87400e8e8e7 service nova] [instance: b353b5e9-500c-42d3-a87f-880a624febec] Received unexpected event network-vif-plugged-08b7016b-dc44-4206-a00c-5da943b82a38 for instance with vm_state building and task_state spawning. [ 1116.646626] env[65726]: DEBUG oslo_concurrency.lockutils [None req-96c35e52-5464-43ed-b438-b393e0533d7f tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Releasing lock "refresh_cache-b353b5e9-500c-42d3-a87f-880a624febec" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1116.647084] env[65726]: DEBUG nova.compute.manager [None req-96c35e52-5464-43ed-b438-b393e0533d7f tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: b353b5e9-500c-42d3-a87f-880a624febec] Instance network_info: |[{"id": "08b7016b-dc44-4206-a00c-5da943b82a38", "address": "fa:16:3e:c5:e4:14", "network": {"id": "eb3278b7-c190-441b-84d2-305f9b896382", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1754571022-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "96149159e18e44f9bf3453e67681f224", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f77ff7a1-209c-4f3f-b2a0-fd817741e739", "external-id": "nsx-vlan-transportzone-935", "segmentation_id": 935, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap08b7016b-dc", "ovs_interfaceid": "08b7016b-dc44-4206-a00c-5da943b82a38", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1116.647646] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-96c35e52-5464-43ed-b438-b393e0533d7f tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: b353b5e9-500c-42d3-a87f-880a624febec] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c5:e4:14', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f77ff7a1-209c-4f3f-b2a0-fd817741e739', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '08b7016b-dc44-4206-a00c-5da943b82a38', 'vif_model': 'vmxnet3'}] {{(pid=65726) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1116.655872] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-96c35e52-5464-43ed-b438-b393e0533d7f tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1116.656181] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b353b5e9-500c-42d3-a87f-880a624febec] Creating VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1116.656458] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d6ccb92c-878b-4d40-ad44-42090d5f47d1 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.679678] env[65726]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1116.679678] env[65726]: value = "task-5116757" [ 1116.679678] env[65726]: _type = "Task" [ 1116.679678] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1116.691138] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116757, 'name': CreateVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1116.717147] env[65726]: DEBUG oslo_vmware.api [None req-04bd2146-ec5d-4e36-832f-8911f884caa3 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Task: {'id': task-5116754, 'name': PowerOffVM_Task, 'duration_secs': 0.360514} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1116.717464] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-04bd2146-ec5d-4e36-832f-8911f884caa3 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: ad4f00b5-3765-494f-8132-8b17b29d219a] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1116.717654] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-04bd2146-ec5d-4e36-832f-8911f884caa3 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: ad4f00b5-3765-494f-8132-8b17b29d219a] Unregistering the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1116.717960] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ffe74d37-dd5e-4dd6-a56c-1aa72b8cebe5 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.801876] env[65726]: DEBUG oslo_vmware.api [None req-022a70bf-a10f-4483-94dc-011169b6c2d5 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': task-5116755, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1116.803950] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-04bd2146-ec5d-4e36-832f-8911f884caa3 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: ad4f00b5-3765-494f-8132-8b17b29d219a] Unregistered the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1116.804148] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-04bd2146-ec5d-4e36-832f-8911f884caa3 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: ad4f00b5-3765-494f-8132-8b17b29d219a] Deleting contents of the VM from datastore datastore2 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1116.805143] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-04bd2146-ec5d-4e36-832f-8911f884caa3 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Deleting the datastore file [datastore2] ad4f00b5-3765-494f-8132-8b17b29d219a {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1116.805143] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2a6af2b9-fcd7-4a60-a8e6-4efac23818b8 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.813150] env[65726]: DEBUG oslo_vmware.api [None req-04bd2146-ec5d-4e36-832f-8911f884caa3 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Waiting for the task: (returnval){ [ 1116.813150] env[65726]: value = "task-5116759" [ 1116.813150] env[65726]: _type = "Task" [ 1116.813150] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1116.825105] env[65726]: WARNING neutronclient.v2_0.client [None req-aea91b5d-face-4f98-b031-6ac6f0d1a83b tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1116.828205] env[65726]: DEBUG oslo_vmware.api [None req-04bd2146-ec5d-4e36-832f-8911f884caa3 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Task: {'id': task-5116759, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1116.857414] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-b72da630-f8d0-435a-9469-cadbe319a3e9 tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] [instance: 4809fcf6-59d4-409c-bd8d-981dc85a686f] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1116.857808] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8fc04042-30e8-4c24-98a4-0b5abdd9af07 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.867437] env[65726]: DEBUG oslo_vmware.api [None req-dcfb1735-1ab3-432a-b25b-08fdbc5ed9b1 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Task: {'id': task-5116756, 'name': SuspendVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1116.869964] env[65726]: DEBUG nova.network.neutron [None req-aea91b5d-face-4f98-b031-6ac6f0d1a83b tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 0361ae8f-634d-4161-a594-7f93dbb1b7cd] Port 2bfd7147-87fd-4a19-b28b-eb95f51f4d00 binding to destination host cpu-1 is already ACTIVE {{(pid=65726) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3236}} [ 1116.871992] env[65726]: DEBUG oslo_vmware.api [None req-b72da630-f8d0-435a-9469-cadbe319a3e9 tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Waiting for the task: (returnval){ [ 1116.871992] env[65726]: value = "task-5116760" [ 1116.871992] env[65726]: _type = "Task" [ 1116.871992] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1116.884464] env[65726]: DEBUG oslo_vmware.api [None req-b72da630-f8d0-435a-9469-cadbe319a3e9 tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Task: {'id': task-5116760, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1116.905187] env[65726]: DEBUG oslo_vmware.api [None req-e0232c78-b0b0-436b-b3db-dd24c81ff983 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52d35f0f-98c8-ad5b-4e3a-9359e3658f74, 'name': SearchDatastore_Task, 'duration_secs': 0.011522} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1116.905470] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e0232c78-b0b0-436b-b3db-dd24c81ff983 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1116.905754] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-e0232c78-b0b0-436b-b3db-dd24c81ff983 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore2] e7156a10-631a-4cc5-a544-88ce37763d02/e7156a10-631a-4cc5-a544-88ce37763d02.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1116.906058] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b3486c3a-e00c-4b80-ac9e-a2d61241fa91 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.915739] env[65726]: DEBUG oslo_vmware.api [None req-e0232c78-b0b0-436b-b3db-dd24c81ff983 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Waiting for the task: (returnval){ [ 1116.915739] env[65726]: value = "task-5116761" [ 1116.915739] env[65726]: _type = "Task" [ 1116.915739] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1116.927430] env[65726]: DEBUG oslo_vmware.api [None req-e0232c78-b0b0-436b-b3db-dd24c81ff983 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': task-5116761, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1117.054282] env[65726]: DEBUG nova.compute.utils [None req-ad75268a-0b50-4f80-9f3a-c89856a157c4 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Using /dev/sd instead of None {{(pid=65726) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1117.056367] env[65726]: DEBUG nova.compute.manager [None req-ad75268a-0b50-4f80-9f3a-c89856a157c4 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 378c1d18-4bb6-4245-80bf-370fb7af0575] Allocating IP information in the background. {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1117.056880] env[65726]: DEBUG nova.network.neutron [None req-ad75268a-0b50-4f80-9f3a-c89856a157c4 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 378c1d18-4bb6-4245-80bf-370fb7af0575] allocate_for_instance() {{(pid=65726) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 1117.057286] env[65726]: WARNING neutronclient.v2_0.client [None req-ad75268a-0b50-4f80-9f3a-c89856a157c4 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1117.057637] env[65726]: WARNING neutronclient.v2_0.client [None req-ad75268a-0b50-4f80-9f3a-c89856a157c4 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1117.058380] env[65726]: WARNING openstack [None req-ad75268a-0b50-4f80-9f3a-c89856a157c4 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1117.058789] env[65726]: WARNING openstack [None req-ad75268a-0b50-4f80-9f3a-c89856a157c4 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1117.122508] env[65726]: DEBUG nova.policy [None req-ad75268a-0b50-4f80-9f3a-c89856a157c4 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b251fcbd04044fcda192bc3febac1a90', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4eece77569624f90bf64e5c51974173f', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65726) authorize /opt/stack/nova/nova/policy.py:192}} [ 1117.183841] env[65726]: DEBUG nova.compute.manager [req-203dc1f1-8f0b-40aa-bb22-cc9ef191946e req-214e1ccf-94fd-4ef9-a4f9-917ec8080b05 service nova] [instance: 51f5337d-2e23-4b7c-b590-76be6bd82411] Received event network-changed-fb580bad-fd60-471d-95f2-066352417659 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1117.184249] env[65726]: DEBUG nova.compute.manager [req-203dc1f1-8f0b-40aa-bb22-cc9ef191946e req-214e1ccf-94fd-4ef9-a4f9-917ec8080b05 service nova] [instance: 51f5337d-2e23-4b7c-b590-76be6bd82411] Refreshing instance network info cache due to event network-changed-fb580bad-fd60-471d-95f2-066352417659. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 1117.184514] env[65726]: DEBUG oslo_concurrency.lockutils [req-203dc1f1-8f0b-40aa-bb22-cc9ef191946e req-214e1ccf-94fd-4ef9-a4f9-917ec8080b05 service nova] Acquiring lock "refresh_cache-51f5337d-2e23-4b7c-b590-76be6bd82411" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1117.184699] env[65726]: DEBUG oslo_concurrency.lockutils [req-203dc1f1-8f0b-40aa-bb22-cc9ef191946e req-214e1ccf-94fd-4ef9-a4f9-917ec8080b05 service nova] Acquired lock "refresh_cache-51f5337d-2e23-4b7c-b590-76be6bd82411" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1117.184983] env[65726]: DEBUG nova.network.neutron [req-203dc1f1-8f0b-40aa-bb22-cc9ef191946e req-214e1ccf-94fd-4ef9-a4f9-917ec8080b05 service nova] [instance: 51f5337d-2e23-4b7c-b590-76be6bd82411] Refreshing network info cache for port fb580bad-fd60-471d-95f2-066352417659 {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 1117.203363] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116757, 'name': CreateVM_Task, 'duration_secs': 0.41751} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1117.205194] env[65726]: DEBUG nova.compute.manager [req-63bfaf01-e600-427f-9f86-051495d4a694 req-4afacf0c-f46b-4a35-883a-8041eee268c8 service nova] [instance: b353b5e9-500c-42d3-a87f-880a624febec] Received event network-changed-08b7016b-dc44-4206-a00c-5da943b82a38 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1117.208127] env[65726]: DEBUG nova.compute.manager [req-63bfaf01-e600-427f-9f86-051495d4a694 req-4afacf0c-f46b-4a35-883a-8041eee268c8 service nova] [instance: b353b5e9-500c-42d3-a87f-880a624febec] Refreshing instance network info cache due to event network-changed-08b7016b-dc44-4206-a00c-5da943b82a38. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 1117.208127] env[65726]: DEBUG oslo_concurrency.lockutils [req-63bfaf01-e600-427f-9f86-051495d4a694 req-4afacf0c-f46b-4a35-883a-8041eee268c8 service nova] Acquiring lock "refresh_cache-b353b5e9-500c-42d3-a87f-880a624febec" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1117.208127] env[65726]: DEBUG oslo_concurrency.lockutils [req-63bfaf01-e600-427f-9f86-051495d4a694 req-4afacf0c-f46b-4a35-883a-8041eee268c8 service nova] Acquired lock "refresh_cache-b353b5e9-500c-42d3-a87f-880a624febec" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1117.208127] env[65726]: DEBUG nova.network.neutron [req-63bfaf01-e600-427f-9f86-051495d4a694 req-4afacf0c-f46b-4a35-883a-8041eee268c8 service nova] [instance: b353b5e9-500c-42d3-a87f-880a624febec] Refreshing network info cache for port 08b7016b-dc44-4206-a00c-5da943b82a38 {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 1117.208127] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b353b5e9-500c-42d3-a87f-880a624febec] Created VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1117.208639] env[65726]: WARNING openstack [None req-96c35e52-5464-43ed-b438-b393e0533d7f tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1117.209028] env[65726]: WARNING openstack [None req-96c35e52-5464-43ed-b438-b393e0533d7f tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1117.215220] env[65726]: DEBUG oslo_concurrency.lockutils [None req-96c35e52-5464-43ed-b438-b393e0533d7f tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1117.215657] env[65726]: DEBUG oslo_concurrency.lockutils [None req-96c35e52-5464-43ed-b438-b393e0533d7f tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1117.216076] env[65726]: DEBUG oslo_concurrency.lockutils [None req-96c35e52-5464-43ed-b438-b393e0533d7f tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1117.220095] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dda31586-390a-482e-b7ff-ebc1c42ee5b7 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.225638] env[65726]: DEBUG oslo_vmware.api [None req-96c35e52-5464-43ed-b438-b393e0533d7f tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Waiting for the task: (returnval){ [ 1117.225638] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5229431d-7e0c-ef10-6fb8-f3b2774684da" [ 1117.225638] env[65726]: _type = "Task" [ 1117.225638] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1117.236841] env[65726]: DEBUG oslo_vmware.api [None req-96c35e52-5464-43ed-b438-b393e0533d7f tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5229431d-7e0c-ef10-6fb8-f3b2774684da, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1117.300094] env[65726]: DEBUG oslo_vmware.api [None req-022a70bf-a10f-4483-94dc-011169b6c2d5 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': task-5116755, 'name': ReconfigVM_Task, 'duration_secs': 0.621556} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1117.301205] env[65726]: WARNING openstack [None req-022a70bf-a10f-4483-94dc-011169b6c2d5 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1117.301457] env[65726]: WARNING openstack [None req-022a70bf-a10f-4483-94dc-011169b6c2d5 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1117.309099] env[65726]: DEBUG oslo_concurrency.lockutils [None req-022a70bf-a10f-4483-94dc-011169b6c2d5 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Releasing lock "51f5337d-2e23-4b7c-b590-76be6bd82411" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1117.309099] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-022a70bf-a10f-4483-94dc-011169b6c2d5 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: 51f5337d-2e23-4b7c-b590-76be6bd82411] Reconfigured VM to attach interface {{(pid=65726) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 1117.327660] env[65726]: DEBUG oslo_vmware.api [None req-04bd2146-ec5d-4e36-832f-8911f884caa3 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Task: {'id': task-5116759, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.372048} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1117.327957] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-04bd2146-ec5d-4e36-832f-8911f884caa3 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Deleted the datastore file {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1117.328244] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-04bd2146-ec5d-4e36-832f-8911f884caa3 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: ad4f00b5-3765-494f-8132-8b17b29d219a] Deleted contents of the VM from datastore datastore2 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1117.328343] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-04bd2146-ec5d-4e36-832f-8911f884caa3 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: ad4f00b5-3765-494f-8132-8b17b29d219a] Instance destroyed {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1117.328524] env[65726]: INFO nova.compute.manager [None req-04bd2146-ec5d-4e36-832f-8911f884caa3 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: ad4f00b5-3765-494f-8132-8b17b29d219a] Took 1.14 seconds to destroy the instance on the hypervisor. [ 1117.328814] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-04bd2146-ec5d-4e36-832f-8911f884caa3 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1117.329054] env[65726]: DEBUG nova.compute.manager [-] [instance: ad4f00b5-3765-494f-8132-8b17b29d219a] Deallocating network for instance {{(pid=65726) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1117.329178] env[65726]: DEBUG nova.network.neutron [-] [instance: ad4f00b5-3765-494f-8132-8b17b29d219a] deallocate_for_instance() {{(pid=65726) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 1117.331589] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1117.331893] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1117.368091] env[65726]: DEBUG oslo_vmware.api [None req-dcfb1735-1ab3-432a-b25b-08fdbc5ed9b1 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Task: {'id': task-5116756, 'name': SuspendVM_Task} progress is 58%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1117.385796] env[65726]: DEBUG oslo_vmware.api [None req-b72da630-f8d0-435a-9469-cadbe319a3e9 tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Task: {'id': task-5116760, 'name': PowerOffVM_Task, 'duration_secs': 0.24838} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1117.386400] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-b72da630-f8d0-435a-9469-cadbe319a3e9 tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] [instance: 4809fcf6-59d4-409c-bd8d-981dc85a686f] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1117.387368] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-115bb700-6a8e-4303-b373-9aaa6eb1aac9 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.424187] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20088db7-4a1a-465e-a649-7befccdbd79b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.440039] env[65726]: DEBUG oslo_vmware.api [None req-e0232c78-b0b0-436b-b3db-dd24c81ff983 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': task-5116761, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1117.493090] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-b72da630-f8d0-435a-9469-cadbe319a3e9 tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] [instance: 4809fcf6-59d4-409c-bd8d-981dc85a686f] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1117.493090] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-96c81e98-4268-49ff-9f7e-f344490c3941 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.501682] env[65726]: DEBUG oslo_vmware.api [None req-b72da630-f8d0-435a-9469-cadbe319a3e9 tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Waiting for the task: (returnval){ [ 1117.501682] env[65726]: value = "task-5116762" [ 1117.501682] env[65726]: _type = "Task" [ 1117.501682] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1117.518434] env[65726]: DEBUG nova.network.neutron [None req-ad75268a-0b50-4f80-9f3a-c89856a157c4 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 378c1d18-4bb6-4245-80bf-370fb7af0575] Successfully created port: 4765f6c2-d161-4e36-a6a8-3c8d39c47a40 {{(pid=65726) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1117.522629] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-b72da630-f8d0-435a-9469-cadbe319a3e9 tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] [instance: 4809fcf6-59d4-409c-bd8d-981dc85a686f] VM already powered off {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1117.522860] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-b72da630-f8d0-435a-9469-cadbe319a3e9 tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] [instance: 4809fcf6-59d4-409c-bd8d-981dc85a686f] Processing image b52362a3-ee8a-4cbf-b06f-513b0cc8f95c {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1117.523176] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b72da630-f8d0-435a-9469-cadbe319a3e9 tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1117.523326] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b72da630-f8d0-435a-9469-cadbe319a3e9 tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1117.523519] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-b72da630-f8d0-435a-9469-cadbe319a3e9 tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1117.527514] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1117.527906] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1117.537455] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8d991e99-c7eb-40f8-a1b6-8a8240e36778 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.566030] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-b72da630-f8d0-435a-9469-cadbe319a3e9 tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1117.566030] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-b72da630-f8d0-435a-9469-cadbe319a3e9 tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65726) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1117.566799] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e2961604-82f4-4f42-9d89-379c48009d69 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.569764] env[65726]: DEBUG nova.compute.manager [None req-ad75268a-0b50-4f80-9f3a-c89856a157c4 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 378c1d18-4bb6-4245-80bf-370fb7af0575] Start building block device mappings for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1117.576957] env[65726]: DEBUG oslo_vmware.api [None req-b72da630-f8d0-435a-9469-cadbe319a3e9 tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Waiting for the task: (returnval){ [ 1117.576957] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5256501e-e70c-d425-b103-5df125337188" [ 1117.576957] env[65726]: _type = "Task" [ 1117.576957] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1117.586705] env[65726]: DEBUG oslo_vmware.api [None req-b72da630-f8d0-435a-9469-cadbe319a3e9 tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5256501e-e70c-d425-b103-5df125337188, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1117.697045] env[65726]: WARNING openstack [req-203dc1f1-8f0b-40aa-bb22-cc9ef191946e req-214e1ccf-94fd-4ef9-a4f9-917ec8080b05 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1117.697450] env[65726]: WARNING openstack [req-203dc1f1-8f0b-40aa-bb22-cc9ef191946e req-214e1ccf-94fd-4ef9-a4f9-917ec8080b05 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1117.718798] env[65726]: WARNING openstack [req-63bfaf01-e600-427f-9f86-051495d4a694 req-4afacf0c-f46b-4a35-883a-8041eee268c8 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1117.719441] env[65726]: WARNING openstack [req-63bfaf01-e600-427f-9f86-051495d4a694 req-4afacf0c-f46b-4a35-883a-8041eee268c8 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1117.738475] env[65726]: DEBUG oslo_vmware.api [None req-96c35e52-5464-43ed-b438-b393e0533d7f tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5229431d-7e0c-ef10-6fb8-f3b2774684da, 'name': SearchDatastore_Task, 'duration_secs': 0.082903} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1117.738777] env[65726]: DEBUG oslo_concurrency.lockutils [None req-96c35e52-5464-43ed-b438-b393e0533d7f tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1117.739086] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-96c35e52-5464-43ed-b438-b393e0533d7f tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: b353b5e9-500c-42d3-a87f-880a624febec] Processing image b52362a3-ee8a-4cbf-b06f-513b0cc8f95c {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1117.739363] env[65726]: DEBUG oslo_concurrency.lockutils [None req-96c35e52-5464-43ed-b438-b393e0533d7f tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1117.739512] env[65726]: DEBUG oslo_concurrency.lockutils [None req-96c35e52-5464-43ed-b438-b393e0533d7f tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1117.739693] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-96c35e52-5464-43ed-b438-b393e0533d7f tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1117.739962] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-bd2b72ce-0baf-4945-bbdf-50c90ddeca91 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.749385] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-96c35e52-5464-43ed-b438-b393e0533d7f tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1117.749581] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-96c35e52-5464-43ed-b438-b393e0533d7f tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65726) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1117.750330] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dc91f81a-3164-49f8-b802-0b18e14339e2 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.756086] env[65726]: DEBUG oslo_vmware.api [None req-96c35e52-5464-43ed-b438-b393e0533d7f tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Waiting for the task: (returnval){ [ 1117.756086] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5257fe3f-0758-ebe3-cbb9-439c49a216e6" [ 1117.756086] env[65726]: _type = "Task" [ 1117.756086] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1117.764878] env[65726]: DEBUG oslo_vmware.api [None req-96c35e52-5464-43ed-b438-b393e0533d7f tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5257fe3f-0758-ebe3-cbb9-439c49a216e6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1117.823204] env[65726]: DEBUG oslo_concurrency.lockutils [None req-022a70bf-a10f-4483-94dc-011169b6c2d5 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Lock "interface-51f5337d-2e23-4b7c-b590-76be6bd82411-fb580bad-fd60-471d-95f2-066352417659" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 6.816s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1117.865389] env[65726]: DEBUG oslo_vmware.api [None req-dcfb1735-1ab3-432a-b25b-08fdbc5ed9b1 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Task: {'id': task-5116756, 'name': SuspendVM_Task, 'duration_secs': 1.210602} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1117.865955] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-dcfb1735-1ab3-432a-b25b-08fdbc5ed9b1 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] [instance: 6fd70b4a-63e4-4258-9cff-f3b582500b9f] Suspended the VM {{(pid=65726) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1166}} [ 1117.866194] env[65726]: DEBUG nova.compute.manager [None req-dcfb1735-1ab3-432a-b25b-08fdbc5ed9b1 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] [instance: 6fd70b4a-63e4-4258-9cff-f3b582500b9f] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1117.866988] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4576072-5fb6-465f-8028-659b918c68f2 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.902334] env[65726]: DEBUG oslo_concurrency.lockutils [None req-aea91b5d-face-4f98-b031-6ac6f0d1a83b tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Acquiring lock "0361ae8f-634d-4161-a594-7f93dbb1b7cd-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1117.902493] env[65726]: DEBUG oslo_concurrency.lockutils [None req-aea91b5d-face-4f98-b031-6ac6f0d1a83b tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Lock "0361ae8f-634d-4161-a594-7f93dbb1b7cd-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1117.902534] env[65726]: DEBUG oslo_concurrency.lockutils [None req-aea91b5d-face-4f98-b031-6ac6f0d1a83b tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Lock "0361ae8f-634d-4161-a594-7f93dbb1b7cd-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1117.929985] env[65726]: DEBUG oslo_vmware.api [None req-e0232c78-b0b0-436b-b3db-dd24c81ff983 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': task-5116761, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.671961} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1117.930278] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-e0232c78-b0b0-436b-b3db-dd24c81ff983 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore2] e7156a10-631a-4cc5-a544-88ce37763d02/e7156a10-631a-4cc5-a544-88ce37763d02.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1117.930493] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-e0232c78-b0b0-436b-b3db-dd24c81ff983 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: e7156a10-631a-4cc5-a544-88ce37763d02] Extending root virtual disk to 1048576 {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1117.930832] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-6cef3d95-96f0-496c-b46d-af80009df6cc {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.938718] env[65726]: DEBUG oslo_vmware.api [None req-e0232c78-b0b0-436b-b3db-dd24c81ff983 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Waiting for the task: (returnval){ [ 1117.938718] env[65726]: value = "task-5116763" [ 1117.938718] env[65726]: _type = "Task" [ 1117.938718] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1117.955085] env[65726]: DEBUG oslo_vmware.api [None req-e0232c78-b0b0-436b-b3db-dd24c81ff983 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': task-5116763, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1118.090317] env[65726]: DEBUG oslo_vmware.api [None req-b72da630-f8d0-435a-9469-cadbe319a3e9 tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5256501e-e70c-d425-b103-5df125337188, 'name': SearchDatastore_Task, 'duration_secs': 0.013961} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1118.091495] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2909d120-8e90-415b-9047-bb5bf1eb5eb5 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.098661] env[65726]: DEBUG oslo_vmware.api [None req-b72da630-f8d0-435a-9469-cadbe319a3e9 tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Waiting for the task: (returnval){ [ 1118.098661] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52465085-8bae-64d3-5c79-3215443d0a8b" [ 1118.098661] env[65726]: _type = "Task" [ 1118.098661] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1118.108699] env[65726]: DEBUG oslo_vmware.api [None req-b72da630-f8d0-435a-9469-cadbe319a3e9 tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52465085-8bae-64d3-5c79-3215443d0a8b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1118.269745] env[65726]: DEBUG oslo_vmware.api [None req-96c35e52-5464-43ed-b438-b393e0533d7f tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5257fe3f-0758-ebe3-cbb9-439c49a216e6, 'name': SearchDatastore_Task, 'duration_secs': 0.009762} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1118.271949] env[65726]: WARNING openstack [req-203dc1f1-8f0b-40aa-bb22-cc9ef191946e req-214e1ccf-94fd-4ef9-a4f9-917ec8080b05 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1118.272540] env[65726]: WARNING openstack [req-203dc1f1-8f0b-40aa-bb22-cc9ef191946e req-214e1ccf-94fd-4ef9-a4f9-917ec8080b05 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1118.281246] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-66ce4040-9d86-4de7-b2e3-d797c2d51c2e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.288239] env[65726]: DEBUG oslo_vmware.api [None req-96c35e52-5464-43ed-b438-b393e0533d7f tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Waiting for the task: (returnval){ [ 1118.288239] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]526c97f5-9a4d-654f-09e0-903ced60bbb1" [ 1118.288239] env[65726]: _type = "Task" [ 1118.288239] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1118.299694] env[65726]: DEBUG oslo_vmware.api [None req-96c35e52-5464-43ed-b438-b393e0533d7f tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]526c97f5-9a4d-654f-09e0-903ced60bbb1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1118.315492] env[65726]: WARNING openstack [req-63bfaf01-e600-427f-9f86-051495d4a694 req-4afacf0c-f46b-4a35-883a-8041eee268c8 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1118.315872] env[65726]: WARNING openstack [req-63bfaf01-e600-427f-9f86-051495d4a694 req-4afacf0c-f46b-4a35-883a-8041eee268c8 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1118.451590] env[65726]: DEBUG oslo_vmware.api [None req-e0232c78-b0b0-436b-b3db-dd24c81ff983 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': task-5116763, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.118826} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1118.451907] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-e0232c78-b0b0-436b-b3db-dd24c81ff983 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: e7156a10-631a-4cc5-a544-88ce37763d02] Extended root virtual disk {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1118.452768] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c43dd96f-2986-45b6-9323-f721d4ab34c6 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.455717] env[65726]: DEBUG nova.network.neutron [-] [instance: ad4f00b5-3765-494f-8132-8b17b29d219a] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1118.483273] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-e0232c78-b0b0-436b-b3db-dd24c81ff983 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: e7156a10-631a-4cc5-a544-88ce37763d02] Reconfiguring VM instance instance-00000066 to attach disk [datastore2] e7156a10-631a-4cc5-a544-88ce37763d02/e7156a10-631a-4cc5-a544-88ce37763d02.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1118.484588] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d7b26253-4c0b-4d98-95ed-c14c6918da1f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.506083] env[65726]: DEBUG oslo_vmware.api [None req-e0232c78-b0b0-436b-b3db-dd24c81ff983 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Waiting for the task: (returnval){ [ 1118.506083] env[65726]: value = "task-5116764" [ 1118.506083] env[65726]: _type = "Task" [ 1118.506083] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1118.516611] env[65726]: DEBUG oslo_vmware.api [None req-e0232c78-b0b0-436b-b3db-dd24c81ff983 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': task-5116764, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1118.586533] env[65726]: DEBUG nova.compute.manager [None req-ad75268a-0b50-4f80-9f3a-c89856a157c4 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 378c1d18-4bb6-4245-80bf-370fb7af0575] Start spawning the instance on the hypervisor. {{(pid=65726) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1118.618689] env[65726]: DEBUG oslo_vmware.api [None req-b72da630-f8d0-435a-9469-cadbe319a3e9 tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52465085-8bae-64d3-5c79-3215443d0a8b, 'name': SearchDatastore_Task, 'duration_secs': 0.011141} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1118.619220] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b72da630-f8d0-435a-9469-cadbe319a3e9 tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1118.620261] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-b72da630-f8d0-435a-9469-cadbe319a3e9 tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Copying virtual disk from [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] 4809fcf6-59d4-409c-bd8d-981dc85a686f/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c-rescue.vmdk. {{(pid=65726) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 1118.620550] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-fedca58b-82dc-4b0a-94b9-f0dcb847dffb {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.629563] env[65726]: DEBUG nova.virt.hardware [None req-ad75268a-0b50-4f80-9f3a-c89856a157c4 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T19:28:34Z,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1539f06fa3534e90acbc3a60c4b8bd3f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T19:28:35Z,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1118.629795] env[65726]: DEBUG nova.virt.hardware [None req-ad75268a-0b50-4f80-9f3a-c89856a157c4 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1118.633406] env[65726]: DEBUG nova.virt.hardware [None req-ad75268a-0b50-4f80-9f3a-c89856a157c4 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1118.633813] env[65726]: DEBUG nova.virt.hardware [None req-ad75268a-0b50-4f80-9f3a-c89856a157c4 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1118.634212] env[65726]: DEBUG nova.virt.hardware [None req-ad75268a-0b50-4f80-9f3a-c89856a157c4 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1118.634426] env[65726]: DEBUG nova.virt.hardware [None req-ad75268a-0b50-4f80-9f3a-c89856a157c4 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1118.634695] env[65726]: DEBUG nova.virt.hardware [None req-ad75268a-0b50-4f80-9f3a-c89856a157c4 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1118.634866] env[65726]: DEBUG nova.virt.hardware [None req-ad75268a-0b50-4f80-9f3a-c89856a157c4 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1118.635071] env[65726]: DEBUG nova.virt.hardware [None req-ad75268a-0b50-4f80-9f3a-c89856a157c4 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1118.635282] env[65726]: DEBUG nova.virt.hardware [None req-ad75268a-0b50-4f80-9f3a-c89856a157c4 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1118.635492] env[65726]: DEBUG nova.virt.hardware [None req-ad75268a-0b50-4f80-9f3a-c89856a157c4 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1118.637170] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b93f2bdc-607b-43da-946b-36fbbf289529 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.642085] env[65726]: DEBUG oslo_vmware.api [None req-b72da630-f8d0-435a-9469-cadbe319a3e9 tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Waiting for the task: (returnval){ [ 1118.642085] env[65726]: value = "task-5116765" [ 1118.642085] env[65726]: _type = "Task" [ 1118.642085] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1118.655320] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5bc36c1c-8610-46f6-9d8b-493047330978 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.661528] env[65726]: DEBUG oslo_vmware.api [None req-b72da630-f8d0-435a-9469-cadbe319a3e9 tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Task: {'id': task-5116765, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1118.801875] env[65726]: DEBUG oslo_vmware.api [None req-96c35e52-5464-43ed-b438-b393e0533d7f tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]526c97f5-9a4d-654f-09e0-903ced60bbb1, 'name': SearchDatastore_Task, 'duration_secs': 0.011751} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1118.801875] env[65726]: DEBUG oslo_concurrency.lockutils [None req-96c35e52-5464-43ed-b438-b393e0533d7f tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1118.802109] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-96c35e52-5464-43ed-b438-b393e0533d7f tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore2] b353b5e9-500c-42d3-a87f-880a624febec/b353b5e9-500c-42d3-a87f-880a624febec.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1118.802505] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4e378d10-812f-461f-9bd9-2b052a0ad0d6 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.813126] env[65726]: DEBUG oslo_vmware.api [None req-96c35e52-5464-43ed-b438-b393e0533d7f tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Waiting for the task: (returnval){ [ 1118.813126] env[65726]: value = "task-5116766" [ 1118.813126] env[65726]: _type = "Task" [ 1118.813126] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1118.824820] env[65726]: DEBUG oslo_vmware.api [None req-96c35e52-5464-43ed-b438-b393e0533d7f tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Task: {'id': task-5116766, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1118.858530] env[65726]: WARNING openstack [req-203dc1f1-8f0b-40aa-bb22-cc9ef191946e req-214e1ccf-94fd-4ef9-a4f9-917ec8080b05 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1118.858987] env[65726]: WARNING openstack [req-203dc1f1-8f0b-40aa-bb22-cc9ef191946e req-214e1ccf-94fd-4ef9-a4f9-917ec8080b05 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1118.914029] env[65726]: WARNING openstack [None req-aea91b5d-face-4f98-b031-6ac6f0d1a83b tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1118.914327] env[65726]: WARNING openstack [None req-aea91b5d-face-4f98-b031-6ac6f0d1a83b tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1118.958545] env[65726]: INFO nova.compute.manager [-] [instance: ad4f00b5-3765-494f-8132-8b17b29d219a] Took 1.63 seconds to deallocate network for instance. [ 1119.019167] env[65726]: DEBUG oslo_vmware.api [None req-e0232c78-b0b0-436b-b3db-dd24c81ff983 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': task-5116764, 'name': ReconfigVM_Task, 'duration_secs': 0.288312} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1119.019862] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-e0232c78-b0b0-436b-b3db-dd24c81ff983 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: e7156a10-631a-4cc5-a544-88ce37763d02] Reconfigured VM instance instance-00000066 to attach disk [datastore2] e7156a10-631a-4cc5-a544-88ce37763d02/e7156a10-631a-4cc5-a544-88ce37763d02.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1119.020582] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-42b1429c-6e9b-4333-88ba-819e2a6e8aea {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.030951] env[65726]: DEBUG oslo_vmware.api [None req-e0232c78-b0b0-436b-b3db-dd24c81ff983 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Waiting for the task: (returnval){ [ 1119.030951] env[65726]: value = "task-5116767" [ 1119.030951] env[65726]: _type = "Task" [ 1119.030951] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1119.044177] env[65726]: DEBUG oslo_vmware.api [None req-e0232c78-b0b0-436b-b3db-dd24c81ff983 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': task-5116767, 'name': Rename_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1119.151455] env[65726]: DEBUG nova.network.neutron [None req-ad75268a-0b50-4f80-9f3a-c89856a157c4 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 378c1d18-4bb6-4245-80bf-370fb7af0575] Successfully updated port: 4765f6c2-d161-4e36-a6a8-3c8d39c47a40 {{(pid=65726) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 1119.160235] env[65726]: DEBUG oslo_vmware.api [None req-b72da630-f8d0-435a-9469-cadbe319a3e9 tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Task: {'id': task-5116765, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.527169} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1119.160569] env[65726]: INFO nova.virt.vmwareapi.ds_util [None req-b72da630-f8d0-435a-9469-cadbe319a3e9 tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Copied virtual disk from [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] 4809fcf6-59d4-409c-bd8d-981dc85a686f/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c-rescue.vmdk. [ 1119.161517] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3eb3052b-7130-4804-814d-58e1004d71fd {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.192404] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-b72da630-f8d0-435a-9469-cadbe319a3e9 tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] [instance: 4809fcf6-59d4-409c-bd8d-981dc85a686f] Reconfiguring VM instance instance-00000065 to attach disk [datastore1] 4809fcf6-59d4-409c-bd8d-981dc85a686f/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c-rescue.vmdk or device None with type thin {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1119.196188] env[65726]: WARNING openstack [req-63bfaf01-e600-427f-9f86-051495d4a694 req-4afacf0c-f46b-4a35-883a-8041eee268c8 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1119.196765] env[65726]: WARNING openstack [req-63bfaf01-e600-427f-9f86-051495d4a694 req-4afacf0c-f46b-4a35-883a-8041eee268c8 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1119.205065] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-148a525a-f879-4642-9179-ab495512b7ad {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.226366] env[65726]: DEBUG oslo_vmware.api [None req-b72da630-f8d0-435a-9469-cadbe319a3e9 tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Waiting for the task: (returnval){ [ 1119.226366] env[65726]: value = "task-5116768" [ 1119.226366] env[65726]: _type = "Task" [ 1119.226366] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1119.243649] env[65726]: DEBUG oslo_vmware.api [None req-b72da630-f8d0-435a-9469-cadbe319a3e9 tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Task: {'id': task-5116768, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1119.254917] env[65726]: DEBUG oslo_concurrency.lockutils [None req-aea91b5d-face-4f98-b031-6ac6f0d1a83b tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Acquiring lock "refresh_cache-0361ae8f-634d-4161-a594-7f93dbb1b7cd" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1119.254917] env[65726]: DEBUG oslo_concurrency.lockutils [None req-aea91b5d-face-4f98-b031-6ac6f0d1a83b tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Acquired lock "refresh_cache-0361ae8f-634d-4161-a594-7f93dbb1b7cd" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1119.254917] env[65726]: DEBUG nova.network.neutron [None req-aea91b5d-face-4f98-b031-6ac6f0d1a83b tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 0361ae8f-634d-4161-a594-7f93dbb1b7cd] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1119.327697] env[65726]: DEBUG oslo_vmware.api [None req-96c35e52-5464-43ed-b438-b393e0533d7f tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Task: {'id': task-5116766, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1119.370737] env[65726]: DEBUG nova.network.neutron [req-203dc1f1-8f0b-40aa-bb22-cc9ef191946e req-214e1ccf-94fd-4ef9-a4f9-917ec8080b05 service nova] [instance: 51f5337d-2e23-4b7c-b590-76be6bd82411] Updated VIF entry in instance network info cache for port fb580bad-fd60-471d-95f2-066352417659. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 1119.371858] env[65726]: DEBUG nova.network.neutron [req-203dc1f1-8f0b-40aa-bb22-cc9ef191946e req-214e1ccf-94fd-4ef9-a4f9-917ec8080b05 service nova] [instance: 51f5337d-2e23-4b7c-b590-76be6bd82411] Updating instance_info_cache with network_info: [{"id": "7407bcd9-3202-4798-9eaa-252fb3fdb21f", "address": "fa:16:3e:79:d2:9d", "network": {"id": "51a66448-0f72-4266-8868-856b48be839f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-155658961-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.213", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d46ccb00794f458b85da4a93879139ee", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b1f3e6c3-5584-4852-9017-476ab8ac4946", "external-id": "nsx-vlan-transportzone-304", "segmentation_id": 304, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7407bcd9-32", "ovs_interfaceid": "7407bcd9-3202-4798-9eaa-252fb3fdb21f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "fb580bad-fd60-471d-95f2-066352417659", "address": "fa:16:3e:32:0d:d6", "network": {"id": "51a66448-0f72-4266-8868-856b48be839f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-155658961-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d46ccb00794f458b85da4a93879139ee", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b1f3e6c3-5584-4852-9017-476ab8ac4946", "external-id": "nsx-vlan-transportzone-304", "segmentation_id": 304, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfb580bad-fd", "ovs_interfaceid": "fb580bad-fd60-471d-95f2-066352417659", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1119.420350] env[65726]: DEBUG nova.network.neutron [req-63bfaf01-e600-427f-9f86-051495d4a694 req-4afacf0c-f46b-4a35-883a-8041eee268c8 service nova] [instance: b353b5e9-500c-42d3-a87f-880a624febec] Updated VIF entry in instance network info cache for port 08b7016b-dc44-4206-a00c-5da943b82a38. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 1119.420976] env[65726]: DEBUG nova.network.neutron [req-63bfaf01-e600-427f-9f86-051495d4a694 req-4afacf0c-f46b-4a35-883a-8041eee268c8 service nova] [instance: b353b5e9-500c-42d3-a87f-880a624febec] Updating instance_info_cache with network_info: [{"id": "08b7016b-dc44-4206-a00c-5da943b82a38", "address": "fa:16:3e:c5:e4:14", "network": {"id": "eb3278b7-c190-441b-84d2-305f9b896382", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1754571022-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "96149159e18e44f9bf3453e67681f224", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f77ff7a1-209c-4f3f-b2a0-fd817741e739", "external-id": "nsx-vlan-transportzone-935", "segmentation_id": 935, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap08b7016b-dc", "ovs_interfaceid": "08b7016b-dc44-4206-a00c-5da943b82a38", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1119.468516] env[65726]: DEBUG oslo_concurrency.lockutils [None req-04bd2146-ec5d-4e36-832f-8911f884caa3 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1119.468804] env[65726]: DEBUG oslo_concurrency.lockutils [None req-04bd2146-ec5d-4e36-832f-8911f884caa3 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1119.469140] env[65726]: DEBUG nova.objects.instance [None req-04bd2146-ec5d-4e36-832f-8911f884caa3 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Lazy-loading 'resources' on Instance uuid ad4f00b5-3765-494f-8132-8b17b29d219a {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1119.546691] env[65726]: DEBUG oslo_vmware.api [None req-e0232c78-b0b0-436b-b3db-dd24c81ff983 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': task-5116767, 'name': Rename_Task, 'duration_secs': 0.234718} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1119.547598] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-e0232c78-b0b0-436b-b3db-dd24c81ff983 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: e7156a10-631a-4cc5-a544-88ce37763d02] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1119.547896] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-96ac9025-7682-4e8e-b57e-71be97b50710 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.559844] env[65726]: DEBUG oslo_vmware.api [None req-e0232c78-b0b0-436b-b3db-dd24c81ff983 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Waiting for the task: (returnval){ [ 1119.559844] env[65726]: value = "task-5116769" [ 1119.559844] env[65726]: _type = "Task" [ 1119.559844] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1119.571200] env[65726]: DEBUG oslo_vmware.api [None req-e0232c78-b0b0-436b-b3db-dd24c81ff983 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': task-5116769, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1119.655978] env[65726]: DEBUG oslo_concurrency.lockutils [None req-ad75268a-0b50-4f80-9f3a-c89856a157c4 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Acquiring lock "refresh_cache-378c1d18-4bb6-4245-80bf-370fb7af0575" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1119.656403] env[65726]: DEBUG oslo_concurrency.lockutils [None req-ad75268a-0b50-4f80-9f3a-c89856a157c4 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Acquired lock "refresh_cache-378c1d18-4bb6-4245-80bf-370fb7af0575" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1119.656727] env[65726]: DEBUG nova.network.neutron [None req-ad75268a-0b50-4f80-9f3a-c89856a157c4 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 378c1d18-4bb6-4245-80bf-370fb7af0575] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1119.699179] env[65726]: DEBUG nova.compute.manager [req-f1d08b0b-7e4c-4185-a08b-f8af209576fa req-14a4037a-f5f0-4730-a8b9-2b34b962b453 service nova] [instance: 378c1d18-4bb6-4245-80bf-370fb7af0575] Received event network-vif-plugged-4765f6c2-d161-4e36-a6a8-3c8d39c47a40 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1119.699380] env[65726]: DEBUG oslo_concurrency.lockutils [req-f1d08b0b-7e4c-4185-a08b-f8af209576fa req-14a4037a-f5f0-4730-a8b9-2b34b962b453 service nova] Acquiring lock "378c1d18-4bb6-4245-80bf-370fb7af0575-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1119.699641] env[65726]: DEBUG oslo_concurrency.lockutils [req-f1d08b0b-7e4c-4185-a08b-f8af209576fa req-14a4037a-f5f0-4730-a8b9-2b34b962b453 service nova] Lock "378c1d18-4bb6-4245-80bf-370fb7af0575-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1119.699904] env[65726]: DEBUG oslo_concurrency.lockutils [req-f1d08b0b-7e4c-4185-a08b-f8af209576fa req-14a4037a-f5f0-4730-a8b9-2b34b962b453 service nova] Lock "378c1d18-4bb6-4245-80bf-370fb7af0575-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1119.700183] env[65726]: DEBUG nova.compute.manager [req-f1d08b0b-7e4c-4185-a08b-f8af209576fa req-14a4037a-f5f0-4730-a8b9-2b34b962b453 service nova] [instance: 378c1d18-4bb6-4245-80bf-370fb7af0575] No waiting events found dispatching network-vif-plugged-4765f6c2-d161-4e36-a6a8-3c8d39c47a40 {{(pid=65726) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1119.700393] env[65726]: WARNING nova.compute.manager [req-f1d08b0b-7e4c-4185-a08b-f8af209576fa req-14a4037a-f5f0-4730-a8b9-2b34b962b453 service nova] [instance: 378c1d18-4bb6-4245-80bf-370fb7af0575] Received unexpected event network-vif-plugged-4765f6c2-d161-4e36-a6a8-3c8d39c47a40 for instance with vm_state building and task_state spawning. [ 1119.700600] env[65726]: DEBUG nova.compute.manager [req-f1d08b0b-7e4c-4185-a08b-f8af209576fa req-14a4037a-f5f0-4730-a8b9-2b34b962b453 service nova] [instance: 378c1d18-4bb6-4245-80bf-370fb7af0575] Received event network-changed-4765f6c2-d161-4e36-a6a8-3c8d39c47a40 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1119.700768] env[65726]: DEBUG nova.compute.manager [req-f1d08b0b-7e4c-4185-a08b-f8af209576fa req-14a4037a-f5f0-4730-a8b9-2b34b962b453 service nova] [instance: 378c1d18-4bb6-4245-80bf-370fb7af0575] Refreshing instance network info cache due to event network-changed-4765f6c2-d161-4e36-a6a8-3c8d39c47a40. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 1119.701077] env[65726]: DEBUG oslo_concurrency.lockutils [req-f1d08b0b-7e4c-4185-a08b-f8af209576fa req-14a4037a-f5f0-4730-a8b9-2b34b962b453 service nova] Acquiring lock "refresh_cache-378c1d18-4bb6-4245-80bf-370fb7af0575" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1119.719012] env[65726]: DEBUG nova.compute.manager [req-4214de48-c780-4156-8ecb-c2211987914d req-89bee8d2-9dd1-4a3b-acf8-12d314606eee service nova] [instance: ad4f00b5-3765-494f-8132-8b17b29d219a] Received event network-vif-deleted-508e0342-f4cf-409d-96e4-742e92925ba6 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1119.737751] env[65726]: DEBUG oslo_vmware.api [None req-b72da630-f8d0-435a-9469-cadbe319a3e9 tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Task: {'id': task-5116768, 'name': ReconfigVM_Task, 'duration_secs': 0.413757} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1119.738073] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-b72da630-f8d0-435a-9469-cadbe319a3e9 tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] [instance: 4809fcf6-59d4-409c-bd8d-981dc85a686f] Reconfigured VM instance instance-00000065 to attach disk [datastore1] 4809fcf6-59d4-409c-bd8d-981dc85a686f/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c-rescue.vmdk or device None with type thin {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1119.738976] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff511ff8-10d8-4df5-a702-6dee92204ff6 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.763328] env[65726]: WARNING openstack [None req-aea91b5d-face-4f98-b031-6ac6f0d1a83b tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1119.763860] env[65726]: WARNING openstack [None req-aea91b5d-face-4f98-b031-6ac6f0d1a83b tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1119.777720] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2d1646bc-371c-4400-a6d4-7efdb47f2734 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.797523] env[65726]: DEBUG oslo_vmware.api [None req-b72da630-f8d0-435a-9469-cadbe319a3e9 tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Waiting for the task: (returnval){ [ 1119.797523] env[65726]: value = "task-5116770" [ 1119.797523] env[65726]: _type = "Task" [ 1119.797523] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1119.807864] env[65726]: DEBUG oslo_vmware.api [None req-b72da630-f8d0-435a-9469-cadbe319a3e9 tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Task: {'id': task-5116770, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1119.826285] env[65726]: DEBUG oslo_vmware.api [None req-96c35e52-5464-43ed-b438-b393e0533d7f tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Task: {'id': task-5116766, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.790882} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1119.826600] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-96c35e52-5464-43ed-b438-b393e0533d7f tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore2] b353b5e9-500c-42d3-a87f-880a624febec/b353b5e9-500c-42d3-a87f-880a624febec.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1119.826939] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-96c35e52-5464-43ed-b438-b393e0533d7f tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: b353b5e9-500c-42d3-a87f-880a624febec] Extending root virtual disk to 1048576 {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1119.827361] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-24f2f0b8-e251-4dbe-a301-7db74159449c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.840896] env[65726]: DEBUG oslo_vmware.api [None req-96c35e52-5464-43ed-b438-b393e0533d7f tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Waiting for the task: (returnval){ [ 1119.840896] env[65726]: value = "task-5116771" [ 1119.840896] env[65726]: _type = "Task" [ 1119.840896] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1119.852042] env[65726]: DEBUG oslo_vmware.api [None req-96c35e52-5464-43ed-b438-b393e0533d7f tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Task: {'id': task-5116771, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1119.876922] env[65726]: DEBUG oslo_concurrency.lockutils [req-203dc1f1-8f0b-40aa-bb22-cc9ef191946e req-214e1ccf-94fd-4ef9-a4f9-917ec8080b05 service nova] Releasing lock "refresh_cache-51f5337d-2e23-4b7c-b590-76be6bd82411" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1119.923812] env[65726]: DEBUG oslo_concurrency.lockutils [req-63bfaf01-e600-427f-9f86-051495d4a694 req-4afacf0c-f46b-4a35-883a-8041eee268c8 service nova] Releasing lock "refresh_cache-b353b5e9-500c-42d3-a87f-880a624febec" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1120.010139] env[65726]: WARNING openstack [None req-aea91b5d-face-4f98-b031-6ac6f0d1a83b tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1120.010385] env[65726]: WARNING openstack [None req-aea91b5d-face-4f98-b031-6ac6f0d1a83b tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1120.078876] env[65726]: DEBUG oslo_vmware.api [None req-e0232c78-b0b0-436b-b3db-dd24c81ff983 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': task-5116769, 'name': PowerOnVM_Task} progress is 87%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1120.155207] env[65726]: WARNING openstack [None req-aea91b5d-face-4f98-b031-6ac6f0d1a83b tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1120.155604] env[65726]: WARNING openstack [None req-aea91b5d-face-4f98-b031-6ac6f0d1a83b tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1120.163601] env[65726]: WARNING openstack [None req-ad75268a-0b50-4f80-9f3a-c89856a157c4 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1120.164070] env[65726]: WARNING openstack [None req-ad75268a-0b50-4f80-9f3a-c89856a157c4 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1120.168915] env[65726]: DEBUG nova.network.neutron [None req-ad75268a-0b50-4f80-9f3a-c89856a157c4 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 378c1d18-4bb6-4245-80bf-370fb7af0575] Instance cache missing network info. {{(pid=65726) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 1120.248265] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f708868c-827d-43ac-b170-3acd0b38dbb7 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.258570] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-baf90821-f898-4911-a906-a7307d323625 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.264313] env[65726]: DEBUG nova.network.neutron [None req-aea91b5d-face-4f98-b031-6ac6f0d1a83b tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 0361ae8f-634d-4161-a594-7f93dbb1b7cd] Updating instance_info_cache with network_info: [{"id": "2bfd7147-87fd-4a19-b28b-eb95f51f4d00", "address": "fa:16:3e:d6:07:64", "network": {"id": "9dd55fb6-0dd2-484d-b94d-fbdccb958c79", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1005598846-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f93d852e2c904f42981cbfff882558e9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "74f30339-6421-4654-bddb-81d7f34db9d7", "external-id": "nsx-vlan-transportzone-899", "segmentation_id": 899, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2bfd7147-87", "ovs_interfaceid": "2bfd7147-87fd-4a19-b28b-eb95f51f4d00", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1120.296608] env[65726]: DEBUG oslo_concurrency.lockutils [None req-bb18e015-a533-4756-8784-6ff16584fde4 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Acquiring lock "interface-51f5337d-2e23-4b7c-b590-76be6bd82411-fb580bad-fd60-471d-95f2-066352417659" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1120.296891] env[65726]: DEBUG oslo_concurrency.lockutils [None req-bb18e015-a533-4756-8784-6ff16584fde4 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Lock "interface-51f5337d-2e23-4b7c-b590-76be6bd82411-fb580bad-fd60-471d-95f2-066352417659" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.001s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1120.300870] env[65726]: WARNING openstack [None req-ad75268a-0b50-4f80-9f3a-c89856a157c4 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1120.301507] env[65726]: WARNING openstack [None req-ad75268a-0b50-4f80-9f3a-c89856a157c4 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1120.309778] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9aba76e6-0bc7-4796-928d-e2194cd73196 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.328059] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b31ee8c4-3275-4777-b544-bc0c63a5aef7 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.332762] env[65726]: DEBUG oslo_vmware.api [None req-b72da630-f8d0-435a-9469-cadbe319a3e9 tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Task: {'id': task-5116770, 'name': ReconfigVM_Task, 'duration_secs': 0.186542} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1120.333075] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-b72da630-f8d0-435a-9469-cadbe319a3e9 tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] [instance: 4809fcf6-59d4-409c-bd8d-981dc85a686f] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1120.333783] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6f00011a-4c0a-405d-abe6-306a325e2c28 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.344456] env[65726]: DEBUG nova.compute.provider_tree [None req-04bd2146-ec5d-4e36-832f-8911f884caa3 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1120.355714] env[65726]: DEBUG oslo_vmware.api [None req-b72da630-f8d0-435a-9469-cadbe319a3e9 tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Waiting for the task: (returnval){ [ 1120.355714] env[65726]: value = "task-5116772" [ 1120.355714] env[65726]: _type = "Task" [ 1120.355714] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1120.364971] env[65726]: DEBUG oslo_vmware.api [None req-96c35e52-5464-43ed-b438-b393e0533d7f tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Task: {'id': task-5116771, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.08296} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1120.365767] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-96c35e52-5464-43ed-b438-b393e0533d7f tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: b353b5e9-500c-42d3-a87f-880a624febec] Extended root virtual disk {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1120.366879] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9fd47036-506b-4561-8872-6c66446f59d1 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.375858] env[65726]: INFO nova.compute.manager [None req-ba4dbb33-514e-4e75-8287-9941e479b176 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] [instance: 6fd70b4a-63e4-4258-9cff-f3b582500b9f] Resuming [ 1120.376744] env[65726]: DEBUG nova.objects.instance [None req-ba4dbb33-514e-4e75-8287-9941e479b176 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Lazy-loading 'flavor' on Instance uuid 6fd70b4a-63e4-4258-9cff-f3b582500b9f {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1120.380674] env[65726]: DEBUG oslo_vmware.api [None req-b72da630-f8d0-435a-9469-cadbe319a3e9 tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Task: {'id': task-5116772, 'name': PowerOnVM_Task} progress is 33%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1120.403292] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-96c35e52-5464-43ed-b438-b393e0533d7f tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: b353b5e9-500c-42d3-a87f-880a624febec] Reconfiguring VM instance instance-00000067 to attach disk [datastore2] b353b5e9-500c-42d3-a87f-880a624febec/b353b5e9-500c-42d3-a87f-880a624febec.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1120.404403] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-015c5e5d-9b82-4f96-bbb5-4b3deb998379 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.423012] env[65726]: WARNING openstack [None req-ad75268a-0b50-4f80-9f3a-c89856a157c4 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1120.423455] env[65726]: WARNING openstack [None req-ad75268a-0b50-4f80-9f3a-c89856a157c4 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1120.438593] env[65726]: DEBUG oslo_vmware.api [None req-96c35e52-5464-43ed-b438-b393e0533d7f tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Waiting for the task: (returnval){ [ 1120.438593] env[65726]: value = "task-5116773" [ 1120.438593] env[65726]: _type = "Task" [ 1120.438593] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1120.448406] env[65726]: DEBUG oslo_vmware.api [None req-96c35e52-5464-43ed-b438-b393e0533d7f tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Task: {'id': task-5116773, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1120.555865] env[65726]: DEBUG nova.network.neutron [None req-ad75268a-0b50-4f80-9f3a-c89856a157c4 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 378c1d18-4bb6-4245-80bf-370fb7af0575] Updating instance_info_cache with network_info: [{"id": "4765f6c2-d161-4e36-a6a8-3c8d39c47a40", "address": "fa:16:3e:3c:38:b2", "network": {"id": "0d0f561d-ac41-4ef5-9926-c707409799f7", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1134321217-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4eece77569624f90bf64e5c51974173f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2ce62383-8e84-4e26-955b-74c11392f4c9", "external-id": "nsx-vlan-transportzone-215", "segmentation_id": 215, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4765f6c2-d1", "ovs_interfaceid": "4765f6c2-d161-4e36-a6a8-3c8d39c47a40", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1120.570991] env[65726]: DEBUG oslo_vmware.api [None req-e0232c78-b0b0-436b-b3db-dd24c81ff983 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': task-5116769, 'name': PowerOnVM_Task, 'duration_secs': 0.975658} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1120.573221] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-e0232c78-b0b0-436b-b3db-dd24c81ff983 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: e7156a10-631a-4cc5-a544-88ce37763d02] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1120.573477] env[65726]: INFO nova.compute.manager [None req-e0232c78-b0b0-436b-b3db-dd24c81ff983 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: e7156a10-631a-4cc5-a544-88ce37763d02] Took 8.56 seconds to spawn the instance on the hypervisor. [ 1120.573674] env[65726]: DEBUG nova.compute.manager [None req-e0232c78-b0b0-436b-b3db-dd24c81ff983 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: e7156a10-631a-4cc5-a544-88ce37763d02] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1120.574952] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f72ade0-4585-4ede-b1d8-efa5c418a05e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.767709] env[65726]: DEBUG oslo_concurrency.lockutils [None req-aea91b5d-face-4f98-b031-6ac6f0d1a83b tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Releasing lock "refresh_cache-0361ae8f-634d-4161-a594-7f93dbb1b7cd" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1120.809928] env[65726]: DEBUG oslo_concurrency.lockutils [None req-bb18e015-a533-4756-8784-6ff16584fde4 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Acquiring lock "51f5337d-2e23-4b7c-b590-76be6bd82411" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1120.810467] env[65726]: DEBUG oslo_concurrency.lockutils [None req-bb18e015-a533-4756-8784-6ff16584fde4 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Acquired lock "51f5337d-2e23-4b7c-b590-76be6bd82411" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1120.811060] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aad2eee1-5c95-4f6c-b5fb-076ab94351ba {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.830300] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfa66d06-eea2-4eb5-9666-335c582d426a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.858519] env[65726]: DEBUG nova.scheduler.client.report [None req-04bd2146-ec5d-4e36-832f-8911f884caa3 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1120.862355] env[65726]: WARNING openstack [None req-bb18e015-a533-4756-8784-6ff16584fde4 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1120.862712] env[65726]: WARNING openstack [None req-bb18e015-a533-4756-8784-6ff16584fde4 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1120.872967] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-bb18e015-a533-4756-8784-6ff16584fde4 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: 51f5337d-2e23-4b7c-b590-76be6bd82411] Reconfiguring VM to detach interface {{(pid=65726) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1968}} [ 1120.876483] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3cc047ff-7480-4fb0-9e72-94e375117018 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.899740] env[65726]: DEBUG oslo_vmware.api [None req-b72da630-f8d0-435a-9469-cadbe319a3e9 tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Task: {'id': task-5116772, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1120.901212] env[65726]: DEBUG oslo_vmware.api [None req-bb18e015-a533-4756-8784-6ff16584fde4 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Waiting for the task: (returnval){ [ 1120.901212] env[65726]: value = "task-5116774" [ 1120.901212] env[65726]: _type = "Task" [ 1120.901212] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1120.910909] env[65726]: DEBUG oslo_vmware.api [None req-bb18e015-a533-4756-8784-6ff16584fde4 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': task-5116774, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1120.951070] env[65726]: DEBUG oslo_vmware.api [None req-96c35e52-5464-43ed-b438-b393e0533d7f tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Task: {'id': task-5116773, 'name': ReconfigVM_Task, 'duration_secs': 0.320788} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1120.951411] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-96c35e52-5464-43ed-b438-b393e0533d7f tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: b353b5e9-500c-42d3-a87f-880a624febec] Reconfigured VM instance instance-00000067 to attach disk [datastore2] b353b5e9-500c-42d3-a87f-880a624febec/b353b5e9-500c-42d3-a87f-880a624febec.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1120.952106] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-69929aa1-7b76-4618-b979-d12f71746f0d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.959650] env[65726]: DEBUG oslo_vmware.api [None req-96c35e52-5464-43ed-b438-b393e0533d7f tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Waiting for the task: (returnval){ [ 1120.959650] env[65726]: value = "task-5116775" [ 1120.959650] env[65726]: _type = "Task" [ 1120.959650] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1120.968454] env[65726]: DEBUG oslo_vmware.api [None req-96c35e52-5464-43ed-b438-b393e0533d7f tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Task: {'id': task-5116775, 'name': Rename_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1121.059266] env[65726]: DEBUG oslo_concurrency.lockutils [None req-ad75268a-0b50-4f80-9f3a-c89856a157c4 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Releasing lock "refresh_cache-378c1d18-4bb6-4245-80bf-370fb7af0575" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1121.059571] env[65726]: DEBUG nova.compute.manager [None req-ad75268a-0b50-4f80-9f3a-c89856a157c4 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 378c1d18-4bb6-4245-80bf-370fb7af0575] Instance network_info: |[{"id": "4765f6c2-d161-4e36-a6a8-3c8d39c47a40", "address": "fa:16:3e:3c:38:b2", "network": {"id": "0d0f561d-ac41-4ef5-9926-c707409799f7", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1134321217-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4eece77569624f90bf64e5c51974173f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2ce62383-8e84-4e26-955b-74c11392f4c9", "external-id": "nsx-vlan-transportzone-215", "segmentation_id": 215, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4765f6c2-d1", "ovs_interfaceid": "4765f6c2-d161-4e36-a6a8-3c8d39c47a40", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1121.060063] env[65726]: DEBUG oslo_concurrency.lockutils [req-f1d08b0b-7e4c-4185-a08b-f8af209576fa req-14a4037a-f5f0-4730-a8b9-2b34b962b453 service nova] Acquired lock "refresh_cache-378c1d18-4bb6-4245-80bf-370fb7af0575" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1121.060455] env[65726]: DEBUG nova.network.neutron [req-f1d08b0b-7e4c-4185-a08b-f8af209576fa req-14a4037a-f5f0-4730-a8b9-2b34b962b453 service nova] [instance: 378c1d18-4bb6-4245-80bf-370fb7af0575] Refreshing network info cache for port 4765f6c2-d161-4e36-a6a8-3c8d39c47a40 {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 1121.062047] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-ad75268a-0b50-4f80-9f3a-c89856a157c4 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 378c1d18-4bb6-4245-80bf-370fb7af0575] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:3c:38:b2', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '2ce62383-8e84-4e26-955b-74c11392f4c9', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4765f6c2-d161-4e36-a6a8-3c8d39c47a40', 'vif_model': 'vmxnet3'}] {{(pid=65726) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1121.071432] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-ad75268a-0b50-4f80-9f3a-c89856a157c4 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1121.072027] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 378c1d18-4bb6-4245-80bf-370fb7af0575] Creating VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1121.074050] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-774c0612-4040-4ecd-9cb7-00ac90d57274 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.100701] env[65726]: INFO nova.compute.manager [None req-e0232c78-b0b0-436b-b3db-dd24c81ff983 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: e7156a10-631a-4cc5-a544-88ce37763d02] Took 15.23 seconds to build instance. [ 1121.104220] env[65726]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1121.104220] env[65726]: value = "task-5116776" [ 1121.104220] env[65726]: _type = "Task" [ 1121.104220] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1121.115714] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116776, 'name': CreateVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1121.309022] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7684ec47-6d09-452d-b4c0-4082e521bfbe {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.335330] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84b6736d-9be7-4e11-917f-12eedf4337d7 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.345817] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-aea91b5d-face-4f98-b031-6ac6f0d1a83b tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 0361ae8f-634d-4161-a594-7f93dbb1b7cd] Updating instance '0361ae8f-634d-4161-a594-7f93dbb1b7cd' progress to 83 {{(pid=65726) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1121.377774] env[65726]: DEBUG oslo_concurrency.lockutils [None req-04bd2146-ec5d-4e36-832f-8911f884caa3 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.908s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1121.381734] env[65726]: DEBUG oslo_vmware.api [None req-b72da630-f8d0-435a-9469-cadbe319a3e9 tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Task: {'id': task-5116772, 'name': PowerOnVM_Task, 'duration_secs': 1.015187} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1121.383078] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-b72da630-f8d0-435a-9469-cadbe319a3e9 tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] [instance: 4809fcf6-59d4-409c-bd8d-981dc85a686f] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1121.385835] env[65726]: DEBUG nova.compute.manager [None req-b72da630-f8d0-435a-9469-cadbe319a3e9 tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] [instance: 4809fcf6-59d4-409c-bd8d-981dc85a686f] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1121.386728] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67a58510-1580-46ef-a31e-5d40b2fbf5cd {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.415038] env[65726]: DEBUG oslo_vmware.api [None req-bb18e015-a533-4756-8784-6ff16584fde4 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': task-5116774, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1121.437661] env[65726]: INFO nova.scheduler.client.report [None req-04bd2146-ec5d-4e36-832f-8911f884caa3 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Deleted allocations for instance ad4f00b5-3765-494f-8132-8b17b29d219a [ 1121.474998] env[65726]: DEBUG oslo_vmware.api [None req-96c35e52-5464-43ed-b438-b393e0533d7f tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Task: {'id': task-5116775, 'name': Rename_Task, 'duration_secs': 0.143037} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1121.474998] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-96c35e52-5464-43ed-b438-b393e0533d7f tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: b353b5e9-500c-42d3-a87f-880a624febec] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1121.475492] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-156c1227-51e4-415b-ab9c-0b749f78fd25 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.488248] env[65726]: DEBUG oslo_vmware.api [None req-96c35e52-5464-43ed-b438-b393e0533d7f tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Waiting for the task: (returnval){ [ 1121.488248] env[65726]: value = "task-5116777" [ 1121.488248] env[65726]: _type = "Task" [ 1121.488248] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1121.495662] env[65726]: DEBUG oslo_vmware.api [None req-96c35e52-5464-43ed-b438-b393e0533d7f tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Task: {'id': task-5116777, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1121.572580] env[65726]: WARNING openstack [req-f1d08b0b-7e4c-4185-a08b-f8af209576fa req-14a4037a-f5f0-4730-a8b9-2b34b962b453 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1121.573059] env[65726]: WARNING openstack [req-f1d08b0b-7e4c-4185-a08b-f8af209576fa req-14a4037a-f5f0-4730-a8b9-2b34b962b453 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1121.605508] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e0232c78-b0b0-436b-b3db-dd24c81ff983 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Lock "e7156a10-631a-4cc5-a544-88ce37763d02" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.743s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1121.618284] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116776, 'name': CreateVM_Task, 'duration_secs': 0.344369} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1121.618284] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 378c1d18-4bb6-4245-80bf-370fb7af0575] Created VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1121.618284] env[65726]: WARNING openstack [None req-ad75268a-0b50-4f80-9f3a-c89856a157c4 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1121.618284] env[65726]: WARNING openstack [None req-ad75268a-0b50-4f80-9f3a-c89856a157c4 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1121.622984] env[65726]: DEBUG oslo_concurrency.lockutils [None req-ad75268a-0b50-4f80-9f3a-c89856a157c4 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1121.623170] env[65726]: DEBUG oslo_concurrency.lockutils [None req-ad75268a-0b50-4f80-9f3a-c89856a157c4 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1121.623498] env[65726]: DEBUG oslo_concurrency.lockutils [None req-ad75268a-0b50-4f80-9f3a-c89856a157c4 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1121.624058] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a4294c48-9e37-451b-8d0a-d37b7a53a85d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.630083] env[65726]: DEBUG oslo_vmware.api [None req-ad75268a-0b50-4f80-9f3a-c89856a157c4 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Waiting for the task: (returnval){ [ 1121.630083] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]525e7311-5a9c-3ce1-1e60-e3161ed2e9f4" [ 1121.630083] env[65726]: _type = "Task" [ 1121.630083] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1121.638756] env[65726]: DEBUG oslo_vmware.api [None req-ad75268a-0b50-4f80-9f3a-c89856a157c4 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]525e7311-5a9c-3ce1-1e60-e3161ed2e9f4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1121.814822] env[65726]: WARNING openstack [req-f1d08b0b-7e4c-4185-a08b-f8af209576fa req-14a4037a-f5f0-4730-a8b9-2b34b962b453 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1121.815549] env[65726]: WARNING openstack [req-f1d08b0b-7e4c-4185-a08b-f8af209576fa req-14a4037a-f5f0-4730-a8b9-2b34b962b453 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1121.855132] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-aea91b5d-face-4f98-b031-6ac6f0d1a83b tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 0361ae8f-634d-4161-a594-7f93dbb1b7cd] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1121.855468] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0a7e071a-28dd-403a-a3eb-7805ff69bf7d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.864396] env[65726]: DEBUG oslo_vmware.api [None req-aea91b5d-face-4f98-b031-6ac6f0d1a83b tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Waiting for the task: (returnval){ [ 1121.864396] env[65726]: value = "task-5116778" [ 1121.864396] env[65726]: _type = "Task" [ 1121.864396] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1121.874560] env[65726]: DEBUG oslo_vmware.api [None req-aea91b5d-face-4f98-b031-6ac6f0d1a83b tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': task-5116778, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1121.904307] env[65726]: DEBUG oslo_concurrency.lockutils [None req-ba4dbb33-514e-4e75-8287-9941e479b176 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Acquiring lock "refresh_cache-6fd70b4a-63e4-4258-9cff-f3b582500b9f" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1121.904590] env[65726]: DEBUG oslo_concurrency.lockutils [None req-ba4dbb33-514e-4e75-8287-9941e479b176 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Acquired lock "refresh_cache-6fd70b4a-63e4-4258-9cff-f3b582500b9f" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1121.904868] env[65726]: DEBUG nova.network.neutron [None req-ba4dbb33-514e-4e75-8287-9941e479b176 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] [instance: 6fd70b4a-63e4-4258-9cff-f3b582500b9f] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1121.927063] env[65726]: DEBUG oslo_vmware.api [None req-bb18e015-a533-4756-8784-6ff16584fde4 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': task-5116774, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1121.948134] env[65726]: DEBUG oslo_concurrency.lockutils [None req-04bd2146-ec5d-4e36-832f-8911f884caa3 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Lock "ad4f00b5-3765-494f-8132-8b17b29d219a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.271s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1122.000405] env[65726]: DEBUG oslo_vmware.api [None req-96c35e52-5464-43ed-b438-b393e0533d7f tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Task: {'id': task-5116777, 'name': PowerOnVM_Task, 'duration_secs': 0.498665} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1122.000757] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-96c35e52-5464-43ed-b438-b393e0533d7f tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: b353b5e9-500c-42d3-a87f-880a624febec] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1122.000988] env[65726]: INFO nova.compute.manager [None req-96c35e52-5464-43ed-b438-b393e0533d7f tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: b353b5e9-500c-42d3-a87f-880a624febec] Took 7.65 seconds to spawn the instance on the hypervisor. [ 1122.001190] env[65726]: DEBUG nova.compute.manager [None req-96c35e52-5464-43ed-b438-b393e0533d7f tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: b353b5e9-500c-42d3-a87f-880a624febec] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1122.002023] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2cfdaded-6216-4aa5-9482-e9df3dc4c97a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.053051] env[65726]: WARNING openstack [req-f1d08b0b-7e4c-4185-a08b-f8af209576fa req-14a4037a-f5f0-4730-a8b9-2b34b962b453 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1122.053521] env[65726]: WARNING openstack [req-f1d08b0b-7e4c-4185-a08b-f8af209576fa req-14a4037a-f5f0-4730-a8b9-2b34b962b453 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1122.140421] env[65726]: DEBUG oslo_vmware.api [None req-ad75268a-0b50-4f80-9f3a-c89856a157c4 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]525e7311-5a9c-3ce1-1e60-e3161ed2e9f4, 'name': SearchDatastore_Task, 'duration_secs': 0.013712} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1122.140949] env[65726]: DEBUG oslo_concurrency.lockutils [None req-ad75268a-0b50-4f80-9f3a-c89856a157c4 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1122.140949] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-ad75268a-0b50-4f80-9f3a-c89856a157c4 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 378c1d18-4bb6-4245-80bf-370fb7af0575] Processing image b52362a3-ee8a-4cbf-b06f-513b0cc8f95c {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1122.141880] env[65726]: DEBUG oslo_concurrency.lockutils [None req-ad75268a-0b50-4f80-9f3a-c89856a157c4 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1122.141880] env[65726]: DEBUG oslo_concurrency.lockutils [None req-ad75268a-0b50-4f80-9f3a-c89856a157c4 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1122.141880] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-ad75268a-0b50-4f80-9f3a-c89856a157c4 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1122.141880] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6ea481f7-a527-407c-b0ce-8125bc6289dd {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.154652] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-ad75268a-0b50-4f80-9f3a-c89856a157c4 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1122.154777] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-ad75268a-0b50-4f80-9f3a-c89856a157c4 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65726) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1122.155659] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bb9a7d2d-4827-4b7b-ad4d-9116de5e8814 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.165386] env[65726]: DEBUG oslo_vmware.api [None req-ad75268a-0b50-4f80-9f3a-c89856a157c4 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Waiting for the task: (returnval){ [ 1122.165386] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]522c203c-7035-5668-ae42-1f696650e87b" [ 1122.165386] env[65726]: _type = "Task" [ 1122.165386] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1122.176898] env[65726]: DEBUG oslo_vmware.api [None req-ad75268a-0b50-4f80-9f3a-c89856a157c4 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]522c203c-7035-5668-ae42-1f696650e87b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1122.186559] env[65726]: DEBUG nova.network.neutron [req-f1d08b0b-7e4c-4185-a08b-f8af209576fa req-14a4037a-f5f0-4730-a8b9-2b34b962b453 service nova] [instance: 378c1d18-4bb6-4245-80bf-370fb7af0575] Updated VIF entry in instance network info cache for port 4765f6c2-d161-4e36-a6a8-3c8d39c47a40. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 1122.187256] env[65726]: DEBUG nova.network.neutron [req-f1d08b0b-7e4c-4185-a08b-f8af209576fa req-14a4037a-f5f0-4730-a8b9-2b34b962b453 service nova] [instance: 378c1d18-4bb6-4245-80bf-370fb7af0575] Updating instance_info_cache with network_info: [{"id": "4765f6c2-d161-4e36-a6a8-3c8d39c47a40", "address": "fa:16:3e:3c:38:b2", "network": {"id": "0d0f561d-ac41-4ef5-9926-c707409799f7", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1134321217-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4eece77569624f90bf64e5c51974173f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2ce62383-8e84-4e26-955b-74c11392f4c9", "external-id": "nsx-vlan-transportzone-215", "segmentation_id": 215, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4765f6c2-d1", "ovs_interfaceid": "4765f6c2-d161-4e36-a6a8-3c8d39c47a40", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1122.376098] env[65726]: DEBUG oslo_vmware.api [None req-aea91b5d-face-4f98-b031-6ac6f0d1a83b tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': task-5116778, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1122.413711] env[65726]: WARNING openstack [None req-ba4dbb33-514e-4e75-8287-9941e479b176 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1122.413711] env[65726]: WARNING openstack [None req-ba4dbb33-514e-4e75-8287-9941e479b176 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1122.437022] env[65726]: DEBUG oslo_vmware.api [None req-bb18e015-a533-4756-8784-6ff16584fde4 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': task-5116774, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1122.483276] env[65726]: INFO nova.compute.manager [None req-116f601c-d8e3-4bed-a9f0-abae1fb1e4c5 tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] [instance: 4809fcf6-59d4-409c-bd8d-981dc85a686f] Unrescuing [ 1122.483609] env[65726]: DEBUG oslo_concurrency.lockutils [None req-116f601c-d8e3-4bed-a9f0-abae1fb1e4c5 tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Acquiring lock "refresh_cache-4809fcf6-59d4-409c-bd8d-981dc85a686f" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1122.483609] env[65726]: DEBUG oslo_concurrency.lockutils [None req-116f601c-d8e3-4bed-a9f0-abae1fb1e4c5 tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Acquired lock "refresh_cache-4809fcf6-59d4-409c-bd8d-981dc85a686f" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1122.483779] env[65726]: DEBUG nova.network.neutron [None req-116f601c-d8e3-4bed-a9f0-abae1fb1e4c5 tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] [instance: 4809fcf6-59d4-409c-bd8d-981dc85a686f] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1122.529102] env[65726]: INFO nova.compute.manager [None req-96c35e52-5464-43ed-b438-b393e0533d7f tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: b353b5e9-500c-42d3-a87f-880a624febec] Took 13.48 seconds to build instance. [ 1122.600420] env[65726]: WARNING openstack [None req-ba4dbb33-514e-4e75-8287-9941e479b176 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1122.600850] env[65726]: WARNING openstack [None req-ba4dbb33-514e-4e75-8287-9941e479b176 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1122.678575] env[65726]: DEBUG oslo_vmware.api [None req-ad75268a-0b50-4f80-9f3a-c89856a157c4 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]522c203c-7035-5668-ae42-1f696650e87b, 'name': SearchDatastore_Task, 'duration_secs': 0.01304} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1122.679328] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f1319014-7566-4a14-b70d-50ad9f504b94 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.686970] env[65726]: DEBUG oslo_vmware.api [None req-ad75268a-0b50-4f80-9f3a-c89856a157c4 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Waiting for the task: (returnval){ [ 1122.686970] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]525ffac8-70f4-40d4-cb99-2a054254fa87" [ 1122.686970] env[65726]: _type = "Task" [ 1122.686970] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1122.693017] env[65726]: DEBUG oslo_concurrency.lockutils [req-f1d08b0b-7e4c-4185-a08b-f8af209576fa req-14a4037a-f5f0-4730-a8b9-2b34b962b453 service nova] Releasing lock "refresh_cache-378c1d18-4bb6-4245-80bf-370fb7af0575" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1122.703185] env[65726]: DEBUG oslo_vmware.api [None req-ad75268a-0b50-4f80-9f3a-c89856a157c4 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]525ffac8-70f4-40d4-cb99-2a054254fa87, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1122.767646] env[65726]: WARNING openstack [None req-ba4dbb33-514e-4e75-8287-9941e479b176 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1122.770139] env[65726]: WARNING openstack [None req-ba4dbb33-514e-4e75-8287-9941e479b176 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1122.803511] env[65726]: DEBUG nova.compute.manager [None req-a2dc650f-ed4c-4edd-a0f8-7cba2e214975 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: e7156a10-631a-4cc5-a544-88ce37763d02] Stashing vm_state: active {{(pid=65726) _prep_resize /opt/stack/nova/nova/compute/manager.py:6193}} [ 1122.878129] env[65726]: DEBUG oslo_vmware.api [None req-aea91b5d-face-4f98-b031-6ac6f0d1a83b tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': task-5116778, 'name': PowerOnVM_Task, 'duration_secs': 0.720024} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1122.878284] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-aea91b5d-face-4f98-b031-6ac6f0d1a83b tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 0361ae8f-634d-4161-a594-7f93dbb1b7cd] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1122.878991] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-aea91b5d-face-4f98-b031-6ac6f0d1a83b tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 0361ae8f-634d-4161-a594-7f93dbb1b7cd] Updating instance '0361ae8f-634d-4161-a594-7f93dbb1b7cd' progress to 100 {{(pid=65726) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1122.933680] env[65726]: DEBUG oslo_vmware.api [None req-bb18e015-a533-4756-8784-6ff16584fde4 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': task-5116774, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1122.992364] env[65726]: WARNING openstack [None req-116f601c-d8e3-4bed-a9f0-abae1fb1e4c5 tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1122.992609] env[65726]: WARNING openstack [None req-116f601c-d8e3-4bed-a9f0-abae1fb1e4c5 tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1123.031567] env[65726]: DEBUG oslo_concurrency.lockutils [None req-96c35e52-5464-43ed-b438-b393e0533d7f tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Lock "b353b5e9-500c-42d3-a87f-880a624febec" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.998s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1123.165703] env[65726]: DEBUG nova.network.neutron [None req-ba4dbb33-514e-4e75-8287-9941e479b176 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] [instance: 6fd70b4a-63e4-4258-9cff-f3b582500b9f] Updating instance_info_cache with network_info: [{"id": "153b4c9d-d01d-4254-9aa6-040705be347a", "address": "fa:16:3e:17:18:d9", "network": {"id": "93e38d7e-b4d7-4d4a-8393-64ab2f6769de", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1445539922-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "06e221dc693640929cac7bc5af45948f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "399f3826-705c-45f7-9fe0-3a08a945151a", "external-id": "nsx-vlan-transportzone-936", "segmentation_id": 936, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap153b4c9d-d0", "ovs_interfaceid": "153b4c9d-d01d-4254-9aa6-040705be347a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1123.198826] env[65726]: DEBUG nova.compute.manager [req-d8a65e4e-ada6-42c3-a183-a7a898237f6d req-97f0bb60-c103-44b0-aa0f-1abf6c529a5b service nova] [instance: b353b5e9-500c-42d3-a87f-880a624febec] Received event network-changed-08b7016b-dc44-4206-a00c-5da943b82a38 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1123.200039] env[65726]: DEBUG nova.compute.manager [req-d8a65e4e-ada6-42c3-a183-a7a898237f6d req-97f0bb60-c103-44b0-aa0f-1abf6c529a5b service nova] [instance: b353b5e9-500c-42d3-a87f-880a624febec] Refreshing instance network info cache due to event network-changed-08b7016b-dc44-4206-a00c-5da943b82a38. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 1123.200039] env[65726]: DEBUG oslo_concurrency.lockutils [req-d8a65e4e-ada6-42c3-a183-a7a898237f6d req-97f0bb60-c103-44b0-aa0f-1abf6c529a5b service nova] Acquiring lock "refresh_cache-b353b5e9-500c-42d3-a87f-880a624febec" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1123.200039] env[65726]: DEBUG oslo_concurrency.lockutils [req-d8a65e4e-ada6-42c3-a183-a7a898237f6d req-97f0bb60-c103-44b0-aa0f-1abf6c529a5b service nova] Acquired lock "refresh_cache-b353b5e9-500c-42d3-a87f-880a624febec" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1123.200039] env[65726]: DEBUG nova.network.neutron [req-d8a65e4e-ada6-42c3-a183-a7a898237f6d req-97f0bb60-c103-44b0-aa0f-1abf6c529a5b service nova] [instance: b353b5e9-500c-42d3-a87f-880a624febec] Refreshing network info cache for port 08b7016b-dc44-4206-a00c-5da943b82a38 {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 1123.206033] env[65726]: DEBUG oslo_vmware.api [None req-ad75268a-0b50-4f80-9f3a-c89856a157c4 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]525ffac8-70f4-40d4-cb99-2a054254fa87, 'name': SearchDatastore_Task, 'duration_secs': 0.020721} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1123.206556] env[65726]: DEBUG oslo_concurrency.lockutils [None req-ad75268a-0b50-4f80-9f3a-c89856a157c4 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1123.206895] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-ad75268a-0b50-4f80-9f3a-c89856a157c4 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] 378c1d18-4bb6-4245-80bf-370fb7af0575/378c1d18-4bb6-4245-80bf-370fb7af0575.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1123.207221] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4e98c456-f927-4550-a2e0-a486e43bc57a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.216944] env[65726]: DEBUG oslo_vmware.api [None req-ad75268a-0b50-4f80-9f3a-c89856a157c4 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Waiting for the task: (returnval){ [ 1123.216944] env[65726]: value = "task-5116779" [ 1123.216944] env[65726]: _type = "Task" [ 1123.216944] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1123.233157] env[65726]: DEBUG oslo_vmware.api [None req-ad75268a-0b50-4f80-9f3a-c89856a157c4 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5116779, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1123.262712] env[65726]: WARNING openstack [None req-116f601c-d8e3-4bed-a9f0-abae1fb1e4c5 tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1123.263254] env[65726]: WARNING openstack [None req-116f601c-d8e3-4bed-a9f0-abae1fb1e4c5 tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1123.333120] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a2dc650f-ed4c-4edd-a0f8-7cba2e214975 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1123.333546] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a2dc650f-ed4c-4edd-a0f8-7cba2e214975 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 0.001s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1123.402501] env[65726]: WARNING openstack [None req-116f601c-d8e3-4bed-a9f0-abae1fb1e4c5 tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1123.403031] env[65726]: WARNING openstack [None req-116f601c-d8e3-4bed-a9f0-abae1fb1e4c5 tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1123.433768] env[65726]: DEBUG oslo_vmware.api [None req-bb18e015-a533-4756-8784-6ff16584fde4 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': task-5116774, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1123.558654] env[65726]: DEBUG nova.network.neutron [None req-116f601c-d8e3-4bed-a9f0-abae1fb1e4c5 tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] [instance: 4809fcf6-59d4-409c-bd8d-981dc85a686f] Updating instance_info_cache with network_info: [{"id": "13050a31-7165-4b22-83cc-d8ce9a41eadf", "address": "fa:16:3e:02:ed:66", "network": {"id": "53facc8b-af7e-44de-8c6f-6bea798b3f0b", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-2084216438-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "eb811e38506b4297bf01bc5f2e098370", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "19671de9-8b5b-4710-adc3-7419f3c0f171", "external-id": "nsx-vlan-transportzone-421", "segmentation_id": 421, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap13050a31-71", "ovs_interfaceid": "13050a31-7165-4b22-83cc-d8ce9a41eadf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1123.669716] env[65726]: DEBUG oslo_concurrency.lockutils [None req-ba4dbb33-514e-4e75-8287-9941e479b176 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Releasing lock "refresh_cache-6fd70b4a-63e4-4258-9cff-f3b582500b9f" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1123.670893] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-165f552b-ea0a-4e96-956a-1ca56936f1b7 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.680374] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-ba4dbb33-514e-4e75-8287-9941e479b176 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] [instance: 6fd70b4a-63e4-4258-9cff-f3b582500b9f] Resuming the VM {{(pid=65726) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1183}} [ 1123.680682] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ce4620bc-f5ea-4535-bc17-853a22686ffc {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.691670] env[65726]: DEBUG oslo_vmware.api [None req-ba4dbb33-514e-4e75-8287-9941e479b176 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Waiting for the task: (returnval){ [ 1123.691670] env[65726]: value = "task-5116783" [ 1123.691670] env[65726]: _type = "Task" [ 1123.691670] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1123.704048] env[65726]: DEBUG oslo_vmware.api [None req-ba4dbb33-514e-4e75-8287-9941e479b176 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Task: {'id': task-5116783, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1123.710041] env[65726]: WARNING openstack [req-d8a65e4e-ada6-42c3-a183-a7a898237f6d req-97f0bb60-c103-44b0-aa0f-1abf6c529a5b service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1123.710041] env[65726]: WARNING openstack [req-d8a65e4e-ada6-42c3-a183-a7a898237f6d req-97f0bb60-c103-44b0-aa0f-1abf6c529a5b service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1123.732090] env[65726]: DEBUG oslo_vmware.api [None req-ad75268a-0b50-4f80-9f3a-c89856a157c4 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5116779, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1123.839704] env[65726]: INFO nova.compute.claims [None req-a2dc650f-ed4c-4edd-a0f8-7cba2e214975 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: e7156a10-631a-4cc5-a544-88ce37763d02] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1123.863322] env[65726]: WARNING openstack [req-d8a65e4e-ada6-42c3-a183-a7a898237f6d req-97f0bb60-c103-44b0-aa0f-1abf6c529a5b service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1123.864120] env[65726]: WARNING openstack [req-d8a65e4e-ada6-42c3-a183-a7a898237f6d req-97f0bb60-c103-44b0-aa0f-1abf6c529a5b service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1123.935854] env[65726]: DEBUG oslo_vmware.api [None req-bb18e015-a533-4756-8784-6ff16584fde4 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': task-5116774, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1123.941267] env[65726]: WARNING openstack [req-d8a65e4e-ada6-42c3-a183-a7a898237f6d req-97f0bb60-c103-44b0-aa0f-1abf6c529a5b service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1123.941855] env[65726]: WARNING openstack [req-d8a65e4e-ada6-42c3-a183-a7a898237f6d req-97f0bb60-c103-44b0-aa0f-1abf6c529a5b service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1124.043487] env[65726]: DEBUG nova.network.neutron [req-d8a65e4e-ada6-42c3-a183-a7a898237f6d req-97f0bb60-c103-44b0-aa0f-1abf6c529a5b service nova] [instance: b353b5e9-500c-42d3-a87f-880a624febec] Updated VIF entry in instance network info cache for port 08b7016b-dc44-4206-a00c-5da943b82a38. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 1124.043709] env[65726]: DEBUG nova.network.neutron [req-d8a65e4e-ada6-42c3-a183-a7a898237f6d req-97f0bb60-c103-44b0-aa0f-1abf6c529a5b service nova] [instance: b353b5e9-500c-42d3-a87f-880a624febec] Updating instance_info_cache with network_info: [{"id": "08b7016b-dc44-4206-a00c-5da943b82a38", "address": "fa:16:3e:c5:e4:14", "network": {"id": "eb3278b7-c190-441b-84d2-305f9b896382", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1754571022-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.144", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "96149159e18e44f9bf3453e67681f224", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f77ff7a1-209c-4f3f-b2a0-fd817741e739", "external-id": "nsx-vlan-transportzone-935", "segmentation_id": 935, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap08b7016b-dc", "ovs_interfaceid": "08b7016b-dc44-4206-a00c-5da943b82a38", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1124.064354] env[65726]: DEBUG oslo_concurrency.lockutils [None req-116f601c-d8e3-4bed-a9f0-abae1fb1e4c5 tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Releasing lock "refresh_cache-4809fcf6-59d4-409c-bd8d-981dc85a686f" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1124.064354] env[65726]: DEBUG nova.objects.instance [None req-116f601c-d8e3-4bed-a9f0-abae1fb1e4c5 tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Lazy-loading 'flavor' on Instance uuid 4809fcf6-59d4-409c-bd8d-981dc85a686f {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1124.202418] env[65726]: DEBUG oslo_vmware.api [None req-ba4dbb33-514e-4e75-8287-9941e479b176 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Task: {'id': task-5116783, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1124.228836] env[65726]: DEBUG oslo_vmware.api [None req-ad75268a-0b50-4f80-9f3a-c89856a157c4 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5116779, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.678612} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1124.229167] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-ad75268a-0b50-4f80-9f3a-c89856a157c4 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] 378c1d18-4bb6-4245-80bf-370fb7af0575/378c1d18-4bb6-4245-80bf-370fb7af0575.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1124.229389] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-ad75268a-0b50-4f80-9f3a-c89856a157c4 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 378c1d18-4bb6-4245-80bf-370fb7af0575] Extending root virtual disk to 1048576 {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1124.229720] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3198912c-8063-4ab2-8d91-ec0a67043959 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.238023] env[65726]: DEBUG oslo_vmware.api [None req-ad75268a-0b50-4f80-9f3a-c89856a157c4 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Waiting for the task: (returnval){ [ 1124.238023] env[65726]: value = "task-5116784" [ 1124.238023] env[65726]: _type = "Task" [ 1124.238023] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1124.247991] env[65726]: DEBUG oslo_vmware.api [None req-ad75268a-0b50-4f80-9f3a-c89856a157c4 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5116784, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1124.349575] env[65726]: INFO nova.compute.resource_tracker [None req-a2dc650f-ed4c-4edd-a0f8-7cba2e214975 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: e7156a10-631a-4cc5-a544-88ce37763d02] Updating resource usage from migration 59ac4637-e900-45a3-bb6b-29b88535da20 [ 1124.445133] env[65726]: DEBUG oslo_vmware.api [None req-bb18e015-a533-4756-8784-6ff16584fde4 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': task-5116774, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1124.547677] env[65726]: DEBUG oslo_concurrency.lockutils [req-d8a65e4e-ada6-42c3-a183-a7a898237f6d req-97f0bb60-c103-44b0-aa0f-1abf6c529a5b service nova] Releasing lock "refresh_cache-b353b5e9-500c-42d3-a87f-880a624febec" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1124.569754] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e18b726c-8a51-4436-91da-a83d9fc78f56 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.574565] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74b82c07-b59d-48ec-9ee4-b425fa2286b6 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.598010] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-116f601c-d8e3-4bed-a9f0-abae1fb1e4c5 tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] [instance: 4809fcf6-59d4-409c-bd8d-981dc85a686f] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1124.601031] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c43c6673-9b65-46de-8561-6f8951d4ff51 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.604209] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9778e2d2-a0b3-4c45-909c-f89adfd3e963 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.639905] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc755163-6c49-483b-930a-184a94ccd9d0 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.644329] env[65726]: DEBUG oslo_vmware.api [None req-116f601c-d8e3-4bed-a9f0-abae1fb1e4c5 tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Waiting for the task: (returnval){ [ 1124.644329] env[65726]: value = "task-5116785" [ 1124.644329] env[65726]: _type = "Task" [ 1124.644329] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1124.650504] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc424aff-47dd-4a3a-900f-264965d69fcc {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.659228] env[65726]: DEBUG oslo_vmware.api [None req-116f601c-d8e3-4bed-a9f0-abae1fb1e4c5 tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Task: {'id': task-5116785, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1124.670500] env[65726]: DEBUG nova.compute.provider_tree [None req-a2dc650f-ed4c-4edd-a0f8-7cba2e214975 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1124.713241] env[65726]: DEBUG oslo_vmware.api [None req-ba4dbb33-514e-4e75-8287-9941e479b176 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Task: {'id': task-5116783, 'name': PowerOnVM_Task} progress is 93%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1124.751293] env[65726]: DEBUG oslo_vmware.api [None req-ad75268a-0b50-4f80-9f3a-c89856a157c4 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5116784, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.072159} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1124.751635] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-ad75268a-0b50-4f80-9f3a-c89856a157c4 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 378c1d18-4bb6-4245-80bf-370fb7af0575] Extended root virtual disk {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1124.752348] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9cec773a-acd6-4f82-bcaf-193db864320e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.782952] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-ad75268a-0b50-4f80-9f3a-c89856a157c4 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 378c1d18-4bb6-4245-80bf-370fb7af0575] Reconfiguring VM instance instance-00000068 to attach disk [datastore1] 378c1d18-4bb6-4245-80bf-370fb7af0575/378c1d18-4bb6-4245-80bf-370fb7af0575.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1124.783738] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2b7e497d-d962-4988-9227-69a913f53d82 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.810752] env[65726]: DEBUG oslo_vmware.api [None req-ad75268a-0b50-4f80-9f3a-c89856a157c4 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Waiting for the task: (returnval){ [ 1124.810752] env[65726]: value = "task-5116786" [ 1124.810752] env[65726]: _type = "Task" [ 1124.810752] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1124.825171] env[65726]: DEBUG oslo_vmware.api [None req-ad75268a-0b50-4f80-9f3a-c89856a157c4 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5116786, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1124.937691] env[65726]: DEBUG oslo_vmware.api [None req-bb18e015-a533-4756-8784-6ff16584fde4 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': task-5116774, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1125.156774] env[65726]: DEBUG oslo_vmware.api [None req-116f601c-d8e3-4bed-a9f0-abae1fb1e4c5 tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Task: {'id': task-5116785, 'name': PowerOffVM_Task, 'duration_secs': 0.349738} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1125.156774] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-116f601c-d8e3-4bed-a9f0-abae1fb1e4c5 tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] [instance: 4809fcf6-59d4-409c-bd8d-981dc85a686f] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1125.162284] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-116f601c-d8e3-4bed-a9f0-abae1fb1e4c5 tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] [instance: 4809fcf6-59d4-409c-bd8d-981dc85a686f] Reconfiguring VM instance instance-00000065 to detach disk 2001 {{(pid=65726) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1125.162684] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-118921d5-a6fb-48dc-a0eb-71d3cbdf385d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.177069] env[65726]: DEBUG nova.scheduler.client.report [None req-a2dc650f-ed4c-4edd-a0f8-7cba2e214975 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1125.187542] env[65726]: DEBUG oslo_vmware.api [None req-116f601c-d8e3-4bed-a9f0-abae1fb1e4c5 tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Waiting for the task: (returnval){ [ 1125.187542] env[65726]: value = "task-5116787" [ 1125.187542] env[65726]: _type = "Task" [ 1125.187542] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1125.199745] env[65726]: DEBUG oslo_vmware.api [None req-116f601c-d8e3-4bed-a9f0-abae1fb1e4c5 tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Task: {'id': task-5116787, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1125.206036] env[65726]: DEBUG oslo_vmware.api [None req-ba4dbb33-514e-4e75-8287-9941e479b176 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Task: {'id': task-5116783, 'name': PowerOnVM_Task, 'duration_secs': 1.153997} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1125.206227] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-ba4dbb33-514e-4e75-8287-9941e479b176 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] [instance: 6fd70b4a-63e4-4258-9cff-f3b582500b9f] Resumed the VM {{(pid=65726) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1188}} [ 1125.206296] env[65726]: DEBUG nova.compute.manager [None req-ba4dbb33-514e-4e75-8287-9941e479b176 tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] [instance: 6fd70b4a-63e4-4258-9cff-f3b582500b9f] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1125.207244] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfd9c41f-0635-40f5-9d1e-056cd1b3a2c0 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.215050] env[65726]: DEBUG oslo_concurrency.lockutils [None req-96ccefc5-467e-42ef-be69-e76eb66b409b tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Acquiring lock "0361ae8f-634d-4161-a594-7f93dbb1b7cd" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1125.215428] env[65726]: DEBUG oslo_concurrency.lockutils [None req-96ccefc5-467e-42ef-be69-e76eb66b409b tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Lock "0361ae8f-634d-4161-a594-7f93dbb1b7cd" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.001s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1125.215699] env[65726]: DEBUG nova.compute.manager [None req-96ccefc5-467e-42ef-be69-e76eb66b409b tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 0361ae8f-634d-4161-a594-7f93dbb1b7cd] Going to confirm migration 5 {{(pid=65726) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:5307}} [ 1125.322222] env[65726]: DEBUG oslo_vmware.api [None req-ad75268a-0b50-4f80-9f3a-c89856a157c4 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5116786, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1125.440839] env[65726]: DEBUG oslo_vmware.api [None req-bb18e015-a533-4756-8784-6ff16584fde4 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': task-5116774, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1125.523020] env[65726]: DEBUG oslo_concurrency.lockutils [None req-72f2a102-8ae8-4f21-a769-e63022914c54 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Acquiring lock "3bc25953-21b1-4729-af09-e7211fd8b2c6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1125.523419] env[65726]: DEBUG oslo_concurrency.lockutils [None req-72f2a102-8ae8-4f21-a769-e63022914c54 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Lock "3bc25953-21b1-4729-af09-e7211fd8b2c6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1125.683191] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a2dc650f-ed4c-4edd-a0f8-7cba2e214975 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.349s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1125.683427] env[65726]: INFO nova.compute.manager [None req-a2dc650f-ed4c-4edd-a0f8-7cba2e214975 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: e7156a10-631a-4cc5-a544-88ce37763d02] Migrating [ 1125.707964] env[65726]: DEBUG oslo_vmware.api [None req-116f601c-d8e3-4bed-a9f0-abae1fb1e4c5 tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Task: {'id': task-5116787, 'name': ReconfigVM_Task, 'duration_secs': 0.45919} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1125.708231] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-116f601c-d8e3-4bed-a9f0-abae1fb1e4c5 tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] [instance: 4809fcf6-59d4-409c-bd8d-981dc85a686f] Reconfigured VM instance instance-00000065 to detach disk 2001 {{(pid=65726) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1125.708407] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-116f601c-d8e3-4bed-a9f0-abae1fb1e4c5 tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] [instance: 4809fcf6-59d4-409c-bd8d-981dc85a686f] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1125.708672] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2f794158-c8e2-4efd-bd2d-fae0aa6b3c2a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.721446] env[65726]: DEBUG oslo_vmware.api [None req-116f601c-d8e3-4bed-a9f0-abae1fb1e4c5 tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Waiting for the task: (returnval){ [ 1125.721446] env[65726]: value = "task-5116788" [ 1125.721446] env[65726]: _type = "Task" [ 1125.721446] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1125.722378] env[65726]: WARNING openstack [None req-96ccefc5-467e-42ef-be69-e76eb66b409b tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1125.722747] env[65726]: WARNING openstack [None req-96ccefc5-467e-42ef-be69-e76eb66b409b tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1125.744088] env[65726]: DEBUG oslo_vmware.api [None req-116f601c-d8e3-4bed-a9f0-abae1fb1e4c5 tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Task: {'id': task-5116788, 'name': PowerOnVM_Task} progress is 33%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1125.769063] env[65726]: WARNING openstack [None req-96ccefc5-467e-42ef-be69-e76eb66b409b tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1125.769616] env[65726]: WARNING openstack [None req-96ccefc5-467e-42ef-be69-e76eb66b409b tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1125.776022] env[65726]: DEBUG oslo_concurrency.lockutils [None req-96ccefc5-467e-42ef-be69-e76eb66b409b tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Acquiring lock "refresh_cache-0361ae8f-634d-4161-a594-7f93dbb1b7cd" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1125.776249] env[65726]: DEBUG oslo_concurrency.lockutils [None req-96ccefc5-467e-42ef-be69-e76eb66b409b tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Acquired lock "refresh_cache-0361ae8f-634d-4161-a594-7f93dbb1b7cd" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1125.776518] env[65726]: DEBUG nova.network.neutron [None req-96ccefc5-467e-42ef-be69-e76eb66b409b tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 0361ae8f-634d-4161-a594-7f93dbb1b7cd] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1125.776687] env[65726]: DEBUG nova.objects.instance [None req-96ccefc5-467e-42ef-be69-e76eb66b409b tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Lazy-loading 'info_cache' on Instance uuid 0361ae8f-634d-4161-a594-7f93dbb1b7cd {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1125.824680] env[65726]: DEBUG oslo_vmware.api [None req-ad75268a-0b50-4f80-9f3a-c89856a157c4 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5116786, 'name': ReconfigVM_Task, 'duration_secs': 0.683971} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1125.824680] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-ad75268a-0b50-4f80-9f3a-c89856a157c4 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 378c1d18-4bb6-4245-80bf-370fb7af0575] Reconfigured VM instance instance-00000068 to attach disk [datastore1] 378c1d18-4bb6-4245-80bf-370fb7af0575/378c1d18-4bb6-4245-80bf-370fb7af0575.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1125.824916] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a660437d-3147-4bff-8d75-c092cf372ef0 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.832605] env[65726]: DEBUG oslo_vmware.api [None req-ad75268a-0b50-4f80-9f3a-c89856a157c4 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Waiting for the task: (returnval){ [ 1125.832605] env[65726]: value = "task-5116789" [ 1125.832605] env[65726]: _type = "Task" [ 1125.832605] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1125.841704] env[65726]: DEBUG oslo_vmware.api [None req-ad75268a-0b50-4f80-9f3a-c89856a157c4 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5116789, 'name': Rename_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1125.937043] env[65726]: DEBUG oslo_vmware.api [None req-bb18e015-a533-4756-8784-6ff16584fde4 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': task-5116774, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1126.026104] env[65726]: DEBUG nova.compute.manager [None req-72f2a102-8ae8-4f21-a769-e63022914c54 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: 3bc25953-21b1-4729-af09-e7211fd8b2c6] Starting instance... {{(pid=65726) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 1126.204589] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a2dc650f-ed4c-4edd-a0f8-7cba2e214975 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Acquiring lock "refresh_cache-e7156a10-631a-4cc5-a544-88ce37763d02" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1126.204877] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a2dc650f-ed4c-4edd-a0f8-7cba2e214975 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Acquired lock "refresh_cache-e7156a10-631a-4cc5-a544-88ce37763d02" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1126.205067] env[65726]: DEBUG nova.network.neutron [None req-a2dc650f-ed4c-4edd-a0f8-7cba2e214975 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: e7156a10-631a-4cc5-a544-88ce37763d02] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1126.243773] env[65726]: DEBUG oslo_vmware.api [None req-116f601c-d8e3-4bed-a9f0-abae1fb1e4c5 tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Task: {'id': task-5116788, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1126.347196] env[65726]: DEBUG oslo_vmware.api [None req-ad75268a-0b50-4f80-9f3a-c89856a157c4 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5116789, 'name': Rename_Task, 'duration_secs': 0.298098} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1126.347575] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-ad75268a-0b50-4f80-9f3a-c89856a157c4 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 378c1d18-4bb6-4245-80bf-370fb7af0575] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1126.347884] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e73ee04a-4e67-4020-b284-25f3bca981f2 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.355367] env[65726]: DEBUG oslo_vmware.api [None req-ad75268a-0b50-4f80-9f3a-c89856a157c4 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Waiting for the task: (returnval){ [ 1126.355367] env[65726]: value = "task-5116790" [ 1126.355367] env[65726]: _type = "Task" [ 1126.355367] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1126.366947] env[65726]: DEBUG oslo_vmware.api [None req-ad75268a-0b50-4f80-9f3a-c89856a157c4 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5116790, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1126.437272] env[65726]: DEBUG oslo_vmware.api [None req-bb18e015-a533-4756-8784-6ff16584fde4 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': task-5116774, 'name': ReconfigVM_Task} progress is 18%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1126.560586] env[65726]: DEBUG oslo_concurrency.lockutils [None req-72f2a102-8ae8-4f21-a769-e63022914c54 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1126.560586] env[65726]: DEBUG oslo_concurrency.lockutils [None req-72f2a102-8ae8-4f21-a769-e63022914c54 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1126.561820] env[65726]: INFO nova.compute.claims [None req-72f2a102-8ae8-4f21-a769-e63022914c54 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: 3bc25953-21b1-4729-af09-e7211fd8b2c6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1126.708545] env[65726]: WARNING openstack [None req-a2dc650f-ed4c-4edd-a0f8-7cba2e214975 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1126.708941] env[65726]: WARNING openstack [None req-a2dc650f-ed4c-4edd-a0f8-7cba2e214975 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1126.743710] env[65726]: DEBUG oslo_vmware.api [None req-116f601c-d8e3-4bed-a9f0-abae1fb1e4c5 tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Task: {'id': task-5116788, 'name': PowerOnVM_Task, 'duration_secs': 0.858059} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1126.743929] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-116f601c-d8e3-4bed-a9f0-abae1fb1e4c5 tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] [instance: 4809fcf6-59d4-409c-bd8d-981dc85a686f] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1126.744210] env[65726]: DEBUG nova.compute.manager [None req-116f601c-d8e3-4bed-a9f0-abae1fb1e4c5 tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] [instance: 4809fcf6-59d4-409c-bd8d-981dc85a686f] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1126.745120] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84dcd4ed-902e-49db-a4be-186f7c6507b2 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.785143] env[65726]: WARNING openstack [None req-96ccefc5-467e-42ef-be69-e76eb66b409b tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1126.785716] env[65726]: WARNING openstack [None req-96ccefc5-467e-42ef-be69-e76eb66b409b tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1126.835654] env[65726]: WARNING openstack [None req-a2dc650f-ed4c-4edd-a0f8-7cba2e214975 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1126.835654] env[65726]: WARNING openstack [None req-a2dc650f-ed4c-4edd-a0f8-7cba2e214975 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1126.866496] env[65726]: DEBUG oslo_vmware.api [None req-ad75268a-0b50-4f80-9f3a-c89856a157c4 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5116790, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1126.888335] env[65726]: WARNING openstack [None req-96ccefc5-467e-42ef-be69-e76eb66b409b tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1126.888730] env[65726]: WARNING openstack [None req-96ccefc5-467e-42ef-be69-e76eb66b409b tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1126.937382] env[65726]: DEBUG oslo_vmware.api [None req-bb18e015-a533-4756-8784-6ff16584fde4 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': task-5116774, 'name': ReconfigVM_Task, 'duration_secs': 5.858703} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1126.942054] env[65726]: DEBUG oslo_concurrency.lockutils [None req-bb18e015-a533-4756-8784-6ff16584fde4 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Releasing lock "51f5337d-2e23-4b7c-b590-76be6bd82411" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1126.942165] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-bb18e015-a533-4756-8784-6ff16584fde4 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: 51f5337d-2e23-4b7c-b590-76be6bd82411] Reconfigured VM to detach interface {{(pid=65726) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1978}} [ 1126.942810] env[65726]: WARNING openstack [None req-bb18e015-a533-4756-8784-6ff16584fde4 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1126.943220] env[65726]: WARNING openstack [None req-bb18e015-a533-4756-8784-6ff16584fde4 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1126.956589] env[65726]: WARNING openstack [None req-a2dc650f-ed4c-4edd-a0f8-7cba2e214975 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1126.956589] env[65726]: WARNING openstack [None req-a2dc650f-ed4c-4edd-a0f8-7cba2e214975 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1126.992498] env[65726]: WARNING openstack [None req-bb18e015-a533-4756-8784-6ff16584fde4 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1126.993749] env[65726]: WARNING openstack [None req-bb18e015-a533-4756-8784-6ff16584fde4 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1127.004302] env[65726]: WARNING openstack [None req-96ccefc5-467e-42ef-be69-e76eb66b409b tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1127.004717] env[65726]: WARNING openstack [None req-96ccefc5-467e-42ef-be69-e76eb66b409b tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1127.106044] env[65726]: DEBUG nova.network.neutron [None req-a2dc650f-ed4c-4edd-a0f8-7cba2e214975 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: e7156a10-631a-4cc5-a544-88ce37763d02] Updating instance_info_cache with network_info: [{"id": "44a1554d-d380-49b2-ad33-9a251114c93c", "address": "fa:16:3e:63:0a:a8", "network": {"id": "4c972d9d-db3b-4779-ac39-39d9a2d141dd", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1418913861-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cbb17a09b35c4c22ade5c4082b10bda9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "50cf0a70-948d-4611-af05-94c1483064ed", "external-id": "nsx-vlan-transportzone-536", "segmentation_id": 536, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap44a1554d-d3", "ovs_interfaceid": "44a1554d-d380-49b2-ad33-9a251114c93c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1127.150188] env[65726]: DEBUG nova.network.neutron [None req-96ccefc5-467e-42ef-be69-e76eb66b409b tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 0361ae8f-634d-4161-a594-7f93dbb1b7cd] Updating instance_info_cache with network_info: [{"id": "2bfd7147-87fd-4a19-b28b-eb95f51f4d00", "address": "fa:16:3e:d6:07:64", "network": {"id": "9dd55fb6-0dd2-484d-b94d-fbdccb958c79", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1005598846-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f93d852e2c904f42981cbfff882558e9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "74f30339-6421-4654-bddb-81d7f34db9d7", "external-id": "nsx-vlan-transportzone-899", "segmentation_id": 899, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2bfd7147-87", "ovs_interfaceid": "2bfd7147-87fd-4a19-b28b-eb95f51f4d00", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1127.373223] env[65726]: DEBUG oslo_vmware.api [None req-ad75268a-0b50-4f80-9f3a-c89856a157c4 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5116790, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1127.522123] env[65726]: DEBUG oslo_concurrency.lockutils [None req-8a669c02-f06b-4fc5-8c97-9ab50c8d780e tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Acquiring lock "4809fcf6-59d4-409c-bd8d-981dc85a686f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1127.522123] env[65726]: DEBUG oslo_concurrency.lockutils [None req-8a669c02-f06b-4fc5-8c97-9ab50c8d780e tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Lock "4809fcf6-59d4-409c-bd8d-981dc85a686f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1127.522123] env[65726]: DEBUG oslo_concurrency.lockutils [None req-8a669c02-f06b-4fc5-8c97-9ab50c8d780e tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Acquiring lock "4809fcf6-59d4-409c-bd8d-981dc85a686f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1127.522123] env[65726]: DEBUG oslo_concurrency.lockutils [None req-8a669c02-f06b-4fc5-8c97-9ab50c8d780e tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Lock "4809fcf6-59d4-409c-bd8d-981dc85a686f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1127.522465] env[65726]: DEBUG oslo_concurrency.lockutils [None req-8a669c02-f06b-4fc5-8c97-9ab50c8d780e tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Lock "4809fcf6-59d4-409c-bd8d-981dc85a686f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1127.524924] env[65726]: INFO nova.compute.manager [None req-8a669c02-f06b-4fc5-8c97-9ab50c8d780e tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] [instance: 4809fcf6-59d4-409c-bd8d-981dc85a686f] Terminating instance [ 1127.610531] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a2dc650f-ed4c-4edd-a0f8-7cba2e214975 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Releasing lock "refresh_cache-e7156a10-631a-4cc5-a544-88ce37763d02" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1127.657275] env[65726]: DEBUG oslo_concurrency.lockutils [None req-96ccefc5-467e-42ef-be69-e76eb66b409b tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Releasing lock "refresh_cache-0361ae8f-634d-4161-a594-7f93dbb1b7cd" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1127.657554] env[65726]: DEBUG nova.objects.instance [None req-96ccefc5-467e-42ef-be69-e76eb66b409b tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Lazy-loading 'migration_context' on Instance uuid 0361ae8f-634d-4161-a594-7f93dbb1b7cd {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1127.765340] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95879063-fc38-43aa-91fd-b6bf7fa8eef3 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.773952] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9744074-8e29-49d7-bbbc-b805d404f328 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.806146] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a17f2ddd-909a-4ded-82a4-469a7611f00d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.814626] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a95f5a2-e60d-4325-92ee-01b71b0e425e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.830263] env[65726]: DEBUG nova.compute.provider_tree [None req-72f2a102-8ae8-4f21-a769-e63022914c54 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1127.868097] env[65726]: DEBUG oslo_vmware.api [None req-ad75268a-0b50-4f80-9f3a-c89856a157c4 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5116790, 'name': PowerOnVM_Task, 'duration_secs': 1.058804} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1127.868398] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-ad75268a-0b50-4f80-9f3a-c89856a157c4 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 378c1d18-4bb6-4245-80bf-370fb7af0575] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1127.868618] env[65726]: INFO nova.compute.manager [None req-ad75268a-0b50-4f80-9f3a-c89856a157c4 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 378c1d18-4bb6-4245-80bf-370fb7af0575] Took 9.28 seconds to spawn the instance on the hypervisor. [ 1127.868832] env[65726]: DEBUG nova.compute.manager [None req-ad75268a-0b50-4f80-9f3a-c89856a157c4 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 378c1d18-4bb6-4245-80bf-370fb7af0575] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1127.869681] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0cf9ffb-b670-493f-ad6b-968b78750b0d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.030217] env[65726]: DEBUG nova.compute.manager [None req-8a669c02-f06b-4fc5-8c97-9ab50c8d780e tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] [instance: 4809fcf6-59d4-409c-bd8d-981dc85a686f] Start destroying the instance on the hypervisor. {{(pid=65726) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1128.030478] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-8a669c02-f06b-4fc5-8c97-9ab50c8d780e tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] [instance: 4809fcf6-59d4-409c-bd8d-981dc85a686f] Destroying instance {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1128.031422] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12c380d5-408e-4b52-99cb-348b7b83725d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.039840] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-8a669c02-f06b-4fc5-8c97-9ab50c8d780e tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] [instance: 4809fcf6-59d4-409c-bd8d-981dc85a686f] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1128.040102] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3ee4da14-d8ae-4a4c-9ce9-c9d6c9b999d9 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.046392] env[65726]: DEBUG oslo_vmware.api [None req-8a669c02-f06b-4fc5-8c97-9ab50c8d780e tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Waiting for the task: (returnval){ [ 1128.046392] env[65726]: value = "task-5116792" [ 1128.046392] env[65726]: _type = "Task" [ 1128.046392] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1128.055788] env[65726]: DEBUG oslo_vmware.api [None req-8a669c02-f06b-4fc5-8c97-9ab50c8d780e tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Task: {'id': task-5116792, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1128.162078] env[65726]: DEBUG nova.objects.base [None req-96ccefc5-467e-42ef-be69-e76eb66b409b tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Object Instance<0361ae8f-634d-4161-a594-7f93dbb1b7cd> lazy-loaded attributes: info_cache,migration_context {{(pid=65726) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1128.162479] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbed6bff-40f9-4a70-b280-151ba2134964 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.183781] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-77a1a4ac-98a7-49da-9c09-db4f24a60f0b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.190059] env[65726]: DEBUG oslo_vmware.api [None req-96ccefc5-467e-42ef-be69-e76eb66b409b tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Waiting for the task: (returnval){ [ 1128.190059] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52d666a8-6292-663e-9681-5f500945cca9" [ 1128.190059] env[65726]: _type = "Task" [ 1128.190059] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1128.198730] env[65726]: DEBUG oslo_vmware.api [None req-96ccefc5-467e-42ef-be69-e76eb66b409b tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52d666a8-6292-663e-9681-5f500945cca9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1128.336338] env[65726]: DEBUG nova.scheduler.client.report [None req-72f2a102-8ae8-4f21-a769-e63022914c54 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1128.391202] env[65726]: INFO nova.compute.manager [None req-ad75268a-0b50-4f80-9f3a-c89856a157c4 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 378c1d18-4bb6-4245-80bf-370fb7af0575] Took 14.50 seconds to build instance. [ 1128.408897] env[65726]: DEBUG oslo_concurrency.lockutils [None req-bb18e015-a533-4756-8784-6ff16584fde4 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Acquiring lock "refresh_cache-51f5337d-2e23-4b7c-b590-76be6bd82411" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1128.409146] env[65726]: DEBUG oslo_concurrency.lockutils [None req-bb18e015-a533-4756-8784-6ff16584fde4 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Acquired lock "refresh_cache-51f5337d-2e23-4b7c-b590-76be6bd82411" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1128.409298] env[65726]: DEBUG nova.network.neutron [None req-bb18e015-a533-4756-8784-6ff16584fde4 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: 51f5337d-2e23-4b7c-b590-76be6bd82411] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1128.557792] env[65726]: DEBUG oslo_vmware.api [None req-8a669c02-f06b-4fc5-8c97-9ab50c8d780e tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Task: {'id': task-5116792, 'name': PowerOffVM_Task, 'duration_secs': 0.365547} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1128.557792] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-8a669c02-f06b-4fc5-8c97-9ab50c8d780e tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] [instance: 4809fcf6-59d4-409c-bd8d-981dc85a686f] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1128.558024] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-8a669c02-f06b-4fc5-8c97-9ab50c8d780e tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] [instance: 4809fcf6-59d4-409c-bd8d-981dc85a686f] Unregistering the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1128.558169] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1ac37f80-f8f4-4e6a-b6cc-c79777d1132c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.650551] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-8a669c02-f06b-4fc5-8c97-9ab50c8d780e tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] [instance: 4809fcf6-59d4-409c-bd8d-981dc85a686f] Unregistered the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1128.651082] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-8a669c02-f06b-4fc5-8c97-9ab50c8d780e tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] [instance: 4809fcf6-59d4-409c-bd8d-981dc85a686f] Deleting contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1128.651194] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-8a669c02-f06b-4fc5-8c97-9ab50c8d780e tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Deleting the datastore file [datastore1] 4809fcf6-59d4-409c-bd8d-981dc85a686f {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1128.651473] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f0f24af7-89bb-4cff-848b-e62ce7cb79e0 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.659630] env[65726]: DEBUG oslo_vmware.api [None req-8a669c02-f06b-4fc5-8c97-9ab50c8d780e tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Waiting for the task: (returnval){ [ 1128.659630] env[65726]: value = "task-5116795" [ 1128.659630] env[65726]: _type = "Task" [ 1128.659630] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1128.668555] env[65726]: DEBUG oslo_vmware.api [None req-8a669c02-f06b-4fc5-8c97-9ab50c8d780e tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Task: {'id': task-5116795, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1128.700449] env[65726]: DEBUG oslo_vmware.api [None req-96ccefc5-467e-42ef-be69-e76eb66b409b tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52d666a8-6292-663e-9681-5f500945cca9, 'name': SearchDatastore_Task, 'duration_secs': 0.015948} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1128.701144] env[65726]: DEBUG oslo_concurrency.lockutils [None req-96ccefc5-467e-42ef-be69-e76eb66b409b tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1128.843489] env[65726]: DEBUG oslo_concurrency.lockutils [None req-72f2a102-8ae8-4f21-a769-e63022914c54 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.283s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1128.844072] env[65726]: DEBUG nova.compute.manager [None req-72f2a102-8ae8-4f21-a769-e63022914c54 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: 3bc25953-21b1-4729-af09-e7211fd8b2c6] Start building networks asynchronously for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1128.847009] env[65726]: DEBUG oslo_concurrency.lockutils [None req-96ccefc5-467e-42ef-be69-e76eb66b409b tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 0.146s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1128.894338] env[65726]: DEBUG oslo_concurrency.lockutils [None req-ad75268a-0b50-4f80-9f3a-c89856a157c4 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Lock "378c1d18-4bb6-4245-80bf-370fb7af0575" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.018s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1128.914169] env[65726]: WARNING openstack [None req-bb18e015-a533-4756-8784-6ff16584fde4 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1128.914576] env[65726]: WARNING openstack [None req-bb18e015-a533-4756-8784-6ff16584fde4 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1129.130272] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b474cd82-3077-4941-8b49-6a0958a96178 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.152717] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-a2dc650f-ed4c-4edd-a0f8-7cba2e214975 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: e7156a10-631a-4cc5-a544-88ce37763d02] Updating instance 'e7156a10-631a-4cc5-a544-88ce37763d02' progress to 0 {{(pid=65726) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1129.170166] env[65726]: DEBUG oslo_vmware.api [None req-8a669c02-f06b-4fc5-8c97-9ab50c8d780e tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Task: {'id': task-5116795, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.200437} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1129.170493] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-8a669c02-f06b-4fc5-8c97-9ab50c8d780e tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Deleted the datastore file {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1129.170716] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-8a669c02-f06b-4fc5-8c97-9ab50c8d780e tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] [instance: 4809fcf6-59d4-409c-bd8d-981dc85a686f] Deleted contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1129.170942] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-8a669c02-f06b-4fc5-8c97-9ab50c8d780e tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] [instance: 4809fcf6-59d4-409c-bd8d-981dc85a686f] Instance destroyed {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1129.171162] env[65726]: INFO nova.compute.manager [None req-8a669c02-f06b-4fc5-8c97-9ab50c8d780e tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] [instance: 4809fcf6-59d4-409c-bd8d-981dc85a686f] Took 1.14 seconds to destroy the instance on the hypervisor. [ 1129.171451] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-8a669c02-f06b-4fc5-8c97-9ab50c8d780e tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1129.171679] env[65726]: DEBUG nova.compute.manager [-] [instance: 4809fcf6-59d4-409c-bd8d-981dc85a686f] Deallocating network for instance {{(pid=65726) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1129.171808] env[65726]: DEBUG nova.network.neutron [-] [instance: 4809fcf6-59d4-409c-bd8d-981dc85a686f] deallocate_for_instance() {{(pid=65726) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 1129.172526] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1129.172973] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1129.238037] env[65726]: DEBUG nova.compute.manager [req-ce573974-21ab-4748-a87f-44eae5ae519b req-f1bf5ee7-25e6-4a95-b4c2-d3720a8362c5 service nova] [instance: 51f5337d-2e23-4b7c-b590-76be6bd82411] Received event network-changed-7407bcd9-3202-4798-9eaa-252fb3fdb21f {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1129.238490] env[65726]: DEBUG nova.compute.manager [req-ce573974-21ab-4748-a87f-44eae5ae519b req-f1bf5ee7-25e6-4a95-b4c2-d3720a8362c5 service nova] [instance: 51f5337d-2e23-4b7c-b590-76be6bd82411] Refreshing instance network info cache due to event network-changed-7407bcd9-3202-4798-9eaa-252fb3fdb21f. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 1129.238568] env[65726]: DEBUG oslo_concurrency.lockutils [req-ce573974-21ab-4748-a87f-44eae5ae519b req-f1bf5ee7-25e6-4a95-b4c2-d3720a8362c5 service nova] Acquiring lock "refresh_cache-51f5337d-2e23-4b7c-b590-76be6bd82411" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1129.274542] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1129.274872] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1129.292275] env[65726]: WARNING openstack [None req-bb18e015-a533-4756-8784-6ff16584fde4 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1129.292275] env[65726]: WARNING openstack [None req-bb18e015-a533-4756-8784-6ff16584fde4 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1129.354035] env[65726]: DEBUG nova.compute.utils [None req-72f2a102-8ae8-4f21-a769-e63022914c54 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Using /dev/sd instead of None {{(pid=65726) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1129.359987] env[65726]: DEBUG nova.compute.manager [None req-72f2a102-8ae8-4f21-a769-e63022914c54 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: 3bc25953-21b1-4729-af09-e7211fd8b2c6] Allocating IP information in the background. {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1129.360089] env[65726]: DEBUG nova.network.neutron [None req-72f2a102-8ae8-4f21-a769-e63022914c54 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: 3bc25953-21b1-4729-af09-e7211fd8b2c6] allocate_for_instance() {{(pid=65726) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 1129.361065] env[65726]: WARNING neutronclient.v2_0.client [None req-72f2a102-8ae8-4f21-a769-e63022914c54 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1129.361065] env[65726]: WARNING neutronclient.v2_0.client [None req-72f2a102-8ae8-4f21-a769-e63022914c54 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1129.364302] env[65726]: WARNING openstack [None req-72f2a102-8ae8-4f21-a769-e63022914c54 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1129.364302] env[65726]: WARNING openstack [None req-72f2a102-8ae8-4f21-a769-e63022914c54 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1129.536588] env[65726]: DEBUG nova.policy [None req-72f2a102-8ae8-4f21-a769-e63022914c54 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f3050294b8f34a0f97d10b038b048779', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a090d553766847e58e3231b966c92565', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65726) authorize /opt/stack/nova/nova/policy.py:192}} [ 1129.583720] env[65726]: WARNING openstack [None req-bb18e015-a533-4756-8784-6ff16584fde4 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1129.584232] env[65726]: WARNING openstack [None req-bb18e015-a533-4756-8784-6ff16584fde4 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1129.595207] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b661bc9-24b4-4596-bc98-c13d8d8eabed {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.609507] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4499a1a-203c-4a38-900b-de7d615a6624 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.650492] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f2db326-4c60-4421-9095-dfb6fddde280 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.660591] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-a2dc650f-ed4c-4edd-a0f8-7cba2e214975 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: e7156a10-631a-4cc5-a544-88ce37763d02] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1129.660999] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1bb1804a-834c-401d-8a8e-a8c70b840dae {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.663959] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e4e4791-d970-4118-8449-776b5576bcf4 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.679405] env[65726]: DEBUG nova.compute.manager [req-9533a06e-0dff-4d02-84fe-e9164e7366e8 req-590a3fca-5a4f-4da2-8e3c-f2e611f11bf5 service nova] [instance: 378c1d18-4bb6-4245-80bf-370fb7af0575] Received event network-changed-4765f6c2-d161-4e36-a6a8-3c8d39c47a40 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1129.679621] env[65726]: DEBUG nova.compute.manager [req-9533a06e-0dff-4d02-84fe-e9164e7366e8 req-590a3fca-5a4f-4da2-8e3c-f2e611f11bf5 service nova] [instance: 378c1d18-4bb6-4245-80bf-370fb7af0575] Refreshing instance network info cache due to event network-changed-4765f6c2-d161-4e36-a6a8-3c8d39c47a40. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 1129.679802] env[65726]: DEBUG oslo_concurrency.lockutils [req-9533a06e-0dff-4d02-84fe-e9164e7366e8 req-590a3fca-5a4f-4da2-8e3c-f2e611f11bf5 service nova] Acquiring lock "refresh_cache-378c1d18-4bb6-4245-80bf-370fb7af0575" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1129.680828] env[65726]: DEBUG oslo_concurrency.lockutils [req-9533a06e-0dff-4d02-84fe-e9164e7366e8 req-590a3fca-5a4f-4da2-8e3c-f2e611f11bf5 service nova] Acquired lock "refresh_cache-378c1d18-4bb6-4245-80bf-370fb7af0575" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1129.680828] env[65726]: DEBUG nova.network.neutron [req-9533a06e-0dff-4d02-84fe-e9164e7366e8 req-590a3fca-5a4f-4da2-8e3c-f2e611f11bf5 service nova] [instance: 378c1d18-4bb6-4245-80bf-370fb7af0575] Refreshing network info cache for port 4765f6c2-d161-4e36-a6a8-3c8d39c47a40 {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 1129.692711] env[65726]: DEBUG nova.compute.provider_tree [None req-96ccefc5-467e-42ef-be69-e76eb66b409b tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1129.694858] env[65726]: DEBUG oslo_vmware.api [None req-a2dc650f-ed4c-4edd-a0f8-7cba2e214975 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Waiting for the task: (returnval){ [ 1129.694858] env[65726]: value = "task-5116796" [ 1129.694858] env[65726]: _type = "Task" [ 1129.694858] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1129.698465] env[65726]: INFO nova.network.neutron [None req-bb18e015-a533-4756-8784-6ff16584fde4 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: 51f5337d-2e23-4b7c-b590-76be6bd82411] Port fb580bad-fd60-471d-95f2-066352417659 from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 1129.698933] env[65726]: DEBUG nova.network.neutron [None req-bb18e015-a533-4756-8784-6ff16584fde4 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: 51f5337d-2e23-4b7c-b590-76be6bd82411] Updating instance_info_cache with network_info: [{"id": "7407bcd9-3202-4798-9eaa-252fb3fdb21f", "address": "fa:16:3e:79:d2:9d", "network": {"id": "51a66448-0f72-4266-8868-856b48be839f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-155658961-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d46ccb00794f458b85da4a93879139ee", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b1f3e6c3-5584-4852-9017-476ab8ac4946", "external-id": "nsx-vlan-transportzone-304", "segmentation_id": 304, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7407bcd9-32", "ovs_interfaceid": "7407bcd9-3202-4798-9eaa-252fb3fdb21f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1129.714373] env[65726]: DEBUG oslo_vmware.api [None req-a2dc650f-ed4c-4edd-a0f8-7cba2e214975 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': task-5116796, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1129.851178] env[65726]: DEBUG nova.network.neutron [None req-72f2a102-8ae8-4f21-a769-e63022914c54 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: 3bc25953-21b1-4729-af09-e7211fd8b2c6] Successfully created port: a7e1795d-67d0-4f32-baea-744806988e4b {{(pid=65726) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1129.861159] env[65726]: DEBUG nova.compute.manager [None req-72f2a102-8ae8-4f21-a769-e63022914c54 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: 3bc25953-21b1-4729-af09-e7211fd8b2c6] Start building block device mappings for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1130.044529] env[65726]: DEBUG nova.network.neutron [-] [instance: 4809fcf6-59d4-409c-bd8d-981dc85a686f] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1130.180665] env[65726]: DEBUG oslo_concurrency.lockutils [None req-263ff8f3-91c3-4e15-8f8e-307b6d4d0871 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Acquiring lock "interface-0d42fa75-20e9-4646-9b08-17015b7f068c-fb580bad-fd60-471d-95f2-066352417659" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1130.182956] env[65726]: DEBUG oslo_concurrency.lockutils [None req-263ff8f3-91c3-4e15-8f8e-307b6d4d0871 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Lock "interface-0d42fa75-20e9-4646-9b08-17015b7f068c-fb580bad-fd60-471d-95f2-066352417659" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.001s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1130.182956] env[65726]: DEBUG nova.objects.instance [None req-263ff8f3-91c3-4e15-8f8e-307b6d4d0871 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Lazy-loading 'flavor' on Instance uuid 0d42fa75-20e9-4646-9b08-17015b7f068c {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1130.197364] env[65726]: DEBUG nova.scheduler.client.report [None req-96ccefc5-467e-42ef-be69-e76eb66b409b tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1130.202173] env[65726]: WARNING openstack [req-9533a06e-0dff-4d02-84fe-e9164e7366e8 req-590a3fca-5a4f-4da2-8e3c-f2e611f11bf5 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1130.202577] env[65726]: WARNING openstack [req-9533a06e-0dff-4d02-84fe-e9164e7366e8 req-590a3fca-5a4f-4da2-8e3c-f2e611f11bf5 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1130.210902] env[65726]: DEBUG oslo_concurrency.lockutils [None req-bb18e015-a533-4756-8784-6ff16584fde4 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Releasing lock "refresh_cache-51f5337d-2e23-4b7c-b590-76be6bd82411" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1130.219214] env[65726]: DEBUG oslo_concurrency.lockutils [req-ce573974-21ab-4748-a87f-44eae5ae519b req-f1bf5ee7-25e6-4a95-b4c2-d3720a8362c5 service nova] Acquired lock "refresh_cache-51f5337d-2e23-4b7c-b590-76be6bd82411" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1130.219214] env[65726]: DEBUG nova.network.neutron [req-ce573974-21ab-4748-a87f-44eae5ae519b req-f1bf5ee7-25e6-4a95-b4c2-d3720a8362c5 service nova] [instance: 51f5337d-2e23-4b7c-b590-76be6bd82411] Refreshing network info cache for port 7407bcd9-3202-4798-9eaa-252fb3fdb21f {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 1130.226977] env[65726]: DEBUG oslo_vmware.api [None req-a2dc650f-ed4c-4edd-a0f8-7cba2e214975 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': task-5116796, 'name': PowerOffVM_Task, 'duration_secs': 0.430936} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1130.227361] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-a2dc650f-ed4c-4edd-a0f8-7cba2e214975 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: e7156a10-631a-4cc5-a544-88ce37763d02] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1130.227499] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-a2dc650f-ed4c-4edd-a0f8-7cba2e214975 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: e7156a10-631a-4cc5-a544-88ce37763d02] Updating instance 'e7156a10-631a-4cc5-a544-88ce37763d02' progress to 17 {{(pid=65726) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1130.336431] env[65726]: WARNING openstack [req-9533a06e-0dff-4d02-84fe-e9164e7366e8 req-590a3fca-5a4f-4da2-8e3c-f2e611f11bf5 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1130.336709] env[65726]: WARNING openstack [req-9533a06e-0dff-4d02-84fe-e9164e7366e8 req-590a3fca-5a4f-4da2-8e3c-f2e611f11bf5 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1130.419609] env[65726]: WARNING openstack [req-9533a06e-0dff-4d02-84fe-e9164e7366e8 req-590a3fca-5a4f-4da2-8e3c-f2e611f11bf5 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1130.420052] env[65726]: WARNING openstack [req-9533a06e-0dff-4d02-84fe-e9164e7366e8 req-590a3fca-5a4f-4da2-8e3c-f2e611f11bf5 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1130.522842] env[65726]: DEBUG nova.network.neutron [req-9533a06e-0dff-4d02-84fe-e9164e7366e8 req-590a3fca-5a4f-4da2-8e3c-f2e611f11bf5 service nova] [instance: 378c1d18-4bb6-4245-80bf-370fb7af0575] Updated VIF entry in instance network info cache for port 4765f6c2-d161-4e36-a6a8-3c8d39c47a40. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 1130.523488] env[65726]: DEBUG nova.network.neutron [req-9533a06e-0dff-4d02-84fe-e9164e7366e8 req-590a3fca-5a4f-4da2-8e3c-f2e611f11bf5 service nova] [instance: 378c1d18-4bb6-4245-80bf-370fb7af0575] Updating instance_info_cache with network_info: [{"id": "4765f6c2-d161-4e36-a6a8-3c8d39c47a40", "address": "fa:16:3e:3c:38:b2", "network": {"id": "0d0f561d-ac41-4ef5-9926-c707409799f7", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1134321217-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.160", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4eece77569624f90bf64e5c51974173f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2ce62383-8e84-4e26-955b-74c11392f4c9", "external-id": "nsx-vlan-transportzone-215", "segmentation_id": 215, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4765f6c2-d1", "ovs_interfaceid": "4765f6c2-d161-4e36-a6a8-3c8d39c47a40", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1130.536360] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9b5f9575-8a6d-44da-8cc3-16504101fcc2 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Acquiring lock "aa6ce489-c62f-4481-87b7-e74242aeb8ca" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1130.536620] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9b5f9575-8a6d-44da-8cc3-16504101fcc2 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Lock "aa6ce489-c62f-4481-87b7-e74242aeb8ca" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1130.536836] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9b5f9575-8a6d-44da-8cc3-16504101fcc2 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Acquiring lock "aa6ce489-c62f-4481-87b7-e74242aeb8ca-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1130.537026] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9b5f9575-8a6d-44da-8cc3-16504101fcc2 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Lock "aa6ce489-c62f-4481-87b7-e74242aeb8ca-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1130.537194] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9b5f9575-8a6d-44da-8cc3-16504101fcc2 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Lock "aa6ce489-c62f-4481-87b7-e74242aeb8ca-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1130.540778] env[65726]: INFO nova.compute.manager [None req-9b5f9575-8a6d-44da-8cc3-16504101fcc2 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] [instance: aa6ce489-c62f-4481-87b7-e74242aeb8ca] Terminating instance [ 1130.547396] env[65726]: INFO nova.compute.manager [-] [instance: 4809fcf6-59d4-409c-bd8d-981dc85a686f] Took 1.38 seconds to deallocate network for instance. [ 1130.688069] env[65726]: WARNING openstack [None req-263ff8f3-91c3-4e15-8f8e-307b6d4d0871 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1130.688566] env[65726]: WARNING openstack [None req-263ff8f3-91c3-4e15-8f8e-307b6d4d0871 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1130.718851] env[65726]: DEBUG oslo_concurrency.lockutils [None req-bb18e015-a533-4756-8784-6ff16584fde4 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Lock "interface-51f5337d-2e23-4b7c-b590-76be6bd82411-fb580bad-fd60-471d-95f2-066352417659" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 10.422s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1130.725933] env[65726]: WARNING openstack [req-ce573974-21ab-4748-a87f-44eae5ae519b req-f1bf5ee7-25e6-4a95-b4c2-d3720a8362c5 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1130.725933] env[65726]: WARNING openstack [req-ce573974-21ab-4748-a87f-44eae5ae519b req-f1bf5ee7-25e6-4a95-b4c2-d3720a8362c5 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1130.738259] env[65726]: DEBUG nova.virt.hardware [None req-a2dc650f-ed4c-4edd-a0f8-7cba2e214975 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1130.738564] env[65726]: DEBUG nova.virt.hardware [None req-a2dc650f-ed4c-4edd-a0f8-7cba2e214975 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1130.738728] env[65726]: DEBUG nova.virt.hardware [None req-a2dc650f-ed4c-4edd-a0f8-7cba2e214975 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1130.738906] env[65726]: DEBUG nova.virt.hardware [None req-a2dc650f-ed4c-4edd-a0f8-7cba2e214975 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1130.739095] env[65726]: DEBUG nova.virt.hardware [None req-a2dc650f-ed4c-4edd-a0f8-7cba2e214975 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1130.739274] env[65726]: DEBUG nova.virt.hardware [None req-a2dc650f-ed4c-4edd-a0f8-7cba2e214975 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1130.739588] env[65726]: DEBUG nova.virt.hardware [None req-a2dc650f-ed4c-4edd-a0f8-7cba2e214975 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1130.739809] env[65726]: DEBUG nova.virt.hardware [None req-a2dc650f-ed4c-4edd-a0f8-7cba2e214975 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1130.740048] env[65726]: DEBUG nova.virt.hardware [None req-a2dc650f-ed4c-4edd-a0f8-7cba2e214975 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1130.740253] env[65726]: DEBUG nova.virt.hardware [None req-a2dc650f-ed4c-4edd-a0f8-7cba2e214975 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1130.740483] env[65726]: DEBUG nova.virt.hardware [None req-a2dc650f-ed4c-4edd-a0f8-7cba2e214975 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1130.748103] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cdea21a3-22bd-44a5-912f-479fbe1bbad7 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.769603] env[65726]: DEBUG oslo_vmware.api [None req-a2dc650f-ed4c-4edd-a0f8-7cba2e214975 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Waiting for the task: (returnval){ [ 1130.769603] env[65726]: value = "task-5116798" [ 1130.769603] env[65726]: _type = "Task" [ 1130.769603] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1130.780174] env[65726]: DEBUG oslo_vmware.api [None req-a2dc650f-ed4c-4edd-a0f8-7cba2e214975 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': task-5116798, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1130.882023] env[65726]: DEBUG nova.compute.manager [None req-72f2a102-8ae8-4f21-a769-e63022914c54 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: 3bc25953-21b1-4729-af09-e7211fd8b2c6] Start spawning the instance on the hypervisor. {{(pid=65726) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1130.916885] env[65726]: DEBUG nova.virt.hardware [None req-72f2a102-8ae8-4f21-a769-e63022914c54 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T19:28:34Z,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1539f06fa3534e90acbc3a60c4b8bd3f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T19:28:35Z,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1130.917155] env[65726]: DEBUG nova.virt.hardware [None req-72f2a102-8ae8-4f21-a769-e63022914c54 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1130.917311] env[65726]: DEBUG nova.virt.hardware [None req-72f2a102-8ae8-4f21-a769-e63022914c54 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1130.917486] env[65726]: DEBUG nova.virt.hardware [None req-72f2a102-8ae8-4f21-a769-e63022914c54 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1130.917629] env[65726]: DEBUG nova.virt.hardware [None req-72f2a102-8ae8-4f21-a769-e63022914c54 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1130.917769] env[65726]: DEBUG nova.virt.hardware [None req-72f2a102-8ae8-4f21-a769-e63022914c54 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1130.917975] env[65726]: DEBUG nova.virt.hardware [None req-72f2a102-8ae8-4f21-a769-e63022914c54 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1130.918142] env[65726]: DEBUG nova.virt.hardware [None req-72f2a102-8ae8-4f21-a769-e63022914c54 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1130.918320] env[65726]: DEBUG nova.virt.hardware [None req-72f2a102-8ae8-4f21-a769-e63022914c54 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1130.918477] env[65726]: DEBUG nova.virt.hardware [None req-72f2a102-8ae8-4f21-a769-e63022914c54 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1130.918645] env[65726]: DEBUG nova.virt.hardware [None req-72f2a102-8ae8-4f21-a769-e63022914c54 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1130.919594] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4b69a2b-e747-4919-8450-c4e0e2b3fa2b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.928681] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e684f810-da42-462b-b991-8dbfbda529fc {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.935563] env[65726]: DEBUG nova.objects.instance [None req-263ff8f3-91c3-4e15-8f8e-307b6d4d0871 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Lazy-loading 'pci_requests' on Instance uuid 0d42fa75-20e9-4646-9b08-17015b7f068c {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1130.948729] env[65726]: WARNING openstack [req-ce573974-21ab-4748-a87f-44eae5ae519b req-f1bf5ee7-25e6-4a95-b4c2-d3720a8362c5 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1130.949128] env[65726]: WARNING openstack [req-ce573974-21ab-4748-a87f-44eae5ae519b req-f1bf5ee7-25e6-4a95-b4c2-d3720a8362c5 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1131.029020] env[65726]: DEBUG oslo_concurrency.lockutils [req-9533a06e-0dff-4d02-84fe-e9164e7366e8 req-590a3fca-5a4f-4da2-8e3c-f2e611f11bf5 service nova] Releasing lock "refresh_cache-378c1d18-4bb6-4245-80bf-370fb7af0575" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1131.029020] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e3b3b0a6-575e-4b59-bef3-4eef65e9267a tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Acquiring lock "6fd70b4a-63e4-4258-9cff-f3b582500b9f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1131.029020] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e3b3b0a6-575e-4b59-bef3-4eef65e9267a tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Lock "6fd70b4a-63e4-4258-9cff-f3b582500b9f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1131.029020] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e3b3b0a6-575e-4b59-bef3-4eef65e9267a tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Acquiring lock "6fd70b4a-63e4-4258-9cff-f3b582500b9f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1131.029020] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e3b3b0a6-575e-4b59-bef3-4eef65e9267a tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Lock "6fd70b4a-63e4-4258-9cff-f3b582500b9f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1131.029020] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e3b3b0a6-575e-4b59-bef3-4eef65e9267a tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Lock "6fd70b4a-63e4-4258-9cff-f3b582500b9f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1131.030685] env[65726]: INFO nova.compute.manager [None req-e3b3b0a6-575e-4b59-bef3-4eef65e9267a tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] [instance: 6fd70b4a-63e4-4258-9cff-f3b582500b9f] Terminating instance [ 1131.045618] env[65726]: DEBUG nova.compute.manager [None req-9b5f9575-8a6d-44da-8cc3-16504101fcc2 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] [instance: aa6ce489-c62f-4481-87b7-e74242aeb8ca] Start destroying the instance on the hypervisor. {{(pid=65726) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1131.045843] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-9b5f9575-8a6d-44da-8cc3-16504101fcc2 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] [instance: aa6ce489-c62f-4481-87b7-e74242aeb8ca] Destroying instance {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1131.046778] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03475b09-7edd-402a-93d6-57589f846932 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.053325] env[65726]: WARNING openstack [req-ce573974-21ab-4748-a87f-44eae5ae519b req-f1bf5ee7-25e6-4a95-b4c2-d3720a8362c5 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1131.053775] env[65726]: WARNING openstack [req-ce573974-21ab-4748-a87f-44eae5ae519b req-f1bf5ee7-25e6-4a95-b4c2-d3720a8362c5 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1131.063139] env[65726]: DEBUG oslo_concurrency.lockutils [None req-8a669c02-f06b-4fc5-8c97-9ab50c8d780e tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1131.068761] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-9b5f9575-8a6d-44da-8cc3-16504101fcc2 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] [instance: aa6ce489-c62f-4481-87b7-e74242aeb8ca] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1131.069431] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-bfc2596d-dc49-4bc7-a8ae-e82ca22e51e6 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.078089] env[65726]: DEBUG oslo_vmware.api [None req-9b5f9575-8a6d-44da-8cc3-16504101fcc2 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Waiting for the task: (returnval){ [ 1131.078089] env[65726]: value = "task-5116799" [ 1131.078089] env[65726]: _type = "Task" [ 1131.078089] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1131.088145] env[65726]: DEBUG oslo_vmware.api [None req-9b5f9575-8a6d-44da-8cc3-16504101fcc2 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Task: {'id': task-5116799, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1131.218516] env[65726]: DEBUG nova.network.neutron [req-ce573974-21ab-4748-a87f-44eae5ae519b req-f1bf5ee7-25e6-4a95-b4c2-d3720a8362c5 service nova] [instance: 51f5337d-2e23-4b7c-b590-76be6bd82411] Updated VIF entry in instance network info cache for port 7407bcd9-3202-4798-9eaa-252fb3fdb21f. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 1131.218902] env[65726]: DEBUG nova.network.neutron [req-ce573974-21ab-4748-a87f-44eae5ae519b req-f1bf5ee7-25e6-4a95-b4c2-d3720a8362c5 service nova] [instance: 51f5337d-2e23-4b7c-b590-76be6bd82411] Updating instance_info_cache with network_info: [{"id": "7407bcd9-3202-4798-9eaa-252fb3fdb21f", "address": "fa:16:3e:79:d2:9d", "network": {"id": "51a66448-0f72-4266-8868-856b48be839f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-155658961-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d46ccb00794f458b85da4a93879139ee", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b1f3e6c3-5584-4852-9017-476ab8ac4946", "external-id": "nsx-vlan-transportzone-304", "segmentation_id": 304, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7407bcd9-32", "ovs_interfaceid": "7407bcd9-3202-4798-9eaa-252fb3fdb21f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1131.221203] env[65726]: DEBUG oslo_concurrency.lockutils [None req-96ccefc5-467e-42ef-be69-e76eb66b409b tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.374s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1131.224438] env[65726]: DEBUG oslo_concurrency.lockutils [None req-8a669c02-f06b-4fc5-8c97-9ab50c8d780e tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.161s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1131.224719] env[65726]: DEBUG nova.objects.instance [None req-8a669c02-f06b-4fc5-8c97-9ab50c8d780e tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Lazy-loading 'resources' on Instance uuid 4809fcf6-59d4-409c-bd8d-981dc85a686f {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1131.280680] env[65726]: DEBUG oslo_vmware.api [None req-a2dc650f-ed4c-4edd-a0f8-7cba2e214975 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': task-5116798, 'name': ReconfigVM_Task, 'duration_secs': 0.427509} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1131.281189] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-a2dc650f-ed4c-4edd-a0f8-7cba2e214975 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: e7156a10-631a-4cc5-a544-88ce37763d02] Updating instance 'e7156a10-631a-4cc5-a544-88ce37763d02' progress to 33 {{(pid=65726) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1131.351471] env[65726]: DEBUG nova.compute.manager [req-a86d6fd6-aee6-4792-9710-f64a4c1c7ea5 req-d4f6c307-145d-4909-830e-4ce7991ab855 service nova] [instance: 0d42fa75-20e9-4646-9b08-17015b7f068c] Received event network-changed-7a95cb45-ebfc-4ac0-ad58-4bb57170e47a {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1131.351677] env[65726]: DEBUG nova.compute.manager [req-a86d6fd6-aee6-4792-9710-f64a4c1c7ea5 req-d4f6c307-145d-4909-830e-4ce7991ab855 service nova] [instance: 0d42fa75-20e9-4646-9b08-17015b7f068c] Refreshing instance network info cache due to event network-changed-7a95cb45-ebfc-4ac0-ad58-4bb57170e47a. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 1131.351894] env[65726]: DEBUG oslo_concurrency.lockutils [req-a86d6fd6-aee6-4792-9710-f64a4c1c7ea5 req-d4f6c307-145d-4909-830e-4ce7991ab855 service nova] Acquiring lock "refresh_cache-0d42fa75-20e9-4646-9b08-17015b7f068c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1131.352057] env[65726]: DEBUG oslo_concurrency.lockutils [req-a86d6fd6-aee6-4792-9710-f64a4c1c7ea5 req-d4f6c307-145d-4909-830e-4ce7991ab855 service nova] Acquired lock "refresh_cache-0d42fa75-20e9-4646-9b08-17015b7f068c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1131.352291] env[65726]: DEBUG nova.network.neutron [req-a86d6fd6-aee6-4792-9710-f64a4c1c7ea5 req-d4f6c307-145d-4909-830e-4ce7991ab855 service nova] [instance: 0d42fa75-20e9-4646-9b08-17015b7f068c] Refreshing network info cache for port 7a95cb45-ebfc-4ac0-ad58-4bb57170e47a {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 1131.438545] env[65726]: DEBUG nova.objects.base [None req-263ff8f3-91c3-4e15-8f8e-307b6d4d0871 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Object Instance<0d42fa75-20e9-4646-9b08-17015b7f068c> lazy-loaded attributes: flavor,pci_requests {{(pid=65726) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1131.438732] env[65726]: DEBUG nova.network.neutron [None req-263ff8f3-91c3-4e15-8f8e-307b6d4d0871 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: 0d42fa75-20e9-4646-9b08-17015b7f068c] allocate_for_instance() {{(pid=65726) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 1131.439056] env[65726]: WARNING neutronclient.v2_0.client [None req-263ff8f3-91c3-4e15-8f8e-307b6d4d0871 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1131.439366] env[65726]: WARNING neutronclient.v2_0.client [None req-263ff8f3-91c3-4e15-8f8e-307b6d4d0871 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1131.439955] env[65726]: WARNING openstack [None req-263ff8f3-91c3-4e15-8f8e-307b6d4d0871 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1131.440326] env[65726]: WARNING openstack [None req-263ff8f3-91c3-4e15-8f8e-307b6d4d0871 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1131.479436] env[65726]: DEBUG nova.network.neutron [None req-72f2a102-8ae8-4f21-a769-e63022914c54 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: 3bc25953-21b1-4729-af09-e7211fd8b2c6] Successfully updated port: a7e1795d-67d0-4f32-baea-744806988e4b {{(pid=65726) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 1131.532310] env[65726]: DEBUG nova.policy [None req-263ff8f3-91c3-4e15-8f8e-307b6d4d0871 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9b65b9194933469c908c7cc04478444b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd46ccb00794f458b85da4a93879139ee', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65726) authorize /opt/stack/nova/nova/policy.py:192}} [ 1131.535671] env[65726]: DEBUG nova.compute.manager [None req-e3b3b0a6-575e-4b59-bef3-4eef65e9267a tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] [instance: 6fd70b4a-63e4-4258-9cff-f3b582500b9f] Start destroying the instance on the hypervisor. {{(pid=65726) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1131.535868] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-e3b3b0a6-575e-4b59-bef3-4eef65e9267a tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] [instance: 6fd70b4a-63e4-4258-9cff-f3b582500b9f] Destroying instance {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1131.537016] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-869842fd-e1c3-4635-8c2d-fd4f62adea6a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.545274] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-e3b3b0a6-575e-4b59-bef3-4eef65e9267a tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] [instance: 6fd70b4a-63e4-4258-9cff-f3b582500b9f] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1131.545533] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4baa0a63-32cb-4f77-84c2-6f2a9660c12d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.552947] env[65726]: DEBUG oslo_vmware.api [None req-e3b3b0a6-575e-4b59-bef3-4eef65e9267a tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Waiting for the task: (returnval){ [ 1131.552947] env[65726]: value = "task-5116800" [ 1131.552947] env[65726]: _type = "Task" [ 1131.552947] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1131.562125] env[65726]: DEBUG oslo_vmware.api [None req-e3b3b0a6-575e-4b59-bef3-4eef65e9267a tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Task: {'id': task-5116800, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1131.588906] env[65726]: DEBUG oslo_vmware.api [None req-9b5f9575-8a6d-44da-8cc3-16504101fcc2 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Task: {'id': task-5116799, 'name': PowerOffVM_Task, 'duration_secs': 0.239609} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1131.589233] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-9b5f9575-8a6d-44da-8cc3-16504101fcc2 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] [instance: aa6ce489-c62f-4481-87b7-e74242aeb8ca] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1131.589425] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-9b5f9575-8a6d-44da-8cc3-16504101fcc2 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] [instance: aa6ce489-c62f-4481-87b7-e74242aeb8ca] Unregistering the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1131.589695] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-15e3a158-dc0d-49fc-8308-c0ddf631cffa {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.668539] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-9b5f9575-8a6d-44da-8cc3-16504101fcc2 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] [instance: aa6ce489-c62f-4481-87b7-e74242aeb8ca] Unregistered the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1131.668539] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-9b5f9575-8a6d-44da-8cc3-16504101fcc2 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] [instance: aa6ce489-c62f-4481-87b7-e74242aeb8ca] Deleting contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1131.668539] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-9b5f9575-8a6d-44da-8cc3-16504101fcc2 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Deleting the datastore file [datastore1] aa6ce489-c62f-4481-87b7-e74242aeb8ca {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1131.668539] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0632c1db-e544-4f6b-9672-a9c5fe511791 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.675381] env[65726]: DEBUG oslo_vmware.api [None req-9b5f9575-8a6d-44da-8cc3-16504101fcc2 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Waiting for the task: (returnval){ [ 1131.675381] env[65726]: value = "task-5116802" [ 1131.675381] env[65726]: _type = "Task" [ 1131.675381] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1131.683636] env[65726]: DEBUG oslo_vmware.api [None req-9b5f9575-8a6d-44da-8cc3-16504101fcc2 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Task: {'id': task-5116802, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1131.725706] env[65726]: DEBUG oslo_concurrency.lockutils [req-ce573974-21ab-4748-a87f-44eae5ae519b req-f1bf5ee7-25e6-4a95-b4c2-d3720a8362c5 service nova] Releasing lock "refresh_cache-51f5337d-2e23-4b7c-b590-76be6bd82411" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1131.790906] env[65726]: DEBUG nova.virt.hardware [None req-a2dc650f-ed4c-4edd-a0f8-7cba2e214975 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1131.791519] env[65726]: DEBUG nova.virt.hardware [None req-a2dc650f-ed4c-4edd-a0f8-7cba2e214975 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1131.791519] env[65726]: DEBUG nova.virt.hardware [None req-a2dc650f-ed4c-4edd-a0f8-7cba2e214975 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1131.791519] env[65726]: DEBUG nova.virt.hardware [None req-a2dc650f-ed4c-4edd-a0f8-7cba2e214975 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1131.791656] env[65726]: DEBUG nova.virt.hardware [None req-a2dc650f-ed4c-4edd-a0f8-7cba2e214975 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1131.791767] env[65726]: DEBUG nova.virt.hardware [None req-a2dc650f-ed4c-4edd-a0f8-7cba2e214975 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1131.791978] env[65726]: DEBUG nova.virt.hardware [None req-a2dc650f-ed4c-4edd-a0f8-7cba2e214975 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1131.792187] env[65726]: DEBUG nova.virt.hardware [None req-a2dc650f-ed4c-4edd-a0f8-7cba2e214975 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1131.792351] env[65726]: DEBUG nova.virt.hardware [None req-a2dc650f-ed4c-4edd-a0f8-7cba2e214975 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1131.792515] env[65726]: DEBUG nova.virt.hardware [None req-a2dc650f-ed4c-4edd-a0f8-7cba2e214975 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1131.792687] env[65726]: DEBUG nova.virt.hardware [None req-a2dc650f-ed4c-4edd-a0f8-7cba2e214975 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1131.799112] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-a2dc650f-ed4c-4edd-a0f8-7cba2e214975 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: e7156a10-631a-4cc5-a544-88ce37763d02] Reconfiguring VM instance instance-00000066 to detach disk 2000 {{(pid=65726) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1131.800360] env[65726]: INFO nova.scheduler.client.report [None req-96ccefc5-467e-42ef-be69-e76eb66b409b tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Deleted allocation for migration 995b2821-290f-47ee-adb1-3c8a9e262c05 [ 1131.801671] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ef61cae0-0f3b-4d73-a25a-53c0036a5a93 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.826863] env[65726]: DEBUG oslo_vmware.api [None req-a2dc650f-ed4c-4edd-a0f8-7cba2e214975 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Waiting for the task: (returnval){ [ 1131.826863] env[65726]: value = "task-5116803" [ 1131.826863] env[65726]: _type = "Task" [ 1131.826863] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1131.836492] env[65726]: DEBUG oslo_vmware.api [None req-a2dc650f-ed4c-4edd-a0f8-7cba2e214975 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': task-5116803, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1131.855652] env[65726]: WARNING openstack [req-a86d6fd6-aee6-4792-9710-f64a4c1c7ea5 req-d4f6c307-145d-4909-830e-4ce7991ab855 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1131.856092] env[65726]: WARNING openstack [req-a86d6fd6-aee6-4792-9710-f64a4c1c7ea5 req-d4f6c307-145d-4909-830e-4ce7991ab855 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1131.951924] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca586aa5-af99-4dc5-9821-f2bce4b47639 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.964037] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c47fa3a4-2027-4cd9-b03b-8a418123acd7 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.005535] env[65726]: DEBUG oslo_concurrency.lockutils [None req-72f2a102-8ae8-4f21-a769-e63022914c54 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Acquiring lock "refresh_cache-3bc25953-21b1-4729-af09-e7211fd8b2c6" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1132.005535] env[65726]: DEBUG oslo_concurrency.lockutils [None req-72f2a102-8ae8-4f21-a769-e63022914c54 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Acquired lock "refresh_cache-3bc25953-21b1-4729-af09-e7211fd8b2c6" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1132.005811] env[65726]: DEBUG nova.network.neutron [None req-72f2a102-8ae8-4f21-a769-e63022914c54 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: 3bc25953-21b1-4729-af09-e7211fd8b2c6] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1132.013959] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6adbadf6-9536-47b0-9d9a-e951195fed98 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.023225] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f642c9a-beba-4a0a-b2e4-abb0d2ae32c8 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.038365] env[65726]: DEBUG nova.compute.provider_tree [None req-8a669c02-f06b-4fc5-8c97-9ab50c8d780e tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1132.053324] env[65726]: WARNING openstack [req-a86d6fd6-aee6-4792-9710-f64a4c1c7ea5 req-d4f6c307-145d-4909-830e-4ce7991ab855 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1132.053713] env[65726]: WARNING openstack [req-a86d6fd6-aee6-4792-9710-f64a4c1c7ea5 req-d4f6c307-145d-4909-830e-4ce7991ab855 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1132.070042] env[65726]: DEBUG oslo_vmware.api [None req-e3b3b0a6-575e-4b59-bef3-4eef65e9267a tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Task: {'id': task-5116800, 'name': PowerOffVM_Task, 'duration_secs': 0.25633} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1132.070329] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-e3b3b0a6-575e-4b59-bef3-4eef65e9267a tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] [instance: 6fd70b4a-63e4-4258-9cff-f3b582500b9f] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1132.070493] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-e3b3b0a6-575e-4b59-bef3-4eef65e9267a tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] [instance: 6fd70b4a-63e4-4258-9cff-f3b582500b9f] Unregistering the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1132.070751] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5359b2ab-f5d6-4c28-96c6-cbfb4cd17dca {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.137809] env[65726]: WARNING openstack [req-a86d6fd6-aee6-4792-9710-f64a4c1c7ea5 req-d4f6c307-145d-4909-830e-4ce7991ab855 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1132.138287] env[65726]: WARNING openstack [req-a86d6fd6-aee6-4792-9710-f64a4c1c7ea5 req-d4f6c307-145d-4909-830e-4ce7991ab855 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1132.149015] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-e3b3b0a6-575e-4b59-bef3-4eef65e9267a tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] [instance: 6fd70b4a-63e4-4258-9cff-f3b582500b9f] Unregistered the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1132.149410] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-e3b3b0a6-575e-4b59-bef3-4eef65e9267a tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] [instance: 6fd70b4a-63e4-4258-9cff-f3b582500b9f] Deleting contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1132.149451] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-e3b3b0a6-575e-4b59-bef3-4eef65e9267a tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Deleting the datastore file [datastore1] 6fd70b4a-63e4-4258-9cff-f3b582500b9f {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1132.149719] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0aa167f8-e90b-4974-8c82-4a2097611245 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.156369] env[65726]: DEBUG oslo_vmware.api [None req-e3b3b0a6-575e-4b59-bef3-4eef65e9267a tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Waiting for the task: (returnval){ [ 1132.156369] env[65726]: value = "task-5116805" [ 1132.156369] env[65726]: _type = "Task" [ 1132.156369] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1132.165508] env[65726]: DEBUG oslo_vmware.api [None req-e3b3b0a6-575e-4b59-bef3-4eef65e9267a tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Task: {'id': task-5116805, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1132.186290] env[65726]: DEBUG oslo_vmware.api [None req-9b5f9575-8a6d-44da-8cc3-16504101fcc2 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Task: {'id': task-5116802, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.251171} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1132.186666] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-9b5f9575-8a6d-44da-8cc3-16504101fcc2 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Deleted the datastore file {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1132.186724] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-9b5f9575-8a6d-44da-8cc3-16504101fcc2 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] [instance: aa6ce489-c62f-4481-87b7-e74242aeb8ca] Deleted contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1132.186926] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-9b5f9575-8a6d-44da-8cc3-16504101fcc2 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] [instance: aa6ce489-c62f-4481-87b7-e74242aeb8ca] Instance destroyed {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1132.187210] env[65726]: INFO nova.compute.manager [None req-9b5f9575-8a6d-44da-8cc3-16504101fcc2 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] [instance: aa6ce489-c62f-4481-87b7-e74242aeb8ca] Took 1.14 seconds to destroy the instance on the hypervisor. [ 1132.187543] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-9b5f9575-8a6d-44da-8cc3-16504101fcc2 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1132.187761] env[65726]: DEBUG nova.compute.manager [-] [instance: aa6ce489-c62f-4481-87b7-e74242aeb8ca] Deallocating network for instance {{(pid=65726) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1132.187873] env[65726]: DEBUG nova.network.neutron [-] [instance: aa6ce489-c62f-4481-87b7-e74242aeb8ca] deallocate_for_instance() {{(pid=65726) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 1132.188423] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1132.188731] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1132.226176] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1132.226551] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1132.263386] env[65726]: DEBUG nova.network.neutron [req-a86d6fd6-aee6-4792-9710-f64a4c1c7ea5 req-d4f6c307-145d-4909-830e-4ce7991ab855 service nova] [instance: 0d42fa75-20e9-4646-9b08-17015b7f068c] Updated VIF entry in instance network info cache for port 7a95cb45-ebfc-4ac0-ad58-4bb57170e47a. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 1132.263783] env[65726]: DEBUG nova.network.neutron [req-a86d6fd6-aee6-4792-9710-f64a4c1c7ea5 req-d4f6c307-145d-4909-830e-4ce7991ab855 service nova] [instance: 0d42fa75-20e9-4646-9b08-17015b7f068c] Updating instance_info_cache with network_info: [{"id": "7a95cb45-ebfc-4ac0-ad58-4bb57170e47a", "address": "fa:16:3e:2b:ea:81", "network": {"id": "51a66448-0f72-4266-8868-856b48be839f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-155658961-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.213", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d46ccb00794f458b85da4a93879139ee", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b1f3e6c3-5584-4852-9017-476ab8ac4946", "external-id": "nsx-vlan-transportzone-304", "segmentation_id": 304, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7a95cb45-eb", "ovs_interfaceid": "7a95cb45-ebfc-4ac0-ad58-4bb57170e47a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1132.324350] env[65726]: DEBUG oslo_concurrency.lockutils [None req-96ccefc5-467e-42ef-be69-e76eb66b409b tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Lock "0361ae8f-634d-4161-a594-7f93dbb1b7cd" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 7.109s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1132.342860] env[65726]: DEBUG oslo_vmware.api [None req-a2dc650f-ed4c-4edd-a0f8-7cba2e214975 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': task-5116803, 'name': ReconfigVM_Task, 'duration_secs': 0.168752} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1132.343382] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-a2dc650f-ed4c-4edd-a0f8-7cba2e214975 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: e7156a10-631a-4cc5-a544-88ce37763d02] Reconfigured VM instance instance-00000066 to detach disk 2000 {{(pid=65726) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1132.344709] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4506f8fc-eda5-412e-b2a0-6303fa4ad8e0 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.376538] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-a2dc650f-ed4c-4edd-a0f8-7cba2e214975 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: e7156a10-631a-4cc5-a544-88ce37763d02] Reconfiguring VM instance instance-00000066 to attach disk [datastore2] e7156a10-631a-4cc5-a544-88ce37763d02/e7156a10-631a-4cc5-a544-88ce37763d02.vmdk or device None with type thin {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1132.377259] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e55d71f0-2604-43ee-ab35-f3db1b4c085b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.398263] env[65726]: DEBUG oslo_vmware.api [None req-a2dc650f-ed4c-4edd-a0f8-7cba2e214975 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Waiting for the task: (returnval){ [ 1132.398263] env[65726]: value = "task-5116806" [ 1132.398263] env[65726]: _type = "Task" [ 1132.398263] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1132.408862] env[65726]: DEBUG oslo_vmware.api [None req-a2dc650f-ed4c-4edd-a0f8-7cba2e214975 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': task-5116806, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1132.509722] env[65726]: WARNING openstack [None req-72f2a102-8ae8-4f21-a769-e63022914c54 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1132.510343] env[65726]: WARNING openstack [None req-72f2a102-8ae8-4f21-a769-e63022914c54 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1132.516233] env[65726]: DEBUG nova.network.neutron [None req-72f2a102-8ae8-4f21-a769-e63022914c54 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: 3bc25953-21b1-4729-af09-e7211fd8b2c6] Instance cache missing network info. {{(pid=65726) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 1132.541656] env[65726]: DEBUG nova.scheduler.client.report [None req-8a669c02-f06b-4fc5-8c97-9ab50c8d780e tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1132.669568] env[65726]: DEBUG oslo_vmware.api [None req-e3b3b0a6-575e-4b59-bef3-4eef65e9267a tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Task: {'id': task-5116805, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.193685} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1132.669878] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-e3b3b0a6-575e-4b59-bef3-4eef65e9267a tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Deleted the datastore file {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1132.670112] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-e3b3b0a6-575e-4b59-bef3-4eef65e9267a tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] [instance: 6fd70b4a-63e4-4258-9cff-f3b582500b9f] Deleted contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1132.670333] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-e3b3b0a6-575e-4b59-bef3-4eef65e9267a tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] [instance: 6fd70b4a-63e4-4258-9cff-f3b582500b9f] Instance destroyed {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1132.670532] env[65726]: INFO nova.compute.manager [None req-e3b3b0a6-575e-4b59-bef3-4eef65e9267a tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] [instance: 6fd70b4a-63e4-4258-9cff-f3b582500b9f] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1132.670879] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-e3b3b0a6-575e-4b59-bef3-4eef65e9267a tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1132.671094] env[65726]: DEBUG nova.compute.manager [-] [instance: 6fd70b4a-63e4-4258-9cff-f3b582500b9f] Deallocating network for instance {{(pid=65726) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1132.671273] env[65726]: DEBUG nova.network.neutron [-] [instance: 6fd70b4a-63e4-4258-9cff-f3b582500b9f] deallocate_for_instance() {{(pid=65726) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 1132.671810] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1132.672109] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1132.767561] env[65726]: DEBUG oslo_concurrency.lockutils [req-a86d6fd6-aee6-4792-9710-f64a4c1c7ea5 req-d4f6c307-145d-4909-830e-4ce7991ab855 service nova] Releasing lock "refresh_cache-0d42fa75-20e9-4646-9b08-17015b7f068c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1132.767983] env[65726]: DEBUG nova.compute.manager [req-a86d6fd6-aee6-4792-9710-f64a4c1c7ea5 req-d4f6c307-145d-4909-830e-4ce7991ab855 service nova] [instance: 4809fcf6-59d4-409c-bd8d-981dc85a686f] Received event network-vif-deleted-13050a31-7165-4b22-83cc-d8ce9a41eadf {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1132.787900] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1132.788230] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1132.802322] env[65726]: WARNING openstack [None req-72f2a102-8ae8-4f21-a769-e63022914c54 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1132.802684] env[65726]: WARNING openstack [None req-72f2a102-8ae8-4f21-a769-e63022914c54 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1132.911933] env[65726]: DEBUG oslo_vmware.api [None req-a2dc650f-ed4c-4edd-a0f8-7cba2e214975 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': task-5116806, 'name': ReconfigVM_Task, 'duration_secs': 0.427047} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1132.912442] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-a2dc650f-ed4c-4edd-a0f8-7cba2e214975 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: e7156a10-631a-4cc5-a544-88ce37763d02] Reconfigured VM instance instance-00000066 to attach disk [datastore2] e7156a10-631a-4cc5-a544-88ce37763d02/e7156a10-631a-4cc5-a544-88ce37763d02.vmdk or device None with type thin {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1132.916138] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-a2dc650f-ed4c-4edd-a0f8-7cba2e214975 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: e7156a10-631a-4cc5-a544-88ce37763d02] Updating instance 'e7156a10-631a-4cc5-a544-88ce37763d02' progress to 50 {{(pid=65726) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1133.048390] env[65726]: DEBUG oslo_concurrency.lockutils [None req-8a669c02-f06b-4fc5-8c97-9ab50c8d780e tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.824s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1133.076163] env[65726]: INFO nova.scheduler.client.report [None req-8a669c02-f06b-4fc5-8c97-9ab50c8d780e tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Deleted allocations for instance 4809fcf6-59d4-409c-bd8d-981dc85a686f [ 1133.148895] env[65726]: WARNING openstack [None req-72f2a102-8ae8-4f21-a769-e63022914c54 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1133.149301] env[65726]: WARNING openstack [None req-72f2a102-8ae8-4f21-a769-e63022914c54 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1133.161980] env[65726]: DEBUG nova.network.neutron [None req-263ff8f3-91c3-4e15-8f8e-307b6d4d0871 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: 0d42fa75-20e9-4646-9b08-17015b7f068c] Successfully updated port: fb580bad-fd60-471d-95f2-066352417659 {{(pid=65726) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 1133.195684] env[65726]: DEBUG nova.network.neutron [-] [instance: aa6ce489-c62f-4481-87b7-e74242aeb8ca] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1133.262556] env[65726]: DEBUG nova.network.neutron [None req-72f2a102-8ae8-4f21-a769-e63022914c54 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: 3bc25953-21b1-4729-af09-e7211fd8b2c6] Updating instance_info_cache with network_info: [{"id": "a7e1795d-67d0-4f32-baea-744806988e4b", "address": "fa:16:3e:06:9b:44", "network": {"id": "ad1373b5-e0ed-4ee6-a97c-af931afbab35", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-909418417-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a090d553766847e58e3231b966c92565", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5f60c972-a72d-4c5f-a250-faadfd6eafbe", "external-id": "nsx-vlan-transportzone-932", "segmentation_id": 932, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa7e1795d-67", "ovs_interfaceid": "a7e1795d-67d0-4f32-baea-744806988e4b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1133.347088] env[65726]: DEBUG nova.compute.manager [req-18f77a2f-a212-43f4-b7b1-29a2aa96e9a6 req-c048b3ff-9a15-43f0-8a2e-f09f6fd334c4 service nova] [instance: 0d42fa75-20e9-4646-9b08-17015b7f068c] Received event network-vif-plugged-fb580bad-fd60-471d-95f2-066352417659 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1133.347327] env[65726]: DEBUG oslo_concurrency.lockutils [req-18f77a2f-a212-43f4-b7b1-29a2aa96e9a6 req-c048b3ff-9a15-43f0-8a2e-f09f6fd334c4 service nova] Acquiring lock "0d42fa75-20e9-4646-9b08-17015b7f068c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1133.347537] env[65726]: DEBUG oslo_concurrency.lockutils [req-18f77a2f-a212-43f4-b7b1-29a2aa96e9a6 req-c048b3ff-9a15-43f0-8a2e-f09f6fd334c4 service nova] Lock "0d42fa75-20e9-4646-9b08-17015b7f068c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1133.347705] env[65726]: DEBUG oslo_concurrency.lockutils [req-18f77a2f-a212-43f4-b7b1-29a2aa96e9a6 req-c048b3ff-9a15-43f0-8a2e-f09f6fd334c4 service nova] Lock "0d42fa75-20e9-4646-9b08-17015b7f068c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1133.348053] env[65726]: DEBUG nova.compute.manager [req-18f77a2f-a212-43f4-b7b1-29a2aa96e9a6 req-c048b3ff-9a15-43f0-8a2e-f09f6fd334c4 service nova] [instance: 0d42fa75-20e9-4646-9b08-17015b7f068c] No waiting events found dispatching network-vif-plugged-fb580bad-fd60-471d-95f2-066352417659 {{(pid=65726) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1133.348270] env[65726]: WARNING nova.compute.manager [req-18f77a2f-a212-43f4-b7b1-29a2aa96e9a6 req-c048b3ff-9a15-43f0-8a2e-f09f6fd334c4 service nova] [instance: 0d42fa75-20e9-4646-9b08-17015b7f068c] Received unexpected event network-vif-plugged-fb580bad-fd60-471d-95f2-066352417659 for instance with vm_state active and task_state None. [ 1133.422869] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84037a4b-e33e-435e-bd1e-3edb12f80cc3 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.444344] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ee8a1dd-3dbe-4b58-8170-076e9e4de505 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.465830] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-a2dc650f-ed4c-4edd-a0f8-7cba2e214975 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: e7156a10-631a-4cc5-a544-88ce37763d02] Updating instance 'e7156a10-631a-4cc5-a544-88ce37763d02' progress to 67 {{(pid=65726) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1133.552489] env[65726]: DEBUG nova.network.neutron [-] [instance: 6fd70b4a-63e4-4258-9cff-f3b582500b9f] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1133.586972] env[65726]: DEBUG oslo_concurrency.lockutils [None req-8a669c02-f06b-4fc5-8c97-9ab50c8d780e tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Lock "4809fcf6-59d4-409c-bd8d-981dc85a686f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.065s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1133.593255] env[65726]: DEBUG nova.compute.manager [req-a5f0bfc9-cd7a-47f5-8e70-b874645fa024 req-19e45910-c980-43ba-8cce-a69279d074ce service nova] [instance: 3bc25953-21b1-4729-af09-e7211fd8b2c6] Received event network-vif-plugged-a7e1795d-67d0-4f32-baea-744806988e4b {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1133.594546] env[65726]: DEBUG oslo_concurrency.lockutils [req-a5f0bfc9-cd7a-47f5-8e70-b874645fa024 req-19e45910-c980-43ba-8cce-a69279d074ce service nova] Acquiring lock "3bc25953-21b1-4729-af09-e7211fd8b2c6-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1133.595410] env[65726]: DEBUG oslo_concurrency.lockutils [req-a5f0bfc9-cd7a-47f5-8e70-b874645fa024 req-19e45910-c980-43ba-8cce-a69279d074ce service nova] Lock "3bc25953-21b1-4729-af09-e7211fd8b2c6-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1133.595806] env[65726]: DEBUG oslo_concurrency.lockutils [req-a5f0bfc9-cd7a-47f5-8e70-b874645fa024 req-19e45910-c980-43ba-8cce-a69279d074ce service nova] Lock "3bc25953-21b1-4729-af09-e7211fd8b2c6-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1133.596416] env[65726]: DEBUG nova.compute.manager [req-a5f0bfc9-cd7a-47f5-8e70-b874645fa024 req-19e45910-c980-43ba-8cce-a69279d074ce service nova] [instance: 3bc25953-21b1-4729-af09-e7211fd8b2c6] No waiting events found dispatching network-vif-plugged-a7e1795d-67d0-4f32-baea-744806988e4b {{(pid=65726) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1133.597138] env[65726]: WARNING nova.compute.manager [req-a5f0bfc9-cd7a-47f5-8e70-b874645fa024 req-19e45910-c980-43ba-8cce-a69279d074ce service nova] [instance: 3bc25953-21b1-4729-af09-e7211fd8b2c6] Received unexpected event network-vif-plugged-a7e1795d-67d0-4f32-baea-744806988e4b for instance with vm_state building and task_state spawning. [ 1133.597783] env[65726]: DEBUG nova.compute.manager [req-a5f0bfc9-cd7a-47f5-8e70-b874645fa024 req-19e45910-c980-43ba-8cce-a69279d074ce service nova] [instance: 3bc25953-21b1-4729-af09-e7211fd8b2c6] Received event network-changed-a7e1795d-67d0-4f32-baea-744806988e4b {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1133.598159] env[65726]: DEBUG nova.compute.manager [req-a5f0bfc9-cd7a-47f5-8e70-b874645fa024 req-19e45910-c980-43ba-8cce-a69279d074ce service nova] [instance: 3bc25953-21b1-4729-af09-e7211fd8b2c6] Refreshing instance network info cache due to event network-changed-a7e1795d-67d0-4f32-baea-744806988e4b. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 1133.598901] env[65726]: DEBUG oslo_concurrency.lockutils [req-a5f0bfc9-cd7a-47f5-8e70-b874645fa024 req-19e45910-c980-43ba-8cce-a69279d074ce service nova] Acquiring lock "refresh_cache-3bc25953-21b1-4729-af09-e7211fd8b2c6" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1133.667661] env[65726]: DEBUG oslo_concurrency.lockutils [None req-263ff8f3-91c3-4e15-8f8e-307b6d4d0871 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Acquiring lock "refresh_cache-0d42fa75-20e9-4646-9b08-17015b7f068c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1133.667661] env[65726]: DEBUG oslo_concurrency.lockutils [None req-263ff8f3-91c3-4e15-8f8e-307b6d4d0871 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Acquired lock "refresh_cache-0d42fa75-20e9-4646-9b08-17015b7f068c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1133.667661] env[65726]: DEBUG nova.network.neutron [None req-263ff8f3-91c3-4e15-8f8e-307b6d4d0871 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: 0d42fa75-20e9-4646-9b08-17015b7f068c] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1133.698287] env[65726]: INFO nova.compute.manager [-] [instance: aa6ce489-c62f-4481-87b7-e74242aeb8ca] Took 1.51 seconds to deallocate network for instance. [ 1133.766775] env[65726]: DEBUG oslo_concurrency.lockutils [None req-72f2a102-8ae8-4f21-a769-e63022914c54 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Releasing lock "refresh_cache-3bc25953-21b1-4729-af09-e7211fd8b2c6" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1133.766775] env[65726]: DEBUG nova.compute.manager [None req-72f2a102-8ae8-4f21-a769-e63022914c54 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: 3bc25953-21b1-4729-af09-e7211fd8b2c6] Instance network_info: |[{"id": "a7e1795d-67d0-4f32-baea-744806988e4b", "address": "fa:16:3e:06:9b:44", "network": {"id": "ad1373b5-e0ed-4ee6-a97c-af931afbab35", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-909418417-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a090d553766847e58e3231b966c92565", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5f60c972-a72d-4c5f-a250-faadfd6eafbe", "external-id": "nsx-vlan-transportzone-932", "segmentation_id": 932, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa7e1795d-67", "ovs_interfaceid": "a7e1795d-67d0-4f32-baea-744806988e4b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1133.766775] env[65726]: DEBUG oslo_concurrency.lockutils [req-a5f0bfc9-cd7a-47f5-8e70-b874645fa024 req-19e45910-c980-43ba-8cce-a69279d074ce service nova] Acquired lock "refresh_cache-3bc25953-21b1-4729-af09-e7211fd8b2c6" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1133.767183] env[65726]: DEBUG nova.network.neutron [req-a5f0bfc9-cd7a-47f5-8e70-b874645fa024 req-19e45910-c980-43ba-8cce-a69279d074ce service nova] [instance: 3bc25953-21b1-4729-af09-e7211fd8b2c6] Refreshing network info cache for port a7e1795d-67d0-4f32-baea-744806988e4b {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 1133.768360] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-72f2a102-8ae8-4f21-a769-e63022914c54 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: 3bc25953-21b1-4729-af09-e7211fd8b2c6] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:06:9b:44', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '5f60c972-a72d-4c5f-a250-faadfd6eafbe', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a7e1795d-67d0-4f32-baea-744806988e4b', 'vif_model': 'vmxnet3'}] {{(pid=65726) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1133.778172] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-72f2a102-8ae8-4f21-a769-e63022914c54 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1133.779284] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3bc25953-21b1-4729-af09-e7211fd8b2c6] Creating VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1133.779538] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-4f1ea513-6e59-4296-aeb0-f1916675a9d4 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.801969] env[65726]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1133.801969] env[65726]: value = "task-5116808" [ 1133.801969] env[65726]: _type = "Task" [ 1133.801969] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1133.813313] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116808, 'name': CreateVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1133.974446] env[65726]: WARNING neutronclient.v2_0.client [None req-a2dc650f-ed4c-4edd-a0f8-7cba2e214975 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1134.019869] env[65726]: DEBUG nova.network.neutron [None req-a2dc650f-ed4c-4edd-a0f8-7cba2e214975 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: e7156a10-631a-4cc5-a544-88ce37763d02] Port 44a1554d-d380-49b2-ad33-9a251114c93c binding to destination host cpu-1 is already ACTIVE {{(pid=65726) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3236}} [ 1134.060540] env[65726]: INFO nova.compute.manager [-] [instance: 6fd70b4a-63e4-4258-9cff-f3b582500b9f] Took 1.39 seconds to deallocate network for instance. [ 1134.060689] env[65726]: DEBUG oslo_concurrency.lockutils [None req-668445ef-7aad-4f12-b535-5fc35338444d tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Acquiring lock "3b4d9461-747a-461c-b231-a0de02f0cb83" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1134.060912] env[65726]: DEBUG oslo_concurrency.lockutils [None req-668445ef-7aad-4f12-b535-5fc35338444d tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Lock "3b4d9461-747a-461c-b231-a0de02f0cb83" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1134.061206] env[65726]: DEBUG oslo_concurrency.lockutils [None req-668445ef-7aad-4f12-b535-5fc35338444d tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Acquiring lock "3b4d9461-747a-461c-b231-a0de02f0cb83-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1134.061402] env[65726]: DEBUG oslo_concurrency.lockutils [None req-668445ef-7aad-4f12-b535-5fc35338444d tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Lock "3b4d9461-747a-461c-b231-a0de02f0cb83-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1134.061569] env[65726]: DEBUG oslo_concurrency.lockutils [None req-668445ef-7aad-4f12-b535-5fc35338444d tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Lock "3b4d9461-747a-461c-b231-a0de02f0cb83-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1134.065321] env[65726]: INFO nova.compute.manager [None req-668445ef-7aad-4f12-b535-5fc35338444d tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] [instance: 3b4d9461-747a-461c-b231-a0de02f0cb83] Terminating instance [ 1134.172505] env[65726]: WARNING openstack [None req-263ff8f3-91c3-4e15-8f8e-307b6d4d0871 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1134.173205] env[65726]: WARNING openstack [None req-263ff8f3-91c3-4e15-8f8e-307b6d4d0871 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1134.206697] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9b5f9575-8a6d-44da-8cc3-16504101fcc2 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1134.207115] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9b5f9575-8a6d-44da-8cc3-16504101fcc2 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1134.207460] env[65726]: DEBUG nova.objects.instance [None req-9b5f9575-8a6d-44da-8cc3-16504101fcc2 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Lazy-loading 'resources' on Instance uuid aa6ce489-c62f-4481-87b7-e74242aeb8ca {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1134.220511] env[65726]: WARNING nova.network.neutron [None req-263ff8f3-91c3-4e15-8f8e-307b6d4d0871 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: 0d42fa75-20e9-4646-9b08-17015b7f068c] 51a66448-0f72-4266-8868-856b48be839f already exists in list: networks containing: ['51a66448-0f72-4266-8868-856b48be839f']. ignoring it [ 1134.259462] env[65726]: WARNING openstack [None req-263ff8f3-91c3-4e15-8f8e-307b6d4d0871 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1134.259880] env[65726]: WARNING openstack [None req-263ff8f3-91c3-4e15-8f8e-307b6d4d0871 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1134.280507] env[65726]: WARNING openstack [req-a5f0bfc9-cd7a-47f5-8e70-b874645fa024 req-19e45910-c980-43ba-8cce-a69279d074ce service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1134.280905] env[65726]: WARNING openstack [req-a5f0bfc9-cd7a-47f5-8e70-b874645fa024 req-19e45910-c980-43ba-8cce-a69279d074ce service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1134.314046] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116808, 'name': CreateVM_Task} progress is 25%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1134.328766] env[65726]: WARNING openstack [None req-263ff8f3-91c3-4e15-8f8e-307b6d4d0871 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1134.328943] env[65726]: WARNING openstack [None req-263ff8f3-91c3-4e15-8f8e-307b6d4d0871 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1134.401480] env[65726]: WARNING openstack [req-a5f0bfc9-cd7a-47f5-8e70-b874645fa024 req-19e45910-c980-43ba-8cce-a69279d074ce service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1134.402073] env[65726]: WARNING openstack [req-a5f0bfc9-cd7a-47f5-8e70-b874645fa024 req-19e45910-c980-43ba-8cce-a69279d074ce service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1134.464790] env[65726]: WARNING openstack [None req-263ff8f3-91c3-4e15-8f8e-307b6d4d0871 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1134.465469] env[65726]: WARNING openstack [None req-263ff8f3-91c3-4e15-8f8e-307b6d4d0871 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1134.534910] env[65726]: WARNING openstack [req-a5f0bfc9-cd7a-47f5-8e70-b874645fa024 req-19e45910-c980-43ba-8cce-a69279d074ce service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1134.535306] env[65726]: WARNING openstack [req-a5f0bfc9-cd7a-47f5-8e70-b874645fa024 req-19e45910-c980-43ba-8cce-a69279d074ce service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1134.546883] env[65726]: WARNING openstack [None req-263ff8f3-91c3-4e15-8f8e-307b6d4d0871 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1134.547270] env[65726]: WARNING openstack [None req-263ff8f3-91c3-4e15-8f8e-307b6d4d0871 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1134.569197] env[65726]: DEBUG nova.compute.manager [None req-668445ef-7aad-4f12-b535-5fc35338444d tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] [instance: 3b4d9461-747a-461c-b231-a0de02f0cb83] Start destroying the instance on the hypervisor. {{(pid=65726) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1134.569462] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-668445ef-7aad-4f12-b535-5fc35338444d tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] [instance: 3b4d9461-747a-461c-b231-a0de02f0cb83] Destroying instance {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1134.570447] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afbff651-2801-4365-b9c2-f5e634e886b6 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.575512] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e3b3b0a6-575e-4b59-bef3-4eef65e9267a tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1134.581408] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-668445ef-7aad-4f12-b535-5fc35338444d tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] [instance: 3b4d9461-747a-461c-b231-a0de02f0cb83] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1134.581593] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-cc513453-ed35-4ff4-bac3-77f8542a34e6 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.589405] env[65726]: DEBUG oslo_vmware.api [None req-668445ef-7aad-4f12-b535-5fc35338444d tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Waiting for the task: (returnval){ [ 1134.589405] env[65726]: value = "task-5116809" [ 1134.589405] env[65726]: _type = "Task" [ 1134.589405] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1134.604629] env[65726]: DEBUG oslo_vmware.api [None req-668445ef-7aad-4f12-b535-5fc35338444d tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Task: {'id': task-5116809, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1134.642549] env[65726]: DEBUG nova.network.neutron [req-a5f0bfc9-cd7a-47f5-8e70-b874645fa024 req-19e45910-c980-43ba-8cce-a69279d074ce service nova] [instance: 3bc25953-21b1-4729-af09-e7211fd8b2c6] Updated VIF entry in instance network info cache for port a7e1795d-67d0-4f32-baea-744806988e4b. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 1134.642952] env[65726]: DEBUG nova.network.neutron [req-a5f0bfc9-cd7a-47f5-8e70-b874645fa024 req-19e45910-c980-43ba-8cce-a69279d074ce service nova] [instance: 3bc25953-21b1-4729-af09-e7211fd8b2c6] Updating instance_info_cache with network_info: [{"id": "a7e1795d-67d0-4f32-baea-744806988e4b", "address": "fa:16:3e:06:9b:44", "network": {"id": "ad1373b5-e0ed-4ee6-a97c-af931afbab35", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-909418417-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a090d553766847e58e3231b966c92565", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5f60c972-a72d-4c5f-a250-faadfd6eafbe", "external-id": "nsx-vlan-transportzone-932", "segmentation_id": 932, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa7e1795d-67", "ovs_interfaceid": "a7e1795d-67d0-4f32-baea-744806988e4b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1134.665786] env[65726]: DEBUG nova.network.neutron [None req-263ff8f3-91c3-4e15-8f8e-307b6d4d0871 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: 0d42fa75-20e9-4646-9b08-17015b7f068c] Updating instance_info_cache with network_info: [{"id": "7a95cb45-ebfc-4ac0-ad58-4bb57170e47a", "address": "fa:16:3e:2b:ea:81", "network": {"id": "51a66448-0f72-4266-8868-856b48be839f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-155658961-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.213", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d46ccb00794f458b85da4a93879139ee", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b1f3e6c3-5584-4852-9017-476ab8ac4946", "external-id": "nsx-vlan-transportzone-304", "segmentation_id": 304, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7a95cb45-eb", "ovs_interfaceid": "7a95cb45-ebfc-4ac0-ad58-4bb57170e47a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "fb580bad-fd60-471d-95f2-066352417659", "address": "fa:16:3e:32:0d:d6", "network": {"id": "51a66448-0f72-4266-8868-856b48be839f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-155658961-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d46ccb00794f458b85da4a93879139ee", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b1f3e6c3-5584-4852-9017-476ab8ac4946", "external-id": "nsx-vlan-transportzone-304", "segmentation_id": 304, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfb580bad-fd", "ovs_interfaceid": "fb580bad-fd60-471d-95f2-066352417659", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1134.821293] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116808, 'name': CreateVM_Task, 'duration_secs': 0.736097} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1134.821293] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3bc25953-21b1-4729-af09-e7211fd8b2c6] Created VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1134.825020] env[65726]: WARNING openstack [None req-72f2a102-8ae8-4f21-a769-e63022914c54 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1134.825020] env[65726]: WARNING openstack [None req-72f2a102-8ae8-4f21-a769-e63022914c54 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1134.827882] env[65726]: DEBUG oslo_concurrency.lockutils [None req-72f2a102-8ae8-4f21-a769-e63022914c54 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1134.828118] env[65726]: DEBUG oslo_concurrency.lockutils [None req-72f2a102-8ae8-4f21-a769-e63022914c54 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1134.828545] env[65726]: DEBUG oslo_concurrency.lockutils [None req-72f2a102-8ae8-4f21-a769-e63022914c54 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1134.828898] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c90ec191-95df-4cd4-beda-5535f4f5971d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.837413] env[65726]: DEBUG oslo_vmware.api [None req-72f2a102-8ae8-4f21-a769-e63022914c54 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Waiting for the task: (returnval){ [ 1134.837413] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52da8b9a-d4d9-b703-2781-523d247d6251" [ 1134.837413] env[65726]: _type = "Task" [ 1134.837413] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1134.851616] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a7d158c8-609b-4687-83b0-c5b67c744acc tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Acquiring lock "0361ae8f-634d-4161-a594-7f93dbb1b7cd" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1134.851891] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a7d158c8-609b-4687-83b0-c5b67c744acc tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Lock "0361ae8f-634d-4161-a594-7f93dbb1b7cd" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1134.852121] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a7d158c8-609b-4687-83b0-c5b67c744acc tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Acquiring lock "0361ae8f-634d-4161-a594-7f93dbb1b7cd-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1134.852322] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a7d158c8-609b-4687-83b0-c5b67c744acc tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Lock "0361ae8f-634d-4161-a594-7f93dbb1b7cd-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1134.852487] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a7d158c8-609b-4687-83b0-c5b67c744acc tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Lock "0361ae8f-634d-4161-a594-7f93dbb1b7cd-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1134.854482] env[65726]: DEBUG oslo_vmware.api [None req-72f2a102-8ae8-4f21-a769-e63022914c54 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52da8b9a-d4d9-b703-2781-523d247d6251, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1134.855462] env[65726]: INFO nova.compute.manager [None req-a7d158c8-609b-4687-83b0-c5b67c744acc tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 0361ae8f-634d-4161-a594-7f93dbb1b7cd] Terminating instance [ 1134.922918] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37e90871-0ccf-48da-ac69-f729389f7513 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.932348] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43942bf7-6416-4d45-9385-5f9f158dd5c2 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.968930] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9504252a-a062-4f7b-915f-19e6b6ce6c6b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.978922] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3073605f-dbe8-46ce-8da3-2645ff8970ea {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.994779] env[65726]: DEBUG nova.compute.provider_tree [None req-9b5f9575-8a6d-44da-8cc3-16504101fcc2 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1135.044925] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a2dc650f-ed4c-4edd-a0f8-7cba2e214975 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Acquiring lock "e7156a10-631a-4cc5-a544-88ce37763d02-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1135.045126] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a2dc650f-ed4c-4edd-a0f8-7cba2e214975 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Lock "e7156a10-631a-4cc5-a544-88ce37763d02-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1135.045337] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a2dc650f-ed4c-4edd-a0f8-7cba2e214975 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Lock "e7156a10-631a-4cc5-a544-88ce37763d02-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1135.101050] env[65726]: DEBUG oslo_vmware.api [None req-668445ef-7aad-4f12-b535-5fc35338444d tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Task: {'id': task-5116809, 'name': PowerOffVM_Task, 'duration_secs': 0.443414} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1135.101050] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-668445ef-7aad-4f12-b535-5fc35338444d tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] [instance: 3b4d9461-747a-461c-b231-a0de02f0cb83] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1135.101050] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-668445ef-7aad-4f12-b535-5fc35338444d tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] [instance: 3b4d9461-747a-461c-b231-a0de02f0cb83] Unregistering the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1135.101050] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-45943ecd-6c68-4e73-ba37-49a9c0dd79bc {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.147761] env[65726]: DEBUG oslo_concurrency.lockutils [req-a5f0bfc9-cd7a-47f5-8e70-b874645fa024 req-19e45910-c980-43ba-8cce-a69279d074ce service nova] Releasing lock "refresh_cache-3bc25953-21b1-4729-af09-e7211fd8b2c6" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1135.148317] env[65726]: DEBUG nova.compute.manager [req-a5f0bfc9-cd7a-47f5-8e70-b874645fa024 req-19e45910-c980-43ba-8cce-a69279d074ce service nova] [instance: aa6ce489-c62f-4481-87b7-e74242aeb8ca] Received event network-vif-deleted-d5a805f5-b6a8-42ec-92a0-4a652b51c1e6 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1135.148317] env[65726]: DEBUG nova.compute.manager [req-a5f0bfc9-cd7a-47f5-8e70-b874645fa024 req-19e45910-c980-43ba-8cce-a69279d074ce service nova] [instance: 6fd70b4a-63e4-4258-9cff-f3b582500b9f] Received event network-vif-deleted-153b4c9d-d01d-4254-9aa6-040705be347a {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1135.169317] env[65726]: DEBUG oslo_concurrency.lockutils [None req-263ff8f3-91c3-4e15-8f8e-307b6d4d0871 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Releasing lock "refresh_cache-0d42fa75-20e9-4646-9b08-17015b7f068c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1135.170050] env[65726]: DEBUG oslo_concurrency.lockutils [None req-263ff8f3-91c3-4e15-8f8e-307b6d4d0871 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Acquiring lock "0d42fa75-20e9-4646-9b08-17015b7f068c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1135.170210] env[65726]: DEBUG oslo_concurrency.lockutils [None req-263ff8f3-91c3-4e15-8f8e-307b6d4d0871 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Acquired lock "0d42fa75-20e9-4646-9b08-17015b7f068c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1135.171181] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-044e305b-1d43-47d1-ad80-590ed48f4fea {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.190372] env[65726]: DEBUG nova.virt.hardware [None req-263ff8f3-91c3-4e15-8f8e-307b6d4d0871 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1135.190638] env[65726]: DEBUG nova.virt.hardware [None req-263ff8f3-91c3-4e15-8f8e-307b6d4d0871 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1135.190813] env[65726]: DEBUG nova.virt.hardware [None req-263ff8f3-91c3-4e15-8f8e-307b6d4d0871 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1135.191121] env[65726]: DEBUG nova.virt.hardware [None req-263ff8f3-91c3-4e15-8f8e-307b6d4d0871 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1135.191287] env[65726]: DEBUG nova.virt.hardware [None req-263ff8f3-91c3-4e15-8f8e-307b6d4d0871 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1135.191436] env[65726]: DEBUG nova.virt.hardware [None req-263ff8f3-91c3-4e15-8f8e-307b6d4d0871 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1135.191648] env[65726]: DEBUG nova.virt.hardware [None req-263ff8f3-91c3-4e15-8f8e-307b6d4d0871 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1135.191804] env[65726]: DEBUG nova.virt.hardware [None req-263ff8f3-91c3-4e15-8f8e-307b6d4d0871 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1135.191969] env[65726]: DEBUG nova.virt.hardware [None req-263ff8f3-91c3-4e15-8f8e-307b6d4d0871 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1135.192143] env[65726]: DEBUG nova.virt.hardware [None req-263ff8f3-91c3-4e15-8f8e-307b6d4d0871 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1135.192321] env[65726]: DEBUG nova.virt.hardware [None req-263ff8f3-91c3-4e15-8f8e-307b6d4d0871 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1135.201611] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-263ff8f3-91c3-4e15-8f8e-307b6d4d0871 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: 0d42fa75-20e9-4646-9b08-17015b7f068c] Reconfiguring VM to attach interface {{(pid=65726) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 1135.203687] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d08006bd-0991-47a4-8fbc-798599fbe242 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.221556] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-668445ef-7aad-4f12-b535-5fc35338444d tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] [instance: 3b4d9461-747a-461c-b231-a0de02f0cb83] Unregistered the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1135.221782] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-668445ef-7aad-4f12-b535-5fc35338444d tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] [instance: 3b4d9461-747a-461c-b231-a0de02f0cb83] Deleting contents of the VM from datastore datastore2 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1135.221961] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-668445ef-7aad-4f12-b535-5fc35338444d tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Deleting the datastore file [datastore2] 3b4d9461-747a-461c-b231-a0de02f0cb83 {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1135.222274] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2a8e5462-1a8f-4c8b-9954-bb1473a661eb {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.230819] env[65726]: DEBUG oslo_vmware.api [None req-668445ef-7aad-4f12-b535-5fc35338444d tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Waiting for the task: (returnval){ [ 1135.230819] env[65726]: value = "task-5116812" [ 1135.230819] env[65726]: _type = "Task" [ 1135.230819] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1135.232810] env[65726]: DEBUG oslo_vmware.api [None req-263ff8f3-91c3-4e15-8f8e-307b6d4d0871 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Waiting for the task: (returnval){ [ 1135.232810] env[65726]: value = "task-5116813" [ 1135.232810] env[65726]: _type = "Task" [ 1135.232810] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1135.245341] env[65726]: DEBUG oslo_vmware.api [None req-263ff8f3-91c3-4e15-8f8e-307b6d4d0871 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': task-5116813, 'name': ReconfigVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1135.248611] env[65726]: DEBUG oslo_vmware.api [None req-668445ef-7aad-4f12-b535-5fc35338444d tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Task: {'id': task-5116812, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1135.349259] env[65726]: DEBUG oslo_vmware.api [None req-72f2a102-8ae8-4f21-a769-e63022914c54 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52da8b9a-d4d9-b703-2781-523d247d6251, 'name': SearchDatastore_Task, 'duration_secs': 0.018076} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1135.349567] env[65726]: DEBUG oslo_concurrency.lockutils [None req-72f2a102-8ae8-4f21-a769-e63022914c54 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1135.349809] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-72f2a102-8ae8-4f21-a769-e63022914c54 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: 3bc25953-21b1-4729-af09-e7211fd8b2c6] Processing image b52362a3-ee8a-4cbf-b06f-513b0cc8f95c {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1135.350068] env[65726]: DEBUG oslo_concurrency.lockutils [None req-72f2a102-8ae8-4f21-a769-e63022914c54 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1135.350221] env[65726]: DEBUG oslo_concurrency.lockutils [None req-72f2a102-8ae8-4f21-a769-e63022914c54 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1135.350406] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-72f2a102-8ae8-4f21-a769-e63022914c54 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1135.350703] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2e59d208-b994-4589-8535-184f5666c4e9 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.363665] env[65726]: DEBUG nova.compute.manager [None req-a7d158c8-609b-4687-83b0-c5b67c744acc tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 0361ae8f-634d-4161-a594-7f93dbb1b7cd] Start destroying the instance on the hypervisor. {{(pid=65726) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1135.363982] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-a7d158c8-609b-4687-83b0-c5b67c744acc tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 0361ae8f-634d-4161-a594-7f93dbb1b7cd] Destroying instance {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1135.366474] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a12abf77-27f5-472d-a7b7-59e568445ec3 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.370159] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-72f2a102-8ae8-4f21-a769-e63022914c54 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1135.370391] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-72f2a102-8ae8-4f21-a769-e63022914c54 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65726) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1135.371232] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c6156aa2-8dce-43b7-9e50-32e465f8ce14 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.378557] env[65726]: DEBUG oslo_vmware.api [None req-72f2a102-8ae8-4f21-a769-e63022914c54 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Waiting for the task: (returnval){ [ 1135.378557] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5249b7ea-4e9c-e2fd-8e36-174f439ab63c" [ 1135.378557] env[65726]: _type = "Task" [ 1135.378557] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1135.381937] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-a7d158c8-609b-4687-83b0-c5b67c744acc tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 0361ae8f-634d-4161-a594-7f93dbb1b7cd] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1135.386512] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e88157fa-6869-4152-a0b1-e6b7afdfe49c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.397051] env[65726]: DEBUG oslo_vmware.api [None req-72f2a102-8ae8-4f21-a769-e63022914c54 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5249b7ea-4e9c-e2fd-8e36-174f439ab63c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1135.398183] env[65726]: DEBUG oslo_vmware.api [None req-a7d158c8-609b-4687-83b0-c5b67c744acc tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Waiting for the task: (returnval){ [ 1135.398183] env[65726]: value = "task-5116814" [ 1135.398183] env[65726]: _type = "Task" [ 1135.398183] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1135.408579] env[65726]: DEBUG oslo_vmware.api [None req-a7d158c8-609b-4687-83b0-c5b67c744acc tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': task-5116814, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1135.498220] env[65726]: DEBUG nova.scheduler.client.report [None req-9b5f9575-8a6d-44da-8cc3-16504101fcc2 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1135.724044] env[65726]: DEBUG nova.compute.manager [req-121a2eef-9487-4e80-aa0e-f4640b6bcf62 req-cd6ee88f-0971-418a-8eb4-83e91bffc982 service nova] [instance: 0d42fa75-20e9-4646-9b08-17015b7f068c] Received event network-changed-fb580bad-fd60-471d-95f2-066352417659 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1135.724044] env[65726]: DEBUG nova.compute.manager [req-121a2eef-9487-4e80-aa0e-f4640b6bcf62 req-cd6ee88f-0971-418a-8eb4-83e91bffc982 service nova] [instance: 0d42fa75-20e9-4646-9b08-17015b7f068c] Refreshing instance network info cache due to event network-changed-fb580bad-fd60-471d-95f2-066352417659. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 1135.724296] env[65726]: DEBUG oslo_concurrency.lockutils [req-121a2eef-9487-4e80-aa0e-f4640b6bcf62 req-cd6ee88f-0971-418a-8eb4-83e91bffc982 service nova] Acquiring lock "refresh_cache-0d42fa75-20e9-4646-9b08-17015b7f068c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1135.724296] env[65726]: DEBUG oslo_concurrency.lockutils [req-121a2eef-9487-4e80-aa0e-f4640b6bcf62 req-cd6ee88f-0971-418a-8eb4-83e91bffc982 service nova] Acquired lock "refresh_cache-0d42fa75-20e9-4646-9b08-17015b7f068c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1135.724688] env[65726]: DEBUG nova.network.neutron [req-121a2eef-9487-4e80-aa0e-f4640b6bcf62 req-cd6ee88f-0971-418a-8eb4-83e91bffc982 service nova] [instance: 0d42fa75-20e9-4646-9b08-17015b7f068c] Refreshing network info cache for port fb580bad-fd60-471d-95f2-066352417659 {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 1135.749614] env[65726]: DEBUG oslo_vmware.api [None req-668445ef-7aad-4f12-b535-5fc35338444d tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Task: {'id': task-5116812, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.269225} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1135.753061] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-668445ef-7aad-4f12-b535-5fc35338444d tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Deleted the datastore file {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1135.753352] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-668445ef-7aad-4f12-b535-5fc35338444d tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] [instance: 3b4d9461-747a-461c-b231-a0de02f0cb83] Deleted contents of the VM from datastore datastore2 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1135.753557] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-668445ef-7aad-4f12-b535-5fc35338444d tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] [instance: 3b4d9461-747a-461c-b231-a0de02f0cb83] Instance destroyed {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1135.753734] env[65726]: INFO nova.compute.manager [None req-668445ef-7aad-4f12-b535-5fc35338444d tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] [instance: 3b4d9461-747a-461c-b231-a0de02f0cb83] Took 1.18 seconds to destroy the instance on the hypervisor. [ 1135.753981] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-668445ef-7aad-4f12-b535-5fc35338444d tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1135.754219] env[65726]: DEBUG oslo_vmware.api [None req-263ff8f3-91c3-4e15-8f8e-307b6d4d0871 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': task-5116813, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1135.754723] env[65726]: DEBUG nova.compute.manager [-] [instance: 3b4d9461-747a-461c-b231-a0de02f0cb83] Deallocating network for instance {{(pid=65726) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1135.754834] env[65726]: DEBUG nova.network.neutron [-] [instance: 3b4d9461-747a-461c-b231-a0de02f0cb83] deallocate_for_instance() {{(pid=65726) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 1135.755401] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1135.755692] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1135.795985] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1135.796286] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1135.902484] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9230f0c2-55f9-4dd0-82d8-cd3277e657f9 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Acquiring lock "7c1188ce-8718-4719-8631-e59e7915b7aa" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1135.902484] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9230f0c2-55f9-4dd0-82d8-cd3277e657f9 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Lock "7c1188ce-8718-4719-8631-e59e7915b7aa" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1135.903901] env[65726]: DEBUG oslo_vmware.api [None req-72f2a102-8ae8-4f21-a769-e63022914c54 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5249b7ea-4e9c-e2fd-8e36-174f439ab63c, 'name': SearchDatastore_Task, 'duration_secs': 0.02502} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1135.911728] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bc1202cf-c05d-4bd4-9e1d-aad47609723a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.928254] env[65726]: DEBUG oslo_vmware.api [None req-a7d158c8-609b-4687-83b0-c5b67c744acc tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': task-5116814, 'name': PowerOffVM_Task, 'duration_secs': 0.255806} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1135.928710] env[65726]: DEBUG oslo_vmware.api [None req-72f2a102-8ae8-4f21-a769-e63022914c54 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Waiting for the task: (returnval){ [ 1135.928710] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52b6a571-b3fd-89d6-cda6-cc9324464c5f" [ 1135.928710] env[65726]: _type = "Task" [ 1135.928710] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1135.929092] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-a7d158c8-609b-4687-83b0-c5b67c744acc tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 0361ae8f-634d-4161-a594-7f93dbb1b7cd] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1135.929419] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-a7d158c8-609b-4687-83b0-c5b67c744acc tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 0361ae8f-634d-4161-a594-7f93dbb1b7cd] Unregistering the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1135.929801] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b585d750-e174-4b17-882f-6b881c5f3e16 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.942965] env[65726]: DEBUG oslo_vmware.api [None req-72f2a102-8ae8-4f21-a769-e63022914c54 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52b6a571-b3fd-89d6-cda6-cc9324464c5f, 'name': SearchDatastore_Task, 'duration_secs': 0.018682} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1135.943378] env[65726]: DEBUG oslo_concurrency.lockutils [None req-72f2a102-8ae8-4f21-a769-e63022914c54 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1135.943798] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-72f2a102-8ae8-4f21-a769-e63022914c54 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] 3bc25953-21b1-4729-af09-e7211fd8b2c6/3bc25953-21b1-4729-af09-e7211fd8b2c6.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1135.944245] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-427046ec-1875-4e10-98fe-720ed55a0e0e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.951806] env[65726]: DEBUG oslo_vmware.api [None req-72f2a102-8ae8-4f21-a769-e63022914c54 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Waiting for the task: (returnval){ [ 1135.951806] env[65726]: value = "task-5116816" [ 1135.951806] env[65726]: _type = "Task" [ 1135.951806] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1135.961710] env[65726]: DEBUG oslo_vmware.api [None req-72f2a102-8ae8-4f21-a769-e63022914c54 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Task: {'id': task-5116816, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1136.004633] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9b5f9575-8a6d-44da-8cc3-16504101fcc2 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.797s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1136.008640] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e3b3b0a6-575e-4b59-bef3-4eef65e9267a tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.433s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1136.008902] env[65726]: DEBUG nova.objects.instance [None req-e3b3b0a6-575e-4b59-bef3-4eef65e9267a tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Lazy-loading 'resources' on Instance uuid 6fd70b4a-63e4-4258-9cff-f3b582500b9f {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1136.010327] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-a7d158c8-609b-4687-83b0-c5b67c744acc tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 0361ae8f-634d-4161-a594-7f93dbb1b7cd] Unregistered the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1136.010920] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-a7d158c8-609b-4687-83b0-c5b67c744acc tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 0361ae8f-634d-4161-a594-7f93dbb1b7cd] Deleting contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1136.010920] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-a7d158c8-609b-4687-83b0-c5b67c744acc tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Deleting the datastore file [datastore1] 0361ae8f-634d-4161-a594-7f93dbb1b7cd {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1136.011213] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-057d6791-ece9-469f-b670-ba575521b7d0 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.025210] env[65726]: DEBUG oslo_vmware.api [None req-a7d158c8-609b-4687-83b0-c5b67c744acc tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Waiting for the task: (returnval){ [ 1136.025210] env[65726]: value = "task-5116817" [ 1136.025210] env[65726]: _type = "Task" [ 1136.025210] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1136.038344] env[65726]: DEBUG oslo_vmware.api [None req-a7d158c8-609b-4687-83b0-c5b67c744acc tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': task-5116817, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1136.041479] env[65726]: INFO nova.scheduler.client.report [None req-9b5f9575-8a6d-44da-8cc3-16504101fcc2 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Deleted allocations for instance aa6ce489-c62f-4481-87b7-e74242aeb8ca [ 1136.050251] env[65726]: DEBUG nova.compute.manager [req-e1bde330-7594-429a-8c2f-e8541912c789 req-0f5ea3b0-09a2-4ebd-bfca-6864a4956279 service nova] [instance: 3b4d9461-747a-461c-b231-a0de02f0cb83] Received event network-vif-deleted-20c6b53e-2740-46bc-a94f-59a15f8b2c61 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1136.050595] env[65726]: INFO nova.compute.manager [req-e1bde330-7594-429a-8c2f-e8541912c789 req-0f5ea3b0-09a2-4ebd-bfca-6864a4956279 service nova] [instance: 3b4d9461-747a-461c-b231-a0de02f0cb83] Neutron deleted interface 20c6b53e-2740-46bc-a94f-59a15f8b2c61; detaching it from the instance and deleting it from the info cache [ 1136.050881] env[65726]: DEBUG nova.network.neutron [req-e1bde330-7594-429a-8c2f-e8541912c789 req-0f5ea3b0-09a2-4ebd-bfca-6864a4956279 service nova] [instance: 3b4d9461-747a-461c-b231-a0de02f0cb83] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1136.054743] env[65726]: WARNING openstack [None req-a2dc650f-ed4c-4edd-a0f8-7cba2e214975 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1136.055335] env[65726]: WARNING openstack [None req-a2dc650f-ed4c-4edd-a0f8-7cba2e214975 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1136.103502] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a2dc650f-ed4c-4edd-a0f8-7cba2e214975 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Acquiring lock "refresh_cache-e7156a10-631a-4cc5-a544-88ce37763d02" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1136.103703] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a2dc650f-ed4c-4edd-a0f8-7cba2e214975 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Acquired lock "refresh_cache-e7156a10-631a-4cc5-a544-88ce37763d02" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1136.103886] env[65726]: DEBUG nova.network.neutron [None req-a2dc650f-ed4c-4edd-a0f8-7cba2e214975 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: e7156a10-631a-4cc5-a544-88ce37763d02] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1136.227979] env[65726]: WARNING openstack [req-121a2eef-9487-4e80-aa0e-f4640b6bcf62 req-cd6ee88f-0971-418a-8eb4-83e91bffc982 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1136.228382] env[65726]: WARNING openstack [req-121a2eef-9487-4e80-aa0e-f4640b6bcf62 req-cd6ee88f-0971-418a-8eb4-83e91bffc982 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1136.251445] env[65726]: DEBUG oslo_vmware.api [None req-263ff8f3-91c3-4e15-8f8e-307b6d4d0871 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': task-5116813, 'name': ReconfigVM_Task, 'duration_secs': 0.69503} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1136.252361] env[65726]: WARNING openstack [None req-263ff8f3-91c3-4e15-8f8e-307b6d4d0871 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1136.252725] env[65726]: WARNING openstack [None req-263ff8f3-91c3-4e15-8f8e-307b6d4d0871 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1136.259330] env[65726]: DEBUG oslo_concurrency.lockutils [None req-263ff8f3-91c3-4e15-8f8e-307b6d4d0871 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Releasing lock "0d42fa75-20e9-4646-9b08-17015b7f068c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1136.259752] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-263ff8f3-91c3-4e15-8f8e-307b6d4d0871 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: 0d42fa75-20e9-4646-9b08-17015b7f068c] Reconfigured VM to attach interface {{(pid=65726) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 1136.367958] env[65726]: WARNING openstack [req-121a2eef-9487-4e80-aa0e-f4640b6bcf62 req-cd6ee88f-0971-418a-8eb4-83e91bffc982 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1136.368473] env[65726]: WARNING openstack [req-121a2eef-9487-4e80-aa0e-f4640b6bcf62 req-cd6ee88f-0971-418a-8eb4-83e91bffc982 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1136.406129] env[65726]: DEBUG nova.compute.manager [None req-9230f0c2-55f9-4dd0-82d8-cd3277e657f9 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 7c1188ce-8718-4719-8631-e59e7915b7aa] Starting instance... {{(pid=65726) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 1136.467109] env[65726]: DEBUG oslo_vmware.api [None req-72f2a102-8ae8-4f21-a769-e63022914c54 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Task: {'id': task-5116816, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1136.476962] env[65726]: WARNING openstack [req-121a2eef-9487-4e80-aa0e-f4640b6bcf62 req-cd6ee88f-0971-418a-8eb4-83e91bffc982 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1136.477391] env[65726]: WARNING openstack [req-121a2eef-9487-4e80-aa0e-f4640b6bcf62 req-cd6ee88f-0971-418a-8eb4-83e91bffc982 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1136.526550] env[65726]: DEBUG nova.network.neutron [-] [instance: 3b4d9461-747a-461c-b231-a0de02f0cb83] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1136.541672] env[65726]: DEBUG oslo_vmware.api [None req-a7d158c8-609b-4687-83b0-c5b67c744acc tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': task-5116817, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.517912} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1136.547582] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-a7d158c8-609b-4687-83b0-c5b67c744acc tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Deleted the datastore file {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1136.548082] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-a7d158c8-609b-4687-83b0-c5b67c744acc tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 0361ae8f-634d-4161-a594-7f93dbb1b7cd] Deleted contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1136.548082] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-a7d158c8-609b-4687-83b0-c5b67c744acc tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 0361ae8f-634d-4161-a594-7f93dbb1b7cd] Instance destroyed {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1136.548289] env[65726]: INFO nova.compute.manager [None req-a7d158c8-609b-4687-83b0-c5b67c744acc tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 0361ae8f-634d-4161-a594-7f93dbb1b7cd] Took 1.18 seconds to destroy the instance on the hypervisor. [ 1136.548613] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-a7d158c8-609b-4687-83b0-c5b67c744acc tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1136.552128] env[65726]: DEBUG nova.compute.manager [-] [instance: 0361ae8f-634d-4161-a594-7f93dbb1b7cd] Deallocating network for instance {{(pid=65726) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1136.552302] env[65726]: DEBUG nova.network.neutron [-] [instance: 0361ae8f-634d-4161-a594-7f93dbb1b7cd] deallocate_for_instance() {{(pid=65726) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 1136.553084] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1136.553084] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1136.564964] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9b5f9575-8a6d-44da-8cc3-16504101fcc2 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Lock "aa6ce489-c62f-4481-87b7-e74242aeb8ca" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.028s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1136.566238] env[65726]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-49c165b0-87a5-4c15-a4c1-0aecfd2dd90a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.570459] env[65726]: DEBUG oslo_service.periodic_task [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=65726) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1136.570459] env[65726]: DEBUG oslo_service.periodic_task [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=65726) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1136.570459] env[65726]: DEBUG oslo_service.periodic_task [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=65726) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1136.570459] env[65726]: DEBUG oslo_service.periodic_task [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=65726) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1136.570592] env[65726]: DEBUG oslo_service.periodic_task [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=65726) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1136.570730] env[65726]: DEBUG oslo_service.periodic_task [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=65726) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1136.571066] env[65726]: DEBUG oslo_service.periodic_task [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=65726) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1136.571066] env[65726]: DEBUG nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=65726) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11251}} [ 1136.571805] env[65726]: DEBUG oslo_service.periodic_task [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Running periodic task ComputeManager.update_available_resource {{(pid=65726) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1136.583730] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1fd86ca1-0f8c-4f4c-aa6c-231d389f2b37 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.604050] env[65726]: DEBUG nova.network.neutron [req-121a2eef-9487-4e80-aa0e-f4640b6bcf62 req-cd6ee88f-0971-418a-8eb4-83e91bffc982 service nova] [instance: 0d42fa75-20e9-4646-9b08-17015b7f068c] Updated VIF entry in instance network info cache for port fb580bad-fd60-471d-95f2-066352417659. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 1136.605042] env[65726]: DEBUG nova.network.neutron [req-121a2eef-9487-4e80-aa0e-f4640b6bcf62 req-cd6ee88f-0971-418a-8eb4-83e91bffc982 service nova] [instance: 0d42fa75-20e9-4646-9b08-17015b7f068c] Updating instance_info_cache with network_info: [{"id": "7a95cb45-ebfc-4ac0-ad58-4bb57170e47a", "address": "fa:16:3e:2b:ea:81", "network": {"id": "51a66448-0f72-4266-8868-856b48be839f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-155658961-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.213", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d46ccb00794f458b85da4a93879139ee", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b1f3e6c3-5584-4852-9017-476ab8ac4946", "external-id": "nsx-vlan-transportzone-304", "segmentation_id": 304, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7a95cb45-eb", "ovs_interfaceid": "7a95cb45-ebfc-4ac0-ad58-4bb57170e47a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "fb580bad-fd60-471d-95f2-066352417659", "address": "fa:16:3e:32:0d:d6", "network": {"id": "51a66448-0f72-4266-8868-856b48be839f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-155658961-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d46ccb00794f458b85da4a93879139ee", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b1f3e6c3-5584-4852-9017-476ab8ac4946", "external-id": "nsx-vlan-transportzone-304", "segmentation_id": 304, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfb580bad-fd", "ovs_interfaceid": "fb580bad-fd60-471d-95f2-066352417659", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1136.609442] env[65726]: WARNING openstack [None req-a2dc650f-ed4c-4edd-a0f8-7cba2e214975 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1136.610561] env[65726]: WARNING openstack [None req-a2dc650f-ed4c-4edd-a0f8-7cba2e214975 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1136.622568] env[65726]: DEBUG oslo_concurrency.lockutils [req-121a2eef-9487-4e80-aa0e-f4640b6bcf62 req-cd6ee88f-0971-418a-8eb4-83e91bffc982 service nova] Releasing lock "refresh_cache-0d42fa75-20e9-4646-9b08-17015b7f068c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1136.625557] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1136.625829] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1136.658672] env[65726]: DEBUG nova.compute.manager [req-e1bde330-7594-429a-8c2f-e8541912c789 req-0f5ea3b0-09a2-4ebd-bfca-6864a4956279 service nova] [instance: 3b4d9461-747a-461c-b231-a0de02f0cb83] Detach interface failed, port_id=20c6b53e-2740-46bc-a94f-59a15f8b2c61, reason: Instance 3b4d9461-747a-461c-b231-a0de02f0cb83 could not be found. {{(pid=65726) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11669}} [ 1136.743215] env[65726]: WARNING openstack [None req-a2dc650f-ed4c-4edd-a0f8-7cba2e214975 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1136.743672] env[65726]: WARNING openstack [None req-a2dc650f-ed4c-4edd-a0f8-7cba2e214975 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1136.765384] env[65726]: DEBUG oslo_concurrency.lockutils [None req-263ff8f3-91c3-4e15-8f8e-307b6d4d0871 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Lock "interface-0d42fa75-20e9-4646-9b08-17015b7f068c-fb580bad-fd60-471d-95f2-066352417659" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 6.584s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1136.767140] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99101ba7-4006-48b4-bdeb-0225137e7107 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.781359] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27c6a105-5578-4c13-aa74-f91494922361 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.821843] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b74d759-a791-4d62-9001-66de3865f984 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.830622] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ea86a35-2eeb-4759-97f3-999d532c8019 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.846748] env[65726]: DEBUG nova.compute.provider_tree [None req-e3b3b0a6-575e-4b59-bef3-4eef65e9267a tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1136.853570] env[65726]: WARNING openstack [None req-a2dc650f-ed4c-4edd-a0f8-7cba2e214975 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1136.853959] env[65726]: WARNING openstack [None req-a2dc650f-ed4c-4edd-a0f8-7cba2e214975 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1136.863775] env[65726]: DEBUG nova.scheduler.client.report [None req-e3b3b0a6-575e-4b59-bef3-4eef65e9267a tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1136.928899] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9230f0c2-55f9-4dd0-82d8-cd3277e657f9 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1136.952388] env[65726]: DEBUG nova.network.neutron [None req-a2dc650f-ed4c-4edd-a0f8-7cba2e214975 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: e7156a10-631a-4cc5-a544-88ce37763d02] Updating instance_info_cache with network_info: [{"id": "44a1554d-d380-49b2-ad33-9a251114c93c", "address": "fa:16:3e:63:0a:a8", "network": {"id": "4c972d9d-db3b-4779-ac39-39d9a2d141dd", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1418913861-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cbb17a09b35c4c22ade5c4082b10bda9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "50cf0a70-948d-4611-af05-94c1483064ed", "external-id": "nsx-vlan-transportzone-536", "segmentation_id": 536, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap44a1554d-d3", "ovs_interfaceid": "44a1554d-d380-49b2-ad33-9a251114c93c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1136.964477] env[65726]: DEBUG oslo_vmware.api [None req-72f2a102-8ae8-4f21-a769-e63022914c54 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Task: {'id': task-5116816, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.636351} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1136.965653] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-72f2a102-8ae8-4f21-a769-e63022914c54 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] 3bc25953-21b1-4729-af09-e7211fd8b2c6/3bc25953-21b1-4729-af09-e7211fd8b2c6.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1136.965876] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-72f2a102-8ae8-4f21-a769-e63022914c54 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: 3bc25953-21b1-4729-af09-e7211fd8b2c6] Extending root virtual disk to 1048576 {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1136.966498] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-37746483-6132-41c3-85cb-08281ee1c2ba {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.976599] env[65726]: DEBUG oslo_vmware.api [None req-72f2a102-8ae8-4f21-a769-e63022914c54 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Waiting for the task: (returnval){ [ 1136.976599] env[65726]: value = "task-5116819" [ 1136.976599] env[65726]: _type = "Task" [ 1136.976599] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1136.986874] env[65726]: DEBUG oslo_vmware.api [None req-72f2a102-8ae8-4f21-a769-e63022914c54 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Task: {'id': task-5116819, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1137.033989] env[65726]: INFO nova.compute.manager [-] [instance: 3b4d9461-747a-461c-b231-a0de02f0cb83] Took 1.28 seconds to deallocate network for instance. [ 1137.077670] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1137.364234] env[65726]: DEBUG nova.network.neutron [-] [instance: 0361ae8f-634d-4161-a594-7f93dbb1b7cd] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1137.369978] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e3b3b0a6-575e-4b59-bef3-4eef65e9267a tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.361s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1137.372910] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9230f0c2-55f9-4dd0-82d8-cd3277e657f9 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.444s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1137.375542] env[65726]: INFO nova.compute.claims [None req-9230f0c2-55f9-4dd0-82d8-cd3277e657f9 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 7c1188ce-8718-4719-8631-e59e7915b7aa] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1137.395279] env[65726]: INFO nova.scheduler.client.report [None req-e3b3b0a6-575e-4b59-bef3-4eef65e9267a tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Deleted allocations for instance 6fd70b4a-63e4-4258-9cff-f3b582500b9f [ 1137.458252] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a2dc650f-ed4c-4edd-a0f8-7cba2e214975 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Releasing lock "refresh_cache-e7156a10-631a-4cc5-a544-88ce37763d02" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1137.488326] env[65726]: DEBUG oslo_vmware.api [None req-72f2a102-8ae8-4f21-a769-e63022914c54 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Task: {'id': task-5116819, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.109543} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1137.489287] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-72f2a102-8ae8-4f21-a769-e63022914c54 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: 3bc25953-21b1-4729-af09-e7211fd8b2c6] Extended root virtual disk {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1137.490332] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d88af3f3-a9fb-4537-a5cb-8fd9d17cd8df {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.515553] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-72f2a102-8ae8-4f21-a769-e63022914c54 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: 3bc25953-21b1-4729-af09-e7211fd8b2c6] Reconfiguring VM instance instance-00000069 to attach disk [datastore1] 3bc25953-21b1-4729-af09-e7211fd8b2c6/3bc25953-21b1-4729-af09-e7211fd8b2c6.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1137.516364] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-dd2a3ed6-4b79-486f-8143-6aa2d4ebc58e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.541435] env[65726]: DEBUG oslo_concurrency.lockutils [None req-668445ef-7aad-4f12-b535-5fc35338444d tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1137.541772] env[65726]: DEBUG oslo_vmware.api [None req-72f2a102-8ae8-4f21-a769-e63022914c54 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Waiting for the task: (returnval){ [ 1137.541772] env[65726]: value = "task-5116820" [ 1137.541772] env[65726]: _type = "Task" [ 1137.541772] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1137.552465] env[65726]: DEBUG oslo_vmware.api [None req-72f2a102-8ae8-4f21-a769-e63022914c54 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Task: {'id': task-5116820, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1137.867661] env[65726]: INFO nova.compute.manager [-] [instance: 0361ae8f-634d-4161-a594-7f93dbb1b7cd] Took 1.32 seconds to deallocate network for instance. [ 1137.904522] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e3b3b0a6-575e-4b59-bef3-4eef65e9267a tempest-ServersNegativeTestJSON-292131928 tempest-ServersNegativeTestJSON-292131928-project-member] Lock "6fd70b4a-63e4-4258-9cff-f3b582500b9f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.877s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1137.988233] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d5263a0-8b43-4a4c-9fc4-9a357ea9ab25 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.008879] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61400e53-8905-4ffc-862b-5190b5d06a55 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.018021] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-a2dc650f-ed4c-4edd-a0f8-7cba2e214975 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: e7156a10-631a-4cc5-a544-88ce37763d02] Updating instance 'e7156a10-631a-4cc5-a544-88ce37763d02' progress to 83 {{(pid=65726) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1138.054851] env[65726]: DEBUG oslo_vmware.api [None req-72f2a102-8ae8-4f21-a769-e63022914c54 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Task: {'id': task-5116820, 'name': ReconfigVM_Task, 'duration_secs': 0.433743} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1138.055189] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-72f2a102-8ae8-4f21-a769-e63022914c54 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: 3bc25953-21b1-4729-af09-e7211fd8b2c6] Reconfigured VM instance instance-00000069 to attach disk [datastore1] 3bc25953-21b1-4729-af09-e7211fd8b2c6/3bc25953-21b1-4729-af09-e7211fd8b2c6.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1138.055874] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f03d570c-e851-490e-bfff-e769e33e0dfa {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.063325] env[65726]: DEBUG oslo_vmware.api [None req-72f2a102-8ae8-4f21-a769-e63022914c54 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Waiting for the task: (returnval){ [ 1138.063325] env[65726]: value = "task-5116821" [ 1138.063325] env[65726]: _type = "Task" [ 1138.063325] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1138.075413] env[65726]: DEBUG oslo_vmware.api [None req-72f2a102-8ae8-4f21-a769-e63022914c54 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Task: {'id': task-5116821, 'name': Rename_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1138.089716] env[65726]: DEBUG nova.compute.manager [req-aa38673a-7115-4cdd-81fd-ace82d090ad1 req-5364cc52-85af-4a5a-8a24-4c6911c3cf3f service nova] [instance: 0361ae8f-634d-4161-a594-7f93dbb1b7cd] Received event network-vif-deleted-2bfd7147-87fd-4a19-b28b-eb95f51f4d00 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1138.380568] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a7d158c8-609b-4687-83b0-c5b67c744acc tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1138.381223] env[65726]: DEBUG oslo_concurrency.lockutils [None req-d3029c9c-9ccf-4268-b57c-8efb1da60ab1 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Acquiring lock "interface-0d42fa75-20e9-4646-9b08-17015b7f068c-fb580bad-fd60-471d-95f2-066352417659" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1138.381486] env[65726]: DEBUG oslo_concurrency.lockutils [None req-d3029c9c-9ccf-4268-b57c-8efb1da60ab1 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Lock "interface-0d42fa75-20e9-4646-9b08-17015b7f068c-fb580bad-fd60-471d-95f2-066352417659" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1138.531058] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-a2dc650f-ed4c-4edd-a0f8-7cba2e214975 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: e7156a10-631a-4cc5-a544-88ce37763d02] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1138.531058] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b2f11efc-55cc-4da2-9206-0c0d133ba90f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.539078] env[65726]: DEBUG oslo_vmware.api [None req-a2dc650f-ed4c-4edd-a0f8-7cba2e214975 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Waiting for the task: (returnval){ [ 1138.539078] env[65726]: value = "task-5116822" [ 1138.539078] env[65726]: _type = "Task" [ 1138.539078] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1138.555421] env[65726]: DEBUG oslo_vmware.api [None req-a2dc650f-ed4c-4edd-a0f8-7cba2e214975 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': task-5116822, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1138.578338] env[65726]: DEBUG oslo_vmware.api [None req-72f2a102-8ae8-4f21-a769-e63022914c54 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Task: {'id': task-5116821, 'name': Rename_Task, 'duration_secs': 0.194205} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1138.582578] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-72f2a102-8ae8-4f21-a769-e63022914c54 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: 3bc25953-21b1-4729-af09-e7211fd8b2c6] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1138.583281] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9a9903e6-b581-41df-874c-45e88cbd6d7c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.593390] env[65726]: DEBUG oslo_vmware.api [None req-72f2a102-8ae8-4f21-a769-e63022914c54 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Waiting for the task: (returnval){ [ 1138.593390] env[65726]: value = "task-5116823" [ 1138.593390] env[65726]: _type = "Task" [ 1138.593390] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1138.601521] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53176613-d65a-4a54-bf30-55d1696f949e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.616533] env[65726]: DEBUG oslo_vmware.api [None req-72f2a102-8ae8-4f21-a769-e63022914c54 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Task: {'id': task-5116823, 'name': PowerOnVM_Task} progress is 33%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1138.620989] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-927de874-ae53-4428-9c01-24be32624b72 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.666579] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7059c97-fb98-48c8-948c-03e53c5445d9 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.675538] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c593feb-caa8-4d7e-8e65-dd985ea6683f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.690442] env[65726]: DEBUG nova.compute.provider_tree [None req-9230f0c2-55f9-4dd0-82d8-cd3277e657f9 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1138.887969] env[65726]: DEBUG oslo_concurrency.lockutils [None req-d3029c9c-9ccf-4268-b57c-8efb1da60ab1 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Acquiring lock "0d42fa75-20e9-4646-9b08-17015b7f068c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1138.888220] env[65726]: DEBUG oslo_concurrency.lockutils [None req-d3029c9c-9ccf-4268-b57c-8efb1da60ab1 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Acquired lock "0d42fa75-20e9-4646-9b08-17015b7f068c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1138.892425] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cdd57eee-f97a-45a2-a807-07a9867ae66f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.913650] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eada7221-8aff-4709-94f7-c9d01c307ebd {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.942418] env[65726]: WARNING openstack [None req-d3029c9c-9ccf-4268-b57c-8efb1da60ab1 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1138.942861] env[65726]: WARNING openstack [None req-d3029c9c-9ccf-4268-b57c-8efb1da60ab1 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1138.954356] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-d3029c9c-9ccf-4268-b57c-8efb1da60ab1 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: 0d42fa75-20e9-4646-9b08-17015b7f068c] Reconfiguring VM to detach interface {{(pid=65726) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1968}} [ 1138.954509] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4c0c6ac0-5959-4ac2-b8ba-eb76e8af82f6 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.978401] env[65726]: DEBUG oslo_vmware.api [None req-d3029c9c-9ccf-4268-b57c-8efb1da60ab1 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Waiting for the task: (returnval){ [ 1138.978401] env[65726]: value = "task-5116825" [ 1138.978401] env[65726]: _type = "Task" [ 1138.978401] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1138.988946] env[65726]: DEBUG oslo_vmware.api [None req-d3029c9c-9ccf-4268-b57c-8efb1da60ab1 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': task-5116825, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1139.050426] env[65726]: DEBUG oslo_vmware.api [None req-a2dc650f-ed4c-4edd-a0f8-7cba2e214975 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': task-5116822, 'name': PowerOnVM_Task, 'duration_secs': 0.508601} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1139.050739] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-a2dc650f-ed4c-4edd-a0f8-7cba2e214975 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: e7156a10-631a-4cc5-a544-88ce37763d02] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1139.050921] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-a2dc650f-ed4c-4edd-a0f8-7cba2e214975 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: e7156a10-631a-4cc5-a544-88ce37763d02] Updating instance 'e7156a10-631a-4cc5-a544-88ce37763d02' progress to 100 {{(pid=65726) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1139.106342] env[65726]: DEBUG oslo_vmware.api [None req-72f2a102-8ae8-4f21-a769-e63022914c54 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Task: {'id': task-5116823, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1139.194061] env[65726]: DEBUG nova.scheduler.client.report [None req-9230f0c2-55f9-4dd0-82d8-cd3277e657f9 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1139.492346] env[65726]: DEBUG oslo_vmware.api [None req-d3029c9c-9ccf-4268-b57c-8efb1da60ab1 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': task-5116825, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1139.542430] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b778c6b3-2a71-4d65-8347-d07f00992900 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Acquiring lock "5bfbaf25-5aed-46bc-97fc-f138d08ebbde" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1139.542639] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b778c6b3-2a71-4d65-8347-d07f00992900 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Lock "5bfbaf25-5aed-46bc-97fc-f138d08ebbde" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1139.607846] env[65726]: DEBUG oslo_vmware.api [None req-72f2a102-8ae8-4f21-a769-e63022914c54 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Task: {'id': task-5116823, 'name': PowerOnVM_Task, 'duration_secs': 0.697207} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1139.608203] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-72f2a102-8ae8-4f21-a769-e63022914c54 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: 3bc25953-21b1-4729-af09-e7211fd8b2c6] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1139.608453] env[65726]: INFO nova.compute.manager [None req-72f2a102-8ae8-4f21-a769-e63022914c54 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: 3bc25953-21b1-4729-af09-e7211fd8b2c6] Took 8.73 seconds to spawn the instance on the hypervisor. [ 1139.608638] env[65726]: DEBUG nova.compute.manager [None req-72f2a102-8ae8-4f21-a769-e63022914c54 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: 3bc25953-21b1-4729-af09-e7211fd8b2c6] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1139.609867] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c4ceb9d-90a1-441d-9378-a139b57fcc23 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.700840] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9230f0c2-55f9-4dd0-82d8-cd3277e657f9 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.328s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1139.701391] env[65726]: DEBUG nova.compute.manager [None req-9230f0c2-55f9-4dd0-82d8-cd3277e657f9 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 7c1188ce-8718-4719-8631-e59e7915b7aa] Start building networks asynchronously for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1139.705044] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 2.627s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1139.705224] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1139.705376] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=65726) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:937}} [ 1139.705668] env[65726]: DEBUG oslo_concurrency.lockutils [None req-668445ef-7aad-4f12-b535-5fc35338444d tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.164s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1139.705874] env[65726]: DEBUG nova.objects.instance [None req-668445ef-7aad-4f12-b535-5fc35338444d tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Lazy-loading 'resources' on Instance uuid 3b4d9461-747a-461c-b231-a0de02f0cb83 {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1139.707398] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4aa03c35-4ad8-4779-b823-463ca9d81d24 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.718490] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c24d8852-98bb-43f4-800f-504ceb3ee585 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.737119] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e374dbd7-cfb8-437f-a9cf-9e58d15b7392 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.744503] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7c4eebe-bc01-48d6-b7c8-94a287424200 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.780991] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=179083MB free_disk=95GB free_vcpus=48 pci_devices=None {{(pid=65726) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1136}} [ 1139.781161] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1139.990840] env[65726]: DEBUG oslo_vmware.api [None req-d3029c9c-9ccf-4268-b57c-8efb1da60ab1 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': task-5116825, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1140.045633] env[65726]: DEBUG nova.compute.manager [None req-b778c6b3-2a71-4d65-8347-d07f00992900 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] [instance: 5bfbaf25-5aed-46bc-97fc-f138d08ebbde] Starting instance... {{(pid=65726) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 1140.132288] env[65726]: INFO nova.compute.manager [None req-72f2a102-8ae8-4f21-a769-e63022914c54 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: 3bc25953-21b1-4729-af09-e7211fd8b2c6] Took 13.60 seconds to build instance. [ 1140.207768] env[65726]: DEBUG nova.compute.utils [None req-9230f0c2-55f9-4dd0-82d8-cd3277e657f9 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Using /dev/sd instead of None {{(pid=65726) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1140.209181] env[65726]: DEBUG nova.compute.manager [None req-9230f0c2-55f9-4dd0-82d8-cd3277e657f9 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 7c1188ce-8718-4719-8631-e59e7915b7aa] Allocating IP information in the background. {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1140.209380] env[65726]: DEBUG nova.network.neutron [None req-9230f0c2-55f9-4dd0-82d8-cd3277e657f9 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 7c1188ce-8718-4719-8631-e59e7915b7aa] allocate_for_instance() {{(pid=65726) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 1140.209697] env[65726]: WARNING neutronclient.v2_0.client [None req-9230f0c2-55f9-4dd0-82d8-cd3277e657f9 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1140.210038] env[65726]: WARNING neutronclient.v2_0.client [None req-9230f0c2-55f9-4dd0-82d8-cd3277e657f9 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1140.210721] env[65726]: WARNING openstack [None req-9230f0c2-55f9-4dd0-82d8-cd3277e657f9 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1140.211173] env[65726]: WARNING openstack [None req-9230f0c2-55f9-4dd0-82d8-cd3277e657f9 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1140.288664] env[65726]: DEBUG nova.policy [None req-9230f0c2-55f9-4dd0-82d8-cd3277e657f9 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a82797497789484a8fbf88a23a449c95', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f93d852e2c904f42981cbfff882558e9', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65726) authorize /opt/stack/nova/nova/policy.py:192}} [ 1140.396019] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55423322-35ee-400f-af8c-877468c75f50 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.405185] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0aec291b-2d71-4e5b-854e-151395e8f834 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.442183] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b7846c5-82fc-4ffc-a014-a00ebf450300 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.451263] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d57fb6ef-b4c2-44c8-aa8d-ee90aa055549 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.467759] env[65726]: DEBUG nova.compute.provider_tree [None req-668445ef-7aad-4f12-b535-5fc35338444d tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1140.490900] env[65726]: DEBUG oslo_vmware.api [None req-d3029c9c-9ccf-4268-b57c-8efb1da60ab1 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': task-5116825, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1140.577596] env[65726]: DEBUG nova.network.neutron [None req-9230f0c2-55f9-4dd0-82d8-cd3277e657f9 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 7c1188ce-8718-4719-8631-e59e7915b7aa] Successfully created port: e8f51770-ca83-4aa3-87c1-ef47653d9bb2 {{(pid=65726) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1140.588416] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b778c6b3-2a71-4d65-8347-d07f00992900 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1140.634935] env[65726]: DEBUG oslo_concurrency.lockutils [None req-72f2a102-8ae8-4f21-a769-e63022914c54 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Lock "3bc25953-21b1-4729-af09-e7211fd8b2c6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.111s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1140.710967] env[65726]: DEBUG nova.compute.manager [req-36cf57e4-3972-4bbb-89d5-6b37856bce0b req-63db5e10-c107-43af-b052-805ff0d71fc4 service nova] [instance: 3bc25953-21b1-4729-af09-e7211fd8b2c6] Received event network-changed-a7e1795d-67d0-4f32-baea-744806988e4b {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1140.711200] env[65726]: DEBUG nova.compute.manager [req-36cf57e4-3972-4bbb-89d5-6b37856bce0b req-63db5e10-c107-43af-b052-805ff0d71fc4 service nova] [instance: 3bc25953-21b1-4729-af09-e7211fd8b2c6] Refreshing instance network info cache due to event network-changed-a7e1795d-67d0-4f32-baea-744806988e4b. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 1140.711415] env[65726]: DEBUG oslo_concurrency.lockutils [req-36cf57e4-3972-4bbb-89d5-6b37856bce0b req-63db5e10-c107-43af-b052-805ff0d71fc4 service nova] Acquiring lock "refresh_cache-3bc25953-21b1-4729-af09-e7211fd8b2c6" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1140.711555] env[65726]: DEBUG oslo_concurrency.lockutils [req-36cf57e4-3972-4bbb-89d5-6b37856bce0b req-63db5e10-c107-43af-b052-805ff0d71fc4 service nova] Acquired lock "refresh_cache-3bc25953-21b1-4729-af09-e7211fd8b2c6" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1140.711709] env[65726]: DEBUG nova.network.neutron [req-36cf57e4-3972-4bbb-89d5-6b37856bce0b req-63db5e10-c107-43af-b052-805ff0d71fc4 service nova] [instance: 3bc25953-21b1-4729-af09-e7211fd8b2c6] Refreshing network info cache for port a7e1795d-67d0-4f32-baea-744806988e4b {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 1140.718599] env[65726]: DEBUG nova.compute.manager [None req-9230f0c2-55f9-4dd0-82d8-cd3277e657f9 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 7c1188ce-8718-4719-8631-e59e7915b7aa] Start building block device mappings for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1140.837908] env[65726]: DEBUG oslo_concurrency.lockutils [None req-aad3fddf-3318-4cfd-8f31-0db9fa8fe254 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Acquiring lock "e7156a10-631a-4cc5-a544-88ce37763d02" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1140.837908] env[65726]: DEBUG oslo_concurrency.lockutils [None req-aad3fddf-3318-4cfd-8f31-0db9fa8fe254 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Lock "e7156a10-631a-4cc5-a544-88ce37763d02" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1140.837908] env[65726]: DEBUG nova.compute.manager [None req-aad3fddf-3318-4cfd-8f31-0db9fa8fe254 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: e7156a10-631a-4cc5-a544-88ce37763d02] Going to confirm migration 6 {{(pid=65726) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:5307}} [ 1140.985831] env[65726]: DEBUG nova.scheduler.client.report [None req-668445ef-7aad-4f12-b535-5fc35338444d tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1140.996426] env[65726]: DEBUG oslo_vmware.api [None req-d3029c9c-9ccf-4268-b57c-8efb1da60ab1 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': task-5116825, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1141.216804] env[65726]: WARNING openstack [req-36cf57e4-3972-4bbb-89d5-6b37856bce0b req-63db5e10-c107-43af-b052-805ff0d71fc4 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1141.217236] env[65726]: WARNING openstack [req-36cf57e4-3972-4bbb-89d5-6b37856bce0b req-63db5e10-c107-43af-b052-805ff0d71fc4 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1141.347723] env[65726]: WARNING openstack [None req-aad3fddf-3318-4cfd-8f31-0db9fa8fe254 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1141.348349] env[65726]: WARNING openstack [None req-aad3fddf-3318-4cfd-8f31-0db9fa8fe254 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1141.359960] env[65726]: WARNING openstack [req-36cf57e4-3972-4bbb-89d5-6b37856bce0b req-63db5e10-c107-43af-b052-805ff0d71fc4 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1141.360403] env[65726]: WARNING openstack [req-36cf57e4-3972-4bbb-89d5-6b37856bce0b req-63db5e10-c107-43af-b052-805ff0d71fc4 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1141.389100] env[65726]: WARNING openstack [None req-aad3fddf-3318-4cfd-8f31-0db9fa8fe254 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1141.390441] env[65726]: WARNING openstack [None req-aad3fddf-3318-4cfd-8f31-0db9fa8fe254 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1141.396329] env[65726]: DEBUG oslo_concurrency.lockutils [None req-aad3fddf-3318-4cfd-8f31-0db9fa8fe254 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Acquiring lock "refresh_cache-e7156a10-631a-4cc5-a544-88ce37763d02" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1141.396525] env[65726]: DEBUG oslo_concurrency.lockutils [None req-aad3fddf-3318-4cfd-8f31-0db9fa8fe254 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Acquired lock "refresh_cache-e7156a10-631a-4cc5-a544-88ce37763d02" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1141.396732] env[65726]: DEBUG nova.network.neutron [None req-aad3fddf-3318-4cfd-8f31-0db9fa8fe254 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: e7156a10-631a-4cc5-a544-88ce37763d02] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1141.396911] env[65726]: DEBUG nova.objects.instance [None req-aad3fddf-3318-4cfd-8f31-0db9fa8fe254 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Lazy-loading 'info_cache' on Instance uuid e7156a10-631a-4cc5-a544-88ce37763d02 {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1141.460190] env[65726]: WARNING openstack [req-36cf57e4-3972-4bbb-89d5-6b37856bce0b req-63db5e10-c107-43af-b052-805ff0d71fc4 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1141.460605] env[65726]: WARNING openstack [req-36cf57e4-3972-4bbb-89d5-6b37856bce0b req-63db5e10-c107-43af-b052-805ff0d71fc4 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1141.492573] env[65726]: DEBUG oslo_concurrency.lockutils [None req-668445ef-7aad-4f12-b535-5fc35338444d tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.787s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1141.494788] env[65726]: DEBUG oslo_vmware.api [None req-d3029c9c-9ccf-4268-b57c-8efb1da60ab1 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': task-5116825, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1141.495384] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a7d158c8-609b-4687-83b0-c5b67c744acc tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.115s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1141.495497] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a7d158c8-609b-4687-83b0-c5b67c744acc tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1141.497665] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 1.716s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1141.525357] env[65726]: INFO nova.scheduler.client.report [None req-668445ef-7aad-4f12-b535-5fc35338444d tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Deleted allocations for instance 3b4d9461-747a-461c-b231-a0de02f0cb83 [ 1141.526656] env[65726]: INFO nova.scheduler.client.report [None req-a7d158c8-609b-4687-83b0-c5b67c744acc tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Deleted allocations for instance 0361ae8f-634d-4161-a594-7f93dbb1b7cd [ 1141.571819] env[65726]: DEBUG nova.network.neutron [req-36cf57e4-3972-4bbb-89d5-6b37856bce0b req-63db5e10-c107-43af-b052-805ff0d71fc4 service nova] [instance: 3bc25953-21b1-4729-af09-e7211fd8b2c6] Updated VIF entry in instance network info cache for port a7e1795d-67d0-4f32-baea-744806988e4b. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 1141.572235] env[65726]: DEBUG nova.network.neutron [req-36cf57e4-3972-4bbb-89d5-6b37856bce0b req-63db5e10-c107-43af-b052-805ff0d71fc4 service nova] [instance: 3bc25953-21b1-4729-af09-e7211fd8b2c6] Updating instance_info_cache with network_info: [{"id": "a7e1795d-67d0-4f32-baea-744806988e4b", "address": "fa:16:3e:06:9b:44", "network": {"id": "ad1373b5-e0ed-4ee6-a97c-af931afbab35", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-909418417-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.178", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a090d553766847e58e3231b966c92565", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5f60c972-a72d-4c5f-a250-faadfd6eafbe", "external-id": "nsx-vlan-transportzone-932", "segmentation_id": 932, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa7e1795d-67", "ovs_interfaceid": "a7e1795d-67d0-4f32-baea-744806988e4b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1141.729906] env[65726]: DEBUG nova.compute.manager [None req-9230f0c2-55f9-4dd0-82d8-cd3277e657f9 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 7c1188ce-8718-4719-8631-e59e7915b7aa] Start spawning the instance on the hypervisor. {{(pid=65726) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1141.762765] env[65726]: DEBUG nova.virt.hardware [None req-9230f0c2-55f9-4dd0-82d8-cd3277e657f9 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T19:28:34Z,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1539f06fa3534e90acbc3a60c4b8bd3f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T19:28:35Z,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1141.763019] env[65726]: DEBUG nova.virt.hardware [None req-9230f0c2-55f9-4dd0-82d8-cd3277e657f9 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1141.763238] env[65726]: DEBUG nova.virt.hardware [None req-9230f0c2-55f9-4dd0-82d8-cd3277e657f9 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1141.763450] env[65726]: DEBUG nova.virt.hardware [None req-9230f0c2-55f9-4dd0-82d8-cd3277e657f9 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1141.763594] env[65726]: DEBUG nova.virt.hardware [None req-9230f0c2-55f9-4dd0-82d8-cd3277e657f9 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1141.763735] env[65726]: DEBUG nova.virt.hardware [None req-9230f0c2-55f9-4dd0-82d8-cd3277e657f9 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1141.763982] env[65726]: DEBUG nova.virt.hardware [None req-9230f0c2-55f9-4dd0-82d8-cd3277e657f9 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1141.764230] env[65726]: DEBUG nova.virt.hardware [None req-9230f0c2-55f9-4dd0-82d8-cd3277e657f9 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1141.764416] env[65726]: DEBUG nova.virt.hardware [None req-9230f0c2-55f9-4dd0-82d8-cd3277e657f9 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1141.764602] env[65726]: DEBUG nova.virt.hardware [None req-9230f0c2-55f9-4dd0-82d8-cd3277e657f9 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1141.764774] env[65726]: DEBUG nova.virt.hardware [None req-9230f0c2-55f9-4dd0-82d8-cd3277e657f9 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1141.765659] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a5affe9-80f2-4280-9781-f744493f4446 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.775559] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da82d3ce-ff6a-479d-a339-d3839a9ba085 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.992545] env[65726]: DEBUG oslo_vmware.api [None req-d3029c9c-9ccf-4268-b57c-8efb1da60ab1 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': task-5116825, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1142.037740] env[65726]: DEBUG oslo_concurrency.lockutils [None req-668445ef-7aad-4f12-b535-5fc35338444d tempest-ServerRescueTestJSON-335691270 tempest-ServerRescueTestJSON-335691270-project-member] Lock "3b4d9461-747a-461c-b231-a0de02f0cb83" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.977s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1142.039067] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a7d158c8-609b-4687-83b0-c5b67c744acc tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Lock "0361ae8f-634d-4161-a594-7f93dbb1b7cd" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.187s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1142.075198] env[65726]: DEBUG oslo_concurrency.lockutils [req-36cf57e4-3972-4bbb-89d5-6b37856bce0b req-63db5e10-c107-43af-b052-805ff0d71fc4 service nova] Releasing lock "refresh_cache-3bc25953-21b1-4729-af09-e7211fd8b2c6" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1142.104998] env[65726]: DEBUG nova.network.neutron [None req-9230f0c2-55f9-4dd0-82d8-cd3277e657f9 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 7c1188ce-8718-4719-8631-e59e7915b7aa] Successfully updated port: e8f51770-ca83-4aa3-87c1-ef47653d9bb2 {{(pid=65726) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 1142.320209] env[65726]: DEBUG nova.compute.manager [req-03dc4527-7b1c-488b-a8dc-ccdedec21688 req-92ac51ac-5434-4576-9ca4-ebca7bbf943e service nova] [instance: 7c1188ce-8718-4719-8631-e59e7915b7aa] Received event network-vif-plugged-e8f51770-ca83-4aa3-87c1-ef47653d9bb2 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1142.320437] env[65726]: DEBUG oslo_concurrency.lockutils [req-03dc4527-7b1c-488b-a8dc-ccdedec21688 req-92ac51ac-5434-4576-9ca4-ebca7bbf943e service nova] Acquiring lock "7c1188ce-8718-4719-8631-e59e7915b7aa-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1142.320698] env[65726]: DEBUG oslo_concurrency.lockutils [req-03dc4527-7b1c-488b-a8dc-ccdedec21688 req-92ac51ac-5434-4576-9ca4-ebca7bbf943e service nova] Lock "7c1188ce-8718-4719-8631-e59e7915b7aa-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1142.320965] env[65726]: DEBUG oslo_concurrency.lockutils [req-03dc4527-7b1c-488b-a8dc-ccdedec21688 req-92ac51ac-5434-4576-9ca4-ebca7bbf943e service nova] Lock "7c1188ce-8718-4719-8631-e59e7915b7aa-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1142.321310] env[65726]: DEBUG nova.compute.manager [req-03dc4527-7b1c-488b-a8dc-ccdedec21688 req-92ac51ac-5434-4576-9ca4-ebca7bbf943e service nova] [instance: 7c1188ce-8718-4719-8631-e59e7915b7aa] No waiting events found dispatching network-vif-plugged-e8f51770-ca83-4aa3-87c1-ef47653d9bb2 {{(pid=65726) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1142.321484] env[65726]: WARNING nova.compute.manager [req-03dc4527-7b1c-488b-a8dc-ccdedec21688 req-92ac51ac-5434-4576-9ca4-ebca7bbf943e service nova] [instance: 7c1188ce-8718-4719-8631-e59e7915b7aa] Received unexpected event network-vif-plugged-e8f51770-ca83-4aa3-87c1-ef47653d9bb2 for instance with vm_state building and task_state spawning. [ 1142.404115] env[65726]: WARNING openstack [None req-aad3fddf-3318-4cfd-8f31-0db9fa8fe254 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1142.404510] env[65726]: WARNING openstack [None req-aad3fddf-3318-4cfd-8f31-0db9fa8fe254 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1142.498750] env[65726]: DEBUG oslo_vmware.api [None req-d3029c9c-9ccf-4268-b57c-8efb1da60ab1 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': task-5116825, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1142.507763] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Applying migration context for instance e7156a10-631a-4cc5-a544-88ce37763d02 as it has an incoming, in-progress migration 59ac4637-e900-45a3-bb6b-29b88535da20. Migration status is finished {{(pid=65726) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1046}} [ 1142.508987] env[65726]: INFO nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] [instance: e7156a10-631a-4cc5-a544-88ce37763d02] Updating resource usage from migration 59ac4637-e900-45a3-bb6b-29b88535da20 [ 1142.527598] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Instance 51f5337d-2e23-4b7c-b590-76be6bd82411 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65726) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1142.527865] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Instance 0d42fa75-20e9-4646-9b08-17015b7f068c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65726) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1142.529059] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Instance b353b5e9-500c-42d3-a87f-880a624febec actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65726) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1142.529059] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Instance 378c1d18-4bb6-4245-80bf-370fb7af0575 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65726) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1142.529059] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Migration 59ac4637-e900-45a3-bb6b-29b88535da20 is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65726) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1745}} [ 1142.529059] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Instance e7156a10-631a-4cc5-a544-88ce37763d02 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=65726) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1142.529059] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Instance 3bc25953-21b1-4729-af09-e7211fd8b2c6 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65726) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1142.529059] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Instance 7c1188ce-8718-4719-8631-e59e7915b7aa actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65726) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1142.570552] env[65726]: WARNING openstack [None req-aad3fddf-3318-4cfd-8f31-0db9fa8fe254 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1142.570980] env[65726]: WARNING openstack [None req-aad3fddf-3318-4cfd-8f31-0db9fa8fe254 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1142.609185] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9230f0c2-55f9-4dd0-82d8-cd3277e657f9 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Acquiring lock "refresh_cache-7c1188ce-8718-4719-8631-e59e7915b7aa" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1142.609185] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9230f0c2-55f9-4dd0-82d8-cd3277e657f9 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Acquired lock "refresh_cache-7c1188ce-8718-4719-8631-e59e7915b7aa" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1142.609185] env[65726]: DEBUG nova.network.neutron [None req-9230f0c2-55f9-4dd0-82d8-cd3277e657f9 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 7c1188ce-8718-4719-8631-e59e7915b7aa] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1142.638739] env[65726]: WARNING openstack [None req-aad3fddf-3318-4cfd-8f31-0db9fa8fe254 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1142.639130] env[65726]: WARNING openstack [None req-aad3fddf-3318-4cfd-8f31-0db9fa8fe254 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1142.720435] env[65726]: DEBUG nova.network.neutron [None req-aad3fddf-3318-4cfd-8f31-0db9fa8fe254 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: e7156a10-631a-4cc5-a544-88ce37763d02] Updating instance_info_cache with network_info: [{"id": "44a1554d-d380-49b2-ad33-9a251114c93c", "address": "fa:16:3e:63:0a:a8", "network": {"id": "4c972d9d-db3b-4779-ac39-39d9a2d141dd", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1418913861-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cbb17a09b35c4c22ade5c4082b10bda9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "50cf0a70-948d-4611-af05-94c1483064ed", "external-id": "nsx-vlan-transportzone-536", "segmentation_id": 536, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap44a1554d-d3", "ovs_interfaceid": "44a1554d-d380-49b2-ad33-9a251114c93c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1142.935018] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3a8ffaf6-e93e-4fc3-a889-1a35a6c01d6d tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] Acquiring lock "47e52bc1-c18e-4aa2-82a2-e4cb030a7a30" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1142.935245] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3a8ffaf6-e93e-4fc3-a889-1a35a6c01d6d tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] Lock "47e52bc1-c18e-4aa2-82a2-e4cb030a7a30" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1142.993682] env[65726]: DEBUG oslo_vmware.api [None req-d3029c9c-9ccf-4268-b57c-8efb1da60ab1 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': task-5116825, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1143.032654] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Instance 5bfbaf25-5aed-46bc-97fc-f138d08ebbde has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65726) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 1143.116245] env[65726]: WARNING openstack [None req-9230f0c2-55f9-4dd0-82d8-cd3277e657f9 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1143.116245] env[65726]: WARNING openstack [None req-9230f0c2-55f9-4dd0-82d8-cd3277e657f9 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1143.119576] env[65726]: DEBUG nova.network.neutron [None req-9230f0c2-55f9-4dd0-82d8-cd3277e657f9 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 7c1188ce-8718-4719-8631-e59e7915b7aa] Instance cache missing network info. {{(pid=65726) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 1143.194073] env[65726]: WARNING openstack [None req-9230f0c2-55f9-4dd0-82d8-cd3277e657f9 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1143.194073] env[65726]: WARNING openstack [None req-9230f0c2-55f9-4dd0-82d8-cd3277e657f9 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1143.227191] env[65726]: DEBUG oslo_concurrency.lockutils [None req-aad3fddf-3318-4cfd-8f31-0db9fa8fe254 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Releasing lock "refresh_cache-e7156a10-631a-4cc5-a544-88ce37763d02" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1143.227191] env[65726]: DEBUG nova.objects.instance [None req-aad3fddf-3318-4cfd-8f31-0db9fa8fe254 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Lazy-loading 'migration_context' on Instance uuid e7156a10-631a-4cc5-a544-88ce37763d02 {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1143.263383] env[65726]: WARNING openstack [None req-9230f0c2-55f9-4dd0-82d8-cd3277e657f9 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1143.263837] env[65726]: WARNING openstack [None req-9230f0c2-55f9-4dd0-82d8-cd3277e657f9 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1143.359565] env[65726]: DEBUG nova.network.neutron [None req-9230f0c2-55f9-4dd0-82d8-cd3277e657f9 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 7c1188ce-8718-4719-8631-e59e7915b7aa] Updating instance_info_cache with network_info: [{"id": "e8f51770-ca83-4aa3-87c1-ef47653d9bb2", "address": "fa:16:3e:8e:ce:ad", "network": {"id": "9dd55fb6-0dd2-484d-b94d-fbdccb958c79", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1005598846-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f93d852e2c904f42981cbfff882558e9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "74f30339-6421-4654-bddb-81d7f34db9d7", "external-id": "nsx-vlan-transportzone-899", "segmentation_id": 899, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape8f51770-ca", "ovs_interfaceid": "e8f51770-ca83-4aa3-87c1-ef47653d9bb2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1143.438330] env[65726]: DEBUG nova.compute.manager [None req-3a8ffaf6-e93e-4fc3-a889-1a35a6c01d6d tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] [instance: 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30] Starting instance... {{(pid=65726) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 1143.495887] env[65726]: DEBUG oslo_vmware.api [None req-d3029c9c-9ccf-4268-b57c-8efb1da60ab1 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': task-5116825, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1143.536692] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Instance 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65726) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 1143.536768] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Total usable vcpus: 48, total allocated vcpus: 8 {{(pid=65726) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1159}} [ 1143.537520] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2112MB phys_disk=100GB used_disk=8GB total_vcpus=48 used_vcpus=8 pci_stats=[] stats={'failed_builds': '0', 'num_instances': '7', 'num_vm_active': '5', 'num_task_None': '5', 'num_os_type_None': '7', 'num_proj_d46ccb00794f458b85da4a93879139ee': '2', 'io_workload': '1', 'num_vm_resized': '1', 'num_task_deleting': '1', 'num_proj_cbb17a09b35c4c22ade5c4082b10bda9': '1', 'num_proj_96149159e18e44f9bf3453e67681f224': '1', 'num_proj_4eece77569624f90bf64e5c51974173f': '1', 'num_proj_a090d553766847e58e3231b966c92565': '1', 'num_vm_building': '1', 'num_task_spawning': '1', 'num_proj_f93d852e2c904f42981cbfff882558e9': '1'} {{(pid=65726) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1168}} [ 1143.689089] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-156bee1e-c286-404a-8afb-9311b3bdbd8f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.698381] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b37f82b9-5ef8-450a-b599-d04fa4482522 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.733263] env[65726]: DEBUG nova.objects.base [None req-aad3fddf-3318-4cfd-8f31-0db9fa8fe254 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Object Instance lazy-loaded attributes: info_cache,migration_context {{(pid=65726) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1143.734446] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7b6bb81-c895-46d6-be88-032bb0af08c9 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.737561] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2d065df-4a1f-45f2-9824-e3f1566823fd {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.759805] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-95c14c45-19b5-43fb-807b-04e98493f0a9 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.763079] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1131092-c9da-485c-acda-60fb9bcfd120 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.771044] env[65726]: DEBUG oslo_vmware.api [None req-aad3fddf-3318-4cfd-8f31-0db9fa8fe254 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Waiting for the task: (returnval){ [ 1143.771044] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]526a0558-9daa-851f-bb3e-cfd4861c2eed" [ 1143.771044] env[65726]: _type = "Task" [ 1143.771044] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1143.779119] env[65726]: DEBUG nova.compute.provider_tree [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1143.791497] env[65726]: DEBUG oslo_vmware.api [None req-aad3fddf-3318-4cfd-8f31-0db9fa8fe254 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]526a0558-9daa-851f-bb3e-cfd4861c2eed, 'name': SearchDatastore_Task, 'duration_secs': 0.016451} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1143.792559] env[65726]: DEBUG oslo_concurrency.lockutils [None req-aad3fddf-3318-4cfd-8f31-0db9fa8fe254 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1143.862641] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9230f0c2-55f9-4dd0-82d8-cd3277e657f9 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Releasing lock "refresh_cache-7c1188ce-8718-4719-8631-e59e7915b7aa" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1143.863075] env[65726]: DEBUG nova.compute.manager [None req-9230f0c2-55f9-4dd0-82d8-cd3277e657f9 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 7c1188ce-8718-4719-8631-e59e7915b7aa] Instance network_info: |[{"id": "e8f51770-ca83-4aa3-87c1-ef47653d9bb2", "address": "fa:16:3e:8e:ce:ad", "network": {"id": "9dd55fb6-0dd2-484d-b94d-fbdccb958c79", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1005598846-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f93d852e2c904f42981cbfff882558e9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "74f30339-6421-4654-bddb-81d7f34db9d7", "external-id": "nsx-vlan-transportzone-899", "segmentation_id": 899, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape8f51770-ca", "ovs_interfaceid": "e8f51770-ca83-4aa3-87c1-ef47653d9bb2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1143.863591] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-9230f0c2-55f9-4dd0-82d8-cd3277e657f9 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 7c1188ce-8718-4719-8631-e59e7915b7aa] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:8e:ce:ad', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '74f30339-6421-4654-bddb-81d7f34db9d7', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e8f51770-ca83-4aa3-87c1-ef47653d9bb2', 'vif_model': 'vmxnet3'}] {{(pid=65726) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1143.871370] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-9230f0c2-55f9-4dd0-82d8-cd3277e657f9 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1143.871614] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7c1188ce-8718-4719-8631-e59e7915b7aa] Creating VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1143.871854] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-4a4d32e8-9f97-4948-b7e4-8bf32007a8f1 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.893231] env[65726]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1143.893231] env[65726]: value = "task-5116826" [ 1143.893231] env[65726]: _type = "Task" [ 1143.893231] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1143.901605] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116826, 'name': CreateVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1143.959129] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3a8ffaf6-e93e-4fc3-a889-1a35a6c01d6d tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1143.994969] env[65726]: DEBUG oslo_vmware.api [None req-d3029c9c-9ccf-4268-b57c-8efb1da60ab1 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': task-5116825, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1144.286182] env[65726]: DEBUG nova.scheduler.client.report [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1144.354342] env[65726]: DEBUG nova.compute.manager [req-85184f13-740d-4d96-b22b-ae6591bd66ec req-52634347-a767-4bce-8815-31d817c32da5 service nova] [instance: 7c1188ce-8718-4719-8631-e59e7915b7aa] Received event network-changed-e8f51770-ca83-4aa3-87c1-ef47653d9bb2 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1144.354342] env[65726]: DEBUG nova.compute.manager [req-85184f13-740d-4d96-b22b-ae6591bd66ec req-52634347-a767-4bce-8815-31d817c32da5 service nova] [instance: 7c1188ce-8718-4719-8631-e59e7915b7aa] Refreshing instance network info cache due to event network-changed-e8f51770-ca83-4aa3-87c1-ef47653d9bb2. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 1144.354493] env[65726]: DEBUG oslo_concurrency.lockutils [req-85184f13-740d-4d96-b22b-ae6591bd66ec req-52634347-a767-4bce-8815-31d817c32da5 service nova] Acquiring lock "refresh_cache-7c1188ce-8718-4719-8631-e59e7915b7aa" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1144.354606] env[65726]: DEBUG oslo_concurrency.lockutils [req-85184f13-740d-4d96-b22b-ae6591bd66ec req-52634347-a767-4bce-8815-31d817c32da5 service nova] Acquired lock "refresh_cache-7c1188ce-8718-4719-8631-e59e7915b7aa" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1144.354812] env[65726]: DEBUG nova.network.neutron [req-85184f13-740d-4d96-b22b-ae6591bd66ec req-52634347-a767-4bce-8815-31d817c32da5 service nova] [instance: 7c1188ce-8718-4719-8631-e59e7915b7aa] Refreshing network info cache for port e8f51770-ca83-4aa3-87c1-ef47653d9bb2 {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 1144.405092] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116826, 'name': CreateVM_Task} progress is 25%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1144.498726] env[65726]: DEBUG oslo_vmware.api [None req-d3029c9c-9ccf-4268-b57c-8efb1da60ab1 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': task-5116825, 'name': ReconfigVM_Task} progress is 18%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1144.791344] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=65726) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1097}} [ 1144.791605] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 3.294s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1144.792228] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b778c6b3-2a71-4d65-8347-d07f00992900 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.204s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1144.793519] env[65726]: INFO nova.compute.claims [None req-b778c6b3-2a71-4d65-8347-d07f00992900 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] [instance: 5bfbaf25-5aed-46bc-97fc-f138d08ebbde] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1144.796365] env[65726]: DEBUG oslo_service.periodic_task [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=65726) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1144.796513] env[65726]: DEBUG nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Cleaning up deleted instances {{(pid=65726) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11932}} [ 1144.861083] env[65726]: WARNING openstack [req-85184f13-740d-4d96-b22b-ae6591bd66ec req-52634347-a767-4bce-8815-31d817c32da5 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1144.861083] env[65726]: WARNING openstack [req-85184f13-740d-4d96-b22b-ae6591bd66ec req-52634347-a767-4bce-8815-31d817c32da5 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1144.907042] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116826, 'name': CreateVM_Task, 'duration_secs': 0.671242} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1144.907536] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7c1188ce-8718-4719-8631-e59e7915b7aa] Created VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1144.908447] env[65726]: WARNING openstack [None req-9230f0c2-55f9-4dd0-82d8-cd3277e657f9 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1144.909097] env[65726]: WARNING openstack [None req-9230f0c2-55f9-4dd0-82d8-cd3277e657f9 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1144.914570] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9230f0c2-55f9-4dd0-82d8-cd3277e657f9 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1144.914930] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9230f0c2-55f9-4dd0-82d8-cd3277e657f9 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1144.915475] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9230f0c2-55f9-4dd0-82d8-cd3277e657f9 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1144.916367] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1c2175b7-db01-47a5-b094-0104e48a0694 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.923076] env[65726]: DEBUG oslo_vmware.api [None req-9230f0c2-55f9-4dd0-82d8-cd3277e657f9 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Waiting for the task: (returnval){ [ 1144.923076] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5227fee3-5017-b38f-9580-e307cd007428" [ 1144.923076] env[65726]: _type = "Task" [ 1144.923076] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1144.931894] env[65726]: DEBUG oslo_vmware.api [None req-9230f0c2-55f9-4dd0-82d8-cd3277e657f9 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5227fee3-5017-b38f-9580-e307cd007428, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1144.961927] env[65726]: WARNING openstack [req-85184f13-740d-4d96-b22b-ae6591bd66ec req-52634347-a767-4bce-8815-31d817c32da5 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1144.961927] env[65726]: WARNING openstack [req-85184f13-740d-4d96-b22b-ae6591bd66ec req-52634347-a767-4bce-8815-31d817c32da5 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1144.999542] env[65726]: DEBUG oslo_vmware.api [None req-d3029c9c-9ccf-4268-b57c-8efb1da60ab1 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': task-5116825, 'name': ReconfigVM_Task, 'duration_secs': 5.847641} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1144.999785] env[65726]: DEBUG oslo_concurrency.lockutils [None req-d3029c9c-9ccf-4268-b57c-8efb1da60ab1 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Releasing lock "0d42fa75-20e9-4646-9b08-17015b7f068c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1144.999988] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-d3029c9c-9ccf-4268-b57c-8efb1da60ab1 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: 0d42fa75-20e9-4646-9b08-17015b7f068c] Reconfigured VM to detach interface {{(pid=65726) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1978}} [ 1145.000872] env[65726]: WARNING openstack [None req-d3029c9c-9ccf-4268-b57c-8efb1da60ab1 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1145.001288] env[65726]: WARNING openstack [None req-d3029c9c-9ccf-4268-b57c-8efb1da60ab1 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1145.023078] env[65726]: WARNING openstack [req-85184f13-740d-4d96-b22b-ae6591bd66ec req-52634347-a767-4bce-8815-31d817c32da5 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1145.024404] env[65726]: WARNING openstack [req-85184f13-740d-4d96-b22b-ae6591bd66ec req-52634347-a767-4bce-8815-31d817c32da5 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1145.047306] env[65726]: WARNING openstack [None req-d3029c9c-9ccf-4268-b57c-8efb1da60ab1 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1145.047719] env[65726]: WARNING openstack [None req-d3029c9c-9ccf-4268-b57c-8efb1da60ab1 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1145.136854] env[65726]: DEBUG nova.network.neutron [req-85184f13-740d-4d96-b22b-ae6591bd66ec req-52634347-a767-4bce-8815-31d817c32da5 service nova] [instance: 7c1188ce-8718-4719-8631-e59e7915b7aa] Updated VIF entry in instance network info cache for port e8f51770-ca83-4aa3-87c1-ef47653d9bb2. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 1145.137298] env[65726]: DEBUG nova.network.neutron [req-85184f13-740d-4d96-b22b-ae6591bd66ec req-52634347-a767-4bce-8815-31d817c32da5 service nova] [instance: 7c1188ce-8718-4719-8631-e59e7915b7aa] Updating instance_info_cache with network_info: [{"id": "e8f51770-ca83-4aa3-87c1-ef47653d9bb2", "address": "fa:16:3e:8e:ce:ad", "network": {"id": "9dd55fb6-0dd2-484d-b94d-fbdccb958c79", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1005598846-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f93d852e2c904f42981cbfff882558e9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "74f30339-6421-4654-bddb-81d7f34db9d7", "external-id": "nsx-vlan-transportzone-899", "segmentation_id": 899, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape8f51770-ca", "ovs_interfaceid": "e8f51770-ca83-4aa3-87c1-ef47653d9bb2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1145.313510] env[65726]: DEBUG nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] There are 59 instances to clean {{(pid=65726) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11941}} [ 1145.313777] env[65726]: DEBUG nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] [instance: 4809fcf6-59d4-409c-bd8d-981dc85a686f] Instance has had 0 of 5 cleanup attempts {{(pid=65726) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 1145.436231] env[65726]: DEBUG oslo_vmware.api [None req-9230f0c2-55f9-4dd0-82d8-cd3277e657f9 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5227fee3-5017-b38f-9580-e307cd007428, 'name': SearchDatastore_Task, 'duration_secs': 0.038983} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1145.436563] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9230f0c2-55f9-4dd0-82d8-cd3277e657f9 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1145.436801] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-9230f0c2-55f9-4dd0-82d8-cd3277e657f9 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 7c1188ce-8718-4719-8631-e59e7915b7aa] Processing image b52362a3-ee8a-4cbf-b06f-513b0cc8f95c {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1145.437745] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9230f0c2-55f9-4dd0-82d8-cd3277e657f9 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1145.437745] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9230f0c2-55f9-4dd0-82d8-cd3277e657f9 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1145.437745] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-9230f0c2-55f9-4dd0-82d8-cd3277e657f9 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1145.437745] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4320caf4-227e-4f4e-b1db-cef64dc11ec5 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.449271] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-9230f0c2-55f9-4dd0-82d8-cd3277e657f9 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1145.449555] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-9230f0c2-55f9-4dd0-82d8-cd3277e657f9 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65726) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1145.450598] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a36a33d5-145c-42de-b9ef-01b1a9811695 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.458226] env[65726]: DEBUG oslo_vmware.api [None req-9230f0c2-55f9-4dd0-82d8-cd3277e657f9 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Waiting for the task: (returnval){ [ 1145.458226] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]528e5226-a4dc-c3c1-e3a0-7b44eaa2f73b" [ 1145.458226] env[65726]: _type = "Task" [ 1145.458226] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1145.475484] env[65726]: DEBUG oslo_vmware.api [None req-9230f0c2-55f9-4dd0-82d8-cd3277e657f9 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]528e5226-a4dc-c3c1-e3a0-7b44eaa2f73b, 'name': SearchDatastore_Task, 'duration_secs': 0.012169} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1145.475484] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-48fb407b-f332-489b-9b75-b01e21ff1bbf {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.482238] env[65726]: DEBUG oslo_vmware.api [None req-9230f0c2-55f9-4dd0-82d8-cd3277e657f9 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Waiting for the task: (returnval){ [ 1145.482238] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]526558d3-dfc3-d885-81d2-e178c9c8ee90" [ 1145.482238] env[65726]: _type = "Task" [ 1145.482238] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1145.491870] env[65726]: DEBUG oslo_vmware.api [None req-9230f0c2-55f9-4dd0-82d8-cd3277e657f9 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]526558d3-dfc3-d885-81d2-e178c9c8ee90, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1145.640979] env[65726]: DEBUG oslo_concurrency.lockutils [req-85184f13-740d-4d96-b22b-ae6591bd66ec req-52634347-a767-4bce-8815-31d817c32da5 service nova] Releasing lock "refresh_cache-7c1188ce-8718-4719-8631-e59e7915b7aa" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1145.820905] env[65726]: DEBUG nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] [instance: 0361ae8f-634d-4161-a594-7f93dbb1b7cd] Instance has had 0 of 5 cleanup attempts {{(pid=65726) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 1145.995893] env[65726]: DEBUG oslo_vmware.api [None req-9230f0c2-55f9-4dd0-82d8-cd3277e657f9 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]526558d3-dfc3-d885-81d2-e178c9c8ee90, 'name': SearchDatastore_Task, 'duration_secs': 0.011681} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1145.997403] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9230f0c2-55f9-4dd0-82d8-cd3277e657f9 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1145.997841] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-9230f0c2-55f9-4dd0-82d8-cd3277e657f9 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] 7c1188ce-8718-4719-8631-e59e7915b7aa/7c1188ce-8718-4719-8631-e59e7915b7aa.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1145.999355] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5ac71ea-32d0-40dd-abce-14003d72545d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.002063] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7dbf60f8-3e1c-4b7d-aef0-5ac3b5c1e5bb {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.010081] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a0a07ab-efb4-4734-8e20-b727f8e67fd7 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.016628] env[65726]: DEBUG oslo_vmware.api [None req-9230f0c2-55f9-4dd0-82d8-cd3277e657f9 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Waiting for the task: (returnval){ [ 1146.016628] env[65726]: value = "task-5116827" [ 1146.016628] env[65726]: _type = "Task" [ 1146.016628] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1146.053883] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46005b32-3bdb-4b32-a504-78a62021a534 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.056682] env[65726]: DEBUG oslo_vmware.api [None req-9230f0c2-55f9-4dd0-82d8-cd3277e657f9 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': task-5116827, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1146.062638] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8429e0a0-c6a8-4fb0-ac07-db33a7dd7fc3 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.068731] env[65726]: DEBUG oslo_concurrency.lockutils [None req-8ea0406c-db1c-4929-acf9-e0ba5e267d6a tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Acquiring lock "0d42fa75-20e9-4646-9b08-17015b7f068c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1146.069148] env[65726]: DEBUG oslo_concurrency.lockutils [None req-8ea0406c-db1c-4929-acf9-e0ba5e267d6a tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Lock "0d42fa75-20e9-4646-9b08-17015b7f068c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1146.069400] env[65726]: DEBUG oslo_concurrency.lockutils [None req-8ea0406c-db1c-4929-acf9-e0ba5e267d6a tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Acquiring lock "0d42fa75-20e9-4646-9b08-17015b7f068c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1146.069598] env[65726]: DEBUG oslo_concurrency.lockutils [None req-8ea0406c-db1c-4929-acf9-e0ba5e267d6a tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Lock "0d42fa75-20e9-4646-9b08-17015b7f068c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1146.069761] env[65726]: DEBUG oslo_concurrency.lockutils [None req-8ea0406c-db1c-4929-acf9-e0ba5e267d6a tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Lock "0d42fa75-20e9-4646-9b08-17015b7f068c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1146.079402] env[65726]: DEBUG nova.compute.provider_tree [None req-b778c6b3-2a71-4d65-8347-d07f00992900 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1146.081121] env[65726]: INFO nova.compute.manager [None req-8ea0406c-db1c-4929-acf9-e0ba5e267d6a tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: 0d42fa75-20e9-4646-9b08-17015b7f068c] Terminating instance [ 1146.327274] env[65726]: DEBUG nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] [instance: d8801049-fc34-4b2f-8d5e-b9459c3e03a6] Instance has had 0 of 5 cleanup attempts {{(pid=65726) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 1146.390921] env[65726]: DEBUG oslo_concurrency.lockutils [None req-d3029c9c-9ccf-4268-b57c-8efb1da60ab1 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Acquiring lock "refresh_cache-0d42fa75-20e9-4646-9b08-17015b7f068c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1146.391189] env[65726]: DEBUG oslo_concurrency.lockutils [None req-d3029c9c-9ccf-4268-b57c-8efb1da60ab1 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Acquired lock "refresh_cache-0d42fa75-20e9-4646-9b08-17015b7f068c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1146.391447] env[65726]: DEBUG nova.network.neutron [None req-d3029c9c-9ccf-4268-b57c-8efb1da60ab1 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: 0d42fa75-20e9-4646-9b08-17015b7f068c] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1146.531104] env[65726]: DEBUG oslo_vmware.api [None req-9230f0c2-55f9-4dd0-82d8-cd3277e657f9 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': task-5116827, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1146.586217] env[65726]: DEBUG nova.scheduler.client.report [None req-b778c6b3-2a71-4d65-8347-d07f00992900 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1146.590539] env[65726]: DEBUG nova.compute.manager [None req-8ea0406c-db1c-4929-acf9-e0ba5e267d6a tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: 0d42fa75-20e9-4646-9b08-17015b7f068c] Start destroying the instance on the hypervisor. {{(pid=65726) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1146.590759] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-8ea0406c-db1c-4929-acf9-e0ba5e267d6a tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: 0d42fa75-20e9-4646-9b08-17015b7f068c] Destroying instance {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1146.592077] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe751836-4fbf-4016-91f5-27b7fba442bb {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.603361] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-8ea0406c-db1c-4929-acf9-e0ba5e267d6a tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: 0d42fa75-20e9-4646-9b08-17015b7f068c] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1146.603668] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f3d19430-4cf3-4eaa-a010-662fb1b524af {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.614686] env[65726]: DEBUG oslo_vmware.api [None req-8ea0406c-db1c-4929-acf9-e0ba5e267d6a tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Waiting for the task: (returnval){ [ 1146.614686] env[65726]: value = "task-5116828" [ 1146.614686] env[65726]: _type = "Task" [ 1146.614686] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1146.625043] env[65726]: DEBUG oslo_vmware.api [None req-8ea0406c-db1c-4929-acf9-e0ba5e267d6a tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': task-5116828, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1146.830896] env[65726]: DEBUG nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] [instance: b0fec5ff-8c1e-41f9-b30a-6253a34baeec] Instance has had 0 of 5 cleanup attempts {{(pid=65726) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 1146.894518] env[65726]: WARNING openstack [None req-d3029c9c-9ccf-4268-b57c-8efb1da60ab1 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1146.894932] env[65726]: WARNING openstack [None req-d3029c9c-9ccf-4268-b57c-8efb1da60ab1 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1147.015371] env[65726]: WARNING openstack [None req-d3029c9c-9ccf-4268-b57c-8efb1da60ab1 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1147.015743] env[65726]: WARNING openstack [None req-d3029c9c-9ccf-4268-b57c-8efb1da60ab1 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1147.044939] env[65726]: DEBUG oslo_vmware.api [None req-9230f0c2-55f9-4dd0-82d8-cd3277e657f9 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': task-5116827, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.664471} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1147.045305] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-9230f0c2-55f9-4dd0-82d8-cd3277e657f9 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] 7c1188ce-8718-4719-8631-e59e7915b7aa/7c1188ce-8718-4719-8631-e59e7915b7aa.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1147.045756] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-9230f0c2-55f9-4dd0-82d8-cd3277e657f9 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 7c1188ce-8718-4719-8631-e59e7915b7aa] Extending root virtual disk to 1048576 {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1147.046190] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f4822ca0-ffb9-4730-8380-a03c1c3cb6de {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.054690] env[65726]: DEBUG oslo_vmware.api [None req-9230f0c2-55f9-4dd0-82d8-cd3277e657f9 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Waiting for the task: (returnval){ [ 1147.054690] env[65726]: value = "task-5116829" [ 1147.054690] env[65726]: _type = "Task" [ 1147.054690] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1147.068897] env[65726]: DEBUG oslo_vmware.api [None req-9230f0c2-55f9-4dd0-82d8-cd3277e657f9 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': task-5116829, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1147.093114] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b778c6b3-2a71-4d65-8347-d07f00992900 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.301s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1147.093721] env[65726]: DEBUG nova.compute.manager [None req-b778c6b3-2a71-4d65-8347-d07f00992900 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] [instance: 5bfbaf25-5aed-46bc-97fc-f138d08ebbde] Start building networks asynchronously for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1147.100318] env[65726]: WARNING openstack [None req-d3029c9c-9ccf-4268-b57c-8efb1da60ab1 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1147.101569] env[65726]: WARNING openstack [None req-d3029c9c-9ccf-4268-b57c-8efb1da60ab1 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1147.108321] env[65726]: DEBUG oslo_concurrency.lockutils [None req-aad3fddf-3318-4cfd-8f31-0db9fa8fe254 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 3.316s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1147.126173] env[65726]: DEBUG oslo_vmware.api [None req-8ea0406c-db1c-4929-acf9-e0ba5e267d6a tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': task-5116828, 'name': PowerOffVM_Task, 'duration_secs': 0.243506} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1147.127122] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-8ea0406c-db1c-4929-acf9-e0ba5e267d6a tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: 0d42fa75-20e9-4646-9b08-17015b7f068c] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1147.127297] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-8ea0406c-db1c-4929-acf9-e0ba5e267d6a tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: 0d42fa75-20e9-4646-9b08-17015b7f068c] Unregistering the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1147.127626] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-14606d1d-876a-4c4a-960b-8dc7929bf574 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.194788] env[65726]: INFO nova.network.neutron [None req-d3029c9c-9ccf-4268-b57c-8efb1da60ab1 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: 0d42fa75-20e9-4646-9b08-17015b7f068c] Port fb580bad-fd60-471d-95f2-066352417659 from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 1147.195228] env[65726]: DEBUG nova.network.neutron [None req-d3029c9c-9ccf-4268-b57c-8efb1da60ab1 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: 0d42fa75-20e9-4646-9b08-17015b7f068c] Updating instance_info_cache with network_info: [{"id": "7a95cb45-ebfc-4ac0-ad58-4bb57170e47a", "address": "fa:16:3e:2b:ea:81", "network": {"id": "51a66448-0f72-4266-8868-856b48be839f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-155658961-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.213", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d46ccb00794f458b85da4a93879139ee", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b1f3e6c3-5584-4852-9017-476ab8ac4946", "external-id": "nsx-vlan-transportzone-304", "segmentation_id": 304, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7a95cb45-eb", "ovs_interfaceid": "7a95cb45-ebfc-4ac0-ad58-4bb57170e47a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1147.197731] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-8ea0406c-db1c-4929-acf9-e0ba5e267d6a tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: 0d42fa75-20e9-4646-9b08-17015b7f068c] Unregistered the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1147.197973] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-8ea0406c-db1c-4929-acf9-e0ba5e267d6a tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: 0d42fa75-20e9-4646-9b08-17015b7f068c] Deleting contents of the VM from datastore datastore2 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1147.198109] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-8ea0406c-db1c-4929-acf9-e0ba5e267d6a tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Deleting the datastore file [datastore2] 0d42fa75-20e9-4646-9b08-17015b7f068c {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1147.198631] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9838e418-9cb8-4109-b060-a9d8fa535fb9 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.207885] env[65726]: DEBUG oslo_vmware.api [None req-8ea0406c-db1c-4929-acf9-e0ba5e267d6a tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Waiting for the task: (returnval){ [ 1147.207885] env[65726]: value = "task-5116831" [ 1147.207885] env[65726]: _type = "Task" [ 1147.207885] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1147.218705] env[65726]: DEBUG oslo_vmware.api [None req-8ea0406c-db1c-4929-acf9-e0ba5e267d6a tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': task-5116831, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1147.334762] env[65726]: DEBUG nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] [instance: 3b4d9461-747a-461c-b231-a0de02f0cb83] Instance has had 0 of 5 cleanup attempts {{(pid=65726) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 1147.568760] env[65726]: DEBUG oslo_vmware.api [None req-9230f0c2-55f9-4dd0-82d8-cd3277e657f9 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': task-5116829, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.097734} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1147.570307] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-9230f0c2-55f9-4dd0-82d8-cd3277e657f9 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 7c1188ce-8718-4719-8631-e59e7915b7aa] Extended root virtual disk {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1147.570307] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f5d61f3-87c5-4e88-b050-778ff4034ac4 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.597491] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-9230f0c2-55f9-4dd0-82d8-cd3277e657f9 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 7c1188ce-8718-4719-8631-e59e7915b7aa] Reconfiguring VM instance instance-0000006a to attach disk [datastore1] 7c1188ce-8718-4719-8631-e59e7915b7aa/7c1188ce-8718-4719-8631-e59e7915b7aa.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1147.597832] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1df3bb27-62b7-4bba-8928-f3bfa279be29 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.613647] env[65726]: DEBUG nova.compute.utils [None req-b778c6b3-2a71-4d65-8347-d07f00992900 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Using /dev/sd instead of None {{(pid=65726) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1147.618459] env[65726]: DEBUG nova.compute.manager [None req-b778c6b3-2a71-4d65-8347-d07f00992900 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] [instance: 5bfbaf25-5aed-46bc-97fc-f138d08ebbde] Allocating IP information in the background. {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1147.618459] env[65726]: DEBUG nova.network.neutron [None req-b778c6b3-2a71-4d65-8347-d07f00992900 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] [instance: 5bfbaf25-5aed-46bc-97fc-f138d08ebbde] allocate_for_instance() {{(pid=65726) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 1147.618682] env[65726]: WARNING neutronclient.v2_0.client [None req-b778c6b3-2a71-4d65-8347-d07f00992900 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1147.618960] env[65726]: WARNING neutronclient.v2_0.client [None req-b778c6b3-2a71-4d65-8347-d07f00992900 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1147.620036] env[65726]: WARNING openstack [None req-b778c6b3-2a71-4d65-8347-d07f00992900 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1147.620036] env[65726]: WARNING openstack [None req-b778c6b3-2a71-4d65-8347-d07f00992900 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1147.634373] env[65726]: DEBUG oslo_vmware.api [None req-9230f0c2-55f9-4dd0-82d8-cd3277e657f9 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Waiting for the task: (returnval){ [ 1147.634373] env[65726]: value = "task-5116832" [ 1147.634373] env[65726]: _type = "Task" [ 1147.634373] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1147.644362] env[65726]: DEBUG oslo_vmware.api [None req-9230f0c2-55f9-4dd0-82d8-cd3277e657f9 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': task-5116832, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1147.678143] env[65726]: DEBUG nova.policy [None req-b778c6b3-2a71-4d65-8347-d07f00992900 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'cc5f9964e96a460fadfec7a9fe8e861b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4ffd45f4a7a041199a4fc7f69f5e7e9b', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65726) authorize /opt/stack/nova/nova/policy.py:192}} [ 1147.699598] env[65726]: DEBUG oslo_concurrency.lockutils [None req-d3029c9c-9ccf-4268-b57c-8efb1da60ab1 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Releasing lock "refresh_cache-0d42fa75-20e9-4646-9b08-17015b7f068c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1147.724955] env[65726]: DEBUG oslo_vmware.api [None req-8ea0406c-db1c-4929-acf9-e0ba5e267d6a tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': task-5116831, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.305933} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1147.725677] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-8ea0406c-db1c-4929-acf9-e0ba5e267d6a tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Deleted the datastore file {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1147.725677] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-8ea0406c-db1c-4929-acf9-e0ba5e267d6a tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: 0d42fa75-20e9-4646-9b08-17015b7f068c] Deleted contents of the VM from datastore datastore2 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1147.725862] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-8ea0406c-db1c-4929-acf9-e0ba5e267d6a tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: 0d42fa75-20e9-4646-9b08-17015b7f068c] Instance destroyed {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1147.726023] env[65726]: INFO nova.compute.manager [None req-8ea0406c-db1c-4929-acf9-e0ba5e267d6a tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: 0d42fa75-20e9-4646-9b08-17015b7f068c] Took 1.14 seconds to destroy the instance on the hypervisor. [ 1147.726303] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-8ea0406c-db1c-4929-acf9-e0ba5e267d6a tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1147.726896] env[65726]: DEBUG nova.compute.manager [-] [instance: 0d42fa75-20e9-4646-9b08-17015b7f068c] Deallocating network for instance {{(pid=65726) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1147.726935] env[65726]: DEBUG nova.network.neutron [-] [instance: 0d42fa75-20e9-4646-9b08-17015b7f068c] deallocate_for_instance() {{(pid=65726) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 1147.727630] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1147.728500] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1147.769536] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1147.769877] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1147.840214] env[65726]: DEBUG nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] [instance: 5cdf1d61-b817-4986-a48c-804a2374497a] Instance has had 0 of 5 cleanup attempts {{(pid=65726) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 1147.873866] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4c3b7cf-4923-42dc-a16c-9da0895d3036 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.883993] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2f2b80b-d5c1-416e-9ad7-f23c287678fd {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.920884] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-058f2900-d571-4e4a-8555-c766aaa15beb {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.930160] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f290a61c-83fb-4df0-95b3-398a9d130327 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.947937] env[65726]: DEBUG nova.compute.provider_tree [None req-aad3fddf-3318-4cfd-8f31-0db9fa8fe254 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1147.964609] env[65726]: DEBUG nova.network.neutron [None req-b778c6b3-2a71-4d65-8347-d07f00992900 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] [instance: 5bfbaf25-5aed-46bc-97fc-f138d08ebbde] Successfully created port: c41be8b4-e252-47c3-8529-cd6a7ce6a8d3 {{(pid=65726) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1148.011178] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e8a62c22-e16d-473f-bc47-d51d9d4949ce tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Acquiring lock "9c0599e3-383f-46ab-809b-944cc3a4d206" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1148.011485] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e8a62c22-e16d-473f-bc47-d51d9d4949ce tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Lock "9c0599e3-383f-46ab-809b-944cc3a4d206" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1148.118924] env[65726]: DEBUG nova.compute.manager [None req-b778c6b3-2a71-4d65-8347-d07f00992900 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] [instance: 5bfbaf25-5aed-46bc-97fc-f138d08ebbde] Start building block device mappings for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1148.156564] env[65726]: DEBUG oslo_vmware.api [None req-9230f0c2-55f9-4dd0-82d8-cd3277e657f9 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': task-5116832, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1148.203682] env[65726]: DEBUG oslo_concurrency.lockutils [None req-d3029c9c-9ccf-4268-b57c-8efb1da60ab1 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Lock "interface-0d42fa75-20e9-4646-9b08-17015b7f068c-fb580bad-fd60-471d-95f2-066352417659" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 9.822s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1148.343809] env[65726]: DEBUG nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] [instance: 48923e3d-b00d-4034-bacf-7be82009fb08] Instance has had 0 of 5 cleanup attempts {{(pid=65726) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 1148.450713] env[65726]: DEBUG nova.scheduler.client.report [None req-aad3fddf-3318-4cfd-8f31-0db9fa8fe254 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1148.513872] env[65726]: DEBUG nova.compute.manager [None req-e8a62c22-e16d-473f-bc47-d51d9d4949ce tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 9c0599e3-383f-46ab-809b-944cc3a4d206] Starting instance... {{(pid=65726) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 1148.644889] env[65726]: DEBUG oslo_vmware.api [None req-9230f0c2-55f9-4dd0-82d8-cd3277e657f9 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': task-5116832, 'name': ReconfigVM_Task, 'duration_secs': 0.874891} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1148.645236] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-9230f0c2-55f9-4dd0-82d8-cd3277e657f9 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 7c1188ce-8718-4719-8631-e59e7915b7aa] Reconfigured VM instance instance-0000006a to attach disk [datastore1] 7c1188ce-8718-4719-8631-e59e7915b7aa/7c1188ce-8718-4719-8631-e59e7915b7aa.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1148.645900] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0b92600a-1f9f-4617-b9cb-908ac8489de7 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.652394] env[65726]: DEBUG oslo_vmware.api [None req-9230f0c2-55f9-4dd0-82d8-cd3277e657f9 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Waiting for the task: (returnval){ [ 1148.652394] env[65726]: value = "task-5116833" [ 1148.652394] env[65726]: _type = "Task" [ 1148.652394] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1148.662362] env[65726]: DEBUG oslo_vmware.api [None req-9230f0c2-55f9-4dd0-82d8-cd3277e657f9 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': task-5116833, 'name': Rename_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1148.850156] env[65726]: DEBUG nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] [instance: f947529e-c930-4420-91a7-1e677b5ac2f1] Instance has had 0 of 5 cleanup attempts {{(pid=65726) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 1149.040741] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e8a62c22-e16d-473f-bc47-d51d9d4949ce tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1149.129058] env[65726]: DEBUG nova.compute.manager [None req-b778c6b3-2a71-4d65-8347-d07f00992900 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] [instance: 5bfbaf25-5aed-46bc-97fc-f138d08ebbde] Start spawning the instance on the hypervisor. {{(pid=65726) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1149.159277] env[65726]: DEBUG nova.virt.hardware [None req-b778c6b3-2a71-4d65-8347-d07f00992900 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T19:28:34Z,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1539f06fa3534e90acbc3a60c4b8bd3f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T19:28:35Z,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1149.159536] env[65726]: DEBUG nova.virt.hardware [None req-b778c6b3-2a71-4d65-8347-d07f00992900 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1149.159698] env[65726]: DEBUG nova.virt.hardware [None req-b778c6b3-2a71-4d65-8347-d07f00992900 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1149.159872] env[65726]: DEBUG nova.virt.hardware [None req-b778c6b3-2a71-4d65-8347-d07f00992900 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1149.160019] env[65726]: DEBUG nova.virt.hardware [None req-b778c6b3-2a71-4d65-8347-d07f00992900 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1149.160165] env[65726]: DEBUG nova.virt.hardware [None req-b778c6b3-2a71-4d65-8347-d07f00992900 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1149.160372] env[65726]: DEBUG nova.virt.hardware [None req-b778c6b3-2a71-4d65-8347-d07f00992900 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1149.160523] env[65726]: DEBUG nova.virt.hardware [None req-b778c6b3-2a71-4d65-8347-d07f00992900 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1149.160685] env[65726]: DEBUG nova.virt.hardware [None req-b778c6b3-2a71-4d65-8347-d07f00992900 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1149.160841] env[65726]: DEBUG nova.virt.hardware [None req-b778c6b3-2a71-4d65-8347-d07f00992900 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1149.161017] env[65726]: DEBUG nova.virt.hardware [None req-b778c6b3-2a71-4d65-8347-d07f00992900 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1149.161999] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc7e7879-1c84-465b-85be-00749f7d1350 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.172793] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2206ffe-8147-460a-934b-d3310bbb1170 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.178435] env[65726]: DEBUG oslo_vmware.api [None req-9230f0c2-55f9-4dd0-82d8-cd3277e657f9 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': task-5116833, 'name': Rename_Task, 'duration_secs': 0.186842} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1149.178697] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-9230f0c2-55f9-4dd0-82d8-cd3277e657f9 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 7c1188ce-8718-4719-8631-e59e7915b7aa] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1149.179282] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-cb35451b-c348-49d3-bac8-c36047a5e834 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.190268] env[65726]: DEBUG oslo_vmware.api [None req-9230f0c2-55f9-4dd0-82d8-cd3277e657f9 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Waiting for the task: (returnval){ [ 1149.190268] env[65726]: value = "task-5116834" [ 1149.190268] env[65726]: _type = "Task" [ 1149.190268] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1149.197945] env[65726]: DEBUG oslo_vmware.api [None req-9230f0c2-55f9-4dd0-82d8-cd3277e657f9 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': task-5116834, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1149.200335] env[65726]: DEBUG nova.compute.manager [req-24ec1753-64e3-4a79-9b2f-eca173d1e220 req-2907172f-51c5-423d-8a43-e4e95b23d792 service nova] [instance: 0d42fa75-20e9-4646-9b08-17015b7f068c] Received event network-vif-deleted-7a95cb45-ebfc-4ac0-ad58-4bb57170e47a {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1149.200519] env[65726]: INFO nova.compute.manager [req-24ec1753-64e3-4a79-9b2f-eca173d1e220 req-2907172f-51c5-423d-8a43-e4e95b23d792 service nova] [instance: 0d42fa75-20e9-4646-9b08-17015b7f068c] Neutron deleted interface 7a95cb45-ebfc-4ac0-ad58-4bb57170e47a; detaching it from the instance and deleting it from the info cache [ 1149.200681] env[65726]: DEBUG nova.network.neutron [req-24ec1753-64e3-4a79-9b2f-eca173d1e220 req-2907172f-51c5-423d-8a43-e4e95b23d792 service nova] [instance: 0d42fa75-20e9-4646-9b08-17015b7f068c] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1149.353587] env[65726]: DEBUG nova.network.neutron [None req-ea88cc1c-dde5-4150-976d-12646abf77ca None None] [instance: 0d42fa75-20e9-4646-9b08-17015b7f068c] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1149.354900] env[65726]: DEBUG nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] [instance: ad4f00b5-3765-494f-8132-8b17b29d219a] Instance has had 0 of 5 cleanup attempts {{(pid=65726) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 1149.415400] env[65726]: DEBUG nova.compute.manager [req-cc6d3841-c5bb-4468-b590-8b0751fef9e9 req-985e15e0-4cc5-4fe4-8ad7-87c475850b26 service nova] [instance: 5bfbaf25-5aed-46bc-97fc-f138d08ebbde] Received event network-vif-plugged-c41be8b4-e252-47c3-8529-cd6a7ce6a8d3 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1149.416862] env[65726]: DEBUG oslo_concurrency.lockutils [req-cc6d3841-c5bb-4468-b590-8b0751fef9e9 req-985e15e0-4cc5-4fe4-8ad7-87c475850b26 service nova] Acquiring lock "5bfbaf25-5aed-46bc-97fc-f138d08ebbde-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1149.416862] env[65726]: DEBUG oslo_concurrency.lockutils [req-cc6d3841-c5bb-4468-b590-8b0751fef9e9 req-985e15e0-4cc5-4fe4-8ad7-87c475850b26 service nova] Lock "5bfbaf25-5aed-46bc-97fc-f138d08ebbde-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1149.416862] env[65726]: DEBUG oslo_concurrency.lockutils [req-cc6d3841-c5bb-4468-b590-8b0751fef9e9 req-985e15e0-4cc5-4fe4-8ad7-87c475850b26 service nova] Lock "5bfbaf25-5aed-46bc-97fc-f138d08ebbde-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1149.416862] env[65726]: DEBUG nova.compute.manager [req-cc6d3841-c5bb-4468-b590-8b0751fef9e9 req-985e15e0-4cc5-4fe4-8ad7-87c475850b26 service nova] [instance: 5bfbaf25-5aed-46bc-97fc-f138d08ebbde] No waiting events found dispatching network-vif-plugged-c41be8b4-e252-47c3-8529-cd6a7ce6a8d3 {{(pid=65726) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1149.417284] env[65726]: WARNING nova.compute.manager [req-cc6d3841-c5bb-4468-b590-8b0751fef9e9 req-985e15e0-4cc5-4fe4-8ad7-87c475850b26 service nova] [instance: 5bfbaf25-5aed-46bc-97fc-f138d08ebbde] Received unexpected event network-vif-plugged-c41be8b4-e252-47c3-8529-cd6a7ce6a8d3 for instance with vm_state building and task_state spawning. [ 1149.463034] env[65726]: DEBUG oslo_concurrency.lockutils [None req-aad3fddf-3318-4cfd-8f31-0db9fa8fe254 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.355s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1149.466627] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3a8ffaf6-e93e-4fc3-a889-1a35a6c01d6d tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.507s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1149.468997] env[65726]: INFO nova.compute.claims [None req-3a8ffaf6-e93e-4fc3-a889-1a35a6c01d6d tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] [instance: 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1149.537986] env[65726]: DEBUG nova.network.neutron [None req-b778c6b3-2a71-4d65-8347-d07f00992900 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] [instance: 5bfbaf25-5aed-46bc-97fc-f138d08ebbde] Successfully updated port: c41be8b4-e252-47c3-8529-cd6a7ce6a8d3 {{(pid=65726) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 1149.703705] env[65726]: DEBUG oslo_vmware.api [None req-9230f0c2-55f9-4dd0-82d8-cd3277e657f9 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': task-5116834, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1149.703959] env[65726]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1a6d6c53-adde-4b02-bb22-59d2a3888d7e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.716278] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0b8491d-69a2-44c2-a982-b26266e27855 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.751811] env[65726]: DEBUG nova.compute.manager [req-24ec1753-64e3-4a79-9b2f-eca173d1e220 req-2907172f-51c5-423d-8a43-e4e95b23d792 service nova] [instance: 0d42fa75-20e9-4646-9b08-17015b7f068c] Detach interface failed, port_id=7a95cb45-ebfc-4ac0-ad58-4bb57170e47a, reason: Instance 0d42fa75-20e9-4646-9b08-17015b7f068c could not be found. {{(pid=65726) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11669}} [ 1149.858395] env[65726]: INFO nova.compute.manager [None req-ea88cc1c-dde5-4150-976d-12646abf77ca None None] [instance: 0d42fa75-20e9-4646-9b08-17015b7f068c] Took 2.13 seconds to deallocate network for instance. [ 1149.858807] env[65726]: DEBUG nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] [instance: df08d166-7811-49b5-9c16-a3434229d623] Instance has had 0 of 5 cleanup attempts {{(pid=65726) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 1150.033946] env[65726]: INFO nova.scheduler.client.report [None req-aad3fddf-3318-4cfd-8f31-0db9fa8fe254 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Deleted allocation for migration 59ac4637-e900-45a3-bb6b-29b88535da20 [ 1150.044660] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b778c6b3-2a71-4d65-8347-d07f00992900 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Acquiring lock "refresh_cache-5bfbaf25-5aed-46bc-97fc-f138d08ebbde" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1150.044839] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b778c6b3-2a71-4d65-8347-d07f00992900 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Acquired lock "refresh_cache-5bfbaf25-5aed-46bc-97fc-f138d08ebbde" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1150.045019] env[65726]: DEBUG nova.network.neutron [None req-b778c6b3-2a71-4d65-8347-d07f00992900 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] [instance: 5bfbaf25-5aed-46bc-97fc-f138d08ebbde] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1150.203663] env[65726]: DEBUG oslo_vmware.api [None req-9230f0c2-55f9-4dd0-82d8-cd3277e657f9 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': task-5116834, 'name': PowerOnVM_Task, 'duration_secs': 0.596743} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1150.204229] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-9230f0c2-55f9-4dd0-82d8-cd3277e657f9 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 7c1188ce-8718-4719-8631-e59e7915b7aa] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1150.204565] env[65726]: INFO nova.compute.manager [None req-9230f0c2-55f9-4dd0-82d8-cd3277e657f9 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 7c1188ce-8718-4719-8631-e59e7915b7aa] Took 8.47 seconds to spawn the instance on the hypervisor. [ 1150.204860] env[65726]: DEBUG nova.compute.manager [None req-9230f0c2-55f9-4dd0-82d8-cd3277e657f9 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 7c1188ce-8718-4719-8631-e59e7915b7aa] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1150.206103] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d292adc-4979-47d3-9912-4aa50e0af94b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.364116] env[65726]: DEBUG nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] [instance: 2c812714-dccc-4d1e-bdb5-c11e446949c3] Instance has had 0 of 5 cleanup attempts {{(pid=65726) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 1150.367262] env[65726]: DEBUG oslo_concurrency.lockutils [None req-8ea0406c-db1c-4929-acf9-e0ba5e267d6a tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1150.540187] env[65726]: DEBUG oslo_concurrency.lockutils [None req-aad3fddf-3318-4cfd-8f31-0db9fa8fe254 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Lock "e7156a10-631a-4cc5-a544-88ce37763d02" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 9.703s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1150.548829] env[65726]: WARNING openstack [None req-b778c6b3-2a71-4d65-8347-d07f00992900 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1150.549214] env[65726]: WARNING openstack [None req-b778c6b3-2a71-4d65-8347-d07f00992900 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1150.554675] env[65726]: DEBUG nova.network.neutron [None req-b778c6b3-2a71-4d65-8347-d07f00992900 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] [instance: 5bfbaf25-5aed-46bc-97fc-f138d08ebbde] Instance cache missing network info. {{(pid=65726) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 1150.620537] env[65726]: WARNING openstack [None req-b778c6b3-2a71-4d65-8347-d07f00992900 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1150.620910] env[65726]: WARNING openstack [None req-b778c6b3-2a71-4d65-8347-d07f00992900 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1150.629838] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9cad091-6db7-4395-be65-d7584bcad85e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.637778] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b38b6925-e304-4d52-9233-41fbe1374c4f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.671218] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-faa58170-7951-44d9-8fa8-a5b2ba249f58 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.678680] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2646169e-3d3c-4e23-b8cd-b71e05e212d5 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.693689] env[65726]: DEBUG nova.compute.provider_tree [None req-3a8ffaf6-e93e-4fc3-a889-1a35a6c01d6d tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1150.706664] env[65726]: WARNING openstack [None req-b778c6b3-2a71-4d65-8347-d07f00992900 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1150.707028] env[65726]: WARNING openstack [None req-b778c6b3-2a71-4d65-8347-d07f00992900 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1150.722045] env[65726]: INFO nova.compute.manager [None req-9230f0c2-55f9-4dd0-82d8-cd3277e657f9 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 7c1188ce-8718-4719-8631-e59e7915b7aa] Took 13.81 seconds to build instance. [ 1150.793649] env[65726]: DEBUG nova.network.neutron [None req-b778c6b3-2a71-4d65-8347-d07f00992900 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] [instance: 5bfbaf25-5aed-46bc-97fc-f138d08ebbde] Updating instance_info_cache with network_info: [{"id": "c41be8b4-e252-47c3-8529-cd6a7ce6a8d3", "address": "fa:16:3e:dd:aa:e3", "network": {"id": "547d7b16-4591-4d80-9e9b-c8a9a3a78dab", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-2077874536-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4ffd45f4a7a041199a4fc7f69f5e7e9b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2ed91b7b-b4ec-486d-ab34-af0afb7ec691", "external-id": "nsx-vlan-transportzone-75", "segmentation_id": 75, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc41be8b4-e2", "ovs_interfaceid": "c41be8b4-e252-47c3-8529-cd6a7ce6a8d3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1150.868917] env[65726]: DEBUG nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] [instance: 2b8eb60b-dc9b-47a4-9ddc-205f2af69888] Instance has had 0 of 5 cleanup attempts {{(pid=65726) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 1151.196907] env[65726]: DEBUG nova.scheduler.client.report [None req-3a8ffaf6-e93e-4fc3-a889-1a35a6c01d6d tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1151.224053] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9230f0c2-55f9-4dd0-82d8-cd3277e657f9 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Lock "7c1188ce-8718-4719-8631-e59e7915b7aa" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.322s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1151.297681] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b778c6b3-2a71-4d65-8347-d07f00992900 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Releasing lock "refresh_cache-5bfbaf25-5aed-46bc-97fc-f138d08ebbde" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1151.297681] env[65726]: DEBUG nova.compute.manager [None req-b778c6b3-2a71-4d65-8347-d07f00992900 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] [instance: 5bfbaf25-5aed-46bc-97fc-f138d08ebbde] Instance network_info: |[{"id": "c41be8b4-e252-47c3-8529-cd6a7ce6a8d3", "address": "fa:16:3e:dd:aa:e3", "network": {"id": "547d7b16-4591-4d80-9e9b-c8a9a3a78dab", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-2077874536-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4ffd45f4a7a041199a4fc7f69f5e7e9b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2ed91b7b-b4ec-486d-ab34-af0afb7ec691", "external-id": "nsx-vlan-transportzone-75", "segmentation_id": 75, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc41be8b4-e2", "ovs_interfaceid": "c41be8b4-e252-47c3-8529-cd6a7ce6a8d3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1151.298087] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-b778c6b3-2a71-4d65-8347-d07f00992900 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] [instance: 5bfbaf25-5aed-46bc-97fc-f138d08ebbde] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:dd:aa:e3', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '2ed91b7b-b4ec-486d-ab34-af0afb7ec691', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c41be8b4-e252-47c3-8529-cd6a7ce6a8d3', 'vif_model': 'vmxnet3'}] {{(pid=65726) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1151.307049] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-b778c6b3-2a71-4d65-8347-d07f00992900 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1151.307049] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5bfbaf25-5aed-46bc-97fc-f138d08ebbde] Creating VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1151.307049] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e1c6afaf-58fb-4cf6-9c42-a9b416274541 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.329734] env[65726]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1151.329734] env[65726]: value = "task-5116835" [ 1151.329734] env[65726]: _type = "Task" [ 1151.329734] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1151.338589] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116835, 'name': CreateVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1151.373127] env[65726]: DEBUG nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] [instance: fb8eac12-aedd-426f-9dd3-6c52f9d4c71d] Instance has had 0 of 5 cleanup attempts {{(pid=65726) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 1151.505716] env[65726]: DEBUG nova.compute.manager [req-e33f016c-8c93-4d04-8b0c-7a1238e578ea req-df957d9e-bb7d-45c2-ad43-c206bc2f1991 service nova] [instance: 5bfbaf25-5aed-46bc-97fc-f138d08ebbde] Received event network-changed-c41be8b4-e252-47c3-8529-cd6a7ce6a8d3 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1151.505920] env[65726]: DEBUG nova.compute.manager [req-e33f016c-8c93-4d04-8b0c-7a1238e578ea req-df957d9e-bb7d-45c2-ad43-c206bc2f1991 service nova] [instance: 5bfbaf25-5aed-46bc-97fc-f138d08ebbde] Refreshing instance network info cache due to event network-changed-c41be8b4-e252-47c3-8529-cd6a7ce6a8d3. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 1151.506253] env[65726]: DEBUG oslo_concurrency.lockutils [req-e33f016c-8c93-4d04-8b0c-7a1238e578ea req-df957d9e-bb7d-45c2-ad43-c206bc2f1991 service nova] Acquiring lock "refresh_cache-5bfbaf25-5aed-46bc-97fc-f138d08ebbde" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1151.506447] env[65726]: DEBUG oslo_concurrency.lockutils [req-e33f016c-8c93-4d04-8b0c-7a1238e578ea req-df957d9e-bb7d-45c2-ad43-c206bc2f1991 service nova] Acquired lock "refresh_cache-5bfbaf25-5aed-46bc-97fc-f138d08ebbde" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1151.506637] env[65726]: DEBUG nova.network.neutron [req-e33f016c-8c93-4d04-8b0c-7a1238e578ea req-df957d9e-bb7d-45c2-ad43-c206bc2f1991 service nova] [instance: 5bfbaf25-5aed-46bc-97fc-f138d08ebbde] Refreshing network info cache for port c41be8b4-e252-47c3-8529-cd6a7ce6a8d3 {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 1151.701903] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3a8ffaf6-e93e-4fc3-a889-1a35a6c01d6d tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.236s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1151.702330] env[65726]: DEBUG nova.compute.manager [None req-3a8ffaf6-e93e-4fc3-a889-1a35a6c01d6d tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] [instance: 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30] Start building networks asynchronously for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1151.705421] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e8a62c22-e16d-473f-bc47-d51d9d4949ce tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.665s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1151.706832] env[65726]: INFO nova.compute.claims [None req-e8a62c22-e16d-473f-bc47-d51d9d4949ce tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 9c0599e3-383f-46ab-809b-944cc3a4d206] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1151.840013] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116835, 'name': CreateVM_Task, 'duration_secs': 0.346859} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1151.840204] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5bfbaf25-5aed-46bc-97fc-f138d08ebbde] Created VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1151.840968] env[65726]: WARNING openstack [None req-b778c6b3-2a71-4d65-8347-d07f00992900 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1151.841332] env[65726]: WARNING openstack [None req-b778c6b3-2a71-4d65-8347-d07f00992900 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1151.846488] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b778c6b3-2a71-4d65-8347-d07f00992900 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1151.846644] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b778c6b3-2a71-4d65-8347-d07f00992900 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1151.846960] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b778c6b3-2a71-4d65-8347-d07f00992900 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1151.847250] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b26d2753-fcf0-474e-9a7d-a8196bc6584e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.852623] env[65726]: DEBUG oslo_vmware.api [None req-b778c6b3-2a71-4d65-8347-d07f00992900 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Waiting for the task: (returnval){ [ 1151.852623] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5294b6b5-1e6b-e1de-35fc-8660b6314d23" [ 1151.852623] env[65726]: _type = "Task" [ 1151.852623] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1151.863058] env[65726]: DEBUG oslo_concurrency.lockutils [None req-aad3fddf-3318-4cfd-8f31-0db9fa8fe254 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Acquiring lock "e7156a10-631a-4cc5-a544-88ce37763d02" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1151.863322] env[65726]: DEBUG oslo_concurrency.lockutils [None req-aad3fddf-3318-4cfd-8f31-0db9fa8fe254 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Lock "e7156a10-631a-4cc5-a544-88ce37763d02" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1151.863515] env[65726]: DEBUG oslo_concurrency.lockutils [None req-aad3fddf-3318-4cfd-8f31-0db9fa8fe254 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Acquiring lock "e7156a10-631a-4cc5-a544-88ce37763d02-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1151.863689] env[65726]: DEBUG oslo_concurrency.lockutils [None req-aad3fddf-3318-4cfd-8f31-0db9fa8fe254 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Lock "e7156a10-631a-4cc5-a544-88ce37763d02-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1151.863848] env[65726]: DEBUG oslo_concurrency.lockutils [None req-aad3fddf-3318-4cfd-8f31-0db9fa8fe254 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Lock "e7156a10-631a-4cc5-a544-88ce37763d02-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1151.865410] env[65726]: DEBUG oslo_vmware.api [None req-b778c6b3-2a71-4d65-8347-d07f00992900 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5294b6b5-1e6b-e1de-35fc-8660b6314d23, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1151.867154] env[65726]: INFO nova.compute.manager [None req-aad3fddf-3318-4cfd-8f31-0db9fa8fe254 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: e7156a10-631a-4cc5-a544-88ce37763d02] Terminating instance [ 1151.877915] env[65726]: DEBUG nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] [instance: 699127a9-f40f-4c1c-ba4a-625af097350b] Instance has had 0 of 5 cleanup attempts {{(pid=65726) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 1152.010037] env[65726]: WARNING openstack [req-e33f016c-8c93-4d04-8b0c-7a1238e578ea req-df957d9e-bb7d-45c2-ad43-c206bc2f1991 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1152.010454] env[65726]: WARNING openstack [req-e33f016c-8c93-4d04-8b0c-7a1238e578ea req-df957d9e-bb7d-45c2-ad43-c206bc2f1991 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1152.112619] env[65726]: WARNING openstack [req-e33f016c-8c93-4d04-8b0c-7a1238e578ea req-df957d9e-bb7d-45c2-ad43-c206bc2f1991 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1152.113022] env[65726]: WARNING openstack [req-e33f016c-8c93-4d04-8b0c-7a1238e578ea req-df957d9e-bb7d-45c2-ad43-c206bc2f1991 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1152.173618] env[65726]: WARNING openstack [req-e33f016c-8c93-4d04-8b0c-7a1238e578ea req-df957d9e-bb7d-45c2-ad43-c206bc2f1991 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1152.174010] env[65726]: WARNING openstack [req-e33f016c-8c93-4d04-8b0c-7a1238e578ea req-df957d9e-bb7d-45c2-ad43-c206bc2f1991 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1152.213751] env[65726]: DEBUG nova.compute.utils [None req-3a8ffaf6-e93e-4fc3-a889-1a35a6c01d6d tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] Using /dev/sd instead of None {{(pid=65726) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1152.219732] env[65726]: DEBUG nova.compute.manager [None req-3a8ffaf6-e93e-4fc3-a889-1a35a6c01d6d tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] [instance: 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30] Allocating IP information in the background. {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1152.219928] env[65726]: DEBUG nova.network.neutron [None req-3a8ffaf6-e93e-4fc3-a889-1a35a6c01d6d tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] [instance: 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30] allocate_for_instance() {{(pid=65726) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 1152.220275] env[65726]: WARNING neutronclient.v2_0.client [None req-3a8ffaf6-e93e-4fc3-a889-1a35a6c01d6d tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1152.220782] env[65726]: WARNING neutronclient.v2_0.client [None req-3a8ffaf6-e93e-4fc3-a889-1a35a6c01d6d tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1152.221509] env[65726]: WARNING openstack [None req-3a8ffaf6-e93e-4fc3-a889-1a35a6c01d6d tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1152.221872] env[65726]: WARNING openstack [None req-3a8ffaf6-e93e-4fc3-a889-1a35a6c01d6d tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1152.230266] env[65726]: DEBUG nova.compute.manager [None req-3a8ffaf6-e93e-4fc3-a889-1a35a6c01d6d tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] [instance: 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30] Start building block device mappings for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1152.266256] env[65726]: DEBUG nova.policy [None req-3a8ffaf6-e93e-4fc3-a889-1a35a6c01d6d tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2ae0a28b338a4a22a0e4b6b427a28ef2', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1e3203c117de4d4d9c4c90436801ff3f', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65726) authorize /opt/stack/nova/nova/policy.py:192}} [ 1152.270097] env[65726]: DEBUG nova.network.neutron [req-e33f016c-8c93-4d04-8b0c-7a1238e578ea req-df957d9e-bb7d-45c2-ad43-c206bc2f1991 service nova] [instance: 5bfbaf25-5aed-46bc-97fc-f138d08ebbde] Updated VIF entry in instance network info cache for port c41be8b4-e252-47c3-8529-cd6a7ce6a8d3. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 1152.270476] env[65726]: DEBUG nova.network.neutron [req-e33f016c-8c93-4d04-8b0c-7a1238e578ea req-df957d9e-bb7d-45c2-ad43-c206bc2f1991 service nova] [instance: 5bfbaf25-5aed-46bc-97fc-f138d08ebbde] Updating instance_info_cache with network_info: [{"id": "c41be8b4-e252-47c3-8529-cd6a7ce6a8d3", "address": "fa:16:3e:dd:aa:e3", "network": {"id": "547d7b16-4591-4d80-9e9b-c8a9a3a78dab", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-2077874536-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4ffd45f4a7a041199a4fc7f69f5e7e9b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2ed91b7b-b4ec-486d-ab34-af0afb7ec691", "external-id": "nsx-vlan-transportzone-75", "segmentation_id": 75, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc41be8b4-e2", "ovs_interfaceid": "c41be8b4-e252-47c3-8529-cd6a7ce6a8d3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1152.364073] env[65726]: DEBUG oslo_vmware.api [None req-b778c6b3-2a71-4d65-8347-d07f00992900 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5294b6b5-1e6b-e1de-35fc-8660b6314d23, 'name': SearchDatastore_Task, 'duration_secs': 0.011845} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1152.364418] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b778c6b3-2a71-4d65-8347-d07f00992900 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1152.365022] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-b778c6b3-2a71-4d65-8347-d07f00992900 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] [instance: 5bfbaf25-5aed-46bc-97fc-f138d08ebbde] Processing image b52362a3-ee8a-4cbf-b06f-513b0cc8f95c {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1152.365273] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b778c6b3-2a71-4d65-8347-d07f00992900 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1152.365412] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b778c6b3-2a71-4d65-8347-d07f00992900 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1152.366028] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-b778c6b3-2a71-4d65-8347-d07f00992900 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1152.366028] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b87d8aff-b275-4e54-9e43-7b71050b24d6 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.370811] env[65726]: DEBUG nova.compute.manager [None req-aad3fddf-3318-4cfd-8f31-0db9fa8fe254 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: e7156a10-631a-4cc5-a544-88ce37763d02] Start destroying the instance on the hypervisor. {{(pid=65726) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1152.371109] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-aad3fddf-3318-4cfd-8f31-0db9fa8fe254 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: e7156a10-631a-4cc5-a544-88ce37763d02] Destroying instance {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1152.371989] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b507292-cbd4-4415-8ad8-1a1d016c1be5 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.376867] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-b778c6b3-2a71-4d65-8347-d07f00992900 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1152.377252] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-b778c6b3-2a71-4d65-8347-d07f00992900 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65726) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1152.378406] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cc2deba6-937d-495c-a553-401a48dc1d27 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.383999] env[65726]: DEBUG nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] [instance: a286ba4f-a83e-4c30-8079-ba88e3ba5696] Instance has had 0 of 5 cleanup attempts {{(pid=65726) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 1152.386344] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-aad3fddf-3318-4cfd-8f31-0db9fa8fe254 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: e7156a10-631a-4cc5-a544-88ce37763d02] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1152.387150] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-80fe7ecb-3f2f-4f16-b423-6493ab33420f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.390803] env[65726]: DEBUG oslo_vmware.api [None req-b778c6b3-2a71-4d65-8347-d07f00992900 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Waiting for the task: (returnval){ [ 1152.390803] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]521c0930-9b4f-5573-4368-2dc986cf8487" [ 1152.390803] env[65726]: _type = "Task" [ 1152.390803] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1152.398572] env[65726]: DEBUG oslo_vmware.api [None req-aad3fddf-3318-4cfd-8f31-0db9fa8fe254 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Waiting for the task: (returnval){ [ 1152.398572] env[65726]: value = "task-5116836" [ 1152.398572] env[65726]: _type = "Task" [ 1152.398572] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1152.402643] env[65726]: DEBUG oslo_vmware.api [None req-b778c6b3-2a71-4d65-8347-d07f00992900 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]521c0930-9b4f-5573-4368-2dc986cf8487, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1152.416926] env[65726]: DEBUG oslo_vmware.api [None req-aad3fddf-3318-4cfd-8f31-0db9fa8fe254 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': task-5116836, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1152.629449] env[65726]: DEBUG nova.network.neutron [None req-3a8ffaf6-e93e-4fc3-a889-1a35a6c01d6d tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] [instance: 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30] Successfully created port: e0f18f03-6a1b-4d5e-9b8e-68051f61ad8b {{(pid=65726) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1152.735671] env[65726]: INFO nova.virt.block_device [None req-3a8ffaf6-e93e-4fc3-a889-1a35a6c01d6d tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] [instance: 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30] Booting with volume 39918e7e-1d4a-45df-911f-46ed9091f96b at /dev/sda [ 1152.777180] env[65726]: DEBUG oslo_concurrency.lockutils [req-e33f016c-8c93-4d04-8b0c-7a1238e578ea req-df957d9e-bb7d-45c2-ad43-c206bc2f1991 service nova] Releasing lock "refresh_cache-5bfbaf25-5aed-46bc-97fc-f138d08ebbde" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1152.777180] env[65726]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ea8d906c-93fd-43bd-8fc1-b342b2b2595b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.790209] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86144d5b-1f83-44bb-b3c2-cecfb4eede0d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.825344] env[65726]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-eb676a74-babf-42bc-b188-ef35e46e0158 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.835907] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-333aee99-faf3-4ea8-bdbe-225cbf11f05b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.875647] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da932e2a-dad5-4819-90c8-6b2b79f9207b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.884509] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41c15d17-a65d-4f71-bea2-ff15120e0a17 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.890771] env[65726]: DEBUG nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] [instance: d8468efa-c2d3-4dce-ab89-fc077011e3d8] Instance has had 0 of 5 cleanup attempts {{(pid=65726) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 1152.900454] env[65726]: DEBUG nova.virt.block_device [None req-3a8ffaf6-e93e-4fc3-a889-1a35a6c01d6d tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] [instance: 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30] Updating existing volume attachment record: 460c5ba7-8f25-40cb-a2a3-ca1443ab65dc {{(pid=65726) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1152.907229] env[65726]: DEBUG oslo_vmware.api [None req-b778c6b3-2a71-4d65-8347-d07f00992900 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]521c0930-9b4f-5573-4368-2dc986cf8487, 'name': SearchDatastore_Task, 'duration_secs': 0.012155} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1152.915048] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-435490f7-95e2-453b-99fe-5a1fe69ff81a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.925360] env[65726]: DEBUG oslo_vmware.api [None req-aad3fddf-3318-4cfd-8f31-0db9fa8fe254 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': task-5116836, 'name': PowerOffVM_Task, 'duration_secs': 0.258069} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1152.928990] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-aad3fddf-3318-4cfd-8f31-0db9fa8fe254 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: e7156a10-631a-4cc5-a544-88ce37763d02] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1152.929201] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-aad3fddf-3318-4cfd-8f31-0db9fa8fe254 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: e7156a10-631a-4cc5-a544-88ce37763d02] Unregistering the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1152.929998] env[65726]: DEBUG oslo_vmware.api [None req-b778c6b3-2a71-4d65-8347-d07f00992900 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Waiting for the task: (returnval){ [ 1152.929998] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52eaff85-8a15-af30-d3fa-38206f101c10" [ 1152.929998] env[65726]: _type = "Task" [ 1152.929998] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1152.932617] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-74bb9f4e-5c41-43bc-b7ad-81fdad83929c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.934793] env[65726]: DEBUG nova.compute.manager [None req-168947ab-de46-4966-89ee-e38c38aa73bc tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 7c1188ce-8718-4719-8631-e59e7915b7aa] Stashing vm_state: active {{(pid=65726) _prep_resize /opt/stack/nova/nova/compute/manager.py:6193}} [ 1152.948881] env[65726]: DEBUG oslo_vmware.api [None req-b778c6b3-2a71-4d65-8347-d07f00992900 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52eaff85-8a15-af30-d3fa-38206f101c10, 'name': SearchDatastore_Task, 'duration_secs': 0.011067} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1152.949156] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b778c6b3-2a71-4d65-8347-d07f00992900 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1152.949429] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-b778c6b3-2a71-4d65-8347-d07f00992900 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore2] 5bfbaf25-5aed-46bc-97fc-f138d08ebbde/5bfbaf25-5aed-46bc-97fc-f138d08ebbde.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1152.949818] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7bd3bf72-79f7-4bec-aa41-7b5f8d383581 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.960441] env[65726]: DEBUG oslo_vmware.api [None req-b778c6b3-2a71-4d65-8347-d07f00992900 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Waiting for the task: (returnval){ [ 1152.960441] env[65726]: value = "task-5116838" [ 1152.960441] env[65726]: _type = "Task" [ 1152.960441] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1152.970639] env[65726]: DEBUG oslo_vmware.api [None req-b778c6b3-2a71-4d65-8347-d07f00992900 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Task: {'id': task-5116838, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1152.973076] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c7a0ea5-e365-41f7-8ff0-cffa3e081aac {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.981354] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b16314c-9c7e-44f4-a61b-02ab8554e587 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.014914] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e589f668-4bb3-45a4-ba21-44d47c47697a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.021513] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-aad3fddf-3318-4cfd-8f31-0db9fa8fe254 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: e7156a10-631a-4cc5-a544-88ce37763d02] Unregistered the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1153.021774] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-aad3fddf-3318-4cfd-8f31-0db9fa8fe254 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: e7156a10-631a-4cc5-a544-88ce37763d02] Deleting contents of the VM from datastore datastore2 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1153.021989] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-aad3fddf-3318-4cfd-8f31-0db9fa8fe254 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Deleting the datastore file [datastore2] e7156a10-631a-4cc5-a544-88ce37763d02 {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1153.024508] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e31e55f7-94fb-4f5b-825c-89fd28e1847b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.028285] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9fe7878-3ec5-4b5a-bee8-04f8784b4f7c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.043141] env[65726]: DEBUG nova.compute.provider_tree [None req-e8a62c22-e16d-473f-bc47-d51d9d4949ce tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1153.046643] env[65726]: DEBUG oslo_vmware.api [None req-aad3fddf-3318-4cfd-8f31-0db9fa8fe254 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Waiting for the task: (returnval){ [ 1153.046643] env[65726]: value = "task-5116839" [ 1153.046643] env[65726]: _type = "Task" [ 1153.046643] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1153.056672] env[65726]: DEBUG oslo_vmware.api [None req-aad3fddf-3318-4cfd-8f31-0db9fa8fe254 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': task-5116839, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1153.396890] env[65726]: DEBUG nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] [instance: a56be6fa-6027-46ee-9aa8-332fe4222d68] Instance has had 0 of 5 cleanup attempts {{(pid=65726) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 1153.466145] env[65726]: DEBUG oslo_concurrency.lockutils [None req-168947ab-de46-4966-89ee-e38c38aa73bc tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1153.471852] env[65726]: DEBUG oslo_vmware.api [None req-b778c6b3-2a71-4d65-8347-d07f00992900 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Task: {'id': task-5116838, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1153.548923] env[65726]: DEBUG nova.scheduler.client.report [None req-e8a62c22-e16d-473f-bc47-d51d9d4949ce tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1153.563243] env[65726]: DEBUG oslo_vmware.api [None req-aad3fddf-3318-4cfd-8f31-0db9fa8fe254 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Task: {'id': task-5116839, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.498769} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1153.563384] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-aad3fddf-3318-4cfd-8f31-0db9fa8fe254 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Deleted the datastore file {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1153.563586] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-aad3fddf-3318-4cfd-8f31-0db9fa8fe254 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: e7156a10-631a-4cc5-a544-88ce37763d02] Deleted contents of the VM from datastore datastore2 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1153.563759] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-aad3fddf-3318-4cfd-8f31-0db9fa8fe254 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: e7156a10-631a-4cc5-a544-88ce37763d02] Instance destroyed {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1153.563929] env[65726]: INFO nova.compute.manager [None req-aad3fddf-3318-4cfd-8f31-0db9fa8fe254 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] [instance: e7156a10-631a-4cc5-a544-88ce37763d02] Took 1.19 seconds to destroy the instance on the hypervisor. [ 1153.564196] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-aad3fddf-3318-4cfd-8f31-0db9fa8fe254 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1153.564395] env[65726]: DEBUG nova.compute.manager [-] [instance: e7156a10-631a-4cc5-a544-88ce37763d02] Deallocating network for instance {{(pid=65726) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1153.564502] env[65726]: DEBUG nova.network.neutron [-] [instance: e7156a10-631a-4cc5-a544-88ce37763d02] deallocate_for_instance() {{(pid=65726) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 1153.565022] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1153.565283] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1153.640840] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1153.641140] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1153.901333] env[65726]: DEBUG nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] [instance: f9a40d36-dda9-454c-8774-ed011f3f50ae] Instance has had 0 of 5 cleanup attempts {{(pid=65726) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 1153.975822] env[65726]: DEBUG oslo_vmware.api [None req-b778c6b3-2a71-4d65-8347-d07f00992900 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Task: {'id': task-5116838, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.571323} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1153.978763] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-b778c6b3-2a71-4d65-8347-d07f00992900 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore2] 5bfbaf25-5aed-46bc-97fc-f138d08ebbde/5bfbaf25-5aed-46bc-97fc-f138d08ebbde.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1153.978763] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-b778c6b3-2a71-4d65-8347-d07f00992900 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] [instance: 5bfbaf25-5aed-46bc-97fc-f138d08ebbde] Extending root virtual disk to 1048576 {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1153.978763] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-01b7ce68-8bfa-446e-a959-417ca28b640d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.988020] env[65726]: DEBUG oslo_vmware.api [None req-b778c6b3-2a71-4d65-8347-d07f00992900 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Waiting for the task: (returnval){ [ 1153.988020] env[65726]: value = "task-5116840" [ 1153.988020] env[65726]: _type = "Task" [ 1153.988020] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1153.997447] env[65726]: DEBUG oslo_vmware.api [None req-b778c6b3-2a71-4d65-8347-d07f00992900 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Task: {'id': task-5116840, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1154.033105] env[65726]: DEBUG nova.compute.manager [req-2df6d046-8d1d-49ca-83eb-c110445c412c req-a3dee949-f353-4eb2-9b4c-040d04734613 service nova] [instance: e7156a10-631a-4cc5-a544-88ce37763d02] Received event network-vif-deleted-44a1554d-d380-49b2-ad33-9a251114c93c {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1154.033385] env[65726]: INFO nova.compute.manager [req-2df6d046-8d1d-49ca-83eb-c110445c412c req-a3dee949-f353-4eb2-9b4c-040d04734613 service nova] [instance: e7156a10-631a-4cc5-a544-88ce37763d02] Neutron deleted interface 44a1554d-d380-49b2-ad33-9a251114c93c; detaching it from the instance and deleting it from the info cache [ 1154.033618] env[65726]: DEBUG nova.network.neutron [req-2df6d046-8d1d-49ca-83eb-c110445c412c req-a3dee949-f353-4eb2-9b4c-040d04734613 service nova] [instance: e7156a10-631a-4cc5-a544-88ce37763d02] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1154.058171] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e8a62c22-e16d-473f-bc47-d51d9d4949ce tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.352s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1154.059072] env[65726]: DEBUG nova.compute.manager [None req-e8a62c22-e16d-473f-bc47-d51d9d4949ce tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 9c0599e3-383f-46ab-809b-944cc3a4d206] Start building networks asynchronously for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1154.062624] env[65726]: DEBUG oslo_concurrency.lockutils [None req-8ea0406c-db1c-4929-acf9-e0ba5e267d6a tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.695s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1154.062940] env[65726]: DEBUG nova.objects.instance [None req-8ea0406c-db1c-4929-acf9-e0ba5e267d6a tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Lazy-loading 'resources' on Instance uuid 0d42fa75-20e9-4646-9b08-17015b7f068c {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1154.087112] env[65726]: DEBUG nova.compute.manager [req-d542c7c8-d18a-4364-874d-6ef2a1286d92 req-9711cbb0-e300-471b-97eb-ecca0d6ad299 service nova] [instance: 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30] Received event network-vif-plugged-e0f18f03-6a1b-4d5e-9b8e-68051f61ad8b {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1154.087112] env[65726]: DEBUG oslo_concurrency.lockutils [req-d542c7c8-d18a-4364-874d-6ef2a1286d92 req-9711cbb0-e300-471b-97eb-ecca0d6ad299 service nova] Acquiring lock "47e52bc1-c18e-4aa2-82a2-e4cb030a7a30-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1154.087271] env[65726]: DEBUG oslo_concurrency.lockutils [req-d542c7c8-d18a-4364-874d-6ef2a1286d92 req-9711cbb0-e300-471b-97eb-ecca0d6ad299 service nova] Lock "47e52bc1-c18e-4aa2-82a2-e4cb030a7a30-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1154.087438] env[65726]: DEBUG oslo_concurrency.lockutils [req-d542c7c8-d18a-4364-874d-6ef2a1286d92 req-9711cbb0-e300-471b-97eb-ecca0d6ad299 service nova] Lock "47e52bc1-c18e-4aa2-82a2-e4cb030a7a30-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1154.087602] env[65726]: DEBUG nova.compute.manager [req-d542c7c8-d18a-4364-874d-6ef2a1286d92 req-9711cbb0-e300-471b-97eb-ecca0d6ad299 service nova] [instance: 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30] No waiting events found dispatching network-vif-plugged-e0f18f03-6a1b-4d5e-9b8e-68051f61ad8b {{(pid=65726) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1154.087766] env[65726]: WARNING nova.compute.manager [req-d542c7c8-d18a-4364-874d-6ef2a1286d92 req-9711cbb0-e300-471b-97eb-ecca0d6ad299 service nova] [instance: 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30] Received unexpected event network-vif-plugged-e0f18f03-6a1b-4d5e-9b8e-68051f61ad8b for instance with vm_state building and task_state block_device_mapping. [ 1154.169703] env[65726]: DEBUG nova.network.neutron [None req-3a8ffaf6-e93e-4fc3-a889-1a35a6c01d6d tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] [instance: 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30] Successfully updated port: e0f18f03-6a1b-4d5e-9b8e-68051f61ad8b {{(pid=65726) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 1154.404074] env[65726]: DEBUG nova.network.neutron [-] [instance: e7156a10-631a-4cc5-a544-88ce37763d02] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1154.405694] env[65726]: DEBUG nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] [instance: 406c59e7-8eb8-4cfd-8e1e-e151ab908571] Instance has had 0 of 5 cleanup attempts {{(pid=65726) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 1154.496315] env[65726]: DEBUG oslo_vmware.api [None req-b778c6b3-2a71-4d65-8347-d07f00992900 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Task: {'id': task-5116840, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.083586} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1154.496641] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-b778c6b3-2a71-4d65-8347-d07f00992900 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] [instance: 5bfbaf25-5aed-46bc-97fc-f138d08ebbde] Extended root virtual disk {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1154.497470] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae9dfaf3-d209-4692-bcca-a1bd48e550a5 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.523769] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-b778c6b3-2a71-4d65-8347-d07f00992900 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] [instance: 5bfbaf25-5aed-46bc-97fc-f138d08ebbde] Reconfiguring VM instance instance-0000006b to attach disk [datastore2] 5bfbaf25-5aed-46bc-97fc-f138d08ebbde/5bfbaf25-5aed-46bc-97fc-f138d08ebbde.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1154.524367] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-83c67582-6759-4993-8b1f-fe3a98e08741 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.539091] env[65726]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ae4c1a30-79c2-4062-a9f2-b6292ebed72c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.550017] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3410e4b-64a6-4a0f-a65b-897d2b747d31 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.562773] env[65726]: DEBUG oslo_vmware.api [None req-b778c6b3-2a71-4d65-8347-d07f00992900 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Waiting for the task: (returnval){ [ 1154.562773] env[65726]: value = "task-5116841" [ 1154.562773] env[65726]: _type = "Task" [ 1154.562773] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1154.569870] env[65726]: DEBUG nova.compute.utils [None req-e8a62c22-e16d-473f-bc47-d51d9d4949ce tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Using /dev/sd instead of None {{(pid=65726) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1154.575054] env[65726]: DEBUG nova.compute.manager [None req-e8a62c22-e16d-473f-bc47-d51d9d4949ce tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 9c0599e3-383f-46ab-809b-944cc3a4d206] Allocating IP information in the background. {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1154.575626] env[65726]: DEBUG nova.network.neutron [None req-e8a62c22-e16d-473f-bc47-d51d9d4949ce tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 9c0599e3-383f-46ab-809b-944cc3a4d206] allocate_for_instance() {{(pid=65726) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 1154.575699] env[65726]: WARNING neutronclient.v2_0.client [None req-e8a62c22-e16d-473f-bc47-d51d9d4949ce tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1154.576065] env[65726]: WARNING neutronclient.v2_0.client [None req-e8a62c22-e16d-473f-bc47-d51d9d4949ce tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1154.576745] env[65726]: WARNING openstack [None req-e8a62c22-e16d-473f-bc47-d51d9d4949ce tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1154.577484] env[65726]: WARNING openstack [None req-e8a62c22-e16d-473f-bc47-d51d9d4949ce tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1154.596674] env[65726]: DEBUG nova.compute.manager [req-2df6d046-8d1d-49ca-83eb-c110445c412c req-a3dee949-f353-4eb2-9b4c-040d04734613 service nova] [instance: e7156a10-631a-4cc5-a544-88ce37763d02] Detach interface failed, port_id=44a1554d-d380-49b2-ad33-9a251114c93c, reason: Instance e7156a10-631a-4cc5-a544-88ce37763d02 could not be found. {{(pid=65726) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11669}} [ 1154.602739] env[65726]: DEBUG oslo_vmware.api [None req-b778c6b3-2a71-4d65-8347-d07f00992900 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Task: {'id': task-5116841, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1154.649134] env[65726]: DEBUG nova.policy [None req-e8a62c22-e16d-473f-bc47-d51d9d4949ce tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ccb3513b5b4f453f842139ec1815c4af', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f3f799aa94f64e8cb26d93f6124efd81', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65726) authorize /opt/stack/nova/nova/policy.py:192}} [ 1154.672145] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3a8ffaf6-e93e-4fc3-a889-1a35a6c01d6d tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] Acquiring lock "refresh_cache-47e52bc1-c18e-4aa2-82a2-e4cb030a7a30" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1154.672449] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3a8ffaf6-e93e-4fc3-a889-1a35a6c01d6d tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] Acquired lock "refresh_cache-47e52bc1-c18e-4aa2-82a2-e4cb030a7a30" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1154.673508] env[65726]: DEBUG nova.network.neutron [None req-3a8ffaf6-e93e-4fc3-a889-1a35a6c01d6d tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] [instance: 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1154.748291] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f4ae950-651c-48b3-826d-944dfee97f57 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.757182] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36f1ece1-1e3f-4ebc-b70c-66eb41f2f83d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.791136] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3042c8a8-fb9b-4f40-8b16-9d0cddd1ef38 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.800568] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c7101d4-6837-4507-a4d9-97e7b71983ce {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.818676] env[65726]: DEBUG nova.compute.provider_tree [None req-8ea0406c-db1c-4929-acf9-e0ba5e267d6a tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1154.904213] env[65726]: DEBUG nova.network.neutron [None req-e8a62c22-e16d-473f-bc47-d51d9d4949ce tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 9c0599e3-383f-46ab-809b-944cc3a4d206] Successfully created port: c5730368-8477-43fb-ae16-4a4e35a403ab {{(pid=65726) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1154.908712] env[65726]: INFO nova.compute.manager [-] [instance: e7156a10-631a-4cc5-a544-88ce37763d02] Took 1.34 seconds to deallocate network for instance. [ 1154.909091] env[65726]: DEBUG nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] [instance: 54e85161-5f63-405f-83f8-490c70645a3b] Instance has had 0 of 5 cleanup attempts {{(pid=65726) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 1155.025692] env[65726]: DEBUG nova.compute.manager [None req-3a8ffaf6-e93e-4fc3-a889-1a35a6c01d6d tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] [instance: 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30] Start spawning the instance on the hypervisor. {{(pid=65726) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1155.026282] env[65726]: DEBUG nova.virt.hardware [None req-3a8ffaf6-e93e-4fc3-a889-1a35a6c01d6d tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1155.026491] env[65726]: DEBUG nova.virt.hardware [None req-3a8ffaf6-e93e-4fc3-a889-1a35a6c01d6d tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1155.026723] env[65726]: DEBUG nova.virt.hardware [None req-3a8ffaf6-e93e-4fc3-a889-1a35a6c01d6d tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1155.026911] env[65726]: DEBUG nova.virt.hardware [None req-3a8ffaf6-e93e-4fc3-a889-1a35a6c01d6d tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1155.027120] env[65726]: DEBUG nova.virt.hardware [None req-3a8ffaf6-e93e-4fc3-a889-1a35a6c01d6d tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1155.027353] env[65726]: DEBUG nova.virt.hardware [None req-3a8ffaf6-e93e-4fc3-a889-1a35a6c01d6d tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1155.027698] env[65726]: DEBUG nova.virt.hardware [None req-3a8ffaf6-e93e-4fc3-a889-1a35a6c01d6d tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1155.028083] env[65726]: DEBUG nova.virt.hardware [None req-3a8ffaf6-e93e-4fc3-a889-1a35a6c01d6d tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1155.028469] env[65726]: DEBUG nova.virt.hardware [None req-3a8ffaf6-e93e-4fc3-a889-1a35a6c01d6d tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1155.028605] env[65726]: DEBUG nova.virt.hardware [None req-3a8ffaf6-e93e-4fc3-a889-1a35a6c01d6d tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1155.028787] env[65726]: DEBUG nova.virt.hardware [None req-3a8ffaf6-e93e-4fc3-a889-1a35a6c01d6d tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1155.030172] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75812f60-d105-4b90-b44b-597bb5f09f62 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.039354] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28d9ba79-adb6-4bc9-b0c4-29174ca1485c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.074536] env[65726]: DEBUG oslo_vmware.api [None req-b778c6b3-2a71-4d65-8347-d07f00992900 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Task: {'id': task-5116841, 'name': ReconfigVM_Task, 'duration_secs': 0.304792} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1155.074860] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-b778c6b3-2a71-4d65-8347-d07f00992900 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] [instance: 5bfbaf25-5aed-46bc-97fc-f138d08ebbde] Reconfigured VM instance instance-0000006b to attach disk [datastore2] 5bfbaf25-5aed-46bc-97fc-f138d08ebbde/5bfbaf25-5aed-46bc-97fc-f138d08ebbde.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1155.075789] env[65726]: DEBUG nova.compute.manager [None req-e8a62c22-e16d-473f-bc47-d51d9d4949ce tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 9c0599e3-383f-46ab-809b-944cc3a4d206] Start building block device mappings for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1155.078261] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0721703b-ace9-4ade-8c7a-e25dd9ebc684 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.087350] env[65726]: DEBUG oslo_vmware.api [None req-b778c6b3-2a71-4d65-8347-d07f00992900 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Waiting for the task: (returnval){ [ 1155.087350] env[65726]: value = "task-5116842" [ 1155.087350] env[65726]: _type = "Task" [ 1155.087350] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1155.097207] env[65726]: DEBUG oslo_vmware.api [None req-b778c6b3-2a71-4d65-8347-d07f00992900 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Task: {'id': task-5116842, 'name': Rename_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1155.176993] env[65726]: WARNING openstack [None req-3a8ffaf6-e93e-4fc3-a889-1a35a6c01d6d tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1155.177435] env[65726]: WARNING openstack [None req-3a8ffaf6-e93e-4fc3-a889-1a35a6c01d6d tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1155.183146] env[65726]: DEBUG nova.network.neutron [None req-3a8ffaf6-e93e-4fc3-a889-1a35a6c01d6d tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] [instance: 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30] Instance cache missing network info. {{(pid=65726) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 1155.252710] env[65726]: WARNING openstack [None req-3a8ffaf6-e93e-4fc3-a889-1a35a6c01d6d tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1155.253124] env[65726]: WARNING openstack [None req-3a8ffaf6-e93e-4fc3-a889-1a35a6c01d6d tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1155.323294] env[65726]: WARNING openstack [None req-3a8ffaf6-e93e-4fc3-a889-1a35a6c01d6d tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1155.323741] env[65726]: WARNING openstack [None req-3a8ffaf6-e93e-4fc3-a889-1a35a6c01d6d tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1155.331896] env[65726]: DEBUG nova.scheduler.client.report [None req-8ea0406c-db1c-4929-acf9-e0ba5e267d6a tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1155.413336] env[65726]: DEBUG nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] [instance: aa6ce489-c62f-4481-87b7-e74242aeb8ca] Instance has had 0 of 5 cleanup attempts {{(pid=65726) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 1155.417254] env[65726]: DEBUG oslo_concurrency.lockutils [None req-aad3fddf-3318-4cfd-8f31-0db9fa8fe254 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1155.424113] env[65726]: DEBUG nova.network.neutron [None req-3a8ffaf6-e93e-4fc3-a889-1a35a6c01d6d tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] [instance: 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30] Updating instance_info_cache with network_info: [{"id": "e0f18f03-6a1b-4d5e-9b8e-68051f61ad8b", "address": "fa:16:3e:39:d3:8c", "network": {"id": "ba97c3c2-c46b-4ce9-8f4d-b20857d6f8ec", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-992165785-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1e3203c117de4d4d9c4c90436801ff3f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f856fca-9fb5-41ea-a057-ac4193bd323d", "external-id": "nsx-vlan-transportzone-148", "segmentation_id": 148, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape0f18f03-6a", "ovs_interfaceid": "e0f18f03-6a1b-4d5e-9b8e-68051f61ad8b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1155.598275] env[65726]: DEBUG oslo_vmware.api [None req-b778c6b3-2a71-4d65-8347-d07f00992900 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Task: {'id': task-5116842, 'name': Rename_Task, 'duration_secs': 0.261249} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1155.598599] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-b778c6b3-2a71-4d65-8347-d07f00992900 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] [instance: 5bfbaf25-5aed-46bc-97fc-f138d08ebbde] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1155.598889] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a6401675-860d-43e8-a21f-471464321aec {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.607614] env[65726]: DEBUG oslo_vmware.api [None req-b778c6b3-2a71-4d65-8347-d07f00992900 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Waiting for the task: (returnval){ [ 1155.607614] env[65726]: value = "task-5116843" [ 1155.607614] env[65726]: _type = "Task" [ 1155.607614] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1155.617730] env[65726]: DEBUG oslo_vmware.api [None req-b778c6b3-2a71-4d65-8347-d07f00992900 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Task: {'id': task-5116843, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1155.837437] env[65726]: DEBUG oslo_concurrency.lockutils [None req-8ea0406c-db1c-4929-acf9-e0ba5e267d6a tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.775s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1155.839870] env[65726]: DEBUG oslo_concurrency.lockutils [None req-168947ab-de46-4966-89ee-e38c38aa73bc tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 2.374s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1155.861175] env[65726]: INFO nova.scheduler.client.report [None req-8ea0406c-db1c-4929-acf9-e0ba5e267d6a tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Deleted allocations for instance 0d42fa75-20e9-4646-9b08-17015b7f068c [ 1155.919368] env[65726]: DEBUG nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] [instance: 73f5b23e-4463-4068-8994-dd2752a2abc9] Instance has had 0 of 5 cleanup attempts {{(pid=65726) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 1155.926077] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3a8ffaf6-e93e-4fc3-a889-1a35a6c01d6d tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] Releasing lock "refresh_cache-47e52bc1-c18e-4aa2-82a2-e4cb030a7a30" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1155.926428] env[65726]: DEBUG nova.compute.manager [None req-3a8ffaf6-e93e-4fc3-a889-1a35a6c01d6d tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] [instance: 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30] Instance network_info: |[{"id": "e0f18f03-6a1b-4d5e-9b8e-68051f61ad8b", "address": "fa:16:3e:39:d3:8c", "network": {"id": "ba97c3c2-c46b-4ce9-8f4d-b20857d6f8ec", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-992165785-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1e3203c117de4d4d9c4c90436801ff3f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f856fca-9fb5-41ea-a057-ac4193bd323d", "external-id": "nsx-vlan-transportzone-148", "segmentation_id": 148, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape0f18f03-6a", "ovs_interfaceid": "e0f18f03-6a1b-4d5e-9b8e-68051f61ad8b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1155.926942] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-3a8ffaf6-e93e-4fc3-a889-1a35a6c01d6d tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] [instance: 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:39:d3:8c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '9f856fca-9fb5-41ea-a057-ac4193bd323d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e0f18f03-6a1b-4d5e-9b8e-68051f61ad8b', 'vif_model': 'vmxnet3'}] {{(pid=65726) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1155.934842] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-3a8ffaf6-e93e-4fc3-a889-1a35a6c01d6d tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] Creating folder: Project (1e3203c117de4d4d9c4c90436801ff3f). Parent ref: group-v995008. {{(pid=65726) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1155.937063] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5272d5ea-1b43-4942-bf47-c4b7e115fa0d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.951896] env[65726]: WARNING suds.client [-] Web service reported a SOAP processing fault using an unexpected HTTP status code 200. Reporting as an internal server error. [ 1155.952101] env[65726]: DEBUG oslo_vmware.api [-] Fault list: [DuplicateName] {{(pid=65726) _invoke_api /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:337}} [ 1155.952776] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-3a8ffaf6-e93e-4fc3-a889-1a35a6c01d6d tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] Folder already exists: Project (1e3203c117de4d4d9c4c90436801ff3f). Parent ref: group-v995008. {{(pid=65726) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1609}} [ 1155.953037] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-3a8ffaf6-e93e-4fc3-a889-1a35a6c01d6d tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] Creating folder: Instances. Parent ref: group-v995303. {{(pid=65726) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1155.953413] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-1b4019cb-c8dc-4986-b211-c644196d543d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.966023] env[65726]: INFO nova.virt.vmwareapi.vm_util [None req-3a8ffaf6-e93e-4fc3-a889-1a35a6c01d6d tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] Created folder: Instances in parent group-v995303. [ 1155.966317] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-3a8ffaf6-e93e-4fc3-a889-1a35a6c01d6d tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1155.966523] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30] Creating VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1155.966748] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-db9f6504-3f2b-422c-9c87-76f6b12fe33b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.988964] env[65726]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1155.988964] env[65726]: value = "task-5116846" [ 1155.988964] env[65726]: _type = "Task" [ 1155.988964] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1155.997731] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116846, 'name': CreateVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1156.087771] env[65726]: DEBUG nova.compute.manager [None req-e8a62c22-e16d-473f-bc47-d51d9d4949ce tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 9c0599e3-383f-46ab-809b-944cc3a4d206] Start spawning the instance on the hypervisor. {{(pid=65726) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1156.120733] env[65726]: DEBUG oslo_vmware.api [None req-b778c6b3-2a71-4d65-8347-d07f00992900 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Task: {'id': task-5116843, 'name': PowerOnVM_Task, 'duration_secs': 0.483848} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1156.123575] env[65726]: DEBUG nova.virt.hardware [None req-e8a62c22-e16d-473f-bc47-d51d9d4949ce tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T19:28:34Z,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1539f06fa3534e90acbc3a60c4b8bd3f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T19:28:35Z,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1156.123818] env[65726]: DEBUG nova.virt.hardware [None req-e8a62c22-e16d-473f-bc47-d51d9d4949ce tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1156.123969] env[65726]: DEBUG nova.virt.hardware [None req-e8a62c22-e16d-473f-bc47-d51d9d4949ce tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1156.124171] env[65726]: DEBUG nova.virt.hardware [None req-e8a62c22-e16d-473f-bc47-d51d9d4949ce tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1156.124323] env[65726]: DEBUG nova.virt.hardware [None req-e8a62c22-e16d-473f-bc47-d51d9d4949ce tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1156.124508] env[65726]: DEBUG nova.virt.hardware [None req-e8a62c22-e16d-473f-bc47-d51d9d4949ce tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1156.124772] env[65726]: DEBUG nova.virt.hardware [None req-e8a62c22-e16d-473f-bc47-d51d9d4949ce tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1156.124877] env[65726]: DEBUG nova.virt.hardware [None req-e8a62c22-e16d-473f-bc47-d51d9d4949ce tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1156.125044] env[65726]: DEBUG nova.virt.hardware [None req-e8a62c22-e16d-473f-bc47-d51d9d4949ce tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1156.125204] env[65726]: DEBUG nova.virt.hardware [None req-e8a62c22-e16d-473f-bc47-d51d9d4949ce tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1156.125372] env[65726]: DEBUG nova.virt.hardware [None req-e8a62c22-e16d-473f-bc47-d51d9d4949ce tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1156.125676] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-b778c6b3-2a71-4d65-8347-d07f00992900 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] [instance: 5bfbaf25-5aed-46bc-97fc-f138d08ebbde] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1156.125880] env[65726]: INFO nova.compute.manager [None req-b778c6b3-2a71-4d65-8347-d07f00992900 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] [instance: 5bfbaf25-5aed-46bc-97fc-f138d08ebbde] Took 7.00 seconds to spawn the instance on the hypervisor. [ 1156.126071] env[65726]: DEBUG nova.compute.manager [None req-b778c6b3-2a71-4d65-8347-d07f00992900 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] [instance: 5bfbaf25-5aed-46bc-97fc-f138d08ebbde] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1156.127102] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b6e1b2e-9775-4c9e-91cc-d7007ea18a00 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.131220] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59f82cc2-8430-4b4e-a919-d7a18621b6a3 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.134915] env[65726]: DEBUG nova.compute.manager [req-4b225834-3772-4029-a8f6-4aec203a9f68 req-9f2eeae5-4c74-4443-8139-7a314cca881b service nova] [instance: 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30] Received event network-changed-e0f18f03-6a1b-4d5e-9b8e-68051f61ad8b {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1156.135105] env[65726]: DEBUG nova.compute.manager [req-4b225834-3772-4029-a8f6-4aec203a9f68 req-9f2eeae5-4c74-4443-8139-7a314cca881b service nova] [instance: 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30] Refreshing instance network info cache due to event network-changed-e0f18f03-6a1b-4d5e-9b8e-68051f61ad8b. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 1156.135325] env[65726]: DEBUG oslo_concurrency.lockutils [req-4b225834-3772-4029-a8f6-4aec203a9f68 req-9f2eeae5-4c74-4443-8139-7a314cca881b service nova] Acquiring lock "refresh_cache-47e52bc1-c18e-4aa2-82a2-e4cb030a7a30" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1156.135913] env[65726]: DEBUG oslo_concurrency.lockutils [req-4b225834-3772-4029-a8f6-4aec203a9f68 req-9f2eeae5-4c74-4443-8139-7a314cca881b service nova] Acquired lock "refresh_cache-47e52bc1-c18e-4aa2-82a2-e4cb030a7a30" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1156.135913] env[65726]: DEBUG nova.network.neutron [req-4b225834-3772-4029-a8f6-4aec203a9f68 req-9f2eeae5-4c74-4443-8139-7a314cca881b service nova] [instance: 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30] Refreshing network info cache for port e0f18f03-6a1b-4d5e-9b8e-68051f61ad8b {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 1156.148938] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4892f06e-b0a5-43fb-988a-65e65d4b39fb {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.346081] env[65726]: INFO nova.compute.claims [None req-168947ab-de46-4966-89ee-e38c38aa73bc tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 7c1188ce-8718-4719-8631-e59e7915b7aa] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1156.370806] env[65726]: DEBUG oslo_concurrency.lockutils [None req-8ea0406c-db1c-4929-acf9-e0ba5e267d6a tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Lock "0d42fa75-20e9-4646-9b08-17015b7f068c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 10.301s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1156.424868] env[65726]: DEBUG nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] [instance: 6fd70b4a-63e4-4258-9cff-f3b582500b9f] Instance has had 0 of 5 cleanup attempts {{(pid=65726) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 1156.477963] env[65726]: DEBUG nova.network.neutron [None req-e8a62c22-e16d-473f-bc47-d51d9d4949ce tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 9c0599e3-383f-46ab-809b-944cc3a4d206] Successfully updated port: c5730368-8477-43fb-ae16-4a4e35a403ab {{(pid=65726) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 1156.501988] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116846, 'name': CreateVM_Task} progress is 99%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1156.641148] env[65726]: WARNING openstack [req-4b225834-3772-4029-a8f6-4aec203a9f68 req-9f2eeae5-4c74-4443-8139-7a314cca881b service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1156.641577] env[65726]: WARNING openstack [req-4b225834-3772-4029-a8f6-4aec203a9f68 req-9f2eeae5-4c74-4443-8139-7a314cca881b service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1156.654843] env[65726]: INFO nova.compute.manager [None req-b778c6b3-2a71-4d65-8347-d07f00992900 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] [instance: 5bfbaf25-5aed-46bc-97fc-f138d08ebbde] Took 16.10 seconds to build instance. [ 1156.752336] env[65726]: WARNING openstack [req-4b225834-3772-4029-a8f6-4aec203a9f68 req-9f2eeae5-4c74-4443-8139-7a314cca881b service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1156.752718] env[65726]: WARNING openstack [req-4b225834-3772-4029-a8f6-4aec203a9f68 req-9f2eeae5-4c74-4443-8139-7a314cca881b service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1156.823716] env[65726]: WARNING openstack [req-4b225834-3772-4029-a8f6-4aec203a9f68 req-9f2eeae5-4c74-4443-8139-7a314cca881b service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1156.824101] env[65726]: WARNING openstack [req-4b225834-3772-4029-a8f6-4aec203a9f68 req-9f2eeae5-4c74-4443-8139-7a314cca881b service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1156.852238] env[65726]: INFO nova.compute.resource_tracker [None req-168947ab-de46-4966-89ee-e38c38aa73bc tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 7c1188ce-8718-4719-8631-e59e7915b7aa] Updating resource usage from migration 66575ff4-cc55-4594-bdc2-e957ee22e601 [ 1156.909081] env[65726]: DEBUG nova.network.neutron [req-4b225834-3772-4029-a8f6-4aec203a9f68 req-9f2eeae5-4c74-4443-8139-7a314cca881b service nova] [instance: 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30] Updated VIF entry in instance network info cache for port e0f18f03-6a1b-4d5e-9b8e-68051f61ad8b. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 1156.909491] env[65726]: DEBUG nova.network.neutron [req-4b225834-3772-4029-a8f6-4aec203a9f68 req-9f2eeae5-4c74-4443-8139-7a314cca881b service nova] [instance: 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30] Updating instance_info_cache with network_info: [{"id": "e0f18f03-6a1b-4d5e-9b8e-68051f61ad8b", "address": "fa:16:3e:39:d3:8c", "network": {"id": "ba97c3c2-c46b-4ce9-8f4d-b20857d6f8ec", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-992165785-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1e3203c117de4d4d9c4c90436801ff3f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f856fca-9fb5-41ea-a057-ac4193bd323d", "external-id": "nsx-vlan-transportzone-148", "segmentation_id": 148, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape0f18f03-6a", "ovs_interfaceid": "e0f18f03-6a1b-4d5e-9b8e-68051f61ad8b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1156.928704] env[65726]: DEBUG nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] [instance: df46141c-b545-4e03-b3a3-fd9f5feda0d2] Instance has had 0 of 5 cleanup attempts {{(pid=65726) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 1156.965545] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a25d259a-c453-4078-8dde-1bd0d195d5e7 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Acquiring lock "51f5337d-2e23-4b7c-b590-76be6bd82411" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1156.965918] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a25d259a-c453-4078-8dde-1bd0d195d5e7 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Lock "51f5337d-2e23-4b7c-b590-76be6bd82411" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1156.966189] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a25d259a-c453-4078-8dde-1bd0d195d5e7 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Acquiring lock "51f5337d-2e23-4b7c-b590-76be6bd82411-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1156.966411] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a25d259a-c453-4078-8dde-1bd0d195d5e7 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Lock "51f5337d-2e23-4b7c-b590-76be6bd82411-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1156.966609] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a25d259a-c453-4078-8dde-1bd0d195d5e7 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Lock "51f5337d-2e23-4b7c-b590-76be6bd82411-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1156.971270] env[65726]: INFO nova.compute.manager [None req-a25d259a-c453-4078-8dde-1bd0d195d5e7 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: 51f5337d-2e23-4b7c-b590-76be6bd82411] Terminating instance [ 1156.980601] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e8a62c22-e16d-473f-bc47-d51d9d4949ce tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Acquiring lock "refresh_cache-9c0599e3-383f-46ab-809b-944cc3a4d206" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1156.980750] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e8a62c22-e16d-473f-bc47-d51d9d4949ce tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Acquired lock "refresh_cache-9c0599e3-383f-46ab-809b-944cc3a4d206" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1156.980885] env[65726]: DEBUG nova.network.neutron [None req-e8a62c22-e16d-473f-bc47-d51d9d4949ce tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 9c0599e3-383f-46ab-809b-944cc3a4d206] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1156.999593] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78c43ad0-4944-442a-a98a-f661e5694f56 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.005970] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116846, 'name': CreateVM_Task} progress is 99%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1157.011430] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3314d455-b7ee-42d2-89c5-8ea5fd249853 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.043006] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8582c546-1872-4ded-a97d-8d02035ed3f2 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.052462] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55e7ea53-3634-476c-9eb9-208e3ab6fbe4 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.067359] env[65726]: DEBUG nova.compute.provider_tree [None req-168947ab-de46-4966-89ee-e38c38aa73bc tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1157.157784] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b778c6b3-2a71-4d65-8347-d07f00992900 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Lock "5bfbaf25-5aed-46bc-97fc-f138d08ebbde" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.615s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1157.414024] env[65726]: DEBUG oslo_concurrency.lockutils [req-4b225834-3772-4029-a8f6-4aec203a9f68 req-9f2eeae5-4c74-4443-8139-7a314cca881b service nova] Releasing lock "refresh_cache-47e52bc1-c18e-4aa2-82a2-e4cb030a7a30" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1157.432031] env[65726]: DEBUG nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] [instance: 7aa4c0d9-f7ca-42ba-9fd0-4075cd76752b] Instance has had 0 of 5 cleanup attempts {{(pid=65726) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 1157.475294] env[65726]: DEBUG nova.compute.manager [None req-a25d259a-c453-4078-8dde-1bd0d195d5e7 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: 51f5337d-2e23-4b7c-b590-76be6bd82411] Start destroying the instance on the hypervisor. {{(pid=65726) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1157.475670] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-a25d259a-c453-4078-8dde-1bd0d195d5e7 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: 51f5337d-2e23-4b7c-b590-76be6bd82411] Destroying instance {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1157.477032] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28893aab-e89f-4249-a9dd-a26a00b515c8 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.484960] env[65726]: WARNING openstack [None req-e8a62c22-e16d-473f-bc47-d51d9d4949ce tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1157.485614] env[65726]: WARNING openstack [None req-e8a62c22-e16d-473f-bc47-d51d9d4949ce tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1157.494199] env[65726]: DEBUG nova.network.neutron [None req-e8a62c22-e16d-473f-bc47-d51d9d4949ce tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 9c0599e3-383f-46ab-809b-944cc3a4d206] Instance cache missing network info. {{(pid=65726) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 1157.504866] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-a25d259a-c453-4078-8dde-1bd0d195d5e7 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: 51f5337d-2e23-4b7c-b590-76be6bd82411] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1157.505547] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d8a07893-bc40-439e-86c6-5772d8ef142d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.510596] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116846, 'name': CreateVM_Task, 'duration_secs': 1.42311} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1157.511814] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30] Created VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1157.512198] env[65726]: DEBUG oslo_vmware.api [None req-a25d259a-c453-4078-8dde-1bd0d195d5e7 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Waiting for the task: (returnval){ [ 1157.512198] env[65726]: value = "task-5116847" [ 1157.512198] env[65726]: _type = "Task" [ 1157.512198] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1157.512909] env[65726]: WARNING openstack [None req-3a8ffaf6-e93e-4fc3-a889-1a35a6c01d6d tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1157.513320] env[65726]: WARNING openstack [None req-3a8ffaf6-e93e-4fc3-a889-1a35a6c01d6d tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1157.518578] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-3a8ffaf6-e93e-4fc3-a889-1a35a6c01d6d tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] [instance: 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30] Block device information present: {'root_device_name': '/dev/sda', 'image': [], 'ephemerals': [], 'block_device_mapping': [{'disk_bus': None, 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-995306', 'volume_id': '39918e7e-1d4a-45df-911f-46ed9091f96b', 'name': 'volume-39918e7e-1d4a-45df-911f-46ed9091f96b', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '47e52bc1-c18e-4aa2-82a2-e4cb030a7a30', 'attached_at': '', 'detached_at': '', 'volume_id': '39918e7e-1d4a-45df-911f-46ed9091f96b', 'serial': '39918e7e-1d4a-45df-911f-46ed9091f96b'}, 'delete_on_termination': True, 'device_type': None, 'boot_index': 0, 'mount_device': '/dev/sda', 'attachment_id': '460c5ba7-8f25-40cb-a2a3-ca1443ab65dc', 'guest_format': None, 'volume_type': None}], 'swap': None} {{(pid=65726) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 1157.518817] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-3a8ffaf6-e93e-4fc3-a889-1a35a6c01d6d tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] [instance: 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30] Root volume attach. Driver type: vmdk {{(pid=65726) attach_root_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:661}} [ 1157.519841] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-597d72fd-4387-4391-894d-869e2a32c489 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.540509] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7735a25-7a8b-4609-b3b3-042dff698365 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.543473] env[65726]: DEBUG oslo_vmware.api [None req-a25d259a-c453-4078-8dde-1bd0d195d5e7 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': task-5116847, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1157.548556] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b332a1f2-faf6-4d96-ab77-09bb4c747907 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.555799] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.RelocateVM_Task with opID=oslo.vmware-e987ef9e-6c83-4af3-ac33-0d9bd56e30d8 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.563865] env[65726]: DEBUG oslo_vmware.api [None req-3a8ffaf6-e93e-4fc3-a889-1a35a6c01d6d tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] Waiting for the task: (returnval){ [ 1157.563865] env[65726]: value = "task-5116848" [ 1157.563865] env[65726]: _type = "Task" [ 1157.563865] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1157.570528] env[65726]: DEBUG nova.scheduler.client.report [None req-168947ab-de46-4966-89ee-e38c38aa73bc tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1157.577946] env[65726]: DEBUG oslo_vmware.api [None req-3a8ffaf6-e93e-4fc3-a889-1a35a6c01d6d tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] Task: {'id': task-5116848, 'name': RelocateVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1157.583777] env[65726]: WARNING openstack [None req-e8a62c22-e16d-473f-bc47-d51d9d4949ce tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1157.584268] env[65726]: WARNING openstack [None req-e8a62c22-e16d-473f-bc47-d51d9d4949ce tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1157.673150] env[65726]: WARNING openstack [None req-e8a62c22-e16d-473f-bc47-d51d9d4949ce tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1157.673693] env[65726]: WARNING openstack [None req-e8a62c22-e16d-473f-bc47-d51d9d4949ce tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1157.751134] env[65726]: DEBUG nova.compute.manager [req-f0e6cdf4-be1e-4975-accb-e751173e12af req-5dccef3b-6d0b-4af4-ac28-ec780638f4b2 service nova] [instance: 5bfbaf25-5aed-46bc-97fc-f138d08ebbde] Received event network-changed-c41be8b4-e252-47c3-8529-cd6a7ce6a8d3 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1157.751385] env[65726]: DEBUG nova.compute.manager [req-f0e6cdf4-be1e-4975-accb-e751173e12af req-5dccef3b-6d0b-4af4-ac28-ec780638f4b2 service nova] [instance: 5bfbaf25-5aed-46bc-97fc-f138d08ebbde] Refreshing instance network info cache due to event network-changed-c41be8b4-e252-47c3-8529-cd6a7ce6a8d3. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 1157.751611] env[65726]: DEBUG oslo_concurrency.lockutils [req-f0e6cdf4-be1e-4975-accb-e751173e12af req-5dccef3b-6d0b-4af4-ac28-ec780638f4b2 service nova] Acquiring lock "refresh_cache-5bfbaf25-5aed-46bc-97fc-f138d08ebbde" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1157.751755] env[65726]: DEBUG oslo_concurrency.lockutils [req-f0e6cdf4-be1e-4975-accb-e751173e12af req-5dccef3b-6d0b-4af4-ac28-ec780638f4b2 service nova] Acquired lock "refresh_cache-5bfbaf25-5aed-46bc-97fc-f138d08ebbde" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1157.751924] env[65726]: DEBUG nova.network.neutron [req-f0e6cdf4-be1e-4975-accb-e751173e12af req-5dccef3b-6d0b-4af4-ac28-ec780638f4b2 service nova] [instance: 5bfbaf25-5aed-46bc-97fc-f138d08ebbde] Refreshing network info cache for port c41be8b4-e252-47c3-8529-cd6a7ce6a8d3 {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 1157.787620] env[65726]: DEBUG nova.network.neutron [None req-e8a62c22-e16d-473f-bc47-d51d9d4949ce tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 9c0599e3-383f-46ab-809b-944cc3a4d206] Updating instance_info_cache with network_info: [{"id": "c5730368-8477-43fb-ae16-4a4e35a403ab", "address": "fa:16:3e:13:27:ee", "network": {"id": "69642329-e986-48fa-accf-1f296c08116a", "bridge": "br-int", "label": "tempest-ServersTestJSON-1786538788-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f3f799aa94f64e8cb26d93f6124efd81", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "816c6e38-e200-4544-8c5b-9fc3e16c5761", "external-id": "nsx-vlan-transportzone-195", "segmentation_id": 195, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc5730368-84", "ovs_interfaceid": "c5730368-8477-43fb-ae16-4a4e35a403ab", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1157.935505] env[65726]: DEBUG nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] [instance: 060646cc-f0fb-4e6c-9f67-7b2daa6b6fd6] Instance has had 0 of 5 cleanup attempts {{(pid=65726) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 1158.031091] env[65726]: DEBUG oslo_vmware.api [None req-a25d259a-c453-4078-8dde-1bd0d195d5e7 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': task-5116847, 'name': PowerOffVM_Task, 'duration_secs': 0.251226} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1158.031523] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-a25d259a-c453-4078-8dde-1bd0d195d5e7 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: 51f5337d-2e23-4b7c-b590-76be6bd82411] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1158.031735] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-a25d259a-c453-4078-8dde-1bd0d195d5e7 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: 51f5337d-2e23-4b7c-b590-76be6bd82411] Unregistering the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1158.032094] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-dd438c9c-58fc-4359-8439-159917636c89 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.077881] env[65726]: DEBUG oslo_vmware.api [None req-3a8ffaf6-e93e-4fc3-a889-1a35a6c01d6d tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] Task: {'id': task-5116848, 'name': RelocateVM_Task} progress is 40%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1158.079861] env[65726]: DEBUG oslo_concurrency.lockutils [None req-168947ab-de46-4966-89ee-e38c38aa73bc tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.240s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1158.080082] env[65726]: INFO nova.compute.manager [None req-168947ab-de46-4966-89ee-e38c38aa73bc tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 7c1188ce-8718-4719-8631-e59e7915b7aa] Migrating [ 1158.088108] env[65726]: DEBUG oslo_concurrency.lockutils [None req-aad3fddf-3318-4cfd-8f31-0db9fa8fe254 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.671s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1158.088323] env[65726]: DEBUG oslo_concurrency.lockutils [None req-aad3fddf-3318-4cfd-8f31-0db9fa8fe254 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1158.106421] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-a25d259a-c453-4078-8dde-1bd0d195d5e7 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: 51f5337d-2e23-4b7c-b590-76be6bd82411] Unregistered the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1158.107205] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-a25d259a-c453-4078-8dde-1bd0d195d5e7 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: 51f5337d-2e23-4b7c-b590-76be6bd82411] Deleting contents of the VM from datastore datastore2 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1158.107205] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-a25d259a-c453-4078-8dde-1bd0d195d5e7 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Deleting the datastore file [datastore2] 51f5337d-2e23-4b7c-b590-76be6bd82411 {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1158.107205] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-28485737-0dfb-4886-b5bc-ebf378a0091e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.116695] env[65726]: DEBUG oslo_vmware.api [None req-a25d259a-c453-4078-8dde-1bd0d195d5e7 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Waiting for the task: (returnval){ [ 1158.116695] env[65726]: value = "task-5116850" [ 1158.116695] env[65726]: _type = "Task" [ 1158.116695] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1158.117944] env[65726]: INFO nova.scheduler.client.report [None req-aad3fddf-3318-4cfd-8f31-0db9fa8fe254 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Deleted allocations for instance e7156a10-631a-4cc5-a544-88ce37763d02 [ 1158.134659] env[65726]: DEBUG oslo_vmware.api [None req-a25d259a-c453-4078-8dde-1bd0d195d5e7 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': task-5116850, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1158.163204] env[65726]: DEBUG nova.compute.manager [req-a9cf0701-f551-45e0-921a-2a8b7afde5c9 req-d46d1cc2-87a9-476c-a139-c0207d663de5 service nova] [instance: 9c0599e3-383f-46ab-809b-944cc3a4d206] Received event network-vif-plugged-c5730368-8477-43fb-ae16-4a4e35a403ab {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1158.163472] env[65726]: DEBUG oslo_concurrency.lockutils [req-a9cf0701-f551-45e0-921a-2a8b7afde5c9 req-d46d1cc2-87a9-476c-a139-c0207d663de5 service nova] Acquiring lock "9c0599e3-383f-46ab-809b-944cc3a4d206-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1158.163779] env[65726]: DEBUG oslo_concurrency.lockutils [req-a9cf0701-f551-45e0-921a-2a8b7afde5c9 req-d46d1cc2-87a9-476c-a139-c0207d663de5 service nova] Lock "9c0599e3-383f-46ab-809b-944cc3a4d206-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1158.163982] env[65726]: DEBUG oslo_concurrency.lockutils [req-a9cf0701-f551-45e0-921a-2a8b7afde5c9 req-d46d1cc2-87a9-476c-a139-c0207d663de5 service nova] Lock "9c0599e3-383f-46ab-809b-944cc3a4d206-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1158.164229] env[65726]: DEBUG nova.compute.manager [req-a9cf0701-f551-45e0-921a-2a8b7afde5c9 req-d46d1cc2-87a9-476c-a139-c0207d663de5 service nova] [instance: 9c0599e3-383f-46ab-809b-944cc3a4d206] No waiting events found dispatching network-vif-plugged-c5730368-8477-43fb-ae16-4a4e35a403ab {{(pid=65726) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1158.164450] env[65726]: WARNING nova.compute.manager [req-a9cf0701-f551-45e0-921a-2a8b7afde5c9 req-d46d1cc2-87a9-476c-a139-c0207d663de5 service nova] [instance: 9c0599e3-383f-46ab-809b-944cc3a4d206] Received unexpected event network-vif-plugged-c5730368-8477-43fb-ae16-4a4e35a403ab for instance with vm_state building and task_state spawning. [ 1158.164643] env[65726]: DEBUG nova.compute.manager [req-a9cf0701-f551-45e0-921a-2a8b7afde5c9 req-d46d1cc2-87a9-476c-a139-c0207d663de5 service nova] [instance: 9c0599e3-383f-46ab-809b-944cc3a4d206] Received event network-changed-c5730368-8477-43fb-ae16-4a4e35a403ab {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1158.164857] env[65726]: DEBUG nova.compute.manager [req-a9cf0701-f551-45e0-921a-2a8b7afde5c9 req-d46d1cc2-87a9-476c-a139-c0207d663de5 service nova] [instance: 9c0599e3-383f-46ab-809b-944cc3a4d206] Refreshing instance network info cache due to event network-changed-c5730368-8477-43fb-ae16-4a4e35a403ab. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 1158.165042] env[65726]: DEBUG oslo_concurrency.lockutils [req-a9cf0701-f551-45e0-921a-2a8b7afde5c9 req-d46d1cc2-87a9-476c-a139-c0207d663de5 service nova] Acquiring lock "refresh_cache-9c0599e3-383f-46ab-809b-944cc3a4d206" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1158.254645] env[65726]: WARNING openstack [req-f0e6cdf4-be1e-4975-accb-e751173e12af req-5dccef3b-6d0b-4af4-ac28-ec780638f4b2 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1158.254897] env[65726]: WARNING openstack [req-f0e6cdf4-be1e-4975-accb-e751173e12af req-5dccef3b-6d0b-4af4-ac28-ec780638f4b2 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1158.290303] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e8a62c22-e16d-473f-bc47-d51d9d4949ce tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Releasing lock "refresh_cache-9c0599e3-383f-46ab-809b-944cc3a4d206" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1158.290847] env[65726]: DEBUG nova.compute.manager [None req-e8a62c22-e16d-473f-bc47-d51d9d4949ce tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 9c0599e3-383f-46ab-809b-944cc3a4d206] Instance network_info: |[{"id": "c5730368-8477-43fb-ae16-4a4e35a403ab", "address": "fa:16:3e:13:27:ee", "network": {"id": "69642329-e986-48fa-accf-1f296c08116a", "bridge": "br-int", "label": "tempest-ServersTestJSON-1786538788-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f3f799aa94f64e8cb26d93f6124efd81", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "816c6e38-e200-4544-8c5b-9fc3e16c5761", "external-id": "nsx-vlan-transportzone-195", "segmentation_id": 195, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc5730368-84", "ovs_interfaceid": "c5730368-8477-43fb-ae16-4a4e35a403ab", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1158.291294] env[65726]: DEBUG oslo_concurrency.lockutils [req-a9cf0701-f551-45e0-921a-2a8b7afde5c9 req-d46d1cc2-87a9-476c-a139-c0207d663de5 service nova] Acquired lock "refresh_cache-9c0599e3-383f-46ab-809b-944cc3a4d206" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1158.291544] env[65726]: DEBUG nova.network.neutron [req-a9cf0701-f551-45e0-921a-2a8b7afde5c9 req-d46d1cc2-87a9-476c-a139-c0207d663de5 service nova] [instance: 9c0599e3-383f-46ab-809b-944cc3a4d206] Refreshing network info cache for port c5730368-8477-43fb-ae16-4a4e35a403ab {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 1158.293184] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-e8a62c22-e16d-473f-bc47-d51d9d4949ce tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 9c0599e3-383f-46ab-809b-944cc3a4d206] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:13:27:ee', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '816c6e38-e200-4544-8c5b-9fc3e16c5761', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c5730368-8477-43fb-ae16-4a4e35a403ab', 'vif_model': 'vmxnet3'}] {{(pid=65726) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1158.302308] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-e8a62c22-e16d-473f-bc47-d51d9d4949ce tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Creating folder: Project (f3f799aa94f64e8cb26d93f6124efd81). Parent ref: group-v995008. {{(pid=65726) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1158.308588] env[65726]: WARNING openstack [req-a9cf0701-f551-45e0-921a-2a8b7afde5c9 req-d46d1cc2-87a9-476c-a139-c0207d663de5 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1158.308983] env[65726]: WARNING openstack [req-a9cf0701-f551-45e0-921a-2a8b7afde5c9 req-d46d1cc2-87a9-476c-a139-c0207d663de5 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1158.316541] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f3c3a9f2-465c-4984-b137-d3f4d3acec4e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.331639] env[65726]: INFO nova.virt.vmwareapi.vm_util [None req-e8a62c22-e16d-473f-bc47-d51d9d4949ce tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Created folder: Project (f3f799aa94f64e8cb26d93f6124efd81) in parent group-v995008. [ 1158.331940] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-e8a62c22-e16d-473f-bc47-d51d9d4949ce tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Creating folder: Instances. Parent ref: group-v995312. {{(pid=65726) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1158.332608] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f68050e3-4aa2-47c7-bd5a-d0213528c4ba {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.345625] env[65726]: INFO nova.virt.vmwareapi.vm_util [None req-e8a62c22-e16d-473f-bc47-d51d9d4949ce tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Created folder: Instances in parent group-v995312. [ 1158.346936] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-e8a62c22-e16d-473f-bc47-d51d9d4949ce tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1158.346936] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9c0599e3-383f-46ab-809b-944cc3a4d206] Creating VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1158.346936] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-dabf563d-b65b-4388-902a-be43048f5c63 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.377470] env[65726]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1158.377470] env[65726]: value = "task-5116853" [ 1158.377470] env[65726]: _type = "Task" [ 1158.377470] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1158.387243] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116853, 'name': CreateVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1158.410634] env[65726]: WARNING openstack [req-f0e6cdf4-be1e-4975-accb-e751173e12af req-5dccef3b-6d0b-4af4-ac28-ec780638f4b2 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1158.411131] env[65726]: WARNING openstack [req-f0e6cdf4-be1e-4975-accb-e751173e12af req-5dccef3b-6d0b-4af4-ac28-ec780638f4b2 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1158.439788] env[65726]: DEBUG nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] [instance: 9029549c-1914-4bae-91e2-8812b79051ec] Instance has had 0 of 5 cleanup attempts {{(pid=65726) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 1158.472556] env[65726]: WARNING openstack [req-a9cf0701-f551-45e0-921a-2a8b7afde5c9 req-d46d1cc2-87a9-476c-a139-c0207d663de5 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1158.473128] env[65726]: WARNING openstack [req-a9cf0701-f551-45e0-921a-2a8b7afde5c9 req-d46d1cc2-87a9-476c-a139-c0207d663de5 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1158.494767] env[65726]: WARNING openstack [req-f0e6cdf4-be1e-4975-accb-e751173e12af req-5dccef3b-6d0b-4af4-ac28-ec780638f4b2 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1158.495447] env[65726]: WARNING openstack [req-f0e6cdf4-be1e-4975-accb-e751173e12af req-5dccef3b-6d0b-4af4-ac28-ec780638f4b2 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1158.568942] env[65726]: WARNING openstack [req-a9cf0701-f551-45e0-921a-2a8b7afde5c9 req-d46d1cc2-87a9-476c-a139-c0207d663de5 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1158.569606] env[65726]: WARNING openstack [req-a9cf0701-f551-45e0-921a-2a8b7afde5c9 req-d46d1cc2-87a9-476c-a139-c0207d663de5 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1158.590717] env[65726]: DEBUG oslo_vmware.api [None req-3a8ffaf6-e93e-4fc3-a889-1a35a6c01d6d tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] Task: {'id': task-5116848, 'name': RelocateVM_Task} progress is 53%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1158.599880] env[65726]: DEBUG oslo_concurrency.lockutils [None req-168947ab-de46-4966-89ee-e38c38aa73bc tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Acquiring lock "refresh_cache-7c1188ce-8718-4719-8631-e59e7915b7aa" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1158.599880] env[65726]: DEBUG oslo_concurrency.lockutils [None req-168947ab-de46-4966-89ee-e38c38aa73bc tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Acquired lock "refresh_cache-7c1188ce-8718-4719-8631-e59e7915b7aa" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1158.600299] env[65726]: DEBUG nova.network.neutron [None req-168947ab-de46-4966-89ee-e38c38aa73bc tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 7c1188ce-8718-4719-8631-e59e7915b7aa] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1158.629191] env[65726]: DEBUG nova.network.neutron [req-f0e6cdf4-be1e-4975-accb-e751173e12af req-5dccef3b-6d0b-4af4-ac28-ec780638f4b2 service nova] [instance: 5bfbaf25-5aed-46bc-97fc-f138d08ebbde] Updated VIF entry in instance network info cache for port c41be8b4-e252-47c3-8529-cd6a7ce6a8d3. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 1158.629822] env[65726]: DEBUG nova.network.neutron [req-f0e6cdf4-be1e-4975-accb-e751173e12af req-5dccef3b-6d0b-4af4-ac28-ec780638f4b2 service nova] [instance: 5bfbaf25-5aed-46bc-97fc-f138d08ebbde] Updating instance_info_cache with network_info: [{"id": "c41be8b4-e252-47c3-8529-cd6a7ce6a8d3", "address": "fa:16:3e:dd:aa:e3", "network": {"id": "547d7b16-4591-4d80-9e9b-c8a9a3a78dab", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-2077874536-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.181", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4ffd45f4a7a041199a4fc7f69f5e7e9b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2ed91b7b-b4ec-486d-ab34-af0afb7ec691", "external-id": "nsx-vlan-transportzone-75", "segmentation_id": 75, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc41be8b4-e2", "ovs_interfaceid": "c41be8b4-e252-47c3-8529-cd6a7ce6a8d3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1158.635311] env[65726]: DEBUG oslo_concurrency.lockutils [None req-aad3fddf-3318-4cfd-8f31-0db9fa8fe254 tempest-DeleteServersTestJSON-479881887 tempest-DeleteServersTestJSON-479881887-project-member] Lock "e7156a10-631a-4cc5-a544-88ce37763d02" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.772s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1158.653445] env[65726]: DEBUG oslo_vmware.api [None req-a25d259a-c453-4078-8dde-1bd0d195d5e7 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': task-5116850, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1158.699223] env[65726]: DEBUG nova.network.neutron [req-a9cf0701-f551-45e0-921a-2a8b7afde5c9 req-d46d1cc2-87a9-476c-a139-c0207d663de5 service nova] [instance: 9c0599e3-383f-46ab-809b-944cc3a4d206] Updated VIF entry in instance network info cache for port c5730368-8477-43fb-ae16-4a4e35a403ab. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 1158.699694] env[65726]: DEBUG nova.network.neutron [req-a9cf0701-f551-45e0-921a-2a8b7afde5c9 req-d46d1cc2-87a9-476c-a139-c0207d663de5 service nova] [instance: 9c0599e3-383f-46ab-809b-944cc3a4d206] Updating instance_info_cache with network_info: [{"id": "c5730368-8477-43fb-ae16-4a4e35a403ab", "address": "fa:16:3e:13:27:ee", "network": {"id": "69642329-e986-48fa-accf-1f296c08116a", "bridge": "br-int", "label": "tempest-ServersTestJSON-1786538788-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f3f799aa94f64e8cb26d93f6124efd81", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "816c6e38-e200-4544-8c5b-9fc3e16c5761", "external-id": "nsx-vlan-transportzone-195", "segmentation_id": 195, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc5730368-84", "ovs_interfaceid": "c5730368-8477-43fb-ae16-4a4e35a403ab", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1158.888665] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116853, 'name': CreateVM_Task} progress is 25%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1158.943677] env[65726]: DEBUG nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] [instance: dcde25df-87a5-47fb-94ce-334f68894e04] Instance has had 0 of 5 cleanup attempts {{(pid=65726) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 1159.090871] env[65726]: DEBUG oslo_vmware.api [None req-3a8ffaf6-e93e-4fc3-a889-1a35a6c01d6d tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] Task: {'id': task-5116848, 'name': RelocateVM_Task} progress is 65%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1159.104254] env[65726]: WARNING openstack [None req-168947ab-de46-4966-89ee-e38c38aa73bc tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1159.104814] env[65726]: WARNING openstack [None req-168947ab-de46-4966-89ee-e38c38aa73bc tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1159.133940] env[65726]: DEBUG oslo_vmware.api [None req-a25d259a-c453-4078-8dde-1bd0d195d5e7 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Task: {'id': task-5116850, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.955414} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1159.135104] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-a25d259a-c453-4078-8dde-1bd0d195d5e7 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Deleted the datastore file {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1159.135104] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-a25d259a-c453-4078-8dde-1bd0d195d5e7 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: 51f5337d-2e23-4b7c-b590-76be6bd82411] Deleted contents of the VM from datastore datastore2 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1159.135104] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-a25d259a-c453-4078-8dde-1bd0d195d5e7 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: 51f5337d-2e23-4b7c-b590-76be6bd82411] Instance destroyed {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1159.135104] env[65726]: INFO nova.compute.manager [None req-a25d259a-c453-4078-8dde-1bd0d195d5e7 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] [instance: 51f5337d-2e23-4b7c-b590-76be6bd82411] Took 1.66 seconds to destroy the instance on the hypervisor. [ 1159.135383] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-a25d259a-c453-4078-8dde-1bd0d195d5e7 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1159.135483] env[65726]: DEBUG nova.compute.manager [-] [instance: 51f5337d-2e23-4b7c-b590-76be6bd82411] Deallocating network for instance {{(pid=65726) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1159.135611] env[65726]: DEBUG nova.network.neutron [-] [instance: 51f5337d-2e23-4b7c-b590-76be6bd82411] deallocate_for_instance() {{(pid=65726) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 1159.136500] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1159.136936] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1159.150387] env[65726]: DEBUG oslo_concurrency.lockutils [req-f0e6cdf4-be1e-4975-accb-e751173e12af req-5dccef3b-6d0b-4af4-ac28-ec780638f4b2 service nova] Releasing lock "refresh_cache-5bfbaf25-5aed-46bc-97fc-f138d08ebbde" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1159.197593] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1159.197884] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1159.208087] env[65726]: DEBUG oslo_concurrency.lockutils [req-a9cf0701-f551-45e0-921a-2a8b7afde5c9 req-d46d1cc2-87a9-476c-a139-c0207d663de5 service nova] Releasing lock "refresh_cache-9c0599e3-383f-46ab-809b-944cc3a4d206" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1159.214246] env[65726]: WARNING openstack [None req-168947ab-de46-4966-89ee-e38c38aa73bc tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1159.215039] env[65726]: WARNING openstack [None req-168947ab-de46-4966-89ee-e38c38aa73bc tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1159.284906] env[65726]: WARNING openstack [None req-168947ab-de46-4966-89ee-e38c38aa73bc tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1159.285311] env[65726]: WARNING openstack [None req-168947ab-de46-4966-89ee-e38c38aa73bc tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1159.385762] env[65726]: DEBUG nova.network.neutron [None req-168947ab-de46-4966-89ee-e38c38aa73bc tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 7c1188ce-8718-4719-8631-e59e7915b7aa] Updating instance_info_cache with network_info: [{"id": "e8f51770-ca83-4aa3-87c1-ef47653d9bb2", "address": "fa:16:3e:8e:ce:ad", "network": {"id": "9dd55fb6-0dd2-484d-b94d-fbdccb958c79", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1005598846-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f93d852e2c904f42981cbfff882558e9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "74f30339-6421-4654-bddb-81d7f34db9d7", "external-id": "nsx-vlan-transportzone-899", "segmentation_id": 899, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape8f51770-ca", "ovs_interfaceid": "e8f51770-ca83-4aa3-87c1-ef47653d9bb2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1159.393027] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116853, 'name': CreateVM_Task} progress is 99%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1159.447974] env[65726]: DEBUG nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] [instance: 3088ee5e-0d27-4058-a94b-2e04c3b52add] Instance has had 0 of 5 cleanup attempts {{(pid=65726) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 1159.594522] env[65726]: DEBUG oslo_vmware.api [None req-3a8ffaf6-e93e-4fc3-a889-1a35a6c01d6d tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] Task: {'id': task-5116848, 'name': RelocateVM_Task} progress is 76%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1159.777569] env[65726]: DEBUG nova.compute.manager [req-6876df14-5295-4dd1-9bd8-4cad04718f84 req-41657097-da7a-4663-b43a-f19e89d51d71 service nova] [instance: 51f5337d-2e23-4b7c-b590-76be6bd82411] Received event network-vif-deleted-7407bcd9-3202-4798-9eaa-252fb3fdb21f {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1159.777569] env[65726]: INFO nova.compute.manager [req-6876df14-5295-4dd1-9bd8-4cad04718f84 req-41657097-da7a-4663-b43a-f19e89d51d71 service nova] [instance: 51f5337d-2e23-4b7c-b590-76be6bd82411] Neutron deleted interface 7407bcd9-3202-4798-9eaa-252fb3fdb21f; detaching it from the instance and deleting it from the info cache [ 1159.777819] env[65726]: DEBUG nova.network.neutron [req-6876df14-5295-4dd1-9bd8-4cad04718f84 req-41657097-da7a-4663-b43a-f19e89d51d71 service nova] [instance: 51f5337d-2e23-4b7c-b590-76be6bd82411] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1159.890505] env[65726]: DEBUG oslo_concurrency.lockutils [None req-168947ab-de46-4966-89ee-e38c38aa73bc tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Releasing lock "refresh_cache-7c1188ce-8718-4719-8631-e59e7915b7aa" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1159.891879] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116853, 'name': CreateVM_Task, 'duration_secs': 1.192388} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1159.892069] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9c0599e3-383f-46ab-809b-944cc3a4d206] Created VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1159.893159] env[65726]: WARNING openstack [None req-e8a62c22-e16d-473f-bc47-d51d9d4949ce tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1159.893418] env[65726]: WARNING openstack [None req-e8a62c22-e16d-473f-bc47-d51d9d4949ce tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1159.899394] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e8a62c22-e16d-473f-bc47-d51d9d4949ce tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1159.899575] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e8a62c22-e16d-473f-bc47-d51d9d4949ce tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1159.899922] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e8a62c22-e16d-473f-bc47-d51d9d4949ce tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1159.900278] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-30c6b325-af85-46ab-9289-d66b95bcc468 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.907540] env[65726]: DEBUG oslo_vmware.api [None req-e8a62c22-e16d-473f-bc47-d51d9d4949ce tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Waiting for the task: (returnval){ [ 1159.907540] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]525b66b2-145d-9fd8-2ca1-f4663740c3ca" [ 1159.907540] env[65726]: _type = "Task" [ 1159.907540] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1159.918511] env[65726]: DEBUG oslo_vmware.api [None req-e8a62c22-e16d-473f-bc47-d51d9d4949ce tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]525b66b2-145d-9fd8-2ca1-f4663740c3ca, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1159.929854] env[65726]: DEBUG nova.network.neutron [-] [instance: 51f5337d-2e23-4b7c-b590-76be6bd82411] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1159.951811] env[65726]: DEBUG nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] [instance: 6f91b053-772a-4497-b29d-349b960c55eb] Instance has had 0 of 5 cleanup attempts {{(pid=65726) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 1160.091212] env[65726]: DEBUG oslo_vmware.api [None req-3a8ffaf6-e93e-4fc3-a889-1a35a6c01d6d tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] Task: {'id': task-5116848, 'name': RelocateVM_Task} progress is 89%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1160.280706] env[65726]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ca2f2455-13e8-4e9f-93a3-f28232cda98e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.292272] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72e24f91-e152-4205-b35d-c74a531b056b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.327991] env[65726]: DEBUG nova.compute.manager [req-6876df14-5295-4dd1-9bd8-4cad04718f84 req-41657097-da7a-4663-b43a-f19e89d51d71 service nova] [instance: 51f5337d-2e23-4b7c-b590-76be6bd82411] Detach interface failed, port_id=7407bcd9-3202-4798-9eaa-252fb3fdb21f, reason: Instance 51f5337d-2e23-4b7c-b590-76be6bd82411 could not be found. {{(pid=65726) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11669}} [ 1160.418091] env[65726]: DEBUG oslo_vmware.api [None req-e8a62c22-e16d-473f-bc47-d51d9d4949ce tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]525b66b2-145d-9fd8-2ca1-f4663740c3ca, 'name': SearchDatastore_Task, 'duration_secs': 0.041208} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1160.418396] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e8a62c22-e16d-473f-bc47-d51d9d4949ce tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1160.418624] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-e8a62c22-e16d-473f-bc47-d51d9d4949ce tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 9c0599e3-383f-46ab-809b-944cc3a4d206] Processing image b52362a3-ee8a-4cbf-b06f-513b0cc8f95c {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1160.418954] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e8a62c22-e16d-473f-bc47-d51d9d4949ce tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1160.419058] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e8a62c22-e16d-473f-bc47-d51d9d4949ce tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1160.419181] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-e8a62c22-e16d-473f-bc47-d51d9d4949ce tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1160.419450] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e31f5120-c2f0-4d07-ac54-e2d9af328fa8 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.430105] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-e8a62c22-e16d-473f-bc47-d51d9d4949ce tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1160.430300] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-e8a62c22-e16d-473f-bc47-d51d9d4949ce tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65726) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1160.431053] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-431aa7cb-8af1-4ead-a54a-52b7da78aa1a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.433772] env[65726]: INFO nova.compute.manager [-] [instance: 51f5337d-2e23-4b7c-b590-76be6bd82411] Took 1.30 seconds to deallocate network for instance. [ 1160.442213] env[65726]: DEBUG oslo_vmware.api [None req-e8a62c22-e16d-473f-bc47-d51d9d4949ce tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Waiting for the task: (returnval){ [ 1160.442213] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52c549c7-ebf6-6160-0910-b6eb935e2c11" [ 1160.442213] env[65726]: _type = "Task" [ 1160.442213] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1160.451303] env[65726]: DEBUG oslo_vmware.api [None req-e8a62c22-e16d-473f-bc47-d51d9d4949ce tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52c549c7-ebf6-6160-0910-b6eb935e2c11, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1160.455816] env[65726]: DEBUG nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] [instance: 5a252ef1-93c9-4bff-842b-b64df2bc5d75] Instance has had 0 of 5 cleanup attempts {{(pid=65726) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 1160.591019] env[65726]: DEBUG oslo_vmware.api [None req-3a8ffaf6-e93e-4fc3-a889-1a35a6c01d6d tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] Task: {'id': task-5116848, 'name': RelocateVM_Task} progress is 97%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1160.943302] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a25d259a-c453-4078-8dde-1bd0d195d5e7 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1160.943711] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a25d259a-c453-4078-8dde-1bd0d195d5e7 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1160.944100] env[65726]: DEBUG nova.objects.instance [None req-a25d259a-c453-4078-8dde-1bd0d195d5e7 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Lazy-loading 'resources' on Instance uuid 51f5337d-2e23-4b7c-b590-76be6bd82411 {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1160.955789] env[65726]: DEBUG oslo_vmware.api [None req-e8a62c22-e16d-473f-bc47-d51d9d4949ce tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52c549c7-ebf6-6160-0910-b6eb935e2c11, 'name': SearchDatastore_Task, 'duration_secs': 0.040613} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1160.957283] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-26d1284b-3ac2-4dd4-9376-5ebbff91a147 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.960185] env[65726]: DEBUG nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] [instance: 70375746-06dd-4710-9136-95e9e5759a0e] Instance has had 0 of 5 cleanup attempts {{(pid=65726) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 1160.966418] env[65726]: DEBUG oslo_vmware.api [None req-e8a62c22-e16d-473f-bc47-d51d9d4949ce tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Waiting for the task: (returnval){ [ 1160.966418] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52ea3791-2c0f-b89e-3dd1-702d32f732fd" [ 1160.966418] env[65726]: _type = "Task" [ 1160.966418] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1160.977395] env[65726]: DEBUG oslo_vmware.api [None req-e8a62c22-e16d-473f-bc47-d51d9d4949ce tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52ea3791-2c0f-b89e-3dd1-702d32f732fd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1161.092676] env[65726]: DEBUG oslo_vmware.api [None req-3a8ffaf6-e93e-4fc3-a889-1a35a6c01d6d tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] Task: {'id': task-5116848, 'name': RelocateVM_Task} progress is 98%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1161.407619] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1738a40-b4b7-4f75-bb8f-7057031383c3 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.427712] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-168947ab-de46-4966-89ee-e38c38aa73bc tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 7c1188ce-8718-4719-8631-e59e7915b7aa] Updating instance '7c1188ce-8718-4719-8631-e59e7915b7aa' progress to 0 {{(pid=65726) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1161.464760] env[65726]: DEBUG nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] [instance: 29823bc7-3909-4ab7-8119-91ee59b289c5] Instance has had 0 of 5 cleanup attempts {{(pid=65726) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 1161.479255] env[65726]: DEBUG oslo_vmware.api [None req-e8a62c22-e16d-473f-bc47-d51d9d4949ce tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52ea3791-2c0f-b89e-3dd1-702d32f732fd, 'name': SearchDatastore_Task, 'duration_secs': 0.01121} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1161.480273] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e8a62c22-e16d-473f-bc47-d51d9d4949ce tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1161.480526] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-e8a62c22-e16d-473f-bc47-d51d9d4949ce tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore2] 9c0599e3-383f-46ab-809b-944cc3a4d206/9c0599e3-383f-46ab-809b-944cc3a4d206.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1161.480797] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a4982251-72a9-401b-8fea-936cde215d68 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.490768] env[65726]: DEBUG oslo_vmware.api [None req-e8a62c22-e16d-473f-bc47-d51d9d4949ce tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Waiting for the task: (returnval){ [ 1161.490768] env[65726]: value = "task-5116855" [ 1161.490768] env[65726]: _type = "Task" [ 1161.490768] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1161.504881] env[65726]: DEBUG oslo_vmware.api [None req-e8a62c22-e16d-473f-bc47-d51d9d4949ce tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Task: {'id': task-5116855, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1161.580192] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0cc36d61-cf2b-488c-b0a2-f311530075ad {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.591985] env[65726]: DEBUG oslo_vmware.api [None req-3a8ffaf6-e93e-4fc3-a889-1a35a6c01d6d tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] Task: {'id': task-5116848, 'name': RelocateVM_Task, 'duration_secs': 3.655865} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1161.594094] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-3a8ffaf6-e93e-4fc3-a889-1a35a6c01d6d tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] [instance: 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30] Volume attach. Driver type: vmdk {{(pid=65726) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1161.594324] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-3a8ffaf6-e93e-4fc3-a889-1a35a6c01d6d tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] [instance: 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-995306', 'volume_id': '39918e7e-1d4a-45df-911f-46ed9091f96b', 'name': 'volume-39918e7e-1d4a-45df-911f-46ed9091f96b', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '47e52bc1-c18e-4aa2-82a2-e4cb030a7a30', 'attached_at': '', 'detached_at': '', 'volume_id': '39918e7e-1d4a-45df-911f-46ed9091f96b', 'serial': '39918e7e-1d4a-45df-911f-46ed9091f96b'} {{(pid=65726) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1161.595215] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2beb4d80-80c9-41e8-a68e-063db373a988 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.598381] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3ef9027-5854-4585-9657-631e1ea82b0e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.640386] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0157d2b-0e13-4af2-a37e-d9dd930da346 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.644339] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5568ff7a-ce3b-4606-9cc5-93ada6f9dbee {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.668335] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-3a8ffaf6-e93e-4fc3-a889-1a35a6c01d6d tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] [instance: 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30] Reconfiguring VM instance instance-0000006c to attach disk [datastore2] volume-39918e7e-1d4a-45df-911f-46ed9091f96b/volume-39918e7e-1d4a-45df-911f-46ed9091f96b.vmdk or device None with type thin {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1161.671567] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c8ed953a-3dc4-4845-bb51-742c72fc39b8 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.690593] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6410dc56-5cd5-40d1-b87c-91870c2368be {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.710876] env[65726]: DEBUG nova.compute.provider_tree [None req-a25d259a-c453-4078-8dde-1bd0d195d5e7 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1161.713445] env[65726]: DEBUG oslo_vmware.api [None req-3a8ffaf6-e93e-4fc3-a889-1a35a6c01d6d tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] Waiting for the task: (returnval){ [ 1161.713445] env[65726]: value = "task-5116856" [ 1161.713445] env[65726]: _type = "Task" [ 1161.713445] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1161.723534] env[65726]: DEBUG oslo_vmware.api [None req-3a8ffaf6-e93e-4fc3-a889-1a35a6c01d6d tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] Task: {'id': task-5116856, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1161.935067] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-168947ab-de46-4966-89ee-e38c38aa73bc tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 7c1188ce-8718-4719-8631-e59e7915b7aa] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1161.935625] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e075fdee-062e-457b-9e34-019cac3e57a9 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.944039] env[65726]: DEBUG oslo_vmware.api [None req-168947ab-de46-4966-89ee-e38c38aa73bc tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Waiting for the task: (returnval){ [ 1161.944039] env[65726]: value = "task-5116857" [ 1161.944039] env[65726]: _type = "Task" [ 1161.944039] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1161.955603] env[65726]: DEBUG oslo_vmware.api [None req-168947ab-de46-4966-89ee-e38c38aa73bc tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': task-5116857, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1161.968226] env[65726]: DEBUG nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] [instance: 815ee20b-cb07-4514-bd43-1b74972508a5] Instance has had 0 of 5 cleanup attempts {{(pid=65726) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 1162.002561] env[65726]: DEBUG oslo_vmware.api [None req-e8a62c22-e16d-473f-bc47-d51d9d4949ce tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Task: {'id': task-5116855, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.485973} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1162.002862] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-e8a62c22-e16d-473f-bc47-d51d9d4949ce tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore2] 9c0599e3-383f-46ab-809b-944cc3a4d206/9c0599e3-383f-46ab-809b-944cc3a4d206.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1162.003110] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-e8a62c22-e16d-473f-bc47-d51d9d4949ce tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 9c0599e3-383f-46ab-809b-944cc3a4d206] Extending root virtual disk to 1048576 {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1162.003390] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-8ef51ecf-26d5-4a27-b47a-cdc7c9278cb0 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.012116] env[65726]: DEBUG oslo_vmware.api [None req-e8a62c22-e16d-473f-bc47-d51d9d4949ce tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Waiting for the task: (returnval){ [ 1162.012116] env[65726]: value = "task-5116858" [ 1162.012116] env[65726]: _type = "Task" [ 1162.012116] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1162.025157] env[65726]: DEBUG oslo_vmware.api [None req-e8a62c22-e16d-473f-bc47-d51d9d4949ce tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Task: {'id': task-5116858, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1162.216515] env[65726]: DEBUG nova.scheduler.client.report [None req-a25d259a-c453-4078-8dde-1bd0d195d5e7 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1162.232674] env[65726]: DEBUG oslo_vmware.api [None req-3a8ffaf6-e93e-4fc3-a889-1a35a6c01d6d tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] Task: {'id': task-5116856, 'name': ReconfigVM_Task, 'duration_secs': 0.390627} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1162.233833] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-3a8ffaf6-e93e-4fc3-a889-1a35a6c01d6d tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] [instance: 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30] Reconfigured VM instance instance-0000006c to attach disk [datastore2] volume-39918e7e-1d4a-45df-911f-46ed9091f96b/volume-39918e7e-1d4a-45df-911f-46ed9091f96b.vmdk or device None with type thin {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1162.241503] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b6f971fc-443f-4790-97a9-c2ae3a797ed9 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.264615] env[65726]: DEBUG oslo_vmware.api [None req-3a8ffaf6-e93e-4fc3-a889-1a35a6c01d6d tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] Waiting for the task: (returnval){ [ 1162.264615] env[65726]: value = "task-5116859" [ 1162.264615] env[65726]: _type = "Task" [ 1162.264615] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1162.281217] env[65726]: DEBUG oslo_vmware.api [None req-3a8ffaf6-e93e-4fc3-a889-1a35a6c01d6d tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] Task: {'id': task-5116859, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1162.455054] env[65726]: DEBUG oslo_vmware.api [None req-168947ab-de46-4966-89ee-e38c38aa73bc tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': task-5116857, 'name': PowerOffVM_Task, 'duration_secs': 0.214354} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1162.455345] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-168947ab-de46-4966-89ee-e38c38aa73bc tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 7c1188ce-8718-4719-8631-e59e7915b7aa] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1162.455519] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-168947ab-de46-4966-89ee-e38c38aa73bc tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 7c1188ce-8718-4719-8631-e59e7915b7aa] Updating instance '7c1188ce-8718-4719-8631-e59e7915b7aa' progress to 17 {{(pid=65726) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1162.471509] env[65726]: DEBUG nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] [instance: 4ea2b70b-0cfe-4fbf-817a-baa028b8372a] Instance has had 0 of 5 cleanup attempts {{(pid=65726) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 1162.522587] env[65726]: DEBUG oslo_vmware.api [None req-e8a62c22-e16d-473f-bc47-d51d9d4949ce tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Task: {'id': task-5116858, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.077269} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1162.522975] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-e8a62c22-e16d-473f-bc47-d51d9d4949ce tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 9c0599e3-383f-46ab-809b-944cc3a4d206] Extended root virtual disk {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1162.523859] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1088e49-ba81-4182-a130-c4082818f2f3 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.546195] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-e8a62c22-e16d-473f-bc47-d51d9d4949ce tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 9c0599e3-383f-46ab-809b-944cc3a4d206] Reconfiguring VM instance instance-0000006d to attach disk [datastore2] 9c0599e3-383f-46ab-809b-944cc3a4d206/9c0599e3-383f-46ab-809b-944cc3a4d206.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1162.546805] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8b09fe3e-0ceb-43ca-b7f7-e97f48fdc084 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.566622] env[65726]: DEBUG oslo_vmware.api [None req-e8a62c22-e16d-473f-bc47-d51d9d4949ce tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Waiting for the task: (returnval){ [ 1162.566622] env[65726]: value = "task-5116860" [ 1162.566622] env[65726]: _type = "Task" [ 1162.566622] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1162.575519] env[65726]: DEBUG oslo_vmware.api [None req-e8a62c22-e16d-473f-bc47-d51d9d4949ce tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Task: {'id': task-5116860, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1162.722801] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a25d259a-c453-4078-8dde-1bd0d195d5e7 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.779s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1162.745014] env[65726]: INFO nova.scheduler.client.report [None req-a25d259a-c453-4078-8dde-1bd0d195d5e7 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Deleted allocations for instance 51f5337d-2e23-4b7c-b590-76be6bd82411 [ 1162.777489] env[65726]: DEBUG oslo_vmware.api [None req-3a8ffaf6-e93e-4fc3-a889-1a35a6c01d6d tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] Task: {'id': task-5116859, 'name': ReconfigVM_Task, 'duration_secs': 0.379315} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1162.777846] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-3a8ffaf6-e93e-4fc3-a889-1a35a6c01d6d tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] [instance: 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-995306', 'volume_id': '39918e7e-1d4a-45df-911f-46ed9091f96b', 'name': 'volume-39918e7e-1d4a-45df-911f-46ed9091f96b', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '47e52bc1-c18e-4aa2-82a2-e4cb030a7a30', 'attached_at': '', 'detached_at': '', 'volume_id': '39918e7e-1d4a-45df-911f-46ed9091f96b', 'serial': '39918e7e-1d4a-45df-911f-46ed9091f96b'} {{(pid=65726) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1162.778509] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-093ead2e-51e9-42d6-8d43-a3c4fca614b7 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.785975] env[65726]: DEBUG oslo_vmware.api [None req-3a8ffaf6-e93e-4fc3-a889-1a35a6c01d6d tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] Waiting for the task: (returnval){ [ 1162.785975] env[65726]: value = "task-5116861" [ 1162.785975] env[65726]: _type = "Task" [ 1162.785975] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1162.795265] env[65726]: DEBUG oslo_vmware.api [None req-3a8ffaf6-e93e-4fc3-a889-1a35a6c01d6d tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] Task: {'id': task-5116861, 'name': Rename_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1162.961764] env[65726]: DEBUG nova.virt.hardware [None req-168947ab-de46-4966-89ee-e38c38aa73bc tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1162.962030] env[65726]: DEBUG nova.virt.hardware [None req-168947ab-de46-4966-89ee-e38c38aa73bc tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1162.962193] env[65726]: DEBUG nova.virt.hardware [None req-168947ab-de46-4966-89ee-e38c38aa73bc tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1162.962367] env[65726]: DEBUG nova.virt.hardware [None req-168947ab-de46-4966-89ee-e38c38aa73bc tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1162.962510] env[65726]: DEBUG nova.virt.hardware [None req-168947ab-de46-4966-89ee-e38c38aa73bc tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1162.962650] env[65726]: DEBUG nova.virt.hardware [None req-168947ab-de46-4966-89ee-e38c38aa73bc tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1162.962849] env[65726]: DEBUG nova.virt.hardware [None req-168947ab-de46-4966-89ee-e38c38aa73bc tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1162.962999] env[65726]: DEBUG nova.virt.hardware [None req-168947ab-de46-4966-89ee-e38c38aa73bc tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1162.963211] env[65726]: DEBUG nova.virt.hardware [None req-168947ab-de46-4966-89ee-e38c38aa73bc tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1162.963369] env[65726]: DEBUG nova.virt.hardware [None req-168947ab-de46-4966-89ee-e38c38aa73bc tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1162.963574] env[65726]: DEBUG nova.virt.hardware [None req-168947ab-de46-4966-89ee-e38c38aa73bc tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1162.972704] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f5eb7e5d-dca1-47e2-8299-cfd721911e39 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.995210] env[65726]: DEBUG nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] [instance: 2fe6fc5a-f5c7-4f8e-96df-4e621a252f04] Instance has had 0 of 5 cleanup attempts {{(pid=65726) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 1163.003526] env[65726]: DEBUG oslo_vmware.api [None req-168947ab-de46-4966-89ee-e38c38aa73bc tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Waiting for the task: (returnval){ [ 1163.003526] env[65726]: value = "task-5116862" [ 1163.003526] env[65726]: _type = "Task" [ 1163.003526] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1163.014264] env[65726]: DEBUG oslo_vmware.api [None req-168947ab-de46-4966-89ee-e38c38aa73bc tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': task-5116862, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1163.078357] env[65726]: DEBUG oslo_vmware.api [None req-e8a62c22-e16d-473f-bc47-d51d9d4949ce tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Task: {'id': task-5116860, 'name': ReconfigVM_Task, 'duration_secs': 0.283233} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1163.078696] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-e8a62c22-e16d-473f-bc47-d51d9d4949ce tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 9c0599e3-383f-46ab-809b-944cc3a4d206] Reconfigured VM instance instance-0000006d to attach disk [datastore2] 9c0599e3-383f-46ab-809b-944cc3a4d206/9c0599e3-383f-46ab-809b-944cc3a4d206.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1163.079578] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-095ce4dd-f1aa-4bb5-b228-b68e613022e1 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.088160] env[65726]: DEBUG oslo_vmware.api [None req-e8a62c22-e16d-473f-bc47-d51d9d4949ce tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Waiting for the task: (returnval){ [ 1163.088160] env[65726]: value = "task-5116863" [ 1163.088160] env[65726]: _type = "Task" [ 1163.088160] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1163.099455] env[65726]: DEBUG oslo_vmware.api [None req-e8a62c22-e16d-473f-bc47-d51d9d4949ce tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Task: {'id': task-5116863, 'name': Rename_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1163.254762] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a25d259a-c453-4078-8dde-1bd0d195d5e7 tempest-AttachInterfacesTestJSON-763311326 tempest-AttachInterfacesTestJSON-763311326-project-member] Lock "51f5337d-2e23-4b7c-b590-76be6bd82411" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.289s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1163.299883] env[65726]: DEBUG oslo_vmware.api [None req-3a8ffaf6-e93e-4fc3-a889-1a35a6c01d6d tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] Task: {'id': task-5116861, 'name': Rename_Task} progress is 99%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1163.510678] env[65726]: DEBUG nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] [instance: cf950ec3-9914-4eb9-99db-048b3969bb21] Instance has had 0 of 5 cleanup attempts {{(pid=65726) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 1163.522172] env[65726]: DEBUG oslo_vmware.api [None req-168947ab-de46-4966-89ee-e38c38aa73bc tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': task-5116862, 'name': ReconfigVM_Task, 'duration_secs': 0.431953} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1163.522172] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-168947ab-de46-4966-89ee-e38c38aa73bc tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 7c1188ce-8718-4719-8631-e59e7915b7aa] Updating instance '7c1188ce-8718-4719-8631-e59e7915b7aa' progress to 33 {{(pid=65726) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1163.602404] env[65726]: DEBUG oslo_vmware.api [None req-e8a62c22-e16d-473f-bc47-d51d9d4949ce tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Task: {'id': task-5116863, 'name': Rename_Task, 'duration_secs': 0.476279} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1163.602736] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-e8a62c22-e16d-473f-bc47-d51d9d4949ce tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 9c0599e3-383f-46ab-809b-944cc3a4d206] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1163.603007] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-34d36051-8d4e-49f7-bd33-9c28345d7016 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.613111] env[65726]: DEBUG oslo_vmware.api [None req-e8a62c22-e16d-473f-bc47-d51d9d4949ce tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Waiting for the task: (returnval){ [ 1163.613111] env[65726]: value = "task-5116864" [ 1163.613111] env[65726]: _type = "Task" [ 1163.613111] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1163.626073] env[65726]: DEBUG oslo_vmware.api [None req-e8a62c22-e16d-473f-bc47-d51d9d4949ce tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Task: {'id': task-5116864, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1163.798343] env[65726]: DEBUG oslo_vmware.api [None req-3a8ffaf6-e93e-4fc3-a889-1a35a6c01d6d tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] Task: {'id': task-5116861, 'name': Rename_Task, 'duration_secs': 0.533669} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1163.798829] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-3a8ffaf6-e93e-4fc3-a889-1a35a6c01d6d tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] [instance: 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1163.798984] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-304f9c25-51bb-4ad8-86cc-2249c23a4bb7 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.806876] env[65726]: DEBUG oslo_vmware.api [None req-3a8ffaf6-e93e-4fc3-a889-1a35a6c01d6d tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] Waiting for the task: (returnval){ [ 1163.806876] env[65726]: value = "task-5116865" [ 1163.806876] env[65726]: _type = "Task" [ 1163.806876] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1163.820027] env[65726]: DEBUG oslo_vmware.api [None req-3a8ffaf6-e93e-4fc3-a889-1a35a6c01d6d tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] Task: {'id': task-5116865, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1164.015946] env[65726]: DEBUG nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] [instance: c4cc4f85-cb35-4edc-a58b-adfee0ce1265] Instance has had 0 of 5 cleanup attempts {{(pid=65726) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 1164.032029] env[65726]: DEBUG nova.virt.hardware [None req-168947ab-de46-4966-89ee-e38c38aa73bc tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1164.032029] env[65726]: DEBUG nova.virt.hardware [None req-168947ab-de46-4966-89ee-e38c38aa73bc tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1164.032029] env[65726]: DEBUG nova.virt.hardware [None req-168947ab-de46-4966-89ee-e38c38aa73bc tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1164.032029] env[65726]: DEBUG nova.virt.hardware [None req-168947ab-de46-4966-89ee-e38c38aa73bc tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1164.032424] env[65726]: DEBUG nova.virt.hardware [None req-168947ab-de46-4966-89ee-e38c38aa73bc tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1164.032720] env[65726]: DEBUG nova.virt.hardware [None req-168947ab-de46-4966-89ee-e38c38aa73bc tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1164.033092] env[65726]: DEBUG nova.virt.hardware [None req-168947ab-de46-4966-89ee-e38c38aa73bc tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1164.033356] env[65726]: DEBUG nova.virt.hardware [None req-168947ab-de46-4966-89ee-e38c38aa73bc tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1164.033664] env[65726]: DEBUG nova.virt.hardware [None req-168947ab-de46-4966-89ee-e38c38aa73bc tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1164.033924] env[65726]: DEBUG nova.virt.hardware [None req-168947ab-de46-4966-89ee-e38c38aa73bc tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1164.034201] env[65726]: DEBUG nova.virt.hardware [None req-168947ab-de46-4966-89ee-e38c38aa73bc tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1164.039884] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-168947ab-de46-4966-89ee-e38c38aa73bc tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 7c1188ce-8718-4719-8631-e59e7915b7aa] Reconfiguring VM instance instance-0000006a to detach disk 2000 {{(pid=65726) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1164.040524] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cf907d87-7c75-4d5e-8f36-0f1fad184bb6 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.060908] env[65726]: DEBUG oslo_vmware.api [None req-168947ab-de46-4966-89ee-e38c38aa73bc tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Waiting for the task: (returnval){ [ 1164.060908] env[65726]: value = "task-5116866" [ 1164.060908] env[65726]: _type = "Task" [ 1164.060908] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1164.071255] env[65726]: DEBUG oslo_vmware.api [None req-168947ab-de46-4966-89ee-e38c38aa73bc tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': task-5116866, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1164.132370] env[65726]: DEBUG oslo_vmware.api [None req-e8a62c22-e16d-473f-bc47-d51d9d4949ce tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Task: {'id': task-5116864, 'name': PowerOnVM_Task} progress is 78%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1164.321718] env[65726]: DEBUG oslo_vmware.api [None req-3a8ffaf6-e93e-4fc3-a889-1a35a6c01d6d tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] Task: {'id': task-5116865, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1164.519959] env[65726]: DEBUG nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] [instance: 2cbee359-a1fc-45c0-bcc5-a2c24ded9c69] Instance has had 0 of 5 cleanup attempts {{(pid=65726) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 1164.574872] env[65726]: DEBUG oslo_vmware.api [None req-168947ab-de46-4966-89ee-e38c38aa73bc tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': task-5116866, 'name': ReconfigVM_Task, 'duration_secs': 0.201879} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1164.575200] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-168947ab-de46-4966-89ee-e38c38aa73bc tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 7c1188ce-8718-4719-8631-e59e7915b7aa] Reconfigured VM instance instance-0000006a to detach disk 2000 {{(pid=65726) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1164.576054] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3202700d-977b-4ee5-85d1-e1675ab49f9a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.607064] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-168947ab-de46-4966-89ee-e38c38aa73bc tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 7c1188ce-8718-4719-8631-e59e7915b7aa] Reconfiguring VM instance instance-0000006a to attach disk [datastore1] 7c1188ce-8718-4719-8631-e59e7915b7aa/7c1188ce-8718-4719-8631-e59e7915b7aa.vmdk or device None with type thin {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1164.607064] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-29646ae3-b80e-422c-bb34-1bfe1184a2a0 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.622224] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a3f83649-0195-4086-b546-9898d337f536 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Acquiring lock "b353b5e9-500c-42d3-a87f-880a624febec" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1164.622470] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a3f83649-0195-4086-b546-9898d337f536 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Lock "b353b5e9-500c-42d3-a87f-880a624febec" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1164.622650] env[65726]: INFO nova.compute.manager [None req-a3f83649-0195-4086-b546-9898d337f536 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: b353b5e9-500c-42d3-a87f-880a624febec] Shelving [ 1164.636135] env[65726]: DEBUG oslo_vmware.api [None req-e8a62c22-e16d-473f-bc47-d51d9d4949ce tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Task: {'id': task-5116864, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1164.637246] env[65726]: DEBUG oslo_vmware.api [None req-168947ab-de46-4966-89ee-e38c38aa73bc tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Waiting for the task: (returnval){ [ 1164.637246] env[65726]: value = "task-5116867" [ 1164.637246] env[65726]: _type = "Task" [ 1164.637246] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1164.650040] env[65726]: DEBUG oslo_vmware.api [None req-168947ab-de46-4966-89ee-e38c38aa73bc tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': task-5116867, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1164.821811] env[65726]: DEBUG oslo_vmware.api [None req-3a8ffaf6-e93e-4fc3-a889-1a35a6c01d6d tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] Task: {'id': task-5116865, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1165.025352] env[65726]: DEBUG nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] [instance: d179030a-e5ba-45b1-ad11-4d2f71ed7bd4] Instance has had 0 of 5 cleanup attempts {{(pid=65726) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 1165.133852] env[65726]: DEBUG oslo_vmware.api [None req-e8a62c22-e16d-473f-bc47-d51d9d4949ce tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Task: {'id': task-5116864, 'name': PowerOnVM_Task, 'duration_secs': 1.036981} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1165.135996] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-e8a62c22-e16d-473f-bc47-d51d9d4949ce tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 9c0599e3-383f-46ab-809b-944cc3a4d206] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1165.136207] env[65726]: INFO nova.compute.manager [None req-e8a62c22-e16d-473f-bc47-d51d9d4949ce tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 9c0599e3-383f-46ab-809b-944cc3a4d206] Took 9.05 seconds to spawn the instance on the hypervisor. [ 1165.136378] env[65726]: DEBUG nova.compute.manager [None req-e8a62c22-e16d-473f-bc47-d51d9d4949ce tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 9c0599e3-383f-46ab-809b-944cc3a4d206] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1165.137939] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f2eb3aa-b234-48c1-b44e-535e2446c2ea {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.151390] env[65726]: DEBUG oslo_vmware.api [None req-168947ab-de46-4966-89ee-e38c38aa73bc tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': task-5116867, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1165.319746] env[65726]: DEBUG oslo_vmware.api [None req-3a8ffaf6-e93e-4fc3-a889-1a35a6c01d6d tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] Task: {'id': task-5116865, 'name': PowerOnVM_Task, 'duration_secs': 1.137583} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1165.320412] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-3a8ffaf6-e93e-4fc3-a889-1a35a6c01d6d tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] [instance: 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1165.320708] env[65726]: INFO nova.compute.manager [None req-3a8ffaf6-e93e-4fc3-a889-1a35a6c01d6d tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] [instance: 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30] Took 10.29 seconds to spawn the instance on the hypervisor. [ 1165.320904] env[65726]: DEBUG nova.compute.manager [None req-3a8ffaf6-e93e-4fc3-a889-1a35a6c01d6d tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] [instance: 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1165.321777] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-575d1047-173d-47ae-b305-75cfafec0f76 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.528951] env[65726]: DEBUG nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] [instance: c4177e20-b1bd-4b54-a275-c93582359a07] Instance has had 0 of 5 cleanup attempts {{(pid=65726) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 1165.641785] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-a3f83649-0195-4086-b546-9898d337f536 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: b353b5e9-500c-42d3-a87f-880a624febec] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1165.642274] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-41e8750b-f51c-4084-b8fb-fa64232ab8cb {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.655606] env[65726]: DEBUG oslo_vmware.api [None req-168947ab-de46-4966-89ee-e38c38aa73bc tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': task-5116867, 'name': ReconfigVM_Task, 'duration_secs': 0.72524} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1165.656812] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-168947ab-de46-4966-89ee-e38c38aa73bc tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 7c1188ce-8718-4719-8631-e59e7915b7aa] Reconfigured VM instance instance-0000006a to attach disk [datastore1] 7c1188ce-8718-4719-8631-e59e7915b7aa/7c1188ce-8718-4719-8631-e59e7915b7aa.vmdk or device None with type thin {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1165.657176] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-168947ab-de46-4966-89ee-e38c38aa73bc tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 7c1188ce-8718-4719-8631-e59e7915b7aa] Updating instance '7c1188ce-8718-4719-8631-e59e7915b7aa' progress to 50 {{(pid=65726) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1165.668327] env[65726]: DEBUG oslo_vmware.api [None req-a3f83649-0195-4086-b546-9898d337f536 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Waiting for the task: (returnval){ [ 1165.668327] env[65726]: value = "task-5116868" [ 1165.668327] env[65726]: _type = "Task" [ 1165.668327] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1165.677578] env[65726]: INFO nova.compute.manager [None req-e8a62c22-e16d-473f-bc47-d51d9d4949ce tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 9c0599e3-383f-46ab-809b-944cc3a4d206] Took 16.66 seconds to build instance. [ 1165.685502] env[65726]: DEBUG oslo_vmware.api [None req-a3f83649-0195-4086-b546-9898d337f536 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Task: {'id': task-5116868, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1165.844240] env[65726]: INFO nova.compute.manager [None req-3a8ffaf6-e93e-4fc3-a889-1a35a6c01d6d tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] [instance: 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30] Took 21.90 seconds to build instance. [ 1166.034357] env[65726]: DEBUG nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] [instance: cb8855b6-589d-4863-b86f-7fb4e30fb29e] Instance has had 0 of 5 cleanup attempts {{(pid=65726) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 1166.173131] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01c09365-7884-4896-be65-ccd95f0f086f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.180297] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e8a62c22-e16d-473f-bc47-d51d9d4949ce tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Lock "9c0599e3-383f-46ab-809b-944cc3a4d206" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 18.169s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1166.189311] env[65726]: DEBUG oslo_vmware.api [None req-a3f83649-0195-4086-b546-9898d337f536 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Task: {'id': task-5116868, 'name': PowerOffVM_Task} progress is 100%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1166.205521] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-887d3c3e-ba1a-4990-911a-bb8aa713f3b2 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.231525] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-168947ab-de46-4966-89ee-e38c38aa73bc tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 7c1188ce-8718-4719-8631-e59e7915b7aa] Updating instance '7c1188ce-8718-4719-8631-e59e7915b7aa' progress to 67 {{(pid=65726) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1166.348579] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3a8ffaf6-e93e-4fc3-a889-1a35a6c01d6d tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] Lock "47e52bc1-c18e-4aa2-82a2-e4cb030a7a30" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 23.411s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1166.538916] env[65726]: DEBUG nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] [instance: 44d73b2d-2b6a-4501-9944-432da3c9330d] Instance has had 0 of 5 cleanup attempts {{(pid=65726) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 1166.681665] env[65726]: DEBUG oslo_vmware.api [None req-a3f83649-0195-4086-b546-9898d337f536 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Task: {'id': task-5116868, 'name': PowerOffVM_Task, 'duration_secs': 0.583763} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1166.682387] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-a3f83649-0195-4086-b546-9898d337f536 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: b353b5e9-500c-42d3-a87f-880a624febec] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1166.682750] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-109cde84-41fe-4704-a352-f737ca56bbbe {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.702744] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-356372f5-4cd1-4edd-91da-de5a25bb71b7 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.739631] env[65726]: WARNING neutronclient.v2_0.client [None req-168947ab-de46-4966-89ee-e38c38aa73bc tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1166.943663] env[65726]: DEBUG nova.network.neutron [None req-168947ab-de46-4966-89ee-e38c38aa73bc tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 7c1188ce-8718-4719-8631-e59e7915b7aa] Port e8f51770-ca83-4aa3-87c1-ef47653d9bb2 binding to destination host cpu-1 is already ACTIVE {{(pid=65726) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3236}} [ 1167.041713] env[65726]: DEBUG nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] [instance: 1cfb1bba-1e21-47eb-b50f-a86575846a65] Instance has had 0 of 5 cleanup attempts {{(pid=65726) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 1167.214619] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-a3f83649-0195-4086-b546-9898d337f536 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: b353b5e9-500c-42d3-a87f-880a624febec] Creating Snapshot of the VM instance {{(pid=65726) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1167.215516] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-cd6b4215-a824-4d7d-ab6a-6711375974f0 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.217256] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b7bd70fc-78fa-4df3-bb13-d10a1172d2a5 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Acquiring lock "378c1d18-4bb6-4245-80bf-370fb7af0575" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1167.218023] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b7bd70fc-78fa-4df3-bb13-d10a1172d2a5 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Lock "378c1d18-4bb6-4245-80bf-370fb7af0575" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1167.218023] env[65726]: INFO nova.compute.manager [None req-b7bd70fc-78fa-4df3-bb13-d10a1172d2a5 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 378c1d18-4bb6-4245-80bf-370fb7af0575] Shelving [ 1167.226981] env[65726]: DEBUG oslo_vmware.api [None req-a3f83649-0195-4086-b546-9898d337f536 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Waiting for the task: (returnval){ [ 1167.226981] env[65726]: value = "task-5116869" [ 1167.226981] env[65726]: _type = "Task" [ 1167.226981] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1167.241268] env[65726]: DEBUG oslo_vmware.api [None req-a3f83649-0195-4086-b546-9898d337f536 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Task: {'id': task-5116869, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1167.546436] env[65726]: DEBUG nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] [instance: 01d86089-6b9b-4588-864e-ef91375a1eea] Instance has had 0 of 5 cleanup attempts {{(pid=65726) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 1167.745981] env[65726]: DEBUG oslo_vmware.api [None req-a3f83649-0195-4086-b546-9898d337f536 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Task: {'id': task-5116869, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1167.974362] env[65726]: DEBUG oslo_concurrency.lockutils [None req-168947ab-de46-4966-89ee-e38c38aa73bc tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Acquiring lock "7c1188ce-8718-4719-8631-e59e7915b7aa-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1167.975400] env[65726]: DEBUG oslo_concurrency.lockutils [None req-168947ab-de46-4966-89ee-e38c38aa73bc tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Lock "7c1188ce-8718-4719-8631-e59e7915b7aa-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1167.975400] env[65726]: DEBUG oslo_concurrency.lockutils [None req-168947ab-de46-4966-89ee-e38c38aa73bc tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Lock "7c1188ce-8718-4719-8631-e59e7915b7aa-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1168.047170] env[65726]: DEBUG nova.compute.manager [req-f9643885-9d5d-4071-b255-bfd827d17221 req-abb4de93-f294-41fa-8d27-a8dbd4a12a92 service nova] [instance: 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30] Received event network-changed-e0f18f03-6a1b-4d5e-9b8e-68051f61ad8b {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1168.047170] env[65726]: DEBUG nova.compute.manager [req-f9643885-9d5d-4071-b255-bfd827d17221 req-abb4de93-f294-41fa-8d27-a8dbd4a12a92 service nova] [instance: 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30] Refreshing instance network info cache due to event network-changed-e0f18f03-6a1b-4d5e-9b8e-68051f61ad8b. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 1168.047170] env[65726]: DEBUG oslo_concurrency.lockutils [req-f9643885-9d5d-4071-b255-bfd827d17221 req-abb4de93-f294-41fa-8d27-a8dbd4a12a92 service nova] Acquiring lock "refresh_cache-47e52bc1-c18e-4aa2-82a2-e4cb030a7a30" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1168.047170] env[65726]: DEBUG oslo_concurrency.lockutils [req-f9643885-9d5d-4071-b255-bfd827d17221 req-abb4de93-f294-41fa-8d27-a8dbd4a12a92 service nova] Acquired lock "refresh_cache-47e52bc1-c18e-4aa2-82a2-e4cb030a7a30" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1168.047170] env[65726]: DEBUG nova.network.neutron [req-f9643885-9d5d-4071-b255-bfd827d17221 req-abb4de93-f294-41fa-8d27-a8dbd4a12a92 service nova] [instance: 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30] Refreshing network info cache for port e0f18f03-6a1b-4d5e-9b8e-68051f61ad8b {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 1168.053243] env[65726]: DEBUG nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] [instance: c8be64a3-1bfa-41a3-b9be-68a7e6e052e0] Instance has had 0 of 5 cleanup attempts {{(pid=65726) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 1168.242179] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-b7bd70fc-78fa-4df3-bb13-d10a1172d2a5 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 378c1d18-4bb6-4245-80bf-370fb7af0575] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1168.242488] env[65726]: DEBUG oslo_vmware.api [None req-a3f83649-0195-4086-b546-9898d337f536 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Task: {'id': task-5116869, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1168.242707] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9fd78ea2-3807-47fe-a075-f271f609c5dc {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.252501] env[65726]: DEBUG oslo_vmware.api [None req-b7bd70fc-78fa-4df3-bb13-d10a1172d2a5 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Waiting for the task: (returnval){ [ 1168.252501] env[65726]: value = "task-5116870" [ 1168.252501] env[65726]: _type = "Task" [ 1168.252501] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1168.265328] env[65726]: DEBUG oslo_vmware.api [None req-b7bd70fc-78fa-4df3-bb13-d10a1172d2a5 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5116870, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1168.551211] env[65726]: WARNING openstack [req-f9643885-9d5d-4071-b255-bfd827d17221 req-abb4de93-f294-41fa-8d27-a8dbd4a12a92 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1168.551527] env[65726]: WARNING openstack [req-f9643885-9d5d-4071-b255-bfd827d17221 req-abb4de93-f294-41fa-8d27-a8dbd4a12a92 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1168.558945] env[65726]: DEBUG nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] [instance: 3cbbe146-be3a-43d5-867d-4d669884758c] Instance has had 0 of 5 cleanup attempts {{(pid=65726) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 1168.741021] env[65726]: DEBUG oslo_vmware.api [None req-a3f83649-0195-4086-b546-9898d337f536 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Task: {'id': task-5116869, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1168.764636] env[65726]: DEBUG oslo_vmware.api [None req-b7bd70fc-78fa-4df3-bb13-d10a1172d2a5 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5116870, 'name': PowerOffVM_Task, 'duration_secs': 0.247409} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1168.765733] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-b7bd70fc-78fa-4df3-bb13-d10a1172d2a5 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 378c1d18-4bb6-4245-80bf-370fb7af0575] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1168.766366] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4a11f1a-3f4a-410f-9c1d-463a06f981f5 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.804179] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e49ca4b-0770-46aa-b8e1-7178f82f1930 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.965589] env[65726]: WARNING openstack [req-f9643885-9d5d-4071-b255-bfd827d17221 req-abb4de93-f294-41fa-8d27-a8dbd4a12a92 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1168.965972] env[65726]: WARNING openstack [req-f9643885-9d5d-4071-b255-bfd827d17221 req-abb4de93-f294-41fa-8d27-a8dbd4a12a92 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1168.982844] env[65726]: WARNING openstack [None req-168947ab-de46-4966-89ee-e38c38aa73bc tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1168.983281] env[65726]: WARNING openstack [None req-168947ab-de46-4966-89ee-e38c38aa73bc tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1169.063900] env[65726]: DEBUG nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] [instance: 6c8db442-daaa-4eca-ae24-b9d4afaf3a71] Instance has had 0 of 5 cleanup attempts {{(pid=65726) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 1169.231622] env[65726]: DEBUG oslo_concurrency.lockutils [None req-168947ab-de46-4966-89ee-e38c38aa73bc tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Acquiring lock "refresh_cache-7c1188ce-8718-4719-8631-e59e7915b7aa" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1169.231826] env[65726]: DEBUG oslo_concurrency.lockutils [None req-168947ab-de46-4966-89ee-e38c38aa73bc tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Acquired lock "refresh_cache-7c1188ce-8718-4719-8631-e59e7915b7aa" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1169.231972] env[65726]: DEBUG nova.network.neutron [None req-168947ab-de46-4966-89ee-e38c38aa73bc tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 7c1188ce-8718-4719-8631-e59e7915b7aa] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1169.246563] env[65726]: DEBUG oslo_vmware.api [None req-a3f83649-0195-4086-b546-9898d337f536 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Task: {'id': task-5116869, 'name': CreateSnapshot_Task, 'duration_secs': 1.531875} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1169.247483] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-a3f83649-0195-4086-b546-9898d337f536 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: b353b5e9-500c-42d3-a87f-880a624febec] Created Snapshot of the VM instance {{(pid=65726) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1169.248295] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-506faa06-d209-49fc-b96d-c9d86089a92e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.321107] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-b7bd70fc-78fa-4df3-bb13-d10a1172d2a5 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 378c1d18-4bb6-4245-80bf-370fb7af0575] Creating Snapshot of the VM instance {{(pid=65726) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1169.321742] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-74603d0a-5767-4ac2-9a1b-c56646b54283 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.331109] env[65726]: DEBUG oslo_vmware.api [None req-b7bd70fc-78fa-4df3-bb13-d10a1172d2a5 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Waiting for the task: (returnval){ [ 1169.331109] env[65726]: value = "task-5116871" [ 1169.331109] env[65726]: _type = "Task" [ 1169.331109] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1169.341664] env[65726]: DEBUG oslo_vmware.api [None req-b7bd70fc-78fa-4df3-bb13-d10a1172d2a5 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5116871, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1169.432321] env[65726]: WARNING openstack [req-f9643885-9d5d-4071-b255-bfd827d17221 req-abb4de93-f294-41fa-8d27-a8dbd4a12a92 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1169.432868] env[65726]: WARNING openstack [req-f9643885-9d5d-4071-b255-bfd827d17221 req-abb4de93-f294-41fa-8d27-a8dbd4a12a92 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1169.567604] env[65726]: DEBUG nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] [instance: bc2c12e0-0d06-432f-b42f-be468e3b6ee4] Instance has had 0 of 5 cleanup attempts {{(pid=65726) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 1169.742537] env[65726]: WARNING openstack [None req-168947ab-de46-4966-89ee-e38c38aa73bc tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1169.744093] env[65726]: WARNING openstack [None req-168947ab-de46-4966-89ee-e38c38aa73bc tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1169.768897] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-a3f83649-0195-4086-b546-9898d337f536 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: b353b5e9-500c-42d3-a87f-880a624febec] Creating linked-clone VM from snapshot {{(pid=65726) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1169.770695] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-6843c708-72a7-48d2-a5ce-da1f45106b31 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.781175] env[65726]: DEBUG oslo_vmware.api [None req-a3f83649-0195-4086-b546-9898d337f536 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Waiting for the task: (returnval){ [ 1169.781175] env[65726]: value = "task-5116872" [ 1169.781175] env[65726]: _type = "Task" [ 1169.781175] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1169.792246] env[65726]: DEBUG oslo_vmware.api [None req-a3f83649-0195-4086-b546-9898d337f536 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Task: {'id': task-5116872, 'name': CloneVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1169.841929] env[65726]: DEBUG oslo_vmware.api [None req-b7bd70fc-78fa-4df3-bb13-d10a1172d2a5 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5116871, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1169.921049] env[65726]: DEBUG nova.network.neutron [req-f9643885-9d5d-4071-b255-bfd827d17221 req-abb4de93-f294-41fa-8d27-a8dbd4a12a92 service nova] [instance: 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30] Updated VIF entry in instance network info cache for port e0f18f03-6a1b-4d5e-9b8e-68051f61ad8b. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 1169.921511] env[65726]: DEBUG nova.network.neutron [req-f9643885-9d5d-4071-b255-bfd827d17221 req-abb4de93-f294-41fa-8d27-a8dbd4a12a92 service nova] [instance: 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30] Updating instance_info_cache with network_info: [{"id": "e0f18f03-6a1b-4d5e-9b8e-68051f61ad8b", "address": "fa:16:3e:39:d3:8c", "network": {"id": "ba97c3c2-c46b-4ce9-8f4d-b20857d6f8ec", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-992165785-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.128", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1e3203c117de4d4d9c4c90436801ff3f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f856fca-9fb5-41ea-a057-ac4193bd323d", "external-id": "nsx-vlan-transportzone-148", "segmentation_id": 148, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape0f18f03-6a", "ovs_interfaceid": "e0f18f03-6a1b-4d5e-9b8e-68051f61ad8b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1170.071366] env[65726]: DEBUG nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] [instance: 76249623-6f83-46a3-b8c5-c001111aa698] Instance has had 0 of 5 cleanup attempts {{(pid=65726) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 1170.296917] env[65726]: DEBUG oslo_vmware.api [None req-a3f83649-0195-4086-b546-9898d337f536 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Task: {'id': task-5116872, 'name': CloneVM_Task} progress is 94%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1170.309594] env[65726]: WARNING openstack [None req-168947ab-de46-4966-89ee-e38c38aa73bc tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1170.309594] env[65726]: WARNING openstack [None req-168947ab-de46-4966-89ee-e38c38aa73bc tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1170.326372] env[65726]: DEBUG oslo_concurrency.lockutils [None req-f04e60e2-18c2-41e8-8f5e-88ce749ed252 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Acquiring lock "dc112e67-7c2d-4081-9a53-e4f43f61dcd6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1170.326773] env[65726]: DEBUG oslo_concurrency.lockutils [None req-f04e60e2-18c2-41e8-8f5e-88ce749ed252 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Lock "dc112e67-7c2d-4081-9a53-e4f43f61dcd6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1170.342772] env[65726]: DEBUG oslo_vmware.api [None req-b7bd70fc-78fa-4df3-bb13-d10a1172d2a5 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5116871, 'name': CreateSnapshot_Task, 'duration_secs': 0.992922} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1170.343747] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-b7bd70fc-78fa-4df3-bb13-d10a1172d2a5 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 378c1d18-4bb6-4245-80bf-370fb7af0575] Created Snapshot of the VM instance {{(pid=65726) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1170.344547] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-adc953da-8d51-4172-ab37-da499b20ad00 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.425585] env[65726]: DEBUG oslo_concurrency.lockutils [req-f9643885-9d5d-4071-b255-bfd827d17221 req-abb4de93-f294-41fa-8d27-a8dbd4a12a92 service nova] Releasing lock "refresh_cache-47e52bc1-c18e-4aa2-82a2-e4cb030a7a30" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1170.474062] env[65726]: WARNING openstack [None req-168947ab-de46-4966-89ee-e38c38aa73bc tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1170.474938] env[65726]: WARNING openstack [None req-168947ab-de46-4966-89ee-e38c38aa73bc tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1170.575478] env[65726]: DEBUG nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] [instance: 9fc0c9ee-9dc7-4059-b8d2-a5db965e00c2] Instance has had 0 of 5 cleanup attempts {{(pid=65726) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 1170.796139] env[65726]: DEBUG oslo_vmware.api [None req-a3f83649-0195-4086-b546-9898d337f536 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Task: {'id': task-5116872, 'name': CloneVM_Task} progress is 94%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1170.798721] env[65726]: DEBUG nova.network.neutron [None req-168947ab-de46-4966-89ee-e38c38aa73bc tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 7c1188ce-8718-4719-8631-e59e7915b7aa] Updating instance_info_cache with network_info: [{"id": "e8f51770-ca83-4aa3-87c1-ef47653d9bb2", "address": "fa:16:3e:8e:ce:ad", "network": {"id": "9dd55fb6-0dd2-484d-b94d-fbdccb958c79", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1005598846-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f93d852e2c904f42981cbfff882558e9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "74f30339-6421-4654-bddb-81d7f34db9d7", "external-id": "nsx-vlan-transportzone-899", "segmentation_id": 899, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape8f51770-ca", "ovs_interfaceid": "e8f51770-ca83-4aa3-87c1-ef47653d9bb2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1170.830493] env[65726]: DEBUG nova.compute.manager [None req-f04e60e2-18c2-41e8-8f5e-88ce749ed252 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: dc112e67-7c2d-4081-9a53-e4f43f61dcd6] Starting instance... {{(pid=65726) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 1170.874962] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-b7bd70fc-78fa-4df3-bb13-d10a1172d2a5 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 378c1d18-4bb6-4245-80bf-370fb7af0575] Creating linked-clone VM from snapshot {{(pid=65726) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1170.876506] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-553faf55-ad96-42d2-a3e1-c3047f48cc11 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.888837] env[65726]: DEBUG oslo_vmware.api [None req-b7bd70fc-78fa-4df3-bb13-d10a1172d2a5 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Waiting for the task: (returnval){ [ 1170.888837] env[65726]: value = "task-5116873" [ 1170.888837] env[65726]: _type = "Task" [ 1170.888837] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1170.899195] env[65726]: DEBUG oslo_vmware.api [None req-b7bd70fc-78fa-4df3-bb13-d10a1172d2a5 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5116873, 'name': CloneVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1171.079135] env[65726]: DEBUG nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] [instance: 6d2b86c2-0a44-433f-8b3f-2e9c7f693baa] Instance has had 0 of 5 cleanup attempts {{(pid=65726) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 1171.294425] env[65726]: DEBUG oslo_vmware.api [None req-a3f83649-0195-4086-b546-9898d337f536 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Task: {'id': task-5116872, 'name': CloneVM_Task, 'duration_secs': 1.385664} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1171.294773] env[65726]: INFO nova.virt.vmwareapi.vmops [None req-a3f83649-0195-4086-b546-9898d337f536 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: b353b5e9-500c-42d3-a87f-880a624febec] Created linked-clone VM from snapshot [ 1171.295666] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a4f18f3-eae4-4602-8630-a51e0245a203 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1171.304721] env[65726]: DEBUG oslo_concurrency.lockutils [None req-168947ab-de46-4966-89ee-e38c38aa73bc tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Releasing lock "refresh_cache-7c1188ce-8718-4719-8631-e59e7915b7aa" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1171.308527] env[65726]: DEBUG nova.virt.vmwareapi.images [None req-a3f83649-0195-4086-b546-9898d337f536 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: b353b5e9-500c-42d3-a87f-880a624febec] Uploading image 745a44b6-0c3f-4396-a82a-6e9472e10925 {{(pid=65726) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 1171.342555] env[65726]: DEBUG oslo_vmware.rw_handles [None req-a3f83649-0195-4086-b546-9898d337f536 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1171.342555] env[65726]: value = "vm-995316" [ 1171.342555] env[65726]: _type = "VirtualMachine" [ 1171.342555] env[65726]: }. {{(pid=65726) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1171.342555] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-34bdb1d6-8311-44b6-b49a-5a89125c55e1 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1171.354618] env[65726]: DEBUG oslo_vmware.rw_handles [None req-a3f83649-0195-4086-b546-9898d337f536 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Lease: (returnval){ [ 1171.354618] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]525bf1cc-c128-28f1-ddbb-8d52c1cfd96e" [ 1171.354618] env[65726]: _type = "HttpNfcLease" [ 1171.354618] env[65726]: } obtained for exporting VM: (result){ [ 1171.354618] env[65726]: value = "vm-995316" [ 1171.354618] env[65726]: _type = "VirtualMachine" [ 1171.354618] env[65726]: }. {{(pid=65726) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1171.355702] env[65726]: DEBUG oslo_vmware.api [None req-a3f83649-0195-4086-b546-9898d337f536 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Waiting for the lease: (returnval){ [ 1171.355702] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]525bf1cc-c128-28f1-ddbb-8d52c1cfd96e" [ 1171.355702] env[65726]: _type = "HttpNfcLease" [ 1171.355702] env[65726]: } to be ready. {{(pid=65726) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1171.364045] env[65726]: DEBUG oslo_concurrency.lockutils [None req-f04e60e2-18c2-41e8-8f5e-88ce749ed252 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1171.364045] env[65726]: DEBUG oslo_concurrency.lockutils [None req-f04e60e2-18c2-41e8-8f5e-88ce749ed252 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1171.364738] env[65726]: INFO nova.compute.claims [None req-f04e60e2-18c2-41e8-8f5e-88ce749ed252 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: dc112e67-7c2d-4081-9a53-e4f43f61dcd6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1171.369878] env[65726]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1171.369878] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]525bf1cc-c128-28f1-ddbb-8d52c1cfd96e" [ 1171.369878] env[65726]: _type = "HttpNfcLease" [ 1171.369878] env[65726]: } is initializing. {{(pid=65726) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1171.401361] env[65726]: DEBUG oslo_vmware.api [None req-b7bd70fc-78fa-4df3-bb13-d10a1172d2a5 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5116873, 'name': CloneVM_Task} progress is 94%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1171.582990] env[65726]: DEBUG nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] [instance: 895797e4-2941-44cd-aab1-67afa4fac02d] Instance has had 0 of 5 cleanup attempts {{(pid=65726) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 1171.663140] env[65726]: DEBUG oslo_concurrency.lockutils [None req-d8745e4a-3a23-4daa-80f7-d677b94a15ab tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Acquiring lock "e3255f3b-028f-4a0b-b621-873730417936" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1171.663370] env[65726]: DEBUG oslo_concurrency.lockutils [None req-d8745e4a-3a23-4daa-80f7-d677b94a15ab tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Lock "e3255f3b-028f-4a0b-b621-873730417936" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1171.838983] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b92f969e-efa2-4e53-ba0d-80408fdf3b02 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1171.864512] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d70b7527-7d22-490c-ab1e-63a3cc2a653e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1171.877496] env[65726]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1171.877496] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]525bf1cc-c128-28f1-ddbb-8d52c1cfd96e" [ 1171.877496] env[65726]: _type = "HttpNfcLease" [ 1171.877496] env[65726]: } is ready. {{(pid=65726) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1171.877834] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-168947ab-de46-4966-89ee-e38c38aa73bc tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 7c1188ce-8718-4719-8631-e59e7915b7aa] Updating instance '7c1188ce-8718-4719-8631-e59e7915b7aa' progress to 83 {{(pid=65726) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1171.881358] env[65726]: DEBUG oslo_vmware.rw_handles [None req-a3f83649-0195-4086-b546-9898d337f536 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1171.881358] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]525bf1cc-c128-28f1-ddbb-8d52c1cfd96e" [ 1171.881358] env[65726]: _type = "HttpNfcLease" [ 1171.881358] env[65726]: }. {{(pid=65726) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1171.882204] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4875679d-bc59-409e-b5f5-ebbfaf05bb15 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1171.891807] env[65726]: DEBUG oslo_vmware.rw_handles [None req-a3f83649-0195-4086-b546-9898d337f536 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52cf9848-25af-835a-954d-ad2a7bf6bcc4/disk-0.vmdk from lease info. {{(pid=65726) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1171.892133] env[65726]: DEBUG oslo_vmware.rw_handles [None req-a3f83649-0195-4086-b546-9898d337f536 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52cf9848-25af-835a-954d-ad2a7bf6bcc4/disk-0.vmdk for reading. {{(pid=65726) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1171.973037] env[65726]: DEBUG oslo_vmware.api [None req-b7bd70fc-78fa-4df3-bb13-d10a1172d2a5 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5116873, 'name': CloneVM_Task} progress is 94%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1172.009943] env[65726]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-71d22cff-23a0-4d5c-a332-5efcdf3f6897 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.086700] env[65726]: DEBUG nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] [instance: c108f2a5-031b-47ee-9a5b-d62c2f42c26b] Instance has had 0 of 5 cleanup attempts {{(pid=65726) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 1172.168212] env[65726]: DEBUG nova.compute.manager [None req-d8745e4a-3a23-4daa-80f7-d677b94a15ab tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: e3255f3b-028f-4a0b-b621-873730417936] Starting instance... {{(pid=65726) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 1172.387758] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-168947ab-de46-4966-89ee-e38c38aa73bc tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 7c1188ce-8718-4719-8631-e59e7915b7aa] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1172.388139] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3f72feeb-0485-43bb-a6f4-c0e28f56273b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.398693] env[65726]: DEBUG oslo_vmware.api [None req-168947ab-de46-4966-89ee-e38c38aa73bc tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Waiting for the task: (returnval){ [ 1172.398693] env[65726]: value = "task-5116875" [ 1172.398693] env[65726]: _type = "Task" [ 1172.398693] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1172.408785] env[65726]: DEBUG oslo_vmware.api [None req-b7bd70fc-78fa-4df3-bb13-d10a1172d2a5 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5116873, 'name': CloneVM_Task} progress is 95%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1172.422078] env[65726]: DEBUG oslo_vmware.api [None req-168947ab-de46-4966-89ee-e38c38aa73bc tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': task-5116875, 'name': PowerOnVM_Task} progress is 33%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1172.591910] env[65726]: DEBUG nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] [instance: 608cc0ea-3f6b-4b4f-83c1-01aa50999d63] Instance has had 0 of 5 cleanup attempts {{(pid=65726) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 1172.599025] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee4af43c-f3e6-4d1c-a7a6-2ae250573ce2 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.609174] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8066d2c6-63d7-463a-8454-d34bb4fad8d5 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.647706] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0e2f082-119d-4424-b33c-6172070a550c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.658688] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8e52035-a483-412d-a21f-8375bbe41eae {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.679555] env[65726]: DEBUG nova.compute.provider_tree [None req-f04e60e2-18c2-41e8-8f5e-88ce749ed252 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1172.696334] env[65726]: DEBUG oslo_concurrency.lockutils [None req-d8745e4a-3a23-4daa-80f7-d677b94a15ab tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1172.907567] env[65726]: DEBUG oslo_vmware.api [None req-b7bd70fc-78fa-4df3-bb13-d10a1172d2a5 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5116873, 'name': CloneVM_Task, 'duration_secs': 1.878808} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1172.912355] env[65726]: INFO nova.virt.vmwareapi.vmops [None req-b7bd70fc-78fa-4df3-bb13-d10a1172d2a5 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 378c1d18-4bb6-4245-80bf-370fb7af0575] Created linked-clone VM from snapshot [ 1172.913994] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c169582-77f0-4ee8-8036-026b2b5c8665 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.923633] env[65726]: DEBUG oslo_vmware.api [None req-168947ab-de46-4966-89ee-e38c38aa73bc tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': task-5116875, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1172.927669] env[65726]: DEBUG nova.virt.vmwareapi.images [None req-b7bd70fc-78fa-4df3-bb13-d10a1172d2a5 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 378c1d18-4bb6-4245-80bf-370fb7af0575] Uploading image 1d5801e6-cfd0-4c6a-8fa0-57ee72cc963a {{(pid=65726) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 1172.955867] env[65726]: DEBUG oslo_vmware.rw_handles [None req-b7bd70fc-78fa-4df3-bb13-d10a1172d2a5 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1172.955867] env[65726]: value = "vm-995318" [ 1172.955867] env[65726]: _type = "VirtualMachine" [ 1172.955867] env[65726]: }. {{(pid=65726) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1172.956444] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-a22c3086-4e19-476b-a4d6-76930f664dee {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.967072] env[65726]: DEBUG oslo_vmware.rw_handles [None req-b7bd70fc-78fa-4df3-bb13-d10a1172d2a5 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Lease: (returnval){ [ 1172.967072] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5234bbca-ac5c-8bb7-7885-d90e2e49f382" [ 1172.967072] env[65726]: _type = "HttpNfcLease" [ 1172.967072] env[65726]: } obtained for exporting VM: (result){ [ 1172.967072] env[65726]: value = "vm-995318" [ 1172.967072] env[65726]: _type = "VirtualMachine" [ 1172.967072] env[65726]: }. {{(pid=65726) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1172.967582] env[65726]: DEBUG oslo_vmware.api [None req-b7bd70fc-78fa-4df3-bb13-d10a1172d2a5 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Waiting for the lease: (returnval){ [ 1172.967582] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5234bbca-ac5c-8bb7-7885-d90e2e49f382" [ 1172.967582] env[65726]: _type = "HttpNfcLease" [ 1172.967582] env[65726]: } to be ready. {{(pid=65726) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1172.980289] env[65726]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1172.980289] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5234bbca-ac5c-8bb7-7885-d90e2e49f382" [ 1172.980289] env[65726]: _type = "HttpNfcLease" [ 1172.980289] env[65726]: } is initializing. {{(pid=65726) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1173.103594] env[65726]: DEBUG nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] [instance: d7905c4f-1ba3-43b3-b4b3-aebacd2f9cc3] Instance has had 0 of 5 cleanup attempts {{(pid=65726) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 1173.188376] env[65726]: DEBUG nova.scheduler.client.report [None req-f04e60e2-18c2-41e8-8f5e-88ce749ed252 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1173.384870] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3b1f8e43-10c3-447e-a11b-a1499eca83f6 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Acquiring lock "55ee97c3-b014-48eb-b41b-fc5aa16bc09e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1173.385349] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3b1f8e43-10c3-447e-a11b-a1499eca83f6 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Lock "55ee97c3-b014-48eb-b41b-fc5aa16bc09e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1173.420780] env[65726]: DEBUG oslo_vmware.api [None req-168947ab-de46-4966-89ee-e38c38aa73bc tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': task-5116875, 'name': PowerOnVM_Task, 'duration_secs': 0.635772} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1173.422099] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-168947ab-de46-4966-89ee-e38c38aa73bc tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 7c1188ce-8718-4719-8631-e59e7915b7aa] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1173.425025] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-168947ab-de46-4966-89ee-e38c38aa73bc tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 7c1188ce-8718-4719-8631-e59e7915b7aa] Updating instance '7c1188ce-8718-4719-8631-e59e7915b7aa' progress to 100 {{(pid=65726) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1173.478858] env[65726]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1173.478858] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5234bbca-ac5c-8bb7-7885-d90e2e49f382" [ 1173.478858] env[65726]: _type = "HttpNfcLease" [ 1173.478858] env[65726]: } is ready. {{(pid=65726) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1173.479279] env[65726]: DEBUG oslo_vmware.rw_handles [None req-b7bd70fc-78fa-4df3-bb13-d10a1172d2a5 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1173.479279] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5234bbca-ac5c-8bb7-7885-d90e2e49f382" [ 1173.479279] env[65726]: _type = "HttpNfcLease" [ 1173.479279] env[65726]: }. {{(pid=65726) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1173.480235] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f757e932-41a6-4c4f-8616-fd35e6fbfc57 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.489629] env[65726]: DEBUG oslo_vmware.rw_handles [None req-b7bd70fc-78fa-4df3-bb13-d10a1172d2a5 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52b93164-06a5-d34b-5e6c-a27fd01cc0a1/disk-0.vmdk from lease info. {{(pid=65726) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1173.490056] env[65726]: DEBUG oslo_vmware.rw_handles [None req-b7bd70fc-78fa-4df3-bb13-d10a1172d2a5 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52b93164-06a5-d34b-5e6c-a27fd01cc0a1/disk-0.vmdk for reading. {{(pid=65726) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1173.609156] env[65726]: DEBUG nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] [instance: 274ab469-61a9-4b7e-852c-074c871e3abf] Instance has had 0 of 5 cleanup attempts {{(pid=65726) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 1173.639170] env[65726]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-5b02a556-724d-45f4-9b38-c3a1d2feb383 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.695631] env[65726]: DEBUG oslo_concurrency.lockutils [None req-f04e60e2-18c2-41e8-8f5e-88ce749ed252 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.333s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1173.696081] env[65726]: DEBUG nova.compute.manager [None req-f04e60e2-18c2-41e8-8f5e-88ce749ed252 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: dc112e67-7c2d-4081-9a53-e4f43f61dcd6] Start building networks asynchronously for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1173.699609] env[65726]: DEBUG oslo_concurrency.lockutils [None req-d8745e4a-3a23-4daa-80f7-d677b94a15ab tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.003s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1173.701297] env[65726]: INFO nova.compute.claims [None req-d8745e4a-3a23-4daa-80f7-d677b94a15ab tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: e3255f3b-028f-4a0b-b621-873730417936] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1173.889072] env[65726]: DEBUG nova.compute.manager [None req-3b1f8e43-10c3-447e-a11b-a1499eca83f6 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 55ee97c3-b014-48eb-b41b-fc5aa16bc09e] Starting instance... {{(pid=65726) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 1174.113251] env[65726]: DEBUG nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] [instance: f2d9090c-988f-43f4-9c81-7aa718a3438a] Instance has had 0 of 5 cleanup attempts {{(pid=65726) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 1174.201512] env[65726]: DEBUG nova.compute.utils [None req-f04e60e2-18c2-41e8-8f5e-88ce749ed252 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Using /dev/sd instead of None {{(pid=65726) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1174.203121] env[65726]: DEBUG nova.compute.manager [None req-f04e60e2-18c2-41e8-8f5e-88ce749ed252 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: dc112e67-7c2d-4081-9a53-e4f43f61dcd6] Allocating IP information in the background. {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1174.203795] env[65726]: DEBUG nova.network.neutron [None req-f04e60e2-18c2-41e8-8f5e-88ce749ed252 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: dc112e67-7c2d-4081-9a53-e4f43f61dcd6] allocate_for_instance() {{(pid=65726) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 1174.203911] env[65726]: WARNING neutronclient.v2_0.client [None req-f04e60e2-18c2-41e8-8f5e-88ce749ed252 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1174.204314] env[65726]: WARNING neutronclient.v2_0.client [None req-f04e60e2-18c2-41e8-8f5e-88ce749ed252 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1174.205015] env[65726]: WARNING openstack [None req-f04e60e2-18c2-41e8-8f5e-88ce749ed252 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1174.205528] env[65726]: WARNING openstack [None req-f04e60e2-18c2-41e8-8f5e-88ce749ed252 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1174.271933] env[65726]: DEBUG nova.policy [None req-f04e60e2-18c2-41e8-8f5e-88ce749ed252 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ccb3513b5b4f453f842139ec1815c4af', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f3f799aa94f64e8cb26d93f6124efd81', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65726) authorize /opt/stack/nova/nova/policy.py:192}} [ 1174.415935] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3b1f8e43-10c3-447e-a11b-a1499eca83f6 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1174.615139] env[65726]: DEBUG nova.network.neutron [None req-f04e60e2-18c2-41e8-8f5e-88ce749ed252 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: dc112e67-7c2d-4081-9a53-e4f43f61dcd6] Successfully created port: 46483adf-29bb-4abf-b767-42ff3c0b3c20 {{(pid=65726) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1174.619590] env[65726]: DEBUG nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] [instance: 0e064341-4e4a-407b-8c26-3eb04b409029] Instance has had 0 of 5 cleanup attempts {{(pid=65726) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 1174.714806] env[65726]: DEBUG nova.compute.manager [None req-f04e60e2-18c2-41e8-8f5e-88ce749ed252 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: dc112e67-7c2d-4081-9a53-e4f43f61dcd6] Start building block device mappings for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1174.915252] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b8d5838-3a32-4eb6-b2a0-3ac8c03c3d0e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.924283] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13d3ee32-1142-457e-a961-a36413d4bba7 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.967902] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7483779-53e5-4370-85e4-3522a8d4ab34 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.977763] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0b55fc2-e64c-4e7b-ad4c-9e0ddd23b619 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.996207] env[65726]: DEBUG nova.compute.provider_tree [None req-d8745e4a-3a23-4daa-80f7-d677b94a15ab tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1175.123313] env[65726]: DEBUG oslo_service.periodic_task [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=65726) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1175.123717] env[65726]: DEBUG nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Cleaning up deleted instances with incomplete migration {{(pid=65726) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11970}} [ 1175.361762] env[65726]: DEBUG oslo_concurrency.lockutils [None req-ea123e2e-31cf-42d0-af88-34d2c73324cf tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Acquiring lock "7c1188ce-8718-4719-8631-e59e7915b7aa" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1175.362173] env[65726]: DEBUG oslo_concurrency.lockutils [None req-ea123e2e-31cf-42d0-af88-34d2c73324cf tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Lock "7c1188ce-8718-4719-8631-e59e7915b7aa" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.001s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1175.362433] env[65726]: DEBUG nova.compute.manager [None req-ea123e2e-31cf-42d0-af88-34d2c73324cf tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 7c1188ce-8718-4719-8631-e59e7915b7aa] Going to confirm migration 7 {{(pid=65726) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:5307}} [ 1175.500506] env[65726]: DEBUG nova.scheduler.client.report [None req-d8745e4a-3a23-4daa-80f7-d677b94a15ab tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1175.627147] env[65726]: DEBUG oslo_service.periodic_task [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=65726) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1175.729065] env[65726]: DEBUG nova.compute.manager [None req-f04e60e2-18c2-41e8-8f5e-88ce749ed252 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: dc112e67-7c2d-4081-9a53-e4f43f61dcd6] Start spawning the instance on the hypervisor. {{(pid=65726) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1175.873183] env[65726]: WARNING openstack [None req-ea123e2e-31cf-42d0-af88-34d2c73324cf tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1175.873655] env[65726]: WARNING openstack [None req-ea123e2e-31cf-42d0-af88-34d2c73324cf tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1175.911544] env[65726]: WARNING openstack [None req-ea123e2e-31cf-42d0-af88-34d2c73324cf tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1175.911961] env[65726]: WARNING openstack [None req-ea123e2e-31cf-42d0-af88-34d2c73324cf tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1175.917446] env[65726]: DEBUG oslo_concurrency.lockutils [None req-ea123e2e-31cf-42d0-af88-34d2c73324cf tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Acquiring lock "refresh_cache-7c1188ce-8718-4719-8631-e59e7915b7aa" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1175.917608] env[65726]: DEBUG oslo_concurrency.lockutils [None req-ea123e2e-31cf-42d0-af88-34d2c73324cf tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Acquired lock "refresh_cache-7c1188ce-8718-4719-8631-e59e7915b7aa" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1175.917776] env[65726]: DEBUG nova.network.neutron [None req-ea123e2e-31cf-42d0-af88-34d2c73324cf tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 7c1188ce-8718-4719-8631-e59e7915b7aa] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1175.917952] env[65726]: DEBUG nova.objects.instance [None req-ea123e2e-31cf-42d0-af88-34d2c73324cf tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Lazy-loading 'info_cache' on Instance uuid 7c1188ce-8718-4719-8631-e59e7915b7aa {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1176.006158] env[65726]: DEBUG oslo_concurrency.lockutils [None req-d8745e4a-3a23-4daa-80f7-d677b94a15ab tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.307s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1176.006833] env[65726]: DEBUG nova.compute.manager [None req-d8745e4a-3a23-4daa-80f7-d677b94a15ab tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: e3255f3b-028f-4a0b-b621-873730417936] Start building networks asynchronously for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1176.010598] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3b1f8e43-10c3-447e-a11b-a1499eca83f6 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.594s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1176.012313] env[65726]: INFO nova.compute.claims [None req-3b1f8e43-10c3-447e-a11b-a1499eca83f6 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 55ee97c3-b014-48eb-b41b-fc5aa16bc09e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1176.127177] env[65726]: DEBUG nova.compute.manager [req-67a76416-56e9-449c-812f-3007e77c4833 req-d3a03c91-4592-43f3-b352-16491bd3efe5 service nova] [instance: dc112e67-7c2d-4081-9a53-e4f43f61dcd6] Received event network-vif-plugged-46483adf-29bb-4abf-b767-42ff3c0b3c20 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1176.127177] env[65726]: DEBUG oslo_concurrency.lockutils [req-67a76416-56e9-449c-812f-3007e77c4833 req-d3a03c91-4592-43f3-b352-16491bd3efe5 service nova] Acquiring lock "dc112e67-7c2d-4081-9a53-e4f43f61dcd6-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1176.127399] env[65726]: DEBUG oslo_concurrency.lockutils [req-67a76416-56e9-449c-812f-3007e77c4833 req-d3a03c91-4592-43f3-b352-16491bd3efe5 service nova] Lock "dc112e67-7c2d-4081-9a53-e4f43f61dcd6-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1176.127487] env[65726]: DEBUG oslo_concurrency.lockutils [req-67a76416-56e9-449c-812f-3007e77c4833 req-d3a03c91-4592-43f3-b352-16491bd3efe5 service nova] Lock "dc112e67-7c2d-4081-9a53-e4f43f61dcd6-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1176.127649] env[65726]: DEBUG nova.compute.manager [req-67a76416-56e9-449c-812f-3007e77c4833 req-d3a03c91-4592-43f3-b352-16491bd3efe5 service nova] [instance: dc112e67-7c2d-4081-9a53-e4f43f61dcd6] No waiting events found dispatching network-vif-plugged-46483adf-29bb-4abf-b767-42ff3c0b3c20 {{(pid=65726) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1176.127808] env[65726]: WARNING nova.compute.manager [req-67a76416-56e9-449c-812f-3007e77c4833 req-d3a03c91-4592-43f3-b352-16491bd3efe5 service nova] [instance: dc112e67-7c2d-4081-9a53-e4f43f61dcd6] Received unexpected event network-vif-plugged-46483adf-29bb-4abf-b767-42ff3c0b3c20 for instance with vm_state building and task_state spawning. [ 1176.231179] env[65726]: DEBUG nova.network.neutron [None req-f04e60e2-18c2-41e8-8f5e-88ce749ed252 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: dc112e67-7c2d-4081-9a53-e4f43f61dcd6] Successfully updated port: 46483adf-29bb-4abf-b767-42ff3c0b3c20 {{(pid=65726) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 1176.512368] env[65726]: DEBUG nova.compute.utils [None req-d8745e4a-3a23-4daa-80f7-d677b94a15ab tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Using /dev/sd instead of None {{(pid=65726) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1176.513906] env[65726]: DEBUG nova.compute.manager [None req-d8745e4a-3a23-4daa-80f7-d677b94a15ab tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: e3255f3b-028f-4a0b-b621-873730417936] Allocating IP information in the background. {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1176.514133] env[65726]: DEBUG nova.network.neutron [None req-d8745e4a-3a23-4daa-80f7-d677b94a15ab tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: e3255f3b-028f-4a0b-b621-873730417936] allocate_for_instance() {{(pid=65726) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 1176.514461] env[65726]: WARNING neutronclient.v2_0.client [None req-d8745e4a-3a23-4daa-80f7-d677b94a15ab tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1176.514811] env[65726]: WARNING neutronclient.v2_0.client [None req-d8745e4a-3a23-4daa-80f7-d677b94a15ab tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1176.515416] env[65726]: WARNING openstack [None req-d8745e4a-3a23-4daa-80f7-d677b94a15ab tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1176.515745] env[65726]: WARNING openstack [None req-d8745e4a-3a23-4daa-80f7-d677b94a15ab tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1176.578019] env[65726]: DEBUG nova.policy [None req-d8745e4a-3a23-4daa-80f7-d677b94a15ab tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5143cc185c7645b48304c8ebc5c854de', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'acd800e0438940bc80ae3c6b672db4ee', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65726) authorize /opt/stack/nova/nova/policy.py:192}} [ 1176.735074] env[65726]: DEBUG oslo_concurrency.lockutils [None req-f04e60e2-18c2-41e8-8f5e-88ce749ed252 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Acquiring lock "refresh_cache-dc112e67-7c2d-4081-9a53-e4f43f61dcd6" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1176.735408] env[65726]: DEBUG oslo_concurrency.lockutils [None req-f04e60e2-18c2-41e8-8f5e-88ce749ed252 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Acquired lock "refresh_cache-dc112e67-7c2d-4081-9a53-e4f43f61dcd6" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1176.735480] env[65726]: DEBUG nova.network.neutron [None req-f04e60e2-18c2-41e8-8f5e-88ce749ed252 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: dc112e67-7c2d-4081-9a53-e4f43f61dcd6] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1176.844887] env[65726]: DEBUG nova.network.neutron [None req-d8745e4a-3a23-4daa-80f7-d677b94a15ab tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: e3255f3b-028f-4a0b-b621-873730417936] Successfully created port: 03c9e6fb-2435-43be-bb55-8afecdcf38ff {{(pid=65726) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1176.925818] env[65726]: WARNING openstack [None req-ea123e2e-31cf-42d0-af88-34d2c73324cf tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1176.926079] env[65726]: WARNING openstack [None req-ea123e2e-31cf-42d0-af88-34d2c73324cf tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1177.023635] env[65726]: DEBUG nova.compute.manager [None req-d8745e4a-3a23-4daa-80f7-d677b94a15ab tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: e3255f3b-028f-4a0b-b621-873730417936] Start building block device mappings for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1177.036208] env[65726]: WARNING openstack [None req-ea123e2e-31cf-42d0-af88-34d2c73324cf tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1177.036886] env[65726]: WARNING openstack [None req-ea123e2e-31cf-42d0-af88-34d2c73324cf tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1177.156117] env[65726]: WARNING openstack [None req-ea123e2e-31cf-42d0-af88-34d2c73324cf tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1177.156621] env[65726]: WARNING openstack [None req-ea123e2e-31cf-42d0-af88-34d2c73324cf tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1177.219152] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc1023fe-d76a-42c6-aca8-a3d07decb462 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1177.228392] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84091c2b-7c8c-479b-92e5-4532ab6fd59f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1177.264313] env[65726]: WARNING openstack [None req-f04e60e2-18c2-41e8-8f5e-88ce749ed252 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1177.264741] env[65726]: WARNING openstack [None req-f04e60e2-18c2-41e8-8f5e-88ce749ed252 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1177.271366] env[65726]: DEBUG nova.network.neutron [None req-f04e60e2-18c2-41e8-8f5e-88ce749ed252 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: dc112e67-7c2d-4081-9a53-e4f43f61dcd6] Instance cache missing network info. {{(pid=65726) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 1177.276189] env[65726]: DEBUG nova.network.neutron [None req-ea123e2e-31cf-42d0-af88-34d2c73324cf tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 7c1188ce-8718-4719-8631-e59e7915b7aa] Updating instance_info_cache with network_info: [{"id": "e8f51770-ca83-4aa3-87c1-ef47653d9bb2", "address": "fa:16:3e:8e:ce:ad", "network": {"id": "9dd55fb6-0dd2-484d-b94d-fbdccb958c79", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1005598846-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f93d852e2c904f42981cbfff882558e9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "74f30339-6421-4654-bddb-81d7f34db9d7", "external-id": "nsx-vlan-transportzone-899", "segmentation_id": 899, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape8f51770-ca", "ovs_interfaceid": "e8f51770-ca83-4aa3-87c1-ef47653d9bb2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1177.278486] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b035778a-92ee-485f-8374-ad9a0f0c1da3 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1177.287211] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d67a6e1c-1125-4231-9a40-5164ff7aa708 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1177.304053] env[65726]: DEBUG nova.compute.provider_tree [None req-3b1f8e43-10c3-447e-a11b-a1499eca83f6 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1177.339994] env[65726]: WARNING openstack [None req-f04e60e2-18c2-41e8-8f5e-88ce749ed252 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1177.340479] env[65726]: WARNING openstack [None req-f04e60e2-18c2-41e8-8f5e-88ce749ed252 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1177.409650] env[65726]: WARNING openstack [None req-f04e60e2-18c2-41e8-8f5e-88ce749ed252 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1177.410193] env[65726]: WARNING openstack [None req-f04e60e2-18c2-41e8-8f5e-88ce749ed252 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1177.519871] env[65726]: DEBUG nova.network.neutron [None req-f04e60e2-18c2-41e8-8f5e-88ce749ed252 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: dc112e67-7c2d-4081-9a53-e4f43f61dcd6] Updating instance_info_cache with network_info: [{"id": "46483adf-29bb-4abf-b767-42ff3c0b3c20", "address": "fa:16:3e:72:9c:f1", "network": {"id": "69642329-e986-48fa-accf-1f296c08116a", "bridge": "br-int", "label": "tempest-ServersTestJSON-1786538788-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f3f799aa94f64e8cb26d93f6124efd81", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "816c6e38-e200-4544-8c5b-9fc3e16c5761", "external-id": "nsx-vlan-transportzone-195", "segmentation_id": 195, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap46483adf-29", "ovs_interfaceid": "46483adf-29bb-4abf-b767-42ff3c0b3c20", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1177.782946] env[65726]: DEBUG oslo_concurrency.lockutils [None req-ea123e2e-31cf-42d0-af88-34d2c73324cf tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Releasing lock "refresh_cache-7c1188ce-8718-4719-8631-e59e7915b7aa" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1177.782946] env[65726]: DEBUG nova.objects.instance [None req-ea123e2e-31cf-42d0-af88-34d2c73324cf tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Lazy-loading 'migration_context' on Instance uuid 7c1188ce-8718-4719-8631-e59e7915b7aa {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1177.808750] env[65726]: DEBUG nova.scheduler.client.report [None req-3b1f8e43-10c3-447e-a11b-a1499eca83f6 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1178.022627] env[65726]: DEBUG oslo_concurrency.lockutils [None req-f04e60e2-18c2-41e8-8f5e-88ce749ed252 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Releasing lock "refresh_cache-dc112e67-7c2d-4081-9a53-e4f43f61dcd6" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1178.023151] env[65726]: DEBUG nova.compute.manager [None req-f04e60e2-18c2-41e8-8f5e-88ce749ed252 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: dc112e67-7c2d-4081-9a53-e4f43f61dcd6] Instance network_info: |[{"id": "46483adf-29bb-4abf-b767-42ff3c0b3c20", "address": "fa:16:3e:72:9c:f1", "network": {"id": "69642329-e986-48fa-accf-1f296c08116a", "bridge": "br-int", "label": "tempest-ServersTestJSON-1786538788-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f3f799aa94f64e8cb26d93f6124efd81", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "816c6e38-e200-4544-8c5b-9fc3e16c5761", "external-id": "nsx-vlan-transportzone-195", "segmentation_id": 195, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap46483adf-29", "ovs_interfaceid": "46483adf-29bb-4abf-b767-42ff3c0b3c20", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1178.036897] env[65726]: DEBUG nova.compute.manager [None req-d8745e4a-3a23-4daa-80f7-d677b94a15ab tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: e3255f3b-028f-4a0b-b621-873730417936] Start spawning the instance on the hypervisor. {{(pid=65726) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1178.162309] env[65726]: DEBUG nova.compute.manager [req-e80faa67-7d72-446d-a95f-b7df01910f15 req-c947abb4-f644-4ef8-bd26-8e7f860456da service nova] [instance: dc112e67-7c2d-4081-9a53-e4f43f61dcd6] Received event network-changed-46483adf-29bb-4abf-b767-42ff3c0b3c20 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1178.162502] env[65726]: DEBUG nova.compute.manager [req-e80faa67-7d72-446d-a95f-b7df01910f15 req-c947abb4-f644-4ef8-bd26-8e7f860456da service nova] [instance: dc112e67-7c2d-4081-9a53-e4f43f61dcd6] Refreshing instance network info cache due to event network-changed-46483adf-29bb-4abf-b767-42ff3c0b3c20. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 1178.162711] env[65726]: DEBUG oslo_concurrency.lockutils [req-e80faa67-7d72-446d-a95f-b7df01910f15 req-c947abb4-f644-4ef8-bd26-8e7f860456da service nova] Acquiring lock "refresh_cache-dc112e67-7c2d-4081-9a53-e4f43f61dcd6" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1178.162850] env[65726]: DEBUG oslo_concurrency.lockutils [req-e80faa67-7d72-446d-a95f-b7df01910f15 req-c947abb4-f644-4ef8-bd26-8e7f860456da service nova] Acquired lock "refresh_cache-dc112e67-7c2d-4081-9a53-e4f43f61dcd6" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1178.162993] env[65726]: DEBUG nova.network.neutron [req-e80faa67-7d72-446d-a95f-b7df01910f15 req-c947abb4-f644-4ef8-bd26-8e7f860456da service nova] [instance: dc112e67-7c2d-4081-9a53-e4f43f61dcd6] Refreshing network info cache for port 46483adf-29bb-4abf-b767-42ff3c0b3c20 {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 1179.178907] env[65726]: DEBUG nova.network.neutron [None req-d8745e4a-3a23-4daa-80f7-d677b94a15ab tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: e3255f3b-028f-4a0b-b621-873730417936] Successfully updated port: 03c9e6fb-2435-43be-bb55-8afecdcf38ff {{(pid=65726) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 1179.181427] env[65726]: DEBUG nova.objects.base [None req-ea123e2e-31cf-42d0-af88-34d2c73324cf tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Object Instance<7c1188ce-8718-4719-8631-e59e7915b7aa> lazy-loaded attributes: info_cache,migration_context {{(pid=65726) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1179.182689] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3b1f8e43-10c3-447e-a11b-a1499eca83f6 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.172s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1179.183324] env[65726]: DEBUG nova.compute.manager [None req-3b1f8e43-10c3-447e-a11b-a1499eca83f6 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 55ee97c3-b014-48eb-b41b-fc5aa16bc09e] Start building networks asynchronously for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1179.187935] env[65726]: WARNING openstack [req-e80faa67-7d72-446d-a95f-b7df01910f15 req-c947abb4-f644-4ef8-bd26-8e7f860456da service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1179.188506] env[65726]: WARNING openstack [req-e80faa67-7d72-446d-a95f-b7df01910f15 req-c947abb4-f644-4ef8-bd26-8e7f860456da service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1179.196350] env[65726]: DEBUG oslo_concurrency.lockutils [None req-d8745e4a-3a23-4daa-80f7-d677b94a15ab tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Acquiring lock "refresh_cache-e3255f3b-028f-4a0b-b621-873730417936" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1179.196517] env[65726]: DEBUG oslo_concurrency.lockutils [None req-d8745e4a-3a23-4daa-80f7-d677b94a15ab tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Acquired lock "refresh_cache-e3255f3b-028f-4a0b-b621-873730417936" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1179.196655] env[65726]: DEBUG nova.network.neutron [None req-d8745e4a-3a23-4daa-80f7-d677b94a15ab tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: e3255f3b-028f-4a0b-b621-873730417936] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1179.198564] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34ac9618-a057-4566-843a-57161d8b383c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1179.220985] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c54ea498-8ef3-436a-83e9-1753f4187f40 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1179.232024] env[65726]: DEBUG oslo_vmware.api [None req-ea123e2e-31cf-42d0-af88-34d2c73324cf tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Waiting for the task: (returnval){ [ 1179.232024] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5290750c-c54b-d70f-6b19-bb5332916cab" [ 1179.232024] env[65726]: _type = "Task" [ 1179.232024] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1179.241115] env[65726]: DEBUG oslo_vmware.api [None req-ea123e2e-31cf-42d0-af88-34d2c73324cf tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5290750c-c54b-d70f-6b19-bb5332916cab, 'name': SearchDatastore_Task, 'duration_secs': 0.010588} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1179.243281] env[65726]: DEBUG oslo_concurrency.lockutils [None req-ea123e2e-31cf-42d0-af88-34d2c73324cf tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1179.243559] env[65726]: DEBUG oslo_concurrency.lockutils [None req-ea123e2e-31cf-42d0-af88-34d2c73324cf tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1179.342286] env[65726]: WARNING openstack [req-e80faa67-7d72-446d-a95f-b7df01910f15 req-c947abb4-f644-4ef8-bd26-8e7f860456da service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1179.342747] env[65726]: WARNING openstack [req-e80faa67-7d72-446d-a95f-b7df01910f15 req-c947abb4-f644-4ef8-bd26-8e7f860456da service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1179.364593] env[65726]: DEBUG oslo_concurrency.lockutils [None req-252f0171-117c-4650-8dad-6b950e3f6801 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Acquiring lock "3bc25953-21b1-4729-af09-e7211fd8b2c6" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1179.364900] env[65726]: DEBUG oslo_concurrency.lockutils [None req-252f0171-117c-4650-8dad-6b950e3f6801 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Lock "3bc25953-21b1-4729-af09-e7211fd8b2c6" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1179.416373] env[65726]: WARNING openstack [req-e80faa67-7d72-446d-a95f-b7df01910f15 req-c947abb4-f644-4ef8-bd26-8e7f860456da service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1179.416857] env[65726]: WARNING openstack [req-e80faa67-7d72-446d-a95f-b7df01910f15 req-c947abb4-f644-4ef8-bd26-8e7f860456da service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1179.530094] env[65726]: DEBUG nova.network.neutron [req-e80faa67-7d72-446d-a95f-b7df01910f15 req-c947abb4-f644-4ef8-bd26-8e7f860456da service nova] [instance: dc112e67-7c2d-4081-9a53-e4f43f61dcd6] Updated VIF entry in instance network info cache for port 46483adf-29bb-4abf-b767-42ff3c0b3c20. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 1179.530472] env[65726]: DEBUG nova.network.neutron [req-e80faa67-7d72-446d-a95f-b7df01910f15 req-c947abb4-f644-4ef8-bd26-8e7f860456da service nova] [instance: dc112e67-7c2d-4081-9a53-e4f43f61dcd6] Updating instance_info_cache with network_info: [{"id": "46483adf-29bb-4abf-b767-42ff3c0b3c20", "address": "fa:16:3e:72:9c:f1", "network": {"id": "69642329-e986-48fa-accf-1f296c08116a", "bridge": "br-int", "label": "tempest-ServersTestJSON-1786538788-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f3f799aa94f64e8cb26d93f6124efd81", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "816c6e38-e200-4544-8c5b-9fc3e16c5761", "external-id": "nsx-vlan-transportzone-195", "segmentation_id": 195, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap46483adf-29", "ovs_interfaceid": "46483adf-29bb-4abf-b767-42ff3c0b3c20", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1179.698815] env[65726]: DEBUG nova.compute.utils [None req-3b1f8e43-10c3-447e-a11b-a1499eca83f6 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Using /dev/sd instead of None {{(pid=65726) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1179.700983] env[65726]: DEBUG nova.compute.manager [None req-3b1f8e43-10c3-447e-a11b-a1499eca83f6 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 55ee97c3-b014-48eb-b41b-fc5aa16bc09e] Allocating IP information in the background. {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1179.701359] env[65726]: DEBUG nova.network.neutron [None req-3b1f8e43-10c3-447e-a11b-a1499eca83f6 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 55ee97c3-b014-48eb-b41b-fc5aa16bc09e] allocate_for_instance() {{(pid=65726) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 1179.701887] env[65726]: WARNING neutronclient.v2_0.client [None req-3b1f8e43-10c3-447e-a11b-a1499eca83f6 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1179.702427] env[65726]: WARNING neutronclient.v2_0.client [None req-3b1f8e43-10c3-447e-a11b-a1499eca83f6 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1179.703357] env[65726]: WARNING openstack [None req-3b1f8e43-10c3-447e-a11b-a1499eca83f6 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1179.703969] env[65726]: WARNING openstack [None req-3b1f8e43-10c3-447e-a11b-a1499eca83f6 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1179.717041] env[65726]: DEBUG nova.compute.manager [None req-3b1f8e43-10c3-447e-a11b-a1499eca83f6 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 55ee97c3-b014-48eb-b41b-fc5aa16bc09e] Start building block device mappings for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1179.721100] env[65726]: WARNING openstack [None req-d8745e4a-3a23-4daa-80f7-d677b94a15ab tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1179.721614] env[65726]: WARNING openstack [None req-d8745e4a-3a23-4daa-80f7-d677b94a15ab tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1179.727455] env[65726]: DEBUG nova.network.neutron [None req-d8745e4a-3a23-4daa-80f7-d677b94a15ab tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: e3255f3b-028f-4a0b-b621-873730417936] Instance cache missing network info. {{(pid=65726) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 1179.764203] env[65726]: DEBUG nova.policy [None req-3b1f8e43-10c3-447e-a11b-a1499eca83f6 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6ab5b309207a46bb9d95998ef0a7a46f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6b56f93ba0e14521921484ac2c785c43', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65726) authorize /opt/stack/nova/nova/policy.py:192}} [ 1179.792573] env[65726]: WARNING openstack [None req-d8745e4a-3a23-4daa-80f7-d677b94a15ab tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1179.793172] env[65726]: WARNING openstack [None req-d8745e4a-3a23-4daa-80f7-d677b94a15ab tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1179.868640] env[65726]: DEBUG nova.compute.utils [None req-252f0171-117c-4650-8dad-6b950e3f6801 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Using /dev/sd instead of None {{(pid=65726) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1179.880929] env[65726]: WARNING openstack [None req-d8745e4a-3a23-4daa-80f7-d677b94a15ab tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1179.881333] env[65726]: WARNING openstack [None req-d8745e4a-3a23-4daa-80f7-d677b94a15ab tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1179.968235] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9cdbfe89-af91-49bc-b366-f51bf3a7d36f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1179.982664] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-951a1467-667e-47fd-8054-c8ceabaa9f0e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1180.019212] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec58a40a-2c71-43b4-a60b-366d7ac9b4f9 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1180.031989] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd852fe1-9d39-4373-9bcd-46e154c74680 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1180.038211] env[65726]: DEBUG oslo_concurrency.lockutils [req-e80faa67-7d72-446d-a95f-b7df01910f15 req-c947abb4-f644-4ef8-bd26-8e7f860456da service nova] Releasing lock "refresh_cache-dc112e67-7c2d-4081-9a53-e4f43f61dcd6" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1180.046613] env[65726]: DEBUG nova.network.neutron [None req-d8745e4a-3a23-4daa-80f7-d677b94a15ab tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: e3255f3b-028f-4a0b-b621-873730417936] Updating instance_info_cache with network_info: [{"id": "03c9e6fb-2435-43be-bb55-8afecdcf38ff", "address": "fa:16:3e:d2:0f:96", "network": {"id": "a899453d-5f43-4650-9642-b33f54a4d2eb", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-614861139-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "acd800e0438940bc80ae3c6b672db4ee", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "205fb402-8eaf-4b61-8f57-8f216024179a", "external-id": "nsx-vlan-transportzone-78", "segmentation_id": 78, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap03c9e6fb-24", "ovs_interfaceid": "03c9e6fb-2435-43be-bb55-8afecdcf38ff", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1180.063367] env[65726]: DEBUG nova.compute.provider_tree [None req-ea123e2e-31cf-42d0-af88-34d2c73324cf tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1180.074369] env[65726]: DEBUG nova.network.neutron [None req-3b1f8e43-10c3-447e-a11b-a1499eca83f6 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 55ee97c3-b014-48eb-b41b-fc5aa16bc09e] Successfully created port: 65a8678e-a808-4eb1-9cb2-2f86b5fbd4b9 {{(pid=65726) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1180.273258] env[65726]: DEBUG nova.compute.manager [req-45301573-bdd0-4708-8fa3-bde7b7f5ebc5 req-ab4a7536-88a5-4c2c-a158-06d70d824422 service nova] [instance: e3255f3b-028f-4a0b-b621-873730417936] Received event network-vif-plugged-03c9e6fb-2435-43be-bb55-8afecdcf38ff {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1180.273543] env[65726]: DEBUG oslo_concurrency.lockutils [req-45301573-bdd0-4708-8fa3-bde7b7f5ebc5 req-ab4a7536-88a5-4c2c-a158-06d70d824422 service nova] Acquiring lock "e3255f3b-028f-4a0b-b621-873730417936-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1180.274984] env[65726]: DEBUG oslo_concurrency.lockutils [req-45301573-bdd0-4708-8fa3-bde7b7f5ebc5 req-ab4a7536-88a5-4c2c-a158-06d70d824422 service nova] Lock "e3255f3b-028f-4a0b-b621-873730417936-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1180.274984] env[65726]: DEBUG oslo_concurrency.lockutils [req-45301573-bdd0-4708-8fa3-bde7b7f5ebc5 req-ab4a7536-88a5-4c2c-a158-06d70d824422 service nova] Lock "e3255f3b-028f-4a0b-b621-873730417936-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1180.274984] env[65726]: DEBUG nova.compute.manager [req-45301573-bdd0-4708-8fa3-bde7b7f5ebc5 req-ab4a7536-88a5-4c2c-a158-06d70d824422 service nova] [instance: e3255f3b-028f-4a0b-b621-873730417936] No waiting events found dispatching network-vif-plugged-03c9e6fb-2435-43be-bb55-8afecdcf38ff {{(pid=65726) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1180.274984] env[65726]: WARNING nova.compute.manager [req-45301573-bdd0-4708-8fa3-bde7b7f5ebc5 req-ab4a7536-88a5-4c2c-a158-06d70d824422 service nova] [instance: e3255f3b-028f-4a0b-b621-873730417936] Received unexpected event network-vif-plugged-03c9e6fb-2435-43be-bb55-8afecdcf38ff for instance with vm_state building and task_state spawning. [ 1180.274984] env[65726]: DEBUG nova.compute.manager [req-45301573-bdd0-4708-8fa3-bde7b7f5ebc5 req-ab4a7536-88a5-4c2c-a158-06d70d824422 service nova] [instance: e3255f3b-028f-4a0b-b621-873730417936] Received event network-changed-03c9e6fb-2435-43be-bb55-8afecdcf38ff {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1180.274984] env[65726]: DEBUG nova.compute.manager [req-45301573-bdd0-4708-8fa3-bde7b7f5ebc5 req-ab4a7536-88a5-4c2c-a158-06d70d824422 service nova] [instance: e3255f3b-028f-4a0b-b621-873730417936] Refreshing instance network info cache due to event network-changed-03c9e6fb-2435-43be-bb55-8afecdcf38ff. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 1180.275461] env[65726]: DEBUG oslo_concurrency.lockutils [req-45301573-bdd0-4708-8fa3-bde7b7f5ebc5 req-ab4a7536-88a5-4c2c-a158-06d70d824422 service nova] Acquiring lock "refresh_cache-e3255f3b-028f-4a0b-b621-873730417936" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1180.375227] env[65726]: DEBUG oslo_concurrency.lockutils [None req-252f0171-117c-4650-8dad-6b950e3f6801 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Lock "3bc25953-21b1-4729-af09-e7211fd8b2c6" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.010s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1180.567014] env[65726]: DEBUG oslo_concurrency.lockutils [None req-d8745e4a-3a23-4daa-80f7-d677b94a15ab tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Releasing lock "refresh_cache-e3255f3b-028f-4a0b-b621-873730417936" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1180.567449] env[65726]: DEBUG nova.compute.manager [None req-d8745e4a-3a23-4daa-80f7-d677b94a15ab tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: e3255f3b-028f-4a0b-b621-873730417936] Instance network_info: |[{"id": "03c9e6fb-2435-43be-bb55-8afecdcf38ff", "address": "fa:16:3e:d2:0f:96", "network": {"id": "a899453d-5f43-4650-9642-b33f54a4d2eb", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-614861139-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "acd800e0438940bc80ae3c6b672db4ee", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "205fb402-8eaf-4b61-8f57-8f216024179a", "external-id": "nsx-vlan-transportzone-78", "segmentation_id": 78, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap03c9e6fb-24", "ovs_interfaceid": "03c9e6fb-2435-43be-bb55-8afecdcf38ff", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1180.568403] env[65726]: DEBUG nova.scheduler.client.report [None req-ea123e2e-31cf-42d0-af88-34d2c73324cf tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1180.571960] env[65726]: DEBUG oslo_concurrency.lockutils [req-45301573-bdd0-4708-8fa3-bde7b7f5ebc5 req-ab4a7536-88a5-4c2c-a158-06d70d824422 service nova] Acquired lock "refresh_cache-e3255f3b-028f-4a0b-b621-873730417936" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1180.572190] env[65726]: DEBUG nova.network.neutron [req-45301573-bdd0-4708-8fa3-bde7b7f5ebc5 req-ab4a7536-88a5-4c2c-a158-06d70d824422 service nova] [instance: e3255f3b-028f-4a0b-b621-873730417936] Refreshing network info cache for port 03c9e6fb-2435-43be-bb55-8afecdcf38ff {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 1180.738461] env[65726]: DEBUG nova.compute.manager [None req-3b1f8e43-10c3-447e-a11b-a1499eca83f6 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 55ee97c3-b014-48eb-b41b-fc5aa16bc09e] Start spawning the instance on the hypervisor. {{(pid=65726) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1181.079124] env[65726]: WARNING openstack [req-45301573-bdd0-4708-8fa3-bde7b7f5ebc5 req-ab4a7536-88a5-4c2c-a158-06d70d824422 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1181.079584] env[65726]: WARNING openstack [req-45301573-bdd0-4708-8fa3-bde7b7f5ebc5 req-ab4a7536-88a5-4c2c-a158-06d70d824422 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1181.219450] env[65726]: WARNING openstack [req-45301573-bdd0-4708-8fa3-bde7b7f5ebc5 req-ab4a7536-88a5-4c2c-a158-06d70d824422 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1181.219837] env[65726]: WARNING openstack [req-45301573-bdd0-4708-8fa3-bde7b7f5ebc5 req-ab4a7536-88a5-4c2c-a158-06d70d824422 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1181.343191] env[65726]: WARNING openstack [req-45301573-bdd0-4708-8fa3-bde7b7f5ebc5 req-ab4a7536-88a5-4c2c-a158-06d70d824422 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1181.347031] env[65726]: WARNING openstack [req-45301573-bdd0-4708-8fa3-bde7b7f5ebc5 req-ab4a7536-88a5-4c2c-a158-06d70d824422 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1181.454303] env[65726]: DEBUG oslo_concurrency.lockutils [None req-252f0171-117c-4650-8dad-6b950e3f6801 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Acquiring lock "3bc25953-21b1-4729-af09-e7211fd8b2c6" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1181.457322] env[65726]: DEBUG oslo_concurrency.lockutils [None req-252f0171-117c-4650-8dad-6b950e3f6801 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Lock "3bc25953-21b1-4729-af09-e7211fd8b2c6" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.003s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1181.457758] env[65726]: INFO nova.compute.manager [None req-252f0171-117c-4650-8dad-6b950e3f6801 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: 3bc25953-21b1-4729-af09-e7211fd8b2c6] Attaching volume b58c28a5-1144-47d9-96ea-c0a59a979c77 to /dev/sdb [ 1181.468260] env[65726]: DEBUG nova.network.neutron [req-45301573-bdd0-4708-8fa3-bde7b7f5ebc5 req-ab4a7536-88a5-4c2c-a158-06d70d824422 service nova] [instance: e3255f3b-028f-4a0b-b621-873730417936] Updated VIF entry in instance network info cache for port 03c9e6fb-2435-43be-bb55-8afecdcf38ff. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 1181.468642] env[65726]: DEBUG nova.network.neutron [req-45301573-bdd0-4708-8fa3-bde7b7f5ebc5 req-ab4a7536-88a5-4c2c-a158-06d70d824422 service nova] [instance: e3255f3b-028f-4a0b-b621-873730417936] Updating instance_info_cache with network_info: [{"id": "03c9e6fb-2435-43be-bb55-8afecdcf38ff", "address": "fa:16:3e:d2:0f:96", "network": {"id": "a899453d-5f43-4650-9642-b33f54a4d2eb", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-614861139-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "acd800e0438940bc80ae3c6b672db4ee", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "205fb402-8eaf-4b61-8f57-8f216024179a", "external-id": "nsx-vlan-transportzone-78", "segmentation_id": 78, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap03c9e6fb-24", "ovs_interfaceid": "03c9e6fb-2435-43be-bb55-8afecdcf38ff", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1181.500060] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1de44f0-9932-4de9-b0ac-a5faf9cd2ad5 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1181.510735] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bfd95277-08c0-4566-a02c-8873a352c245 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1181.527468] env[65726]: DEBUG nova.virt.block_device [None req-252f0171-117c-4650-8dad-6b950e3f6801 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: 3bc25953-21b1-4729-af09-e7211fd8b2c6] Updating existing volume attachment record: 3713f069-cedb-48f8-918b-950256b63f98 {{(pid=65726) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1181.591786] env[65726]: DEBUG oslo_concurrency.lockutils [None req-ea123e2e-31cf-42d0-af88-34d2c73324cf tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.348s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1181.711415] env[65726]: DEBUG nova.network.neutron [None req-3b1f8e43-10c3-447e-a11b-a1499eca83f6 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 55ee97c3-b014-48eb-b41b-fc5aa16bc09e] Successfully updated port: 65a8678e-a808-4eb1-9cb2-2f86b5fbd4b9 {{(pid=65726) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 1181.733732] env[65726]: DEBUG nova.virt.hardware [None req-f04e60e2-18c2-41e8-8f5e-88ce749ed252 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T19:28:34Z,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1539f06fa3534e90acbc3a60c4b8bd3f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T19:28:35Z,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1181.734274] env[65726]: DEBUG nova.virt.hardware [None req-f04e60e2-18c2-41e8-8f5e-88ce749ed252 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1181.734687] env[65726]: DEBUG nova.virt.hardware [None req-f04e60e2-18c2-41e8-8f5e-88ce749ed252 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1181.736445] env[65726]: DEBUG nova.virt.hardware [None req-f04e60e2-18c2-41e8-8f5e-88ce749ed252 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1181.737123] env[65726]: DEBUG nova.virt.hardware [None req-f04e60e2-18c2-41e8-8f5e-88ce749ed252 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1181.737299] env[65726]: DEBUG nova.virt.hardware [None req-f04e60e2-18c2-41e8-8f5e-88ce749ed252 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1181.737536] env[65726]: DEBUG nova.virt.hardware [None req-f04e60e2-18c2-41e8-8f5e-88ce749ed252 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1181.737743] env[65726]: DEBUG nova.virt.hardware [None req-f04e60e2-18c2-41e8-8f5e-88ce749ed252 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1181.737994] env[65726]: DEBUG nova.virt.hardware [None req-f04e60e2-18c2-41e8-8f5e-88ce749ed252 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1181.738193] env[65726]: DEBUG nova.virt.hardware [None req-f04e60e2-18c2-41e8-8f5e-88ce749ed252 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1181.738390] env[65726]: DEBUG nova.virt.hardware [None req-f04e60e2-18c2-41e8-8f5e-88ce749ed252 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1181.741196] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-860e143b-680e-44af-a82b-58500f2042e6 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1181.752085] env[65726]: DEBUG nova.virt.hardware [None req-d8745e4a-3a23-4daa-80f7-d677b94a15ab tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T19:28:34Z,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1539f06fa3534e90acbc3a60c4b8bd3f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T19:28:35Z,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1181.752342] env[65726]: DEBUG nova.virt.hardware [None req-d8745e4a-3a23-4daa-80f7-d677b94a15ab tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1181.752490] env[65726]: DEBUG nova.virt.hardware [None req-d8745e4a-3a23-4daa-80f7-d677b94a15ab tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1181.752664] env[65726]: DEBUG nova.virt.hardware [None req-d8745e4a-3a23-4daa-80f7-d677b94a15ab tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1181.752804] env[65726]: DEBUG nova.virt.hardware [None req-d8745e4a-3a23-4daa-80f7-d677b94a15ab tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1181.752948] env[65726]: DEBUG nova.virt.hardware [None req-d8745e4a-3a23-4daa-80f7-d677b94a15ab tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1181.753167] env[65726]: DEBUG nova.virt.hardware [None req-d8745e4a-3a23-4daa-80f7-d677b94a15ab tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1181.753317] env[65726]: DEBUG nova.virt.hardware [None req-d8745e4a-3a23-4daa-80f7-d677b94a15ab tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1181.753479] env[65726]: DEBUG nova.virt.hardware [None req-d8745e4a-3a23-4daa-80f7-d677b94a15ab tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1181.753650] env[65726]: DEBUG nova.virt.hardware [None req-d8745e4a-3a23-4daa-80f7-d677b94a15ab tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1181.753840] env[65726]: DEBUG nova.virt.hardware [None req-d8745e4a-3a23-4daa-80f7-d677b94a15ab tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1181.755179] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be0cd53f-892a-454e-bec5-152e036215f5 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1181.761802] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f522f96f-e789-4750-80d6-6cc5872843ef {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1181.771990] env[65726]: DEBUG nova.virt.hardware [None req-3b1f8e43-10c3-447e-a11b-a1499eca83f6 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T19:28:34Z,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1539f06fa3534e90acbc3a60c4b8bd3f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T19:28:35Z,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1181.772293] env[65726]: DEBUG nova.virt.hardware [None req-3b1f8e43-10c3-447e-a11b-a1499eca83f6 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1181.772474] env[65726]: DEBUG nova.virt.hardware [None req-3b1f8e43-10c3-447e-a11b-a1499eca83f6 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1181.772653] env[65726]: DEBUG nova.virt.hardware [None req-3b1f8e43-10c3-447e-a11b-a1499eca83f6 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1181.772796] env[65726]: DEBUG nova.virt.hardware [None req-3b1f8e43-10c3-447e-a11b-a1499eca83f6 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1181.772936] env[65726]: DEBUG nova.virt.hardware [None req-3b1f8e43-10c3-447e-a11b-a1499eca83f6 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1181.773165] env[65726]: DEBUG nova.virt.hardware [None req-3b1f8e43-10c3-447e-a11b-a1499eca83f6 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1181.773317] env[65726]: DEBUG nova.virt.hardware [None req-3b1f8e43-10c3-447e-a11b-a1499eca83f6 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1181.773510] env[65726]: DEBUG nova.virt.hardware [None req-3b1f8e43-10c3-447e-a11b-a1499eca83f6 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1181.773702] env[65726]: DEBUG nova.virt.hardware [None req-3b1f8e43-10c3-447e-a11b-a1499eca83f6 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1181.773897] env[65726]: DEBUG nova.virt.hardware [None req-3b1f8e43-10c3-447e-a11b-a1499eca83f6 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1181.783756] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b20fbe9-68d2-4fcd-8fc3-5bc367c69ebd {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1181.790341] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a1f9b89-1a34-4b00-a82f-3631ce1ff5d1 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1181.794082] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-f04e60e2-18c2-41e8-8f5e-88ce749ed252 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: dc112e67-7c2d-4081-9a53-e4f43f61dcd6] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:72:9c:f1', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '816c6e38-e200-4544-8c5b-9fc3e16c5761', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '46483adf-29bb-4abf-b767-42ff3c0b3c20', 'vif_model': 'vmxnet3'}] {{(pid=65726) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1181.801450] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-f04e60e2-18c2-41e8-8f5e-88ce749ed252 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1181.802125] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: dc112e67-7c2d-4081-9a53-e4f43f61dcd6] Creating VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1181.803367] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f89c4ecc-4394-431a-8175-72b47c252901 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1181.824078] env[65726]: DEBUG oslo_vmware.rw_handles [None req-a3f83649-0195-4086-b546-9898d337f536 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52cf9848-25af-835a-954d-ad2a7bf6bcc4/disk-0.vmdk. {{(pid=65726) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1181.832820] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a980d71-4553-4a28-9885-bcfbfb58e00b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1181.836166] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-d8745e4a-3a23-4daa-80f7-d677b94a15ab tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: e3255f3b-028f-4a0b-b621-873730417936] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d2:0f:96', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '205fb402-8eaf-4b61-8f57-8f216024179a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '03c9e6fb-2435-43be-bb55-8afecdcf38ff', 'vif_model': 'vmxnet3'}] {{(pid=65726) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1181.843375] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-d8745e4a-3a23-4daa-80f7-d677b94a15ab tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Creating folder: Project (acd800e0438940bc80ae3c6b672db4ee). Parent ref: group-v995008. {{(pid=65726) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1181.845248] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46e7117e-db26-477f-8390-425797012d84 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1181.851737] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c37a77a9-0396-478f-923e-9ebcb30649f7 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1181.855436] env[65726]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1181.855436] env[65726]: value = "task-5116878" [ 1181.855436] env[65726]: _type = "Task" [ 1181.855436] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1181.863572] env[65726]: DEBUG oslo_vmware.rw_handles [None req-a3f83649-0195-4086-b546-9898d337f536 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52cf9848-25af-835a-954d-ad2a7bf6bcc4/disk-0.vmdk is in state: ready. {{(pid=65726) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1181.863771] env[65726]: ERROR oslo_vmware.rw_handles [None req-a3f83649-0195-4086-b546-9898d337f536 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52cf9848-25af-835a-954d-ad2a7bf6bcc4/disk-0.vmdk due to incomplete transfer. [ 1181.872875] env[65726]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-7ee1d578-4643-47a9-aa57-d4939befabe7 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1181.877048] env[65726]: INFO nova.virt.vmwareapi.vm_util [None req-d8745e4a-3a23-4daa-80f7-d677b94a15ab tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Created folder: Project (acd800e0438940bc80ae3c6b672db4ee) in parent group-v995008. [ 1181.877048] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-d8745e4a-3a23-4daa-80f7-d677b94a15ab tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Creating folder: Instances. Parent ref: group-v995320. {{(pid=65726) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1181.877476] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-3eafa2a5-810f-467e-a14e-ba7f31ed5b0d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1181.882663] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116878, 'name': CreateVM_Task} progress is 10%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1181.885256] env[65726]: DEBUG oslo_vmware.rw_handles [None req-a3f83649-0195-4086-b546-9898d337f536 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52cf9848-25af-835a-954d-ad2a7bf6bcc4/disk-0.vmdk. {{(pid=65726) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1181.885256] env[65726]: DEBUG nova.virt.vmwareapi.images [None req-a3f83649-0195-4086-b546-9898d337f536 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: b353b5e9-500c-42d3-a87f-880a624febec] Uploaded image 745a44b6-0c3f-4396-a82a-6e9472e10925 to the Glance image server {{(pid=65726) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 1181.888429] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-a3f83649-0195-4086-b546-9898d337f536 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: b353b5e9-500c-42d3-a87f-880a624febec] Destroying the VM {{(pid=65726) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 1181.888738] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-17ba99c4-631f-40cb-9c36-1e177cab9ffe {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1181.893887] env[65726]: INFO nova.virt.vmwareapi.vm_util [None req-d8745e4a-3a23-4daa-80f7-d677b94a15ab tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Created folder: Instances in parent group-v995320. [ 1181.894156] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-d8745e4a-3a23-4daa-80f7-d677b94a15ab tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1181.894363] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e3255f3b-028f-4a0b-b621-873730417936] Creating VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1181.894626] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-0e6025b2-7d3f-42f1-b3f4-82337f97af23 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1181.913917] env[65726]: DEBUG oslo_vmware.api [None req-a3f83649-0195-4086-b546-9898d337f536 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Waiting for the task: (returnval){ [ 1181.913917] env[65726]: value = "task-5116881" [ 1181.913917] env[65726]: _type = "Task" [ 1181.913917] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1181.919285] env[65726]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1181.919285] env[65726]: value = "task-5116882" [ 1181.919285] env[65726]: _type = "Task" [ 1181.919285] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1181.925932] env[65726]: DEBUG oslo_vmware.api [None req-a3f83649-0195-4086-b546-9898d337f536 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Task: {'id': task-5116881, 'name': Destroy_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1181.931697] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116882, 'name': CreateVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1181.971654] env[65726]: DEBUG oslo_concurrency.lockutils [req-45301573-bdd0-4708-8fa3-bde7b7f5ebc5 req-ab4a7536-88a5-4c2c-a158-06d70d824422 service nova] Releasing lock "refresh_cache-e3255f3b-028f-4a0b-b621-873730417936" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1182.162794] env[65726]: INFO nova.scheduler.client.report [None req-ea123e2e-31cf-42d0-af88-34d2c73324cf tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Deleted allocation for migration 66575ff4-cc55-4594-bdc2-e957ee22e601 [ 1182.215697] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3b1f8e43-10c3-447e-a11b-a1499eca83f6 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Acquiring lock "refresh_cache-55ee97c3-b014-48eb-b41b-fc5aa16bc09e" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1182.215905] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3b1f8e43-10c3-447e-a11b-a1499eca83f6 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Acquired lock "refresh_cache-55ee97c3-b014-48eb-b41b-fc5aa16bc09e" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1182.216236] env[65726]: DEBUG nova.network.neutron [None req-3b1f8e43-10c3-447e-a11b-a1499eca83f6 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 55ee97c3-b014-48eb-b41b-fc5aa16bc09e] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1182.313119] env[65726]: DEBUG nova.compute.manager [req-667957f8-9b62-4db4-b93b-1eda3adbf91e req-6b7ce011-e8fa-4475-8f71-30734af92d19 service nova] [instance: 55ee97c3-b014-48eb-b41b-fc5aa16bc09e] Received event network-vif-plugged-65a8678e-a808-4eb1-9cb2-2f86b5fbd4b9 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1182.313357] env[65726]: DEBUG oslo_concurrency.lockutils [req-667957f8-9b62-4db4-b93b-1eda3adbf91e req-6b7ce011-e8fa-4475-8f71-30734af92d19 service nova] Acquiring lock "55ee97c3-b014-48eb-b41b-fc5aa16bc09e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1182.313357] env[65726]: DEBUG oslo_concurrency.lockutils [req-667957f8-9b62-4db4-b93b-1eda3adbf91e req-6b7ce011-e8fa-4475-8f71-30734af92d19 service nova] Lock "55ee97c3-b014-48eb-b41b-fc5aa16bc09e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1182.313901] env[65726]: DEBUG oslo_concurrency.lockutils [req-667957f8-9b62-4db4-b93b-1eda3adbf91e req-6b7ce011-e8fa-4475-8f71-30734af92d19 service nova] Lock "55ee97c3-b014-48eb-b41b-fc5aa16bc09e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1182.313901] env[65726]: DEBUG nova.compute.manager [req-667957f8-9b62-4db4-b93b-1eda3adbf91e req-6b7ce011-e8fa-4475-8f71-30734af92d19 service nova] [instance: 55ee97c3-b014-48eb-b41b-fc5aa16bc09e] No waiting events found dispatching network-vif-plugged-65a8678e-a808-4eb1-9cb2-2f86b5fbd4b9 {{(pid=65726) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1182.314302] env[65726]: WARNING nova.compute.manager [req-667957f8-9b62-4db4-b93b-1eda3adbf91e req-6b7ce011-e8fa-4475-8f71-30734af92d19 service nova] [instance: 55ee97c3-b014-48eb-b41b-fc5aa16bc09e] Received unexpected event network-vif-plugged-65a8678e-a808-4eb1-9cb2-2f86b5fbd4b9 for instance with vm_state building and task_state spawning. [ 1182.314673] env[65726]: DEBUG nova.compute.manager [req-667957f8-9b62-4db4-b93b-1eda3adbf91e req-6b7ce011-e8fa-4475-8f71-30734af92d19 service nova] [instance: 55ee97c3-b014-48eb-b41b-fc5aa16bc09e] Received event network-changed-65a8678e-a808-4eb1-9cb2-2f86b5fbd4b9 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1182.315031] env[65726]: DEBUG nova.compute.manager [req-667957f8-9b62-4db4-b93b-1eda3adbf91e req-6b7ce011-e8fa-4475-8f71-30734af92d19 service nova] [instance: 55ee97c3-b014-48eb-b41b-fc5aa16bc09e] Refreshing instance network info cache due to event network-changed-65a8678e-a808-4eb1-9cb2-2f86b5fbd4b9. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 1182.315267] env[65726]: DEBUG oslo_concurrency.lockutils [req-667957f8-9b62-4db4-b93b-1eda3adbf91e req-6b7ce011-e8fa-4475-8f71-30734af92d19 service nova] Acquiring lock "refresh_cache-55ee97c3-b014-48eb-b41b-fc5aa16bc09e" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1182.369214] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116878, 'name': CreateVM_Task, 'duration_secs': 0.511574} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1182.369559] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: dc112e67-7c2d-4081-9a53-e4f43f61dcd6] Created VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1182.370653] env[65726]: WARNING openstack [None req-f04e60e2-18c2-41e8-8f5e-88ce749ed252 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1182.370786] env[65726]: WARNING openstack [None req-f04e60e2-18c2-41e8-8f5e-88ce749ed252 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1182.377102] env[65726]: DEBUG oslo_concurrency.lockutils [None req-f04e60e2-18c2-41e8-8f5e-88ce749ed252 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1182.377257] env[65726]: DEBUG oslo_concurrency.lockutils [None req-f04e60e2-18c2-41e8-8f5e-88ce749ed252 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1182.377608] env[65726]: DEBUG oslo_concurrency.lockutils [None req-f04e60e2-18c2-41e8-8f5e-88ce749ed252 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1182.377964] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1e863d22-0bff-4f9c-a7fa-0b7521659a5c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1182.387343] env[65726]: DEBUG oslo_vmware.api [None req-f04e60e2-18c2-41e8-8f5e-88ce749ed252 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Waiting for the task: (returnval){ [ 1182.387343] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]527ba114-855f-484a-8b20-15d620f0067c" [ 1182.387343] env[65726]: _type = "Task" [ 1182.387343] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1182.396887] env[65726]: DEBUG oslo_vmware.api [None req-f04e60e2-18c2-41e8-8f5e-88ce749ed252 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]527ba114-855f-484a-8b20-15d620f0067c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1182.431395] env[65726]: DEBUG oslo_vmware.api [None req-a3f83649-0195-4086-b546-9898d337f536 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Task: {'id': task-5116881, 'name': Destroy_Task} progress is 100%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1182.435315] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116882, 'name': CreateVM_Task, 'duration_secs': 0.44292} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1182.435565] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e3255f3b-028f-4a0b-b621-873730417936] Created VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1182.436475] env[65726]: WARNING openstack [None req-d8745e4a-3a23-4daa-80f7-d677b94a15ab tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1182.436903] env[65726]: WARNING openstack [None req-d8745e4a-3a23-4daa-80f7-d677b94a15ab tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1182.442154] env[65726]: DEBUG oslo_concurrency.lockutils [None req-d8745e4a-3a23-4daa-80f7-d677b94a15ab tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1182.669406] env[65726]: DEBUG oslo_concurrency.lockutils [None req-ea123e2e-31cf-42d0-af88-34d2c73324cf tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Lock "7c1188ce-8718-4719-8631-e59e7915b7aa" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 7.307s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1182.721072] env[65726]: WARNING openstack [None req-3b1f8e43-10c3-447e-a11b-a1499eca83f6 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1182.721072] env[65726]: WARNING openstack [None req-3b1f8e43-10c3-447e-a11b-a1499eca83f6 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1182.726379] env[65726]: DEBUG nova.network.neutron [None req-3b1f8e43-10c3-447e-a11b-a1499eca83f6 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 55ee97c3-b014-48eb-b41b-fc5aa16bc09e] Instance cache missing network info. {{(pid=65726) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 1182.738753] env[65726]: DEBUG oslo_concurrency.lockutils [None req-f3e961d9-f8e2-4b47-8ba1-7abd521b41a7 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Acquiring lock "7c1188ce-8718-4719-8631-e59e7915b7aa" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1182.739135] env[65726]: DEBUG oslo_concurrency.lockutils [None req-f3e961d9-f8e2-4b47-8ba1-7abd521b41a7 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Lock "7c1188ce-8718-4719-8631-e59e7915b7aa" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1182.739443] env[65726]: DEBUG oslo_concurrency.lockutils [None req-f3e961d9-f8e2-4b47-8ba1-7abd521b41a7 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Acquiring lock "7c1188ce-8718-4719-8631-e59e7915b7aa-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1182.739802] env[65726]: DEBUG oslo_concurrency.lockutils [None req-f3e961d9-f8e2-4b47-8ba1-7abd521b41a7 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Lock "7c1188ce-8718-4719-8631-e59e7915b7aa-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1182.739967] env[65726]: DEBUG oslo_concurrency.lockutils [None req-f3e961d9-f8e2-4b47-8ba1-7abd521b41a7 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Lock "7c1188ce-8718-4719-8631-e59e7915b7aa-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1182.743063] env[65726]: INFO nova.compute.manager [None req-f3e961d9-f8e2-4b47-8ba1-7abd521b41a7 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 7c1188ce-8718-4719-8631-e59e7915b7aa] Terminating instance [ 1182.784143] env[65726]: WARNING openstack [None req-3b1f8e43-10c3-447e-a11b-a1499eca83f6 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1182.784551] env[65726]: WARNING openstack [None req-3b1f8e43-10c3-447e-a11b-a1499eca83f6 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1182.873355] env[65726]: WARNING openstack [None req-3b1f8e43-10c3-447e-a11b-a1499eca83f6 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1182.873667] env[65726]: WARNING openstack [None req-3b1f8e43-10c3-447e-a11b-a1499eca83f6 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1182.901250] env[65726]: DEBUG oslo_vmware.api [None req-f04e60e2-18c2-41e8-8f5e-88ce749ed252 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]527ba114-855f-484a-8b20-15d620f0067c, 'name': SearchDatastore_Task, 'duration_secs': 0.016732} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1182.901673] env[65726]: DEBUG oslo_concurrency.lockutils [None req-f04e60e2-18c2-41e8-8f5e-88ce749ed252 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1182.901926] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-f04e60e2-18c2-41e8-8f5e-88ce749ed252 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: dc112e67-7c2d-4081-9a53-e4f43f61dcd6] Processing image b52362a3-ee8a-4cbf-b06f-513b0cc8f95c {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1182.902202] env[65726]: DEBUG oslo_concurrency.lockutils [None req-f04e60e2-18c2-41e8-8f5e-88ce749ed252 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1182.902425] env[65726]: DEBUG oslo_concurrency.lockutils [None req-f04e60e2-18c2-41e8-8f5e-88ce749ed252 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1182.902691] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-f04e60e2-18c2-41e8-8f5e-88ce749ed252 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1182.902918] env[65726]: DEBUG oslo_concurrency.lockutils [None req-d8745e4a-3a23-4daa-80f7-d677b94a15ab tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1182.903333] env[65726]: DEBUG oslo_concurrency.lockutils [None req-d8745e4a-3a23-4daa-80f7-d677b94a15ab tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1182.903593] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b5ac78ed-1bc0-41ea-9742-993df78c283f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1182.906118] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f05ebb6b-6a7c-4923-8849-befee376342f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1182.912268] env[65726]: DEBUG oslo_vmware.api [None req-d8745e4a-3a23-4daa-80f7-d677b94a15ab tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Waiting for the task: (returnval){ [ 1182.912268] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5214c1eb-abbf-6767-cb80-1e29d1631e9e" [ 1182.912268] env[65726]: _type = "Task" [ 1182.912268] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1182.920499] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-f04e60e2-18c2-41e8-8f5e-88ce749ed252 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1182.920695] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-f04e60e2-18c2-41e8-8f5e-88ce749ed252 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65726) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1182.929805] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f7057f5a-a902-47e7-adf2-1a7b3e87bc70 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1182.932607] env[65726]: DEBUG oslo_vmware.api [None req-d8745e4a-3a23-4daa-80f7-d677b94a15ab tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5214c1eb-abbf-6767-cb80-1e29d1631e9e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1182.936049] env[65726]: DEBUG oslo_vmware.api [None req-a3f83649-0195-4086-b546-9898d337f536 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Task: {'id': task-5116881, 'name': Destroy_Task, 'duration_secs': 0.563198} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1182.936671] env[65726]: INFO nova.virt.vmwareapi.vm_util [None req-a3f83649-0195-4086-b546-9898d337f536 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: b353b5e9-500c-42d3-a87f-880a624febec] Destroyed the VM [ 1182.936912] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-a3f83649-0195-4086-b546-9898d337f536 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: b353b5e9-500c-42d3-a87f-880a624febec] Deleting Snapshot of the VM instance {{(pid=65726) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1182.937192] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-b9607a68-2944-4200-a6e8-8af2377d5501 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1182.940286] env[65726]: DEBUG oslo_vmware.api [None req-f04e60e2-18c2-41e8-8f5e-88ce749ed252 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Waiting for the task: (returnval){ [ 1182.940286] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]521b9862-320f-cb8d-c817-8aec1bb81588" [ 1182.940286] env[65726]: _type = "Task" [ 1182.940286] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1182.947050] env[65726]: DEBUG oslo_vmware.api [None req-a3f83649-0195-4086-b546-9898d337f536 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Waiting for the task: (returnval){ [ 1182.947050] env[65726]: value = "task-5116883" [ 1182.947050] env[65726]: _type = "Task" [ 1182.947050] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1182.950829] env[65726]: DEBUG oslo_vmware.api [None req-f04e60e2-18c2-41e8-8f5e-88ce749ed252 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]521b9862-320f-cb8d-c817-8aec1bb81588, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1182.961035] env[65726]: DEBUG oslo_vmware.api [None req-a3f83649-0195-4086-b546-9898d337f536 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Task: {'id': task-5116883, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1182.978020] env[65726]: DEBUG nova.network.neutron [None req-3b1f8e43-10c3-447e-a11b-a1499eca83f6 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 55ee97c3-b014-48eb-b41b-fc5aa16bc09e] Updating instance_info_cache with network_info: [{"id": "65a8678e-a808-4eb1-9cb2-2f86b5fbd4b9", "address": "fa:16:3e:ba:7b:9c", "network": {"id": "336fe51f-971a-447d-8e68-505640e0db1b", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1030314703-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6b56f93ba0e14521921484ac2c785c43", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8ee1c76d-1a61-4546-85cb-d4bd3c1b35ef", "external-id": "nsx-vlan-transportzone-161", "segmentation_id": 161, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap65a8678e-a8", "ovs_interfaceid": "65a8678e-a808-4eb1-9cb2-2f86b5fbd4b9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1183.165214] env[65726]: DEBUG oslo_vmware.rw_handles [None req-b7bd70fc-78fa-4df3-bb13-d10a1172d2a5 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52b93164-06a5-d34b-5e6c-a27fd01cc0a1/disk-0.vmdk. {{(pid=65726) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1183.166396] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d052ae91-278a-4b2f-ae15-dfcdcdf29c02 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.174170] env[65726]: DEBUG oslo_vmware.rw_handles [None req-b7bd70fc-78fa-4df3-bb13-d10a1172d2a5 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52b93164-06a5-d34b-5e6c-a27fd01cc0a1/disk-0.vmdk is in state: ready. {{(pid=65726) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1183.174419] env[65726]: ERROR oslo_vmware.rw_handles [None req-b7bd70fc-78fa-4df3-bb13-d10a1172d2a5 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52b93164-06a5-d34b-5e6c-a27fd01cc0a1/disk-0.vmdk due to incomplete transfer. [ 1183.174796] env[65726]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-6f66bc81-b909-4d0b-bcce-e8a95a11eb31 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.184047] env[65726]: DEBUG oslo_vmware.rw_handles [None req-b7bd70fc-78fa-4df3-bb13-d10a1172d2a5 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52b93164-06a5-d34b-5e6c-a27fd01cc0a1/disk-0.vmdk. {{(pid=65726) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1183.184297] env[65726]: DEBUG nova.virt.vmwareapi.images [None req-b7bd70fc-78fa-4df3-bb13-d10a1172d2a5 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 378c1d18-4bb6-4245-80bf-370fb7af0575] Uploaded image 1d5801e6-cfd0-4c6a-8fa0-57ee72cc963a to the Glance image server {{(pid=65726) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 1183.186859] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-b7bd70fc-78fa-4df3-bb13-d10a1172d2a5 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 378c1d18-4bb6-4245-80bf-370fb7af0575] Destroying the VM {{(pid=65726) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 1183.187189] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-82b10e97-595b-4222-848f-e91a4f208e4b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.195164] env[65726]: DEBUG oslo_vmware.api [None req-b7bd70fc-78fa-4df3-bb13-d10a1172d2a5 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Waiting for the task: (returnval){ [ 1183.195164] env[65726]: value = "task-5116884" [ 1183.195164] env[65726]: _type = "Task" [ 1183.195164] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1183.205044] env[65726]: DEBUG oslo_vmware.api [None req-b7bd70fc-78fa-4df3-bb13-d10a1172d2a5 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5116884, 'name': Destroy_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1183.248226] env[65726]: DEBUG nova.compute.manager [None req-f3e961d9-f8e2-4b47-8ba1-7abd521b41a7 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 7c1188ce-8718-4719-8631-e59e7915b7aa] Start destroying the instance on the hypervisor. {{(pid=65726) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1183.248472] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-f3e961d9-f8e2-4b47-8ba1-7abd521b41a7 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 7c1188ce-8718-4719-8631-e59e7915b7aa] Destroying instance {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1183.250054] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03b72172-d22b-4d6b-a059-e5d84b7ed2d8 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.260087] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-f3e961d9-f8e2-4b47-8ba1-7abd521b41a7 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 7c1188ce-8718-4719-8631-e59e7915b7aa] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1183.260409] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-aa2b39d1-cea2-401e-8b4c-c0e2f68d0b1e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.269180] env[65726]: DEBUG oslo_vmware.api [None req-f3e961d9-f8e2-4b47-8ba1-7abd521b41a7 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Waiting for the task: (returnval){ [ 1183.269180] env[65726]: value = "task-5116885" [ 1183.269180] env[65726]: _type = "Task" [ 1183.269180] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1183.280569] env[65726]: DEBUG oslo_vmware.api [None req-f3e961d9-f8e2-4b47-8ba1-7abd521b41a7 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': task-5116885, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1183.425852] env[65726]: DEBUG oslo_vmware.api [None req-d8745e4a-3a23-4daa-80f7-d677b94a15ab tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5214c1eb-abbf-6767-cb80-1e29d1631e9e, 'name': SearchDatastore_Task, 'duration_secs': 0.019968} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1183.426337] env[65726]: DEBUG oslo_concurrency.lockutils [None req-d8745e4a-3a23-4daa-80f7-d677b94a15ab tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1183.426684] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-d8745e4a-3a23-4daa-80f7-d677b94a15ab tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: e3255f3b-028f-4a0b-b621-873730417936] Processing image b52362a3-ee8a-4cbf-b06f-513b0cc8f95c {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1183.427075] env[65726]: DEBUG oslo_concurrency.lockutils [None req-d8745e4a-3a23-4daa-80f7-d677b94a15ab tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1183.453982] env[65726]: DEBUG oslo_vmware.api [None req-f04e60e2-18c2-41e8-8f5e-88ce749ed252 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]521b9862-320f-cb8d-c817-8aec1bb81588, 'name': SearchDatastore_Task, 'duration_secs': 0.011703} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1183.458117] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-973076b8-e63e-46fb-a3e7-32d5dbc2a6fc {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.464262] env[65726]: DEBUG oslo_vmware.api [None req-f04e60e2-18c2-41e8-8f5e-88ce749ed252 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Waiting for the task: (returnval){ [ 1183.464262] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]525f8b18-d624-6ca5-d19f-d3b67739dfd1" [ 1183.464262] env[65726]: _type = "Task" [ 1183.464262] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1183.467795] env[65726]: DEBUG oslo_vmware.api [None req-a3f83649-0195-4086-b546-9898d337f536 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Task: {'id': task-5116883, 'name': RemoveSnapshot_Task} progress is 100%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1183.477255] env[65726]: DEBUG oslo_vmware.api [None req-f04e60e2-18c2-41e8-8f5e-88ce749ed252 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]525f8b18-d624-6ca5-d19f-d3b67739dfd1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1183.479917] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3b1f8e43-10c3-447e-a11b-a1499eca83f6 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Releasing lock "refresh_cache-55ee97c3-b014-48eb-b41b-fc5aa16bc09e" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1183.480511] env[65726]: DEBUG nova.compute.manager [None req-3b1f8e43-10c3-447e-a11b-a1499eca83f6 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 55ee97c3-b014-48eb-b41b-fc5aa16bc09e] Instance network_info: |[{"id": "65a8678e-a808-4eb1-9cb2-2f86b5fbd4b9", "address": "fa:16:3e:ba:7b:9c", "network": {"id": "336fe51f-971a-447d-8e68-505640e0db1b", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1030314703-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6b56f93ba0e14521921484ac2c785c43", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8ee1c76d-1a61-4546-85cb-d4bd3c1b35ef", "external-id": "nsx-vlan-transportzone-161", "segmentation_id": 161, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap65a8678e-a8", "ovs_interfaceid": "65a8678e-a808-4eb1-9cb2-2f86b5fbd4b9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1183.480780] env[65726]: DEBUG oslo_concurrency.lockutils [req-667957f8-9b62-4db4-b93b-1eda3adbf91e req-6b7ce011-e8fa-4475-8f71-30734af92d19 service nova] Acquired lock "refresh_cache-55ee97c3-b014-48eb-b41b-fc5aa16bc09e" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1183.481040] env[65726]: DEBUG nova.network.neutron [req-667957f8-9b62-4db4-b93b-1eda3adbf91e req-6b7ce011-e8fa-4475-8f71-30734af92d19 service nova] [instance: 55ee97c3-b014-48eb-b41b-fc5aa16bc09e] Refreshing network info cache for port 65a8678e-a808-4eb1-9cb2-2f86b5fbd4b9 {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 1183.482595] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-3b1f8e43-10c3-447e-a11b-a1499eca83f6 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 55ee97c3-b014-48eb-b41b-fc5aa16bc09e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ba:7b:9c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '8ee1c76d-1a61-4546-85cb-d4bd3c1b35ef', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '65a8678e-a808-4eb1-9cb2-2f86b5fbd4b9', 'vif_model': 'vmxnet3'}] {{(pid=65726) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1183.490937] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-3b1f8e43-10c3-447e-a11b-a1499eca83f6 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Creating folder: Project (6b56f93ba0e14521921484ac2c785c43). Parent ref: group-v995008. {{(pid=65726) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1183.491634] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-371a8993-a49e-464d-afd6-1185130cdd20 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.503595] env[65726]: INFO nova.virt.vmwareapi.vm_util [None req-3b1f8e43-10c3-447e-a11b-a1499eca83f6 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Created folder: Project (6b56f93ba0e14521921484ac2c785c43) in parent group-v995008. [ 1183.503898] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-3b1f8e43-10c3-447e-a11b-a1499eca83f6 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Creating folder: Instances. Parent ref: group-v995324. {{(pid=65726) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1183.504295] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-76d5fd9b-a452-40a0-b83f-0438af1bf2b2 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.518770] env[65726]: INFO nova.virt.vmwareapi.vm_util [None req-3b1f8e43-10c3-447e-a11b-a1499eca83f6 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Created folder: Instances in parent group-v995324. [ 1183.519120] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-3b1f8e43-10c3-447e-a11b-a1499eca83f6 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1183.519377] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 55ee97c3-b014-48eb-b41b-fc5aa16bc09e] Creating VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1183.519642] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c887298e-8965-47ac-a28e-1d184b594cc3 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.543110] env[65726]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1183.543110] env[65726]: value = "task-5116888" [ 1183.543110] env[65726]: _type = "Task" [ 1183.543110] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1183.552263] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116888, 'name': CreateVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1183.640899] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b0eb83ae-2fbb-47ee-abe7-62f2ae1b54da tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Acquiring lock "6956bda7-5657-45d7-8f80-c6809bd836b7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1183.641167] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b0eb83ae-2fbb-47ee-abe7-62f2ae1b54da tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Lock "6956bda7-5657-45d7-8f80-c6809bd836b7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1183.707067] env[65726]: DEBUG oslo_vmware.api [None req-b7bd70fc-78fa-4df3-bb13-d10a1172d2a5 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5116884, 'name': Destroy_Task, 'duration_secs': 0.45028} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1183.707067] env[65726]: INFO nova.virt.vmwareapi.vm_util [None req-b7bd70fc-78fa-4df3-bb13-d10a1172d2a5 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 378c1d18-4bb6-4245-80bf-370fb7af0575] Destroyed the VM [ 1183.707293] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-b7bd70fc-78fa-4df3-bb13-d10a1172d2a5 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 378c1d18-4bb6-4245-80bf-370fb7af0575] Deleting Snapshot of the VM instance {{(pid=65726) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1183.707445] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-ec264006-28f1-4228-bf06-5525f79dce4a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.715360] env[65726]: DEBUG oslo_vmware.api [None req-b7bd70fc-78fa-4df3-bb13-d10a1172d2a5 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Waiting for the task: (returnval){ [ 1183.715360] env[65726]: value = "task-5116889" [ 1183.715360] env[65726]: _type = "Task" [ 1183.715360] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1183.724801] env[65726]: DEBUG oslo_vmware.api [None req-b7bd70fc-78fa-4df3-bb13-d10a1172d2a5 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5116889, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1183.782395] env[65726]: DEBUG oslo_vmware.api [None req-f3e961d9-f8e2-4b47-8ba1-7abd521b41a7 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': task-5116885, 'name': PowerOffVM_Task, 'duration_secs': 0.237306} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1183.782667] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-f3e961d9-f8e2-4b47-8ba1-7abd521b41a7 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 7c1188ce-8718-4719-8631-e59e7915b7aa] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1183.782828] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-f3e961d9-f8e2-4b47-8ba1-7abd521b41a7 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 7c1188ce-8718-4719-8631-e59e7915b7aa] Unregistering the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1183.783550] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-35c588ff-c1d7-4fbc-b1d0-68e92645da71 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.857964] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-f3e961d9-f8e2-4b47-8ba1-7abd521b41a7 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 7c1188ce-8718-4719-8631-e59e7915b7aa] Unregistered the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1183.858315] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-f3e961d9-f8e2-4b47-8ba1-7abd521b41a7 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 7c1188ce-8718-4719-8631-e59e7915b7aa] Deleting contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1183.858499] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-f3e961d9-f8e2-4b47-8ba1-7abd521b41a7 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Deleting the datastore file [datastore1] 7c1188ce-8718-4719-8631-e59e7915b7aa {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1183.858814] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-af03dd71-fb6f-4200-ae93-b7a818ed3c0f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.866552] env[65726]: DEBUG oslo_vmware.api [None req-f3e961d9-f8e2-4b47-8ba1-7abd521b41a7 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Waiting for the task: (returnval){ [ 1183.866552] env[65726]: value = "task-5116892" [ 1183.866552] env[65726]: _type = "Task" [ 1183.866552] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1183.875418] env[65726]: DEBUG oslo_vmware.api [None req-f3e961d9-f8e2-4b47-8ba1-7abd521b41a7 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': task-5116892, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1183.963487] env[65726]: DEBUG oslo_vmware.api [None req-a3f83649-0195-4086-b546-9898d337f536 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Task: {'id': task-5116883, 'name': RemoveSnapshot_Task, 'duration_secs': 0.525467} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1183.963801] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-a3f83649-0195-4086-b546-9898d337f536 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: b353b5e9-500c-42d3-a87f-880a624febec] Deleted Snapshot of the VM instance {{(pid=65726) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1183.964096] env[65726]: DEBUG nova.compute.manager [None req-a3f83649-0195-4086-b546-9898d337f536 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: b353b5e9-500c-42d3-a87f-880a624febec] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1183.964881] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15a77cb7-dace-4307-a2fb-aad37278546e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.980711] env[65726]: DEBUG oslo_vmware.api [None req-f04e60e2-18c2-41e8-8f5e-88ce749ed252 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]525f8b18-d624-6ca5-d19f-d3b67739dfd1, 'name': SearchDatastore_Task, 'duration_secs': 0.026149} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1183.981103] env[65726]: DEBUG oslo_concurrency.lockutils [None req-f04e60e2-18c2-41e8-8f5e-88ce749ed252 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1183.981255] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-f04e60e2-18c2-41e8-8f5e-88ce749ed252 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore2] dc112e67-7c2d-4081-9a53-e4f43f61dcd6/dc112e67-7c2d-4081-9a53-e4f43f61dcd6.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1183.981539] env[65726]: DEBUG oslo_concurrency.lockutils [None req-d8745e4a-3a23-4daa-80f7-d677b94a15ab tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1183.981761] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-d8745e4a-3a23-4daa-80f7-d677b94a15ab tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1183.982105] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b79c76b3-81a1-4c36-b1c8-3aeb6ecb7e69 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.984585] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-295f6e06-3fd8-42c6-b8c9-58fb55d756ab {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.993562] env[65726]: WARNING openstack [req-667957f8-9b62-4db4-b93b-1eda3adbf91e req-6b7ce011-e8fa-4475-8f71-30734af92d19 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1183.993895] env[65726]: WARNING openstack [req-667957f8-9b62-4db4-b93b-1eda3adbf91e req-6b7ce011-e8fa-4475-8f71-30734af92d19 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1184.000387] env[65726]: DEBUG oslo_vmware.api [None req-f04e60e2-18c2-41e8-8f5e-88ce749ed252 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Waiting for the task: (returnval){ [ 1184.000387] env[65726]: value = "task-5116893" [ 1184.000387] env[65726]: _type = "Task" [ 1184.000387] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1184.002020] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-d8745e4a-3a23-4daa-80f7-d677b94a15ab tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1184.002203] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-d8745e4a-3a23-4daa-80f7-d677b94a15ab tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65726) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1184.005829] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c9ed65e1-10e0-45c2-ad35-322478f3f407 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1184.015846] env[65726]: DEBUG oslo_vmware.api [None req-f04e60e2-18c2-41e8-8f5e-88ce749ed252 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Task: {'id': task-5116893, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1184.016232] env[65726]: DEBUG oslo_vmware.api [None req-d8745e4a-3a23-4daa-80f7-d677b94a15ab tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Waiting for the task: (returnval){ [ 1184.016232] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52350912-cb00-78de-4652-fa8c626e649a" [ 1184.016232] env[65726]: _type = "Task" [ 1184.016232] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1184.027144] env[65726]: DEBUG oslo_vmware.api [None req-d8745e4a-3a23-4daa-80f7-d677b94a15ab tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52350912-cb00-78de-4652-fa8c626e649a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1184.053137] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116888, 'name': CreateVM_Task} progress is 99%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1184.107849] env[65726]: WARNING openstack [req-667957f8-9b62-4db4-b93b-1eda3adbf91e req-6b7ce011-e8fa-4475-8f71-30734af92d19 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1184.108244] env[65726]: WARNING openstack [req-667957f8-9b62-4db4-b93b-1eda3adbf91e req-6b7ce011-e8fa-4475-8f71-30734af92d19 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1184.143472] env[65726]: DEBUG nova.compute.manager [None req-b0eb83ae-2fbb-47ee-abe7-62f2ae1b54da tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 6956bda7-5657-45d7-8f80-c6809bd836b7] Starting instance... {{(pid=65726) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 1184.173593] env[65726]: WARNING openstack [req-667957f8-9b62-4db4-b93b-1eda3adbf91e req-6b7ce011-e8fa-4475-8f71-30734af92d19 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1184.174042] env[65726]: WARNING openstack [req-667957f8-9b62-4db4-b93b-1eda3adbf91e req-6b7ce011-e8fa-4475-8f71-30734af92d19 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1184.230543] env[65726]: DEBUG oslo_vmware.api [None req-b7bd70fc-78fa-4df3-bb13-d10a1172d2a5 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5116889, 'name': RemoveSnapshot_Task, 'duration_secs': 0.408603} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1184.230795] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-b7bd70fc-78fa-4df3-bb13-d10a1172d2a5 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 378c1d18-4bb6-4245-80bf-370fb7af0575] Deleted Snapshot of the VM instance {{(pid=65726) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1184.231133] env[65726]: DEBUG nova.compute.manager [None req-b7bd70fc-78fa-4df3-bb13-d10a1172d2a5 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 378c1d18-4bb6-4245-80bf-370fb7af0575] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1184.231961] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4fd187e-8463-4d43-b2a3-a02e899c642c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1184.257216] env[65726]: DEBUG nova.network.neutron [req-667957f8-9b62-4db4-b93b-1eda3adbf91e req-6b7ce011-e8fa-4475-8f71-30734af92d19 service nova] [instance: 55ee97c3-b014-48eb-b41b-fc5aa16bc09e] Updated VIF entry in instance network info cache for port 65a8678e-a808-4eb1-9cb2-2f86b5fbd4b9. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 1184.257666] env[65726]: DEBUG nova.network.neutron [req-667957f8-9b62-4db4-b93b-1eda3adbf91e req-6b7ce011-e8fa-4475-8f71-30734af92d19 service nova] [instance: 55ee97c3-b014-48eb-b41b-fc5aa16bc09e] Updating instance_info_cache with network_info: [{"id": "65a8678e-a808-4eb1-9cb2-2f86b5fbd4b9", "address": "fa:16:3e:ba:7b:9c", "network": {"id": "336fe51f-971a-447d-8e68-505640e0db1b", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1030314703-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6b56f93ba0e14521921484ac2c785c43", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8ee1c76d-1a61-4546-85cb-d4bd3c1b35ef", "external-id": "nsx-vlan-transportzone-161", "segmentation_id": 161, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap65a8678e-a8", "ovs_interfaceid": "65a8678e-a808-4eb1-9cb2-2f86b5fbd4b9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1184.377363] env[65726]: DEBUG oslo_vmware.api [None req-f3e961d9-f8e2-4b47-8ba1-7abd521b41a7 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': task-5116892, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.198237} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1184.377595] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-f3e961d9-f8e2-4b47-8ba1-7abd521b41a7 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Deleted the datastore file {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1184.377757] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-f3e961d9-f8e2-4b47-8ba1-7abd521b41a7 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 7c1188ce-8718-4719-8631-e59e7915b7aa] Deleted contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1184.377911] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-f3e961d9-f8e2-4b47-8ba1-7abd521b41a7 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 7c1188ce-8718-4719-8631-e59e7915b7aa] Instance destroyed {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1184.378075] env[65726]: INFO nova.compute.manager [None req-f3e961d9-f8e2-4b47-8ba1-7abd521b41a7 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 7c1188ce-8718-4719-8631-e59e7915b7aa] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1184.378335] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-f3e961d9-f8e2-4b47-8ba1-7abd521b41a7 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1184.378534] env[65726]: DEBUG nova.compute.manager [-] [instance: 7c1188ce-8718-4719-8631-e59e7915b7aa] Deallocating network for instance {{(pid=65726) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1184.378630] env[65726]: DEBUG nova.network.neutron [-] [instance: 7c1188ce-8718-4719-8631-e59e7915b7aa] deallocate_for_instance() {{(pid=65726) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 1184.379186] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1184.379442] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1184.423757] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1184.424097] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1184.479861] env[65726]: INFO nova.compute.manager [None req-a3f83649-0195-4086-b546-9898d337f536 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: b353b5e9-500c-42d3-a87f-880a624febec] Shelve offloading [ 1184.514104] env[65726]: DEBUG oslo_vmware.api [None req-f04e60e2-18c2-41e8-8f5e-88ce749ed252 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Task: {'id': task-5116893, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1184.527459] env[65726]: DEBUG oslo_vmware.api [None req-d8745e4a-3a23-4daa-80f7-d677b94a15ab tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52350912-cb00-78de-4652-fa8c626e649a, 'name': SearchDatastore_Task, 'duration_secs': 0.021591} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1184.527794] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f6f98bfa-e20d-412d-8b67-e1fc1687d52e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1184.534713] env[65726]: DEBUG oslo_vmware.api [None req-d8745e4a-3a23-4daa-80f7-d677b94a15ab tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Waiting for the task: (returnval){ [ 1184.534713] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52cf7cef-f06f-fbd2-036c-e21f904ebd6f" [ 1184.534713] env[65726]: _type = "Task" [ 1184.534713] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1184.544643] env[65726]: DEBUG oslo_vmware.api [None req-d8745e4a-3a23-4daa-80f7-d677b94a15ab tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52cf7cef-f06f-fbd2-036c-e21f904ebd6f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1184.552695] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116888, 'name': CreateVM_Task} progress is 99%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1184.667345] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b0eb83ae-2fbb-47ee-abe7-62f2ae1b54da tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1184.667641] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b0eb83ae-2fbb-47ee-abe7-62f2ae1b54da tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1184.669278] env[65726]: INFO nova.compute.claims [None req-b0eb83ae-2fbb-47ee-abe7-62f2ae1b54da tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 6956bda7-5657-45d7-8f80-c6809bd836b7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1184.674126] env[65726]: DEBUG nova.compute.manager [req-b72fdf18-4402-47a9-a7a5-51fd1da2bb55 req-14e06bc2-c2c4-424c-b62f-9d5a4301ec62 service nova] [instance: 7c1188ce-8718-4719-8631-e59e7915b7aa] Received event network-vif-deleted-e8f51770-ca83-4aa3-87c1-ef47653d9bb2 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1184.674329] env[65726]: INFO nova.compute.manager [req-b72fdf18-4402-47a9-a7a5-51fd1da2bb55 req-14e06bc2-c2c4-424c-b62f-9d5a4301ec62 service nova] [instance: 7c1188ce-8718-4719-8631-e59e7915b7aa] Neutron deleted interface e8f51770-ca83-4aa3-87c1-ef47653d9bb2; detaching it from the instance and deleting it from the info cache [ 1184.674515] env[65726]: DEBUG nova.network.neutron [req-b72fdf18-4402-47a9-a7a5-51fd1da2bb55 req-14e06bc2-c2c4-424c-b62f-9d5a4301ec62 service nova] [instance: 7c1188ce-8718-4719-8631-e59e7915b7aa] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1184.745070] env[65726]: INFO nova.compute.manager [None req-b7bd70fc-78fa-4df3-bb13-d10a1172d2a5 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 378c1d18-4bb6-4245-80bf-370fb7af0575] Shelve offloading [ 1184.760639] env[65726]: DEBUG oslo_concurrency.lockutils [req-667957f8-9b62-4db4-b93b-1eda3adbf91e req-6b7ce011-e8fa-4475-8f71-30734af92d19 service nova] Releasing lock "refresh_cache-55ee97c3-b014-48eb-b41b-fc5aa16bc09e" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1184.984447] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-a3f83649-0195-4086-b546-9898d337f536 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: b353b5e9-500c-42d3-a87f-880a624febec] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1184.984447] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-67fd3624-a66f-4a84-a9d7-00ba795eff7e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1184.993239] env[65726]: DEBUG oslo_vmware.api [None req-a3f83649-0195-4086-b546-9898d337f536 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Waiting for the task: (returnval){ [ 1184.993239] env[65726]: value = "task-5116894" [ 1184.993239] env[65726]: _type = "Task" [ 1184.993239] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1185.011380] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-a3f83649-0195-4086-b546-9898d337f536 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: b353b5e9-500c-42d3-a87f-880a624febec] VM already powered off {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1185.011789] env[65726]: DEBUG nova.compute.manager [None req-a3f83649-0195-4086-b546-9898d337f536 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: b353b5e9-500c-42d3-a87f-880a624febec] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1185.013325] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1b8d0e8-5a1e-4550-ae50-be96cada7b43 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.032441] env[65726]: DEBUG oslo_vmware.api [None req-f04e60e2-18c2-41e8-8f5e-88ce749ed252 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Task: {'id': task-5116893, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1185.035901] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a3f83649-0195-4086-b546-9898d337f536 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Acquiring lock "refresh_cache-b353b5e9-500c-42d3-a87f-880a624febec" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1185.036255] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a3f83649-0195-4086-b546-9898d337f536 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Acquired lock "refresh_cache-b353b5e9-500c-42d3-a87f-880a624febec" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1185.036914] env[65726]: DEBUG nova.network.neutron [None req-a3f83649-0195-4086-b546-9898d337f536 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: b353b5e9-500c-42d3-a87f-880a624febec] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1185.057277] env[65726]: DEBUG oslo_vmware.api [None req-d8745e4a-3a23-4daa-80f7-d677b94a15ab tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52cf7cef-f06f-fbd2-036c-e21f904ebd6f, 'name': SearchDatastore_Task, 'duration_secs': 0.024224} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1185.058440] env[65726]: DEBUG oslo_concurrency.lockutils [None req-d8745e4a-3a23-4daa-80f7-d677b94a15ab tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1185.058886] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-d8745e4a-3a23-4daa-80f7-d677b94a15ab tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore2] e3255f3b-028f-4a0b-b621-873730417936/e3255f3b-028f-4a0b-b621-873730417936.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1185.059821] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-df92b6b6-7167-4015-9583-3afa9377c44b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.072722] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116888, 'name': CreateVM_Task, 'duration_secs': 1.429976} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1185.073791] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 55ee97c3-b014-48eb-b41b-fc5aa16bc09e] Created VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1185.075204] env[65726]: WARNING openstack [None req-3b1f8e43-10c3-447e-a11b-a1499eca83f6 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1185.075857] env[65726]: WARNING openstack [None req-3b1f8e43-10c3-447e-a11b-a1499eca83f6 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1185.084309] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3b1f8e43-10c3-447e-a11b-a1499eca83f6 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1185.084559] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3b1f8e43-10c3-447e-a11b-a1499eca83f6 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1185.084975] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3b1f8e43-10c3-447e-a11b-a1499eca83f6 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1185.087144] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c66d5f6a-7bd1-4ab4-80f0-b708ea53edf9 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.092147] env[65726]: DEBUG oslo_vmware.api [None req-d8745e4a-3a23-4daa-80f7-d677b94a15ab tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Waiting for the task: (returnval){ [ 1185.092147] env[65726]: value = "task-5116895" [ 1185.092147] env[65726]: _type = "Task" [ 1185.092147] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1185.095260] env[65726]: DEBUG oslo_vmware.api [None req-3b1f8e43-10c3-447e-a11b-a1499eca83f6 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Waiting for the task: (returnval){ [ 1185.095260] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]525055d7-2059-2903-e0c9-65f4594f0366" [ 1185.095260] env[65726]: _type = "Task" [ 1185.095260] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1185.103130] env[65726]: DEBUG oslo_vmware.api [None req-d8745e4a-3a23-4daa-80f7-d677b94a15ab tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': task-5116895, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1185.109662] env[65726]: DEBUG oslo_vmware.api [None req-3b1f8e43-10c3-447e-a11b-a1499eca83f6 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]525055d7-2059-2903-e0c9-65f4594f0366, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1185.147233] env[65726]: DEBUG nova.network.neutron [-] [instance: 7c1188ce-8718-4719-8631-e59e7915b7aa] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1185.177153] env[65726]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-96f04803-f7ef-4e0f-aa17-c9a7f1e0bbeb {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.187848] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e82586b-8767-4ed8-963e-e89e1f31286b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.224308] env[65726]: DEBUG nova.compute.manager [req-b72fdf18-4402-47a9-a7a5-51fd1da2bb55 req-14e06bc2-c2c4-424c-b62f-9d5a4301ec62 service nova] [instance: 7c1188ce-8718-4719-8631-e59e7915b7aa] Detach interface failed, port_id=e8f51770-ca83-4aa3-87c1-ef47653d9bb2, reason: Instance 7c1188ce-8718-4719-8631-e59e7915b7aa could not be found. {{(pid=65726) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11669}} [ 1185.248971] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-b7bd70fc-78fa-4df3-bb13-d10a1172d2a5 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 378c1d18-4bb6-4245-80bf-370fb7af0575] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1185.249328] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6a5a0644-f9e8-4b44-873a-4321b20521ee {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.258349] env[65726]: DEBUG oslo_vmware.api [None req-b7bd70fc-78fa-4df3-bb13-d10a1172d2a5 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Waiting for the task: (returnval){ [ 1185.258349] env[65726]: value = "task-5116896" [ 1185.258349] env[65726]: _type = "Task" [ 1185.258349] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1185.270811] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-b7bd70fc-78fa-4df3-bb13-d10a1172d2a5 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 378c1d18-4bb6-4245-80bf-370fb7af0575] VM already powered off {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1185.271033] env[65726]: DEBUG nova.compute.manager [None req-b7bd70fc-78fa-4df3-bb13-d10a1172d2a5 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 378c1d18-4bb6-4245-80bf-370fb7af0575] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1185.271846] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-090814f7-bbba-4960-ae4e-77622249ce70 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.280065] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b7bd70fc-78fa-4df3-bb13-d10a1172d2a5 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Acquiring lock "refresh_cache-378c1d18-4bb6-4245-80bf-370fb7af0575" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1185.280268] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b7bd70fc-78fa-4df3-bb13-d10a1172d2a5 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Acquired lock "refresh_cache-378c1d18-4bb6-4245-80bf-370fb7af0575" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1185.280430] env[65726]: DEBUG nova.network.neutron [None req-b7bd70fc-78fa-4df3-bb13-d10a1172d2a5 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 378c1d18-4bb6-4245-80bf-370fb7af0575] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1185.517989] env[65726]: DEBUG oslo_vmware.api [None req-f04e60e2-18c2-41e8-8f5e-88ce749ed252 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Task: {'id': task-5116893, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.374102} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1185.517989] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-f04e60e2-18c2-41e8-8f5e-88ce749ed252 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore2] dc112e67-7c2d-4081-9a53-e4f43f61dcd6/dc112e67-7c2d-4081-9a53-e4f43f61dcd6.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1185.518560] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-f04e60e2-18c2-41e8-8f5e-88ce749ed252 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: dc112e67-7c2d-4081-9a53-e4f43f61dcd6] Extending root virtual disk to 1048576 {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1185.518560] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-735bce2f-742a-409c-b5f3-c409957a9cc6 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.526175] env[65726]: DEBUG oslo_vmware.api [None req-f04e60e2-18c2-41e8-8f5e-88ce749ed252 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Waiting for the task: (returnval){ [ 1185.526175] env[65726]: value = "task-5116897" [ 1185.526175] env[65726]: _type = "Task" [ 1185.526175] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1185.537709] env[65726]: DEBUG oslo_vmware.api [None req-f04e60e2-18c2-41e8-8f5e-88ce749ed252 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Task: {'id': task-5116897, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1185.544807] env[65726]: WARNING openstack [None req-a3f83649-0195-4086-b546-9898d337f536 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1185.545229] env[65726]: WARNING openstack [None req-a3f83649-0195-4086-b546-9898d337f536 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1185.601027] env[65726]: DEBUG oslo_vmware.api [None req-d8745e4a-3a23-4daa-80f7-d677b94a15ab tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': task-5116895, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1185.610581] env[65726]: DEBUG oslo_vmware.api [None req-3b1f8e43-10c3-447e-a11b-a1499eca83f6 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]525055d7-2059-2903-e0c9-65f4594f0366, 'name': SearchDatastore_Task, 'duration_secs': 0.064787} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1185.610898] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3b1f8e43-10c3-447e-a11b-a1499eca83f6 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1185.611152] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-3b1f8e43-10c3-447e-a11b-a1499eca83f6 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 55ee97c3-b014-48eb-b41b-fc5aa16bc09e] Processing image b52362a3-ee8a-4cbf-b06f-513b0cc8f95c {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1185.611396] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3b1f8e43-10c3-447e-a11b-a1499eca83f6 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1185.611535] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3b1f8e43-10c3-447e-a11b-a1499eca83f6 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1185.611716] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-3b1f8e43-10c3-447e-a11b-a1499eca83f6 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1185.613059] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-434d3699-b406-4fce-b822-839ea5aa6eac {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.626923] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-3b1f8e43-10c3-447e-a11b-a1499eca83f6 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1185.627148] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-3b1f8e43-10c3-447e-a11b-a1499eca83f6 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65726) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1185.627962] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-75b32ce4-c9d5-4699-bdb9-21662b19eaba {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.635342] env[65726]: DEBUG oslo_vmware.api [None req-3b1f8e43-10c3-447e-a11b-a1499eca83f6 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Waiting for the task: (returnval){ [ 1185.635342] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52a7c5aa-85c3-1055-9f60-e7fa4c5d2dbb" [ 1185.635342] env[65726]: _type = "Task" [ 1185.635342] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1185.645677] env[65726]: DEBUG oslo_vmware.api [None req-3b1f8e43-10c3-447e-a11b-a1499eca83f6 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52a7c5aa-85c3-1055-9f60-e7fa4c5d2dbb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1185.650667] env[65726]: INFO nova.compute.manager [-] [instance: 7c1188ce-8718-4719-8631-e59e7915b7aa] Took 1.27 seconds to deallocate network for instance. [ 1185.700366] env[65726]: WARNING openstack [None req-a3f83649-0195-4086-b546-9898d337f536 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1185.700976] env[65726]: WARNING openstack [None req-a3f83649-0195-4086-b546-9898d337f536 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1185.783303] env[65726]: WARNING openstack [None req-b7bd70fc-78fa-4df3-bb13-d10a1172d2a5 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1185.783791] env[65726]: WARNING openstack [None req-b7bd70fc-78fa-4df3-bb13-d10a1172d2a5 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1185.794257] env[65726]: WARNING openstack [None req-a3f83649-0195-4086-b546-9898d337f536 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1185.794257] env[65726]: WARNING openstack [None req-a3f83649-0195-4086-b546-9898d337f536 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1185.871402] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b1d3950-bf14-42a2-abc2-11c3e574c7bd {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.879288] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c19c239-eb8c-4516-913b-61a6d1ba88bd {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.914161] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5027385-4649-4f24-84ff-fe1ef6a95725 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.918910] env[65726]: DEBUG nova.network.neutron [None req-a3f83649-0195-4086-b546-9898d337f536 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: b353b5e9-500c-42d3-a87f-880a624febec] Updating instance_info_cache with network_info: [{"id": "08b7016b-dc44-4206-a00c-5da943b82a38", "address": "fa:16:3e:c5:e4:14", "network": {"id": "eb3278b7-c190-441b-84d2-305f9b896382", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1754571022-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.144", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "96149159e18e44f9bf3453e67681f224", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f77ff7a1-209c-4f3f-b2a0-fd817741e739", "external-id": "nsx-vlan-transportzone-935", "segmentation_id": 935, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap08b7016b-dc", "ovs_interfaceid": "08b7016b-dc44-4206-a00c-5da943b82a38", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1185.922580] env[65726]: WARNING openstack [None req-b7bd70fc-78fa-4df3-bb13-d10a1172d2a5 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1185.922952] env[65726]: WARNING openstack [None req-b7bd70fc-78fa-4df3-bb13-d10a1172d2a5 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1185.931627] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a3f83649-0195-4086-b546-9898d337f536 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Releasing lock "refresh_cache-b353b5e9-500c-42d3-a87f-880a624febec" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1185.931627] env[65726]: WARNING openstack [None req-a3f83649-0195-4086-b546-9898d337f536 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1185.932034] env[65726]: WARNING openstack [None req-a3f83649-0195-4086-b546-9898d337f536 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1185.937044] env[65726]: WARNING openstack [None req-a3f83649-0195-4086-b546-9898d337f536 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1185.937394] env[65726]: WARNING openstack [None req-a3f83649-0195-4086-b546-9898d337f536 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1185.949259] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c558900-ff1d-4369-9444-adf22829e8a8 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.965771] env[65726]: DEBUG nova.compute.provider_tree [None req-b0eb83ae-2fbb-47ee-abe7-62f2ae1b54da tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1186.011629] env[65726]: WARNING openstack [None req-b7bd70fc-78fa-4df3-bb13-d10a1172d2a5 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1186.011856] env[65726]: WARNING openstack [None req-b7bd70fc-78fa-4df3-bb13-d10a1172d2a5 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1186.042095] env[65726]: DEBUG oslo_vmware.api [None req-f04e60e2-18c2-41e8-8f5e-88ce749ed252 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Task: {'id': task-5116897, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.177122} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1186.042374] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-f04e60e2-18c2-41e8-8f5e-88ce749ed252 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: dc112e67-7c2d-4081-9a53-e4f43f61dcd6] Extended root virtual disk {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1186.043156] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbe1c75b-7177-452b-a049-ee59e2a924f3 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.066323] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-f04e60e2-18c2-41e8-8f5e-88ce749ed252 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: dc112e67-7c2d-4081-9a53-e4f43f61dcd6] Reconfiguring VM instance instance-0000006e to attach disk [datastore2] dc112e67-7c2d-4081-9a53-e4f43f61dcd6/dc112e67-7c2d-4081-9a53-e4f43f61dcd6.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1186.070667] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9530239a-b93b-4c3d-ac6d-82f74b59524c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.089523] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-252f0171-117c-4650-8dad-6b950e3f6801 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: 3bc25953-21b1-4729-af09-e7211fd8b2c6] Volume attach. Driver type: vmdk {{(pid=65726) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1186.089758] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-252f0171-117c-4650-8dad-6b950e3f6801 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: 3bc25953-21b1-4729-af09-e7211fd8b2c6] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-995319', 'volume_id': 'b58c28a5-1144-47d9-96ea-c0a59a979c77', 'name': 'volume-b58c28a5-1144-47d9-96ea-c0a59a979c77', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '3bc25953-21b1-4729-af09-e7211fd8b2c6', 'attached_at': '', 'detached_at': '', 'volume_id': 'b58c28a5-1144-47d9-96ea-c0a59a979c77', 'serial': 'b58c28a5-1144-47d9-96ea-c0a59a979c77'} {{(pid=65726) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1186.090726] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-116fd95a-5447-449a-a038-604a85857394 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.095392] env[65726]: DEBUG oslo_vmware.api [None req-f04e60e2-18c2-41e8-8f5e-88ce749ed252 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Waiting for the task: (returnval){ [ 1186.095392] env[65726]: value = "task-5116898" [ 1186.095392] env[65726]: _type = "Task" [ 1186.095392] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1186.113776] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8cc1a625-8731-4b2d-a615-230e9af3b4a8 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.122898] env[65726]: DEBUG oslo_vmware.api [None req-d8745e4a-3a23-4daa-80f7-d677b94a15ab tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': task-5116895, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.781746} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1186.123540] env[65726]: DEBUG oslo_vmware.api [None req-f04e60e2-18c2-41e8-8f5e-88ce749ed252 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Task: {'id': task-5116898, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1186.124197] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-d8745e4a-3a23-4daa-80f7-d677b94a15ab tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore2] e3255f3b-028f-4a0b-b621-873730417936/e3255f3b-028f-4a0b-b621-873730417936.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1186.124417] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-d8745e4a-3a23-4daa-80f7-d677b94a15ab tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: e3255f3b-028f-4a0b-b621-873730417936] Extending root virtual disk to 1048576 {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1186.124706] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-bc4a027d-4128-45f1-853c-20cacf5b4a5f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.147496] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-252f0171-117c-4650-8dad-6b950e3f6801 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: 3bc25953-21b1-4729-af09-e7211fd8b2c6] Reconfiguring VM instance instance-00000069 to attach disk [datastore2] volume-b58c28a5-1144-47d9-96ea-c0a59a979c77/volume-b58c28a5-1144-47d9-96ea-c0a59a979c77.vmdk or device None with type thin {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1186.150049] env[65726]: DEBUG nova.network.neutron [None req-b7bd70fc-78fa-4df3-bb13-d10a1172d2a5 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 378c1d18-4bb6-4245-80bf-370fb7af0575] Updating instance_info_cache with network_info: [{"id": "4765f6c2-d161-4e36-a6a8-3c8d39c47a40", "address": "fa:16:3e:3c:38:b2", "network": {"id": "0d0f561d-ac41-4ef5-9926-c707409799f7", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1134321217-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.160", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4eece77569624f90bf64e5c51974173f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2ce62383-8e84-4e26-955b-74c11392f4c9", "external-id": "nsx-vlan-transportzone-215", "segmentation_id": 215, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4765f6c2-d1", "ovs_interfaceid": "4765f6c2-d161-4e36-a6a8-3c8d39c47a40", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1186.155077] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ed4c33d8-d66e-44cb-9299-c1538622c6b1 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.173021] env[65726]: DEBUG oslo_concurrency.lockutils [None req-f3e961d9-f8e2-4b47-8ba1-7abd521b41a7 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1186.173729] env[65726]: DEBUG oslo_vmware.api [None req-d8745e4a-3a23-4daa-80f7-d677b94a15ab tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Waiting for the task: (returnval){ [ 1186.173729] env[65726]: value = "task-5116899" [ 1186.173729] env[65726]: _type = "Task" [ 1186.173729] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1186.184503] env[65726]: DEBUG oslo_vmware.api [None req-3b1f8e43-10c3-447e-a11b-a1499eca83f6 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52a7c5aa-85c3-1055-9f60-e7fa4c5d2dbb, 'name': SearchDatastore_Task, 'duration_secs': 0.062442} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1186.187071] env[65726]: DEBUG oslo_vmware.api [None req-252f0171-117c-4650-8dad-6b950e3f6801 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Waiting for the task: (returnval){ [ 1186.187071] env[65726]: value = "task-5116900" [ 1186.187071] env[65726]: _type = "Task" [ 1186.187071] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1186.187531] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8a289b88-c153-44fe-8925-f97a0bea3815 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.199037] env[65726]: DEBUG oslo_vmware.api [None req-d8745e4a-3a23-4daa-80f7-d677b94a15ab tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': task-5116899, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1186.205503] env[65726]: DEBUG oslo_vmware.api [None req-252f0171-117c-4650-8dad-6b950e3f6801 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Task: {'id': task-5116900, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1186.206819] env[65726]: DEBUG oslo_vmware.api [None req-3b1f8e43-10c3-447e-a11b-a1499eca83f6 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Waiting for the task: (returnval){ [ 1186.206819] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5202a5cf-c68c-25ad-d19d-c640b29380be" [ 1186.206819] env[65726]: _type = "Task" [ 1186.206819] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1186.221108] env[65726]: DEBUG oslo_vmware.api [None req-3b1f8e43-10c3-447e-a11b-a1499eca83f6 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5202a5cf-c68c-25ad-d19d-c640b29380be, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1186.340814] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-a3f83649-0195-4086-b546-9898d337f536 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: b353b5e9-500c-42d3-a87f-880a624febec] Destroying instance {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1186.341930] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f541f1ef-d801-4b37-898b-5aa9ed6b6bc3 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.350393] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-a3f83649-0195-4086-b546-9898d337f536 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: b353b5e9-500c-42d3-a87f-880a624febec] Unregistering the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1186.350669] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3712813e-4bd8-4dd8-9776-a6b40b340076 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.434331] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-a3f83649-0195-4086-b546-9898d337f536 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: b353b5e9-500c-42d3-a87f-880a624febec] Unregistered the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1186.434813] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-a3f83649-0195-4086-b546-9898d337f536 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: b353b5e9-500c-42d3-a87f-880a624febec] Deleting contents of the VM from datastore datastore2 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1186.435051] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-a3f83649-0195-4086-b546-9898d337f536 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Deleting the datastore file [datastore2] b353b5e9-500c-42d3-a87f-880a624febec {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1186.435266] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e92d692a-ec7e-46d3-985b-d2cf25752f14 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.444797] env[65726]: DEBUG oslo_vmware.api [None req-a3f83649-0195-4086-b546-9898d337f536 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Waiting for the task: (returnval){ [ 1186.444797] env[65726]: value = "task-5116902" [ 1186.444797] env[65726]: _type = "Task" [ 1186.444797] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1186.455778] env[65726]: DEBUG oslo_vmware.api [None req-a3f83649-0195-4086-b546-9898d337f536 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Task: {'id': task-5116902, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1186.473027] env[65726]: DEBUG nova.scheduler.client.report [None req-b0eb83ae-2fbb-47ee-abe7-62f2ae1b54da tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1186.608802] env[65726]: DEBUG oslo_vmware.api [None req-f04e60e2-18c2-41e8-8f5e-88ce749ed252 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Task: {'id': task-5116898, 'name': ReconfigVM_Task, 'duration_secs': 0.332006} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1186.609203] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-f04e60e2-18c2-41e8-8f5e-88ce749ed252 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: dc112e67-7c2d-4081-9a53-e4f43f61dcd6] Reconfigured VM instance instance-0000006e to attach disk [datastore2] dc112e67-7c2d-4081-9a53-e4f43f61dcd6/dc112e67-7c2d-4081-9a53-e4f43f61dcd6.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1186.609735] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-30858991-eace-4e25-b935-23ea45a3cdfd {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.617279] env[65726]: DEBUG oslo_vmware.api [None req-f04e60e2-18c2-41e8-8f5e-88ce749ed252 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Waiting for the task: (returnval){ [ 1186.617279] env[65726]: value = "task-5116903" [ 1186.617279] env[65726]: _type = "Task" [ 1186.617279] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1186.626788] env[65726]: DEBUG oslo_vmware.api [None req-f04e60e2-18c2-41e8-8f5e-88ce749ed252 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Task: {'id': task-5116903, 'name': Rename_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1186.670675] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b7bd70fc-78fa-4df3-bb13-d10a1172d2a5 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Releasing lock "refresh_cache-378c1d18-4bb6-4245-80bf-370fb7af0575" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1186.671406] env[65726]: WARNING openstack [None req-b7bd70fc-78fa-4df3-bb13-d10a1172d2a5 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1186.671879] env[65726]: WARNING openstack [None req-b7bd70fc-78fa-4df3-bb13-d10a1172d2a5 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1186.678349] env[65726]: WARNING openstack [None req-b7bd70fc-78fa-4df3-bb13-d10a1172d2a5 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1186.678719] env[65726]: WARNING openstack [None req-b7bd70fc-78fa-4df3-bb13-d10a1172d2a5 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1186.697731] env[65726]: DEBUG oslo_vmware.api [None req-d8745e4a-3a23-4daa-80f7-d677b94a15ab tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': task-5116899, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.075332} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1186.698503] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-d8745e4a-3a23-4daa-80f7-d677b94a15ab tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: e3255f3b-028f-4a0b-b621-873730417936] Extended root virtual disk {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1186.699542] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-439742fe-ca5e-4121-a0e3-381705804c63 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.706976] env[65726]: DEBUG oslo_vmware.api [None req-252f0171-117c-4650-8dad-6b950e3f6801 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Task: {'id': task-5116900, 'name': ReconfigVM_Task, 'duration_secs': 0.453205} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1186.708133] env[65726]: DEBUG nova.compute.manager [req-2df2c6b4-6ae9-4af0-a11f-2d6350e4ce08 req-a562b7a5-00d6-4cda-868f-742b20de2b3a service nova] [instance: b353b5e9-500c-42d3-a87f-880a624febec] Received event network-vif-unplugged-08b7016b-dc44-4206-a00c-5da943b82a38 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1186.708334] env[65726]: DEBUG oslo_concurrency.lockutils [req-2df2c6b4-6ae9-4af0-a11f-2d6350e4ce08 req-a562b7a5-00d6-4cda-868f-742b20de2b3a service nova] Acquiring lock "b353b5e9-500c-42d3-a87f-880a624febec-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1186.708531] env[65726]: DEBUG oslo_concurrency.lockutils [req-2df2c6b4-6ae9-4af0-a11f-2d6350e4ce08 req-a562b7a5-00d6-4cda-868f-742b20de2b3a service nova] Lock "b353b5e9-500c-42d3-a87f-880a624febec-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1186.708690] env[65726]: DEBUG oslo_concurrency.lockutils [req-2df2c6b4-6ae9-4af0-a11f-2d6350e4ce08 req-a562b7a5-00d6-4cda-868f-742b20de2b3a service nova] Lock "b353b5e9-500c-42d3-a87f-880a624febec-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1186.708850] env[65726]: DEBUG nova.compute.manager [req-2df2c6b4-6ae9-4af0-a11f-2d6350e4ce08 req-a562b7a5-00d6-4cda-868f-742b20de2b3a service nova] [instance: b353b5e9-500c-42d3-a87f-880a624febec] No waiting events found dispatching network-vif-unplugged-08b7016b-dc44-4206-a00c-5da943b82a38 {{(pid=65726) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1186.709020] env[65726]: WARNING nova.compute.manager [req-2df2c6b4-6ae9-4af0-a11f-2d6350e4ce08 req-a562b7a5-00d6-4cda-868f-742b20de2b3a service nova] [instance: b353b5e9-500c-42d3-a87f-880a624febec] Received unexpected event network-vif-unplugged-08b7016b-dc44-4206-a00c-5da943b82a38 for instance with vm_state shelved and task_state shelving_offloading. [ 1186.709198] env[65726]: DEBUG nova.compute.manager [req-2df2c6b4-6ae9-4af0-a11f-2d6350e4ce08 req-a562b7a5-00d6-4cda-868f-742b20de2b3a service nova] [instance: b353b5e9-500c-42d3-a87f-880a624febec] Received event network-changed-08b7016b-dc44-4206-a00c-5da943b82a38 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1186.709345] env[65726]: DEBUG nova.compute.manager [req-2df2c6b4-6ae9-4af0-a11f-2d6350e4ce08 req-a562b7a5-00d6-4cda-868f-742b20de2b3a service nova] [instance: b353b5e9-500c-42d3-a87f-880a624febec] Refreshing instance network info cache due to event network-changed-08b7016b-dc44-4206-a00c-5da943b82a38. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 1186.709512] env[65726]: DEBUG oslo_concurrency.lockutils [req-2df2c6b4-6ae9-4af0-a11f-2d6350e4ce08 req-a562b7a5-00d6-4cda-868f-742b20de2b3a service nova] Acquiring lock "refresh_cache-b353b5e9-500c-42d3-a87f-880a624febec" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1186.709637] env[65726]: DEBUG oslo_concurrency.lockutils [req-2df2c6b4-6ae9-4af0-a11f-2d6350e4ce08 req-a562b7a5-00d6-4cda-868f-742b20de2b3a service nova] Acquired lock "refresh_cache-b353b5e9-500c-42d3-a87f-880a624febec" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1186.709790] env[65726]: DEBUG nova.network.neutron [req-2df2c6b4-6ae9-4af0-a11f-2d6350e4ce08 req-a562b7a5-00d6-4cda-868f-742b20de2b3a service nova] [instance: b353b5e9-500c-42d3-a87f-880a624febec] Refreshing network info cache for port 08b7016b-dc44-4206-a00c-5da943b82a38 {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 1186.711512] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-252f0171-117c-4650-8dad-6b950e3f6801 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: 3bc25953-21b1-4729-af09-e7211fd8b2c6] Reconfigured VM instance instance-00000069 to attach disk [datastore2] volume-b58c28a5-1144-47d9-96ea-c0a59a979c77/volume-b58c28a5-1144-47d9-96ea-c0a59a979c77.vmdk or device None with type thin {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1186.733959] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-dbd29c2b-4afa-43b6-8c12-4033dcd3094b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.752542] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-d8745e4a-3a23-4daa-80f7-d677b94a15ab tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: e3255f3b-028f-4a0b-b621-873730417936] Reconfiguring VM instance instance-0000006f to attach disk [datastore2] e3255f3b-028f-4a0b-b621-873730417936/e3255f3b-028f-4a0b-b621-873730417936.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1186.753901] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ca0269e5-2b41-4cac-9c12-b1f94252fd5d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.775284] env[65726]: DEBUG oslo_vmware.api [None req-3b1f8e43-10c3-447e-a11b-a1499eca83f6 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5202a5cf-c68c-25ad-d19d-c640b29380be, 'name': SearchDatastore_Task, 'duration_secs': 0.012699} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1186.776905] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3b1f8e43-10c3-447e-a11b-a1499eca83f6 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1186.777075] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-3b1f8e43-10c3-447e-a11b-a1499eca83f6 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore2] 55ee97c3-b014-48eb-b41b-fc5aa16bc09e/55ee97c3-b014-48eb-b41b-fc5aa16bc09e.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1186.778262] env[65726]: DEBUG oslo_vmware.api [None req-252f0171-117c-4650-8dad-6b950e3f6801 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Waiting for the task: (returnval){ [ 1186.778262] env[65726]: value = "task-5116904" [ 1186.778262] env[65726]: _type = "Task" [ 1186.778262] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1186.778627] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-12bdd365-f6e3-49c5-8a6f-874cfec5e5ac {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.780720] env[65726]: DEBUG oslo_vmware.api [None req-d8745e4a-3a23-4daa-80f7-d677b94a15ab tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Waiting for the task: (returnval){ [ 1186.780720] env[65726]: value = "task-5116905" [ 1186.780720] env[65726]: _type = "Task" [ 1186.780720] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1186.792265] env[65726]: DEBUG oslo_vmware.api [None req-3b1f8e43-10c3-447e-a11b-a1499eca83f6 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Waiting for the task: (returnval){ [ 1186.792265] env[65726]: value = "task-5116906" [ 1186.792265] env[65726]: _type = "Task" [ 1186.792265] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1186.802927] env[65726]: DEBUG oslo_vmware.api [None req-252f0171-117c-4650-8dad-6b950e3f6801 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Task: {'id': task-5116904, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1186.802927] env[65726]: DEBUG oslo_vmware.api [None req-d8745e4a-3a23-4daa-80f7-d677b94a15ab tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': task-5116905, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1186.816887] env[65726]: DEBUG oslo_vmware.api [None req-3b1f8e43-10c3-447e-a11b-a1499eca83f6 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': task-5116906, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1186.955934] env[65726]: DEBUG oslo_vmware.api [None req-a3f83649-0195-4086-b546-9898d337f536 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Task: {'id': task-5116902, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.220964} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1186.956239] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-a3f83649-0195-4086-b546-9898d337f536 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Deleted the datastore file {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1186.956400] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-a3f83649-0195-4086-b546-9898d337f536 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: b353b5e9-500c-42d3-a87f-880a624febec] Deleted contents of the VM from datastore datastore2 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1186.956575] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-a3f83649-0195-4086-b546-9898d337f536 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: b353b5e9-500c-42d3-a87f-880a624febec] Instance destroyed {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1186.978980] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b0eb83ae-2fbb-47ee-abe7-62f2ae1b54da tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.311s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1186.980023] env[65726]: DEBUG nova.compute.manager [None req-b0eb83ae-2fbb-47ee-abe7-62f2ae1b54da tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 6956bda7-5657-45d7-8f80-c6809bd836b7] Start building networks asynchronously for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1186.983828] env[65726]: INFO nova.scheduler.client.report [None req-a3f83649-0195-4086-b546-9898d337f536 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Deleted allocations for instance b353b5e9-500c-42d3-a87f-880a624febec [ 1186.987486] env[65726]: DEBUG oslo_concurrency.lockutils [None req-f3e961d9-f8e2-4b47-8ba1-7abd521b41a7 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.814s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1186.987486] env[65726]: DEBUG oslo_concurrency.lockutils [None req-f3e961d9-f8e2-4b47-8ba1-7abd521b41a7 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1187.025822] env[65726]: INFO nova.scheduler.client.report [None req-f3e961d9-f8e2-4b47-8ba1-7abd521b41a7 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Deleted allocations for instance 7c1188ce-8718-4719-8631-e59e7915b7aa [ 1187.132067] env[65726]: DEBUG oslo_vmware.api [None req-f04e60e2-18c2-41e8-8f5e-88ce749ed252 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Task: {'id': task-5116903, 'name': Rename_Task, 'duration_secs': 0.260007} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1187.132371] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-f04e60e2-18c2-41e8-8f5e-88ce749ed252 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: dc112e67-7c2d-4081-9a53-e4f43f61dcd6] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1187.132782] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-95a1b45e-fafe-4cae-a64f-fb5af3db3b35 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1187.142211] env[65726]: DEBUG oslo_vmware.api [None req-f04e60e2-18c2-41e8-8f5e-88ce749ed252 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Waiting for the task: (returnval){ [ 1187.142211] env[65726]: value = "task-5116907" [ 1187.142211] env[65726]: _type = "Task" [ 1187.142211] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1187.152686] env[65726]: DEBUG oslo_vmware.api [None req-f04e60e2-18c2-41e8-8f5e-88ce749ed252 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Task: {'id': task-5116907, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1187.156931] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-b7bd70fc-78fa-4df3-bb13-d10a1172d2a5 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 378c1d18-4bb6-4245-80bf-370fb7af0575] Destroying instance {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1187.157984] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a57652b-20e2-4db5-b6c3-e2fc410c8333 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1187.168958] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-b7bd70fc-78fa-4df3-bb13-d10a1172d2a5 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 378c1d18-4bb6-4245-80bf-370fb7af0575] Unregistering the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1187.169312] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b90c57a4-46c2-4d9f-b45e-f2dc58cdd9d8 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1187.233287] env[65726]: WARNING openstack [req-2df2c6b4-6ae9-4af0-a11f-2d6350e4ce08 req-a562b7a5-00d6-4cda-868f-742b20de2b3a service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1187.233287] env[65726]: WARNING openstack [req-2df2c6b4-6ae9-4af0-a11f-2d6350e4ce08 req-a562b7a5-00d6-4cda-868f-742b20de2b3a service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1187.251408] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-b7bd70fc-78fa-4df3-bb13-d10a1172d2a5 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 378c1d18-4bb6-4245-80bf-370fb7af0575] Unregistered the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1187.252157] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-b7bd70fc-78fa-4df3-bb13-d10a1172d2a5 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 378c1d18-4bb6-4245-80bf-370fb7af0575] Deleting contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1187.252289] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-b7bd70fc-78fa-4df3-bb13-d10a1172d2a5 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Deleting the datastore file [datastore1] 378c1d18-4bb6-4245-80bf-370fb7af0575 {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1187.252623] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-82f8801b-b2f5-4238-a6c3-833057465e5d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1187.261284] env[65726]: DEBUG oslo_vmware.api [None req-b7bd70fc-78fa-4df3-bb13-d10a1172d2a5 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Waiting for the task: (returnval){ [ 1187.261284] env[65726]: value = "task-5116909" [ 1187.261284] env[65726]: _type = "Task" [ 1187.261284] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1187.273808] env[65726]: DEBUG oslo_vmware.api [None req-b7bd70fc-78fa-4df3-bb13-d10a1172d2a5 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5116909, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1187.294881] env[65726]: DEBUG oslo_vmware.api [None req-d8745e4a-3a23-4daa-80f7-d677b94a15ab tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': task-5116905, 'name': ReconfigVM_Task, 'duration_secs': 0.341887} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1187.298750] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-d8745e4a-3a23-4daa-80f7-d677b94a15ab tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: e3255f3b-028f-4a0b-b621-873730417936] Reconfigured VM instance instance-0000006f to attach disk [datastore2] e3255f3b-028f-4a0b-b621-873730417936/e3255f3b-028f-4a0b-b621-873730417936.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1187.299708] env[65726]: DEBUG oslo_vmware.api [None req-252f0171-117c-4650-8dad-6b950e3f6801 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Task: {'id': task-5116904, 'name': ReconfigVM_Task, 'duration_secs': 0.173908} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1187.300055] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-76700431-296a-4230-a0e0-910c742ae2c4 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1187.302224] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-252f0171-117c-4650-8dad-6b950e3f6801 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: 3bc25953-21b1-4729-af09-e7211fd8b2c6] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-995319', 'volume_id': 'b58c28a5-1144-47d9-96ea-c0a59a979c77', 'name': 'volume-b58c28a5-1144-47d9-96ea-c0a59a979c77', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '3bc25953-21b1-4729-af09-e7211fd8b2c6', 'attached_at': '', 'detached_at': '', 'volume_id': 'b58c28a5-1144-47d9-96ea-c0a59a979c77', 'serial': 'b58c28a5-1144-47d9-96ea-c0a59a979c77'} {{(pid=65726) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1187.314684] env[65726]: DEBUG oslo_vmware.api [None req-3b1f8e43-10c3-447e-a11b-a1499eca83f6 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': task-5116906, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1187.316680] env[65726]: DEBUG oslo_vmware.api [None req-d8745e4a-3a23-4daa-80f7-d677b94a15ab tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Waiting for the task: (returnval){ [ 1187.316680] env[65726]: value = "task-5116910" [ 1187.316680] env[65726]: _type = "Task" [ 1187.316680] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1187.328432] env[65726]: DEBUG oslo_vmware.api [None req-d8745e4a-3a23-4daa-80f7-d677b94a15ab tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': task-5116910, 'name': Rename_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1187.386814] env[65726]: WARNING openstack [req-2df2c6b4-6ae9-4af0-a11f-2d6350e4ce08 req-a562b7a5-00d6-4cda-868f-742b20de2b3a service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1187.387256] env[65726]: WARNING openstack [req-2df2c6b4-6ae9-4af0-a11f-2d6350e4ce08 req-a562b7a5-00d6-4cda-868f-742b20de2b3a service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1187.484668] env[65726]: WARNING openstack [req-2df2c6b4-6ae9-4af0-a11f-2d6350e4ce08 req-a562b7a5-00d6-4cda-868f-742b20de2b3a service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1187.485107] env[65726]: WARNING openstack [req-2df2c6b4-6ae9-4af0-a11f-2d6350e4ce08 req-a562b7a5-00d6-4cda-868f-742b20de2b3a service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1187.496124] env[65726]: DEBUG nova.compute.utils [None req-b0eb83ae-2fbb-47ee-abe7-62f2ae1b54da tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Using /dev/sd instead of None {{(pid=65726) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1187.498596] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a3f83649-0195-4086-b546-9898d337f536 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1187.498919] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a3f83649-0195-4086-b546-9898d337f536 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1187.499188] env[65726]: DEBUG nova.objects.instance [None req-a3f83649-0195-4086-b546-9898d337f536 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Lazy-loading 'resources' on Instance uuid b353b5e9-500c-42d3-a87f-880a624febec {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1187.501290] env[65726]: DEBUG nova.compute.manager [None req-b0eb83ae-2fbb-47ee-abe7-62f2ae1b54da tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 6956bda7-5657-45d7-8f80-c6809bd836b7] Allocating IP information in the background. {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1187.501495] env[65726]: DEBUG nova.network.neutron [None req-b0eb83ae-2fbb-47ee-abe7-62f2ae1b54da tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 6956bda7-5657-45d7-8f80-c6809bd836b7] allocate_for_instance() {{(pid=65726) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 1187.501830] env[65726]: WARNING neutronclient.v2_0.client [None req-b0eb83ae-2fbb-47ee-abe7-62f2ae1b54da tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1187.502166] env[65726]: WARNING neutronclient.v2_0.client [None req-b0eb83ae-2fbb-47ee-abe7-62f2ae1b54da tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1187.503272] env[65726]: WARNING openstack [None req-b0eb83ae-2fbb-47ee-abe7-62f2ae1b54da tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1187.503364] env[65726]: WARNING openstack [None req-b0eb83ae-2fbb-47ee-abe7-62f2ae1b54da tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1187.511776] env[65726]: DEBUG nova.objects.instance [None req-a3f83649-0195-4086-b546-9898d337f536 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Lazy-loading 'numa_topology' on Instance uuid b353b5e9-500c-42d3-a87f-880a624febec {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1187.541752] env[65726]: DEBUG oslo_concurrency.lockutils [None req-f3e961d9-f8e2-4b47-8ba1-7abd521b41a7 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Lock "7c1188ce-8718-4719-8631-e59e7915b7aa" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 4.802s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1187.563989] env[65726]: DEBUG nova.policy [None req-b0eb83ae-2fbb-47ee-abe7-62f2ae1b54da tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a82797497789484a8fbf88a23a449c95', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f93d852e2c904f42981cbfff882558e9', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65726) authorize /opt/stack/nova/nova/policy.py:192}} [ 1187.600798] env[65726]: DEBUG nova.network.neutron [req-2df2c6b4-6ae9-4af0-a11f-2d6350e4ce08 req-a562b7a5-00d6-4cda-868f-742b20de2b3a service nova] [instance: b353b5e9-500c-42d3-a87f-880a624febec] Updated VIF entry in instance network info cache for port 08b7016b-dc44-4206-a00c-5da943b82a38. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 1187.601204] env[65726]: DEBUG nova.network.neutron [req-2df2c6b4-6ae9-4af0-a11f-2d6350e4ce08 req-a562b7a5-00d6-4cda-868f-742b20de2b3a service nova] [instance: b353b5e9-500c-42d3-a87f-880a624febec] Updating instance_info_cache with network_info: [{"id": "08b7016b-dc44-4206-a00c-5da943b82a38", "address": "fa:16:3e:c5:e4:14", "network": {"id": "eb3278b7-c190-441b-84d2-305f9b896382", "bridge": null, "label": "tempest-AttachVolumeShelveTestJSON-1754571022-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.144", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "96149159e18e44f9bf3453e67681f224", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap08b7016b-dc", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1187.654031] env[65726]: DEBUG oslo_vmware.api [None req-f04e60e2-18c2-41e8-8f5e-88ce749ed252 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Task: {'id': task-5116907, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1187.771499] env[65726]: DEBUG oslo_vmware.api [None req-b7bd70fc-78fa-4df3-bb13-d10a1172d2a5 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5116909, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.254289} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1187.771816] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-b7bd70fc-78fa-4df3-bb13-d10a1172d2a5 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Deleted the datastore file {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1187.772052] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-b7bd70fc-78fa-4df3-bb13-d10a1172d2a5 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 378c1d18-4bb6-4245-80bf-370fb7af0575] Deleted contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1187.772268] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-b7bd70fc-78fa-4df3-bb13-d10a1172d2a5 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 378c1d18-4bb6-4245-80bf-370fb7af0575] Instance destroyed {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1187.795226] env[65726]: INFO nova.scheduler.client.report [None req-b7bd70fc-78fa-4df3-bb13-d10a1172d2a5 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Deleted allocations for instance 378c1d18-4bb6-4245-80bf-370fb7af0575 [ 1187.818792] env[65726]: DEBUG oslo_vmware.api [None req-3b1f8e43-10c3-447e-a11b-a1499eca83f6 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': task-5116906, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.555426} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1187.822652] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-3b1f8e43-10c3-447e-a11b-a1499eca83f6 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore2] 55ee97c3-b014-48eb-b41b-fc5aa16bc09e/55ee97c3-b014-48eb-b41b-fc5aa16bc09e.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1187.822652] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-3b1f8e43-10c3-447e-a11b-a1499eca83f6 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 55ee97c3-b014-48eb-b41b-fc5aa16bc09e] Extending root virtual disk to 1048576 {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1187.826085] env[65726]: DEBUG nova.network.neutron [None req-b0eb83ae-2fbb-47ee-abe7-62f2ae1b54da tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 6956bda7-5657-45d7-8f80-c6809bd836b7] Successfully created port: dad01f9d-0dc4-483c-8e27-cf7ffeec9519 {{(pid=65726) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1187.828807] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c26b34d7-b0b0-4eaa-96a2-3aa13230f185 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1187.837191] env[65726]: DEBUG oslo_vmware.api [None req-d8745e4a-3a23-4daa-80f7-d677b94a15ab tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': task-5116910, 'name': Rename_Task, 'duration_secs': 0.186424} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1187.839388] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-d8745e4a-3a23-4daa-80f7-d677b94a15ab tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: e3255f3b-028f-4a0b-b621-873730417936] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1187.839721] env[65726]: DEBUG oslo_vmware.api [None req-3b1f8e43-10c3-447e-a11b-a1499eca83f6 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Waiting for the task: (returnval){ [ 1187.839721] env[65726]: value = "task-5116911" [ 1187.839721] env[65726]: _type = "Task" [ 1187.839721] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1187.840603] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-45e2304a-eaf7-43df-875f-1906973928c4 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1187.853635] env[65726]: DEBUG oslo_vmware.api [None req-3b1f8e43-10c3-447e-a11b-a1499eca83f6 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': task-5116911, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1187.856912] env[65726]: DEBUG oslo_vmware.api [None req-d8745e4a-3a23-4daa-80f7-d677b94a15ab tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Waiting for the task: (returnval){ [ 1187.856912] env[65726]: value = "task-5116912" [ 1187.856912] env[65726]: _type = "Task" [ 1187.856912] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1187.870713] env[65726]: DEBUG oslo_vmware.api [None req-d8745e4a-3a23-4daa-80f7-d677b94a15ab tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': task-5116912, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1188.002886] env[65726]: DEBUG nova.compute.manager [None req-b0eb83ae-2fbb-47ee-abe7-62f2ae1b54da tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 6956bda7-5657-45d7-8f80-c6809bd836b7] Start building block device mappings for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1188.015787] env[65726]: DEBUG nova.objects.base [None req-a3f83649-0195-4086-b546-9898d337f536 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Object Instance lazy-loaded attributes: resources,numa_topology {{(pid=65726) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1188.104327] env[65726]: DEBUG oslo_concurrency.lockutils [req-2df2c6b4-6ae9-4af0-a11f-2d6350e4ce08 req-a562b7a5-00d6-4cda-868f-742b20de2b3a service nova] Releasing lock "refresh_cache-b353b5e9-500c-42d3-a87f-880a624febec" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1188.136897] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4af26e3a-434e-45aa-966f-18bc0b8df6b5 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.152456] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fdc5210-c2bc-4ede-9678-7f1c3a768597 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.162704] env[65726]: DEBUG oslo_vmware.api [None req-f04e60e2-18c2-41e8-8f5e-88ce749ed252 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Task: {'id': task-5116907, 'name': PowerOnVM_Task, 'duration_secs': 0.615737} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1188.189903] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-f04e60e2-18c2-41e8-8f5e-88ce749ed252 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: dc112e67-7c2d-4081-9a53-e4f43f61dcd6] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1188.189903] env[65726]: INFO nova.compute.manager [None req-f04e60e2-18c2-41e8-8f5e-88ce749ed252 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: dc112e67-7c2d-4081-9a53-e4f43f61dcd6] Took 12.46 seconds to spawn the instance on the hypervisor. [ 1188.189903] env[65726]: DEBUG nova.compute.manager [None req-f04e60e2-18c2-41e8-8f5e-88ce749ed252 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: dc112e67-7c2d-4081-9a53-e4f43f61dcd6] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1188.190554] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9eaeb35f-96ce-4028-ae40-b668d6037060 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.193805] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f12ac6e-0ef8-4849-a5de-bb16dcf165b1 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.206978] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-242fe00e-c4ab-4f2e-9bea-0e3d86f4d4c8 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.223980] env[65726]: DEBUG nova.compute.provider_tree [None req-a3f83649-0195-4086-b546-9898d337f536 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1188.299796] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b7bd70fc-78fa-4df3-bb13-d10a1172d2a5 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1188.353399] env[65726]: DEBUG oslo_vmware.api [None req-3b1f8e43-10c3-447e-a11b-a1499eca83f6 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': task-5116911, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.071199} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1188.353764] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-3b1f8e43-10c3-447e-a11b-a1499eca83f6 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 55ee97c3-b014-48eb-b41b-fc5aa16bc09e] Extended root virtual disk {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1188.354596] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-557a0857-5e41-4996-b034-359ce4a6f1a6 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.370249] env[65726]: DEBUG nova.objects.instance [None req-252f0171-117c-4650-8dad-6b950e3f6801 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Lazy-loading 'flavor' on Instance uuid 3bc25953-21b1-4729-af09-e7211fd8b2c6 {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1188.381415] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-3b1f8e43-10c3-447e-a11b-a1499eca83f6 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 55ee97c3-b014-48eb-b41b-fc5aa16bc09e] Reconfiguring VM instance instance-00000070 to attach disk [datastore2] 55ee97c3-b014-48eb-b41b-fc5aa16bc09e/55ee97c3-b014-48eb-b41b-fc5aa16bc09e.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1188.383479] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-93435a44-085f-424f-b677-48098b9ae3d1 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.403345] env[65726]: DEBUG oslo_vmware.api [None req-d8745e4a-3a23-4daa-80f7-d677b94a15ab tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': task-5116912, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1188.409782] env[65726]: DEBUG oslo_vmware.api [None req-3b1f8e43-10c3-447e-a11b-a1499eca83f6 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Waiting for the task: (returnval){ [ 1188.409782] env[65726]: value = "task-5116913" [ 1188.409782] env[65726]: _type = "Task" [ 1188.409782] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1188.422655] env[65726]: DEBUG oslo_vmware.api [None req-3b1f8e43-10c3-447e-a11b-a1499eca83f6 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': task-5116913, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1188.715141] env[65726]: INFO nova.compute.manager [None req-f04e60e2-18c2-41e8-8f5e-88ce749ed252 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: dc112e67-7c2d-4081-9a53-e4f43f61dcd6] Took 17.38 seconds to build instance. [ 1188.727271] env[65726]: DEBUG nova.scheduler.client.report [None req-a3f83649-0195-4086-b546-9898d337f536 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1188.868555] env[65726]: DEBUG oslo_vmware.api [None req-d8745e4a-3a23-4daa-80f7-d677b94a15ab tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': task-5116912, 'name': PowerOnVM_Task, 'duration_secs': 0.584147} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1188.868870] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-d8745e4a-3a23-4daa-80f7-d677b94a15ab tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: e3255f3b-028f-4a0b-b621-873730417936] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1188.869079] env[65726]: INFO nova.compute.manager [None req-d8745e4a-3a23-4daa-80f7-d677b94a15ab tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: e3255f3b-028f-4a0b-b621-873730417936] Took 10.83 seconds to spawn the instance on the hypervisor. [ 1188.869260] env[65726]: DEBUG nova.compute.manager [None req-d8745e4a-3a23-4daa-80f7-d677b94a15ab tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: e3255f3b-028f-4a0b-b621-873730417936] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1188.870014] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c865dd21-7b41-485b-a36e-5dfad7da4195 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.884472] env[65726]: DEBUG oslo_concurrency.lockutils [None req-252f0171-117c-4650-8dad-6b950e3f6801 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Lock "3bc25953-21b1-4729-af09-e7211fd8b2c6" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.427s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1188.921435] env[65726]: DEBUG oslo_vmware.api [None req-3b1f8e43-10c3-447e-a11b-a1499eca83f6 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': task-5116913, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1188.932051] env[65726]: DEBUG nova.compute.manager [req-f29e49b6-c3ff-48b8-b115-5b036dae6854 req-3fa96f0e-4f80-4cc3-843c-892f6c347773 service nova] [instance: 378c1d18-4bb6-4245-80bf-370fb7af0575] Received event network-vif-unplugged-4765f6c2-d161-4e36-a6a8-3c8d39c47a40 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1188.932338] env[65726]: DEBUG oslo_concurrency.lockutils [req-f29e49b6-c3ff-48b8-b115-5b036dae6854 req-3fa96f0e-4f80-4cc3-843c-892f6c347773 service nova] Acquiring lock "378c1d18-4bb6-4245-80bf-370fb7af0575-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1188.932537] env[65726]: DEBUG oslo_concurrency.lockutils [req-f29e49b6-c3ff-48b8-b115-5b036dae6854 req-3fa96f0e-4f80-4cc3-843c-892f6c347773 service nova] Lock "378c1d18-4bb6-4245-80bf-370fb7af0575-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1188.932818] env[65726]: DEBUG oslo_concurrency.lockutils [req-f29e49b6-c3ff-48b8-b115-5b036dae6854 req-3fa96f0e-4f80-4cc3-843c-892f6c347773 service nova] Lock "378c1d18-4bb6-4245-80bf-370fb7af0575-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1188.932886] env[65726]: DEBUG nova.compute.manager [req-f29e49b6-c3ff-48b8-b115-5b036dae6854 req-3fa96f0e-4f80-4cc3-843c-892f6c347773 service nova] [instance: 378c1d18-4bb6-4245-80bf-370fb7af0575] No waiting events found dispatching network-vif-unplugged-4765f6c2-d161-4e36-a6a8-3c8d39c47a40 {{(pid=65726) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1188.933048] env[65726]: WARNING nova.compute.manager [req-f29e49b6-c3ff-48b8-b115-5b036dae6854 req-3fa96f0e-4f80-4cc3-843c-892f6c347773 service nova] [instance: 378c1d18-4bb6-4245-80bf-370fb7af0575] Received unexpected event network-vif-unplugged-4765f6c2-d161-4e36-a6a8-3c8d39c47a40 for instance with vm_state shelved_offloaded and task_state unshelving. [ 1188.933210] env[65726]: DEBUG nova.compute.manager [req-f29e49b6-c3ff-48b8-b115-5b036dae6854 req-3fa96f0e-4f80-4cc3-843c-892f6c347773 service nova] [instance: 378c1d18-4bb6-4245-80bf-370fb7af0575] Received event network-changed-4765f6c2-d161-4e36-a6a8-3c8d39c47a40 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1188.933409] env[65726]: DEBUG nova.compute.manager [req-f29e49b6-c3ff-48b8-b115-5b036dae6854 req-3fa96f0e-4f80-4cc3-843c-892f6c347773 service nova] [instance: 378c1d18-4bb6-4245-80bf-370fb7af0575] Refreshing instance network info cache due to event network-changed-4765f6c2-d161-4e36-a6a8-3c8d39c47a40. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 1188.933597] env[65726]: DEBUG oslo_concurrency.lockutils [req-f29e49b6-c3ff-48b8-b115-5b036dae6854 req-3fa96f0e-4f80-4cc3-843c-892f6c347773 service nova] Acquiring lock "refresh_cache-378c1d18-4bb6-4245-80bf-370fb7af0575" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1188.933782] env[65726]: DEBUG oslo_concurrency.lockutils [req-f29e49b6-c3ff-48b8-b115-5b036dae6854 req-3fa96f0e-4f80-4cc3-843c-892f6c347773 service nova] Acquired lock "refresh_cache-378c1d18-4bb6-4245-80bf-370fb7af0575" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1188.933921] env[65726]: DEBUG nova.network.neutron [req-f29e49b6-c3ff-48b8-b115-5b036dae6854 req-3fa96f0e-4f80-4cc3-843c-892f6c347773 service nova] [instance: 378c1d18-4bb6-4245-80bf-370fb7af0575] Refreshing network info cache for port 4765f6c2-d161-4e36-a6a8-3c8d39c47a40 {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 1189.015141] env[65726]: DEBUG nova.compute.manager [None req-b0eb83ae-2fbb-47ee-abe7-62f2ae1b54da tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 6956bda7-5657-45d7-8f80-c6809bd836b7] Start spawning the instance on the hypervisor. {{(pid=65726) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1189.044396] env[65726]: DEBUG nova.virt.hardware [None req-b0eb83ae-2fbb-47ee-abe7-62f2ae1b54da tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T19:28:34Z,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1539f06fa3534e90acbc3a60c4b8bd3f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T19:28:35Z,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1189.044396] env[65726]: DEBUG nova.virt.hardware [None req-b0eb83ae-2fbb-47ee-abe7-62f2ae1b54da tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1189.044785] env[65726]: DEBUG nova.virt.hardware [None req-b0eb83ae-2fbb-47ee-abe7-62f2ae1b54da tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1189.044862] env[65726]: DEBUG nova.virt.hardware [None req-b0eb83ae-2fbb-47ee-abe7-62f2ae1b54da tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1189.045018] env[65726]: DEBUG nova.virt.hardware [None req-b0eb83ae-2fbb-47ee-abe7-62f2ae1b54da tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1189.045222] env[65726]: DEBUG nova.virt.hardware [None req-b0eb83ae-2fbb-47ee-abe7-62f2ae1b54da tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1189.045991] env[65726]: DEBUG nova.virt.hardware [None req-b0eb83ae-2fbb-47ee-abe7-62f2ae1b54da tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1189.045991] env[65726]: DEBUG nova.virt.hardware [None req-b0eb83ae-2fbb-47ee-abe7-62f2ae1b54da tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1189.045991] env[65726]: DEBUG nova.virt.hardware [None req-b0eb83ae-2fbb-47ee-abe7-62f2ae1b54da tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1189.046159] env[65726]: DEBUG nova.virt.hardware [None req-b0eb83ae-2fbb-47ee-abe7-62f2ae1b54da tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1189.046409] env[65726]: DEBUG nova.virt.hardware [None req-b0eb83ae-2fbb-47ee-abe7-62f2ae1b54da tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1189.048405] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a6a891c-3bc5-4d31-a5cd-1e039e370a5b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.060025] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3b27c62-4206-4660-a4cc-0da0726be305 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.217135] env[65726]: DEBUG oslo_concurrency.lockutils [None req-f04e60e2-18c2-41e8-8f5e-88ce749ed252 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Lock "dc112e67-7c2d-4081-9a53-e4f43f61dcd6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 18.890s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1189.231345] env[65726]: DEBUG oslo_concurrency.lockutils [None req-acbda108-3421-436d-befd-ccf16d53d3c3 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Acquiring lock "3bc25953-21b1-4729-af09-e7211fd8b2c6" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1189.231618] env[65726]: DEBUG oslo_concurrency.lockutils [None req-acbda108-3421-436d-befd-ccf16d53d3c3 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Lock "3bc25953-21b1-4729-af09-e7211fd8b2c6" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1189.233434] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a3f83649-0195-4086-b546-9898d337f536 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.735s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1189.235927] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b7bd70fc-78fa-4df3-bb13-d10a1172d2a5 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.936s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1189.236185] env[65726]: DEBUG nova.objects.instance [None req-b7bd70fc-78fa-4df3-bb13-d10a1172d2a5 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Lazy-loading 'resources' on Instance uuid 378c1d18-4bb6-4245-80bf-370fb7af0575 {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1189.346720] env[65726]: DEBUG nova.network.neutron [None req-b0eb83ae-2fbb-47ee-abe7-62f2ae1b54da tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 6956bda7-5657-45d7-8f80-c6809bd836b7] Successfully updated port: dad01f9d-0dc4-483c-8e27-cf7ffeec9519 {{(pid=65726) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 1189.371726] env[65726]: DEBUG nova.compute.manager [req-044dcf91-901b-4d2f-a227-ba5c03b042a1 req-beb4a09b-2fa3-4fd5-be43-e464c8cdee1e service nova] [instance: 6956bda7-5657-45d7-8f80-c6809bd836b7] Received event network-vif-plugged-dad01f9d-0dc4-483c-8e27-cf7ffeec9519 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1189.371726] env[65726]: DEBUG oslo_concurrency.lockutils [req-044dcf91-901b-4d2f-a227-ba5c03b042a1 req-beb4a09b-2fa3-4fd5-be43-e464c8cdee1e service nova] Acquiring lock "6956bda7-5657-45d7-8f80-c6809bd836b7-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1189.371726] env[65726]: DEBUG oslo_concurrency.lockutils [req-044dcf91-901b-4d2f-a227-ba5c03b042a1 req-beb4a09b-2fa3-4fd5-be43-e464c8cdee1e service nova] Lock "6956bda7-5657-45d7-8f80-c6809bd836b7-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1189.371726] env[65726]: DEBUG oslo_concurrency.lockutils [req-044dcf91-901b-4d2f-a227-ba5c03b042a1 req-beb4a09b-2fa3-4fd5-be43-e464c8cdee1e service nova] Lock "6956bda7-5657-45d7-8f80-c6809bd836b7-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1189.372177] env[65726]: DEBUG nova.compute.manager [req-044dcf91-901b-4d2f-a227-ba5c03b042a1 req-beb4a09b-2fa3-4fd5-be43-e464c8cdee1e service nova] [instance: 6956bda7-5657-45d7-8f80-c6809bd836b7] No waiting events found dispatching network-vif-plugged-dad01f9d-0dc4-483c-8e27-cf7ffeec9519 {{(pid=65726) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1189.372463] env[65726]: WARNING nova.compute.manager [req-044dcf91-901b-4d2f-a227-ba5c03b042a1 req-beb4a09b-2fa3-4fd5-be43-e464c8cdee1e service nova] [instance: 6956bda7-5657-45d7-8f80-c6809bd836b7] Received unexpected event network-vif-plugged-dad01f9d-0dc4-483c-8e27-cf7ffeec9519 for instance with vm_state building and task_state spawning. [ 1189.386743] env[65726]: INFO nova.compute.manager [None req-d8745e4a-3a23-4daa-80f7-d677b94a15ab tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: e3255f3b-028f-4a0b-b621-873730417936] Took 16.71 seconds to build instance. [ 1189.421759] env[65726]: DEBUG oslo_vmware.api [None req-3b1f8e43-10c3-447e-a11b-a1499eca83f6 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': task-5116913, 'name': ReconfigVM_Task, 'duration_secs': 0.772214} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1189.422078] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-3b1f8e43-10c3-447e-a11b-a1499eca83f6 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 55ee97c3-b014-48eb-b41b-fc5aa16bc09e] Reconfigured VM instance instance-00000070 to attach disk [datastore2] 55ee97c3-b014-48eb-b41b-fc5aa16bc09e/55ee97c3-b014-48eb-b41b-fc5aa16bc09e.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1189.422721] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e706b059-e9bb-4a07-9c36-cb412242fd20 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.429749] env[65726]: DEBUG oslo_vmware.api [None req-3b1f8e43-10c3-447e-a11b-a1499eca83f6 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Waiting for the task: (returnval){ [ 1189.429749] env[65726]: value = "task-5116914" [ 1189.429749] env[65726]: _type = "Task" [ 1189.429749] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1189.439621] env[65726]: WARNING openstack [req-f29e49b6-c3ff-48b8-b115-5b036dae6854 req-3fa96f0e-4f80-4cc3-843c-892f6c347773 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1189.440035] env[65726]: WARNING openstack [req-f29e49b6-c3ff-48b8-b115-5b036dae6854 req-3fa96f0e-4f80-4cc3-843c-892f6c347773 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1189.446973] env[65726]: DEBUG oslo_vmware.api [None req-3b1f8e43-10c3-447e-a11b-a1499eca83f6 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': task-5116914, 'name': Rename_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1189.636046] env[65726]: WARNING openstack [req-f29e49b6-c3ff-48b8-b115-5b036dae6854 req-3fa96f0e-4f80-4cc3-843c-892f6c347773 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1189.639490] env[65726]: WARNING openstack [req-f29e49b6-c3ff-48b8-b115-5b036dae6854 req-3fa96f0e-4f80-4cc3-843c-892f6c347773 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1189.649457] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a00ae9d0-fb69-476d-95d1-c304afd40cef tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Acquiring lock "dc112e67-7c2d-4081-9a53-e4f43f61dcd6" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1189.649578] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a00ae9d0-fb69-476d-95d1-c304afd40cef tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Lock "dc112e67-7c2d-4081-9a53-e4f43f61dcd6" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1189.649769] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a00ae9d0-fb69-476d-95d1-c304afd40cef tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Acquiring lock "dc112e67-7c2d-4081-9a53-e4f43f61dcd6-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1189.649943] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a00ae9d0-fb69-476d-95d1-c304afd40cef tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Lock "dc112e67-7c2d-4081-9a53-e4f43f61dcd6-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1189.650119] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a00ae9d0-fb69-476d-95d1-c304afd40cef tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Lock "dc112e67-7c2d-4081-9a53-e4f43f61dcd6-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1189.652637] env[65726]: INFO nova.compute.manager [None req-a00ae9d0-fb69-476d-95d1-c304afd40cef tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: dc112e67-7c2d-4081-9a53-e4f43f61dcd6] Terminating instance [ 1189.737483] env[65726]: INFO nova.compute.manager [None req-acbda108-3421-436d-befd-ccf16d53d3c3 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: 3bc25953-21b1-4729-af09-e7211fd8b2c6] Detaching volume b58c28a5-1144-47d9-96ea-c0a59a979c77 [ 1189.741482] env[65726]: DEBUG nova.objects.instance [None req-b7bd70fc-78fa-4df3-bb13-d10a1172d2a5 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Lazy-loading 'numa_topology' on Instance uuid 378c1d18-4bb6-4245-80bf-370fb7af0575 {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1189.747726] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a3f83649-0195-4086-b546-9898d337f536 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Lock "b353b5e9-500c-42d3-a87f-880a624febec" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 25.125s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1189.778234] env[65726]: INFO nova.virt.block_device [None req-acbda108-3421-436d-befd-ccf16d53d3c3 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: 3bc25953-21b1-4729-af09-e7211fd8b2c6] Attempting to driver detach volume b58c28a5-1144-47d9-96ea-c0a59a979c77 from mountpoint /dev/sdb [ 1189.778234] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-acbda108-3421-436d-befd-ccf16d53d3c3 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: 3bc25953-21b1-4729-af09-e7211fd8b2c6] Volume detach. Driver type: vmdk {{(pid=65726) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1189.778234] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-acbda108-3421-436d-befd-ccf16d53d3c3 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: 3bc25953-21b1-4729-af09-e7211fd8b2c6] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-995319', 'volume_id': 'b58c28a5-1144-47d9-96ea-c0a59a979c77', 'name': 'volume-b58c28a5-1144-47d9-96ea-c0a59a979c77', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '3bc25953-21b1-4729-af09-e7211fd8b2c6', 'attached_at': '', 'detached_at': '', 'volume_id': 'b58c28a5-1144-47d9-96ea-c0a59a979c77', 'serial': 'b58c28a5-1144-47d9-96ea-c0a59a979c77'} {{(pid=65726) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1189.778234] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a1a2abe-054b-47c8-b1e8-8e17d8e6ebe7 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.802414] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22e040e3-8ff1-4801-806f-1efcf8fe5ace {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.812037] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70fcd806-932e-4396-8cff-f8ea8d7a6c08 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.838130] env[65726]: WARNING openstack [req-f29e49b6-c3ff-48b8-b115-5b036dae6854 req-3fa96f0e-4f80-4cc3-843c-892f6c347773 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1189.838601] env[65726]: WARNING openstack [req-f29e49b6-c3ff-48b8-b115-5b036dae6854 req-3fa96f0e-4f80-4cc3-843c-892f6c347773 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1189.846898] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08df8673-7e41-42cb-b18b-9f5510ffd69c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.850128] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b0eb83ae-2fbb-47ee-abe7-62f2ae1b54da tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Acquiring lock "refresh_cache-6956bda7-5657-45d7-8f80-c6809bd836b7" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1189.850293] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b0eb83ae-2fbb-47ee-abe7-62f2ae1b54da tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Acquired lock "refresh_cache-6956bda7-5657-45d7-8f80-c6809bd836b7" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1189.850434] env[65726]: DEBUG nova.network.neutron [None req-b0eb83ae-2fbb-47ee-abe7-62f2ae1b54da tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 6956bda7-5657-45d7-8f80-c6809bd836b7] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1189.867119] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e32abc12-7404-4281-87f0-12de5cc1731f tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Acquiring lock "378c1d18-4bb6-4245-80bf-370fb7af0575" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1189.867119] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-acbda108-3421-436d-befd-ccf16d53d3c3 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] The volume has not been displaced from its original location: [datastore2] volume-b58c28a5-1144-47d9-96ea-c0a59a979c77/volume-b58c28a5-1144-47d9-96ea-c0a59a979c77.vmdk. No consolidation needed. {{(pid=65726) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1189.871890] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-acbda108-3421-436d-befd-ccf16d53d3c3 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: 3bc25953-21b1-4729-af09-e7211fd8b2c6] Reconfiguring VM instance instance-00000069 to detach disk 2001 {{(pid=65726) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1189.872923] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2327721b-bc81-4769-9963-62323b2820bc {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.889618] env[65726]: DEBUG oslo_concurrency.lockutils [None req-d8745e4a-3a23-4daa-80f7-d677b94a15ab tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Lock "e3255f3b-028f-4a0b-b621-873730417936" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 18.226s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1189.899389] env[65726]: DEBUG oslo_vmware.api [None req-acbda108-3421-436d-befd-ccf16d53d3c3 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Waiting for the task: (returnval){ [ 1189.899389] env[65726]: value = "task-5116915" [ 1189.899389] env[65726]: _type = "Task" [ 1189.899389] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1189.913188] env[65726]: DEBUG oslo_vmware.api [None req-acbda108-3421-436d-befd-ccf16d53d3c3 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Task: {'id': task-5116915, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1189.936132] env[65726]: DEBUG nova.network.neutron [req-f29e49b6-c3ff-48b8-b115-5b036dae6854 req-3fa96f0e-4f80-4cc3-843c-892f6c347773 service nova] [instance: 378c1d18-4bb6-4245-80bf-370fb7af0575] Updated VIF entry in instance network info cache for port 4765f6c2-d161-4e36-a6a8-3c8d39c47a40. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 1189.936522] env[65726]: DEBUG nova.network.neutron [req-f29e49b6-c3ff-48b8-b115-5b036dae6854 req-3fa96f0e-4f80-4cc3-843c-892f6c347773 service nova] [instance: 378c1d18-4bb6-4245-80bf-370fb7af0575] Updating instance_info_cache with network_info: [{"id": "4765f6c2-d161-4e36-a6a8-3c8d39c47a40", "address": "fa:16:3e:3c:38:b2", "network": {"id": "0d0f561d-ac41-4ef5-9926-c707409799f7", "bridge": null, "label": "tempest-ServerActionsTestOtherB-1134321217-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.160", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4eece77569624f90bf64e5c51974173f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap4765f6c2-d1", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1189.942152] env[65726]: DEBUG oslo_vmware.api [None req-3b1f8e43-10c3-447e-a11b-a1499eca83f6 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': task-5116914, 'name': Rename_Task} progress is 99%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1190.124610] env[65726]: DEBUG oslo_service.periodic_task [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=65726) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1190.124610] env[65726]: DEBUG oslo_service.periodic_task [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=65726) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1190.157505] env[65726]: DEBUG nova.compute.manager [None req-a00ae9d0-fb69-476d-95d1-c304afd40cef tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: dc112e67-7c2d-4081-9a53-e4f43f61dcd6] Start destroying the instance on the hypervisor. {{(pid=65726) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1190.157725] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-a00ae9d0-fb69-476d-95d1-c304afd40cef tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: dc112e67-7c2d-4081-9a53-e4f43f61dcd6] Destroying instance {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1190.158617] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c58224b6-532f-4532-b550-825b40adb244 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.168088] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-a00ae9d0-fb69-476d-95d1-c304afd40cef tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: dc112e67-7c2d-4081-9a53-e4f43f61dcd6] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1190.168363] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-aec2b40a-c9e0-4040-9b40-fb7e03045437 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.176831] env[65726]: DEBUG oslo_vmware.api [None req-a00ae9d0-fb69-476d-95d1-c304afd40cef tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Waiting for the task: (returnval){ [ 1190.176831] env[65726]: value = "task-5116916" [ 1190.176831] env[65726]: _type = "Task" [ 1190.176831] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1190.186793] env[65726]: DEBUG oslo_vmware.api [None req-a00ae9d0-fb69-476d-95d1-c304afd40cef tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Task: {'id': task-5116916, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1190.245184] env[65726]: DEBUG nova.objects.base [None req-b7bd70fc-78fa-4df3-bb13-d10a1172d2a5 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Object Instance<378c1d18-4bb6-4245-80bf-370fb7af0575> lazy-loaded attributes: resources,numa_topology {{(pid=65726) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1190.353474] env[65726]: WARNING openstack [None req-b0eb83ae-2fbb-47ee-abe7-62f2ae1b54da tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1190.353941] env[65726]: WARNING openstack [None req-b0eb83ae-2fbb-47ee-abe7-62f2ae1b54da tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1190.359312] env[65726]: DEBUG nova.network.neutron [None req-b0eb83ae-2fbb-47ee-abe7-62f2ae1b54da tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 6956bda7-5657-45d7-8f80-c6809bd836b7] Instance cache missing network info. {{(pid=65726) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 1190.406732] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16d958c3-3f2a-4b0f-82c9-9be4f25690ff {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.418171] env[65726]: WARNING openstack [None req-b0eb83ae-2fbb-47ee-abe7-62f2ae1b54da tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1190.418558] env[65726]: WARNING openstack [None req-b0eb83ae-2fbb-47ee-abe7-62f2ae1b54da tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1190.425623] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e674781a-dca2-46eb-84e8-00b3b1a9a5ea {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.428811] env[65726]: DEBUG oslo_vmware.api [None req-acbda108-3421-436d-befd-ccf16d53d3c3 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Task: {'id': task-5116915, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1190.460599] env[65726]: DEBUG oslo_concurrency.lockutils [req-f29e49b6-c3ff-48b8-b115-5b036dae6854 req-3fa96f0e-4f80-4cc3-843c-892f6c347773 service nova] Releasing lock "refresh_cache-378c1d18-4bb6-4245-80bf-370fb7af0575" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1190.471228] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4f8caca-1420-4ebc-84ed-4776e4d68cf3 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.471817] env[65726]: DEBUG oslo_vmware.api [None req-3b1f8e43-10c3-447e-a11b-a1499eca83f6 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': task-5116914, 'name': Rename_Task} progress is 99%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1190.478192] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a02c95a2-0ebb-4747-a678-8f08cc139b37 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.494234] env[65726]: DEBUG nova.compute.provider_tree [None req-b7bd70fc-78fa-4df3-bb13-d10a1172d2a5 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1190.505778] env[65726]: WARNING openstack [None req-b0eb83ae-2fbb-47ee-abe7-62f2ae1b54da tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1190.506175] env[65726]: WARNING openstack [None req-b0eb83ae-2fbb-47ee-abe7-62f2ae1b54da tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1190.604889] env[65726]: DEBUG nova.network.neutron [None req-b0eb83ae-2fbb-47ee-abe7-62f2ae1b54da tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 6956bda7-5657-45d7-8f80-c6809bd836b7] Updating instance_info_cache with network_info: [{"id": "dad01f9d-0dc4-483c-8e27-cf7ffeec9519", "address": "fa:16:3e:17:36:eb", "network": {"id": "9dd55fb6-0dd2-484d-b94d-fbdccb958c79", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1005598846-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f93d852e2c904f42981cbfff882558e9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "74f30339-6421-4654-bddb-81d7f34db9d7", "external-id": "nsx-vlan-transportzone-899", "segmentation_id": 899, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdad01f9d-0d", "ovs_interfaceid": "dad01f9d-0dc4-483c-8e27-cf7ffeec9519", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1190.632580] env[65726]: DEBUG oslo_service.periodic_task [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=65726) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1190.632580] env[65726]: DEBUG oslo_service.periodic_task [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=65726) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1190.632771] env[65726]: DEBUG oslo_service.periodic_task [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=65726) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1190.632854] env[65726]: DEBUG oslo_service.periodic_task [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=65726) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1190.632988] env[65726]: DEBUG oslo_service.periodic_task [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=65726) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1190.633142] env[65726]: DEBUG oslo_service.periodic_task [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=65726) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1190.633279] env[65726]: DEBUG nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=65726) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11251}} [ 1190.633806] env[65726]: DEBUG oslo_service.periodic_task [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Running periodic task ComputeManager.update_available_resource {{(pid=65726) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1190.687193] env[65726]: DEBUG oslo_vmware.api [None req-a00ae9d0-fb69-476d-95d1-c304afd40cef tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Task: {'id': task-5116916, 'name': PowerOffVM_Task, 'duration_secs': 0.228152} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1190.687463] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-a00ae9d0-fb69-476d-95d1-c304afd40cef tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: dc112e67-7c2d-4081-9a53-e4f43f61dcd6] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1190.687628] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-a00ae9d0-fb69-476d-95d1-c304afd40cef tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: dc112e67-7c2d-4081-9a53-e4f43f61dcd6] Unregistering the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1190.687889] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0095abf2-4dee-49e5-b0c7-ef68f0ec88c0 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.699200] env[65726]: DEBUG oslo_concurrency.lockutils [None req-70bb45c2-ad53-4082-9238-425c98dcfe4d tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Acquiring lock "b353b5e9-500c-42d3-a87f-880a624febec" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1190.699435] env[65726]: DEBUG oslo_concurrency.lockutils [None req-70bb45c2-ad53-4082-9238-425c98dcfe4d tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Lock "b353b5e9-500c-42d3-a87f-880a624febec" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 0.001s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1190.699604] env[65726]: INFO nova.compute.manager [None req-70bb45c2-ad53-4082-9238-425c98dcfe4d tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: b353b5e9-500c-42d3-a87f-880a624febec] Unshelving [ 1190.755046] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-a00ae9d0-fb69-476d-95d1-c304afd40cef tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: dc112e67-7c2d-4081-9a53-e4f43f61dcd6] Unregistered the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1190.755046] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-a00ae9d0-fb69-476d-95d1-c304afd40cef tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: dc112e67-7c2d-4081-9a53-e4f43f61dcd6] Deleting contents of the VM from datastore datastore2 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1190.755494] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-a00ae9d0-fb69-476d-95d1-c304afd40cef tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Deleting the datastore file [datastore2] dc112e67-7c2d-4081-9a53-e4f43f61dcd6 {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1190.755529] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2619ef8f-ce57-41fe-9e5b-8044f71ead9d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.763033] env[65726]: DEBUG oslo_vmware.api [None req-a00ae9d0-fb69-476d-95d1-c304afd40cef tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Waiting for the task: (returnval){ [ 1190.763033] env[65726]: value = "task-5116918" [ 1190.763033] env[65726]: _type = "Task" [ 1190.763033] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1190.773285] env[65726]: DEBUG oslo_vmware.api [None req-a00ae9d0-fb69-476d-95d1-c304afd40cef tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Task: {'id': task-5116918, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1190.912947] env[65726]: DEBUG oslo_vmware.api [None req-acbda108-3421-436d-befd-ccf16d53d3c3 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Task: {'id': task-5116915, 'name': ReconfigVM_Task, 'duration_secs': 0.895942} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1190.912947] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-acbda108-3421-436d-befd-ccf16d53d3c3 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: 3bc25953-21b1-4729-af09-e7211fd8b2c6] Reconfigured VM instance instance-00000069 to detach disk 2001 {{(pid=65726) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1190.916356] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a9f27498-77f1-4e4b-b6ba-bb6836d2272c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.935934] env[65726]: DEBUG oslo_vmware.api [None req-acbda108-3421-436d-befd-ccf16d53d3c3 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Waiting for the task: (returnval){ [ 1190.935934] env[65726]: value = "task-5116919" [ 1190.935934] env[65726]: _type = "Task" [ 1190.935934] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1190.943470] env[65726]: DEBUG oslo_vmware.api [None req-3b1f8e43-10c3-447e-a11b-a1499eca83f6 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': task-5116914, 'name': Rename_Task, 'duration_secs': 1.163683} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1190.946879] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-3b1f8e43-10c3-447e-a11b-a1499eca83f6 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 55ee97c3-b014-48eb-b41b-fc5aa16bc09e] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1190.947171] env[65726]: DEBUG oslo_vmware.api [None req-acbda108-3421-436d-befd-ccf16d53d3c3 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Task: {'id': task-5116919, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1190.947390] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2da6d60c-764e-4ac1-bf8e-45015236ebc3 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.955065] env[65726]: DEBUG oslo_vmware.api [None req-3b1f8e43-10c3-447e-a11b-a1499eca83f6 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Waiting for the task: (returnval){ [ 1190.955065] env[65726]: value = "task-5116920" [ 1190.955065] env[65726]: _type = "Task" [ 1190.955065] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1190.963508] env[65726]: DEBUG oslo_vmware.api [None req-3b1f8e43-10c3-447e-a11b-a1499eca83f6 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': task-5116920, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1190.998640] env[65726]: DEBUG nova.scheduler.client.report [None req-b7bd70fc-78fa-4df3-bb13-d10a1172d2a5 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1191.107477] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b0eb83ae-2fbb-47ee-abe7-62f2ae1b54da tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Releasing lock "refresh_cache-6956bda7-5657-45d7-8f80-c6809bd836b7" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1191.107963] env[65726]: DEBUG nova.compute.manager [None req-b0eb83ae-2fbb-47ee-abe7-62f2ae1b54da tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 6956bda7-5657-45d7-8f80-c6809bd836b7] Instance network_info: |[{"id": "dad01f9d-0dc4-483c-8e27-cf7ffeec9519", "address": "fa:16:3e:17:36:eb", "network": {"id": "9dd55fb6-0dd2-484d-b94d-fbdccb958c79", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1005598846-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f93d852e2c904f42981cbfff882558e9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "74f30339-6421-4654-bddb-81d7f34db9d7", "external-id": "nsx-vlan-transportzone-899", "segmentation_id": 899, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdad01f9d-0d", "ovs_interfaceid": "dad01f9d-0dc4-483c-8e27-cf7ffeec9519", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1191.108554] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-b0eb83ae-2fbb-47ee-abe7-62f2ae1b54da tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 6956bda7-5657-45d7-8f80-c6809bd836b7] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:17:36:eb', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '74f30339-6421-4654-bddb-81d7f34db9d7', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'dad01f9d-0dc4-483c-8e27-cf7ffeec9519', 'vif_model': 'vmxnet3'}] {{(pid=65726) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1191.119131] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-b0eb83ae-2fbb-47ee-abe7-62f2ae1b54da tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1191.119496] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6956bda7-5657-45d7-8f80-c6809bd836b7] Creating VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1191.119821] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d595784c-cf1d-4d71-b5e9-f480bda4772c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1191.139065] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1191.146266] env[65726]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1191.146266] env[65726]: value = "task-5116921" [ 1191.146266] env[65726]: _type = "Task" [ 1191.146266] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1191.154929] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116921, 'name': CreateVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1191.275491] env[65726]: DEBUG oslo_vmware.api [None req-a00ae9d0-fb69-476d-95d1-c304afd40cef tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Task: {'id': task-5116918, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.126437} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1191.275800] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-a00ae9d0-fb69-476d-95d1-c304afd40cef tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Deleted the datastore file {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1191.275996] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-a00ae9d0-fb69-476d-95d1-c304afd40cef tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: dc112e67-7c2d-4081-9a53-e4f43f61dcd6] Deleted contents of the VM from datastore datastore2 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1191.276184] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-a00ae9d0-fb69-476d-95d1-c304afd40cef tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: dc112e67-7c2d-4081-9a53-e4f43f61dcd6] Instance destroyed {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1191.276368] env[65726]: INFO nova.compute.manager [None req-a00ae9d0-fb69-476d-95d1-c304afd40cef tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: dc112e67-7c2d-4081-9a53-e4f43f61dcd6] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1191.276624] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-a00ae9d0-fb69-476d-95d1-c304afd40cef tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1191.276825] env[65726]: DEBUG nova.compute.manager [-] [instance: dc112e67-7c2d-4081-9a53-e4f43f61dcd6] Deallocating network for instance {{(pid=65726) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1191.276924] env[65726]: DEBUG nova.network.neutron [-] [instance: dc112e67-7c2d-4081-9a53-e4f43f61dcd6] deallocate_for_instance() {{(pid=65726) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 1191.277452] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1191.277707] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1191.316875] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1191.317390] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1191.405954] env[65726]: DEBUG nova.compute.manager [req-1d719559-7e9a-4332-a810-b858c3f518cf req-4bdeb34f-6047-4f5b-9434-27f0c349514a service nova] [instance: 6956bda7-5657-45d7-8f80-c6809bd836b7] Received event network-changed-dad01f9d-0dc4-483c-8e27-cf7ffeec9519 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1191.405954] env[65726]: DEBUG nova.compute.manager [req-1d719559-7e9a-4332-a810-b858c3f518cf req-4bdeb34f-6047-4f5b-9434-27f0c349514a service nova] [instance: 6956bda7-5657-45d7-8f80-c6809bd836b7] Refreshing instance network info cache due to event network-changed-dad01f9d-0dc4-483c-8e27-cf7ffeec9519. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 1191.406113] env[65726]: DEBUG oslo_concurrency.lockutils [req-1d719559-7e9a-4332-a810-b858c3f518cf req-4bdeb34f-6047-4f5b-9434-27f0c349514a service nova] Acquiring lock "refresh_cache-6956bda7-5657-45d7-8f80-c6809bd836b7" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1191.406271] env[65726]: DEBUG oslo_concurrency.lockutils [req-1d719559-7e9a-4332-a810-b858c3f518cf req-4bdeb34f-6047-4f5b-9434-27f0c349514a service nova] Acquired lock "refresh_cache-6956bda7-5657-45d7-8f80-c6809bd836b7" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1191.406526] env[65726]: DEBUG nova.network.neutron [req-1d719559-7e9a-4332-a810-b858c3f518cf req-4bdeb34f-6047-4f5b-9434-27f0c349514a service nova] [instance: 6956bda7-5657-45d7-8f80-c6809bd836b7] Refreshing network info cache for port dad01f9d-0dc4-483c-8e27-cf7ffeec9519 {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 1191.448955] env[65726]: DEBUG oslo_vmware.api [None req-acbda108-3421-436d-befd-ccf16d53d3c3 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Task: {'id': task-5116919, 'name': ReconfigVM_Task, 'duration_secs': 0.202592} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1191.448955] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-acbda108-3421-436d-befd-ccf16d53d3c3 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: 3bc25953-21b1-4729-af09-e7211fd8b2c6] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-995319', 'volume_id': 'b58c28a5-1144-47d9-96ea-c0a59a979c77', 'name': 'volume-b58c28a5-1144-47d9-96ea-c0a59a979c77', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '3bc25953-21b1-4729-af09-e7211fd8b2c6', 'attached_at': '', 'detached_at': '', 'volume_id': 'b58c28a5-1144-47d9-96ea-c0a59a979c77', 'serial': 'b58c28a5-1144-47d9-96ea-c0a59a979c77'} {{(pid=65726) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1191.466239] env[65726]: DEBUG oslo_vmware.api [None req-3b1f8e43-10c3-447e-a11b-a1499eca83f6 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': task-5116920, 'name': PowerOnVM_Task, 'duration_secs': 0.468366} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1191.466423] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-3b1f8e43-10c3-447e-a11b-a1499eca83f6 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 55ee97c3-b014-48eb-b41b-fc5aa16bc09e] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1191.466638] env[65726]: INFO nova.compute.manager [None req-3b1f8e43-10c3-447e-a11b-a1499eca83f6 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 55ee97c3-b014-48eb-b41b-fc5aa16bc09e] Took 10.73 seconds to spawn the instance on the hypervisor. [ 1191.466838] env[65726]: DEBUG nova.compute.manager [None req-3b1f8e43-10c3-447e-a11b-a1499eca83f6 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 55ee97c3-b014-48eb-b41b-fc5aa16bc09e] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1191.467662] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb8eaeb6-bbf9-4652-af12-763d77ff8f22 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1191.505539] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b7bd70fc-78fa-4df3-bb13-d10a1172d2a5 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.269s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1191.507770] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.368s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1191.507770] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1191.507770] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=65726) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:937}} [ 1191.508795] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aae9f087-18cc-4a0c-9ae6-f046a3846b65 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1191.518104] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd4af73d-be22-43ac-a02d-86135b0be464 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1191.534698] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afe44cbb-12e2-4e8f-8b23-f37c2587d6c5 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1191.543829] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c267b658-fab0-4cae-bedc-43158d249863 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1191.580748] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=178865MB free_disk=95GB free_vcpus=48 pci_devices=None {{(pid=65726) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1136}} [ 1191.580938] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1191.581178] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1191.587371] env[65726]: DEBUG nova.compute.manager [req-fc1354a1-d257-4a9f-b3b0-698ef9c9c913 req-2b9d7c91-679f-493a-969a-e28ba05ce6c7 service nova] [instance: dc112e67-7c2d-4081-9a53-e4f43f61dcd6] Received event network-vif-deleted-46483adf-29bb-4abf-b767-42ff3c0b3c20 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1191.587558] env[65726]: INFO nova.compute.manager [req-fc1354a1-d257-4a9f-b3b0-698ef9c9c913 req-2b9d7c91-679f-493a-969a-e28ba05ce6c7 service nova] [instance: dc112e67-7c2d-4081-9a53-e4f43f61dcd6] Neutron deleted interface 46483adf-29bb-4abf-b767-42ff3c0b3c20; detaching it from the instance and deleting it from the info cache [ 1191.587722] env[65726]: DEBUG nova.network.neutron [req-fc1354a1-d257-4a9f-b3b0-698ef9c9c913 req-2b9d7c91-679f-493a-969a-e28ba05ce6c7 service nova] [instance: dc112e67-7c2d-4081-9a53-e4f43f61dcd6] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1191.656751] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116921, 'name': CreateVM_Task, 'duration_secs': 0.340084} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1191.656933] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6956bda7-5657-45d7-8f80-c6809bd836b7] Created VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1191.657715] env[65726]: WARNING openstack [None req-b0eb83ae-2fbb-47ee-abe7-62f2ae1b54da tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1191.658089] env[65726]: WARNING openstack [None req-b0eb83ae-2fbb-47ee-abe7-62f2ae1b54da tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1191.663446] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b0eb83ae-2fbb-47ee-abe7-62f2ae1b54da tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1191.663635] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b0eb83ae-2fbb-47ee-abe7-62f2ae1b54da tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1191.664207] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b0eb83ae-2fbb-47ee-abe7-62f2ae1b54da tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1191.664323] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-486f6edb-dd36-4958-8c7b-97a889730c1c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1191.669699] env[65726]: DEBUG oslo_vmware.api [None req-b0eb83ae-2fbb-47ee-abe7-62f2ae1b54da tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Waiting for the task: (returnval){ [ 1191.669699] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52981599-dc51-6fd0-8dd3-12f0454d2f04" [ 1191.669699] env[65726]: _type = "Task" [ 1191.669699] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1191.678291] env[65726]: DEBUG oslo_vmware.api [None req-b0eb83ae-2fbb-47ee-abe7-62f2ae1b54da tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52981599-dc51-6fd0-8dd3-12f0454d2f04, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1191.726619] env[65726]: DEBUG oslo_concurrency.lockutils [None req-70bb45c2-ad53-4082-9238-425c98dcfe4d tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1191.909582] env[65726]: WARNING openstack [req-1d719559-7e9a-4332-a810-b858c3f518cf req-4bdeb34f-6047-4f5b-9434-27f0c349514a service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1191.909979] env[65726]: WARNING openstack [req-1d719559-7e9a-4332-a810-b858c3f518cf req-4bdeb34f-6047-4f5b-9434-27f0c349514a service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1191.986238] env[65726]: INFO nova.compute.manager [None req-3b1f8e43-10c3-447e-a11b-a1499eca83f6 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 55ee97c3-b014-48eb-b41b-fc5aa16bc09e] Took 17.59 seconds to build instance. [ 1192.003660] env[65726]: DEBUG nova.objects.instance [None req-acbda108-3421-436d-befd-ccf16d53d3c3 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Lazy-loading 'flavor' on Instance uuid 3bc25953-21b1-4729-af09-e7211fd8b2c6 {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1192.015422] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b7bd70fc-78fa-4df3-bb13-d10a1172d2a5 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Lock "378c1d18-4bb6-4245-80bf-370fb7af0575" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 24.798s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1192.016391] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e32abc12-7404-4281-87f0-12de5cc1731f tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Lock "378c1d18-4bb6-4245-80bf-370fb7af0575" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 2.151s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1192.016623] env[65726]: INFO nova.compute.manager [None req-e32abc12-7404-4281-87f0-12de5cc1731f tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 378c1d18-4bb6-4245-80bf-370fb7af0575] Unshelving [ 1192.024378] env[65726]: WARNING openstack [req-1d719559-7e9a-4332-a810-b858c3f518cf req-4bdeb34f-6047-4f5b-9434-27f0c349514a service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1192.024819] env[65726]: WARNING openstack [req-1d719559-7e9a-4332-a810-b858c3f518cf req-4bdeb34f-6047-4f5b-9434-27f0c349514a service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1192.086039] env[65726]: DEBUG nova.network.neutron [-] [instance: dc112e67-7c2d-4081-9a53-e4f43f61dcd6] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1192.093088] env[65726]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e2035258-381d-4187-b5c6-eead598b8e2a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.105038] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37016b64-f239-44d7-99df-bb439ca5d451 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.123715] env[65726]: WARNING openstack [req-1d719559-7e9a-4332-a810-b858c3f518cf req-4bdeb34f-6047-4f5b-9434-27f0c349514a service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1192.125112] env[65726]: WARNING openstack [req-1d719559-7e9a-4332-a810-b858c3f518cf req-4bdeb34f-6047-4f5b-9434-27f0c349514a service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1192.143614] env[65726]: DEBUG nova.compute.manager [req-fc1354a1-d257-4a9f-b3b0-698ef9c9c913 req-2b9d7c91-679f-493a-969a-e28ba05ce6c7 service nova] [instance: dc112e67-7c2d-4081-9a53-e4f43f61dcd6] Detach interface failed, port_id=46483adf-29bb-4abf-b767-42ff3c0b3c20, reason: Instance dc112e67-7c2d-4081-9a53-e4f43f61dcd6 could not be found. {{(pid=65726) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11669}} [ 1192.184577] env[65726]: DEBUG oslo_vmware.api [None req-b0eb83ae-2fbb-47ee-abe7-62f2ae1b54da tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52981599-dc51-6fd0-8dd3-12f0454d2f04, 'name': SearchDatastore_Task, 'duration_secs': 0.038922} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1192.184577] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b0eb83ae-2fbb-47ee-abe7-62f2ae1b54da tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1192.184577] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-b0eb83ae-2fbb-47ee-abe7-62f2ae1b54da tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 6956bda7-5657-45d7-8f80-c6809bd836b7] Processing image b52362a3-ee8a-4cbf-b06f-513b0cc8f95c {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1192.184577] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b0eb83ae-2fbb-47ee-abe7-62f2ae1b54da tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1192.184577] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b0eb83ae-2fbb-47ee-abe7-62f2ae1b54da tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1192.184577] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-b0eb83ae-2fbb-47ee-abe7-62f2ae1b54da tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1192.184577] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7431bdaa-92d1-45f9-94c4-85ec01a0be70 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.198911] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-b0eb83ae-2fbb-47ee-abe7-62f2ae1b54da tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1192.199141] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-b0eb83ae-2fbb-47ee-abe7-62f2ae1b54da tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65726) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1192.200872] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9b3a6adf-9bae-445e-a37f-dddd4ef0a39b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.207639] env[65726]: DEBUG oslo_vmware.api [None req-b0eb83ae-2fbb-47ee-abe7-62f2ae1b54da tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Waiting for the task: (returnval){ [ 1192.207639] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]528a7cab-2736-426e-6731-873a527d6ca6" [ 1192.207639] env[65726]: _type = "Task" [ 1192.207639] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1192.219876] env[65726]: DEBUG oslo_vmware.api [None req-b0eb83ae-2fbb-47ee-abe7-62f2ae1b54da tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]528a7cab-2736-426e-6731-873a527d6ca6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1192.225921] env[65726]: DEBUG nova.network.neutron [req-1d719559-7e9a-4332-a810-b858c3f518cf req-4bdeb34f-6047-4f5b-9434-27f0c349514a service nova] [instance: 6956bda7-5657-45d7-8f80-c6809bd836b7] Updated VIF entry in instance network info cache for port dad01f9d-0dc4-483c-8e27-cf7ffeec9519. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 1192.226385] env[65726]: DEBUG nova.network.neutron [req-1d719559-7e9a-4332-a810-b858c3f518cf req-4bdeb34f-6047-4f5b-9434-27f0c349514a service nova] [instance: 6956bda7-5657-45d7-8f80-c6809bd836b7] Updating instance_info_cache with network_info: [{"id": "dad01f9d-0dc4-483c-8e27-cf7ffeec9519", "address": "fa:16:3e:17:36:eb", "network": {"id": "9dd55fb6-0dd2-484d-b94d-fbdccb958c79", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1005598846-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f93d852e2c904f42981cbfff882558e9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "74f30339-6421-4654-bddb-81d7f34db9d7", "external-id": "nsx-vlan-transportzone-899", "segmentation_id": 899, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdad01f9d-0d", "ovs_interfaceid": "dad01f9d-0dc4-483c-8e27-cf7ffeec9519", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1192.488492] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3b1f8e43-10c3-447e-a11b-a1499eca83f6 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Lock "55ee97c3-b014-48eb-b41b-fc5aa16bc09e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 19.103s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1192.593717] env[65726]: INFO nova.compute.manager [-] [instance: dc112e67-7c2d-4081-9a53-e4f43f61dcd6] Took 1.32 seconds to deallocate network for instance. [ 1192.615241] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Instance 3bc25953-21b1-4729-af09-e7211fd8b2c6 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65726) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1192.615399] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Instance 5bfbaf25-5aed-46bc-97fc-f138d08ebbde actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65726) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1192.615521] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Instance 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30 actively managed on this compute host and has allocations in placement: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65726) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1192.615636] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Instance 9c0599e3-383f-46ab-809b-944cc3a4d206 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65726) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1192.615747] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Instance dc112e67-7c2d-4081-9a53-e4f43f61dcd6 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65726) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1192.615858] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Instance e3255f3b-028f-4a0b-b621-873730417936 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65726) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1192.615966] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Instance 55ee97c3-b014-48eb-b41b-fc5aa16bc09e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65726) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1192.616126] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Instance 6956bda7-5657-45d7-8f80-c6809bd836b7 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65726) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1192.718296] env[65726]: DEBUG oslo_vmware.api [None req-b0eb83ae-2fbb-47ee-abe7-62f2ae1b54da tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]528a7cab-2736-426e-6731-873a527d6ca6, 'name': SearchDatastore_Task, 'duration_secs': 0.042151} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1192.719151] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3efa3a5b-ff8c-4649-8e20-ce6455edcb6b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.724419] env[65726]: DEBUG oslo_vmware.api [None req-b0eb83ae-2fbb-47ee-abe7-62f2ae1b54da tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Waiting for the task: (returnval){ [ 1192.724419] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52d60c61-ef2f-0b18-ca4a-814388a52af6" [ 1192.724419] env[65726]: _type = "Task" [ 1192.724419] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1192.729251] env[65726]: DEBUG oslo_concurrency.lockutils [req-1d719559-7e9a-4332-a810-b858c3f518cf req-4bdeb34f-6047-4f5b-9434-27f0c349514a service nova] Releasing lock "refresh_cache-6956bda7-5657-45d7-8f80-c6809bd836b7" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1192.729541] env[65726]: DEBUG nova.compute.manager [req-1d719559-7e9a-4332-a810-b858c3f518cf req-4bdeb34f-6047-4f5b-9434-27f0c349514a service nova] [instance: e3255f3b-028f-4a0b-b621-873730417936] Received event network-changed-03c9e6fb-2435-43be-bb55-8afecdcf38ff {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1192.729745] env[65726]: DEBUG nova.compute.manager [req-1d719559-7e9a-4332-a810-b858c3f518cf req-4bdeb34f-6047-4f5b-9434-27f0c349514a service nova] [instance: e3255f3b-028f-4a0b-b621-873730417936] Refreshing instance network info cache due to event network-changed-03c9e6fb-2435-43be-bb55-8afecdcf38ff. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 1192.730010] env[65726]: DEBUG oslo_concurrency.lockutils [req-1d719559-7e9a-4332-a810-b858c3f518cf req-4bdeb34f-6047-4f5b-9434-27f0c349514a service nova] Acquiring lock "refresh_cache-e3255f3b-028f-4a0b-b621-873730417936" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1192.730224] env[65726]: DEBUG oslo_concurrency.lockutils [req-1d719559-7e9a-4332-a810-b858c3f518cf req-4bdeb34f-6047-4f5b-9434-27f0c349514a service nova] Acquired lock "refresh_cache-e3255f3b-028f-4a0b-b621-873730417936" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1192.730458] env[65726]: DEBUG nova.network.neutron [req-1d719559-7e9a-4332-a810-b858c3f518cf req-4bdeb34f-6047-4f5b-9434-27f0c349514a service nova] [instance: e3255f3b-028f-4a0b-b621-873730417936] Refreshing network info cache for port 03c9e6fb-2435-43be-bb55-8afecdcf38ff {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 1192.735598] env[65726]: DEBUG oslo_vmware.api [None req-b0eb83ae-2fbb-47ee-abe7-62f2ae1b54da tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52d60c61-ef2f-0b18-ca4a-814388a52af6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1193.011659] env[65726]: DEBUG oslo_concurrency.lockutils [None req-acbda108-3421-436d-befd-ccf16d53d3c3 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Lock "3bc25953-21b1-4729-af09-e7211fd8b2c6" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.780s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1193.039628] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e32abc12-7404-4281-87f0-12de5cc1731f tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1193.104498] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a00ae9d0-fb69-476d-95d1-c304afd40cef tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1193.119576] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Instance 378c1d18-4bb6-4245-80bf-370fb7af0575 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65726) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 1193.237517] env[65726]: WARNING openstack [req-1d719559-7e9a-4332-a810-b858c3f518cf req-4bdeb34f-6047-4f5b-9434-27f0c349514a service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1193.237967] env[65726]: WARNING openstack [req-1d719559-7e9a-4332-a810-b858c3f518cf req-4bdeb34f-6047-4f5b-9434-27f0c349514a service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1193.246127] env[65726]: DEBUG oslo_vmware.api [None req-b0eb83ae-2fbb-47ee-abe7-62f2ae1b54da tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52d60c61-ef2f-0b18-ca4a-814388a52af6, 'name': SearchDatastore_Task, 'duration_secs': 0.010369} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1193.246702] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b0eb83ae-2fbb-47ee-abe7-62f2ae1b54da tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1193.247025] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-b0eb83ae-2fbb-47ee-abe7-62f2ae1b54da tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] 6956bda7-5657-45d7-8f80-c6809bd836b7/6956bda7-5657-45d7-8f80-c6809bd836b7.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1193.247388] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-edb381a9-2172-458a-93c4-52d6a248b9f2 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.256784] env[65726]: DEBUG oslo_vmware.api [None req-b0eb83ae-2fbb-47ee-abe7-62f2ae1b54da tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Waiting for the task: (returnval){ [ 1193.256784] env[65726]: value = "task-5116922" [ 1193.256784] env[65726]: _type = "Task" [ 1193.256784] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1193.269389] env[65726]: DEBUG oslo_vmware.api [None req-b0eb83ae-2fbb-47ee-abe7-62f2ae1b54da tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': task-5116922, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1193.364083] env[65726]: WARNING openstack [req-1d719559-7e9a-4332-a810-b858c3f518cf req-4bdeb34f-6047-4f5b-9434-27f0c349514a service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1193.364512] env[65726]: WARNING openstack [req-1d719559-7e9a-4332-a810-b858c3f518cf req-4bdeb34f-6047-4f5b-9434-27f0c349514a service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1193.438760] env[65726]: WARNING openstack [req-1d719559-7e9a-4332-a810-b858c3f518cf req-4bdeb34f-6047-4f5b-9434-27f0c349514a service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1193.439242] env[65726]: WARNING openstack [req-1d719559-7e9a-4332-a810-b858c3f518cf req-4bdeb34f-6047-4f5b-9434-27f0c349514a service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1193.477550] env[65726]: DEBUG nova.compute.manager [req-14d19ca5-4c01-4ae7-a865-edb2920615e4 req-d7da183e-29e6-455d-8157-7e978f6d62e3 service nova] [instance: 55ee97c3-b014-48eb-b41b-fc5aa16bc09e] Received event network-changed-65a8678e-a808-4eb1-9cb2-2f86b5fbd4b9 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1193.477990] env[65726]: DEBUG nova.compute.manager [req-14d19ca5-4c01-4ae7-a865-edb2920615e4 req-d7da183e-29e6-455d-8157-7e978f6d62e3 service nova] [instance: 55ee97c3-b014-48eb-b41b-fc5aa16bc09e] Refreshing instance network info cache due to event network-changed-65a8678e-a808-4eb1-9cb2-2f86b5fbd4b9. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 1193.478569] env[65726]: DEBUG oslo_concurrency.lockutils [req-14d19ca5-4c01-4ae7-a865-edb2920615e4 req-d7da183e-29e6-455d-8157-7e978f6d62e3 service nova] Acquiring lock "refresh_cache-55ee97c3-b014-48eb-b41b-fc5aa16bc09e" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1193.478569] env[65726]: DEBUG oslo_concurrency.lockutils [req-14d19ca5-4c01-4ae7-a865-edb2920615e4 req-d7da183e-29e6-455d-8157-7e978f6d62e3 service nova] Acquired lock "refresh_cache-55ee97c3-b014-48eb-b41b-fc5aa16bc09e" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1193.478868] env[65726]: DEBUG nova.network.neutron [req-14d19ca5-4c01-4ae7-a865-edb2920615e4 req-d7da183e-29e6-455d-8157-7e978f6d62e3 service nova] [instance: 55ee97c3-b014-48eb-b41b-fc5aa16bc09e] Refreshing network info cache for port 65a8678e-a808-4eb1-9cb2-2f86b5fbd4b9 {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 1193.541866] env[65726]: DEBUG nova.network.neutron [req-1d719559-7e9a-4332-a810-b858c3f518cf req-4bdeb34f-6047-4f5b-9434-27f0c349514a service nova] [instance: e3255f3b-028f-4a0b-b621-873730417936] Updated VIF entry in instance network info cache for port 03c9e6fb-2435-43be-bb55-8afecdcf38ff. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 1193.542391] env[65726]: DEBUG nova.network.neutron [req-1d719559-7e9a-4332-a810-b858c3f518cf req-4bdeb34f-6047-4f5b-9434-27f0c349514a service nova] [instance: e3255f3b-028f-4a0b-b621-873730417936] Updating instance_info_cache with network_info: [{"id": "03c9e6fb-2435-43be-bb55-8afecdcf38ff", "address": "fa:16:3e:d2:0f:96", "network": {"id": "a899453d-5f43-4650-9642-b33f54a4d2eb", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-614861139-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.231", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "acd800e0438940bc80ae3c6b672db4ee", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "205fb402-8eaf-4b61-8f57-8f216024179a", "external-id": "nsx-vlan-transportzone-78", "segmentation_id": 78, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap03c9e6fb-24", "ovs_interfaceid": "03c9e6fb-2435-43be-bb55-8afecdcf38ff", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1193.623164] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Instance b353b5e9-500c-42d3-a87f-880a624febec has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65726) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 1193.623164] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Total usable vcpus: 48, total allocated vcpus: 8 {{(pid=65726) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1159}} [ 1193.623164] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2048MB phys_disk=100GB used_disk=7GB total_vcpus=48 used_vcpus=8 pci_stats=[] stats={'failed_builds': '0', 'num_instances': '8', 'num_vm_active': '7', 'num_task_None': '6', 'num_os_type_None': '8', 'num_proj_a090d553766847e58e3231b966c92565': '1', 'io_workload': '1', 'num_proj_4ffd45f4a7a041199a4fc7f69f5e7e9b': '1', 'num_proj_1e3203c117de4d4d9c4c90436801ff3f': '1', 'num_proj_f3f799aa94f64e8cb26d93f6124efd81': '2', 'num_task_deleting': '1', 'num_proj_acd800e0438940bc80ae3c6b672db4ee': '1', 'num_proj_6b56f93ba0e14521921484ac2c785c43': '1', 'num_vm_building': '1', 'num_task_spawning': '1', 'num_proj_f93d852e2c904f42981cbfff882558e9': '1'} {{(pid=65726) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1168}} [ 1193.770642] env[65726]: DEBUG oslo_vmware.api [None req-b0eb83ae-2fbb-47ee-abe7-62f2ae1b54da tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': task-5116922, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1193.801530] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-169478dd-6506-46b1-b0b7-d3f50502b8e6 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.811029] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c536af1-85be-4c8c-9c64-2084ffffecc9 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.845386] env[65726]: DEBUG oslo_concurrency.lockutils [None req-7cd2a85c-e262-4e84-b25f-a094169151c3 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Acquiring lock "3bc25953-21b1-4729-af09-e7211fd8b2c6" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1193.845633] env[65726]: DEBUG oslo_concurrency.lockutils [None req-7cd2a85c-e262-4e84-b25f-a094169151c3 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Lock "3bc25953-21b1-4729-af09-e7211fd8b2c6" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1193.845850] env[65726]: DEBUG oslo_concurrency.lockutils [None req-7cd2a85c-e262-4e84-b25f-a094169151c3 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Acquiring lock "3bc25953-21b1-4729-af09-e7211fd8b2c6-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1193.846130] env[65726]: DEBUG oslo_concurrency.lockutils [None req-7cd2a85c-e262-4e84-b25f-a094169151c3 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Lock "3bc25953-21b1-4729-af09-e7211fd8b2c6-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1193.846311] env[65726]: DEBUG oslo_concurrency.lockutils [None req-7cd2a85c-e262-4e84-b25f-a094169151c3 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Lock "3bc25953-21b1-4729-af09-e7211fd8b2c6-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1193.849039] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f6027af-c1a0-4f14-bf0f-1f299bd40920 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.852557] env[65726]: INFO nova.compute.manager [None req-7cd2a85c-e262-4e84-b25f-a094169151c3 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: 3bc25953-21b1-4729-af09-e7211fd8b2c6] Terminating instance [ 1193.860984] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa548b42-a3a4-48d0-9555-96669d4951c4 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.877204] env[65726]: DEBUG nova.compute.provider_tree [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1193.982991] env[65726]: WARNING openstack [req-14d19ca5-4c01-4ae7-a865-edb2920615e4 req-d7da183e-29e6-455d-8157-7e978f6d62e3 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1193.983599] env[65726]: WARNING openstack [req-14d19ca5-4c01-4ae7-a865-edb2920615e4 req-d7da183e-29e6-455d-8157-7e978f6d62e3 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1194.045706] env[65726]: DEBUG oslo_concurrency.lockutils [req-1d719559-7e9a-4332-a810-b858c3f518cf req-4bdeb34f-6047-4f5b-9434-27f0c349514a service nova] Releasing lock "refresh_cache-e3255f3b-028f-4a0b-b621-873730417936" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1194.117495] env[65726]: WARNING openstack [req-14d19ca5-4c01-4ae7-a865-edb2920615e4 req-d7da183e-29e6-455d-8157-7e978f6d62e3 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1194.118792] env[65726]: WARNING openstack [req-14d19ca5-4c01-4ae7-a865-edb2920615e4 req-d7da183e-29e6-455d-8157-7e978f6d62e3 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1194.191545] env[65726]: WARNING openstack [req-14d19ca5-4c01-4ae7-a865-edb2920615e4 req-d7da183e-29e6-455d-8157-7e978f6d62e3 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1194.191957] env[65726]: WARNING openstack [req-14d19ca5-4c01-4ae7-a865-edb2920615e4 req-d7da183e-29e6-455d-8157-7e978f6d62e3 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1194.270419] env[65726]: DEBUG oslo_vmware.api [None req-b0eb83ae-2fbb-47ee-abe7-62f2ae1b54da tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': task-5116922, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.819149} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1194.270728] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-b0eb83ae-2fbb-47ee-abe7-62f2ae1b54da tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] 6956bda7-5657-45d7-8f80-c6809bd836b7/6956bda7-5657-45d7-8f80-c6809bd836b7.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1194.270905] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-b0eb83ae-2fbb-47ee-abe7-62f2ae1b54da tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 6956bda7-5657-45d7-8f80-c6809bd836b7] Extending root virtual disk to 1048576 {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1194.271190] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e780a2ca-beaf-4846-9cd7-915249ad3b06 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1194.279624] env[65726]: DEBUG oslo_vmware.api [None req-b0eb83ae-2fbb-47ee-abe7-62f2ae1b54da tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Waiting for the task: (returnval){ [ 1194.279624] env[65726]: value = "task-5116923" [ 1194.279624] env[65726]: _type = "Task" [ 1194.279624] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1194.282095] env[65726]: DEBUG nova.network.neutron [req-14d19ca5-4c01-4ae7-a865-edb2920615e4 req-d7da183e-29e6-455d-8157-7e978f6d62e3 service nova] [instance: 55ee97c3-b014-48eb-b41b-fc5aa16bc09e] Updated VIF entry in instance network info cache for port 65a8678e-a808-4eb1-9cb2-2f86b5fbd4b9. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 1194.282521] env[65726]: DEBUG nova.network.neutron [req-14d19ca5-4c01-4ae7-a865-edb2920615e4 req-d7da183e-29e6-455d-8157-7e978f6d62e3 service nova] [instance: 55ee97c3-b014-48eb-b41b-fc5aa16bc09e] Updating instance_info_cache with network_info: [{"id": "65a8678e-a808-4eb1-9cb2-2f86b5fbd4b9", "address": "fa:16:3e:ba:7b:9c", "network": {"id": "336fe51f-971a-447d-8e68-505640e0db1b", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1030314703-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.218", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6b56f93ba0e14521921484ac2c785c43", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8ee1c76d-1a61-4546-85cb-d4bd3c1b35ef", "external-id": "nsx-vlan-transportzone-161", "segmentation_id": 161, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap65a8678e-a8", "ovs_interfaceid": "65a8678e-a808-4eb1-9cb2-2f86b5fbd4b9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1194.294075] env[65726]: DEBUG oslo_vmware.api [None req-b0eb83ae-2fbb-47ee-abe7-62f2ae1b54da tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': task-5116923, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1194.357949] env[65726]: DEBUG nova.compute.manager [None req-7cd2a85c-e262-4e84-b25f-a094169151c3 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: 3bc25953-21b1-4729-af09-e7211fd8b2c6] Start destroying the instance on the hypervisor. {{(pid=65726) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1194.358238] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-7cd2a85c-e262-4e84-b25f-a094169151c3 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: 3bc25953-21b1-4729-af09-e7211fd8b2c6] Destroying instance {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1194.359187] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-156f5efa-f7d1-40f8-8dd8-bf90f8d5feda {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1194.370799] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-7cd2a85c-e262-4e84-b25f-a094169151c3 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: 3bc25953-21b1-4729-af09-e7211fd8b2c6] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1194.371342] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-379b268a-6cca-4469-8738-42a1f929e37b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1194.378769] env[65726]: DEBUG oslo_vmware.api [None req-7cd2a85c-e262-4e84-b25f-a094169151c3 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Waiting for the task: (returnval){ [ 1194.378769] env[65726]: value = "task-5116924" [ 1194.378769] env[65726]: _type = "Task" [ 1194.378769] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1194.383544] env[65726]: DEBUG nova.scheduler.client.report [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1194.396582] env[65726]: DEBUG oslo_vmware.api [None req-7cd2a85c-e262-4e84-b25f-a094169151c3 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Task: {'id': task-5116924, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1194.790355] env[65726]: DEBUG oslo_concurrency.lockutils [req-14d19ca5-4c01-4ae7-a865-edb2920615e4 req-d7da183e-29e6-455d-8157-7e978f6d62e3 service nova] Releasing lock "refresh_cache-55ee97c3-b014-48eb-b41b-fc5aa16bc09e" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1194.790898] env[65726]: DEBUG oslo_vmware.api [None req-b0eb83ae-2fbb-47ee-abe7-62f2ae1b54da tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': task-5116923, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.103609} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1194.791288] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-b0eb83ae-2fbb-47ee-abe7-62f2ae1b54da tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 6956bda7-5657-45d7-8f80-c6809bd836b7] Extended root virtual disk {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1194.792355] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1be3ad17-90cf-4724-852b-099dbdeb31ee {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1194.818766] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-b0eb83ae-2fbb-47ee-abe7-62f2ae1b54da tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 6956bda7-5657-45d7-8f80-c6809bd836b7] Reconfiguring VM instance instance-00000071 to attach disk [datastore1] 6956bda7-5657-45d7-8f80-c6809bd836b7/6956bda7-5657-45d7-8f80-c6809bd836b7.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1194.819127] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-53370ed3-3dbb-4ef0-9335-15a4d79edbf6 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1194.841043] env[65726]: DEBUG oslo_vmware.api [None req-b0eb83ae-2fbb-47ee-abe7-62f2ae1b54da tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Waiting for the task: (returnval){ [ 1194.841043] env[65726]: value = "task-5116925" [ 1194.841043] env[65726]: _type = "Task" [ 1194.841043] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1194.850774] env[65726]: DEBUG oslo_vmware.api [None req-b0eb83ae-2fbb-47ee-abe7-62f2ae1b54da tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': task-5116925, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1194.888763] env[65726]: DEBUG oslo_vmware.api [None req-7cd2a85c-e262-4e84-b25f-a094169151c3 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Task: {'id': task-5116924, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1194.891696] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=65726) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1097}} [ 1194.891934] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 3.311s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1194.892220] env[65726]: DEBUG oslo_concurrency.lockutils [None req-70bb45c2-ad53-4082-9238-425c98dcfe4d tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.166s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1194.892868] env[65726]: DEBUG nova.objects.instance [None req-70bb45c2-ad53-4082-9238-425c98dcfe4d tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Lazy-loading 'pci_requests' on Instance uuid b353b5e9-500c-42d3-a87f-880a624febec {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1195.352180] env[65726]: DEBUG oslo_vmware.api [None req-b0eb83ae-2fbb-47ee-abe7-62f2ae1b54da tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': task-5116925, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1195.389419] env[65726]: DEBUG oslo_vmware.api [None req-7cd2a85c-e262-4e84-b25f-a094169151c3 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Task: {'id': task-5116924, 'name': PowerOffVM_Task, 'duration_secs': 0.694143} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1195.390250] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-7cd2a85c-e262-4e84-b25f-a094169151c3 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: 3bc25953-21b1-4729-af09-e7211fd8b2c6] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1195.390250] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-7cd2a85c-e262-4e84-b25f-a094169151c3 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: 3bc25953-21b1-4729-af09-e7211fd8b2c6] Unregistering the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1195.390457] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3c6abdb6-1e81-48bc-b734-e822fe393a43 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1195.396922] env[65726]: DEBUG nova.objects.instance [None req-70bb45c2-ad53-4082-9238-425c98dcfe4d tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Lazy-loading 'numa_topology' on Instance uuid b353b5e9-500c-42d3-a87f-880a624febec {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1195.465475] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-7cd2a85c-e262-4e84-b25f-a094169151c3 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: 3bc25953-21b1-4729-af09-e7211fd8b2c6] Unregistered the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1195.465724] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-7cd2a85c-e262-4e84-b25f-a094169151c3 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: 3bc25953-21b1-4729-af09-e7211fd8b2c6] Deleting contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1195.465871] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-7cd2a85c-e262-4e84-b25f-a094169151c3 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Deleting the datastore file [datastore1] 3bc25953-21b1-4729-af09-e7211fd8b2c6 {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1195.466218] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-05ec626f-8946-492b-b520-4b5f050c63a9 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1195.473802] env[65726]: DEBUG oslo_vmware.api [None req-7cd2a85c-e262-4e84-b25f-a094169151c3 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Waiting for the task: (returnval){ [ 1195.473802] env[65726]: value = "task-5116927" [ 1195.473802] env[65726]: _type = "Task" [ 1195.473802] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1195.483780] env[65726]: DEBUG oslo_vmware.api [None req-7cd2a85c-e262-4e84-b25f-a094169151c3 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Task: {'id': task-5116927, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1195.853054] env[65726]: DEBUG oslo_vmware.api [None req-b0eb83ae-2fbb-47ee-abe7-62f2ae1b54da tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': task-5116925, 'name': ReconfigVM_Task, 'duration_secs': 0.513738} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1195.853373] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-b0eb83ae-2fbb-47ee-abe7-62f2ae1b54da tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 6956bda7-5657-45d7-8f80-c6809bd836b7] Reconfigured VM instance instance-00000071 to attach disk [datastore1] 6956bda7-5657-45d7-8f80-c6809bd836b7/6956bda7-5657-45d7-8f80-c6809bd836b7.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1195.854088] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9aec1e94-5a2a-4ce3-a8fb-287f2a185a12 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1195.861266] env[65726]: DEBUG oslo_vmware.api [None req-b0eb83ae-2fbb-47ee-abe7-62f2ae1b54da tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Waiting for the task: (returnval){ [ 1195.861266] env[65726]: value = "task-5116928" [ 1195.861266] env[65726]: _type = "Task" [ 1195.861266] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1195.870201] env[65726]: DEBUG oslo_vmware.api [None req-b0eb83ae-2fbb-47ee-abe7-62f2ae1b54da tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': task-5116928, 'name': Rename_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1195.900076] env[65726]: INFO nova.compute.claims [None req-70bb45c2-ad53-4082-9238-425c98dcfe4d tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: b353b5e9-500c-42d3-a87f-880a624febec] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1195.987056] env[65726]: DEBUG oslo_vmware.api [None req-7cd2a85c-e262-4e84-b25f-a094169151c3 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Task: {'id': task-5116927, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.305375} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1195.987327] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-7cd2a85c-e262-4e84-b25f-a094169151c3 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Deleted the datastore file {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1195.987506] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-7cd2a85c-e262-4e84-b25f-a094169151c3 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: 3bc25953-21b1-4729-af09-e7211fd8b2c6] Deleted contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1195.987678] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-7cd2a85c-e262-4e84-b25f-a094169151c3 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: 3bc25953-21b1-4729-af09-e7211fd8b2c6] Instance destroyed {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1195.987844] env[65726]: INFO nova.compute.manager [None req-7cd2a85c-e262-4e84-b25f-a094169151c3 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] [instance: 3bc25953-21b1-4729-af09-e7211fd8b2c6] Took 1.63 seconds to destroy the instance on the hypervisor. [ 1195.988114] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-7cd2a85c-e262-4e84-b25f-a094169151c3 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1195.988315] env[65726]: DEBUG nova.compute.manager [-] [instance: 3bc25953-21b1-4729-af09-e7211fd8b2c6] Deallocating network for instance {{(pid=65726) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1195.988416] env[65726]: DEBUG nova.network.neutron [-] [instance: 3bc25953-21b1-4729-af09-e7211fd8b2c6] deallocate_for_instance() {{(pid=65726) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 1195.988939] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1195.989188] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1196.062282] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1196.062282] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1196.148465] env[65726]: DEBUG oslo_concurrency.lockutils [None req-4360b5c9-45f8-4acf-8cf6-3673d2897518 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Acquiring lock "5bfbaf25-5aed-46bc-97fc-f138d08ebbde" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1196.149286] env[65726]: DEBUG oslo_concurrency.lockutils [None req-4360b5c9-45f8-4acf-8cf6-3673d2897518 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Lock "5bfbaf25-5aed-46bc-97fc-f138d08ebbde" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.001s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1196.372058] env[65726]: DEBUG oslo_vmware.api [None req-b0eb83ae-2fbb-47ee-abe7-62f2ae1b54da tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': task-5116928, 'name': Rename_Task, 'duration_secs': 0.150611} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1196.372460] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-b0eb83ae-2fbb-47ee-abe7-62f2ae1b54da tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 6956bda7-5657-45d7-8f80-c6809bd836b7] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1196.373039] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-008134dc-b036-4b25-aa16-0335afa9065e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.384275] env[65726]: DEBUG oslo_vmware.api [None req-b0eb83ae-2fbb-47ee-abe7-62f2ae1b54da tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Waiting for the task: (returnval){ [ 1196.384275] env[65726]: value = "task-5116929" [ 1196.384275] env[65726]: _type = "Task" [ 1196.384275] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1196.392676] env[65726]: DEBUG oslo_vmware.api [None req-b0eb83ae-2fbb-47ee-abe7-62f2ae1b54da tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': task-5116929, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1196.518118] env[65726]: DEBUG nova.compute.manager [req-850b5410-d951-4efb-8696-0f8b74c5ad8a req-ad7dd834-7d01-4f84-83a3-027e7418db41 service nova] [instance: 3bc25953-21b1-4729-af09-e7211fd8b2c6] Received event network-vif-deleted-a7e1795d-67d0-4f32-baea-744806988e4b {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1196.518369] env[65726]: INFO nova.compute.manager [req-850b5410-d951-4efb-8696-0f8b74c5ad8a req-ad7dd834-7d01-4f84-83a3-027e7418db41 service nova] [instance: 3bc25953-21b1-4729-af09-e7211fd8b2c6] Neutron deleted interface a7e1795d-67d0-4f32-baea-744806988e4b; detaching it from the instance and deleting it from the info cache [ 1196.518541] env[65726]: DEBUG nova.network.neutron [req-850b5410-d951-4efb-8696-0f8b74c5ad8a req-ad7dd834-7d01-4f84-83a3-027e7418db41 service nova] [instance: 3bc25953-21b1-4729-af09-e7211fd8b2c6] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1196.653350] env[65726]: DEBUG nova.compute.utils [None req-4360b5c9-45f8-4acf-8cf6-3673d2897518 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Using /dev/sd instead of None {{(pid=65726) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1196.894281] env[65726]: DEBUG oslo_vmware.api [None req-b0eb83ae-2fbb-47ee-abe7-62f2ae1b54da tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': task-5116929, 'name': PowerOnVM_Task, 'duration_secs': 0.511454} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1196.894572] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-b0eb83ae-2fbb-47ee-abe7-62f2ae1b54da tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 6956bda7-5657-45d7-8f80-c6809bd836b7] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1196.894776] env[65726]: INFO nova.compute.manager [None req-b0eb83ae-2fbb-47ee-abe7-62f2ae1b54da tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 6956bda7-5657-45d7-8f80-c6809bd836b7] Took 7.88 seconds to spawn the instance on the hypervisor. [ 1196.894950] env[65726]: DEBUG nova.compute.manager [None req-b0eb83ae-2fbb-47ee-abe7-62f2ae1b54da tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 6956bda7-5657-45d7-8f80-c6809bd836b7] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1196.895794] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2410e63-9a16-4145-968f-d1baa5c3eb89 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.993962] env[65726]: DEBUG nova.network.neutron [-] [instance: 3bc25953-21b1-4729-af09-e7211fd8b2c6] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1197.021964] env[65726]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9183d21d-7fc8-4357-9ced-b01fcacac2e1 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1197.034026] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9121a62-3877-4bdf-92d2-0b76fff77cf4 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1197.074513] env[65726]: DEBUG nova.compute.manager [req-850b5410-d951-4efb-8696-0f8b74c5ad8a req-ad7dd834-7d01-4f84-83a3-027e7418db41 service nova] [instance: 3bc25953-21b1-4729-af09-e7211fd8b2c6] Detach interface failed, port_id=a7e1795d-67d0-4f32-baea-744806988e4b, reason: Instance 3bc25953-21b1-4729-af09-e7211fd8b2c6 could not be found. {{(pid=65726) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11669}} [ 1197.076066] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fca4a490-9322-4f89-83fd-15a4b0573a85 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1197.084195] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ae09ead-a4c9-4e00-a6c6-ee31840eb553 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1197.114593] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e10cc817-816d-4847-8d82-0c121e14593b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1197.123914] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb715c6a-c0ec-450c-a3a5-ce98ed934e52 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1197.137959] env[65726]: DEBUG nova.compute.provider_tree [None req-70bb45c2-ad53-4082-9238-425c98dcfe4d tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1197.157349] env[65726]: DEBUG oslo_concurrency.lockutils [None req-4360b5c9-45f8-4acf-8cf6-3673d2897518 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Lock "5bfbaf25-5aed-46bc-97fc-f138d08ebbde" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.008s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1197.415392] env[65726]: INFO nova.compute.manager [None req-b0eb83ae-2fbb-47ee-abe7-62f2ae1b54da tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 6956bda7-5657-45d7-8f80-c6809bd836b7] Took 12.77 seconds to build instance. [ 1197.499404] env[65726]: INFO nova.compute.manager [-] [instance: 3bc25953-21b1-4729-af09-e7211fd8b2c6] Took 1.51 seconds to deallocate network for instance. [ 1197.641345] env[65726]: DEBUG nova.scheduler.client.report [None req-70bb45c2-ad53-4082-9238-425c98dcfe4d tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1197.917685] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b0eb83ae-2fbb-47ee-abe7-62f2ae1b54da tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Lock "6956bda7-5657-45d7-8f80-c6809bd836b7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.276s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1198.007735] env[65726]: DEBUG oslo_concurrency.lockutils [None req-7cd2a85c-e262-4e84-b25f-a094169151c3 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1198.147231] env[65726]: DEBUG oslo_concurrency.lockutils [None req-70bb45c2-ad53-4082-9238-425c98dcfe4d tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.255s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1198.147848] env[65726]: WARNING openstack [None req-70bb45c2-ad53-4082-9238-425c98dcfe4d tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1198.148217] env[65726]: WARNING openstack [None req-70bb45c2-ad53-4082-9238-425c98dcfe4d tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1198.155954] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e32abc12-7404-4281-87f0-12de5cc1731f tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.116s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1198.155954] env[65726]: DEBUG nova.objects.instance [None req-e32abc12-7404-4281-87f0-12de5cc1731f tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Lazy-loading 'pci_requests' on Instance uuid 378c1d18-4bb6-4245-80bf-370fb7af0575 {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1198.223272] env[65726]: DEBUG oslo_concurrency.lockutils [None req-4360b5c9-45f8-4acf-8cf6-3673d2897518 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Acquiring lock "5bfbaf25-5aed-46bc-97fc-f138d08ebbde" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1198.223559] env[65726]: DEBUG oslo_concurrency.lockutils [None req-4360b5c9-45f8-4acf-8cf6-3673d2897518 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Lock "5bfbaf25-5aed-46bc-97fc-f138d08ebbde" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1198.225113] env[65726]: INFO nova.compute.manager [None req-4360b5c9-45f8-4acf-8cf6-3673d2897518 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] [instance: 5bfbaf25-5aed-46bc-97fc-f138d08ebbde] Attaching volume a86c4f74-2f48-4ced-aa11-e174e767f2a7 to /dev/sdb [ 1198.232275] env[65726]: WARNING openstack [None req-70bb45c2-ad53-4082-9238-425c98dcfe4d tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1198.232647] env[65726]: WARNING openstack [None req-70bb45c2-ad53-4082-9238-425c98dcfe4d tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1198.310035] env[65726]: INFO nova.network.neutron [None req-70bb45c2-ad53-4082-9238-425c98dcfe4d tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: b353b5e9-500c-42d3-a87f-880a624febec] Updating port 08b7016b-dc44-4206-a00c-5da943b82a38 with attributes {'binding_host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 1198.315009] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73c2ca8e-159b-4b1c-aeaa-43c257cdc027 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1198.325034] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2970e766-3f60-4d67-a0a1-d0b22dc9df93 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1198.339539] env[65726]: DEBUG nova.virt.block_device [None req-4360b5c9-45f8-4acf-8cf6-3673d2897518 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] [instance: 5bfbaf25-5aed-46bc-97fc-f138d08ebbde] Updating existing volume attachment record: 0a6e989d-34a9-4ee4-aa25-e152cf8ab09a {{(pid=65726) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1198.662156] env[65726]: DEBUG nova.objects.instance [None req-e32abc12-7404-4281-87f0-12de5cc1731f tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Lazy-loading 'numa_topology' on Instance uuid 378c1d18-4bb6-4245-80bf-370fb7af0575 {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1199.055505] env[65726]: DEBUG oslo_concurrency.lockutils [None req-68748ea0-9eb9-4593-95d7-667ae1802835 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Acquiring lock "6956bda7-5657-45d7-8f80-c6809bd836b7" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1199.055791] env[65726]: DEBUG oslo_concurrency.lockutils [None req-68748ea0-9eb9-4593-95d7-667ae1802835 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Lock "6956bda7-5657-45d7-8f80-c6809bd836b7" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1199.056167] env[65726]: DEBUG oslo_concurrency.lockutils [None req-68748ea0-9eb9-4593-95d7-667ae1802835 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Acquiring lock "6956bda7-5657-45d7-8f80-c6809bd836b7-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1199.056376] env[65726]: DEBUG oslo_concurrency.lockutils [None req-68748ea0-9eb9-4593-95d7-667ae1802835 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Lock "6956bda7-5657-45d7-8f80-c6809bd836b7-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1199.056474] env[65726]: DEBUG oslo_concurrency.lockutils [None req-68748ea0-9eb9-4593-95d7-667ae1802835 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Lock "6956bda7-5657-45d7-8f80-c6809bd836b7-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1199.058710] env[65726]: INFO nova.compute.manager [None req-68748ea0-9eb9-4593-95d7-667ae1802835 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 6956bda7-5657-45d7-8f80-c6809bd836b7] Terminating instance [ 1199.165065] env[65726]: INFO nova.compute.claims [None req-e32abc12-7404-4281-87f0-12de5cc1731f tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 378c1d18-4bb6-4245-80bf-370fb7af0575] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1199.515421] env[65726]: DEBUG oslo_service.periodic_task [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Running periodic task ComputeManager._sync_power_states {{(pid=65726) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1199.563205] env[65726]: DEBUG nova.compute.manager [None req-68748ea0-9eb9-4593-95d7-667ae1802835 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 6956bda7-5657-45d7-8f80-c6809bd836b7] Start destroying the instance on the hypervisor. {{(pid=65726) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1199.563466] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-68748ea0-9eb9-4593-95d7-667ae1802835 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 6956bda7-5657-45d7-8f80-c6809bd836b7] Destroying instance {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1199.564480] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab052114-d4b1-43cd-afda-2eb5a1d8aa43 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1199.573439] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-68748ea0-9eb9-4593-95d7-667ae1802835 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 6956bda7-5657-45d7-8f80-c6809bd836b7] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1199.573711] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-aef552a5-f41b-419e-b45e-7bc264019949 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1199.582188] env[65726]: DEBUG oslo_vmware.api [None req-68748ea0-9eb9-4593-95d7-667ae1802835 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Waiting for the task: (returnval){ [ 1199.582188] env[65726]: value = "task-5116931" [ 1199.582188] env[65726]: _type = "Task" [ 1199.582188] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1199.593470] env[65726]: DEBUG oslo_vmware.api [None req-68748ea0-9eb9-4593-95d7-667ae1802835 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': task-5116931, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1199.845766] env[65726]: DEBUG nova.compute.manager [req-8ac68b2c-fa98-4d94-9171-f034916f9b61 req-6804e7d4-38ab-456d-b2e6-fa3fdcb94e76 service nova] [instance: b353b5e9-500c-42d3-a87f-880a624febec] Received event network-vif-plugged-08b7016b-dc44-4206-a00c-5da943b82a38 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1199.846166] env[65726]: DEBUG oslo_concurrency.lockutils [req-8ac68b2c-fa98-4d94-9171-f034916f9b61 req-6804e7d4-38ab-456d-b2e6-fa3fdcb94e76 service nova] Acquiring lock "b353b5e9-500c-42d3-a87f-880a624febec-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1199.846216] env[65726]: DEBUG oslo_concurrency.lockutils [req-8ac68b2c-fa98-4d94-9171-f034916f9b61 req-6804e7d4-38ab-456d-b2e6-fa3fdcb94e76 service nova] Lock "b353b5e9-500c-42d3-a87f-880a624febec-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1199.846374] env[65726]: DEBUG oslo_concurrency.lockutils [req-8ac68b2c-fa98-4d94-9171-f034916f9b61 req-6804e7d4-38ab-456d-b2e6-fa3fdcb94e76 service nova] Lock "b353b5e9-500c-42d3-a87f-880a624febec-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1199.846540] env[65726]: DEBUG nova.compute.manager [req-8ac68b2c-fa98-4d94-9171-f034916f9b61 req-6804e7d4-38ab-456d-b2e6-fa3fdcb94e76 service nova] [instance: b353b5e9-500c-42d3-a87f-880a624febec] No waiting events found dispatching network-vif-plugged-08b7016b-dc44-4206-a00c-5da943b82a38 {{(pid=65726) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1199.846733] env[65726]: WARNING nova.compute.manager [req-8ac68b2c-fa98-4d94-9171-f034916f9b61 req-6804e7d4-38ab-456d-b2e6-fa3fdcb94e76 service nova] [instance: b353b5e9-500c-42d3-a87f-880a624febec] Received unexpected event network-vif-plugged-08b7016b-dc44-4206-a00c-5da943b82a38 for instance with vm_state shelved_offloaded and task_state spawning. [ 1199.906214] env[65726]: DEBUG oslo_concurrency.lockutils [None req-70bb45c2-ad53-4082-9238-425c98dcfe4d tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Acquiring lock "refresh_cache-b353b5e9-500c-42d3-a87f-880a624febec" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1199.906331] env[65726]: DEBUG oslo_concurrency.lockutils [None req-70bb45c2-ad53-4082-9238-425c98dcfe4d tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Acquired lock "refresh_cache-b353b5e9-500c-42d3-a87f-880a624febec" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1199.906466] env[65726]: DEBUG nova.network.neutron [None req-70bb45c2-ad53-4082-9238-425c98dcfe4d tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: b353b5e9-500c-42d3-a87f-880a624febec] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1200.020558] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Getting list of instances from cluster (obj){ [ 1200.020558] env[65726]: value = "domain-c8" [ 1200.020558] env[65726]: _type = "ClusterComputeResource" [ 1200.020558] env[65726]: } {{(pid=65726) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 1200.021643] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48fb8630-a507-40bf-ac3b-10836e06b8c9 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1200.039133] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Got total of 6 instances {{(pid=65726) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 1200.039312] env[65726]: WARNING nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] While synchronizing instance power states, found 10 instances in the database and 6 instances on the hypervisor. [ 1200.039447] env[65726]: DEBUG nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Triggering sync for uuid b353b5e9-500c-42d3-a87f-880a624febec {{(pid=65726) _sync_power_states /opt/stack/nova/nova/compute/manager.py:11043}} [ 1200.039741] env[65726]: DEBUG nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Triggering sync for uuid 378c1d18-4bb6-4245-80bf-370fb7af0575 {{(pid=65726) _sync_power_states /opt/stack/nova/nova/compute/manager.py:11043}} [ 1200.039958] env[65726]: DEBUG nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Triggering sync for uuid 3bc25953-21b1-4729-af09-e7211fd8b2c6 {{(pid=65726) _sync_power_states /opt/stack/nova/nova/compute/manager.py:11043}} [ 1200.040185] env[65726]: DEBUG nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Triggering sync for uuid 5bfbaf25-5aed-46bc-97fc-f138d08ebbde {{(pid=65726) _sync_power_states /opt/stack/nova/nova/compute/manager.py:11043}} [ 1200.040385] env[65726]: DEBUG nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Triggering sync for uuid 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30 {{(pid=65726) _sync_power_states /opt/stack/nova/nova/compute/manager.py:11043}} [ 1200.040583] env[65726]: DEBUG nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Triggering sync for uuid 9c0599e3-383f-46ab-809b-944cc3a4d206 {{(pid=65726) _sync_power_states /opt/stack/nova/nova/compute/manager.py:11043}} [ 1200.040778] env[65726]: DEBUG nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Triggering sync for uuid dc112e67-7c2d-4081-9a53-e4f43f61dcd6 {{(pid=65726) _sync_power_states /opt/stack/nova/nova/compute/manager.py:11043}} [ 1200.040999] env[65726]: DEBUG nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Triggering sync for uuid e3255f3b-028f-4a0b-b621-873730417936 {{(pid=65726) _sync_power_states /opt/stack/nova/nova/compute/manager.py:11043}} [ 1200.041229] env[65726]: DEBUG nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Triggering sync for uuid 55ee97c3-b014-48eb-b41b-fc5aa16bc09e {{(pid=65726) _sync_power_states /opt/stack/nova/nova/compute/manager.py:11043}} [ 1200.041427] env[65726]: DEBUG nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Triggering sync for uuid 6956bda7-5657-45d7-8f80-c6809bd836b7 {{(pid=65726) _sync_power_states /opt/stack/nova/nova/compute/manager.py:11043}} [ 1200.041857] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Acquiring lock "b353b5e9-500c-42d3-a87f-880a624febec" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1200.042146] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Acquiring lock "378c1d18-4bb6-4245-80bf-370fb7af0575" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1200.042369] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Acquiring lock "3bc25953-21b1-4729-af09-e7211fd8b2c6" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1200.042579] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Acquiring lock "5bfbaf25-5aed-46bc-97fc-f138d08ebbde" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1200.042786] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Acquiring lock "47e52bc1-c18e-4aa2-82a2-e4cb030a7a30" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1200.042967] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Lock "47e52bc1-c18e-4aa2-82a2-e4cb030a7a30" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1200.043322] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Acquiring lock "9c0599e3-383f-46ab-809b-944cc3a4d206" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1200.043462] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Lock "9c0599e3-383f-46ab-809b-944cc3a4d206" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1200.043701] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Acquiring lock "dc112e67-7c2d-4081-9a53-e4f43f61dcd6" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1200.043980] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Acquiring lock "e3255f3b-028f-4a0b-b621-873730417936" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1200.044206] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Lock "e3255f3b-028f-4a0b-b621-873730417936" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1200.044444] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Acquiring lock "55ee97c3-b014-48eb-b41b-fc5aa16bc09e" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1200.044614] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Lock "55ee97c3-b014-48eb-b41b-fc5aa16bc09e" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1200.044836] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Acquiring lock "6956bda7-5657-45d7-8f80-c6809bd836b7" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1200.045676] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f450f9f1-8503-4002-8aa6-c092b4c9c740 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1200.048936] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6611f48f-7a2d-40d7-8fd7-61a54f9b8be5 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1200.051656] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7172f25f-acc1-429b-b2c7-32c51237a720 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1200.054451] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffb33957-aaa3-4614-8702-349e0b014193 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1200.094157] env[65726]: DEBUG oslo_vmware.api [None req-68748ea0-9eb9-4593-95d7-667ae1802835 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': task-5116931, 'name': PowerOffVM_Task, 'duration_secs': 0.204422} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1200.094433] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-68748ea0-9eb9-4593-95d7-667ae1802835 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 6956bda7-5657-45d7-8f80-c6809bd836b7] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1200.094600] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-68748ea0-9eb9-4593-95d7-667ae1802835 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 6956bda7-5657-45d7-8f80-c6809bd836b7] Unregistering the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1200.095374] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d4c1b6d8-2e57-4fec-8e2c-675059cb03e0 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1200.098999] env[65726]: DEBUG oslo_service.periodic_task [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=65726) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1200.162945] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-68748ea0-9eb9-4593-95d7-667ae1802835 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 6956bda7-5657-45d7-8f80-c6809bd836b7] Unregistered the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1200.163125] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-68748ea0-9eb9-4593-95d7-667ae1802835 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 6956bda7-5657-45d7-8f80-c6809bd836b7] Deleting contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1200.163232] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-68748ea0-9eb9-4593-95d7-667ae1802835 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Deleting the datastore file [datastore1] 6956bda7-5657-45d7-8f80-c6809bd836b7 {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1200.163526] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-fb987c53-b515-4e86-a16d-22eab472d02c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1200.175341] env[65726]: DEBUG oslo_vmware.api [None req-68748ea0-9eb9-4593-95d7-667ae1802835 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Waiting for the task: (returnval){ [ 1200.175341] env[65726]: value = "task-5116933" [ 1200.175341] env[65726]: _type = "Task" [ 1200.175341] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1200.185259] env[65726]: DEBUG oslo_vmware.api [None req-68748ea0-9eb9-4593-95d7-667ae1802835 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': task-5116933, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1200.319299] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a487ee8-3435-4cf0-b8e3-cee889e0f0a7 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1200.327853] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d76dea2c-8d97-428d-979d-6a0c6f994952 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1200.364930] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f1ca25c-436b-4ff1-ad4f-d74ee27fd695 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1200.373471] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76af742e-2d81-4a18-a4f1-6fef4f54b0ce {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1200.388117] env[65726]: DEBUG nova.compute.provider_tree [None req-e32abc12-7404-4281-87f0-12de5cc1731f tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1200.409532] env[65726]: WARNING openstack [None req-70bb45c2-ad53-4082-9238-425c98dcfe4d tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1200.409892] env[65726]: WARNING openstack [None req-70bb45c2-ad53-4082-9238-425c98dcfe4d tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1200.532649] env[65726]: WARNING openstack [None req-70bb45c2-ad53-4082-9238-425c98dcfe4d tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1200.533127] env[65726]: WARNING openstack [None req-70bb45c2-ad53-4082-9238-425c98dcfe4d tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1200.570178] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Lock "9c0599e3-383f-46ab-809b-944cc3a4d206" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.526s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1200.570884] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Lock "47e52bc1-c18e-4aa2-82a2-e4cb030a7a30" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.528s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1200.571404] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Lock "55ee97c3-b014-48eb-b41b-fc5aa16bc09e" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.527s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1200.571897] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Lock "e3255f3b-028f-4a0b-b621-873730417936" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.528s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1200.602174] env[65726]: WARNING openstack [None req-70bb45c2-ad53-4082-9238-425c98dcfe4d tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1200.602585] env[65726]: WARNING openstack [None req-70bb45c2-ad53-4082-9238-425c98dcfe4d tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1200.686666] env[65726]: DEBUG oslo_vmware.api [None req-68748ea0-9eb9-4593-95d7-667ae1802835 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Task: {'id': task-5116933, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.194526} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1200.688985] env[65726]: DEBUG nova.network.neutron [None req-70bb45c2-ad53-4082-9238-425c98dcfe4d tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: b353b5e9-500c-42d3-a87f-880a624febec] Updating instance_info_cache with network_info: [{"id": "08b7016b-dc44-4206-a00c-5da943b82a38", "address": "fa:16:3e:c5:e4:14", "network": {"id": "eb3278b7-c190-441b-84d2-305f9b896382", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1754571022-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.144", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "96149159e18e44f9bf3453e67681f224", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f77ff7a1-209c-4f3f-b2a0-fd817741e739", "external-id": "nsx-vlan-transportzone-935", "segmentation_id": 935, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap08b7016b-dc", "ovs_interfaceid": "08b7016b-dc44-4206-a00c-5da943b82a38", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1200.690257] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-68748ea0-9eb9-4593-95d7-667ae1802835 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Deleted the datastore file {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1200.690448] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-68748ea0-9eb9-4593-95d7-667ae1802835 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 6956bda7-5657-45d7-8f80-c6809bd836b7] Deleted contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1200.690624] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-68748ea0-9eb9-4593-95d7-667ae1802835 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 6956bda7-5657-45d7-8f80-c6809bd836b7] Instance destroyed {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1200.690792] env[65726]: INFO nova.compute.manager [None req-68748ea0-9eb9-4593-95d7-667ae1802835 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] [instance: 6956bda7-5657-45d7-8f80-c6809bd836b7] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1200.691056] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-68748ea0-9eb9-4593-95d7-667ae1802835 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1200.691503] env[65726]: DEBUG nova.compute.manager [-] [instance: 6956bda7-5657-45d7-8f80-c6809bd836b7] Deallocating network for instance {{(pid=65726) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1200.691602] env[65726]: DEBUG nova.network.neutron [-] [instance: 6956bda7-5657-45d7-8f80-c6809bd836b7] deallocate_for_instance() {{(pid=65726) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 1200.692142] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1200.692405] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1200.730348] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1200.730633] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1200.892400] env[65726]: DEBUG nova.scheduler.client.report [None req-e32abc12-7404-4281-87f0-12de5cc1731f tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1201.192917] env[65726]: DEBUG oslo_concurrency.lockutils [None req-70bb45c2-ad53-4082-9238-425c98dcfe4d tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Releasing lock "refresh_cache-b353b5e9-500c-42d3-a87f-880a624febec" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1201.220227] env[65726]: DEBUG nova.virt.hardware [None req-70bb45c2-ad53-4082-9238-425c98dcfe4d tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='a6fb9397a6398f70388b0cd084b58819',container_format='bare',created_at=2025-12-12T19:40:27Z,direct_url=,disk_format='vmdk',id=745a44b6-0c3f-4396-a82a-6e9472e10925,min_disk=1,min_ram=0,name='tempest-AttachVolumeShelveTestJSON-server-1968981141-shelved',owner='96149159e18e44f9bf3453e67681f224',properties=ImageMetaProps,protected=,size=31669248,status='active',tags=,updated_at=2025-12-12T19:40:45Z,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1201.220583] env[65726]: DEBUG nova.virt.hardware [None req-70bb45c2-ad53-4082-9238-425c98dcfe4d tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1201.220682] env[65726]: DEBUG nova.virt.hardware [None req-70bb45c2-ad53-4082-9238-425c98dcfe4d tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1201.220860] env[65726]: DEBUG nova.virt.hardware [None req-70bb45c2-ad53-4082-9238-425c98dcfe4d tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1201.221012] env[65726]: DEBUG nova.virt.hardware [None req-70bb45c2-ad53-4082-9238-425c98dcfe4d tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1201.221193] env[65726]: DEBUG nova.virt.hardware [None req-70bb45c2-ad53-4082-9238-425c98dcfe4d tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1201.221363] env[65726]: DEBUG nova.virt.hardware [None req-70bb45c2-ad53-4082-9238-425c98dcfe4d tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1201.221507] env[65726]: DEBUG nova.virt.hardware [None req-70bb45c2-ad53-4082-9238-425c98dcfe4d tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1201.221670] env[65726]: DEBUG nova.virt.hardware [None req-70bb45c2-ad53-4082-9238-425c98dcfe4d tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1201.221895] env[65726]: DEBUG nova.virt.hardware [None req-70bb45c2-ad53-4082-9238-425c98dcfe4d tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1201.222103] env[65726]: DEBUG nova.virt.hardware [None req-70bb45c2-ad53-4082-9238-425c98dcfe4d tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1201.222969] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-383411a5-a4fb-4176-97ce-b274eba31314 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1201.232129] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b35cbb6d-8cfb-4ab1-b896-7ea12dcb6b6d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1201.246349] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-70bb45c2-ad53-4082-9238-425c98dcfe4d tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: b353b5e9-500c-42d3-a87f-880a624febec] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c5:e4:14', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f77ff7a1-209c-4f3f-b2a0-fd817741e739', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '08b7016b-dc44-4206-a00c-5da943b82a38', 'vif_model': 'vmxnet3'}] {{(pid=65726) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1201.254529] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-70bb45c2-ad53-4082-9238-425c98dcfe4d tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1201.254807] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b353b5e9-500c-42d3-a87f-880a624febec] Creating VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1201.255356] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f5aa915e-8214-4cf7-b143-62a77bc99474 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1201.274968] env[65726]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1201.274968] env[65726]: value = "task-5116935" [ 1201.274968] env[65726]: _type = "Task" [ 1201.274968] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1201.285984] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116935, 'name': CreateVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1201.398318] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e32abc12-7404-4281-87f0-12de5cc1731f tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.243s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1201.398969] env[65726]: WARNING openstack [None req-e32abc12-7404-4281-87f0-12de5cc1731f tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1201.399434] env[65726]: WARNING openstack [None req-e32abc12-7404-4281-87f0-12de5cc1731f tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1201.406529] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a00ae9d0-fb69-476d-95d1-c304afd40cef tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 8.302s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1201.406769] env[65726]: DEBUG nova.objects.instance [None req-a00ae9d0-fb69-476d-95d1-c304afd40cef tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Lazy-loading 'resources' on Instance uuid dc112e67-7c2d-4081-9a53-e4f43f61dcd6 {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1201.432232] env[65726]: DEBUG nova.network.neutron [-] [instance: 6956bda7-5657-45d7-8f80-c6809bd836b7] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1201.438548] env[65726]: INFO nova.network.neutron [None req-e32abc12-7404-4281-87f0-12de5cc1731f tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 378c1d18-4bb6-4245-80bf-370fb7af0575] Updating port 4765f6c2-d161-4e36-a6a8-3c8d39c47a40 with attributes {'binding_host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 1201.786633] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116935, 'name': CreateVM_Task, 'duration_secs': 0.365164} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1201.786838] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b353b5e9-500c-42d3-a87f-880a624febec] Created VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1201.787631] env[65726]: WARNING openstack [None req-70bb45c2-ad53-4082-9238-425c98dcfe4d tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1201.788026] env[65726]: WARNING openstack [None req-70bb45c2-ad53-4082-9238-425c98dcfe4d tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1201.794286] env[65726]: DEBUG oslo_concurrency.lockutils [None req-70bb45c2-ad53-4082-9238-425c98dcfe4d tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/745a44b6-0c3f-4396-a82a-6e9472e10925" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1201.794483] env[65726]: DEBUG oslo_concurrency.lockutils [None req-70bb45c2-ad53-4082-9238-425c98dcfe4d tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Acquired lock "[datastore1] devstack-image-cache_base/745a44b6-0c3f-4396-a82a-6e9472e10925" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1201.794861] env[65726]: DEBUG oslo_concurrency.lockutils [None req-70bb45c2-ad53-4082-9238-425c98dcfe4d tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/745a44b6-0c3f-4396-a82a-6e9472e10925" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1201.795153] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-477865ae-69b4-4312-9b3c-9ef07a07d5cb {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1201.800370] env[65726]: DEBUG oslo_vmware.api [None req-70bb45c2-ad53-4082-9238-425c98dcfe4d tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Waiting for the task: (returnval){ [ 1201.800370] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5257dc07-a476-a828-eaa5-89be4b5dad97" [ 1201.800370] env[65726]: _type = "Task" [ 1201.800370] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1201.808634] env[65726]: DEBUG oslo_vmware.api [None req-70bb45c2-ad53-4082-9238-425c98dcfe4d tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5257dc07-a476-a828-eaa5-89be4b5dad97, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1201.882725] env[65726]: DEBUG nova.compute.manager [req-5351fd90-eba8-43a8-856f-d26a62b84655 req-a4be1d0f-7ea3-4e30-8592-56899ecbdd7b service nova] [instance: b353b5e9-500c-42d3-a87f-880a624febec] Received event network-changed-08b7016b-dc44-4206-a00c-5da943b82a38 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1201.882925] env[65726]: DEBUG nova.compute.manager [req-5351fd90-eba8-43a8-856f-d26a62b84655 req-a4be1d0f-7ea3-4e30-8592-56899ecbdd7b service nova] [instance: b353b5e9-500c-42d3-a87f-880a624febec] Refreshing instance network info cache due to event network-changed-08b7016b-dc44-4206-a00c-5da943b82a38. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 1201.883162] env[65726]: DEBUG oslo_concurrency.lockutils [req-5351fd90-eba8-43a8-856f-d26a62b84655 req-a4be1d0f-7ea3-4e30-8592-56899ecbdd7b service nova] Acquiring lock "refresh_cache-b353b5e9-500c-42d3-a87f-880a624febec" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1201.883339] env[65726]: DEBUG oslo_concurrency.lockutils [req-5351fd90-eba8-43a8-856f-d26a62b84655 req-a4be1d0f-7ea3-4e30-8592-56899ecbdd7b service nova] Acquired lock "refresh_cache-b353b5e9-500c-42d3-a87f-880a624febec" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1201.883467] env[65726]: DEBUG nova.network.neutron [req-5351fd90-eba8-43a8-856f-d26a62b84655 req-a4be1d0f-7ea3-4e30-8592-56899ecbdd7b service nova] [instance: b353b5e9-500c-42d3-a87f-880a624febec] Refreshing network info cache for port 08b7016b-dc44-4206-a00c-5da943b82a38 {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 1201.935045] env[65726]: INFO nova.compute.manager [-] [instance: 6956bda7-5657-45d7-8f80-c6809bd836b7] Took 1.24 seconds to deallocate network for instance. [ 1202.156800] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25515f07-557c-432f-999e-b0bf9f03377c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1202.165255] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-272a89d3-804c-48d6-b8e4-f10f647f970d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1202.196066] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8c3843b-2df0-4bd2-ab52-876f1481ecd3 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1202.204275] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b17271a-4056-49f0-9a2a-91bc4c7c2537 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1202.220063] env[65726]: DEBUG nova.compute.provider_tree [None req-a00ae9d0-fb69-476d-95d1-c304afd40cef tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1202.313628] env[65726]: DEBUG oslo_concurrency.lockutils [None req-70bb45c2-ad53-4082-9238-425c98dcfe4d tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Releasing lock "[datastore1] devstack-image-cache_base/745a44b6-0c3f-4396-a82a-6e9472e10925" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1202.315050] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-70bb45c2-ad53-4082-9238-425c98dcfe4d tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: b353b5e9-500c-42d3-a87f-880a624febec] Processing image 745a44b6-0c3f-4396-a82a-6e9472e10925 {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1202.315050] env[65726]: DEBUG oslo_concurrency.lockutils [None req-70bb45c2-ad53-4082-9238-425c98dcfe4d tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/745a44b6-0c3f-4396-a82a-6e9472e10925/745a44b6-0c3f-4396-a82a-6e9472e10925.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1202.315050] env[65726]: DEBUG oslo_concurrency.lockutils [None req-70bb45c2-ad53-4082-9238-425c98dcfe4d tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Acquired lock "[datastore1] devstack-image-cache_base/745a44b6-0c3f-4396-a82a-6e9472e10925/745a44b6-0c3f-4396-a82a-6e9472e10925.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1202.315050] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-70bb45c2-ad53-4082-9238-425c98dcfe4d tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1202.315050] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-bb24f790-bb6d-4d35-9ace-f0f5740e5aeb {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1202.325820] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-70bb45c2-ad53-4082-9238-425c98dcfe4d tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1202.326053] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-70bb45c2-ad53-4082-9238-425c98dcfe4d tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65726) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1202.326868] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4ce5834e-6528-4942-910a-f8353dca6f03 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1202.333552] env[65726]: DEBUG oslo_vmware.api [None req-70bb45c2-ad53-4082-9238-425c98dcfe4d tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Waiting for the task: (returnval){ [ 1202.333552] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52cf6576-2635-f7c6-36ef-aa4327a55e08" [ 1202.333552] env[65726]: _type = "Task" [ 1202.333552] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1202.342597] env[65726]: DEBUG oslo_vmware.api [None req-70bb45c2-ad53-4082-9238-425c98dcfe4d tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52cf6576-2635-f7c6-36ef-aa4327a55e08, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1202.386413] env[65726]: WARNING openstack [req-5351fd90-eba8-43a8-856f-d26a62b84655 req-a4be1d0f-7ea3-4e30-8592-56899ecbdd7b service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1202.386831] env[65726]: WARNING openstack [req-5351fd90-eba8-43a8-856f-d26a62b84655 req-a4be1d0f-7ea3-4e30-8592-56899ecbdd7b service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1202.441669] env[65726]: DEBUG oslo_concurrency.lockutils [None req-68748ea0-9eb9-4593-95d7-667ae1802835 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1202.520248] env[65726]: WARNING openstack [req-5351fd90-eba8-43a8-856f-d26a62b84655 req-a4be1d0f-7ea3-4e30-8592-56899ecbdd7b service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1202.520742] env[65726]: WARNING openstack [req-5351fd90-eba8-43a8-856f-d26a62b84655 req-a4be1d0f-7ea3-4e30-8592-56899ecbdd7b service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1202.569526] env[65726]: DEBUG oslo_service.periodic_task [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=65726) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1202.569774] env[65726]: DEBUG oslo_service.periodic_task [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=65726) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1202.569935] env[65726]: DEBUG oslo_service.periodic_task [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=65726) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1202.587833] env[65726]: WARNING openstack [req-5351fd90-eba8-43a8-856f-d26a62b84655 req-a4be1d0f-7ea3-4e30-8592-56899ecbdd7b service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1202.588305] env[65726]: WARNING openstack [req-5351fd90-eba8-43a8-856f-d26a62b84655 req-a4be1d0f-7ea3-4e30-8592-56899ecbdd7b service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1202.691547] env[65726]: DEBUG nova.network.neutron [req-5351fd90-eba8-43a8-856f-d26a62b84655 req-a4be1d0f-7ea3-4e30-8592-56899ecbdd7b service nova] [instance: b353b5e9-500c-42d3-a87f-880a624febec] Updated VIF entry in instance network info cache for port 08b7016b-dc44-4206-a00c-5da943b82a38. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 1202.691964] env[65726]: DEBUG nova.network.neutron [req-5351fd90-eba8-43a8-856f-d26a62b84655 req-a4be1d0f-7ea3-4e30-8592-56899ecbdd7b service nova] [instance: b353b5e9-500c-42d3-a87f-880a624febec] Updating instance_info_cache with network_info: [{"id": "08b7016b-dc44-4206-a00c-5da943b82a38", "address": "fa:16:3e:c5:e4:14", "network": {"id": "eb3278b7-c190-441b-84d2-305f9b896382", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1754571022-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.144", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "96149159e18e44f9bf3453e67681f224", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f77ff7a1-209c-4f3f-b2a0-fd817741e739", "external-id": "nsx-vlan-transportzone-935", "segmentation_id": 935, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap08b7016b-dc", "ovs_interfaceid": "08b7016b-dc44-4206-a00c-5da943b82a38", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1202.723053] env[65726]: DEBUG nova.scheduler.client.report [None req-a00ae9d0-fb69-476d-95d1-c304afd40cef tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1202.844289] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-70bb45c2-ad53-4082-9238-425c98dcfe4d tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: b353b5e9-500c-42d3-a87f-880a624febec] Preparing fetch location {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 1202.844595] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-70bb45c2-ad53-4082-9238-425c98dcfe4d tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: b353b5e9-500c-42d3-a87f-880a624febec] Fetch image to [datastore1] OSTACK_IMG_b1f4d914-8fdf-4b4a-aba5-f24abe6a510f/OSTACK_IMG_b1f4d914-8fdf-4b4a-aba5-f24abe6a510f.vmdk {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 1202.844785] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-70bb45c2-ad53-4082-9238-425c98dcfe4d tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: b353b5e9-500c-42d3-a87f-880a624febec] Downloading stream optimized image 745a44b6-0c3f-4396-a82a-6e9472e10925 to [datastore1] OSTACK_IMG_b1f4d914-8fdf-4b4a-aba5-f24abe6a510f/OSTACK_IMG_b1f4d914-8fdf-4b4a-aba5-f24abe6a510f.vmdk on the data store datastore1 as vApp {{(pid=65726) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 1202.845069] env[65726]: DEBUG nova.virt.vmwareapi.images [None req-70bb45c2-ad53-4082-9238-425c98dcfe4d tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: b353b5e9-500c-42d3-a87f-880a624febec] Downloading image file data 745a44b6-0c3f-4396-a82a-6e9472e10925 to the ESX as VM named 'OSTACK_IMG_b1f4d914-8fdf-4b4a-aba5-f24abe6a510f' {{(pid=65726) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 1202.889260] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-4360b5c9-45f8-4acf-8cf6-3673d2897518 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] [instance: 5bfbaf25-5aed-46bc-97fc-f138d08ebbde] Volume attach. Driver type: vmdk {{(pid=65726) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1202.889600] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-4360b5c9-45f8-4acf-8cf6-3673d2897518 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] [instance: 5bfbaf25-5aed-46bc-97fc-f138d08ebbde] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-995328', 'volume_id': 'a86c4f74-2f48-4ced-aa11-e174e767f2a7', 'name': 'volume-a86c4f74-2f48-4ced-aa11-e174e767f2a7', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '5bfbaf25-5aed-46bc-97fc-f138d08ebbde', 'attached_at': '', 'detached_at': '', 'volume_id': 'a86c4f74-2f48-4ced-aa11-e174e767f2a7', 'serial': 'a86c4f74-2f48-4ced-aa11-e174e767f2a7'} {{(pid=65726) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1202.890519] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01586b46-ccf4-4c0d-b46b-c8e103df34eb {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1202.909699] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e7870e0-698c-4a6b-b7ea-6aa71dbc0382 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1202.937087] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-4360b5c9-45f8-4acf-8cf6-3673d2897518 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] [instance: 5bfbaf25-5aed-46bc-97fc-f138d08ebbde] Reconfiguring VM instance instance-0000006b to attach disk [datastore1] volume-a86c4f74-2f48-4ced-aa11-e174e767f2a7/volume-a86c4f74-2f48-4ced-aa11-e174e767f2a7.vmdk or device None with type thin {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1202.937493] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-325469da-2be9-4add-8c71-2158af363510 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1202.950952] env[65726]: DEBUG oslo_vmware.rw_handles [None req-70bb45c2-ad53-4082-9238-425c98dcfe4d tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 1202.950952] env[65726]: value = "resgroup-9" [ 1202.950952] env[65726]: _type = "ResourcePool" [ 1202.950952] env[65726]: }. {{(pid=65726) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 1202.951290] env[65726]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-a056543a-a4f4-4343-bac0-a71e64a55220 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1202.972640] env[65726]: DEBUG oslo_vmware.rw_handles [None req-70bb45c2-ad53-4082-9238-425c98dcfe4d tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Lease: (returnval){ [ 1202.972640] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]523ebc12-7f61-81a9-f7fc-798cf2c3a015" [ 1202.972640] env[65726]: _type = "HttpNfcLease" [ 1202.972640] env[65726]: } obtained for vApp import into resource pool (val){ [ 1202.972640] env[65726]: value = "resgroup-9" [ 1202.972640] env[65726]: _type = "ResourcePool" [ 1202.972640] env[65726]: }. {{(pid=65726) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 1202.972926] env[65726]: DEBUG oslo_vmware.api [None req-70bb45c2-ad53-4082-9238-425c98dcfe4d tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Waiting for the lease: (returnval){ [ 1202.972926] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]523ebc12-7f61-81a9-f7fc-798cf2c3a015" [ 1202.972926] env[65726]: _type = "HttpNfcLease" [ 1202.972926] env[65726]: } to be ready. {{(pid=65726) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1202.974398] env[65726]: DEBUG oslo_vmware.api [None req-4360b5c9-45f8-4acf-8cf6-3673d2897518 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Waiting for the task: (returnval){ [ 1202.974398] env[65726]: value = "task-5116936" [ 1202.974398] env[65726]: _type = "Task" [ 1202.974398] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1202.979990] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e32abc12-7404-4281-87f0-12de5cc1731f tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Acquiring lock "refresh_cache-378c1d18-4bb6-4245-80bf-370fb7af0575" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1202.979990] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e32abc12-7404-4281-87f0-12de5cc1731f tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Acquired lock "refresh_cache-378c1d18-4bb6-4245-80bf-370fb7af0575" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1202.979990] env[65726]: DEBUG nova.network.neutron [None req-e32abc12-7404-4281-87f0-12de5cc1731f tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 378c1d18-4bb6-4245-80bf-370fb7af0575] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1202.991581] env[65726]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1202.991581] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]523ebc12-7f61-81a9-f7fc-798cf2c3a015" [ 1202.991581] env[65726]: _type = "HttpNfcLease" [ 1202.991581] env[65726]: } is initializing. {{(pid=65726) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1202.995584] env[65726]: DEBUG oslo_vmware.api [None req-4360b5c9-45f8-4acf-8cf6-3673d2897518 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Task: {'id': task-5116936, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1203.196043] env[65726]: DEBUG oslo_concurrency.lockutils [req-5351fd90-eba8-43a8-856f-d26a62b84655 req-a4be1d0f-7ea3-4e30-8592-56899ecbdd7b service nova] Releasing lock "refresh_cache-b353b5e9-500c-42d3-a87f-880a624febec" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1203.196043] env[65726]: DEBUG nova.compute.manager [req-5351fd90-eba8-43a8-856f-d26a62b84655 req-a4be1d0f-7ea3-4e30-8592-56899ecbdd7b service nova] [instance: 6956bda7-5657-45d7-8f80-c6809bd836b7] Received event network-vif-deleted-dad01f9d-0dc4-483c-8e27-cf7ffeec9519 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1203.228709] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a00ae9d0-fb69-476d-95d1-c304afd40cef tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.822s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1203.231158] env[65726]: DEBUG oslo_concurrency.lockutils [None req-7cd2a85c-e262-4e84-b25f-a094169151c3 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 5.224s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1203.231448] env[65726]: DEBUG nova.objects.instance [None req-7cd2a85c-e262-4e84-b25f-a094169151c3 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Lazy-loading 'resources' on Instance uuid 3bc25953-21b1-4729-af09-e7211fd8b2c6 {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1203.253257] env[65726]: INFO nova.scheduler.client.report [None req-a00ae9d0-fb69-476d-95d1-c304afd40cef tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Deleted allocations for instance dc112e67-7c2d-4081-9a53-e4f43f61dcd6 [ 1203.487196] env[65726]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1203.487196] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]523ebc12-7f61-81a9-f7fc-798cf2c3a015" [ 1203.487196] env[65726]: _type = "HttpNfcLease" [ 1203.487196] env[65726]: } is initializing. {{(pid=65726) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1203.490831] env[65726]: WARNING openstack [None req-e32abc12-7404-4281-87f0-12de5cc1731f tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1203.492037] env[65726]: WARNING openstack [None req-e32abc12-7404-4281-87f0-12de5cc1731f tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1203.498355] env[65726]: DEBUG oslo_vmware.api [None req-4360b5c9-45f8-4acf-8cf6-3673d2897518 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Task: {'id': task-5116936, 'name': ReconfigVM_Task, 'duration_secs': 0.377324} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1203.498908] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-4360b5c9-45f8-4acf-8cf6-3673d2897518 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] [instance: 5bfbaf25-5aed-46bc-97fc-f138d08ebbde] Reconfigured VM instance instance-0000006b to attach disk [datastore1] volume-a86c4f74-2f48-4ced-aa11-e174e767f2a7/volume-a86c4f74-2f48-4ced-aa11-e174e767f2a7.vmdk or device None with type thin {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1203.503760] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-103414d9-51e4-4281-aa0c-44973df73091 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1203.521387] env[65726]: DEBUG oslo_vmware.api [None req-4360b5c9-45f8-4acf-8cf6-3673d2897518 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Waiting for the task: (returnval){ [ 1203.521387] env[65726]: value = "task-5116938" [ 1203.521387] env[65726]: _type = "Task" [ 1203.521387] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1203.531936] env[65726]: DEBUG oslo_vmware.api [None req-4360b5c9-45f8-4acf-8cf6-3673d2897518 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Task: {'id': task-5116938, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1203.564498] env[65726]: DEBUG oslo_service.periodic_task [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=65726) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1203.568535] env[65726]: DEBUG oslo_service.periodic_task [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=65726) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1203.627021] env[65726]: WARNING openstack [None req-e32abc12-7404-4281-87f0-12de5cc1731f tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1203.627444] env[65726]: WARNING openstack [None req-e32abc12-7404-4281-87f0-12de5cc1731f tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1203.697491] env[65726]: WARNING openstack [None req-e32abc12-7404-4281-87f0-12de5cc1731f tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1203.698160] env[65726]: WARNING openstack [None req-e32abc12-7404-4281-87f0-12de5cc1731f tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1203.763925] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a00ae9d0-fb69-476d-95d1-c304afd40cef tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Lock "dc112e67-7c2d-4081-9a53-e4f43f61dcd6" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 14.114s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1203.766568] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Lock "dc112e67-7c2d-4081-9a53-e4f43f61dcd6" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 3.723s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1203.766932] env[65726]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-02745c08-44ec-46b1-96cd-32f1d029b2cd {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1203.779148] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3f99461-5d70-494a-9263-4eba71ad09e1 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1203.795098] env[65726]: DEBUG nova.network.neutron [None req-e32abc12-7404-4281-87f0-12de5cc1731f tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 378c1d18-4bb6-4245-80bf-370fb7af0575] Updating instance_info_cache with network_info: [{"id": "4765f6c2-d161-4e36-a6a8-3c8d39c47a40", "address": "fa:16:3e:3c:38:b2", "network": {"id": "0d0f561d-ac41-4ef5-9926-c707409799f7", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1134321217-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.160", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4eece77569624f90bf64e5c51974173f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2ce62383-8e84-4e26-955b-74c11392f4c9", "external-id": "nsx-vlan-transportzone-215", "segmentation_id": 215, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4765f6c2-d1", "ovs_interfaceid": "4765f6c2-d161-4e36-a6a8-3c8d39c47a40", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1203.897130] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b494700-d65a-42d5-8347-6f382a3f92da {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1203.905423] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efe991db-f33e-462e-bd82-c72ead431f9f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1203.937620] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06f869dd-798b-422f-b165-9360554a8187 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1203.945790] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ebca102-fb87-4686-b3d9-2ca7c1ac6d46 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1203.952184] env[65726]: DEBUG nova.compute.manager [req-73a5274e-466f-4dfd-acde-2f557e7d23ee req-6f359cad-e907-4424-8a9c-b6873f8cfbd7 service nova] [instance: 378c1d18-4bb6-4245-80bf-370fb7af0575] Received event network-vif-plugged-4765f6c2-d161-4e36-a6a8-3c8d39c47a40 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1203.952397] env[65726]: DEBUG oslo_concurrency.lockutils [req-73a5274e-466f-4dfd-acde-2f557e7d23ee req-6f359cad-e907-4424-8a9c-b6873f8cfbd7 service nova] Acquiring lock "378c1d18-4bb6-4245-80bf-370fb7af0575-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1203.952604] env[65726]: DEBUG oslo_concurrency.lockutils [req-73a5274e-466f-4dfd-acde-2f557e7d23ee req-6f359cad-e907-4424-8a9c-b6873f8cfbd7 service nova] Lock "378c1d18-4bb6-4245-80bf-370fb7af0575-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1203.952748] env[65726]: DEBUG oslo_concurrency.lockutils [req-73a5274e-466f-4dfd-acde-2f557e7d23ee req-6f359cad-e907-4424-8a9c-b6873f8cfbd7 service nova] Lock "378c1d18-4bb6-4245-80bf-370fb7af0575-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1203.952913] env[65726]: DEBUG nova.compute.manager [req-73a5274e-466f-4dfd-acde-2f557e7d23ee req-6f359cad-e907-4424-8a9c-b6873f8cfbd7 service nova] [instance: 378c1d18-4bb6-4245-80bf-370fb7af0575] No waiting events found dispatching network-vif-plugged-4765f6c2-d161-4e36-a6a8-3c8d39c47a40 {{(pid=65726) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1203.953068] env[65726]: WARNING nova.compute.manager [req-73a5274e-466f-4dfd-acde-2f557e7d23ee req-6f359cad-e907-4424-8a9c-b6873f8cfbd7 service nova] [instance: 378c1d18-4bb6-4245-80bf-370fb7af0575] Received unexpected event network-vif-plugged-4765f6c2-d161-4e36-a6a8-3c8d39c47a40 for instance with vm_state shelved_offloaded and task_state spawning. [ 1203.953233] env[65726]: DEBUG nova.compute.manager [req-73a5274e-466f-4dfd-acde-2f557e7d23ee req-6f359cad-e907-4424-8a9c-b6873f8cfbd7 service nova] [instance: 378c1d18-4bb6-4245-80bf-370fb7af0575] Received event network-changed-4765f6c2-d161-4e36-a6a8-3c8d39c47a40 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1203.953491] env[65726]: DEBUG nova.compute.manager [req-73a5274e-466f-4dfd-acde-2f557e7d23ee req-6f359cad-e907-4424-8a9c-b6873f8cfbd7 service nova] [instance: 378c1d18-4bb6-4245-80bf-370fb7af0575] Refreshing instance network info cache due to event network-changed-4765f6c2-d161-4e36-a6a8-3c8d39c47a40. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 1203.953660] env[65726]: DEBUG oslo_concurrency.lockutils [req-73a5274e-466f-4dfd-acde-2f557e7d23ee req-6f359cad-e907-4424-8a9c-b6873f8cfbd7 service nova] Acquiring lock "refresh_cache-378c1d18-4bb6-4245-80bf-370fb7af0575" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1203.961908] env[65726]: DEBUG nova.compute.provider_tree [None req-7cd2a85c-e262-4e84-b25f-a094169151c3 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1203.985968] env[65726]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1203.985968] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]523ebc12-7f61-81a9-f7fc-798cf2c3a015" [ 1203.985968] env[65726]: _type = "HttpNfcLease" [ 1203.985968] env[65726]: } is ready. {{(pid=65726) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1203.986185] env[65726]: DEBUG oslo_vmware.rw_handles [None req-70bb45c2-ad53-4082-9238-425c98dcfe4d tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1203.986185] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]523ebc12-7f61-81a9-f7fc-798cf2c3a015" [ 1203.986185] env[65726]: _type = "HttpNfcLease" [ 1203.986185] env[65726]: }. {{(pid=65726) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 1203.986915] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98198aed-c95d-477e-a153-dc9ab1b590dd {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1203.995079] env[65726]: DEBUG oslo_vmware.rw_handles [None req-70bb45c2-ad53-4082-9238-425c98dcfe4d tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5290a352-77af-403f-8c95-fd94975727e2/disk-0.vmdk from lease info. {{(pid=65726) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1203.995248] env[65726]: DEBUG oslo_vmware.rw_handles [None req-70bb45c2-ad53-4082-9238-425c98dcfe4d tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Creating HTTP connection to write to file with size = 31669248 and URL = https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5290a352-77af-403f-8c95-fd94975727e2/disk-0.vmdk. {{(pid=65726) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 1204.066054] env[65726]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-9b734124-c99f-4d1b-8c13-b20e2d53c0ab {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1204.068262] env[65726]: DEBUG oslo_vmware.api [None req-4360b5c9-45f8-4acf-8cf6-3673d2897518 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Task: {'id': task-5116938, 'name': ReconfigVM_Task, 'duration_secs': 0.156778} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1204.068262] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-4360b5c9-45f8-4acf-8cf6-3673d2897518 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] [instance: 5bfbaf25-5aed-46bc-97fc-f138d08ebbde] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-995328', 'volume_id': 'a86c4f74-2f48-4ced-aa11-e174e767f2a7', 'name': 'volume-a86c4f74-2f48-4ced-aa11-e174e767f2a7', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '5bfbaf25-5aed-46bc-97fc-f138d08ebbde', 'attached_at': '', 'detached_at': '', 'volume_id': 'a86c4f74-2f48-4ced-aa11-e174e767f2a7', 'serial': 'a86c4f74-2f48-4ced-aa11-e174e767f2a7'} {{(pid=65726) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1204.297675] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e32abc12-7404-4281-87f0-12de5cc1731f tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Releasing lock "refresh_cache-378c1d18-4bb6-4245-80bf-370fb7af0575" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1204.301706] env[65726]: DEBUG oslo_concurrency.lockutils [req-73a5274e-466f-4dfd-acde-2f557e7d23ee req-6f359cad-e907-4424-8a9c-b6873f8cfbd7 service nova] Acquired lock "refresh_cache-378c1d18-4bb6-4245-80bf-370fb7af0575" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1204.301944] env[65726]: DEBUG nova.network.neutron [req-73a5274e-466f-4dfd-acde-2f557e7d23ee req-6f359cad-e907-4424-8a9c-b6873f8cfbd7 service nova] [instance: 378c1d18-4bb6-4245-80bf-370fb7af0575] Refreshing network info cache for port 4765f6c2-d161-4e36-a6a8-3c8d39c47a40 {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 1204.325584] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Lock "dc112e67-7c2d-4081-9a53-e4f43f61dcd6" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.559s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1204.332259] env[65726]: DEBUG nova.virt.hardware [None req-e32abc12-7404-4281-87f0-12de5cc1731f tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='d24646fc04e356779fcca19129a56bf9',container_format='bare',created_at=2025-12-12T19:40:30Z,direct_url=,disk_format='vmdk',id=1d5801e6-cfd0-4c6a-8fa0-57ee72cc963a,min_disk=1,min_ram=0,name='tempest-ServerActionsTestOtherB-server-1532654838-shelved',owner='4eece77569624f90bf64e5c51974173f',properties=ImageMetaProps,protected=,size=31669760,status='active',tags=,updated_at=2025-12-12T19:40:46Z,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1204.333032] env[65726]: DEBUG nova.virt.hardware [None req-e32abc12-7404-4281-87f0-12de5cc1731f tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1204.333032] env[65726]: DEBUG nova.virt.hardware [None req-e32abc12-7404-4281-87f0-12de5cc1731f tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1204.333278] env[65726]: DEBUG nova.virt.hardware [None req-e32abc12-7404-4281-87f0-12de5cc1731f tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1204.333459] env[65726]: DEBUG nova.virt.hardware [None req-e32abc12-7404-4281-87f0-12de5cc1731f tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1204.333630] env[65726]: DEBUG nova.virt.hardware [None req-e32abc12-7404-4281-87f0-12de5cc1731f tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1204.333897] env[65726]: DEBUG nova.virt.hardware [None req-e32abc12-7404-4281-87f0-12de5cc1731f tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1204.334171] env[65726]: DEBUG nova.virt.hardware [None req-e32abc12-7404-4281-87f0-12de5cc1731f tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1204.334417] env[65726]: DEBUG nova.virt.hardware [None req-e32abc12-7404-4281-87f0-12de5cc1731f tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1204.334776] env[65726]: DEBUG nova.virt.hardware [None req-e32abc12-7404-4281-87f0-12de5cc1731f tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1204.335027] env[65726]: DEBUG nova.virt.hardware [None req-e32abc12-7404-4281-87f0-12de5cc1731f tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1204.336433] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-312378e6-4de5-44a7-adee-ada680e24296 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1204.354127] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a82cea8c-a296-4f78-a247-81c22701c915 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1204.359442] env[65726]: INFO nova.compute.manager [None req-4e863eec-f358-4eca-9eee-f999967acc4a tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] [instance: 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30] Rebuilding instance [ 1204.377394] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-e32abc12-7404-4281-87f0-12de5cc1731f tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 378c1d18-4bb6-4245-80bf-370fb7af0575] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:3c:38:b2', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '2ce62383-8e84-4e26-955b-74c11392f4c9', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4765f6c2-d161-4e36-a6a8-3c8d39c47a40', 'vif_model': 'vmxnet3'}] {{(pid=65726) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1204.385220] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-e32abc12-7404-4281-87f0-12de5cc1731f tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1204.392153] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 378c1d18-4bb6-4245-80bf-370fb7af0575] Creating VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1204.392981] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2c7425db-bf60-4d5b-9fbf-d4042c2e2f5d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1204.423797] env[65726]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1204.423797] env[65726]: value = "task-5116939" [ 1204.423797] env[65726]: _type = "Task" [ 1204.423797] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1204.440895] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116939, 'name': CreateVM_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1204.441480] env[65726]: DEBUG nova.compute.manager [None req-4e863eec-f358-4eca-9eee-f999967acc4a tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] [instance: 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1204.444167] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7b05c71-724b-40cd-8d97-4e00fe9b5cb8 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1204.466397] env[65726]: DEBUG nova.scheduler.client.report [None req-7cd2a85c-e262-4e84-b25f-a094169151c3 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1204.569458] env[65726]: DEBUG oslo_service.periodic_task [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=65726) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1204.569748] env[65726]: DEBUG nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=65726) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11251}} [ 1204.778951] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c1e079dd-56f2-4cf7-8616-b6ff59feac77 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Acquiring lock "86c23794-5d82-4a7a-aec6-de91601177dd" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1204.780540] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c1e079dd-56f2-4cf7-8616-b6ff59feac77 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Lock "86c23794-5d82-4a7a-aec6-de91601177dd" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1204.805292] env[65726]: WARNING openstack [req-73a5274e-466f-4dfd-acde-2f557e7d23ee req-6f359cad-e907-4424-8a9c-b6873f8cfbd7 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1204.805663] env[65726]: WARNING openstack [req-73a5274e-466f-4dfd-acde-2f557e7d23ee req-6f359cad-e907-4424-8a9c-b6873f8cfbd7 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1204.940599] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116939, 'name': CreateVM_Task, 'duration_secs': 0.439666} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1204.944714] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 378c1d18-4bb6-4245-80bf-370fb7af0575] Created VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1204.945934] env[65726]: WARNING openstack [None req-e32abc12-7404-4281-87f0-12de5cc1731f tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1204.946551] env[65726]: WARNING openstack [None req-e32abc12-7404-4281-87f0-12de5cc1731f tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1204.955281] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e32abc12-7404-4281-87f0-12de5cc1731f tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1d5801e6-cfd0-4c6a-8fa0-57ee72cc963a" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1204.955533] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e32abc12-7404-4281-87f0-12de5cc1731f tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1d5801e6-cfd0-4c6a-8fa0-57ee72cc963a" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1204.956079] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e32abc12-7404-4281-87f0-12de5cc1731f tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/1d5801e6-cfd0-4c6a-8fa0-57ee72cc963a" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1204.966377] env[65726]: WARNING openstack [req-73a5274e-466f-4dfd-acde-2f557e7d23ee req-6f359cad-e907-4424-8a9c-b6873f8cfbd7 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1204.966956] env[65726]: WARNING openstack [req-73a5274e-466f-4dfd-acde-2f557e7d23ee req-6f359cad-e907-4424-8a9c-b6873f8cfbd7 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1204.980315] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-03ead88a-ff9b-4655-832a-71ba829006d1 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1204.983473] env[65726]: DEBUG oslo_concurrency.lockutils [None req-7cd2a85c-e262-4e84-b25f-a094169151c3 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.752s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1204.994799] env[65726]: DEBUG oslo_concurrency.lockutils [None req-68748ea0-9eb9-4593-95d7-667ae1802835 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.553s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1204.995192] env[65726]: DEBUG nova.objects.instance [None req-68748ea0-9eb9-4593-95d7-667ae1802835 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Lazy-loading 'resources' on Instance uuid 6956bda7-5657-45d7-8f80-c6809bd836b7 {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1205.002221] env[65726]: DEBUG oslo_vmware.api [None req-e32abc12-7404-4281-87f0-12de5cc1731f tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Waiting for the task: (returnval){ [ 1205.002221] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]521666e8-ae75-5357-7595-12337a1eb818" [ 1205.002221] env[65726]: _type = "Task" [ 1205.002221] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1205.016818] env[65726]: DEBUG oslo_vmware.api [None req-e32abc12-7404-4281-87f0-12de5cc1731f tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]521666e8-ae75-5357-7595-12337a1eb818, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1205.019082] env[65726]: INFO nova.scheduler.client.report [None req-7cd2a85c-e262-4e84-b25f-a094169151c3 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Deleted allocations for instance 3bc25953-21b1-4729-af09-e7211fd8b2c6 [ 1205.063128] env[65726]: WARNING openstack [req-73a5274e-466f-4dfd-acde-2f557e7d23ee req-6f359cad-e907-4424-8a9c-b6873f8cfbd7 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1205.063744] env[65726]: WARNING openstack [req-73a5274e-466f-4dfd-acde-2f557e7d23ee req-6f359cad-e907-4424-8a9c-b6873f8cfbd7 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1205.120595] env[65726]: DEBUG nova.objects.instance [None req-4360b5c9-45f8-4acf-8cf6-3673d2897518 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Lazy-loading 'flavor' on Instance uuid 5bfbaf25-5aed-46bc-97fc-f138d08ebbde {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1205.171665] env[65726]: DEBUG nova.network.neutron [req-73a5274e-466f-4dfd-acde-2f557e7d23ee req-6f359cad-e907-4424-8a9c-b6873f8cfbd7 service nova] [instance: 378c1d18-4bb6-4245-80bf-370fb7af0575] Updated VIF entry in instance network info cache for port 4765f6c2-d161-4e36-a6a8-3c8d39c47a40. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 1205.172186] env[65726]: DEBUG nova.network.neutron [req-73a5274e-466f-4dfd-acde-2f557e7d23ee req-6f359cad-e907-4424-8a9c-b6873f8cfbd7 service nova] [instance: 378c1d18-4bb6-4245-80bf-370fb7af0575] Updating instance_info_cache with network_info: [{"id": "4765f6c2-d161-4e36-a6a8-3c8d39c47a40", "address": "fa:16:3e:3c:38:b2", "network": {"id": "0d0f561d-ac41-4ef5-9926-c707409799f7", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1134321217-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.160", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4eece77569624f90bf64e5c51974173f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2ce62383-8e84-4e26-955b-74c11392f4c9", "external-id": "nsx-vlan-transportzone-215", "segmentation_id": 215, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4765f6c2-d1", "ovs_interfaceid": "4765f6c2-d161-4e36-a6a8-3c8d39c47a40", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1205.267505] env[65726]: DEBUG oslo_vmware.rw_handles [None req-70bb45c2-ad53-4082-9238-425c98dcfe4d tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Completed reading data from the image iterator. {{(pid=65726) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 1205.268102] env[65726]: DEBUG oslo_vmware.rw_handles [None req-70bb45c2-ad53-4082-9238-425c98dcfe4d tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5290a352-77af-403f-8c95-fd94975727e2/disk-0.vmdk. {{(pid=65726) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1205.268929] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-621e0cf4-d583-466e-899d-dc4faebb94f4 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1205.279087] env[65726]: DEBUG oslo_vmware.rw_handles [None req-70bb45c2-ad53-4082-9238-425c98dcfe4d tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5290a352-77af-403f-8c95-fd94975727e2/disk-0.vmdk is in state: ready. {{(pid=65726) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1205.279286] env[65726]: DEBUG oslo_vmware.rw_handles [None req-70bb45c2-ad53-4082-9238-425c98dcfe4d tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Releasing lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5290a352-77af-403f-8c95-fd94975727e2/disk-0.vmdk. {{(pid=65726) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 1205.279547] env[65726]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-231579a3-f433-46f2-bba0-1862fe1a56f6 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1205.281353] env[65726]: DEBUG nova.compute.manager [None req-c1e079dd-56f2-4cf7-8616-b6ff59feac77 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 86c23794-5d82-4a7a-aec6-de91601177dd] Starting instance... {{(pid=65726) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 1205.487459] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-4e863eec-f358-4eca-9eee-f999967acc4a tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] [instance: 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1205.488106] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e8e6e17e-9d54-49a1-89bb-d33fdefea05b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1205.497375] env[65726]: DEBUG oslo_vmware.api [None req-4e863eec-f358-4eca-9eee-f999967acc4a tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] Waiting for the task: (returnval){ [ 1205.497375] env[65726]: value = "task-5116940" [ 1205.497375] env[65726]: _type = "Task" [ 1205.497375] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1205.513906] env[65726]: DEBUG oslo_vmware.api [None req-4e863eec-f358-4eca-9eee-f999967acc4a tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] Task: {'id': task-5116940, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1205.520875] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e32abc12-7404-4281-87f0-12de5cc1731f tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1d5801e6-cfd0-4c6a-8fa0-57ee72cc963a" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1205.521250] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-e32abc12-7404-4281-87f0-12de5cc1731f tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 378c1d18-4bb6-4245-80bf-370fb7af0575] Processing image 1d5801e6-cfd0-4c6a-8fa0-57ee72cc963a {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1205.521521] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e32abc12-7404-4281-87f0-12de5cc1731f tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1d5801e6-cfd0-4c6a-8fa0-57ee72cc963a/1d5801e6-cfd0-4c6a-8fa0-57ee72cc963a.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1205.521705] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e32abc12-7404-4281-87f0-12de5cc1731f tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1d5801e6-cfd0-4c6a-8fa0-57ee72cc963a/1d5801e6-cfd0-4c6a-8fa0-57ee72cc963a.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1205.521951] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-e32abc12-7404-4281-87f0-12de5cc1731f tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1205.522248] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-877ec9eb-deb4-4439-9964-01f5cda520de {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1205.531503] env[65726]: DEBUG oslo_concurrency.lockutils [None req-7cd2a85c-e262-4e84-b25f-a094169151c3 tempest-AttachVolumeNegativeTest-1045106260 tempest-AttachVolumeNegativeTest-1045106260-project-member] Lock "3bc25953-21b1-4729-af09-e7211fd8b2c6" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 11.686s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1205.532591] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Lock "3bc25953-21b1-4729-af09-e7211fd8b2c6" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 5.490s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1205.533099] env[65726]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-766438b3-3091-4314-833f-afdc01334f52 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1205.537585] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-e32abc12-7404-4281-87f0-12de5cc1731f tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1205.537786] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-e32abc12-7404-4281-87f0-12de5cc1731f tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65726) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1205.538633] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-777c5a5a-ddfa-4892-9e30-146bd3505b0e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1205.549453] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3cd4c9d1-87a7-438d-aa47-331f0db32ed1 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1205.563590] env[65726]: DEBUG oslo_vmware.api [None req-e32abc12-7404-4281-87f0-12de5cc1731f tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Waiting for the task: (returnval){ [ 1205.563590] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52921164-7c83-0b33-42f4-21b078c4fc2a" [ 1205.563590] env[65726]: _type = "Task" [ 1205.563590] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1205.577316] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-e32abc12-7404-4281-87f0-12de5cc1731f tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 378c1d18-4bb6-4245-80bf-370fb7af0575] Preparing fetch location {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 1205.577580] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-e32abc12-7404-4281-87f0-12de5cc1731f tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 378c1d18-4bb6-4245-80bf-370fb7af0575] Fetch image to [datastore1] OSTACK_IMG_a3fe861b-02a4-441a-9a6d-7f79ff074ca7/OSTACK_IMG_a3fe861b-02a4-441a-9a6d-7f79ff074ca7.vmdk {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 1205.577804] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-e32abc12-7404-4281-87f0-12de5cc1731f tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 378c1d18-4bb6-4245-80bf-370fb7af0575] Downloading stream optimized image 1d5801e6-cfd0-4c6a-8fa0-57ee72cc963a to [datastore1] OSTACK_IMG_a3fe861b-02a4-441a-9a6d-7f79ff074ca7/OSTACK_IMG_a3fe861b-02a4-441a-9a6d-7f79ff074ca7.vmdk on the data store datastore1 as vApp {{(pid=65726) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 1205.577988] env[65726]: DEBUG nova.virt.vmwareapi.images [None req-e32abc12-7404-4281-87f0-12de5cc1731f tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 378c1d18-4bb6-4245-80bf-370fb7af0575] Downloading image file data 1d5801e6-cfd0-4c6a-8fa0-57ee72cc963a to the ESX as VM named 'OSTACK_IMG_a3fe861b-02a4-441a-9a6d-7f79ff074ca7' {{(pid=65726) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 1205.598441] env[65726]: DEBUG oslo_vmware.rw_handles [None req-70bb45c2-ad53-4082-9238-425c98dcfe4d tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Closed VMDK write handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5290a352-77af-403f-8c95-fd94975727e2/disk-0.vmdk. {{(pid=65726) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 1205.598676] env[65726]: INFO nova.virt.vmwareapi.images [None req-70bb45c2-ad53-4082-9238-425c98dcfe4d tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: b353b5e9-500c-42d3-a87f-880a624febec] Downloaded image file data 745a44b6-0c3f-4396-a82a-6e9472e10925 [ 1205.600518] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-817d1bec-0ed0-4002-9a6b-4ba9a91294e4 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1205.623492] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ba2b403b-0512-473f-8acf-792f465373e9 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1205.652825] env[65726]: DEBUG oslo_concurrency.lockutils [None req-4360b5c9-45f8-4acf-8cf6-3673d2897518 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Lock "5bfbaf25-5aed-46bc-97fc-f138d08ebbde" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.429s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1205.658552] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Lock "5bfbaf25-5aed-46bc-97fc-f138d08ebbde" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 5.615s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1205.660853] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61363f0f-7372-4aee-ba08-b608a5204ceb {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1205.663772] env[65726]: INFO nova.virt.vmwareapi.images [None req-70bb45c2-ad53-4082-9238-425c98dcfe4d tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: b353b5e9-500c-42d3-a87f-880a624febec] The imported VM was unregistered [ 1205.667177] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-70bb45c2-ad53-4082-9238-425c98dcfe4d tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: b353b5e9-500c-42d3-a87f-880a624febec] Caching image {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 1205.667479] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-70bb45c2-ad53-4082-9238-425c98dcfe4d tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Creating directory with path [datastore1] devstack-image-cache_base/745a44b6-0c3f-4396-a82a-6e9472e10925 {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1205.668246] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-acfcfabe-ed28-4ab4-b500-fc1f9233f048 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1205.675480] env[65726]: DEBUG oslo_concurrency.lockutils [req-73a5274e-466f-4dfd-acde-2f557e7d23ee req-6f359cad-e907-4424-8a9c-b6873f8cfbd7 service nova] Releasing lock "refresh_cache-378c1d18-4bb6-4245-80bf-370fb7af0575" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1205.694562] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-70bb45c2-ad53-4082-9238-425c98dcfe4d tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Created directory with path [datastore1] devstack-image-cache_base/745a44b6-0c3f-4396-a82a-6e9472e10925 {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1205.695508] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-70bb45c2-ad53-4082-9238-425c98dcfe4d tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Moving virtual disk from [datastore1] OSTACK_IMG_b1f4d914-8fdf-4b4a-aba5-f24abe6a510f/OSTACK_IMG_b1f4d914-8fdf-4b4a-aba5-f24abe6a510f.vmdk to [datastore1] devstack-image-cache_base/745a44b6-0c3f-4396-a82a-6e9472e10925/745a44b6-0c3f-4396-a82a-6e9472e10925.vmdk. {{(pid=65726) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 1205.695508] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-e0ec551e-cf33-4ffc-ae44-9aeffdb069db {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1205.705691] env[65726]: DEBUG oslo_vmware.api [None req-70bb45c2-ad53-4082-9238-425c98dcfe4d tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Waiting for the task: (returnval){ [ 1205.705691] env[65726]: value = "task-5116942" [ 1205.705691] env[65726]: _type = "Task" [ 1205.705691] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1205.712836] env[65726]: DEBUG oslo_vmware.rw_handles [None req-e32abc12-7404-4281-87f0-12de5cc1731f tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 1205.712836] env[65726]: value = "resgroup-9" [ 1205.712836] env[65726]: _type = "ResourcePool" [ 1205.712836] env[65726]: }. {{(pid=65726) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 1205.713531] env[65726]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-bbbdf34d-818b-420b-b4e4-0107e163a527 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1205.736760] env[65726]: DEBUG oslo_vmware.api [None req-70bb45c2-ad53-4082-9238-425c98dcfe4d tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Task: {'id': task-5116942, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1205.738231] env[65726]: DEBUG oslo_vmware.rw_handles [None req-e32abc12-7404-4281-87f0-12de5cc1731f tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Lease: (returnval){ [ 1205.738231] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]525b3f06-a844-e8ed-0c07-3deacc8e16ba" [ 1205.738231] env[65726]: _type = "HttpNfcLease" [ 1205.738231] env[65726]: } obtained for vApp import into resource pool (val){ [ 1205.738231] env[65726]: value = "resgroup-9" [ 1205.738231] env[65726]: _type = "ResourcePool" [ 1205.738231] env[65726]: }. {{(pid=65726) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 1205.738494] env[65726]: DEBUG oslo_vmware.api [None req-e32abc12-7404-4281-87f0-12de5cc1731f tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Waiting for the lease: (returnval){ [ 1205.738494] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]525b3f06-a844-e8ed-0c07-3deacc8e16ba" [ 1205.738494] env[65726]: _type = "HttpNfcLease" [ 1205.738494] env[65726]: } to be ready. {{(pid=65726) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1205.743271] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18d1bbc3-da09-47cf-a621-7b345131c7ba {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1205.749890] env[65726]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1205.749890] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]525b3f06-a844-e8ed-0c07-3deacc8e16ba" [ 1205.749890] env[65726]: _type = "HttpNfcLease" [ 1205.749890] env[65726]: } is initializing. {{(pid=65726) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1205.752842] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c913c52-0c72-409a-b169-ddba94f16f61 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1205.790407] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5da020c-f8c9-49cd-8daf-811c12683a4e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1205.798595] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42e72190-9141-46a2-8407-052dee85d6e4 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1205.817023] env[65726]: DEBUG nova.compute.provider_tree [None req-68748ea0-9eb9-4593-95d7-667ae1802835 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1205.818408] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c1e079dd-56f2-4cf7-8616-b6ff59feac77 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1206.008817] env[65726]: DEBUG oslo_vmware.api [None req-4e863eec-f358-4eca-9eee-f999967acc4a tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] Task: {'id': task-5116940, 'name': PowerOffVM_Task, 'duration_secs': 0.230323} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1206.010114] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-4e863eec-f358-4eca-9eee-f999967acc4a tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] [instance: 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1206.010599] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-4e863eec-f358-4eca-9eee-f999967acc4a tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] [instance: 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1206.010929] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3235f245-70e7-46fa-9df5-41610684ed12 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1206.019288] env[65726]: DEBUG oslo_vmware.api [None req-4e863eec-f358-4eca-9eee-f999967acc4a tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] Waiting for the task: (returnval){ [ 1206.019288] env[65726]: value = "task-5116945" [ 1206.019288] env[65726]: _type = "Task" [ 1206.019288] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1206.035616] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-4e863eec-f358-4eca-9eee-f999967acc4a tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] [instance: 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30] VM already powered off {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1206.035834] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-4e863eec-f358-4eca-9eee-f999967acc4a tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] [instance: 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30] Volume detach. Driver type: vmdk {{(pid=65726) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1206.036094] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-4e863eec-f358-4eca-9eee-f999967acc4a tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] [instance: 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-995306', 'volume_id': '39918e7e-1d4a-45df-911f-46ed9091f96b', 'name': 'volume-39918e7e-1d4a-45df-911f-46ed9091f96b', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '47e52bc1-c18e-4aa2-82a2-e4cb030a7a30', 'attached_at': '', 'detached_at': '', 'volume_id': '39918e7e-1d4a-45df-911f-46ed9091f96b', 'serial': '39918e7e-1d4a-45df-911f-46ed9091f96b'} {{(pid=65726) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1206.037028] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cef43e0b-076d-4eae-a517-5c2c00477e7c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1206.058518] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5137808-b53e-4be0-ae3b-b9b5051bd6dd {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1206.067295] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93bed84f-893b-4656-8627-c53694f4b688 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1206.090233] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa9de0ec-884d-49ae-b689-25f66e9c2c0a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1206.108590] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Lock "3bc25953-21b1-4729-af09-e7211fd8b2c6" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.576s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1206.109077] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-4e863eec-f358-4eca-9eee-f999967acc4a tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] The volume has not been displaced from its original location: [datastore2] volume-39918e7e-1d4a-45df-911f-46ed9091f96b/volume-39918e7e-1d4a-45df-911f-46ed9091f96b.vmdk. No consolidation needed. {{(pid=65726) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1206.115025] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-4e863eec-f358-4eca-9eee-f999967acc4a tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] [instance: 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30] Reconfiguring VM instance instance-0000006c to detach disk 2000 {{(pid=65726) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1206.115449] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c55779a6-404d-4fa9-a0cd-a780e7fe8b47 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1206.137378] env[65726]: DEBUG oslo_vmware.api [None req-4e863eec-f358-4eca-9eee-f999967acc4a tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] Waiting for the task: (returnval){ [ 1206.137378] env[65726]: value = "task-5116946" [ 1206.137378] env[65726]: _type = "Task" [ 1206.137378] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1206.147547] env[65726]: DEBUG oslo_vmware.api [None req-4e863eec-f358-4eca-9eee-f999967acc4a tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] Task: {'id': task-5116946, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1206.183953] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Lock "5bfbaf25-5aed-46bc-97fc-f138d08ebbde" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.526s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1206.218187] env[65726]: DEBUG oslo_vmware.api [None req-70bb45c2-ad53-4082-9238-425c98dcfe4d tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Task: {'id': task-5116942, 'name': MoveVirtualDisk_Task} progress is 15%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1206.248163] env[65726]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1206.248163] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]525b3f06-a844-e8ed-0c07-3deacc8e16ba" [ 1206.248163] env[65726]: _type = "HttpNfcLease" [ 1206.248163] env[65726]: } is initializing. {{(pid=65726) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1206.318535] env[65726]: DEBUG nova.scheduler.client.report [None req-68748ea0-9eb9-4593-95d7-667ae1802835 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1206.439330] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c0cee6b4-64a9-4ea6-9254-dda6dc407a15 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Acquiring lock "5bfbaf25-5aed-46bc-97fc-f138d08ebbde" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1206.439700] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c0cee6b4-64a9-4ea6-9254-dda6dc407a15 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Lock "5bfbaf25-5aed-46bc-97fc-f138d08ebbde" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.001s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1206.570707] env[65726]: DEBUG oslo_service.periodic_task [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Running periodic task ComputeManager.update_available_resource {{(pid=65726) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1206.649303] env[65726]: DEBUG oslo_vmware.api [None req-4e863eec-f358-4eca-9eee-f999967acc4a tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] Task: {'id': task-5116946, 'name': ReconfigVM_Task, 'duration_secs': 0.229306} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1206.649624] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-4e863eec-f358-4eca-9eee-f999967acc4a tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] [instance: 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30] Reconfigured VM instance instance-0000006c to detach disk 2000 {{(pid=65726) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1206.655760] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9f584299-cb87-4962-88be-3e9976199e92 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1206.678582] env[65726]: DEBUG oslo_vmware.api [None req-4e863eec-f358-4eca-9eee-f999967acc4a tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] Waiting for the task: (returnval){ [ 1206.678582] env[65726]: value = "task-5116947" [ 1206.678582] env[65726]: _type = "Task" [ 1206.678582] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1206.690034] env[65726]: DEBUG oslo_vmware.api [None req-4e863eec-f358-4eca-9eee-f999967acc4a tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] Task: {'id': task-5116947, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1206.717368] env[65726]: DEBUG oslo_vmware.api [None req-70bb45c2-ad53-4082-9238-425c98dcfe4d tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Task: {'id': task-5116942, 'name': MoveVirtualDisk_Task} progress is 35%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1206.748691] env[65726]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1206.748691] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]525b3f06-a844-e8ed-0c07-3deacc8e16ba" [ 1206.748691] env[65726]: _type = "HttpNfcLease" [ 1206.748691] env[65726]: } is ready. {{(pid=65726) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1206.748991] env[65726]: DEBUG oslo_vmware.rw_handles [None req-e32abc12-7404-4281-87f0-12de5cc1731f tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1206.748991] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]525b3f06-a844-e8ed-0c07-3deacc8e16ba" [ 1206.748991] env[65726]: _type = "HttpNfcLease" [ 1206.748991] env[65726]: }. {{(pid=65726) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 1206.749963] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac6ddbac-244c-4f16-84c3-5b802a871148 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1206.758484] env[65726]: DEBUG oslo_vmware.rw_handles [None req-e32abc12-7404-4281-87f0-12de5cc1731f tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52f70014-83a7-19d1-0af9-c8a4a1dbbb11/disk-0.vmdk from lease info. {{(pid=65726) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1206.758686] env[65726]: DEBUG oslo_vmware.rw_handles [None req-e32abc12-7404-4281-87f0-12de5cc1731f tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Creating HTTP connection to write to file with size = 31669760 and URL = https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52f70014-83a7-19d1-0af9-c8a4a1dbbb11/disk-0.vmdk. {{(pid=65726) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 1206.823893] env[65726]: DEBUG oslo_concurrency.lockutils [None req-68748ea0-9eb9-4593-95d7-667ae1802835 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.829s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1206.829928] env[65726]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-f9038202-4726-4169-ac59-3b7a7bdfa02a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1206.831379] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c1e079dd-56f2-4cf7-8616-b6ff59feac77 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.013s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1206.833587] env[65726]: INFO nova.compute.claims [None req-c1e079dd-56f2-4cf7-8616-b6ff59feac77 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 86c23794-5d82-4a7a-aec6-de91601177dd] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1206.851101] env[65726]: INFO nova.scheduler.client.report [None req-68748ea0-9eb9-4593-95d7-667ae1802835 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Deleted allocations for instance 6956bda7-5657-45d7-8f80-c6809bd836b7 [ 1206.943338] env[65726]: DEBUG nova.compute.utils [None req-c0cee6b4-64a9-4ea6-9254-dda6dc407a15 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Using /dev/sd instead of None {{(pid=65726) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1207.073569] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1207.191064] env[65726]: DEBUG oslo_vmware.api [None req-4e863eec-f358-4eca-9eee-f999967acc4a tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] Task: {'id': task-5116947, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1207.218111] env[65726]: DEBUG oslo_vmware.api [None req-70bb45c2-ad53-4082-9238-425c98dcfe4d tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Task: {'id': task-5116942, 'name': MoveVirtualDisk_Task} progress is 54%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1207.359805] env[65726]: DEBUG oslo_concurrency.lockutils [None req-68748ea0-9eb9-4593-95d7-667ae1802835 tempest-ServerDiskConfigTestJSON-603391006 tempest-ServerDiskConfigTestJSON-603391006-project-member] Lock "6956bda7-5657-45d7-8f80-c6809bd836b7" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.304s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1207.360961] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Lock "6956bda7-5657-45d7-8f80-c6809bd836b7" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 7.316s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1207.361165] env[65726]: INFO nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] [instance: 6956bda7-5657-45d7-8f80-c6809bd836b7] During sync_power_state the instance has a pending task (deleting). Skip. [ 1207.361363] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Lock "6956bda7-5657-45d7-8f80-c6809bd836b7" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.001s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1207.447188] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c0cee6b4-64a9-4ea6-9254-dda6dc407a15 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Lock "5bfbaf25-5aed-46bc-97fc-f138d08ebbde" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.007s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1207.694561] env[65726]: DEBUG oslo_vmware.api [None req-4e863eec-f358-4eca-9eee-f999967acc4a tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] Task: {'id': task-5116947, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1207.721553] env[65726]: DEBUG oslo_vmware.api [None req-70bb45c2-ad53-4082-9238-425c98dcfe4d tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Task: {'id': task-5116942, 'name': MoveVirtualDisk_Task} progress is 74%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1207.985251] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8317b00c-23ef-4abe-adfb-b63830ef3b76 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1208.000042] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f525cf24-bada-4e9a-bb40-32277c7e00ad {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1208.051610] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fe5a628-0b7a-417a-bb93-44d8e8218bf8 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1208.069455] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc9290ea-daed-4875-aa30-3aa4cf919a57 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1208.092949] env[65726]: DEBUG nova.compute.provider_tree [None req-c1e079dd-56f2-4cf7-8616-b6ff59feac77 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1208.197138] env[65726]: DEBUG oslo_vmware.api [None req-4e863eec-f358-4eca-9eee-f999967acc4a tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] Task: {'id': task-5116947, 'name': ReconfigVM_Task, 'duration_secs': 1.265327} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1208.197138] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-4e863eec-f358-4eca-9eee-f999967acc4a tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] [instance: 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-995306', 'volume_id': '39918e7e-1d4a-45df-911f-46ed9091f96b', 'name': 'volume-39918e7e-1d4a-45df-911f-46ed9091f96b', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '47e52bc1-c18e-4aa2-82a2-e4cb030a7a30', 'attached_at': '', 'detached_at': '', 'volume_id': '39918e7e-1d4a-45df-911f-46ed9091f96b', 'serial': '39918e7e-1d4a-45df-911f-46ed9091f96b'} {{(pid=65726) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1208.197138] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-4e863eec-f358-4eca-9eee-f999967acc4a tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] [instance: 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30] Destroying instance {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1208.197517] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d986c5be-ddc9-47ee-8ec8-c46e7cf3d135 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1208.212459] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-4e863eec-f358-4eca-9eee-f999967acc4a tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] [instance: 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30] Unregistering the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1208.221488] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-88bc8475-59b0-4c4b-97fb-429bf2fe8b81 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1208.233142] env[65726]: DEBUG oslo_vmware.api [None req-70bb45c2-ad53-4082-9238-425c98dcfe4d tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Task: {'id': task-5116942, 'name': MoveVirtualDisk_Task} progress is 94%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1208.310916] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-4e863eec-f358-4eca-9eee-f999967acc4a tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] [instance: 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30] Unregistered the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1208.311324] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-4e863eec-f358-4eca-9eee-f999967acc4a tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] [instance: 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30] Deleting contents of the VM from datastore datastore2 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1208.311599] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-4e863eec-f358-4eca-9eee-f999967acc4a tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] Deleting the datastore file [datastore2] 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30 {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1208.312042] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-cf1a9c69-e2a1-4931-9b89-294da19e450f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1208.322437] env[65726]: DEBUG oslo_vmware.api [None req-4e863eec-f358-4eca-9eee-f999967acc4a tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] Waiting for the task: (returnval){ [ 1208.322437] env[65726]: value = "task-5116949" [ 1208.322437] env[65726]: _type = "Task" [ 1208.322437] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1208.337780] env[65726]: DEBUG oslo_vmware.api [None req-4e863eec-f358-4eca-9eee-f999967acc4a tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] Task: {'id': task-5116949, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1208.350992] env[65726]: DEBUG oslo_vmware.rw_handles [None req-e32abc12-7404-4281-87f0-12de5cc1731f tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Completed reading data from the image iterator. {{(pid=65726) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 1208.351452] env[65726]: DEBUG oslo_vmware.rw_handles [None req-e32abc12-7404-4281-87f0-12de5cc1731f tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52f70014-83a7-19d1-0af9-c8a4a1dbbb11/disk-0.vmdk. {{(pid=65726) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1208.352876] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-151a04ff-78fc-4962-a87b-acb5da273f58 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1208.362556] env[65726]: DEBUG oslo_vmware.rw_handles [None req-e32abc12-7404-4281-87f0-12de5cc1731f tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52f70014-83a7-19d1-0af9-c8a4a1dbbb11/disk-0.vmdk is in state: ready. {{(pid=65726) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1208.362812] env[65726]: DEBUG oslo_vmware.rw_handles [None req-e32abc12-7404-4281-87f0-12de5cc1731f tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Releasing lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52f70014-83a7-19d1-0af9-c8a4a1dbbb11/disk-0.vmdk. {{(pid=65726) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 1208.363262] env[65726]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-6d300e9e-8b4d-456f-b98c-605def1f0a1e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1208.560103] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c0cee6b4-64a9-4ea6-9254-dda6dc407a15 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Acquiring lock "5bfbaf25-5aed-46bc-97fc-f138d08ebbde" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1208.560444] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c0cee6b4-64a9-4ea6-9254-dda6dc407a15 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Lock "5bfbaf25-5aed-46bc-97fc-f138d08ebbde" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1208.560692] env[65726]: INFO nova.compute.manager [None req-c0cee6b4-64a9-4ea6-9254-dda6dc407a15 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] [instance: 5bfbaf25-5aed-46bc-97fc-f138d08ebbde] Attaching volume a465c46c-664d-4aca-85ae-db25aaeeee00 to /dev/sdc [ 1208.581785] env[65726]: DEBUG oslo_vmware.rw_handles [None req-e32abc12-7404-4281-87f0-12de5cc1731f tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Closed VMDK write handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52f70014-83a7-19d1-0af9-c8a4a1dbbb11/disk-0.vmdk. {{(pid=65726) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 1208.582244] env[65726]: INFO nova.virt.vmwareapi.images [None req-e32abc12-7404-4281-87f0-12de5cc1731f tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 378c1d18-4bb6-4245-80bf-370fb7af0575] Downloaded image file data 1d5801e6-cfd0-4c6a-8fa0-57ee72cc963a [ 1208.583263] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5fffefa-7aba-448b-8cf0-300a2c7b565d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1208.602594] env[65726]: DEBUG nova.scheduler.client.report [None req-c1e079dd-56f2-4cf7-8616-b6ff59feac77 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1208.608060] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-fae359f6-ae88-4260-91f6-5a116138e6da {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1208.610371] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-782291f8-0051-4091-993a-99eb6aac5b78 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1208.620529] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38e787e1-a6b2-4d20-b4cc-7efe5203b82d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1208.637795] env[65726]: DEBUG nova.virt.block_device [None req-c0cee6b4-64a9-4ea6-9254-dda6dc407a15 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] [instance: 5bfbaf25-5aed-46bc-97fc-f138d08ebbde] Updating existing volume attachment record: 031deae6-91ec-4308-9f1d-8ea6e6197f17 {{(pid=65726) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1208.644038] env[65726]: INFO nova.virt.vmwareapi.images [None req-e32abc12-7404-4281-87f0-12de5cc1731f tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 378c1d18-4bb6-4245-80bf-370fb7af0575] The imported VM was unregistered [ 1208.644591] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-e32abc12-7404-4281-87f0-12de5cc1731f tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 378c1d18-4bb6-4245-80bf-370fb7af0575] Caching image {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 1208.644829] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-e32abc12-7404-4281-87f0-12de5cc1731f tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Creating directory with path [datastore1] devstack-image-cache_base/1d5801e6-cfd0-4c6a-8fa0-57ee72cc963a {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1208.645158] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-93af02af-a57c-4f08-aafb-c33ebfcc189a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1208.657811] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-e32abc12-7404-4281-87f0-12de5cc1731f tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Created directory with path [datastore1] devstack-image-cache_base/1d5801e6-cfd0-4c6a-8fa0-57ee72cc963a {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1208.657811] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-e32abc12-7404-4281-87f0-12de5cc1731f tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Moving virtual disk from [datastore1] OSTACK_IMG_a3fe861b-02a4-441a-9a6d-7f79ff074ca7/OSTACK_IMG_a3fe861b-02a4-441a-9a6d-7f79ff074ca7.vmdk to [datastore1] devstack-image-cache_base/1d5801e6-cfd0-4c6a-8fa0-57ee72cc963a/1d5801e6-cfd0-4c6a-8fa0-57ee72cc963a.vmdk. {{(pid=65726) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 1208.658205] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-73ba5b85-f1d9-4e8f-bf55-14bac4cece43 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1208.667143] env[65726]: DEBUG oslo_vmware.api [None req-e32abc12-7404-4281-87f0-12de5cc1731f tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Waiting for the task: (returnval){ [ 1208.667143] env[65726]: value = "task-5116951" [ 1208.667143] env[65726]: _type = "Task" [ 1208.667143] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1208.676797] env[65726]: DEBUG oslo_vmware.api [None req-e32abc12-7404-4281-87f0-12de5cc1731f tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5116951, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1208.723512] env[65726]: DEBUG oslo_vmware.api [None req-70bb45c2-ad53-4082-9238-425c98dcfe4d tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Task: {'id': task-5116942, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.719665} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1208.726911] env[65726]: INFO nova.virt.vmwareapi.ds_util [None req-70bb45c2-ad53-4082-9238-425c98dcfe4d tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Moved virtual disk from [datastore1] OSTACK_IMG_b1f4d914-8fdf-4b4a-aba5-f24abe6a510f/OSTACK_IMG_b1f4d914-8fdf-4b4a-aba5-f24abe6a510f.vmdk to [datastore1] devstack-image-cache_base/745a44b6-0c3f-4396-a82a-6e9472e10925/745a44b6-0c3f-4396-a82a-6e9472e10925.vmdk. [ 1208.726911] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-70bb45c2-ad53-4082-9238-425c98dcfe4d tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: b353b5e9-500c-42d3-a87f-880a624febec] Cleaning up location [datastore1] OSTACK_IMG_b1f4d914-8fdf-4b4a-aba5-f24abe6a510f {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 1208.726911] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-70bb45c2-ad53-4082-9238-425c98dcfe4d tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Deleting the datastore file [datastore1] OSTACK_IMG_b1f4d914-8fdf-4b4a-aba5-f24abe6a510f {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1208.726911] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-67d23665-6a34-4972-bd8f-d9a89da23ce5 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1208.731599] env[65726]: DEBUG oslo_vmware.api [None req-70bb45c2-ad53-4082-9238-425c98dcfe4d tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Waiting for the task: (returnval){ [ 1208.731599] env[65726]: value = "task-5116952" [ 1208.731599] env[65726]: _type = "Task" [ 1208.731599] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1208.740442] env[65726]: DEBUG oslo_vmware.api [None req-70bb45c2-ad53-4082-9238-425c98dcfe4d tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Task: {'id': task-5116952, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1208.834636] env[65726]: DEBUG oslo_vmware.api [None req-4e863eec-f358-4eca-9eee-f999967acc4a tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] Task: {'id': task-5116949, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.104114} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1208.834912] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-4e863eec-f358-4eca-9eee-f999967acc4a tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] Deleted the datastore file {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1208.835150] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-4e863eec-f358-4eca-9eee-f999967acc4a tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] [instance: 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30] Deleted contents of the VM from datastore datastore2 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1208.835669] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-4e863eec-f358-4eca-9eee-f999967acc4a tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] [instance: 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30] Instance destroyed {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1208.918251] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-4e863eec-f358-4eca-9eee-f999967acc4a tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] [instance: 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30] Volume detach. Driver type: vmdk {{(pid=65726) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1208.918757] env[65726]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c1e8c842-c952-42f1-87a1-982c1f5cc96f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1208.936711] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-424aa059-762f-441c-8b5d-ab18d63ae624 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1209.009590] env[65726]: ERROR nova.compute.manager [None req-4e863eec-f358-4eca-9eee-f999967acc4a tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] [instance: 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30] Failed to detach volume 39918e7e-1d4a-45df-911f-46ed9091f96b from /dev/sda: nova.exception.InstanceNotFound: Instance 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30 could not be found. [ 1209.009590] env[65726]: ERROR nova.compute.manager [instance: 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30] Traceback (most recent call last): [ 1209.009590] env[65726]: ERROR nova.compute.manager [instance: 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30] File "/opt/stack/nova/nova/compute/manager.py", line 4241, in _do_rebuild_instance [ 1209.009590] env[65726]: ERROR nova.compute.manager [instance: 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30] self.driver.rebuild(**kwargs) [ 1209.009590] env[65726]: ERROR nova.compute.manager [instance: 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30] File "/opt/stack/nova/nova/virt/driver.py", line 533, in rebuild [ 1209.009590] env[65726]: ERROR nova.compute.manager [instance: 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30] raise NotImplementedError() [ 1209.009590] env[65726]: ERROR nova.compute.manager [instance: 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30] NotImplementedError [ 1209.009590] env[65726]: ERROR nova.compute.manager [instance: 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30] [ 1209.009590] env[65726]: ERROR nova.compute.manager [instance: 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30] During handling of the above exception, another exception occurred: [ 1209.009590] env[65726]: ERROR nova.compute.manager [instance: 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30] [ 1209.009590] env[65726]: ERROR nova.compute.manager [instance: 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30] Traceback (most recent call last): [ 1209.009590] env[65726]: ERROR nova.compute.manager [instance: 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30] File "/opt/stack/nova/nova/compute/manager.py", line 3664, in _detach_root_volume [ 1209.009590] env[65726]: ERROR nova.compute.manager [instance: 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30] self.driver.detach_volume(context, old_connection_info, [ 1209.009590] env[65726]: ERROR nova.compute.manager [instance: 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 561, in detach_volume [ 1209.009590] env[65726]: ERROR nova.compute.manager [instance: 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30] return self._volumeops.detach_volume(connection_info, instance) [ 1209.009590] env[65726]: ERROR nova.compute.manager [instance: 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 649, in detach_volume [ 1209.009590] env[65726]: ERROR nova.compute.manager [instance: 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30] self._detach_volume_vmdk(connection_info, instance) [ 1209.009590] env[65726]: ERROR nova.compute.manager [instance: 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 569, in _detach_volume_vmdk [ 1209.009590] env[65726]: ERROR nova.compute.manager [instance: 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30] vm_ref = vm_util.get_vm_ref(self._session, instance) [ 1209.009590] env[65726]: ERROR nova.compute.manager [instance: 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1145, in get_vm_ref [ 1209.009590] env[65726]: ERROR nova.compute.manager [instance: 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30] stable_ref.fetch_moref(session) [ 1209.009590] env[65726]: ERROR nova.compute.manager [instance: 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1136, in fetch_moref [ 1209.009590] env[65726]: ERROR nova.compute.manager [instance: 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30] raise exception.InstanceNotFound(instance_id=self._uuid) [ 1209.009590] env[65726]: ERROR nova.compute.manager [instance: 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30] nova.exception.InstanceNotFound: Instance 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30 could not be found. [ 1209.009590] env[65726]: ERROR nova.compute.manager [instance: 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30] [ 1209.108336] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c1e079dd-56f2-4cf7-8616-b6ff59feac77 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.277s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1209.108867] env[65726]: DEBUG nova.compute.manager [None req-c1e079dd-56f2-4cf7-8616-b6ff59feac77 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 86c23794-5d82-4a7a-aec6-de91601177dd] Start building networks asynchronously for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1209.111725] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 2.038s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1209.111944] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1209.112620] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=65726) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:937}} [ 1209.113218] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-666c73b9-4729-454c-a4ca-8d57a467eba9 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1209.128181] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-562cdf9e-8fdb-4d9f-88d3-6ec63aedefcc {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1209.152580] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21413cfe-ed18-43a0-9644-c55635f50a83 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1209.163136] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f34a4ffb-947d-4293-977b-1de17f1bcbe9 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1209.179864] env[65726]: DEBUG oslo_vmware.api [None req-e32abc12-7404-4281-87f0-12de5cc1731f tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5116951, 'name': MoveVirtualDisk_Task} progress is 18%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1209.214202] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=179560MB free_disk=95GB free_vcpus=48 pci_devices=None {{(pid=65726) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1136}} [ 1209.214202] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1209.214202] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1209.246031] env[65726]: DEBUG oslo_vmware.api [None req-70bb45c2-ad53-4082-9238-425c98dcfe4d tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Task: {'id': task-5116952, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.045175} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1209.246868] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-70bb45c2-ad53-4082-9238-425c98dcfe4d tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Deleted the datastore file {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1209.246868] env[65726]: DEBUG oslo_concurrency.lockutils [None req-70bb45c2-ad53-4082-9238-425c98dcfe4d tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Releasing lock "[datastore1] devstack-image-cache_base/745a44b6-0c3f-4396-a82a-6e9472e10925/745a44b6-0c3f-4396-a82a-6e9472e10925.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1209.246992] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-70bb45c2-ad53-4082-9238-425c98dcfe4d tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/745a44b6-0c3f-4396-a82a-6e9472e10925/745a44b6-0c3f-4396-a82a-6e9472e10925.vmdk to [datastore1] b353b5e9-500c-42d3-a87f-880a624febec/b353b5e9-500c-42d3-a87f-880a624febec.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1209.247259] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c59ce65a-31cf-4cdc-8fcc-3e4aee692553 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1209.259651] env[65726]: DEBUG oslo_vmware.api [None req-70bb45c2-ad53-4082-9238-425c98dcfe4d tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Waiting for the task: (returnval){ [ 1209.259651] env[65726]: value = "task-5116955" [ 1209.259651] env[65726]: _type = "Task" [ 1209.259651] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1209.262089] env[65726]: DEBUG nova.compute.utils [None req-4e863eec-f358-4eca-9eee-f999967acc4a tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] [instance: 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30] Build of instance 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30 aborted: Failed to rebuild volume backed instance. {{(pid=65726) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:431}} [ 1209.264868] env[65726]: ERROR nova.compute.manager [None req-4e863eec-f358-4eca-9eee-f999967acc4a tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] [instance: 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30] Setting instance vm_state to ERROR: nova.exception.BuildAbortException: Build of instance 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30 aborted: Failed to rebuild volume backed instance. [ 1209.264868] env[65726]: ERROR nova.compute.manager [instance: 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30] Traceback (most recent call last): [ 1209.264868] env[65726]: ERROR nova.compute.manager [instance: 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30] File "/opt/stack/nova/nova/compute/manager.py", line 4241, in _do_rebuild_instance [ 1209.264868] env[65726]: ERROR nova.compute.manager [instance: 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30] self.driver.rebuild(**kwargs) [ 1209.264868] env[65726]: ERROR nova.compute.manager [instance: 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30] File "/opt/stack/nova/nova/virt/driver.py", line 533, in rebuild [ 1209.264868] env[65726]: ERROR nova.compute.manager [instance: 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30] raise NotImplementedError() [ 1209.264868] env[65726]: ERROR nova.compute.manager [instance: 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30] NotImplementedError [ 1209.264868] env[65726]: ERROR nova.compute.manager [instance: 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30] [ 1209.264868] env[65726]: ERROR nova.compute.manager [instance: 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30] During handling of the above exception, another exception occurred: [ 1209.264868] env[65726]: ERROR nova.compute.manager [instance: 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30] [ 1209.264868] env[65726]: ERROR nova.compute.manager [instance: 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30] Traceback (most recent call last): [ 1209.264868] env[65726]: ERROR nova.compute.manager [instance: 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30] File "/opt/stack/nova/nova/compute/manager.py", line 3699, in _rebuild_volume_backed_instance [ 1209.264868] env[65726]: ERROR nova.compute.manager [instance: 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30] self._detach_root_volume(context, instance, root_bdm) [ 1209.264868] env[65726]: ERROR nova.compute.manager [instance: 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30] File "/opt/stack/nova/nova/compute/manager.py", line 3678, in _detach_root_volume [ 1209.264868] env[65726]: ERROR nova.compute.manager [instance: 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30] with excutils.save_and_reraise_exception(): [ 1209.264868] env[65726]: ERROR nova.compute.manager [instance: 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 256, in __exit__ [ 1209.264868] env[65726]: ERROR nova.compute.manager [instance: 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30] self.force_reraise() [ 1209.264868] env[65726]: ERROR nova.compute.manager [instance: 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 222, in force_reraise [ 1209.264868] env[65726]: ERROR nova.compute.manager [instance: 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30] raise self.value [ 1209.264868] env[65726]: ERROR nova.compute.manager [instance: 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30] File "/opt/stack/nova/nova/compute/manager.py", line 3664, in _detach_root_volume [ 1209.264868] env[65726]: ERROR nova.compute.manager [instance: 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30] self.driver.detach_volume(context, old_connection_info, [ 1209.264868] env[65726]: ERROR nova.compute.manager [instance: 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 561, in detach_volume [ 1209.264868] env[65726]: ERROR nova.compute.manager [instance: 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30] return self._volumeops.detach_volume(connection_info, instance) [ 1209.264868] env[65726]: ERROR nova.compute.manager [instance: 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 649, in detach_volume [ 1209.264868] env[65726]: ERROR nova.compute.manager [instance: 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30] self._detach_volume_vmdk(connection_info, instance) [ 1209.264868] env[65726]: ERROR nova.compute.manager [instance: 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 569, in _detach_volume_vmdk [ 1209.264868] env[65726]: ERROR nova.compute.manager [instance: 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30] vm_ref = vm_util.get_vm_ref(self._session, instance) [ 1209.264868] env[65726]: ERROR nova.compute.manager [instance: 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1145, in get_vm_ref [ 1209.264868] env[65726]: ERROR nova.compute.manager [instance: 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30] stable_ref.fetch_moref(session) [ 1209.264868] env[65726]: ERROR nova.compute.manager [instance: 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1136, in fetch_moref [ 1209.264868] env[65726]: ERROR nova.compute.manager [instance: 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30] raise exception.InstanceNotFound(instance_id=self._uuid) [ 1209.264868] env[65726]: ERROR nova.compute.manager [instance: 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30] nova.exception.InstanceNotFound: Instance 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30 could not be found. [ 1209.264868] env[65726]: ERROR nova.compute.manager [instance: 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30] [ 1209.264868] env[65726]: ERROR nova.compute.manager [instance: 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30] During handling of the above exception, another exception occurred: [ 1209.264868] env[65726]: ERROR nova.compute.manager [instance: 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30] [ 1209.264868] env[65726]: ERROR nova.compute.manager [instance: 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30] Traceback (most recent call last): [ 1209.264868] env[65726]: ERROR nova.compute.manager [instance: 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30] File "/opt/stack/nova/nova/compute/manager.py", line 11569, in _error_out_instance_on_exception [ 1209.264868] env[65726]: ERROR nova.compute.manager [instance: 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30] yield [ 1209.264868] env[65726]: ERROR nova.compute.manager [instance: 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30] File "/opt/stack/nova/nova/compute/manager.py", line 3967, in rebuild_instance [ 1209.264868] env[65726]: ERROR nova.compute.manager [instance: 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30] self._do_rebuild_instance_with_claim( [ 1209.267156] env[65726]: ERROR nova.compute.manager [instance: 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30] File "/opt/stack/nova/nova/compute/manager.py", line 4053, in _do_rebuild_instance_with_claim [ 1209.267156] env[65726]: ERROR nova.compute.manager [instance: 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30] self._do_rebuild_instance( [ 1209.267156] env[65726]: ERROR nova.compute.manager [instance: 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30] File "/opt/stack/nova/nova/compute/manager.py", line 4245, in _do_rebuild_instance [ 1209.267156] env[65726]: ERROR nova.compute.manager [instance: 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30] self._rebuild_default_impl(**kwargs) [ 1209.267156] env[65726]: ERROR nova.compute.manager [instance: 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30] File "/opt/stack/nova/nova/compute/manager.py", line 3822, in _rebuild_default_impl [ 1209.267156] env[65726]: ERROR nova.compute.manager [instance: 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30] self._rebuild_volume_backed_instance( [ 1209.267156] env[65726]: ERROR nova.compute.manager [instance: 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30] File "/opt/stack/nova/nova/compute/manager.py", line 3714, in _rebuild_volume_backed_instance [ 1209.267156] env[65726]: ERROR nova.compute.manager [instance: 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30] raise exception.BuildAbortException( [ 1209.267156] env[65726]: ERROR nova.compute.manager [instance: 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30] nova.exception.BuildAbortException: Build of instance 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30 aborted: Failed to rebuild volume backed instance. [ 1209.267156] env[65726]: ERROR nova.compute.manager [instance: 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30] [ 1209.282131] env[65726]: DEBUG oslo_vmware.api [None req-70bb45c2-ad53-4082-9238-425c98dcfe4d tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Task: {'id': task-5116955, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1209.614698] env[65726]: DEBUG nova.compute.utils [None req-c1e079dd-56f2-4cf7-8616-b6ff59feac77 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Using /dev/sd instead of None {{(pid=65726) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1209.616807] env[65726]: DEBUG nova.compute.manager [None req-c1e079dd-56f2-4cf7-8616-b6ff59feac77 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 86c23794-5d82-4a7a-aec6-de91601177dd] Allocating IP information in the background. {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1209.616807] env[65726]: DEBUG nova.network.neutron [None req-c1e079dd-56f2-4cf7-8616-b6ff59feac77 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 86c23794-5d82-4a7a-aec6-de91601177dd] allocate_for_instance() {{(pid=65726) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 1209.616807] env[65726]: WARNING neutronclient.v2_0.client [None req-c1e079dd-56f2-4cf7-8616-b6ff59feac77 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1209.617207] env[65726]: WARNING neutronclient.v2_0.client [None req-c1e079dd-56f2-4cf7-8616-b6ff59feac77 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1209.617704] env[65726]: WARNING openstack [None req-c1e079dd-56f2-4cf7-8616-b6ff59feac77 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1209.618079] env[65726]: WARNING openstack [None req-c1e079dd-56f2-4cf7-8616-b6ff59feac77 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1209.676447] env[65726]: DEBUG nova.policy [None req-c1e079dd-56f2-4cf7-8616-b6ff59feac77 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ccb3513b5b4f453f842139ec1815c4af', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f3f799aa94f64e8cb26d93f6124efd81', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65726) authorize /opt/stack/nova/nova/policy.py:192}} [ 1209.688514] env[65726]: DEBUG oslo_vmware.api [None req-e32abc12-7404-4281-87f0-12de5cc1731f tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5116951, 'name': MoveVirtualDisk_Task} progress is 38%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1209.783397] env[65726]: DEBUG oslo_vmware.api [None req-70bb45c2-ad53-4082-9238-425c98dcfe4d tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Task: {'id': task-5116955, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1209.957934] env[65726]: DEBUG nova.network.neutron [None req-c1e079dd-56f2-4cf7-8616-b6ff59feac77 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 86c23794-5d82-4a7a-aec6-de91601177dd] Successfully created port: dd9340e1-39b8-47ca-91f2-eec36706a058 {{(pid=65726) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1210.126671] env[65726]: DEBUG nova.compute.manager [None req-c1e079dd-56f2-4cf7-8616-b6ff59feac77 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 86c23794-5d82-4a7a-aec6-de91601177dd] Start building block device mappings for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1210.181298] env[65726]: DEBUG oslo_vmware.api [None req-e32abc12-7404-4281-87f0-12de5cc1731f tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5116951, 'name': MoveVirtualDisk_Task} progress is 57%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1210.254543] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Instance 5bfbaf25-5aed-46bc-97fc-f138d08ebbde actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65726) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1210.254604] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Instance 9c0599e3-383f-46ab-809b-944cc3a4d206 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65726) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1210.255596] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Instance e3255f3b-028f-4a0b-b621-873730417936 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65726) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1210.255596] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Instance 55ee97c3-b014-48eb-b41b-fc5aa16bc09e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65726) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1210.255596] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Instance 378c1d18-4bb6-4245-80bf-370fb7af0575 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65726) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1210.255596] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Instance b353b5e9-500c-42d3-a87f-880a624febec actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65726) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1210.255596] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Instance 86c23794-5d82-4a7a-aec6-de91601177dd actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65726) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1210.255596] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Total usable vcpus: 48, total allocated vcpus: 8 {{(pid=65726) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1159}} [ 1210.255596] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2048MB phys_disk=100GB used_disk=7GB total_vcpus=48 used_vcpus=8 pci_stats=[] stats={'failed_builds': '0', 'num_instances': '8', 'num_vm_shelved_offloaded': '2', 'num_task_spawning': '2', 'num_os_type_None': '8', 'num_proj_96149159e18e44f9bf3453e67681f224': '1', 'io_workload': '2', 'num_proj_4eece77569624f90bf64e5c51974173f': '1', 'num_vm_active': '5', 'num_task_None': '4', 'num_proj_4ffd45f4a7a041199a4fc7f69f5e7e9b': '1', 'num_task_rebuilding': '1', 'num_proj_1e3203c117de4d4d9c4c90436801ff3f': '1', 'num_proj_f3f799aa94f64e8cb26d93f6124efd81': '2', 'num_proj_acd800e0438940bc80ae3c6b672db4ee': '1', 'num_proj_6b56f93ba0e14521921484ac2c785c43': '1', 'num_vm_building': '1', 'num_task_networking': '1'} {{(pid=65726) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1168}} [ 1210.288517] env[65726]: DEBUG oslo_vmware.api [None req-70bb45c2-ad53-4082-9238-425c98dcfe4d tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Task: {'id': task-5116955, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1210.400644] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19f6b6ef-2c34-4abc-8112-3520e7500c6b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1210.412115] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6717a477-b2cc-43da-83f6-e633daad26d7 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1210.453957] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0dec86f-a03b-4fd9-a249-cc416ad9cd75 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1210.465196] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1778d3c-ef96-4b66-80fa-4d278573a76a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1210.481860] env[65726]: DEBUG nova.compute.provider_tree [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1210.683429] env[65726]: DEBUG oslo_vmware.api [None req-e32abc12-7404-4281-87f0-12de5cc1731f tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5116951, 'name': MoveVirtualDisk_Task} progress is 77%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1210.779085] env[65726]: DEBUG oslo_vmware.api [None req-70bb45c2-ad53-4082-9238-425c98dcfe4d tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Task: {'id': task-5116955, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1210.985750] env[65726]: DEBUG nova.scheduler.client.report [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1211.144114] env[65726]: DEBUG nova.compute.manager [None req-c1e079dd-56f2-4cf7-8616-b6ff59feac77 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 86c23794-5d82-4a7a-aec6-de91601177dd] Start spawning the instance on the hypervisor. {{(pid=65726) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1211.177280] env[65726]: DEBUG nova.virt.hardware [None req-c1e079dd-56f2-4cf7-8616-b6ff59feac77 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T19:28:34Z,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1539f06fa3534e90acbc3a60c4b8bd3f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T19:28:35Z,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1211.177580] env[65726]: DEBUG nova.virt.hardware [None req-c1e079dd-56f2-4cf7-8616-b6ff59feac77 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1211.177736] env[65726]: DEBUG nova.virt.hardware [None req-c1e079dd-56f2-4cf7-8616-b6ff59feac77 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1211.177938] env[65726]: DEBUG nova.virt.hardware [None req-c1e079dd-56f2-4cf7-8616-b6ff59feac77 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1211.178097] env[65726]: DEBUG nova.virt.hardware [None req-c1e079dd-56f2-4cf7-8616-b6ff59feac77 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1211.178259] env[65726]: DEBUG nova.virt.hardware [None req-c1e079dd-56f2-4cf7-8616-b6ff59feac77 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1211.178482] env[65726]: DEBUG nova.virt.hardware [None req-c1e079dd-56f2-4cf7-8616-b6ff59feac77 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1211.178643] env[65726]: DEBUG nova.virt.hardware [None req-c1e079dd-56f2-4cf7-8616-b6ff59feac77 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1211.178807] env[65726]: DEBUG nova.virt.hardware [None req-c1e079dd-56f2-4cf7-8616-b6ff59feac77 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1211.178964] env[65726]: DEBUG nova.virt.hardware [None req-c1e079dd-56f2-4cf7-8616-b6ff59feac77 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1211.179498] env[65726]: DEBUG nova.virt.hardware [None req-c1e079dd-56f2-4cf7-8616-b6ff59feac77 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1211.180149] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-baf2ebea-db55-4564-8bed-a6ea4e267c11 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1211.189362] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5769dcf-66db-4d1f-9dd1-c164c37fc564 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1211.196955] env[65726]: DEBUG oslo_vmware.api [None req-e32abc12-7404-4281-87f0-12de5cc1731f tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5116951, 'name': MoveVirtualDisk_Task} progress is 97%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1211.278985] env[65726]: DEBUG oslo_vmware.api [None req-70bb45c2-ad53-4082-9238-425c98dcfe4d tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Task: {'id': task-5116955, 'name': CopyVirtualDisk_Task} progress is 9%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1211.294801] env[65726]: DEBUG oslo_concurrency.lockutils [None req-4e863eec-f358-4eca-9eee-f999967acc4a tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1211.491486] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=65726) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1097}} [ 1211.491931] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.278s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1211.492296] env[65726]: DEBUG oslo_concurrency.lockutils [None req-4e863eec-f358-4eca-9eee-f999967acc4a tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.198s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1211.597916] env[65726]: DEBUG nova.network.neutron [None req-c1e079dd-56f2-4cf7-8616-b6ff59feac77 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 86c23794-5d82-4a7a-aec6-de91601177dd] Successfully updated port: dd9340e1-39b8-47ca-91f2-eec36706a058 {{(pid=65726) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 1211.620601] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e52a7e91-079b-4066-85ec-420f82f758c2 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1211.630573] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6e7595d-3012-44d8-b8fd-b7ab4c8d5186 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1211.638753] env[65726]: DEBUG nova.compute.manager [req-ee19ded7-096b-4321-bff3-bc7520d40d41 req-624b84b5-d2b5-4043-9a4c-e9fd26a88f47 service nova] [instance: 86c23794-5d82-4a7a-aec6-de91601177dd] Received event network-vif-plugged-dd9340e1-39b8-47ca-91f2-eec36706a058 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1211.639299] env[65726]: DEBUG oslo_concurrency.lockutils [req-ee19ded7-096b-4321-bff3-bc7520d40d41 req-624b84b5-d2b5-4043-9a4c-e9fd26a88f47 service nova] Acquiring lock "86c23794-5d82-4a7a-aec6-de91601177dd-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1211.639631] env[65726]: DEBUG oslo_concurrency.lockutils [req-ee19ded7-096b-4321-bff3-bc7520d40d41 req-624b84b5-d2b5-4043-9a4c-e9fd26a88f47 service nova] Lock "86c23794-5d82-4a7a-aec6-de91601177dd-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1211.639852] env[65726]: DEBUG oslo_concurrency.lockutils [req-ee19ded7-096b-4321-bff3-bc7520d40d41 req-624b84b5-d2b5-4043-9a4c-e9fd26a88f47 service nova] Lock "86c23794-5d82-4a7a-aec6-de91601177dd-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1211.640160] env[65726]: DEBUG nova.compute.manager [req-ee19ded7-096b-4321-bff3-bc7520d40d41 req-624b84b5-d2b5-4043-9a4c-e9fd26a88f47 service nova] [instance: 86c23794-5d82-4a7a-aec6-de91601177dd] No waiting events found dispatching network-vif-plugged-dd9340e1-39b8-47ca-91f2-eec36706a058 {{(pid=65726) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1211.640337] env[65726]: WARNING nova.compute.manager [req-ee19ded7-096b-4321-bff3-bc7520d40d41 req-624b84b5-d2b5-4043-9a4c-e9fd26a88f47 service nova] [instance: 86c23794-5d82-4a7a-aec6-de91601177dd] Received unexpected event network-vif-plugged-dd9340e1-39b8-47ca-91f2-eec36706a058 for instance with vm_state building and task_state spawning. [ 1211.670924] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-214a80fb-5cc9-41c2-b425-d9dfb6f0c1cb {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1211.690220] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9940639f-4be6-46b9-8907-6703df6629fd {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1211.695082] env[65726]: DEBUG oslo_vmware.api [None req-e32abc12-7404-4281-87f0-12de5cc1731f tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5116951, 'name': MoveVirtualDisk_Task, 'duration_secs': 3.006295} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1211.695837] env[65726]: INFO nova.virt.vmwareapi.ds_util [None req-e32abc12-7404-4281-87f0-12de5cc1731f tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Moved virtual disk from [datastore1] OSTACK_IMG_a3fe861b-02a4-441a-9a6d-7f79ff074ca7/OSTACK_IMG_a3fe861b-02a4-441a-9a6d-7f79ff074ca7.vmdk to [datastore1] devstack-image-cache_base/1d5801e6-cfd0-4c6a-8fa0-57ee72cc963a/1d5801e6-cfd0-4c6a-8fa0-57ee72cc963a.vmdk. [ 1211.696042] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-e32abc12-7404-4281-87f0-12de5cc1731f tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 378c1d18-4bb6-4245-80bf-370fb7af0575] Cleaning up location [datastore1] OSTACK_IMG_a3fe861b-02a4-441a-9a6d-7f79ff074ca7 {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 1211.696216] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-e32abc12-7404-4281-87f0-12de5cc1731f tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Deleting the datastore file [datastore1] OSTACK_IMG_a3fe861b-02a4-441a-9a6d-7f79ff074ca7 {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1211.696956] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7a24f707-97c6-46ff-adcb-111e42a28cb7 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1211.710648] env[65726]: DEBUG nova.compute.provider_tree [None req-4e863eec-f358-4eca-9eee-f999967acc4a tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1211.715455] env[65726]: DEBUG oslo_vmware.api [None req-e32abc12-7404-4281-87f0-12de5cc1731f tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Waiting for the task: (returnval){ [ 1211.715455] env[65726]: value = "task-5116958" [ 1211.715455] env[65726]: _type = "Task" [ 1211.715455] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1211.726947] env[65726]: DEBUG oslo_vmware.api [None req-e32abc12-7404-4281-87f0-12de5cc1731f tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5116958, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1211.780608] env[65726]: DEBUG oslo_vmware.api [None req-70bb45c2-ad53-4082-9238-425c98dcfe4d tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Task: {'id': task-5116955, 'name': CopyVirtualDisk_Task} progress is 26%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1211.889025] env[65726]: DEBUG oslo_concurrency.lockutils [None req-76669e62-2e2e-4af0-a67b-7e8917e36e18 tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] Acquiring lock "47e52bc1-c18e-4aa2-82a2-e4cb030a7a30" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1211.889342] env[65726]: DEBUG oslo_concurrency.lockutils [None req-76669e62-2e2e-4af0-a67b-7e8917e36e18 tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] Lock "47e52bc1-c18e-4aa2-82a2-e4cb030a7a30" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1211.889585] env[65726]: DEBUG oslo_concurrency.lockutils [None req-76669e62-2e2e-4af0-a67b-7e8917e36e18 tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] Acquiring lock "47e52bc1-c18e-4aa2-82a2-e4cb030a7a30-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1211.889798] env[65726]: DEBUG oslo_concurrency.lockutils [None req-76669e62-2e2e-4af0-a67b-7e8917e36e18 tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] Lock "47e52bc1-c18e-4aa2-82a2-e4cb030a7a30-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1211.890107] env[65726]: DEBUG oslo_concurrency.lockutils [None req-76669e62-2e2e-4af0-a67b-7e8917e36e18 tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] Lock "47e52bc1-c18e-4aa2-82a2-e4cb030a7a30-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1211.893324] env[65726]: INFO nova.compute.manager [None req-76669e62-2e2e-4af0-a67b-7e8917e36e18 tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] [instance: 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30] Terminating instance [ 1212.100596] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c1e079dd-56f2-4cf7-8616-b6ff59feac77 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Acquiring lock "refresh_cache-86c23794-5d82-4a7a-aec6-de91601177dd" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1212.100785] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c1e079dd-56f2-4cf7-8616-b6ff59feac77 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Acquired lock "refresh_cache-86c23794-5d82-4a7a-aec6-de91601177dd" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1212.100967] env[65726]: DEBUG nova.network.neutron [None req-c1e079dd-56f2-4cf7-8616-b6ff59feac77 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 86c23794-5d82-4a7a-aec6-de91601177dd] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1212.215175] env[65726]: DEBUG nova.scheduler.client.report [None req-4e863eec-f358-4eca-9eee-f999967acc4a tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1212.233595] env[65726]: DEBUG oslo_vmware.api [None req-e32abc12-7404-4281-87f0-12de5cc1731f tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5116958, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.337789} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1212.233595] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-e32abc12-7404-4281-87f0-12de5cc1731f tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Deleted the datastore file {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1212.233595] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e32abc12-7404-4281-87f0-12de5cc1731f tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1d5801e6-cfd0-4c6a-8fa0-57ee72cc963a/1d5801e6-cfd0-4c6a-8fa0-57ee72cc963a.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1212.233595] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-e32abc12-7404-4281-87f0-12de5cc1731f tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/1d5801e6-cfd0-4c6a-8fa0-57ee72cc963a/1d5801e6-cfd0-4c6a-8fa0-57ee72cc963a.vmdk to [datastore1] 378c1d18-4bb6-4245-80bf-370fb7af0575/378c1d18-4bb6-4245-80bf-370fb7af0575.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1212.233595] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e38aa144-1b8f-4d5b-9fee-465edcb7a1d4 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1212.243878] env[65726]: DEBUG oslo_vmware.api [None req-e32abc12-7404-4281-87f0-12de5cc1731f tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Waiting for the task: (returnval){ [ 1212.243878] env[65726]: value = "task-5116959" [ 1212.243878] env[65726]: _type = "Task" [ 1212.243878] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1212.254597] env[65726]: DEBUG oslo_vmware.api [None req-e32abc12-7404-4281-87f0-12de5cc1731f tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5116959, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1212.279246] env[65726]: DEBUG oslo_vmware.api [None req-70bb45c2-ad53-4082-9238-425c98dcfe4d tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Task: {'id': task-5116955, 'name': CopyVirtualDisk_Task} progress is 49%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1212.399138] env[65726]: DEBUG nova.compute.manager [None req-76669e62-2e2e-4af0-a67b-7e8917e36e18 tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] [instance: 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30] Start destroying the instance on the hypervisor. {{(pid=65726) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1212.399488] env[65726]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-fa04365a-d718-4bd7-8b43-7db3efcb73c0 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1212.410126] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-577c0350-2606-40e3-aa8d-74291bb4b98d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1212.447995] env[65726]: WARNING nova.virt.vmwareapi.driver [None req-76669e62-2e2e-4af0-a67b-7e8917e36e18 tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] [instance: 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30] Instance does not exists. Proceeding to delete instance properties on datastore: nova.exception.InstanceNotFound: Instance 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30 could not be found. [ 1212.448208] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-76669e62-2e2e-4af0-a67b-7e8917e36e18 tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] [instance: 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30] Destroying instance {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1212.448635] env[65726]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-cc896a2d-3e0b-46e9-939f-63211f12ad69 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1212.458312] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58f55781-e83b-4cd5-9119-481728e0820e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1212.492068] env[65726]: WARNING nova.virt.vmwareapi.vmops [None req-76669e62-2e2e-4af0-a67b-7e8917e36e18 tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] [instance: 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30 could not be found. [ 1212.492354] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-76669e62-2e2e-4af0-a67b-7e8917e36e18 tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] [instance: 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30] Instance destroyed {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1212.492587] env[65726]: INFO nova.compute.manager [None req-76669e62-2e2e-4af0-a67b-7e8917e36e18 tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] [instance: 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30] Took 0.09 seconds to destroy the instance on the hypervisor. [ 1212.492863] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-76669e62-2e2e-4af0-a67b-7e8917e36e18 tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1212.493169] env[65726]: DEBUG nova.compute.manager [-] [instance: 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30] Deallocating network for instance {{(pid=65726) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1212.493274] env[65726]: DEBUG nova.network.neutron [-] [instance: 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30] deallocate_for_instance() {{(pid=65726) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 1212.493853] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1212.494208] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1212.535058] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1212.535058] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1212.605069] env[65726]: WARNING openstack [None req-c1e079dd-56f2-4cf7-8616-b6ff59feac77 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1212.605709] env[65726]: WARNING openstack [None req-c1e079dd-56f2-4cf7-8616-b6ff59feac77 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1212.613523] env[65726]: DEBUG nova.network.neutron [None req-c1e079dd-56f2-4cf7-8616-b6ff59feac77 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 86c23794-5d82-4a7a-aec6-de91601177dd] Instance cache missing network info. {{(pid=65726) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 1212.695455] env[65726]: WARNING openstack [None req-c1e079dd-56f2-4cf7-8616-b6ff59feac77 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1212.695829] env[65726]: WARNING openstack [None req-c1e079dd-56f2-4cf7-8616-b6ff59feac77 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1212.726213] env[65726]: DEBUG oslo_concurrency.lockutils [None req-4e863eec-f358-4eca-9eee-f999967acc4a tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.234s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1212.726495] env[65726]: INFO nova.compute.manager [None req-4e863eec-f358-4eca-9eee-f999967acc4a tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] [instance: 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30] Successfully reverted task state from rebuilding on failure for instance. [ 1212.756212] env[65726]: DEBUG oslo_vmware.api [None req-e32abc12-7404-4281-87f0-12de5cc1731f tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5116959, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1212.780018] env[65726]: DEBUG oslo_vmware.api [None req-70bb45c2-ad53-4082-9238-425c98dcfe4d tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Task: {'id': task-5116955, 'name': CopyVirtualDisk_Task} progress is 69%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1212.786454] env[65726]: WARNING openstack [None req-c1e079dd-56f2-4cf7-8616-b6ff59feac77 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1212.786880] env[65726]: WARNING openstack [None req-c1e079dd-56f2-4cf7-8616-b6ff59feac77 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1212.986670] env[65726]: DEBUG nova.network.neutron [None req-c1e079dd-56f2-4cf7-8616-b6ff59feac77 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 86c23794-5d82-4a7a-aec6-de91601177dd] Updating instance_info_cache with network_info: [{"id": "dd9340e1-39b8-47ca-91f2-eec36706a058", "address": "fa:16:3e:c4:39:5d", "network": {"id": "69642329-e986-48fa-accf-1f296c08116a", "bridge": "br-int", "label": "tempest-ServersTestJSON-1786538788-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f3f799aa94f64e8cb26d93f6124efd81", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "816c6e38-e200-4544-8c5b-9fc3e16c5761", "external-id": "nsx-vlan-transportzone-195", "segmentation_id": 195, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdd9340e1-39", "ovs_interfaceid": "dd9340e1-39b8-47ca-91f2-eec36706a058", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1213.189644] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-c0cee6b4-64a9-4ea6-9254-dda6dc407a15 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] [instance: 5bfbaf25-5aed-46bc-97fc-f138d08ebbde] Volume attach. Driver type: vmdk {{(pid=65726) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1213.189948] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-c0cee6b4-64a9-4ea6-9254-dda6dc407a15 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] [instance: 5bfbaf25-5aed-46bc-97fc-f138d08ebbde] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-995333', 'volume_id': 'a465c46c-664d-4aca-85ae-db25aaeeee00', 'name': 'volume-a465c46c-664d-4aca-85ae-db25aaeeee00', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '5bfbaf25-5aed-46bc-97fc-f138d08ebbde', 'attached_at': '', 'detached_at': '', 'volume_id': 'a465c46c-664d-4aca-85ae-db25aaeeee00', 'serial': 'a465c46c-664d-4aca-85ae-db25aaeeee00'} {{(pid=65726) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1213.190874] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4f4f151-d3d5-4f1b-a0b2-540b62925c9d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1213.209857] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9eebbd0-9bf9-4466-9feb-29a064f6b7e8 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1213.241087] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-c0cee6b4-64a9-4ea6-9254-dda6dc407a15 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] [instance: 5bfbaf25-5aed-46bc-97fc-f138d08ebbde] Reconfiguring VM instance instance-0000006b to attach disk [datastore1] volume-a465c46c-664d-4aca-85ae-db25aaeeee00/volume-a465c46c-664d-4aca-85ae-db25aaeeee00.vmdk or device None with type thin {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1213.241538] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f658e6b9-9958-4687-a2ad-463f82fab76c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1213.265789] env[65726]: DEBUG oslo_vmware.api [None req-e32abc12-7404-4281-87f0-12de5cc1731f tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5116959, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1213.267646] env[65726]: DEBUG oslo_vmware.api [None req-c0cee6b4-64a9-4ea6-9254-dda6dc407a15 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Waiting for the task: (returnval){ [ 1213.267646] env[65726]: value = "task-5116960" [ 1213.267646] env[65726]: _type = "Task" [ 1213.267646] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1213.282972] env[65726]: DEBUG oslo_vmware.api [None req-c0cee6b4-64a9-4ea6-9254-dda6dc407a15 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Task: {'id': task-5116960, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1213.287529] env[65726]: DEBUG oslo_vmware.api [None req-70bb45c2-ad53-4082-9238-425c98dcfe4d tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Task: {'id': task-5116955, 'name': CopyVirtualDisk_Task} progress is 88%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1213.471417] env[65726]: DEBUG nova.network.neutron [-] [instance: 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1213.497467] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c1e079dd-56f2-4cf7-8616-b6ff59feac77 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Releasing lock "refresh_cache-86c23794-5d82-4a7a-aec6-de91601177dd" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1213.497467] env[65726]: DEBUG nova.compute.manager [None req-c1e079dd-56f2-4cf7-8616-b6ff59feac77 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 86c23794-5d82-4a7a-aec6-de91601177dd] Instance network_info: |[{"id": "dd9340e1-39b8-47ca-91f2-eec36706a058", "address": "fa:16:3e:c4:39:5d", "network": {"id": "69642329-e986-48fa-accf-1f296c08116a", "bridge": "br-int", "label": "tempest-ServersTestJSON-1786538788-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f3f799aa94f64e8cb26d93f6124efd81", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "816c6e38-e200-4544-8c5b-9fc3e16c5761", "external-id": "nsx-vlan-transportzone-195", "segmentation_id": 195, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdd9340e1-39", "ovs_interfaceid": "dd9340e1-39b8-47ca-91f2-eec36706a058", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1213.497467] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-c1e079dd-56f2-4cf7-8616-b6ff59feac77 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 86c23794-5d82-4a7a-aec6-de91601177dd] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c4:39:5d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '816c6e38-e200-4544-8c5b-9fc3e16c5761', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'dd9340e1-39b8-47ca-91f2-eec36706a058', 'vif_model': 'vmxnet3'}] {{(pid=65726) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1213.507878] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-c1e079dd-56f2-4cf7-8616-b6ff59feac77 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1213.507878] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 86c23794-5d82-4a7a-aec6-de91601177dd] Creating VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1213.507878] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-bceb258a-eb0a-44b2-a257-e09de2c1160c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1213.546801] env[65726]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1213.546801] env[65726]: value = "task-5116961" [ 1213.546801] env[65726]: _type = "Task" [ 1213.546801] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1213.563204] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116961, 'name': CreateVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1213.673836] env[65726]: DEBUG nova.compute.manager [req-244cdfaf-04db-4d6c-9730-0f5000754417 req-809a7728-1743-47f8-85c2-84dd9ea79b07 service nova] [instance: 86c23794-5d82-4a7a-aec6-de91601177dd] Received event network-changed-dd9340e1-39b8-47ca-91f2-eec36706a058 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1213.673836] env[65726]: DEBUG nova.compute.manager [req-244cdfaf-04db-4d6c-9730-0f5000754417 req-809a7728-1743-47f8-85c2-84dd9ea79b07 service nova] [instance: 86c23794-5d82-4a7a-aec6-de91601177dd] Refreshing instance network info cache due to event network-changed-dd9340e1-39b8-47ca-91f2-eec36706a058. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 1213.673836] env[65726]: DEBUG oslo_concurrency.lockutils [req-244cdfaf-04db-4d6c-9730-0f5000754417 req-809a7728-1743-47f8-85c2-84dd9ea79b07 service nova] Acquiring lock "refresh_cache-86c23794-5d82-4a7a-aec6-de91601177dd" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1213.674197] env[65726]: DEBUG oslo_concurrency.lockutils [req-244cdfaf-04db-4d6c-9730-0f5000754417 req-809a7728-1743-47f8-85c2-84dd9ea79b07 service nova] Acquired lock "refresh_cache-86c23794-5d82-4a7a-aec6-de91601177dd" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1213.674197] env[65726]: DEBUG nova.network.neutron [req-244cdfaf-04db-4d6c-9730-0f5000754417 req-809a7728-1743-47f8-85c2-84dd9ea79b07 service nova] [instance: 86c23794-5d82-4a7a-aec6-de91601177dd] Refreshing network info cache for port dd9340e1-39b8-47ca-91f2-eec36706a058 {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 1213.767391] env[65726]: DEBUG oslo_vmware.api [None req-e32abc12-7404-4281-87f0-12de5cc1731f tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5116959, 'name': CopyVirtualDisk_Task} progress is 12%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1213.783372] env[65726]: DEBUG oslo_vmware.api [None req-c0cee6b4-64a9-4ea6-9254-dda6dc407a15 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Task: {'id': task-5116960, 'name': ReconfigVM_Task, 'duration_secs': 0.517807} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1213.787016] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-c0cee6b4-64a9-4ea6-9254-dda6dc407a15 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] [instance: 5bfbaf25-5aed-46bc-97fc-f138d08ebbde] Reconfigured VM instance instance-0000006b to attach disk [datastore1] volume-a465c46c-664d-4aca-85ae-db25aaeeee00/volume-a465c46c-664d-4aca-85ae-db25aaeeee00.vmdk or device None with type thin {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1213.792257] env[65726]: DEBUG oslo_vmware.api [None req-70bb45c2-ad53-4082-9238-425c98dcfe4d tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Task: {'id': task-5116955, 'name': CopyVirtualDisk_Task, 'duration_secs': 4.343826} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1213.792623] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a2ef3cf2-f278-4ead-a6e3-c66925854076 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1213.804465] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-70bb45c2-ad53-4082-9238-425c98dcfe4d tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/745a44b6-0c3f-4396-a82a-6e9472e10925/745a44b6-0c3f-4396-a82a-6e9472e10925.vmdk to [datastore1] b353b5e9-500c-42d3-a87f-880a624febec/b353b5e9-500c-42d3-a87f-880a624febec.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1213.805471] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03be7ec6-4654-4a52-9625-3e7639fa76eb {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1213.833402] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-70bb45c2-ad53-4082-9238-425c98dcfe4d tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: b353b5e9-500c-42d3-a87f-880a624febec] Reconfiguring VM instance instance-00000067 to attach disk [datastore1] b353b5e9-500c-42d3-a87f-880a624febec/b353b5e9-500c-42d3-a87f-880a624febec.vmdk or device None with type streamOptimized {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1213.835681] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-daaf5784-d151-4f2b-ad41-eb4e08999a12 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1213.851952] env[65726]: DEBUG oslo_vmware.api [None req-c0cee6b4-64a9-4ea6-9254-dda6dc407a15 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Waiting for the task: (returnval){ [ 1213.851952] env[65726]: value = "task-5116962" [ 1213.851952] env[65726]: _type = "Task" [ 1213.851952] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1213.858700] env[65726]: DEBUG oslo_vmware.api [None req-70bb45c2-ad53-4082-9238-425c98dcfe4d tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Waiting for the task: (returnval){ [ 1213.858700] env[65726]: value = "task-5116963" [ 1213.858700] env[65726]: _type = "Task" [ 1213.858700] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1213.862767] env[65726]: DEBUG oslo_vmware.api [None req-c0cee6b4-64a9-4ea6-9254-dda6dc407a15 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Task: {'id': task-5116962, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1213.871907] env[65726]: DEBUG oslo_vmware.api [None req-70bb45c2-ad53-4082-9238-425c98dcfe4d tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Task: {'id': task-5116963, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1213.974312] env[65726]: INFO nova.compute.manager [-] [instance: 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30] Took 1.48 seconds to deallocate network for instance. [ 1214.058157] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116961, 'name': CreateVM_Task, 'duration_secs': 0.440499} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1214.058157] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 86c23794-5d82-4a7a-aec6-de91601177dd] Created VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1214.058843] env[65726]: WARNING openstack [None req-c1e079dd-56f2-4cf7-8616-b6ff59feac77 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1214.059167] env[65726]: WARNING openstack [None req-c1e079dd-56f2-4cf7-8616-b6ff59feac77 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1214.064742] env[65726]: WARNING openstack [None req-c1e079dd-56f2-4cf7-8616-b6ff59feac77 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1214.065204] env[65726]: WARNING openstack [None req-c1e079dd-56f2-4cf7-8616-b6ff59feac77 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1214.145344] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c1e079dd-56f2-4cf7-8616-b6ff59feac77 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1214.145652] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c1e079dd-56f2-4cf7-8616-b6ff59feac77 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1214.145996] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c1e079dd-56f2-4cf7-8616-b6ff59feac77 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1214.146335] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-738daa1c-cd8a-4c07-a70f-f33acd2aaaac {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1214.152977] env[65726]: DEBUG oslo_vmware.api [None req-c1e079dd-56f2-4cf7-8616-b6ff59feac77 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Waiting for the task: (returnval){ [ 1214.152977] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52382d0c-09f1-e2c3-3012-d7537cfd82b7" [ 1214.152977] env[65726]: _type = "Task" [ 1214.152977] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1214.168567] env[65726]: DEBUG oslo_vmware.api [None req-c1e079dd-56f2-4cf7-8616-b6ff59feac77 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52382d0c-09f1-e2c3-3012-d7537cfd82b7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1214.178750] env[65726]: WARNING openstack [req-244cdfaf-04db-4d6c-9730-0f5000754417 req-809a7728-1743-47f8-85c2-84dd9ea79b07 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1214.178931] env[65726]: WARNING openstack [req-244cdfaf-04db-4d6c-9730-0f5000754417 req-809a7728-1743-47f8-85c2-84dd9ea79b07 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1214.271186] env[65726]: DEBUG oslo_vmware.api [None req-e32abc12-7404-4281-87f0-12de5cc1731f tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5116959, 'name': CopyVirtualDisk_Task} progress is 29%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1214.282889] env[65726]: WARNING openstack [req-244cdfaf-04db-4d6c-9730-0f5000754417 req-809a7728-1743-47f8-85c2-84dd9ea79b07 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1214.283289] env[65726]: WARNING openstack [req-244cdfaf-04db-4d6c-9730-0f5000754417 req-809a7728-1743-47f8-85c2-84dd9ea79b07 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1214.351282] env[65726]: WARNING openstack [req-244cdfaf-04db-4d6c-9730-0f5000754417 req-809a7728-1743-47f8-85c2-84dd9ea79b07 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1214.352587] env[65726]: WARNING openstack [req-244cdfaf-04db-4d6c-9730-0f5000754417 req-809a7728-1743-47f8-85c2-84dd9ea79b07 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1214.372417] env[65726]: DEBUG oslo_vmware.api [None req-c0cee6b4-64a9-4ea6-9254-dda6dc407a15 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Task: {'id': task-5116962, 'name': ReconfigVM_Task, 'duration_secs': 0.18941} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1214.375839] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-c0cee6b4-64a9-4ea6-9254-dda6dc407a15 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] [instance: 5bfbaf25-5aed-46bc-97fc-f138d08ebbde] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-995333', 'volume_id': 'a465c46c-664d-4aca-85ae-db25aaeeee00', 'name': 'volume-a465c46c-664d-4aca-85ae-db25aaeeee00', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '5bfbaf25-5aed-46bc-97fc-f138d08ebbde', 'attached_at': '', 'detached_at': '', 'volume_id': 'a465c46c-664d-4aca-85ae-db25aaeeee00', 'serial': 'a465c46c-664d-4aca-85ae-db25aaeeee00'} {{(pid=65726) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1214.377399] env[65726]: DEBUG oslo_vmware.api [None req-70bb45c2-ad53-4082-9238-425c98dcfe4d tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Task: {'id': task-5116963, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1214.458332] env[65726]: DEBUG nova.network.neutron [req-244cdfaf-04db-4d6c-9730-0f5000754417 req-809a7728-1743-47f8-85c2-84dd9ea79b07 service nova] [instance: 86c23794-5d82-4a7a-aec6-de91601177dd] Updated VIF entry in instance network info cache for port dd9340e1-39b8-47ca-91f2-eec36706a058. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 1214.458731] env[65726]: DEBUG nova.network.neutron [req-244cdfaf-04db-4d6c-9730-0f5000754417 req-809a7728-1743-47f8-85c2-84dd9ea79b07 service nova] [instance: 86c23794-5d82-4a7a-aec6-de91601177dd] Updating instance_info_cache with network_info: [{"id": "dd9340e1-39b8-47ca-91f2-eec36706a058", "address": "fa:16:3e:c4:39:5d", "network": {"id": "69642329-e986-48fa-accf-1f296c08116a", "bridge": "br-int", "label": "tempest-ServersTestJSON-1786538788-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f3f799aa94f64e8cb26d93f6124efd81", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "816c6e38-e200-4544-8c5b-9fc3e16c5761", "external-id": "nsx-vlan-transportzone-195", "segmentation_id": 195, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdd9340e1-39", "ovs_interfaceid": "dd9340e1-39b8-47ca-91f2-eec36706a058", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1214.527474] env[65726]: INFO nova.compute.manager [None req-76669e62-2e2e-4af0-a67b-7e8917e36e18 tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] [instance: 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30] Took 0.55 seconds to detach 1 volumes for instance. [ 1214.530091] env[65726]: DEBUG nova.compute.manager [None req-76669e62-2e2e-4af0-a67b-7e8917e36e18 tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] [instance: 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30] Deleting volume: 39918e7e-1d4a-45df-911f-46ed9091f96b {{(pid=65726) _cleanup_volumes /opt/stack/nova/nova/compute/manager.py:3319}} [ 1214.666458] env[65726]: DEBUG oslo_vmware.api [None req-c1e079dd-56f2-4cf7-8616-b6ff59feac77 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52382d0c-09f1-e2c3-3012-d7537cfd82b7, 'name': SearchDatastore_Task, 'duration_secs': 0.097183} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1214.666702] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c1e079dd-56f2-4cf7-8616-b6ff59feac77 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1214.666943] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-c1e079dd-56f2-4cf7-8616-b6ff59feac77 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 86c23794-5d82-4a7a-aec6-de91601177dd] Processing image b52362a3-ee8a-4cbf-b06f-513b0cc8f95c {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1214.667252] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c1e079dd-56f2-4cf7-8616-b6ff59feac77 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1214.667402] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c1e079dd-56f2-4cf7-8616-b6ff59feac77 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1214.667591] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-c1e079dd-56f2-4cf7-8616-b6ff59feac77 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1214.667880] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b448c287-6dbc-4fe2-828a-264c027127aa {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1214.689026] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-c1e079dd-56f2-4cf7-8616-b6ff59feac77 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1214.689026] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-c1e079dd-56f2-4cf7-8616-b6ff59feac77 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65726) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1214.689026] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-db6ac6dc-e0fe-42d0-b23c-72b651baef86 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1214.695015] env[65726]: DEBUG oslo_vmware.api [None req-c1e079dd-56f2-4cf7-8616-b6ff59feac77 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Waiting for the task: (returnval){ [ 1214.695015] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52ed1133-1af5-9d2f-b845-214da0554803" [ 1214.695015] env[65726]: _type = "Task" [ 1214.695015] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1214.705290] env[65726]: DEBUG oslo_vmware.api [None req-c1e079dd-56f2-4cf7-8616-b6ff59feac77 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52ed1133-1af5-9d2f-b845-214da0554803, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1214.770436] env[65726]: DEBUG oslo_vmware.api [None req-e32abc12-7404-4281-87f0-12de5cc1731f tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5116959, 'name': CopyVirtualDisk_Task} progress is 49%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1214.877843] env[65726]: DEBUG oslo_vmware.api [None req-70bb45c2-ad53-4082-9238-425c98dcfe4d tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Task: {'id': task-5116963, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1214.961589] env[65726]: DEBUG oslo_concurrency.lockutils [req-244cdfaf-04db-4d6c-9730-0f5000754417 req-809a7728-1743-47f8-85c2-84dd9ea79b07 service nova] Releasing lock "refresh_cache-86c23794-5d82-4a7a-aec6-de91601177dd" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1214.961964] env[65726]: DEBUG nova.compute.manager [req-244cdfaf-04db-4d6c-9730-0f5000754417 req-809a7728-1743-47f8-85c2-84dd9ea79b07 service nova] [instance: 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30] Received event network-vif-deleted-e0f18f03-6a1b-4d5e-9b8e-68051f61ad8b {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1215.074269] env[65726]: DEBUG oslo_concurrency.lockutils [None req-76669e62-2e2e-4af0-a67b-7e8917e36e18 tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1215.074602] env[65726]: DEBUG oslo_concurrency.lockutils [None req-76669e62-2e2e-4af0-a67b-7e8917e36e18 tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1215.074842] env[65726]: DEBUG nova.objects.instance [None req-76669e62-2e2e-4af0-a67b-7e8917e36e18 tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] Lazy-loading 'resources' on Instance uuid 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30 {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1215.208223] env[65726]: DEBUG oslo_vmware.api [None req-c1e079dd-56f2-4cf7-8616-b6ff59feac77 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52ed1133-1af5-9d2f-b845-214da0554803, 'name': SearchDatastore_Task, 'duration_secs': 0.088754} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1215.209120] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d345179c-6fa5-43e1-8075-74dbc0ea1d23 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1215.218377] env[65726]: DEBUG oslo_vmware.api [None req-c1e079dd-56f2-4cf7-8616-b6ff59feac77 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Waiting for the task: (returnval){ [ 1215.218377] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]524892af-50ab-638b-8786-71f8dbda38fd" [ 1215.218377] env[65726]: _type = "Task" [ 1215.218377] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1215.226729] env[65726]: DEBUG oslo_vmware.api [None req-c1e079dd-56f2-4cf7-8616-b6ff59feac77 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]524892af-50ab-638b-8786-71f8dbda38fd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1215.268849] env[65726]: DEBUG oslo_vmware.api [None req-e32abc12-7404-4281-87f0-12de5cc1731f tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5116959, 'name': CopyVirtualDisk_Task} progress is 69%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1215.384105] env[65726]: DEBUG oslo_vmware.api [None req-70bb45c2-ad53-4082-9238-425c98dcfe4d tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Task: {'id': task-5116963, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1215.440842] env[65726]: DEBUG nova.objects.instance [None req-c0cee6b4-64a9-4ea6-9254-dda6dc407a15 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Lazy-loading 'flavor' on Instance uuid 5bfbaf25-5aed-46bc-97fc-f138d08ebbde {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1215.732187] env[65726]: DEBUG oslo_vmware.api [None req-c1e079dd-56f2-4cf7-8616-b6ff59feac77 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]524892af-50ab-638b-8786-71f8dbda38fd, 'name': SearchDatastore_Task, 'duration_secs': 0.099149} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1215.732187] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c1e079dd-56f2-4cf7-8616-b6ff59feac77 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1215.732187] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-c1e079dd-56f2-4cf7-8616-b6ff59feac77 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] 86c23794-5d82-4a7a-aec6-de91601177dd/86c23794-5d82-4a7a-aec6-de91601177dd.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1215.732187] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9837980e-d4ee-4a69-9f7f-fca13eed088b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1215.738041] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57161533-934f-46b2-89cf-dd646c397397 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1215.754392] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d594fdf-41e4-4801-b247-06450212765c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1215.758149] env[65726]: DEBUG oslo_vmware.api [None req-c1e079dd-56f2-4cf7-8616-b6ff59feac77 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Waiting for the task: (returnval){ [ 1215.758149] env[65726]: value = "task-5116965" [ 1215.758149] env[65726]: _type = "Task" [ 1215.758149] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1215.804038] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e92215c4-9160-4fd8-a452-04fcd8f87c48 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1215.814315] env[65726]: DEBUG oslo_vmware.api [None req-c1e079dd-56f2-4cf7-8616-b6ff59feac77 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Task: {'id': task-5116965, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1215.814782] env[65726]: DEBUG oslo_vmware.api [None req-e32abc12-7404-4281-87f0-12de5cc1731f tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5116959, 'name': CopyVirtualDisk_Task} progress is 91%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1215.828139] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38c9e391-2c8c-4fe9-96ce-6f5a0a859a5b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1215.853970] env[65726]: DEBUG nova.compute.provider_tree [None req-76669e62-2e2e-4af0-a67b-7e8917e36e18 tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1215.882021] env[65726]: DEBUG oslo_vmware.api [None req-70bb45c2-ad53-4082-9238-425c98dcfe4d tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Task: {'id': task-5116963, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1215.950245] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c0cee6b4-64a9-4ea6-9254-dda6dc407a15 tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Lock "5bfbaf25-5aed-46bc-97fc-f138d08ebbde" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.390s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1216.275037] env[65726]: DEBUG oslo_vmware.api [None req-e32abc12-7404-4281-87f0-12de5cc1731f tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5116959, 'name': CopyVirtualDisk_Task, 'duration_secs': 3.794395} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1216.275587] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-e32abc12-7404-4281-87f0-12de5cc1731f tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/1d5801e6-cfd0-4c6a-8fa0-57ee72cc963a/1d5801e6-cfd0-4c6a-8fa0-57ee72cc963a.vmdk to [datastore1] 378c1d18-4bb6-4245-80bf-370fb7af0575/378c1d18-4bb6-4245-80bf-370fb7af0575.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1216.275965] env[65726]: DEBUG oslo_vmware.api [None req-c1e079dd-56f2-4cf7-8616-b6ff59feac77 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Task: {'id': task-5116965, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1216.276802] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ab35030-7307-41d0-92da-63e5014749d0 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1216.301491] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-e32abc12-7404-4281-87f0-12de5cc1731f tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 378c1d18-4bb6-4245-80bf-370fb7af0575] Reconfiguring VM instance instance-00000068 to attach disk [datastore1] 378c1d18-4bb6-4245-80bf-370fb7af0575/378c1d18-4bb6-4245-80bf-370fb7af0575.vmdk or device None with type streamOptimized {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1216.301849] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-903db457-dcc0-4d77-a0fd-283f28ac0e53 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1216.324901] env[65726]: DEBUG oslo_vmware.api [None req-e32abc12-7404-4281-87f0-12de5cc1731f tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Waiting for the task: (returnval){ [ 1216.324901] env[65726]: value = "task-5116966" [ 1216.324901] env[65726]: _type = "Task" [ 1216.324901] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1216.328668] env[65726]: DEBUG oslo_concurrency.lockutils [None req-ae769e05-5b34-4565-9e07-8924cc10eaad tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Acquiring lock "5bfbaf25-5aed-46bc-97fc-f138d08ebbde" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1216.328911] env[65726]: DEBUG oslo_concurrency.lockutils [None req-ae769e05-5b34-4565-9e07-8924cc10eaad tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Lock "5bfbaf25-5aed-46bc-97fc-f138d08ebbde" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1216.337406] env[65726]: DEBUG oslo_vmware.api [None req-e32abc12-7404-4281-87f0-12de5cc1731f tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5116966, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1216.358223] env[65726]: DEBUG nova.scheduler.client.report [None req-76669e62-2e2e-4af0-a67b-7e8917e36e18 tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1216.384652] env[65726]: DEBUG oslo_vmware.api [None req-70bb45c2-ad53-4082-9238-425c98dcfe4d tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Task: {'id': task-5116963, 'name': ReconfigVM_Task, 'duration_secs': 2.29794} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1216.386076] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-70bb45c2-ad53-4082-9238-425c98dcfe4d tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: b353b5e9-500c-42d3-a87f-880a624febec] Reconfigured VM instance instance-00000067 to attach disk [datastore1] b353b5e9-500c-42d3-a87f-880a624febec/b353b5e9-500c-42d3-a87f-880a624febec.vmdk or device None with type streamOptimized {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1216.386905] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-92e58a03-e214-43af-9d26-3ddf2a8542c6 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1216.399301] env[65726]: DEBUG oslo_vmware.api [None req-70bb45c2-ad53-4082-9238-425c98dcfe4d tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Waiting for the task: (returnval){ [ 1216.399301] env[65726]: value = "task-5116967" [ 1216.399301] env[65726]: _type = "Task" [ 1216.399301] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1216.408638] env[65726]: DEBUG oslo_vmware.api [None req-70bb45c2-ad53-4082-9238-425c98dcfe4d tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Task: {'id': task-5116967, 'name': Rename_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1216.770962] env[65726]: DEBUG oslo_vmware.api [None req-c1e079dd-56f2-4cf7-8616-b6ff59feac77 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Task: {'id': task-5116965, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.827583} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1216.771275] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-c1e079dd-56f2-4cf7-8616-b6ff59feac77 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] 86c23794-5d82-4a7a-aec6-de91601177dd/86c23794-5d82-4a7a-aec6-de91601177dd.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1216.771483] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-c1e079dd-56f2-4cf7-8616-b6ff59feac77 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 86c23794-5d82-4a7a-aec6-de91601177dd] Extending root virtual disk to 1048576 {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1216.771748] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e5d642ac-c309-4bb4-8029-efe7cdb27631 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1216.781012] env[65726]: DEBUG oslo_vmware.api [None req-c1e079dd-56f2-4cf7-8616-b6ff59feac77 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Waiting for the task: (returnval){ [ 1216.781012] env[65726]: value = "task-5116968" [ 1216.781012] env[65726]: _type = "Task" [ 1216.781012] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1216.792313] env[65726]: DEBUG oslo_vmware.api [None req-c1e079dd-56f2-4cf7-8616-b6ff59feac77 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Task: {'id': task-5116968, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1216.832740] env[65726]: INFO nova.compute.manager [None req-ae769e05-5b34-4565-9e07-8924cc10eaad tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] [instance: 5bfbaf25-5aed-46bc-97fc-f138d08ebbde] Detaching volume a86c4f74-2f48-4ced-aa11-e174e767f2a7 [ 1216.841308] env[65726]: DEBUG oslo_vmware.api [None req-e32abc12-7404-4281-87f0-12de5cc1731f tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5116966, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1216.867564] env[65726]: DEBUG oslo_concurrency.lockutils [None req-76669e62-2e2e-4af0-a67b-7e8917e36e18 tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.793s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1216.874896] env[65726]: INFO nova.virt.block_device [None req-ae769e05-5b34-4565-9e07-8924cc10eaad tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] [instance: 5bfbaf25-5aed-46bc-97fc-f138d08ebbde] Attempting to driver detach volume a86c4f74-2f48-4ced-aa11-e174e767f2a7 from mountpoint /dev/sdb [ 1216.877018] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-ae769e05-5b34-4565-9e07-8924cc10eaad tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] [instance: 5bfbaf25-5aed-46bc-97fc-f138d08ebbde] Volume detach. Driver type: vmdk {{(pid=65726) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1216.877018] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-ae769e05-5b34-4565-9e07-8924cc10eaad tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] [instance: 5bfbaf25-5aed-46bc-97fc-f138d08ebbde] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-995328', 'volume_id': 'a86c4f74-2f48-4ced-aa11-e174e767f2a7', 'name': 'volume-a86c4f74-2f48-4ced-aa11-e174e767f2a7', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '5bfbaf25-5aed-46bc-97fc-f138d08ebbde', 'attached_at': '', 'detached_at': '', 'volume_id': 'a86c4f74-2f48-4ced-aa11-e174e767f2a7', 'serial': 'a86c4f74-2f48-4ced-aa11-e174e767f2a7'} {{(pid=65726) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1216.877018] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2921e691-29b8-47f5-a475-a6f9c14e56aa {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1216.922335] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79eefca6-a693-4e41-b6ea-432fd1476a94 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1216.930887] env[65726]: DEBUG oslo_vmware.api [None req-70bb45c2-ad53-4082-9238-425c98dcfe4d tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Task: {'id': task-5116967, 'name': Rename_Task, 'duration_secs': 0.302164} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1216.933091] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-70bb45c2-ad53-4082-9238-425c98dcfe4d tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: b353b5e9-500c-42d3-a87f-880a624febec] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1216.933545] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-77a44dbe-47d6-4cbf-94fc-571f24f1f3f6 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1216.935972] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6fbb7da-90f5-4335-9cc0-ee2c6a024751 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1216.964316] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-855b40a1-1dcb-433a-aa1b-f23afb3e1191 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1216.967554] env[65726]: DEBUG oslo_vmware.api [None req-70bb45c2-ad53-4082-9238-425c98dcfe4d tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Waiting for the task: (returnval){ [ 1216.967554] env[65726]: value = "task-5116969" [ 1216.967554] env[65726]: _type = "Task" [ 1216.967554] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1216.982812] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-ae769e05-5b34-4565-9e07-8924cc10eaad tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] The volume has not been displaced from its original location: [datastore1] volume-a86c4f74-2f48-4ced-aa11-e174e767f2a7/volume-a86c4f74-2f48-4ced-aa11-e174e767f2a7.vmdk. No consolidation needed. {{(pid=65726) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1216.988801] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-ae769e05-5b34-4565-9e07-8924cc10eaad tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] [instance: 5bfbaf25-5aed-46bc-97fc-f138d08ebbde] Reconfiguring VM instance instance-0000006b to detach disk 2001 {{(pid=65726) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1216.989734] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6ebd098c-a5ea-4e32-8965-aa7cbe9b784a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1217.007039] env[65726]: DEBUG oslo_vmware.api [None req-70bb45c2-ad53-4082-9238-425c98dcfe4d tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Task: {'id': task-5116969, 'name': PowerOnVM_Task} progress is 33%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1217.016072] env[65726]: DEBUG oslo_vmware.api [None req-ae769e05-5b34-4565-9e07-8924cc10eaad tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Waiting for the task: (returnval){ [ 1217.016072] env[65726]: value = "task-5116970" [ 1217.016072] env[65726]: _type = "Task" [ 1217.016072] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1217.031187] env[65726]: DEBUG oslo_vmware.api [None req-ae769e05-5b34-4565-9e07-8924cc10eaad tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Task: {'id': task-5116970, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1217.294712] env[65726]: DEBUG oslo_vmware.api [None req-c1e079dd-56f2-4cf7-8616-b6ff59feac77 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Task: {'id': task-5116968, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.095766} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1217.295033] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-c1e079dd-56f2-4cf7-8616-b6ff59feac77 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 86c23794-5d82-4a7a-aec6-de91601177dd] Extended root virtual disk {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1217.296230] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c05a288-f561-4b41-bfd5-16381588dc77 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1217.323844] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-c1e079dd-56f2-4cf7-8616-b6ff59feac77 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 86c23794-5d82-4a7a-aec6-de91601177dd] Reconfiguring VM instance instance-00000072 to attach disk [datastore1] 86c23794-5d82-4a7a-aec6-de91601177dd/86c23794-5d82-4a7a-aec6-de91601177dd.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1217.324464] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cd0ba356-c0f1-47b1-a060-ee7c40982155 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1217.349331] env[65726]: DEBUG oslo_vmware.api [None req-e32abc12-7404-4281-87f0-12de5cc1731f tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5116966, 'name': ReconfigVM_Task, 'duration_secs': 0.583087} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1217.350991] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-e32abc12-7404-4281-87f0-12de5cc1731f tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 378c1d18-4bb6-4245-80bf-370fb7af0575] Reconfigured VM instance instance-00000068 to attach disk [datastore1] 378c1d18-4bb6-4245-80bf-370fb7af0575/378c1d18-4bb6-4245-80bf-370fb7af0575.vmdk or device None with type streamOptimized {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1217.352070] env[65726]: DEBUG oslo_vmware.api [None req-c1e079dd-56f2-4cf7-8616-b6ff59feac77 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Waiting for the task: (returnval){ [ 1217.352070] env[65726]: value = "task-5116971" [ 1217.352070] env[65726]: _type = "Task" [ 1217.352070] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1217.352259] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-bda19763-fe20-41a4-846e-019e3df92845 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1217.367230] env[65726]: DEBUG oslo_vmware.api [None req-c1e079dd-56f2-4cf7-8616-b6ff59feac77 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Task: {'id': task-5116971, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1217.369038] env[65726]: DEBUG oslo_vmware.api [None req-e32abc12-7404-4281-87f0-12de5cc1731f tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Waiting for the task: (returnval){ [ 1217.369038] env[65726]: value = "task-5116972" [ 1217.369038] env[65726]: _type = "Task" [ 1217.369038] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1217.380948] env[65726]: DEBUG oslo_vmware.api [None req-e32abc12-7404-4281-87f0-12de5cc1731f tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5116972, 'name': Rename_Task} progress is 6%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1217.428655] env[65726]: DEBUG oslo_concurrency.lockutils [None req-76669e62-2e2e-4af0-a67b-7e8917e36e18 tempest-ServerActionsV293TestJSON-1478069518 tempest-ServerActionsV293TestJSON-1478069518-project-member] Lock "47e52bc1-c18e-4aa2-82a2-e4cb030a7a30" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.539s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1217.479390] env[65726]: DEBUG oslo_vmware.api [None req-70bb45c2-ad53-4082-9238-425c98dcfe4d tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Task: {'id': task-5116969, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1217.528638] env[65726]: DEBUG oslo_vmware.api [None req-ae769e05-5b34-4565-9e07-8924cc10eaad tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Task: {'id': task-5116970, 'name': ReconfigVM_Task, 'duration_secs': 0.493535} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1217.528933] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-ae769e05-5b34-4565-9e07-8924cc10eaad tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] [instance: 5bfbaf25-5aed-46bc-97fc-f138d08ebbde] Reconfigured VM instance instance-0000006b to detach disk 2001 {{(pid=65726) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1217.534793] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c61a34be-e85a-4967-914e-2c466e3f523e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1217.553049] env[65726]: DEBUG oslo_vmware.api [None req-ae769e05-5b34-4565-9e07-8924cc10eaad tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Waiting for the task: (returnval){ [ 1217.553049] env[65726]: value = "task-5116973" [ 1217.553049] env[65726]: _type = "Task" [ 1217.553049] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1217.567373] env[65726]: DEBUG oslo_vmware.api [None req-ae769e05-5b34-4565-9e07-8924cc10eaad tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Task: {'id': task-5116973, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1217.864514] env[65726]: DEBUG oslo_vmware.api [None req-c1e079dd-56f2-4cf7-8616-b6ff59feac77 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Task: {'id': task-5116971, 'name': ReconfigVM_Task, 'duration_secs': 0.447599} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1217.865061] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-c1e079dd-56f2-4cf7-8616-b6ff59feac77 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 86c23794-5d82-4a7a-aec6-de91601177dd] Reconfigured VM instance instance-00000072 to attach disk [datastore1] 86c23794-5d82-4a7a-aec6-de91601177dd/86c23794-5d82-4a7a-aec6-de91601177dd.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1217.865960] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-74a2dc32-5414-49ad-a9c6-ccc68cfc60f5 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1217.875491] env[65726]: DEBUG oslo_vmware.api [None req-c1e079dd-56f2-4cf7-8616-b6ff59feac77 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Waiting for the task: (returnval){ [ 1217.875491] env[65726]: value = "task-5116974" [ 1217.875491] env[65726]: _type = "Task" [ 1217.875491] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1217.884887] env[65726]: DEBUG oslo_vmware.api [None req-e32abc12-7404-4281-87f0-12de5cc1731f tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5116972, 'name': Rename_Task, 'duration_secs': 0.217844} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1217.885486] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-e32abc12-7404-4281-87f0-12de5cc1731f tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 378c1d18-4bb6-4245-80bf-370fb7af0575] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1217.885810] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-68a66e6e-b95e-4ef9-9d2f-a3f6fc6672a8 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1217.890617] env[65726]: DEBUG oslo_vmware.api [None req-c1e079dd-56f2-4cf7-8616-b6ff59feac77 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Task: {'id': task-5116974, 'name': Rename_Task} progress is 6%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1217.896933] env[65726]: DEBUG oslo_vmware.api [None req-e32abc12-7404-4281-87f0-12de5cc1731f tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Waiting for the task: (returnval){ [ 1217.896933] env[65726]: value = "task-5116975" [ 1217.896933] env[65726]: _type = "Task" [ 1217.896933] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1217.908957] env[65726]: DEBUG oslo_vmware.api [None req-e32abc12-7404-4281-87f0-12de5cc1731f tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5116975, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1217.984585] env[65726]: DEBUG oslo_vmware.api [None req-70bb45c2-ad53-4082-9238-425c98dcfe4d tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Task: {'id': task-5116969, 'name': PowerOnVM_Task, 'duration_secs': 0.707059} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1217.986138] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-70bb45c2-ad53-4082-9238-425c98dcfe4d tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: b353b5e9-500c-42d3-a87f-880a624febec] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1218.064629] env[65726]: DEBUG oslo_vmware.api [None req-ae769e05-5b34-4565-9e07-8924cc10eaad tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Task: {'id': task-5116973, 'name': ReconfigVM_Task, 'duration_secs': 0.176786} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1218.064965] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-ae769e05-5b34-4565-9e07-8924cc10eaad tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] [instance: 5bfbaf25-5aed-46bc-97fc-f138d08ebbde] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-995328', 'volume_id': 'a86c4f74-2f48-4ced-aa11-e174e767f2a7', 'name': 'volume-a86c4f74-2f48-4ced-aa11-e174e767f2a7', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '5bfbaf25-5aed-46bc-97fc-f138d08ebbde', 'attached_at': '', 'detached_at': '', 'volume_id': 'a86c4f74-2f48-4ced-aa11-e174e767f2a7', 'serial': 'a86c4f74-2f48-4ced-aa11-e174e767f2a7'} {{(pid=65726) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1218.113018] env[65726]: DEBUG nova.compute.manager [None req-70bb45c2-ad53-4082-9238-425c98dcfe4d tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: b353b5e9-500c-42d3-a87f-880a624febec] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1218.113986] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6b33f5b-7c03-4cd5-bf25-ed01785b986c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1218.387405] env[65726]: DEBUG oslo_vmware.api [None req-c1e079dd-56f2-4cf7-8616-b6ff59feac77 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Task: {'id': task-5116974, 'name': Rename_Task} progress is 99%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1218.408386] env[65726]: DEBUG oslo_vmware.api [None req-e32abc12-7404-4281-87f0-12de5cc1731f tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5116975, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1218.623921] env[65726]: DEBUG nova.objects.instance [None req-ae769e05-5b34-4565-9e07-8924cc10eaad tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Lazy-loading 'flavor' on Instance uuid 5bfbaf25-5aed-46bc-97fc-f138d08ebbde {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1218.633154] env[65726]: DEBUG oslo_concurrency.lockutils [None req-70bb45c2-ad53-4082-9238-425c98dcfe4d tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Lock "b353b5e9-500c-42d3-a87f-880a624febec" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 27.934s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1218.634291] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Lock "b353b5e9-500c-42d3-a87f-880a624febec" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 18.592s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1218.634548] env[65726]: INFO nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] [instance: b353b5e9-500c-42d3-a87f-880a624febec] During sync_power_state the instance has a pending task (spawning). Skip. [ 1218.635483] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Lock "b353b5e9-500c-42d3-a87f-880a624febec" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.001s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1218.887740] env[65726]: DEBUG oslo_vmware.api [None req-c1e079dd-56f2-4cf7-8616-b6ff59feac77 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Task: {'id': task-5116974, 'name': Rename_Task} progress is 99%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1218.908826] env[65726]: DEBUG oslo_vmware.api [None req-e32abc12-7404-4281-87f0-12de5cc1731f tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5116975, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1219.388094] env[65726]: DEBUG oslo_vmware.api [None req-c1e079dd-56f2-4cf7-8616-b6ff59feac77 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Task: {'id': task-5116974, 'name': Rename_Task, 'duration_secs': 1.142403} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1219.388432] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-c1e079dd-56f2-4cf7-8616-b6ff59feac77 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 86c23794-5d82-4a7a-aec6-de91601177dd] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1219.388689] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a3fbf548-8d8c-423a-a756-3cf5ddcfa8d2 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1219.395623] env[65726]: DEBUG oslo_vmware.api [None req-c1e079dd-56f2-4cf7-8616-b6ff59feac77 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Waiting for the task: (returnval){ [ 1219.395623] env[65726]: value = "task-5116976" [ 1219.395623] env[65726]: _type = "Task" [ 1219.395623] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1219.414594] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c804e386-b05c-462f-b1f2-e4f32318c412 tempest-ServerPasswordTestJSON-84500404 tempest-ServerPasswordTestJSON-84500404-project-member] Acquiring lock "a1dc9762-ad05-4cd4-b3c3-eb644ba4212b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1219.414915] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c804e386-b05c-462f-b1f2-e4f32318c412 tempest-ServerPasswordTestJSON-84500404 tempest-ServerPasswordTestJSON-84500404-project-member] Lock "a1dc9762-ad05-4cd4-b3c3-eb644ba4212b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1219.416194] env[65726]: DEBUG oslo_vmware.api [None req-c1e079dd-56f2-4cf7-8616-b6ff59feac77 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Task: {'id': task-5116976, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1219.416461] env[65726]: DEBUG oslo_vmware.api [None req-e32abc12-7404-4281-87f0-12de5cc1731f tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5116975, 'name': PowerOnVM_Task, 'duration_secs': 1.166435} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1219.417329] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-e32abc12-7404-4281-87f0-12de5cc1731f tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 378c1d18-4bb6-4245-80bf-370fb7af0575] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1219.549853] env[65726]: DEBUG nova.compute.manager [None req-e32abc12-7404-4281-87f0-12de5cc1731f tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 378c1d18-4bb6-4245-80bf-370fb7af0575] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1219.549853] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a706fb3-327e-403e-979c-1eb343e49609 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1219.632026] env[65726]: DEBUG oslo_concurrency.lockutils [None req-ae769e05-5b34-4565-9e07-8924cc10eaad tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Lock "5bfbaf25-5aed-46bc-97fc-f138d08ebbde" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.303s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1219.739782] env[65726]: DEBUG oslo_concurrency.lockutils [None req-339d6311-4d4e-4252-84dd-03d4bb82a42a tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Acquiring lock "5bfbaf25-5aed-46bc-97fc-f138d08ebbde" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1219.740066] env[65726]: DEBUG oslo_concurrency.lockutils [None req-339d6311-4d4e-4252-84dd-03d4bb82a42a tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Lock "5bfbaf25-5aed-46bc-97fc-f138d08ebbde" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1219.906525] env[65726]: DEBUG oslo_vmware.api [None req-c1e079dd-56f2-4cf7-8616-b6ff59feac77 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Task: {'id': task-5116976, 'name': PowerOnVM_Task, 'duration_secs': 0.46487} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1219.906806] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-c1e079dd-56f2-4cf7-8616-b6ff59feac77 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 86c23794-5d82-4a7a-aec6-de91601177dd] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1219.907027] env[65726]: INFO nova.compute.manager [None req-c1e079dd-56f2-4cf7-8616-b6ff59feac77 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 86c23794-5d82-4a7a-aec6-de91601177dd] Took 8.76 seconds to spawn the instance on the hypervisor. [ 1219.907433] env[65726]: DEBUG nova.compute.manager [None req-c1e079dd-56f2-4cf7-8616-b6ff59feac77 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 86c23794-5d82-4a7a-aec6-de91601177dd] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1219.908429] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03c19231-07ed-47b3-87b1-a4db9732e58f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1219.918685] env[65726]: DEBUG nova.compute.manager [None req-c804e386-b05c-462f-b1f2-e4f32318c412 tempest-ServerPasswordTestJSON-84500404 tempest-ServerPasswordTestJSON-84500404-project-member] [instance: a1dc9762-ad05-4cd4-b3c3-eb644ba4212b] Starting instance... {{(pid=65726) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 1220.061083] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e32abc12-7404-4281-87f0-12de5cc1731f tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Lock "378c1d18-4bb6-4245-80bf-370fb7af0575" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 28.045s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1220.062124] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Lock "378c1d18-4bb6-4245-80bf-370fb7af0575" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 20.020s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1220.062317] env[65726]: INFO nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] [instance: 378c1d18-4bb6-4245-80bf-370fb7af0575] During sync_power_state the instance has a pending task (spawning). Skip. [ 1220.062486] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Lock "378c1d18-4bb6-4245-80bf-370fb7af0575" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1220.243495] env[65726]: INFO nova.compute.manager [None req-339d6311-4d4e-4252-84dd-03d4bb82a42a tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] [instance: 5bfbaf25-5aed-46bc-97fc-f138d08ebbde] Detaching volume a465c46c-664d-4aca-85ae-db25aaeeee00 [ 1220.283302] env[65726]: INFO nova.virt.block_device [None req-339d6311-4d4e-4252-84dd-03d4bb82a42a tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] [instance: 5bfbaf25-5aed-46bc-97fc-f138d08ebbde] Attempting to driver detach volume a465c46c-664d-4aca-85ae-db25aaeeee00 from mountpoint /dev/sdc [ 1220.283827] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-339d6311-4d4e-4252-84dd-03d4bb82a42a tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] [instance: 5bfbaf25-5aed-46bc-97fc-f138d08ebbde] Volume detach. Driver type: vmdk {{(pid=65726) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1220.283827] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-339d6311-4d4e-4252-84dd-03d4bb82a42a tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] [instance: 5bfbaf25-5aed-46bc-97fc-f138d08ebbde] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-995333', 'volume_id': 'a465c46c-664d-4aca-85ae-db25aaeeee00', 'name': 'volume-a465c46c-664d-4aca-85ae-db25aaeeee00', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '5bfbaf25-5aed-46bc-97fc-f138d08ebbde', 'attached_at': '', 'detached_at': '', 'volume_id': 'a465c46c-664d-4aca-85ae-db25aaeeee00', 'serial': 'a465c46c-664d-4aca-85ae-db25aaeeee00'} {{(pid=65726) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1220.284716] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a48cf908-c1bb-48a3-a396-2c405b524503 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1220.308285] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0700f62-2f5d-42e9-a505-5da28968f6ab {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1220.316519] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-759dc5ae-5698-40ff-ad68-41302613113d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1220.338597] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-923a2f5f-4fd4-4ceb-b617-abb027d05072 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1220.354628] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-339d6311-4d4e-4252-84dd-03d4bb82a42a tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] The volume has not been displaced from its original location: [datastore1] volume-a465c46c-664d-4aca-85ae-db25aaeeee00/volume-a465c46c-664d-4aca-85ae-db25aaeeee00.vmdk. No consolidation needed. {{(pid=65726) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1220.360068] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-339d6311-4d4e-4252-84dd-03d4bb82a42a tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] [instance: 5bfbaf25-5aed-46bc-97fc-f138d08ebbde] Reconfiguring VM instance instance-0000006b to detach disk 2002 {{(pid=65726) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1220.360386] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3f5bb881-eb19-4d25-9b58-59f9952f89ff {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1220.381067] env[65726]: DEBUG oslo_vmware.api [None req-339d6311-4d4e-4252-84dd-03d4bb82a42a tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Waiting for the task: (returnval){ [ 1220.381067] env[65726]: value = "task-5116977" [ 1220.381067] env[65726]: _type = "Task" [ 1220.381067] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1220.392118] env[65726]: DEBUG oslo_vmware.api [None req-339d6311-4d4e-4252-84dd-03d4bb82a42a tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Task: {'id': task-5116977, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1220.440880] env[65726]: INFO nova.compute.manager [None req-c1e079dd-56f2-4cf7-8616-b6ff59feac77 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 86c23794-5d82-4a7a-aec6-de91601177dd] Took 14.65 seconds to build instance. [ 1220.458531] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c804e386-b05c-462f-b1f2-e4f32318c412 tempest-ServerPasswordTestJSON-84500404 tempest-ServerPasswordTestJSON-84500404-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1220.458767] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c804e386-b05c-462f-b1f2-e4f32318c412 tempest-ServerPasswordTestJSON-84500404 tempest-ServerPasswordTestJSON-84500404-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1220.460723] env[65726]: INFO nova.compute.claims [None req-c804e386-b05c-462f-b1f2-e4f32318c412 tempest-ServerPasswordTestJSON-84500404 tempest-ServerPasswordTestJSON-84500404-project-member] [instance: a1dc9762-ad05-4cd4-b3c3-eb644ba4212b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1220.891155] env[65726]: DEBUG oslo_vmware.api [None req-339d6311-4d4e-4252-84dd-03d4bb82a42a tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Task: {'id': task-5116977, 'name': ReconfigVM_Task, 'duration_secs': 0.246647} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1220.891607] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-339d6311-4d4e-4252-84dd-03d4bb82a42a tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] [instance: 5bfbaf25-5aed-46bc-97fc-f138d08ebbde] Reconfigured VM instance instance-0000006b to detach disk 2002 {{(pid=65726) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1220.897607] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cb5eb805-75a1-444d-8e22-b36f816efc52 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1220.913907] env[65726]: DEBUG oslo_vmware.api [None req-339d6311-4d4e-4252-84dd-03d4bb82a42a tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Waiting for the task: (returnval){ [ 1220.913907] env[65726]: value = "task-5116978" [ 1220.913907] env[65726]: _type = "Task" [ 1220.913907] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1220.923335] env[65726]: DEBUG oslo_vmware.api [None req-339d6311-4d4e-4252-84dd-03d4bb82a42a tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Task: {'id': task-5116978, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1220.943133] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c1e079dd-56f2-4cf7-8616-b6ff59feac77 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Lock "86c23794-5d82-4a7a-aec6-de91601177dd" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.163s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1220.975212] env[65726]: DEBUG oslo_concurrency.lockutils [None req-8f1940db-b066-401a-a91a-8c578c79d9a0 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Acquiring lock "378c1d18-4bb6-4245-80bf-370fb7af0575" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1220.975477] env[65726]: DEBUG oslo_concurrency.lockutils [None req-8f1940db-b066-401a-a91a-8c578c79d9a0 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Lock "378c1d18-4bb6-4245-80bf-370fb7af0575" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1220.975693] env[65726]: DEBUG oslo_concurrency.lockutils [None req-8f1940db-b066-401a-a91a-8c578c79d9a0 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Acquiring lock "378c1d18-4bb6-4245-80bf-370fb7af0575-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1220.975864] env[65726]: DEBUG oslo_concurrency.lockutils [None req-8f1940db-b066-401a-a91a-8c578c79d9a0 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Lock "378c1d18-4bb6-4245-80bf-370fb7af0575-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1220.976062] env[65726]: DEBUG oslo_concurrency.lockutils [None req-8f1940db-b066-401a-a91a-8c578c79d9a0 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Lock "378c1d18-4bb6-4245-80bf-370fb7af0575-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1220.978151] env[65726]: INFO nova.compute.manager [None req-8f1940db-b066-401a-a91a-8c578c79d9a0 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 378c1d18-4bb6-4245-80bf-370fb7af0575] Terminating instance [ 1220.989433] env[65726]: DEBUG oslo_concurrency.lockutils [None req-df9c79d3-14ef-4842-a2ad-e0f00988e3ba tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Acquiring lock "86c23794-5d82-4a7a-aec6-de91601177dd" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1220.989677] env[65726]: DEBUG oslo_concurrency.lockutils [None req-df9c79d3-14ef-4842-a2ad-e0f00988e3ba tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Lock "86c23794-5d82-4a7a-aec6-de91601177dd" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1220.989868] env[65726]: DEBUG oslo_concurrency.lockutils [None req-df9c79d3-14ef-4842-a2ad-e0f00988e3ba tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Acquiring lock "86c23794-5d82-4a7a-aec6-de91601177dd-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1220.990055] env[65726]: DEBUG oslo_concurrency.lockutils [None req-df9c79d3-14ef-4842-a2ad-e0f00988e3ba tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Lock "86c23794-5d82-4a7a-aec6-de91601177dd-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1220.990255] env[65726]: DEBUG oslo_concurrency.lockutils [None req-df9c79d3-14ef-4842-a2ad-e0f00988e3ba tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Lock "86c23794-5d82-4a7a-aec6-de91601177dd-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1220.992198] env[65726]: INFO nova.compute.manager [None req-df9c79d3-14ef-4842-a2ad-e0f00988e3ba tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 86c23794-5d82-4a7a-aec6-de91601177dd] Terminating instance [ 1221.425960] env[65726]: DEBUG oslo_vmware.api [None req-339d6311-4d4e-4252-84dd-03d4bb82a42a tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Task: {'id': task-5116978, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1221.482194] env[65726]: DEBUG nova.compute.manager [None req-8f1940db-b066-401a-a91a-8c578c79d9a0 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 378c1d18-4bb6-4245-80bf-370fb7af0575] Start destroying the instance on the hypervisor. {{(pid=65726) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1221.482522] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-8f1940db-b066-401a-a91a-8c578c79d9a0 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 378c1d18-4bb6-4245-80bf-370fb7af0575] Destroying instance {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1221.483475] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3c4ce16-50ed-47e1-b979-895348226822 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1221.492900] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-8f1940db-b066-401a-a91a-8c578c79d9a0 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 378c1d18-4bb6-4245-80bf-370fb7af0575] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1221.493172] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a92cd2ec-996d-4b53-9b0d-aa1e28541982 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1221.496399] env[65726]: DEBUG nova.compute.manager [None req-df9c79d3-14ef-4842-a2ad-e0f00988e3ba tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 86c23794-5d82-4a7a-aec6-de91601177dd] Start destroying the instance on the hypervisor. {{(pid=65726) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1221.496536] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-df9c79d3-14ef-4842-a2ad-e0f00988e3ba tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 86c23794-5d82-4a7a-aec6-de91601177dd] Destroying instance {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1221.497299] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8deb055e-7060-4bef-9084-5c609b2f4f52 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1221.501542] env[65726]: DEBUG oslo_vmware.api [None req-8f1940db-b066-401a-a91a-8c578c79d9a0 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Waiting for the task: (returnval){ [ 1221.501542] env[65726]: value = "task-5116979" [ 1221.501542] env[65726]: _type = "Task" [ 1221.501542] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1221.509779] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-df9c79d3-14ef-4842-a2ad-e0f00988e3ba tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 86c23794-5d82-4a7a-aec6-de91601177dd] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1221.510506] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b829e96f-f7fd-453b-b843-4bb985d0c330 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1221.515424] env[65726]: DEBUG oslo_vmware.api [None req-8f1940db-b066-401a-a91a-8c578c79d9a0 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5116979, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1221.517099] env[65726]: DEBUG oslo_vmware.api [None req-df9c79d3-14ef-4842-a2ad-e0f00988e3ba tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Waiting for the task: (returnval){ [ 1221.517099] env[65726]: value = "task-5116980" [ 1221.517099] env[65726]: _type = "Task" [ 1221.517099] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1221.528879] env[65726]: DEBUG oslo_vmware.api [None req-df9c79d3-14ef-4842-a2ad-e0f00988e3ba tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Task: {'id': task-5116980, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1221.597783] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe3b47a3-1984-413a-9aed-901a5dbf8882 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1221.606941] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0def27c-ab40-4adf-8a0f-91c0a7112083 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1221.649880] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4990b540-3ec6-4613-990a-8aa11de56293 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1221.660196] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06b3f855-a8d6-4d79-bbb8-944a4ca28df6 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1221.676882] env[65726]: DEBUG nova.compute.provider_tree [None req-c804e386-b05c-462f-b1f2-e4f32318c412 tempest-ServerPasswordTestJSON-84500404 tempest-ServerPasswordTestJSON-84500404-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1221.925548] env[65726]: DEBUG oslo_vmware.api [None req-339d6311-4d4e-4252-84dd-03d4bb82a42a tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Task: {'id': task-5116978, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1222.011980] env[65726]: DEBUG oslo_vmware.api [None req-8f1940db-b066-401a-a91a-8c578c79d9a0 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5116979, 'name': PowerOffVM_Task, 'duration_secs': 0.26733} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1222.012322] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-8f1940db-b066-401a-a91a-8c578c79d9a0 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 378c1d18-4bb6-4245-80bf-370fb7af0575] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1222.012491] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-8f1940db-b066-401a-a91a-8c578c79d9a0 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 378c1d18-4bb6-4245-80bf-370fb7af0575] Unregistering the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1222.012748] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b343551e-105e-405e-be78-89afa404058b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1222.026460] env[65726]: DEBUG oslo_vmware.api [None req-df9c79d3-14ef-4842-a2ad-e0f00988e3ba tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Task: {'id': task-5116980, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1222.103186] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-8f1940db-b066-401a-a91a-8c578c79d9a0 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 378c1d18-4bb6-4245-80bf-370fb7af0575] Unregistered the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1222.103572] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-8f1940db-b066-401a-a91a-8c578c79d9a0 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 378c1d18-4bb6-4245-80bf-370fb7af0575] Deleting contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1222.103880] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-8f1940db-b066-401a-a91a-8c578c79d9a0 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Deleting the datastore file [datastore1] 378c1d18-4bb6-4245-80bf-370fb7af0575 {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1222.104346] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-456746f7-6ba3-4b85-bf86-d6e4d874ba65 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1222.111998] env[65726]: DEBUG oslo_vmware.api [None req-8f1940db-b066-401a-a91a-8c578c79d9a0 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Waiting for the task: (returnval){ [ 1222.111998] env[65726]: value = "task-5116982" [ 1222.111998] env[65726]: _type = "Task" [ 1222.111998] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1222.122369] env[65726]: DEBUG oslo_vmware.api [None req-8f1940db-b066-401a-a91a-8c578c79d9a0 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5116982, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1222.181023] env[65726]: DEBUG nova.scheduler.client.report [None req-c804e386-b05c-462f-b1f2-e4f32318c412 tempest-ServerPasswordTestJSON-84500404 tempest-ServerPasswordTestJSON-84500404-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1222.425813] env[65726]: DEBUG oslo_vmware.api [None req-339d6311-4d4e-4252-84dd-03d4bb82a42a tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Task: {'id': task-5116978, 'name': ReconfigVM_Task, 'duration_secs': 1.243541} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1222.426179] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-339d6311-4d4e-4252-84dd-03d4bb82a42a tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] [instance: 5bfbaf25-5aed-46bc-97fc-f138d08ebbde] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-995333', 'volume_id': 'a465c46c-664d-4aca-85ae-db25aaeeee00', 'name': 'volume-a465c46c-664d-4aca-85ae-db25aaeeee00', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '5bfbaf25-5aed-46bc-97fc-f138d08ebbde', 'attached_at': '', 'detached_at': '', 'volume_id': 'a465c46c-664d-4aca-85ae-db25aaeeee00', 'serial': 'a465c46c-664d-4aca-85ae-db25aaeeee00'} {{(pid=65726) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1222.528409] env[65726]: DEBUG oslo_vmware.api [None req-df9c79d3-14ef-4842-a2ad-e0f00988e3ba tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Task: {'id': task-5116980, 'name': PowerOffVM_Task, 'duration_secs': 0.795088} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1222.528700] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-df9c79d3-14ef-4842-a2ad-e0f00988e3ba tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 86c23794-5d82-4a7a-aec6-de91601177dd] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1222.528871] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-df9c79d3-14ef-4842-a2ad-e0f00988e3ba tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 86c23794-5d82-4a7a-aec6-de91601177dd] Unregistering the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1222.529172] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8af30298-0e93-4108-9698-214704dbd83a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1222.595954] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-df9c79d3-14ef-4842-a2ad-e0f00988e3ba tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 86c23794-5d82-4a7a-aec6-de91601177dd] Unregistered the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1222.596192] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-df9c79d3-14ef-4842-a2ad-e0f00988e3ba tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 86c23794-5d82-4a7a-aec6-de91601177dd] Deleting contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1222.596378] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-df9c79d3-14ef-4842-a2ad-e0f00988e3ba tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Deleting the datastore file [datastore1] 86c23794-5d82-4a7a-aec6-de91601177dd {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1222.596602] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e25411db-0c60-4767-a993-5541d35bd273 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1222.609903] env[65726]: DEBUG oslo_vmware.api [None req-df9c79d3-14ef-4842-a2ad-e0f00988e3ba tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Waiting for the task: (returnval){ [ 1222.609903] env[65726]: value = "task-5116984" [ 1222.609903] env[65726]: _type = "Task" [ 1222.609903] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1222.621759] env[65726]: DEBUG oslo_vmware.api [None req-df9c79d3-14ef-4842-a2ad-e0f00988e3ba tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Task: {'id': task-5116984, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1222.624980] env[65726]: DEBUG oslo_vmware.api [None req-8f1940db-b066-401a-a91a-8c578c79d9a0 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Task: {'id': task-5116982, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.491764} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1222.625250] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-8f1940db-b066-401a-a91a-8c578c79d9a0 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Deleted the datastore file {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1222.625440] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-8f1940db-b066-401a-a91a-8c578c79d9a0 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 378c1d18-4bb6-4245-80bf-370fb7af0575] Deleted contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1222.625622] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-8f1940db-b066-401a-a91a-8c578c79d9a0 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 378c1d18-4bb6-4245-80bf-370fb7af0575] Instance destroyed {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1222.625780] env[65726]: INFO nova.compute.manager [None req-8f1940db-b066-401a-a91a-8c578c79d9a0 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] [instance: 378c1d18-4bb6-4245-80bf-370fb7af0575] Took 1.14 seconds to destroy the instance on the hypervisor. [ 1222.626039] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-8f1940db-b066-401a-a91a-8c578c79d9a0 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1222.626243] env[65726]: DEBUG nova.compute.manager [-] [instance: 378c1d18-4bb6-4245-80bf-370fb7af0575] Deallocating network for instance {{(pid=65726) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1222.626403] env[65726]: DEBUG nova.network.neutron [-] [instance: 378c1d18-4bb6-4245-80bf-370fb7af0575] deallocate_for_instance() {{(pid=65726) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 1222.626872] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1222.627136] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1222.664191] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1222.664736] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1222.687201] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c804e386-b05c-462f-b1f2-e4f32318c412 tempest-ServerPasswordTestJSON-84500404 tempest-ServerPasswordTestJSON-84500404-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.228s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1222.687201] env[65726]: DEBUG nova.compute.manager [None req-c804e386-b05c-462f-b1f2-e4f32318c412 tempest-ServerPasswordTestJSON-84500404 tempest-ServerPasswordTestJSON-84500404-project-member] [instance: a1dc9762-ad05-4cd4-b3c3-eb644ba4212b] Start building networks asynchronously for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1222.969307] env[65726]: DEBUG nova.objects.instance [None req-339d6311-4d4e-4252-84dd-03d4bb82a42a tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Lazy-loading 'flavor' on Instance uuid 5bfbaf25-5aed-46bc-97fc-f138d08ebbde {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1223.118984] env[65726]: DEBUG nova.compute.manager [req-0124c581-e5cb-417c-a84d-5ec9c381d805 req-6fd732a8-b068-473f-823c-3aedcf3e9669 service nova] [instance: 378c1d18-4bb6-4245-80bf-370fb7af0575] Received event network-vif-deleted-4765f6c2-d161-4e36-a6a8-3c8d39c47a40 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1223.119239] env[65726]: INFO nova.compute.manager [req-0124c581-e5cb-417c-a84d-5ec9c381d805 req-6fd732a8-b068-473f-823c-3aedcf3e9669 service nova] [instance: 378c1d18-4bb6-4245-80bf-370fb7af0575] Neutron deleted interface 4765f6c2-d161-4e36-a6a8-3c8d39c47a40; detaching it from the instance and deleting it from the info cache [ 1223.119410] env[65726]: DEBUG nova.network.neutron [req-0124c581-e5cb-417c-a84d-5ec9c381d805 req-6fd732a8-b068-473f-823c-3aedcf3e9669 service nova] [instance: 378c1d18-4bb6-4245-80bf-370fb7af0575] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1223.127557] env[65726]: DEBUG oslo_vmware.api [None req-df9c79d3-14ef-4842-a2ad-e0f00988e3ba tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Task: {'id': task-5116984, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1223.192427] env[65726]: DEBUG nova.compute.utils [None req-c804e386-b05c-462f-b1f2-e4f32318c412 tempest-ServerPasswordTestJSON-84500404 tempest-ServerPasswordTestJSON-84500404-project-member] Using /dev/sd instead of None {{(pid=65726) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1223.193807] env[65726]: DEBUG nova.compute.manager [None req-c804e386-b05c-462f-b1f2-e4f32318c412 tempest-ServerPasswordTestJSON-84500404 tempest-ServerPasswordTestJSON-84500404-project-member] [instance: a1dc9762-ad05-4cd4-b3c3-eb644ba4212b] Allocating IP information in the background. {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1223.193994] env[65726]: DEBUG nova.network.neutron [None req-c804e386-b05c-462f-b1f2-e4f32318c412 tempest-ServerPasswordTestJSON-84500404 tempest-ServerPasswordTestJSON-84500404-project-member] [instance: a1dc9762-ad05-4cd4-b3c3-eb644ba4212b] allocate_for_instance() {{(pid=65726) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 1223.194360] env[65726]: WARNING neutronclient.v2_0.client [None req-c804e386-b05c-462f-b1f2-e4f32318c412 tempest-ServerPasswordTestJSON-84500404 tempest-ServerPasswordTestJSON-84500404-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1223.194700] env[65726]: WARNING neutronclient.v2_0.client [None req-c804e386-b05c-462f-b1f2-e4f32318c412 tempest-ServerPasswordTestJSON-84500404 tempest-ServerPasswordTestJSON-84500404-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1223.195324] env[65726]: WARNING openstack [None req-c804e386-b05c-462f-b1f2-e4f32318c412 tempest-ServerPasswordTestJSON-84500404 tempest-ServerPasswordTestJSON-84500404-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1223.195687] env[65726]: WARNING openstack [None req-c804e386-b05c-462f-b1f2-e4f32318c412 tempest-ServerPasswordTestJSON-84500404 tempest-ServerPasswordTestJSON-84500404-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1223.242637] env[65726]: DEBUG nova.policy [None req-c804e386-b05c-462f-b1f2-e4f32318c412 tempest-ServerPasswordTestJSON-84500404 tempest-ServerPasswordTestJSON-84500404-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '50a6c77a6e584246aeca798c15cf48b9', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '355f2b6dbbd74ea6b1d802a67f4b1a7c', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65726) authorize /opt/stack/nova/nova/policy.py:192}} [ 1223.483291] env[65726]: DEBUG nova.network.neutron [None req-c804e386-b05c-462f-b1f2-e4f32318c412 tempest-ServerPasswordTestJSON-84500404 tempest-ServerPasswordTestJSON-84500404-project-member] [instance: a1dc9762-ad05-4cd4-b3c3-eb644ba4212b] Successfully created port: f2c7e927-2595-4d37-9f20-f76f432cc8f8 {{(pid=65726) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1223.567136] env[65726]: DEBUG nova.network.neutron [-] [instance: 378c1d18-4bb6-4245-80bf-370fb7af0575] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1223.623644] env[65726]: DEBUG oslo_vmware.api [None req-df9c79d3-14ef-4842-a2ad-e0f00988e3ba tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Task: {'id': task-5116984, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.797911} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1223.623644] env[65726]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8f3a3c5b-8c40-4dc4-8dd4-8381bed8be51 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1223.624906] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-df9c79d3-14ef-4842-a2ad-e0f00988e3ba tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Deleted the datastore file {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1223.625147] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-df9c79d3-14ef-4842-a2ad-e0f00988e3ba tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 86c23794-5d82-4a7a-aec6-de91601177dd] Deleted contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1223.625328] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-df9c79d3-14ef-4842-a2ad-e0f00988e3ba tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 86c23794-5d82-4a7a-aec6-de91601177dd] Instance destroyed {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1223.625550] env[65726]: INFO nova.compute.manager [None req-df9c79d3-14ef-4842-a2ad-e0f00988e3ba tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 86c23794-5d82-4a7a-aec6-de91601177dd] Took 2.13 seconds to destroy the instance on the hypervisor. [ 1223.626209] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-df9c79d3-14ef-4842-a2ad-e0f00988e3ba tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1223.626209] env[65726]: DEBUG nova.compute.manager [-] [instance: 86c23794-5d82-4a7a-aec6-de91601177dd] Deallocating network for instance {{(pid=65726) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1223.626209] env[65726]: DEBUG nova.network.neutron [-] [instance: 86c23794-5d82-4a7a-aec6-de91601177dd] deallocate_for_instance() {{(pid=65726) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 1223.626639] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1223.627373] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1223.642548] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11503e57-b363-42c2-b3db-3f4f8c61de44 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1223.666032] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1223.666032] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1223.682469] env[65726]: DEBUG nova.compute.manager [req-0124c581-e5cb-417c-a84d-5ec9c381d805 req-6fd732a8-b068-473f-823c-3aedcf3e9669 service nova] [instance: 378c1d18-4bb6-4245-80bf-370fb7af0575] Detach interface failed, port_id=4765f6c2-d161-4e36-a6a8-3c8d39c47a40, reason: Instance 378c1d18-4bb6-4245-80bf-370fb7af0575 could not be found. {{(pid=65726) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11669}} [ 1223.703553] env[65726]: DEBUG nova.compute.manager [None req-c804e386-b05c-462f-b1f2-e4f32318c412 tempest-ServerPasswordTestJSON-84500404 tempest-ServerPasswordTestJSON-84500404-project-member] [instance: a1dc9762-ad05-4cd4-b3c3-eb644ba4212b] Start building block device mappings for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1223.919132] env[65726]: DEBUG nova.compute.manager [req-4bf6ea00-1f6d-4d74-bcd4-c376435b6f8d req-7f5faf4c-9294-4ecc-81a8-e095c93c2b1c service nova] [instance: 86c23794-5d82-4a7a-aec6-de91601177dd] Received event network-vif-deleted-dd9340e1-39b8-47ca-91f2-eec36706a058 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1223.919132] env[65726]: INFO nova.compute.manager [req-4bf6ea00-1f6d-4d74-bcd4-c376435b6f8d req-7f5faf4c-9294-4ecc-81a8-e095c93c2b1c service nova] [instance: 86c23794-5d82-4a7a-aec6-de91601177dd] Neutron deleted interface dd9340e1-39b8-47ca-91f2-eec36706a058; detaching it from the instance and deleting it from the info cache [ 1223.919132] env[65726]: DEBUG nova.network.neutron [req-4bf6ea00-1f6d-4d74-bcd4-c376435b6f8d req-7f5faf4c-9294-4ecc-81a8-e095c93c2b1c service nova] [instance: 86c23794-5d82-4a7a-aec6-de91601177dd] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1223.977789] env[65726]: DEBUG oslo_concurrency.lockutils [None req-339d6311-4d4e-4252-84dd-03d4bb82a42a tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Lock "5bfbaf25-5aed-46bc-97fc-f138d08ebbde" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 4.238s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1224.070487] env[65726]: INFO nova.compute.manager [-] [instance: 378c1d18-4bb6-4245-80bf-370fb7af0575] Took 1.44 seconds to deallocate network for instance. [ 1224.196250] env[65726]: DEBUG oslo_concurrency.lockutils [None req-eccf3cf7-0fb8-40ae-87b1-15fb50e6db1a tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Acquiring lock "5bfbaf25-5aed-46bc-97fc-f138d08ebbde" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1224.196377] env[65726]: DEBUG oslo_concurrency.lockutils [None req-eccf3cf7-0fb8-40ae-87b1-15fb50e6db1a tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Lock "5bfbaf25-5aed-46bc-97fc-f138d08ebbde" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1224.196558] env[65726]: DEBUG oslo_concurrency.lockutils [None req-eccf3cf7-0fb8-40ae-87b1-15fb50e6db1a tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Acquiring lock "5bfbaf25-5aed-46bc-97fc-f138d08ebbde-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1224.196725] env[65726]: DEBUG oslo_concurrency.lockutils [None req-eccf3cf7-0fb8-40ae-87b1-15fb50e6db1a tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Lock "5bfbaf25-5aed-46bc-97fc-f138d08ebbde-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1224.196894] env[65726]: DEBUG oslo_concurrency.lockutils [None req-eccf3cf7-0fb8-40ae-87b1-15fb50e6db1a tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Lock "5bfbaf25-5aed-46bc-97fc-f138d08ebbde-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1224.199689] env[65726]: INFO nova.compute.manager [None req-eccf3cf7-0fb8-40ae-87b1-15fb50e6db1a tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] [instance: 5bfbaf25-5aed-46bc-97fc-f138d08ebbde] Terminating instance [ 1224.391309] env[65726]: DEBUG nova.network.neutron [-] [instance: 86c23794-5d82-4a7a-aec6-de91601177dd] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1224.421947] env[65726]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-bf1c3e71-d359-49d0-bb8f-b0e6f468e929 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1224.433357] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-749481e4-b007-4330-9bd1-40a140531d7e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1224.465822] env[65726]: DEBUG nova.compute.manager [req-4bf6ea00-1f6d-4d74-bcd4-c376435b6f8d req-7f5faf4c-9294-4ecc-81a8-e095c93c2b1c service nova] [instance: 86c23794-5d82-4a7a-aec6-de91601177dd] Detach interface failed, port_id=dd9340e1-39b8-47ca-91f2-eec36706a058, reason: Instance 86c23794-5d82-4a7a-aec6-de91601177dd could not be found. {{(pid=65726) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11669}} [ 1224.577603] env[65726]: DEBUG oslo_concurrency.lockutils [None req-8f1940db-b066-401a-a91a-8c578c79d9a0 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1224.577892] env[65726]: DEBUG oslo_concurrency.lockutils [None req-8f1940db-b066-401a-a91a-8c578c79d9a0 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1224.578128] env[65726]: DEBUG nova.objects.instance [None req-8f1940db-b066-401a-a91a-8c578c79d9a0 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Lazy-loading 'resources' on Instance uuid 378c1d18-4bb6-4245-80bf-370fb7af0575 {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1224.703310] env[65726]: DEBUG nova.compute.manager [None req-eccf3cf7-0fb8-40ae-87b1-15fb50e6db1a tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] [instance: 5bfbaf25-5aed-46bc-97fc-f138d08ebbde] Start destroying the instance on the hypervisor. {{(pid=65726) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1224.703525] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-eccf3cf7-0fb8-40ae-87b1-15fb50e6db1a tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] [instance: 5bfbaf25-5aed-46bc-97fc-f138d08ebbde] Destroying instance {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1224.704537] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2365c752-52cd-4ec4-b337-3930b0265f4e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1224.713639] env[65726]: DEBUG nova.compute.manager [None req-c804e386-b05c-462f-b1f2-e4f32318c412 tempest-ServerPasswordTestJSON-84500404 tempest-ServerPasswordTestJSON-84500404-project-member] [instance: a1dc9762-ad05-4cd4-b3c3-eb644ba4212b] Start spawning the instance on the hypervisor. {{(pid=65726) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1224.715740] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-eccf3cf7-0fb8-40ae-87b1-15fb50e6db1a tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] [instance: 5bfbaf25-5aed-46bc-97fc-f138d08ebbde] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1224.716239] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d2c25611-1f98-401b-9966-5d4e9d72d110 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1224.723783] env[65726]: DEBUG oslo_vmware.api [None req-eccf3cf7-0fb8-40ae-87b1-15fb50e6db1a tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Waiting for the task: (returnval){ [ 1224.723783] env[65726]: value = "task-5116985" [ 1224.723783] env[65726]: _type = "Task" [ 1224.723783] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1224.732705] env[65726]: DEBUG oslo_vmware.api [None req-eccf3cf7-0fb8-40ae-87b1-15fb50e6db1a tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Task: {'id': task-5116985, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1224.744937] env[65726]: DEBUG nova.virt.hardware [None req-c804e386-b05c-462f-b1f2-e4f32318c412 tempest-ServerPasswordTestJSON-84500404 tempest-ServerPasswordTestJSON-84500404-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T19:28:34Z,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1539f06fa3534e90acbc3a60c4b8bd3f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T19:28:35Z,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1224.745248] env[65726]: DEBUG nova.virt.hardware [None req-c804e386-b05c-462f-b1f2-e4f32318c412 tempest-ServerPasswordTestJSON-84500404 tempest-ServerPasswordTestJSON-84500404-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1224.745410] env[65726]: DEBUG nova.virt.hardware [None req-c804e386-b05c-462f-b1f2-e4f32318c412 tempest-ServerPasswordTestJSON-84500404 tempest-ServerPasswordTestJSON-84500404-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1224.745591] env[65726]: DEBUG nova.virt.hardware [None req-c804e386-b05c-462f-b1f2-e4f32318c412 tempest-ServerPasswordTestJSON-84500404 tempest-ServerPasswordTestJSON-84500404-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1224.745733] env[65726]: DEBUG nova.virt.hardware [None req-c804e386-b05c-462f-b1f2-e4f32318c412 tempest-ServerPasswordTestJSON-84500404 tempest-ServerPasswordTestJSON-84500404-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1224.745876] env[65726]: DEBUG nova.virt.hardware [None req-c804e386-b05c-462f-b1f2-e4f32318c412 tempest-ServerPasswordTestJSON-84500404 tempest-ServerPasswordTestJSON-84500404-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1224.746103] env[65726]: DEBUG nova.virt.hardware [None req-c804e386-b05c-462f-b1f2-e4f32318c412 tempest-ServerPasswordTestJSON-84500404 tempest-ServerPasswordTestJSON-84500404-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1224.746267] env[65726]: DEBUG nova.virt.hardware [None req-c804e386-b05c-462f-b1f2-e4f32318c412 tempest-ServerPasswordTestJSON-84500404 tempest-ServerPasswordTestJSON-84500404-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1224.746535] env[65726]: DEBUG nova.virt.hardware [None req-c804e386-b05c-462f-b1f2-e4f32318c412 tempest-ServerPasswordTestJSON-84500404 tempest-ServerPasswordTestJSON-84500404-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1224.746739] env[65726]: DEBUG nova.virt.hardware [None req-c804e386-b05c-462f-b1f2-e4f32318c412 tempest-ServerPasswordTestJSON-84500404 tempest-ServerPasswordTestJSON-84500404-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1224.746924] env[65726]: DEBUG nova.virt.hardware [None req-c804e386-b05c-462f-b1f2-e4f32318c412 tempest-ServerPasswordTestJSON-84500404 tempest-ServerPasswordTestJSON-84500404-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1224.747856] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25a10315-ce13-4072-b34e-0dccc9a8f5ab {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1224.756214] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82a7199f-fead-4776-925c-cf8120041c7a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1224.894726] env[65726]: INFO nova.compute.manager [-] [instance: 86c23794-5d82-4a7a-aec6-de91601177dd] Took 1.27 seconds to deallocate network for instance. [ 1225.051821] env[65726]: DEBUG nova.network.neutron [None req-c804e386-b05c-462f-b1f2-e4f32318c412 tempest-ServerPasswordTestJSON-84500404 tempest-ServerPasswordTestJSON-84500404-project-member] [instance: a1dc9762-ad05-4cd4-b3c3-eb644ba4212b] Successfully updated port: f2c7e927-2595-4d37-9f20-f76f432cc8f8 {{(pid=65726) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 1225.146980] env[65726]: DEBUG nova.compute.manager [req-a82a795b-4328-432a-bb74-623c88c9bce2 req-ba77c686-020f-46fc-97cf-fa093677678f service nova] [instance: a1dc9762-ad05-4cd4-b3c3-eb644ba4212b] Received event network-vif-plugged-f2c7e927-2595-4d37-9f20-f76f432cc8f8 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1225.147166] env[65726]: DEBUG oslo_concurrency.lockutils [req-a82a795b-4328-432a-bb74-623c88c9bce2 req-ba77c686-020f-46fc-97cf-fa093677678f service nova] Acquiring lock "a1dc9762-ad05-4cd4-b3c3-eb644ba4212b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1225.147374] env[65726]: DEBUG oslo_concurrency.lockutils [req-a82a795b-4328-432a-bb74-623c88c9bce2 req-ba77c686-020f-46fc-97cf-fa093677678f service nova] Lock "a1dc9762-ad05-4cd4-b3c3-eb644ba4212b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1225.147536] env[65726]: DEBUG oslo_concurrency.lockutils [req-a82a795b-4328-432a-bb74-623c88c9bce2 req-ba77c686-020f-46fc-97cf-fa093677678f service nova] Lock "a1dc9762-ad05-4cd4-b3c3-eb644ba4212b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1225.147698] env[65726]: DEBUG nova.compute.manager [req-a82a795b-4328-432a-bb74-623c88c9bce2 req-ba77c686-020f-46fc-97cf-fa093677678f service nova] [instance: a1dc9762-ad05-4cd4-b3c3-eb644ba4212b] No waiting events found dispatching network-vif-plugged-f2c7e927-2595-4d37-9f20-f76f432cc8f8 {{(pid=65726) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1225.147859] env[65726]: WARNING nova.compute.manager [req-a82a795b-4328-432a-bb74-623c88c9bce2 req-ba77c686-020f-46fc-97cf-fa093677678f service nova] [instance: a1dc9762-ad05-4cd4-b3c3-eb644ba4212b] Received unexpected event network-vif-plugged-f2c7e927-2595-4d37-9f20-f76f432cc8f8 for instance with vm_state building and task_state spawning. [ 1225.148170] env[65726]: DEBUG nova.compute.manager [req-a82a795b-4328-432a-bb74-623c88c9bce2 req-ba77c686-020f-46fc-97cf-fa093677678f service nova] [instance: a1dc9762-ad05-4cd4-b3c3-eb644ba4212b] Received event network-changed-f2c7e927-2595-4d37-9f20-f76f432cc8f8 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1225.148448] env[65726]: DEBUG nova.compute.manager [req-a82a795b-4328-432a-bb74-623c88c9bce2 req-ba77c686-020f-46fc-97cf-fa093677678f service nova] [instance: a1dc9762-ad05-4cd4-b3c3-eb644ba4212b] Refreshing instance network info cache due to event network-changed-f2c7e927-2595-4d37-9f20-f76f432cc8f8. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 1225.148702] env[65726]: DEBUG oslo_concurrency.lockutils [req-a82a795b-4328-432a-bb74-623c88c9bce2 req-ba77c686-020f-46fc-97cf-fa093677678f service nova] Acquiring lock "refresh_cache-a1dc9762-ad05-4cd4-b3c3-eb644ba4212b" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1225.148876] env[65726]: DEBUG oslo_concurrency.lockutils [req-a82a795b-4328-432a-bb74-623c88c9bce2 req-ba77c686-020f-46fc-97cf-fa093677678f service nova] Acquired lock "refresh_cache-a1dc9762-ad05-4cd4-b3c3-eb644ba4212b" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1225.148979] env[65726]: DEBUG nova.network.neutron [req-a82a795b-4328-432a-bb74-623c88c9bce2 req-ba77c686-020f-46fc-97cf-fa093677678f service nova] [instance: a1dc9762-ad05-4cd4-b3c3-eb644ba4212b] Refreshing network info cache for port f2c7e927-2595-4d37-9f20-f76f432cc8f8 {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 1225.200651] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f613e82-145d-421a-8ce7-17c0f336ced3 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1225.209027] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-494d8e63-07ce-4310-8536-314cb6c7e60a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1225.243611] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc761cef-2af4-4c97-b23a-4c1cfa709656 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1225.254503] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2a3b0d7-f2f1-4d95-b67c-27c83197ddd8 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1225.258479] env[65726]: DEBUG oslo_vmware.api [None req-eccf3cf7-0fb8-40ae-87b1-15fb50e6db1a tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Task: {'id': task-5116985, 'name': PowerOffVM_Task, 'duration_secs': 0.20391} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1225.258782] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-eccf3cf7-0fb8-40ae-87b1-15fb50e6db1a tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] [instance: 5bfbaf25-5aed-46bc-97fc-f138d08ebbde] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1225.259059] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-eccf3cf7-0fb8-40ae-87b1-15fb50e6db1a tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] [instance: 5bfbaf25-5aed-46bc-97fc-f138d08ebbde] Unregistering the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1225.259605] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-76174226-8935-43d1-9738-7b3affab1666 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1225.269987] env[65726]: DEBUG nova.compute.provider_tree [None req-8f1940db-b066-401a-a91a-8c578c79d9a0 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1225.332159] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-eccf3cf7-0fb8-40ae-87b1-15fb50e6db1a tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] [instance: 5bfbaf25-5aed-46bc-97fc-f138d08ebbde] Unregistered the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1225.332159] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-eccf3cf7-0fb8-40ae-87b1-15fb50e6db1a tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] [instance: 5bfbaf25-5aed-46bc-97fc-f138d08ebbde] Deleting contents of the VM from datastore datastore2 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1225.332159] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-eccf3cf7-0fb8-40ae-87b1-15fb50e6db1a tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Deleting the datastore file [datastore2] 5bfbaf25-5aed-46bc-97fc-f138d08ebbde {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1225.332651] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b5db5b14-7c3a-4ef7-b16d-e2ab2aaec725 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1225.346824] env[65726]: DEBUG oslo_vmware.api [None req-eccf3cf7-0fb8-40ae-87b1-15fb50e6db1a tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Waiting for the task: (returnval){ [ 1225.346824] env[65726]: value = "task-5116987" [ 1225.346824] env[65726]: _type = "Task" [ 1225.346824] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1225.355634] env[65726]: DEBUG oslo_vmware.api [None req-eccf3cf7-0fb8-40ae-87b1-15fb50e6db1a tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Task: {'id': task-5116987, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1225.402879] env[65726]: DEBUG oslo_concurrency.lockutils [None req-df9c79d3-14ef-4842-a2ad-e0f00988e3ba tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1225.555796] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c804e386-b05c-462f-b1f2-e4f32318c412 tempest-ServerPasswordTestJSON-84500404 tempest-ServerPasswordTestJSON-84500404-project-member] Acquiring lock "refresh_cache-a1dc9762-ad05-4cd4-b3c3-eb644ba4212b" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1225.652138] env[65726]: WARNING openstack [req-a82a795b-4328-432a-bb74-623c88c9bce2 req-ba77c686-020f-46fc-97cf-fa093677678f service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1225.652566] env[65726]: WARNING openstack [req-a82a795b-4328-432a-bb74-623c88c9bce2 req-ba77c686-020f-46fc-97cf-fa093677678f service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1225.657701] env[65726]: DEBUG nova.network.neutron [req-a82a795b-4328-432a-bb74-623c88c9bce2 req-ba77c686-020f-46fc-97cf-fa093677678f service nova] [instance: a1dc9762-ad05-4cd4-b3c3-eb644ba4212b] Instance cache missing network info. {{(pid=65726) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 1225.773977] env[65726]: DEBUG nova.scheduler.client.report [None req-8f1940db-b066-401a-a91a-8c578c79d9a0 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1225.781468] env[65726]: DEBUG nova.network.neutron [req-a82a795b-4328-432a-bb74-623c88c9bce2 req-ba77c686-020f-46fc-97cf-fa093677678f service nova] [instance: a1dc9762-ad05-4cd4-b3c3-eb644ba4212b] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1225.857307] env[65726]: DEBUG oslo_vmware.api [None req-eccf3cf7-0fb8-40ae-87b1-15fb50e6db1a tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Task: {'id': task-5116987, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.23923} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1225.857567] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-eccf3cf7-0fb8-40ae-87b1-15fb50e6db1a tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Deleted the datastore file {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1225.857750] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-eccf3cf7-0fb8-40ae-87b1-15fb50e6db1a tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] [instance: 5bfbaf25-5aed-46bc-97fc-f138d08ebbde] Deleted contents of the VM from datastore datastore2 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1225.857994] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-eccf3cf7-0fb8-40ae-87b1-15fb50e6db1a tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] [instance: 5bfbaf25-5aed-46bc-97fc-f138d08ebbde] Instance destroyed {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1225.858126] env[65726]: INFO nova.compute.manager [None req-eccf3cf7-0fb8-40ae-87b1-15fb50e6db1a tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] [instance: 5bfbaf25-5aed-46bc-97fc-f138d08ebbde] Took 1.15 seconds to destroy the instance on the hypervisor. [ 1225.858384] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-eccf3cf7-0fb8-40ae-87b1-15fb50e6db1a tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1225.858586] env[65726]: DEBUG nova.compute.manager [-] [instance: 5bfbaf25-5aed-46bc-97fc-f138d08ebbde] Deallocating network for instance {{(pid=65726) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1225.858694] env[65726]: DEBUG nova.network.neutron [-] [instance: 5bfbaf25-5aed-46bc-97fc-f138d08ebbde] deallocate_for_instance() {{(pid=65726) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 1225.859298] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1225.859556] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1225.909732] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1225.910040] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1226.284958] env[65726]: DEBUG oslo_concurrency.lockutils [req-a82a795b-4328-432a-bb74-623c88c9bce2 req-ba77c686-020f-46fc-97cf-fa093677678f service nova] Releasing lock "refresh_cache-a1dc9762-ad05-4cd4-b3c3-eb644ba4212b" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1226.285801] env[65726]: DEBUG oslo_concurrency.lockutils [None req-8f1940db-b066-401a-a91a-8c578c79d9a0 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.708s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1226.288774] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c804e386-b05c-462f-b1f2-e4f32318c412 tempest-ServerPasswordTestJSON-84500404 tempest-ServerPasswordTestJSON-84500404-project-member] Acquired lock "refresh_cache-a1dc9762-ad05-4cd4-b3c3-eb644ba4212b" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1226.288828] env[65726]: DEBUG nova.network.neutron [None req-c804e386-b05c-462f-b1f2-e4f32318c412 tempest-ServerPasswordTestJSON-84500404 tempest-ServerPasswordTestJSON-84500404-project-member] [instance: a1dc9762-ad05-4cd4-b3c3-eb644ba4212b] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1226.289855] env[65726]: DEBUG oslo_concurrency.lockutils [None req-df9c79d3-14ef-4842-a2ad-e0f00988e3ba tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.887s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1226.290081] env[65726]: DEBUG nova.objects.instance [None req-df9c79d3-14ef-4842-a2ad-e0f00988e3ba tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Lazy-loading 'resources' on Instance uuid 86c23794-5d82-4a7a-aec6-de91601177dd {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1226.314872] env[65726]: INFO nova.scheduler.client.report [None req-8f1940db-b066-401a-a91a-8c578c79d9a0 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Deleted allocations for instance 378c1d18-4bb6-4245-80bf-370fb7af0575 [ 1226.776692] env[65726]: DEBUG nova.network.neutron [-] [instance: 5bfbaf25-5aed-46bc-97fc-f138d08ebbde] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1226.792652] env[65726]: WARNING openstack [None req-c804e386-b05c-462f-b1f2-e4f32318c412 tempest-ServerPasswordTestJSON-84500404 tempest-ServerPasswordTestJSON-84500404-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1226.793037] env[65726]: WARNING openstack [None req-c804e386-b05c-462f-b1f2-e4f32318c412 tempest-ServerPasswordTestJSON-84500404 tempest-ServerPasswordTestJSON-84500404-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1226.798273] env[65726]: DEBUG nova.network.neutron [None req-c804e386-b05c-462f-b1f2-e4f32318c412 tempest-ServerPasswordTestJSON-84500404 tempest-ServerPasswordTestJSON-84500404-project-member] [instance: a1dc9762-ad05-4cd4-b3c3-eb644ba4212b] Instance cache missing network info. {{(pid=65726) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 1226.823490] env[65726]: DEBUG oslo_concurrency.lockutils [None req-8f1940db-b066-401a-a91a-8c578c79d9a0 tempest-ServerActionsTestOtherB-491183096 tempest-ServerActionsTestOtherB-491183096-project-member] Lock "378c1d18-4bb6-4245-80bf-370fb7af0575" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.848s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1226.857756] env[65726]: WARNING openstack [None req-c804e386-b05c-462f-b1f2-e4f32318c412 tempest-ServerPasswordTestJSON-84500404 tempest-ServerPasswordTestJSON-84500404-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1226.858165] env[65726]: WARNING openstack [None req-c804e386-b05c-462f-b1f2-e4f32318c412 tempest-ServerPasswordTestJSON-84500404 tempest-ServerPasswordTestJSON-84500404-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1226.905701] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51945df8-0a88-40e8-aa22-b8c9b001ee9f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1226.915208] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26b30a0c-0512-4073-a575-9674031073ab {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1226.950211] env[65726]: WARNING openstack [None req-c804e386-b05c-462f-b1f2-e4f32318c412 tempest-ServerPasswordTestJSON-84500404 tempest-ServerPasswordTestJSON-84500404-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1226.950576] env[65726]: WARNING openstack [None req-c804e386-b05c-462f-b1f2-e4f32318c412 tempest-ServerPasswordTestJSON-84500404 tempest-ServerPasswordTestJSON-84500404-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1226.959008] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21415a82-89f0-4e47-baef-2d094b3a9501 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1226.967211] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9942234b-c4d8-431e-9d75-68ef3aacc738 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1226.981411] env[65726]: DEBUG nova.compute.provider_tree [None req-df9c79d3-14ef-4842-a2ad-e0f00988e3ba tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1227.037148] env[65726]: DEBUG nova.network.neutron [None req-c804e386-b05c-462f-b1f2-e4f32318c412 tempest-ServerPasswordTestJSON-84500404 tempest-ServerPasswordTestJSON-84500404-project-member] [instance: a1dc9762-ad05-4cd4-b3c3-eb644ba4212b] Updating instance_info_cache with network_info: [{"id": "f2c7e927-2595-4d37-9f20-f76f432cc8f8", "address": "fa:16:3e:94:86:63", "network": {"id": "e4333ba9-f0e4-42f3-803a-2be0960e2772", "bridge": "br-int", "label": "tempest-ServerPasswordTestJSON-408270272-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "355f2b6dbbd74ea6b1d802a67f4b1a7c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a8bbebaf-0fb5-42ae-8d4f-ecd4f46d0244", "external-id": "nsx-vlan-transportzone-296", "segmentation_id": 296, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf2c7e927-25", "ovs_interfaceid": "f2c7e927-2595-4d37-9f20-f76f432cc8f8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1227.180647] env[65726]: DEBUG nova.compute.manager [req-1c5393fd-530a-4cff-a9fa-bc46e62dd8f4 req-5adf1d9f-8310-4359-889c-b1478dfc31c1 service nova] [instance: 5bfbaf25-5aed-46bc-97fc-f138d08ebbde] Received event network-vif-deleted-c41be8b4-e252-47c3-8529-cd6a7ce6a8d3 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1227.279354] env[65726]: INFO nova.compute.manager [-] [instance: 5bfbaf25-5aed-46bc-97fc-f138d08ebbde] Took 1.42 seconds to deallocate network for instance. [ 1227.484497] env[65726]: DEBUG nova.scheduler.client.report [None req-df9c79d3-14ef-4842-a2ad-e0f00988e3ba tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1227.540259] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c804e386-b05c-462f-b1f2-e4f32318c412 tempest-ServerPasswordTestJSON-84500404 tempest-ServerPasswordTestJSON-84500404-project-member] Releasing lock "refresh_cache-a1dc9762-ad05-4cd4-b3c3-eb644ba4212b" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1227.540761] env[65726]: DEBUG nova.compute.manager [None req-c804e386-b05c-462f-b1f2-e4f32318c412 tempest-ServerPasswordTestJSON-84500404 tempest-ServerPasswordTestJSON-84500404-project-member] [instance: a1dc9762-ad05-4cd4-b3c3-eb644ba4212b] Instance network_info: |[{"id": "f2c7e927-2595-4d37-9f20-f76f432cc8f8", "address": "fa:16:3e:94:86:63", "network": {"id": "e4333ba9-f0e4-42f3-803a-2be0960e2772", "bridge": "br-int", "label": "tempest-ServerPasswordTestJSON-408270272-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "355f2b6dbbd74ea6b1d802a67f4b1a7c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a8bbebaf-0fb5-42ae-8d4f-ecd4f46d0244", "external-id": "nsx-vlan-transportzone-296", "segmentation_id": 296, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf2c7e927-25", "ovs_interfaceid": "f2c7e927-2595-4d37-9f20-f76f432cc8f8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1227.541225] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-c804e386-b05c-462f-b1f2-e4f32318c412 tempest-ServerPasswordTestJSON-84500404 tempest-ServerPasswordTestJSON-84500404-project-member] [instance: a1dc9762-ad05-4cd4-b3c3-eb644ba4212b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:94:86:63', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a8bbebaf-0fb5-42ae-8d4f-ecd4f46d0244', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f2c7e927-2595-4d37-9f20-f76f432cc8f8', 'vif_model': 'vmxnet3'}] {{(pid=65726) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1227.549013] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-c804e386-b05c-462f-b1f2-e4f32318c412 tempest-ServerPasswordTestJSON-84500404 tempest-ServerPasswordTestJSON-84500404-project-member] Creating folder: Project (355f2b6dbbd74ea6b1d802a67f4b1a7c). Parent ref: group-v995008. {{(pid=65726) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1227.549323] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9a288af3-2b7f-43c9-b0de-ce737803107b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1227.560710] env[65726]: INFO nova.virt.vmwareapi.vm_util [None req-c804e386-b05c-462f-b1f2-e4f32318c412 tempest-ServerPasswordTestJSON-84500404 tempest-ServerPasswordTestJSON-84500404-project-member] Created folder: Project (355f2b6dbbd74ea6b1d802a67f4b1a7c) in parent group-v995008. [ 1227.560957] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-c804e386-b05c-462f-b1f2-e4f32318c412 tempest-ServerPasswordTestJSON-84500404 tempest-ServerPasswordTestJSON-84500404-project-member] Creating folder: Instances. Parent ref: group-v995335. {{(pid=65726) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1227.561152] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-cc2a1822-593b-4a1e-80dd-50bf92624d17 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1227.572250] env[65726]: INFO nova.virt.vmwareapi.vm_util [None req-c804e386-b05c-462f-b1f2-e4f32318c412 tempest-ServerPasswordTestJSON-84500404 tempest-ServerPasswordTestJSON-84500404-project-member] Created folder: Instances in parent group-v995335. [ 1227.572501] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-c804e386-b05c-462f-b1f2-e4f32318c412 tempest-ServerPasswordTestJSON-84500404 tempest-ServerPasswordTestJSON-84500404-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1227.572717] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a1dc9762-ad05-4cd4-b3c3-eb644ba4212b] Creating VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1227.572934] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2f6576fa-c2d9-4872-9cef-aea994118fe0 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1227.593085] env[65726]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1227.593085] env[65726]: value = "task-5116991" [ 1227.593085] env[65726]: _type = "Task" [ 1227.593085] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1227.601457] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116991, 'name': CreateVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1227.786824] env[65726]: DEBUG oslo_concurrency.lockutils [None req-eccf3cf7-0fb8-40ae-87b1-15fb50e6db1a tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1227.990719] env[65726]: DEBUG oslo_concurrency.lockutils [None req-df9c79d3-14ef-4842-a2ad-e0f00988e3ba tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.700s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1227.993086] env[65726]: DEBUG oslo_concurrency.lockutils [None req-eccf3cf7-0fb8-40ae-87b1-15fb50e6db1a tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.206s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1227.993340] env[65726]: DEBUG nova.objects.instance [None req-eccf3cf7-0fb8-40ae-87b1-15fb50e6db1a tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Lazy-loading 'resources' on Instance uuid 5bfbaf25-5aed-46bc-97fc-f138d08ebbde {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1228.013477] env[65726]: INFO nova.scheduler.client.report [None req-df9c79d3-14ef-4842-a2ad-e0f00988e3ba tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Deleted allocations for instance 86c23794-5d82-4a7a-aec6-de91601177dd [ 1228.104079] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5116991, 'name': CreateVM_Task, 'duration_secs': 0.374418} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1228.104361] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a1dc9762-ad05-4cd4-b3c3-eb644ba4212b] Created VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1228.105205] env[65726]: WARNING openstack [None req-c804e386-b05c-462f-b1f2-e4f32318c412 tempest-ServerPasswordTestJSON-84500404 tempest-ServerPasswordTestJSON-84500404-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1228.105579] env[65726]: WARNING openstack [None req-c804e386-b05c-462f-b1f2-e4f32318c412 tempest-ServerPasswordTestJSON-84500404 tempest-ServerPasswordTestJSON-84500404-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1228.110768] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c804e386-b05c-462f-b1f2-e4f32318c412 tempest-ServerPasswordTestJSON-84500404 tempest-ServerPasswordTestJSON-84500404-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1228.110933] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c804e386-b05c-462f-b1f2-e4f32318c412 tempest-ServerPasswordTestJSON-84500404 tempest-ServerPasswordTestJSON-84500404-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1228.111271] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c804e386-b05c-462f-b1f2-e4f32318c412 tempest-ServerPasswordTestJSON-84500404 tempest-ServerPasswordTestJSON-84500404-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1228.111554] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e5336bdb-a9c2-43f6-be58-942948ec8955 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1228.117122] env[65726]: DEBUG oslo_vmware.api [None req-c804e386-b05c-462f-b1f2-e4f32318c412 tempest-ServerPasswordTestJSON-84500404 tempest-ServerPasswordTestJSON-84500404-project-member] Waiting for the task: (returnval){ [ 1228.117122] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52198cce-806f-63ba-7679-8dbccdc366b7" [ 1228.117122] env[65726]: _type = "Task" [ 1228.117122] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1228.126469] env[65726]: DEBUG oslo_vmware.api [None req-c804e386-b05c-462f-b1f2-e4f32318c412 tempest-ServerPasswordTestJSON-84500404 tempest-ServerPasswordTestJSON-84500404-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52198cce-806f-63ba-7679-8dbccdc366b7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1228.523634] env[65726]: DEBUG oslo_concurrency.lockutils [None req-df9c79d3-14ef-4842-a2ad-e0f00988e3ba tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Lock "86c23794-5d82-4a7a-aec6-de91601177dd" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.534s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1228.601301] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08b5c50b-7a70-442c-9a72-b66d683ac8b5 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1228.610915] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d336681-058b-4fe1-a09b-160aee0020ee {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1228.647030] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30a52920-e76b-468b-8685-56618e81bed0 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1228.655640] env[65726]: DEBUG oslo_vmware.api [None req-c804e386-b05c-462f-b1f2-e4f32318c412 tempest-ServerPasswordTestJSON-84500404 tempest-ServerPasswordTestJSON-84500404-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52198cce-806f-63ba-7679-8dbccdc366b7, 'name': SearchDatastore_Task, 'duration_secs': 0.014473} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1228.657997] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c804e386-b05c-462f-b1f2-e4f32318c412 tempest-ServerPasswordTestJSON-84500404 tempest-ServerPasswordTestJSON-84500404-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1228.658325] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-c804e386-b05c-462f-b1f2-e4f32318c412 tempest-ServerPasswordTestJSON-84500404 tempest-ServerPasswordTestJSON-84500404-project-member] [instance: a1dc9762-ad05-4cd4-b3c3-eb644ba4212b] Processing image b52362a3-ee8a-4cbf-b06f-513b0cc8f95c {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1228.658535] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c804e386-b05c-462f-b1f2-e4f32318c412 tempest-ServerPasswordTestJSON-84500404 tempest-ServerPasswordTestJSON-84500404-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1228.658644] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c804e386-b05c-462f-b1f2-e4f32318c412 tempest-ServerPasswordTestJSON-84500404 tempest-ServerPasswordTestJSON-84500404-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1228.658809] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-c804e386-b05c-462f-b1f2-e4f32318c412 tempest-ServerPasswordTestJSON-84500404 tempest-ServerPasswordTestJSON-84500404-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1228.659195] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-af9b263e-a20d-4bf7-9032-5a80b1571105 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1228.662515] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-410c323a-6944-4f27-89ff-ee0eaa65662e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1228.678200] env[65726]: DEBUG nova.compute.provider_tree [None req-eccf3cf7-0fb8-40ae-87b1-15fb50e6db1a tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1228.680626] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-c804e386-b05c-462f-b1f2-e4f32318c412 tempest-ServerPasswordTestJSON-84500404 tempest-ServerPasswordTestJSON-84500404-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1228.680841] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-c804e386-b05c-462f-b1f2-e4f32318c412 tempest-ServerPasswordTestJSON-84500404 tempest-ServerPasswordTestJSON-84500404-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65726) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1228.685022] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-67a793ff-9bd7-40e5-929d-9e258793ef12 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1228.687791] env[65726]: DEBUG oslo_vmware.api [None req-c804e386-b05c-462f-b1f2-e4f32318c412 tempest-ServerPasswordTestJSON-84500404 tempest-ServerPasswordTestJSON-84500404-project-member] Waiting for the task: (returnval){ [ 1228.687791] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]529ba458-7686-baf6-0e69-6659118d4adb" [ 1228.687791] env[65726]: _type = "Task" [ 1228.687791] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1228.697399] env[65726]: DEBUG oslo_vmware.api [None req-c804e386-b05c-462f-b1f2-e4f32318c412 tempest-ServerPasswordTestJSON-84500404 tempest-ServerPasswordTestJSON-84500404-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]529ba458-7686-baf6-0e69-6659118d4adb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1228.732538] env[65726]: DEBUG oslo_concurrency.lockutils [None req-0d6072ec-c5b5-49a1-a119-57d8363f77ef tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Acquiring lock "e3255f3b-028f-4a0b-b621-873730417936" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1228.732788] env[65726]: DEBUG oslo_concurrency.lockutils [None req-0d6072ec-c5b5-49a1-a119-57d8363f77ef tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Lock "e3255f3b-028f-4a0b-b621-873730417936" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1228.732968] env[65726]: DEBUG nova.compute.manager [None req-0d6072ec-c5b5-49a1-a119-57d8363f77ef tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: e3255f3b-028f-4a0b-b621-873730417936] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1228.734088] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f911f855-3fe2-4b8e-a793-096aeb2ea7f2 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1228.740945] env[65726]: DEBUG nova.compute.manager [None req-0d6072ec-c5b5-49a1-a119-57d8363f77ef tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: e3255f3b-028f-4a0b-b621-873730417936] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=65726) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3459}} [ 1228.741560] env[65726]: DEBUG nova.objects.instance [None req-0d6072ec-c5b5-49a1-a119-57d8363f77ef tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Lazy-loading 'flavor' on Instance uuid e3255f3b-028f-4a0b-b621-873730417936 {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1229.186734] env[65726]: DEBUG nova.scheduler.client.report [None req-eccf3cf7-0fb8-40ae-87b1-15fb50e6db1a tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1229.201204] env[65726]: DEBUG oslo_vmware.api [None req-c804e386-b05c-462f-b1f2-e4f32318c412 tempest-ServerPasswordTestJSON-84500404 tempest-ServerPasswordTestJSON-84500404-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]529ba458-7686-baf6-0e69-6659118d4adb, 'name': SearchDatastore_Task, 'duration_secs': 0.01068} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1229.202674] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f94a2d71-3f96-4446-b6e8-d5e43f330cb2 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1229.208468] env[65726]: DEBUG oslo_vmware.api [None req-c804e386-b05c-462f-b1f2-e4f32318c412 tempest-ServerPasswordTestJSON-84500404 tempest-ServerPasswordTestJSON-84500404-project-member] Waiting for the task: (returnval){ [ 1229.208468] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52d2c1e0-8431-1624-7815-6e602445ef26" [ 1229.208468] env[65726]: _type = "Task" [ 1229.208468] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1229.216410] env[65726]: DEBUG oslo_vmware.api [None req-c804e386-b05c-462f-b1f2-e4f32318c412 tempest-ServerPasswordTestJSON-84500404 tempest-ServerPasswordTestJSON-84500404-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52d2c1e0-8431-1624-7815-6e602445ef26, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1229.563976] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e1322ace-4de0-4198-b5e5-47926fd137bd tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Acquiring lock "a4c229bb-e5f9-41d3-86f9-ca4cfb087558" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1229.564350] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e1322ace-4de0-4198-b5e5-47926fd137bd tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Lock "a4c229bb-e5f9-41d3-86f9-ca4cfb087558" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1229.696527] env[65726]: DEBUG oslo_concurrency.lockutils [None req-eccf3cf7-0fb8-40ae-87b1-15fb50e6db1a tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.703s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1229.719490] env[65726]: DEBUG oslo_vmware.api [None req-c804e386-b05c-462f-b1f2-e4f32318c412 tempest-ServerPasswordTestJSON-84500404 tempest-ServerPasswordTestJSON-84500404-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52d2c1e0-8431-1624-7815-6e602445ef26, 'name': SearchDatastore_Task, 'duration_secs': 0.013333} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1229.719719] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c804e386-b05c-462f-b1f2-e4f32318c412 tempest-ServerPasswordTestJSON-84500404 tempest-ServerPasswordTestJSON-84500404-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1229.719968] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-c804e386-b05c-462f-b1f2-e4f32318c412 tempest-ServerPasswordTestJSON-84500404 tempest-ServerPasswordTestJSON-84500404-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] a1dc9762-ad05-4cd4-b3c3-eb644ba4212b/a1dc9762-ad05-4cd4-b3c3-eb644ba4212b.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1229.721032] env[65726]: INFO nova.scheduler.client.report [None req-eccf3cf7-0fb8-40ae-87b1-15fb50e6db1a tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Deleted allocations for instance 5bfbaf25-5aed-46bc-97fc-f138d08ebbde [ 1229.721987] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-98ba555b-f095-49b8-95ef-23ec973ba2d0 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1229.731086] env[65726]: DEBUG oslo_vmware.api [None req-c804e386-b05c-462f-b1f2-e4f32318c412 tempest-ServerPasswordTestJSON-84500404 tempest-ServerPasswordTestJSON-84500404-project-member] Waiting for the task: (returnval){ [ 1229.731086] env[65726]: value = "task-5116992" [ 1229.731086] env[65726]: _type = "Task" [ 1229.731086] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1229.739942] env[65726]: DEBUG oslo_vmware.api [None req-c804e386-b05c-462f-b1f2-e4f32318c412 tempest-ServerPasswordTestJSON-84500404 tempest-ServerPasswordTestJSON-84500404-project-member] Task: {'id': task-5116992, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1229.749664] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-0d6072ec-c5b5-49a1-a119-57d8363f77ef tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: e3255f3b-028f-4a0b-b621-873730417936] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1229.750687] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8a074c48-f25a-49ab-96de-3ffa821d310f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1229.756884] env[65726]: DEBUG oslo_vmware.api [None req-0d6072ec-c5b5-49a1-a119-57d8363f77ef tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Waiting for the task: (returnval){ [ 1229.756884] env[65726]: value = "task-5116993" [ 1229.756884] env[65726]: _type = "Task" [ 1229.756884] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1229.766146] env[65726]: DEBUG oslo_vmware.api [None req-0d6072ec-c5b5-49a1-a119-57d8363f77ef tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': task-5116993, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1230.067231] env[65726]: DEBUG nova.compute.manager [None req-e1322ace-4de0-4198-b5e5-47926fd137bd tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: a4c229bb-e5f9-41d3-86f9-ca4cfb087558] Starting instance... {{(pid=65726) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 1230.230535] env[65726]: DEBUG oslo_concurrency.lockutils [None req-eccf3cf7-0fb8-40ae-87b1-15fb50e6db1a tempest-AttachVolumeTestJSON-1781541322 tempest-AttachVolumeTestJSON-1781541322-project-member] Lock "5bfbaf25-5aed-46bc-97fc-f138d08ebbde" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.034s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1230.242857] env[65726]: DEBUG oslo_vmware.api [None req-c804e386-b05c-462f-b1f2-e4f32318c412 tempest-ServerPasswordTestJSON-84500404 tempest-ServerPasswordTestJSON-84500404-project-member] Task: {'id': task-5116992, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1230.268028] env[65726]: DEBUG oslo_vmware.api [None req-0d6072ec-c5b5-49a1-a119-57d8363f77ef tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': task-5116993, 'name': PowerOffVM_Task, 'duration_secs': 0.213032} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1230.268362] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-0d6072ec-c5b5-49a1-a119-57d8363f77ef tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: e3255f3b-028f-4a0b-b621-873730417936] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1230.268537] env[65726]: DEBUG nova.compute.manager [None req-0d6072ec-c5b5-49a1-a119-57d8363f77ef tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: e3255f3b-028f-4a0b-b621-873730417936] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1230.269400] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e21b6fcf-badd-4eca-84a1-5d0e5922c505 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1230.590012] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e1322ace-4de0-4198-b5e5-47926fd137bd tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1230.590359] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e1322ace-4de0-4198-b5e5-47926fd137bd tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1230.594061] env[65726]: INFO nova.compute.claims [None req-e1322ace-4de0-4198-b5e5-47926fd137bd tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: a4c229bb-e5f9-41d3-86f9-ca4cfb087558] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1230.746000] env[65726]: DEBUG oslo_vmware.api [None req-c804e386-b05c-462f-b1f2-e4f32318c412 tempest-ServerPasswordTestJSON-84500404 tempest-ServerPasswordTestJSON-84500404-project-member] Task: {'id': task-5116992, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.595801} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1230.746334] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-c804e386-b05c-462f-b1f2-e4f32318c412 tempest-ServerPasswordTestJSON-84500404 tempest-ServerPasswordTestJSON-84500404-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] a1dc9762-ad05-4cd4-b3c3-eb644ba4212b/a1dc9762-ad05-4cd4-b3c3-eb644ba4212b.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1230.746567] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-c804e386-b05c-462f-b1f2-e4f32318c412 tempest-ServerPasswordTestJSON-84500404 tempest-ServerPasswordTestJSON-84500404-project-member] [instance: a1dc9762-ad05-4cd4-b3c3-eb644ba4212b] Extending root virtual disk to 1048576 {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1230.746817] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-7295e0e8-a422-4b16-8216-1abc3da7b5e5 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1230.755551] env[65726]: DEBUG oslo_vmware.api [None req-c804e386-b05c-462f-b1f2-e4f32318c412 tempest-ServerPasswordTestJSON-84500404 tempest-ServerPasswordTestJSON-84500404-project-member] Waiting for the task: (returnval){ [ 1230.755551] env[65726]: value = "task-5116994" [ 1230.755551] env[65726]: _type = "Task" [ 1230.755551] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1230.770240] env[65726]: DEBUG oslo_vmware.api [None req-c804e386-b05c-462f-b1f2-e4f32318c412 tempest-ServerPasswordTestJSON-84500404 tempest-ServerPasswordTestJSON-84500404-project-member] Task: {'id': task-5116994, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1230.782815] env[65726]: DEBUG oslo_concurrency.lockutils [None req-0d6072ec-c5b5-49a1-a119-57d8363f77ef tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Lock "e3255f3b-028f-4a0b-b621-873730417936" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.050s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1231.206028] env[65726]: DEBUG nova.objects.instance [None req-145aabe1-082c-4201-a407-8b4a130e72f3 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Lazy-loading 'flavor' on Instance uuid e3255f3b-028f-4a0b-b621-873730417936 {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1231.265995] env[65726]: DEBUG oslo_vmware.api [None req-c804e386-b05c-462f-b1f2-e4f32318c412 tempest-ServerPasswordTestJSON-84500404 tempest-ServerPasswordTestJSON-84500404-project-member] Task: {'id': task-5116994, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.07706} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1231.266361] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-c804e386-b05c-462f-b1f2-e4f32318c412 tempest-ServerPasswordTestJSON-84500404 tempest-ServerPasswordTestJSON-84500404-project-member] [instance: a1dc9762-ad05-4cd4-b3c3-eb644ba4212b] Extended root virtual disk {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1231.267189] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-965798b8-6f99-41c9-8340-dfddb3d49d34 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1231.290449] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-c804e386-b05c-462f-b1f2-e4f32318c412 tempest-ServerPasswordTestJSON-84500404 tempest-ServerPasswordTestJSON-84500404-project-member] [instance: a1dc9762-ad05-4cd4-b3c3-eb644ba4212b] Reconfiguring VM instance instance-00000073 to attach disk [datastore1] a1dc9762-ad05-4cd4-b3c3-eb644ba4212b/a1dc9762-ad05-4cd4-b3c3-eb644ba4212b.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1231.290996] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6ea4d98d-ee3a-4526-87b4-6770290403cb {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1231.311956] env[65726]: DEBUG oslo_vmware.api [None req-c804e386-b05c-462f-b1f2-e4f32318c412 tempest-ServerPasswordTestJSON-84500404 tempest-ServerPasswordTestJSON-84500404-project-member] Waiting for the task: (returnval){ [ 1231.311956] env[65726]: value = "task-5116996" [ 1231.311956] env[65726]: _type = "Task" [ 1231.311956] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1231.320702] env[65726]: DEBUG oslo_vmware.api [None req-c804e386-b05c-462f-b1f2-e4f32318c412 tempest-ServerPasswordTestJSON-84500404 tempest-ServerPasswordTestJSON-84500404-project-member] Task: {'id': task-5116996, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1231.710768] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a46b6a61-e631-41e2-b595-dffcedc65a71 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Acquiring lock "e9e93672-d509-45b3-89fe-f1c6ff4d8e78" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1231.711053] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a46b6a61-e631-41e2-b595-dffcedc65a71 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Lock "e9e93672-d509-45b3-89fe-f1c6ff4d8e78" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1231.715554] env[65726]: DEBUG oslo_concurrency.lockutils [None req-145aabe1-082c-4201-a407-8b4a130e72f3 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Acquiring lock "refresh_cache-e3255f3b-028f-4a0b-b621-873730417936" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1231.715672] env[65726]: DEBUG oslo_concurrency.lockutils [None req-145aabe1-082c-4201-a407-8b4a130e72f3 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Acquired lock "refresh_cache-e3255f3b-028f-4a0b-b621-873730417936" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1231.715782] env[65726]: DEBUG nova.network.neutron [None req-145aabe1-082c-4201-a407-8b4a130e72f3 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: e3255f3b-028f-4a0b-b621-873730417936] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1231.715953] env[65726]: DEBUG nova.objects.instance [None req-145aabe1-082c-4201-a407-8b4a130e72f3 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Lazy-loading 'info_cache' on Instance uuid e3255f3b-028f-4a0b-b621-873730417936 {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1231.729232] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43ffc2c5-b694-45aa-aaee-8980c6895fe3 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1231.738507] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1382c44d-4a44-4031-84fa-fd3413531ef9 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1231.772194] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d34f72b-7f01-4a2c-8272-7c9a9dd2c211 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1231.780922] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc29bf5a-bddb-4407-9434-d29dfc859d3f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1231.795738] env[65726]: DEBUG nova.compute.provider_tree [None req-e1322ace-4de0-4198-b5e5-47926fd137bd tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1231.822655] env[65726]: DEBUG oslo_vmware.api [None req-c804e386-b05c-462f-b1f2-e4f32318c412 tempest-ServerPasswordTestJSON-84500404 tempest-ServerPasswordTestJSON-84500404-project-member] Task: {'id': task-5116996, 'name': ReconfigVM_Task, 'duration_secs': 0.372054} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1231.823064] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-c804e386-b05c-462f-b1f2-e4f32318c412 tempest-ServerPasswordTestJSON-84500404 tempest-ServerPasswordTestJSON-84500404-project-member] [instance: a1dc9762-ad05-4cd4-b3c3-eb644ba4212b] Reconfigured VM instance instance-00000073 to attach disk [datastore1] a1dc9762-ad05-4cd4-b3c3-eb644ba4212b/a1dc9762-ad05-4cd4-b3c3-eb644ba4212b.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1231.823721] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e580aeaf-dd33-47c5-a5cd-30e5ac1548e2 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1231.833303] env[65726]: DEBUG oslo_vmware.api [None req-c804e386-b05c-462f-b1f2-e4f32318c412 tempest-ServerPasswordTestJSON-84500404 tempest-ServerPasswordTestJSON-84500404-project-member] Waiting for the task: (returnval){ [ 1231.833303] env[65726]: value = "task-5116997" [ 1231.833303] env[65726]: _type = "Task" [ 1231.833303] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1231.842753] env[65726]: DEBUG oslo_vmware.api [None req-c804e386-b05c-462f-b1f2-e4f32318c412 tempest-ServerPasswordTestJSON-84500404 tempest-ServerPasswordTestJSON-84500404-project-member] Task: {'id': task-5116997, 'name': Rename_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1232.219899] env[65726]: DEBUG nova.objects.base [None req-145aabe1-082c-4201-a407-8b4a130e72f3 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Object Instance lazy-loaded attributes: flavor,info_cache {{(pid=65726) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1232.221506] env[65726]: DEBUG nova.compute.manager [None req-a46b6a61-e631-41e2-b595-dffcedc65a71 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: e9e93672-d509-45b3-89fe-f1c6ff4d8e78] Starting instance... {{(pid=65726) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 1232.299631] env[65726]: DEBUG nova.scheduler.client.report [None req-e1322ace-4de0-4198-b5e5-47926fd137bd tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1232.345409] env[65726]: DEBUG oslo_vmware.api [None req-c804e386-b05c-462f-b1f2-e4f32318c412 tempest-ServerPasswordTestJSON-84500404 tempest-ServerPasswordTestJSON-84500404-project-member] Task: {'id': task-5116997, 'name': Rename_Task, 'duration_secs': 0.171928} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1232.345721] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-c804e386-b05c-462f-b1f2-e4f32318c412 tempest-ServerPasswordTestJSON-84500404 tempest-ServerPasswordTestJSON-84500404-project-member] [instance: a1dc9762-ad05-4cd4-b3c3-eb644ba4212b] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1232.345987] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6a08ead9-9266-4235-b09a-9adfffcf8567 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1232.353447] env[65726]: DEBUG oslo_vmware.api [None req-c804e386-b05c-462f-b1f2-e4f32318c412 tempest-ServerPasswordTestJSON-84500404 tempest-ServerPasswordTestJSON-84500404-project-member] Waiting for the task: (returnval){ [ 1232.353447] env[65726]: value = "task-5116998" [ 1232.353447] env[65726]: _type = "Task" [ 1232.353447] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1232.364463] env[65726]: DEBUG oslo_vmware.api [None req-c804e386-b05c-462f-b1f2-e4f32318c412 tempest-ServerPasswordTestJSON-84500404 tempest-ServerPasswordTestJSON-84500404-project-member] Task: {'id': task-5116998, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1232.725518] env[65726]: WARNING openstack [None req-145aabe1-082c-4201-a407-8b4a130e72f3 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1232.725792] env[65726]: WARNING openstack [None req-145aabe1-082c-4201-a407-8b4a130e72f3 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1232.751490] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a46b6a61-e631-41e2-b595-dffcedc65a71 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1232.805779] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e1322ace-4de0-4198-b5e5-47926fd137bd tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.215s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1232.806339] env[65726]: DEBUG nova.compute.manager [None req-e1322ace-4de0-4198-b5e5-47926fd137bd tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: a4c229bb-e5f9-41d3-86f9-ca4cfb087558] Start building networks asynchronously for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1232.809111] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a46b6a61-e631-41e2-b595-dffcedc65a71 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.058s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1232.810613] env[65726]: INFO nova.compute.claims [None req-a46b6a61-e631-41e2-b595-dffcedc65a71 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: e9e93672-d509-45b3-89fe-f1c6ff4d8e78] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1232.861936] env[65726]: WARNING openstack [None req-145aabe1-082c-4201-a407-8b4a130e72f3 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1232.862293] env[65726]: WARNING openstack [None req-145aabe1-082c-4201-a407-8b4a130e72f3 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1232.875761] env[65726]: DEBUG oslo_vmware.api [None req-c804e386-b05c-462f-b1f2-e4f32318c412 tempest-ServerPasswordTestJSON-84500404 tempest-ServerPasswordTestJSON-84500404-project-member] Task: {'id': task-5116998, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1232.967988] env[65726]: WARNING openstack [None req-145aabe1-082c-4201-a407-8b4a130e72f3 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1232.968950] env[65726]: WARNING openstack [None req-145aabe1-082c-4201-a407-8b4a130e72f3 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1233.087175] env[65726]: DEBUG nova.network.neutron [None req-145aabe1-082c-4201-a407-8b4a130e72f3 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: e3255f3b-028f-4a0b-b621-873730417936] Updating instance_info_cache with network_info: [{"id": "03c9e6fb-2435-43be-bb55-8afecdcf38ff", "address": "fa:16:3e:d2:0f:96", "network": {"id": "a899453d-5f43-4650-9642-b33f54a4d2eb", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-614861139-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.231", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "acd800e0438940bc80ae3c6b672db4ee", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "205fb402-8eaf-4b61-8f57-8f216024179a", "external-id": "nsx-vlan-transportzone-78", "segmentation_id": 78, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap03c9e6fb-24", "ovs_interfaceid": "03c9e6fb-2435-43be-bb55-8afecdcf38ff", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1233.315198] env[65726]: DEBUG nova.compute.utils [None req-e1322ace-4de0-4198-b5e5-47926fd137bd tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Using /dev/sd instead of None {{(pid=65726) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1233.320944] env[65726]: DEBUG nova.compute.manager [None req-e1322ace-4de0-4198-b5e5-47926fd137bd tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: a4c229bb-e5f9-41d3-86f9-ca4cfb087558] Allocating IP information in the background. {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1233.321242] env[65726]: DEBUG nova.network.neutron [None req-e1322ace-4de0-4198-b5e5-47926fd137bd tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: a4c229bb-e5f9-41d3-86f9-ca4cfb087558] allocate_for_instance() {{(pid=65726) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 1233.322232] env[65726]: WARNING neutronclient.v2_0.client [None req-e1322ace-4de0-4198-b5e5-47926fd137bd tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1233.322232] env[65726]: WARNING neutronclient.v2_0.client [None req-e1322ace-4de0-4198-b5e5-47926fd137bd tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1233.322571] env[65726]: WARNING openstack [None req-e1322ace-4de0-4198-b5e5-47926fd137bd tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1233.323568] env[65726]: WARNING openstack [None req-e1322ace-4de0-4198-b5e5-47926fd137bd tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1233.370102] env[65726]: DEBUG oslo_vmware.api [None req-c804e386-b05c-462f-b1f2-e4f32318c412 tempest-ServerPasswordTestJSON-84500404 tempest-ServerPasswordTestJSON-84500404-project-member] Task: {'id': task-5116998, 'name': PowerOnVM_Task, 'duration_secs': 0.702248} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1233.374202] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-c804e386-b05c-462f-b1f2-e4f32318c412 tempest-ServerPasswordTestJSON-84500404 tempest-ServerPasswordTestJSON-84500404-project-member] [instance: a1dc9762-ad05-4cd4-b3c3-eb644ba4212b] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1233.374501] env[65726]: INFO nova.compute.manager [None req-c804e386-b05c-462f-b1f2-e4f32318c412 tempest-ServerPasswordTestJSON-84500404 tempest-ServerPasswordTestJSON-84500404-project-member] [instance: a1dc9762-ad05-4cd4-b3c3-eb644ba4212b] Took 8.66 seconds to spawn the instance on the hypervisor. [ 1233.375330] env[65726]: DEBUG nova.compute.manager [None req-c804e386-b05c-462f-b1f2-e4f32318c412 tempest-ServerPasswordTestJSON-84500404 tempest-ServerPasswordTestJSON-84500404-project-member] [instance: a1dc9762-ad05-4cd4-b3c3-eb644ba4212b] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1233.376176] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5a1bc83-a132-4ea4-9570-18c9e9ff9543 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1233.393700] env[65726]: DEBUG nova.policy [None req-e1322ace-4de0-4198-b5e5-47926fd137bd tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ccb3513b5b4f453f842139ec1815c4af', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f3f799aa94f64e8cb26d93f6124efd81', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65726) authorize /opt/stack/nova/nova/policy.py:192}} [ 1233.479897] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70f4c38f-2d19-41ea-a557-88f583e6c6b4 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1233.489947] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e397732-3714-4097-beff-5d9f2259d743 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1233.524543] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69666934-45c4-4ef6-986b-f755455de0be {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1233.532948] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16127611-2bbe-4889-ba5d-513bbc6492e4 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1233.547554] env[65726]: DEBUG nova.compute.provider_tree [None req-a46b6a61-e631-41e2-b595-dffcedc65a71 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1233.590062] env[65726]: DEBUG oslo_concurrency.lockutils [None req-145aabe1-082c-4201-a407-8b4a130e72f3 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Releasing lock "refresh_cache-e3255f3b-028f-4a0b-b621-873730417936" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1233.697858] env[65726]: DEBUG nova.network.neutron [None req-e1322ace-4de0-4198-b5e5-47926fd137bd tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: a4c229bb-e5f9-41d3-86f9-ca4cfb087558] Successfully created port: 016b56b6-52e2-48c7-9a15-c91dda1dee97 {{(pid=65726) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1233.832256] env[65726]: DEBUG nova.compute.manager [None req-e1322ace-4de0-4198-b5e5-47926fd137bd tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: a4c229bb-e5f9-41d3-86f9-ca4cfb087558] Start building block device mappings for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1233.907410] env[65726]: INFO nova.compute.manager [None req-c804e386-b05c-462f-b1f2-e4f32318c412 tempest-ServerPasswordTestJSON-84500404 tempest-ServerPasswordTestJSON-84500404-project-member] [instance: a1dc9762-ad05-4cd4-b3c3-eb644ba4212b] Took 13.48 seconds to build instance. [ 1234.051102] env[65726]: DEBUG nova.scheduler.client.report [None req-a46b6a61-e631-41e2-b595-dffcedc65a71 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1234.410049] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c804e386-b05c-462f-b1f2-e4f32318c412 tempest-ServerPasswordTestJSON-84500404 tempest-ServerPasswordTestJSON-84500404-project-member] Lock "a1dc9762-ad05-4cd4-b3c3-eb644ba4212b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.995s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1234.557056] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a46b6a61-e631-41e2-b595-dffcedc65a71 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 1.748s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1234.557593] env[65726]: DEBUG nova.compute.manager [None req-a46b6a61-e631-41e2-b595-dffcedc65a71 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: e9e93672-d509-45b3-89fe-f1c6ff4d8e78] Start building networks asynchronously for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1234.596503] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-145aabe1-082c-4201-a407-8b4a130e72f3 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: e3255f3b-028f-4a0b-b621-873730417936] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1234.596503] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-efa214f8-2736-4403-b8e0-ddb823cd4c40 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1234.607054] env[65726]: DEBUG oslo_vmware.api [None req-145aabe1-082c-4201-a407-8b4a130e72f3 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Waiting for the task: (returnval){ [ 1234.607054] env[65726]: value = "task-5117000" [ 1234.607054] env[65726]: _type = "Task" [ 1234.607054] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1234.616966] env[65726]: DEBUG oslo_vmware.api [None req-145aabe1-082c-4201-a407-8b4a130e72f3 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': task-5117000, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1234.842352] env[65726]: DEBUG nova.compute.manager [None req-e1322ace-4de0-4198-b5e5-47926fd137bd tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: a4c229bb-e5f9-41d3-86f9-ca4cfb087558] Start spawning the instance on the hypervisor. {{(pid=65726) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1234.872914] env[65726]: DEBUG nova.virt.hardware [None req-e1322ace-4de0-4198-b5e5-47926fd137bd tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T19:28:34Z,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1539f06fa3534e90acbc3a60c4b8bd3f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T19:28:35Z,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1234.873187] env[65726]: DEBUG nova.virt.hardware [None req-e1322ace-4de0-4198-b5e5-47926fd137bd tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1234.873364] env[65726]: DEBUG nova.virt.hardware [None req-e1322ace-4de0-4198-b5e5-47926fd137bd tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1234.873588] env[65726]: DEBUG nova.virt.hardware [None req-e1322ace-4de0-4198-b5e5-47926fd137bd tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1234.873685] env[65726]: DEBUG nova.virt.hardware [None req-e1322ace-4de0-4198-b5e5-47926fd137bd tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1234.873911] env[65726]: DEBUG nova.virt.hardware [None req-e1322ace-4de0-4198-b5e5-47926fd137bd tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1234.874095] env[65726]: DEBUG nova.virt.hardware [None req-e1322ace-4de0-4198-b5e5-47926fd137bd tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1234.874291] env[65726]: DEBUG nova.virt.hardware [None req-e1322ace-4de0-4198-b5e5-47926fd137bd tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1234.874474] env[65726]: DEBUG nova.virt.hardware [None req-e1322ace-4de0-4198-b5e5-47926fd137bd tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1234.874644] env[65726]: DEBUG nova.virt.hardware [None req-e1322ace-4de0-4198-b5e5-47926fd137bd tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1234.874815] env[65726]: DEBUG nova.virt.hardware [None req-e1322ace-4de0-4198-b5e5-47926fd137bd tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1234.875829] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fcff4968-4594-4e36-b37c-e3f982cda6af {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1234.888064] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aea49dfb-eac2-43ff-ba9d-4cdce0f58a74 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1235.039452] env[65726]: DEBUG oslo_concurrency.lockutils [None req-01b26dc0-172c-48e9-b562-30576ec9ec52 tempest-ServerPasswordTestJSON-84500404 tempest-ServerPasswordTestJSON-84500404-project-member] Acquiring lock "a1dc9762-ad05-4cd4-b3c3-eb644ba4212b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1235.039744] env[65726]: DEBUG oslo_concurrency.lockutils [None req-01b26dc0-172c-48e9-b562-30576ec9ec52 tempest-ServerPasswordTestJSON-84500404 tempest-ServerPasswordTestJSON-84500404-project-member] Lock "a1dc9762-ad05-4cd4-b3c3-eb644ba4212b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1235.039962] env[65726]: DEBUG oslo_concurrency.lockutils [None req-01b26dc0-172c-48e9-b562-30576ec9ec52 tempest-ServerPasswordTestJSON-84500404 tempest-ServerPasswordTestJSON-84500404-project-member] Acquiring lock "a1dc9762-ad05-4cd4-b3c3-eb644ba4212b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1235.040453] env[65726]: DEBUG oslo_concurrency.lockutils [None req-01b26dc0-172c-48e9-b562-30576ec9ec52 tempest-ServerPasswordTestJSON-84500404 tempest-ServerPasswordTestJSON-84500404-project-member] Lock "a1dc9762-ad05-4cd4-b3c3-eb644ba4212b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1235.040453] env[65726]: DEBUG oslo_concurrency.lockutils [None req-01b26dc0-172c-48e9-b562-30576ec9ec52 tempest-ServerPasswordTestJSON-84500404 tempest-ServerPasswordTestJSON-84500404-project-member] Lock "a1dc9762-ad05-4cd4-b3c3-eb644ba4212b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1235.042540] env[65726]: INFO nova.compute.manager [None req-01b26dc0-172c-48e9-b562-30576ec9ec52 tempest-ServerPasswordTestJSON-84500404 tempest-ServerPasswordTestJSON-84500404-project-member] [instance: a1dc9762-ad05-4cd4-b3c3-eb644ba4212b] Terminating instance [ 1235.062451] env[65726]: DEBUG nova.compute.utils [None req-a46b6a61-e631-41e2-b595-dffcedc65a71 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Using /dev/sd instead of None {{(pid=65726) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1235.063931] env[65726]: DEBUG nova.compute.manager [None req-a46b6a61-e631-41e2-b595-dffcedc65a71 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: e9e93672-d509-45b3-89fe-f1c6ff4d8e78] Allocating IP information in the background. {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1235.064786] env[65726]: DEBUG nova.network.neutron [None req-a46b6a61-e631-41e2-b595-dffcedc65a71 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: e9e93672-d509-45b3-89fe-f1c6ff4d8e78] allocate_for_instance() {{(pid=65726) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 1235.064991] env[65726]: WARNING neutronclient.v2_0.client [None req-a46b6a61-e631-41e2-b595-dffcedc65a71 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1235.065335] env[65726]: WARNING neutronclient.v2_0.client [None req-a46b6a61-e631-41e2-b595-dffcedc65a71 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1235.065904] env[65726]: WARNING openstack [None req-a46b6a61-e631-41e2-b595-dffcedc65a71 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1235.066447] env[65726]: WARNING openstack [None req-a46b6a61-e631-41e2-b595-dffcedc65a71 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1235.119929] env[65726]: DEBUG oslo_vmware.api [None req-145aabe1-082c-4201-a407-8b4a130e72f3 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': task-5117000, 'name': PowerOnVM_Task, 'duration_secs': 0.419098} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1235.120293] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-145aabe1-082c-4201-a407-8b4a130e72f3 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: e3255f3b-028f-4a0b-b621-873730417936] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1235.120564] env[65726]: DEBUG nova.compute.manager [None req-145aabe1-082c-4201-a407-8b4a130e72f3 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: e3255f3b-028f-4a0b-b621-873730417936] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1235.121478] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b9e6070-b7e4-449a-8e9d-08ca36660723 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1235.134115] env[65726]: DEBUG nova.policy [None req-a46b6a61-e631-41e2-b595-dffcedc65a71 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6ab5b309207a46bb9d95998ef0a7a46f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6b56f93ba0e14521921484ac2c785c43', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65726) authorize /opt/stack/nova/nova/policy.py:192}} [ 1235.148915] env[65726]: DEBUG nova.compute.manager [req-6245f2c6-23c0-4ab9-b83a-9ccc73f5117e req-30a3aa7e-bd5e-44c1-80b6-25d21f76002c service nova] [instance: a4c229bb-e5f9-41d3-86f9-ca4cfb087558] Received event network-vif-plugged-016b56b6-52e2-48c7-9a15-c91dda1dee97 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1235.149109] env[65726]: DEBUG oslo_concurrency.lockutils [req-6245f2c6-23c0-4ab9-b83a-9ccc73f5117e req-30a3aa7e-bd5e-44c1-80b6-25d21f76002c service nova] Acquiring lock "a4c229bb-e5f9-41d3-86f9-ca4cfb087558-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1235.149326] env[65726]: DEBUG oslo_concurrency.lockutils [req-6245f2c6-23c0-4ab9-b83a-9ccc73f5117e req-30a3aa7e-bd5e-44c1-80b6-25d21f76002c service nova] Lock "a4c229bb-e5f9-41d3-86f9-ca4cfb087558-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1235.149495] env[65726]: DEBUG oslo_concurrency.lockutils [req-6245f2c6-23c0-4ab9-b83a-9ccc73f5117e req-30a3aa7e-bd5e-44c1-80b6-25d21f76002c service nova] Lock "a4c229bb-e5f9-41d3-86f9-ca4cfb087558-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1235.149747] env[65726]: DEBUG nova.compute.manager [req-6245f2c6-23c0-4ab9-b83a-9ccc73f5117e req-30a3aa7e-bd5e-44c1-80b6-25d21f76002c service nova] [instance: a4c229bb-e5f9-41d3-86f9-ca4cfb087558] No waiting events found dispatching network-vif-plugged-016b56b6-52e2-48c7-9a15-c91dda1dee97 {{(pid=65726) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1235.149929] env[65726]: WARNING nova.compute.manager [req-6245f2c6-23c0-4ab9-b83a-9ccc73f5117e req-30a3aa7e-bd5e-44c1-80b6-25d21f76002c service nova] [instance: a4c229bb-e5f9-41d3-86f9-ca4cfb087558] Received unexpected event network-vif-plugged-016b56b6-52e2-48c7-9a15-c91dda1dee97 for instance with vm_state building and task_state spawning. [ 1235.248300] env[65726]: DEBUG nova.network.neutron [None req-e1322ace-4de0-4198-b5e5-47926fd137bd tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: a4c229bb-e5f9-41d3-86f9-ca4cfb087558] Successfully updated port: 016b56b6-52e2-48c7-9a15-c91dda1dee97 {{(pid=65726) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 1235.364470] env[65726]: DEBUG nova.network.neutron [None req-a46b6a61-e631-41e2-b595-dffcedc65a71 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: e9e93672-d509-45b3-89fe-f1c6ff4d8e78] Successfully created port: bc612ce3-dca5-4716-9f5c-60a6247e4b42 {{(pid=65726) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1235.546807] env[65726]: DEBUG nova.compute.manager [None req-01b26dc0-172c-48e9-b562-30576ec9ec52 tempest-ServerPasswordTestJSON-84500404 tempest-ServerPasswordTestJSON-84500404-project-member] [instance: a1dc9762-ad05-4cd4-b3c3-eb644ba4212b] Start destroying the instance on the hypervisor. {{(pid=65726) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1235.547176] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-01b26dc0-172c-48e9-b562-30576ec9ec52 tempest-ServerPasswordTestJSON-84500404 tempest-ServerPasswordTestJSON-84500404-project-member] [instance: a1dc9762-ad05-4cd4-b3c3-eb644ba4212b] Destroying instance {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1235.548116] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d9d5465-2f63-48b8-8ff1-21fff449efef {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1235.556826] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-01b26dc0-172c-48e9-b562-30576ec9ec52 tempest-ServerPasswordTestJSON-84500404 tempest-ServerPasswordTestJSON-84500404-project-member] [instance: a1dc9762-ad05-4cd4-b3c3-eb644ba4212b] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1235.557670] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-54d87224-1715-4ec3-8b92-f0e115344130 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1235.564464] env[65726]: DEBUG oslo_vmware.api [None req-01b26dc0-172c-48e9-b562-30576ec9ec52 tempest-ServerPasswordTestJSON-84500404 tempest-ServerPasswordTestJSON-84500404-project-member] Waiting for the task: (returnval){ [ 1235.564464] env[65726]: value = "task-5117001" [ 1235.564464] env[65726]: _type = "Task" [ 1235.564464] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1235.574944] env[65726]: DEBUG nova.compute.manager [None req-a46b6a61-e631-41e2-b595-dffcedc65a71 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: e9e93672-d509-45b3-89fe-f1c6ff4d8e78] Start building block device mappings for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1235.577786] env[65726]: DEBUG oslo_vmware.api [None req-01b26dc0-172c-48e9-b562-30576ec9ec52 tempest-ServerPasswordTestJSON-84500404 tempest-ServerPasswordTestJSON-84500404-project-member] Task: {'id': task-5117001, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1235.753193] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e1322ace-4de0-4198-b5e5-47926fd137bd tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Acquiring lock "refresh_cache-a4c229bb-e5f9-41d3-86f9-ca4cfb087558" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1235.753505] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e1322ace-4de0-4198-b5e5-47926fd137bd tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Acquired lock "refresh_cache-a4c229bb-e5f9-41d3-86f9-ca4cfb087558" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1235.753765] env[65726]: DEBUG nova.network.neutron [None req-e1322ace-4de0-4198-b5e5-47926fd137bd tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: a4c229bb-e5f9-41d3-86f9-ca4cfb087558] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1236.074822] env[65726]: DEBUG oslo_vmware.api [None req-01b26dc0-172c-48e9-b562-30576ec9ec52 tempest-ServerPasswordTestJSON-84500404 tempest-ServerPasswordTestJSON-84500404-project-member] Task: {'id': task-5117001, 'name': PowerOffVM_Task, 'duration_secs': 0.194855} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1236.075162] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-01b26dc0-172c-48e9-b562-30576ec9ec52 tempest-ServerPasswordTestJSON-84500404 tempest-ServerPasswordTestJSON-84500404-project-member] [instance: a1dc9762-ad05-4cd4-b3c3-eb644ba4212b] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1236.075302] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-01b26dc0-172c-48e9-b562-30576ec9ec52 tempest-ServerPasswordTestJSON-84500404 tempest-ServerPasswordTestJSON-84500404-project-member] [instance: a1dc9762-ad05-4cd4-b3c3-eb644ba4212b] Unregistering the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1236.075560] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-497955bf-0186-47a8-bdc0-71d1d4ee96b1 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1236.145042] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-01b26dc0-172c-48e9-b562-30576ec9ec52 tempest-ServerPasswordTestJSON-84500404 tempest-ServerPasswordTestJSON-84500404-project-member] [instance: a1dc9762-ad05-4cd4-b3c3-eb644ba4212b] Unregistered the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1236.145042] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-01b26dc0-172c-48e9-b562-30576ec9ec52 tempest-ServerPasswordTestJSON-84500404 tempest-ServerPasswordTestJSON-84500404-project-member] [instance: a1dc9762-ad05-4cd4-b3c3-eb644ba4212b] Deleting contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1236.145307] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-01b26dc0-172c-48e9-b562-30576ec9ec52 tempest-ServerPasswordTestJSON-84500404 tempest-ServerPasswordTestJSON-84500404-project-member] Deleting the datastore file [datastore1] a1dc9762-ad05-4cd4-b3c3-eb644ba4212b {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1236.145503] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-70357fb1-0db8-4395-a529-7f8f97e45e08 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1236.154599] env[65726]: DEBUG oslo_vmware.api [None req-01b26dc0-172c-48e9-b562-30576ec9ec52 tempest-ServerPasswordTestJSON-84500404 tempest-ServerPasswordTestJSON-84500404-project-member] Waiting for the task: (returnval){ [ 1236.154599] env[65726]: value = "task-5117003" [ 1236.154599] env[65726]: _type = "Task" [ 1236.154599] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1236.163278] env[65726]: DEBUG oslo_vmware.api [None req-01b26dc0-172c-48e9-b562-30576ec9ec52 tempest-ServerPasswordTestJSON-84500404 tempest-ServerPasswordTestJSON-84500404-project-member] Task: {'id': task-5117003, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1236.257838] env[65726]: WARNING openstack [None req-e1322ace-4de0-4198-b5e5-47926fd137bd tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1236.258389] env[65726]: WARNING openstack [None req-e1322ace-4de0-4198-b5e5-47926fd137bd tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1236.263993] env[65726]: DEBUG nova.network.neutron [None req-e1322ace-4de0-4198-b5e5-47926fd137bd tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: a4c229bb-e5f9-41d3-86f9-ca4cfb087558] Instance cache missing network info. {{(pid=65726) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 1236.320638] env[65726]: WARNING openstack [None req-e1322ace-4de0-4198-b5e5-47926fd137bd tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1236.321198] env[65726]: WARNING openstack [None req-e1322ace-4de0-4198-b5e5-47926fd137bd tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1236.391628] env[65726]: WARNING openstack [None req-e1322ace-4de0-4198-b5e5-47926fd137bd tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1236.392032] env[65726]: WARNING openstack [None req-e1322ace-4de0-4198-b5e5-47926fd137bd tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1236.495040] env[65726]: DEBUG nova.network.neutron [None req-e1322ace-4de0-4198-b5e5-47926fd137bd tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: a4c229bb-e5f9-41d3-86f9-ca4cfb087558] Updating instance_info_cache with network_info: [{"id": "016b56b6-52e2-48c7-9a15-c91dda1dee97", "address": "fa:16:3e:7a:d4:c3", "network": {"id": "69642329-e986-48fa-accf-1f296c08116a", "bridge": "br-int", "label": "tempest-ServersTestJSON-1786538788-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f3f799aa94f64e8cb26d93f6124efd81", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "816c6e38-e200-4544-8c5b-9fc3e16c5761", "external-id": "nsx-vlan-transportzone-195", "segmentation_id": 195, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap016b56b6-52", "ovs_interfaceid": "016b56b6-52e2-48c7-9a15-c91dda1dee97", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1236.584721] env[65726]: DEBUG nova.compute.manager [None req-a46b6a61-e631-41e2-b595-dffcedc65a71 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: e9e93672-d509-45b3-89fe-f1c6ff4d8e78] Start spawning the instance on the hypervisor. {{(pid=65726) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1236.612567] env[65726]: DEBUG nova.virt.hardware [None req-a46b6a61-e631-41e2-b595-dffcedc65a71 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T19:28:34Z,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1539f06fa3534e90acbc3a60c4b8bd3f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T19:28:35Z,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1236.612860] env[65726]: DEBUG nova.virt.hardware [None req-a46b6a61-e631-41e2-b595-dffcedc65a71 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1236.613064] env[65726]: DEBUG nova.virt.hardware [None req-a46b6a61-e631-41e2-b595-dffcedc65a71 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1236.613272] env[65726]: DEBUG nova.virt.hardware [None req-a46b6a61-e631-41e2-b595-dffcedc65a71 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1236.613420] env[65726]: DEBUG nova.virt.hardware [None req-a46b6a61-e631-41e2-b595-dffcedc65a71 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1236.613565] env[65726]: DEBUG nova.virt.hardware [None req-a46b6a61-e631-41e2-b595-dffcedc65a71 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1236.613767] env[65726]: DEBUG nova.virt.hardware [None req-a46b6a61-e631-41e2-b595-dffcedc65a71 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1236.614042] env[65726]: DEBUG nova.virt.hardware [None req-a46b6a61-e631-41e2-b595-dffcedc65a71 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1236.614303] env[65726]: DEBUG nova.virt.hardware [None req-a46b6a61-e631-41e2-b595-dffcedc65a71 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1236.614485] env[65726]: DEBUG nova.virt.hardware [None req-a46b6a61-e631-41e2-b595-dffcedc65a71 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1236.614660] env[65726]: DEBUG nova.virt.hardware [None req-a46b6a61-e631-41e2-b595-dffcedc65a71 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1236.615604] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-495d650e-25aa-4c10-b1ec-57211fa9c1d4 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1236.625328] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03f09875-d4e9-46d1-b518-b41f8a57bf9c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1236.665411] env[65726]: DEBUG oslo_vmware.api [None req-01b26dc0-172c-48e9-b562-30576ec9ec52 tempest-ServerPasswordTestJSON-84500404 tempest-ServerPasswordTestJSON-84500404-project-member] Task: {'id': task-5117003, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.207139} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1236.665728] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-01b26dc0-172c-48e9-b562-30576ec9ec52 tempest-ServerPasswordTestJSON-84500404 tempest-ServerPasswordTestJSON-84500404-project-member] Deleted the datastore file {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1236.665945] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-01b26dc0-172c-48e9-b562-30576ec9ec52 tempest-ServerPasswordTestJSON-84500404 tempest-ServerPasswordTestJSON-84500404-project-member] [instance: a1dc9762-ad05-4cd4-b3c3-eb644ba4212b] Deleted contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1236.666150] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-01b26dc0-172c-48e9-b562-30576ec9ec52 tempest-ServerPasswordTestJSON-84500404 tempest-ServerPasswordTestJSON-84500404-project-member] [instance: a1dc9762-ad05-4cd4-b3c3-eb644ba4212b] Instance destroyed {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1236.666330] env[65726]: INFO nova.compute.manager [None req-01b26dc0-172c-48e9-b562-30576ec9ec52 tempest-ServerPasswordTestJSON-84500404 tempest-ServerPasswordTestJSON-84500404-project-member] [instance: a1dc9762-ad05-4cd4-b3c3-eb644ba4212b] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1236.666592] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-01b26dc0-172c-48e9-b562-30576ec9ec52 tempest-ServerPasswordTestJSON-84500404 tempest-ServerPasswordTestJSON-84500404-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1236.666800] env[65726]: DEBUG nova.compute.manager [-] [instance: a1dc9762-ad05-4cd4-b3c3-eb644ba4212b] Deallocating network for instance {{(pid=65726) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1236.666893] env[65726]: DEBUG nova.network.neutron [-] [instance: a1dc9762-ad05-4cd4-b3c3-eb644ba4212b] deallocate_for_instance() {{(pid=65726) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 1236.667566] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1236.667827] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1236.721167] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1236.721495] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1236.908431] env[65726]: DEBUG nova.network.neutron [None req-a46b6a61-e631-41e2-b595-dffcedc65a71 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: e9e93672-d509-45b3-89fe-f1c6ff4d8e78] Successfully updated port: bc612ce3-dca5-4716-9f5c-60a6247e4b42 {{(pid=65726) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 1236.951745] env[65726]: DEBUG nova.compute.manager [req-fb73edc7-2c45-46d3-aa00-586831974ff3 req-534b9851-a46e-4bab-908f-20dd951f5668 service nova] [instance: a1dc9762-ad05-4cd4-b3c3-eb644ba4212b] Received event network-vif-deleted-f2c7e927-2595-4d37-9f20-f76f432cc8f8 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1236.951950] env[65726]: INFO nova.compute.manager [req-fb73edc7-2c45-46d3-aa00-586831974ff3 req-534b9851-a46e-4bab-908f-20dd951f5668 service nova] [instance: a1dc9762-ad05-4cd4-b3c3-eb644ba4212b] Neutron deleted interface f2c7e927-2595-4d37-9f20-f76f432cc8f8; detaching it from the instance and deleting it from the info cache [ 1236.952447] env[65726]: DEBUG nova.network.neutron [req-fb73edc7-2c45-46d3-aa00-586831974ff3 req-534b9851-a46e-4bab-908f-20dd951f5668 service nova] [instance: a1dc9762-ad05-4cd4-b3c3-eb644ba4212b] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1236.998056] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e1322ace-4de0-4198-b5e5-47926fd137bd tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Releasing lock "refresh_cache-a4c229bb-e5f9-41d3-86f9-ca4cfb087558" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1236.998418] env[65726]: DEBUG nova.compute.manager [None req-e1322ace-4de0-4198-b5e5-47926fd137bd tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: a4c229bb-e5f9-41d3-86f9-ca4cfb087558] Instance network_info: |[{"id": "016b56b6-52e2-48c7-9a15-c91dda1dee97", "address": "fa:16:3e:7a:d4:c3", "network": {"id": "69642329-e986-48fa-accf-1f296c08116a", "bridge": "br-int", "label": "tempest-ServersTestJSON-1786538788-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f3f799aa94f64e8cb26d93f6124efd81", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "816c6e38-e200-4544-8c5b-9fc3e16c5761", "external-id": "nsx-vlan-transportzone-195", "segmentation_id": 195, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap016b56b6-52", "ovs_interfaceid": "016b56b6-52e2-48c7-9a15-c91dda1dee97", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1236.998881] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-e1322ace-4de0-4198-b5e5-47926fd137bd tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: a4c229bb-e5f9-41d3-86f9-ca4cfb087558] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:7a:d4:c3', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '816c6e38-e200-4544-8c5b-9fc3e16c5761', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '016b56b6-52e2-48c7-9a15-c91dda1dee97', 'vif_model': 'vmxnet3'}] {{(pid=65726) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1237.006708] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-e1322ace-4de0-4198-b5e5-47926fd137bd tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1237.006924] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a4c229bb-e5f9-41d3-86f9-ca4cfb087558] Creating VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1237.007143] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8a276c69-3946-4152-aa06-b36e26ef46aa {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1237.029385] env[65726]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1237.029385] env[65726]: value = "task-5117005" [ 1237.029385] env[65726]: _type = "Task" [ 1237.029385] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1237.038110] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5117005, 'name': CreateVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1237.180291] env[65726]: DEBUG nova.compute.manager [req-bf55f577-e3f8-4ff6-ba58-a699309176ed req-5335ae9a-3052-45b1-a098-ce00c0d3b287 service nova] [instance: a4c229bb-e5f9-41d3-86f9-ca4cfb087558] Received event network-changed-016b56b6-52e2-48c7-9a15-c91dda1dee97 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1237.180545] env[65726]: DEBUG nova.compute.manager [req-bf55f577-e3f8-4ff6-ba58-a699309176ed req-5335ae9a-3052-45b1-a098-ce00c0d3b287 service nova] [instance: a4c229bb-e5f9-41d3-86f9-ca4cfb087558] Refreshing instance network info cache due to event network-changed-016b56b6-52e2-48c7-9a15-c91dda1dee97. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 1237.180832] env[65726]: DEBUG oslo_concurrency.lockutils [req-bf55f577-e3f8-4ff6-ba58-a699309176ed req-5335ae9a-3052-45b1-a098-ce00c0d3b287 service nova] Acquiring lock "refresh_cache-a4c229bb-e5f9-41d3-86f9-ca4cfb087558" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1237.181037] env[65726]: DEBUG oslo_concurrency.lockutils [req-bf55f577-e3f8-4ff6-ba58-a699309176ed req-5335ae9a-3052-45b1-a098-ce00c0d3b287 service nova] Acquired lock "refresh_cache-a4c229bb-e5f9-41d3-86f9-ca4cfb087558" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1237.181333] env[65726]: DEBUG nova.network.neutron [req-bf55f577-e3f8-4ff6-ba58-a699309176ed req-5335ae9a-3052-45b1-a098-ce00c0d3b287 service nova] [instance: a4c229bb-e5f9-41d3-86f9-ca4cfb087558] Refreshing network info cache for port 016b56b6-52e2-48c7-9a15-c91dda1dee97 {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 1237.414095] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a46b6a61-e631-41e2-b595-dffcedc65a71 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Acquiring lock "refresh_cache-e9e93672-d509-45b3-89fe-f1c6ff4d8e78" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1237.414095] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a46b6a61-e631-41e2-b595-dffcedc65a71 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Acquired lock "refresh_cache-e9e93672-d509-45b3-89fe-f1c6ff4d8e78" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1237.414095] env[65726]: DEBUG nova.network.neutron [None req-a46b6a61-e631-41e2-b595-dffcedc65a71 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: e9e93672-d509-45b3-89fe-f1c6ff4d8e78] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1237.433658] env[65726]: DEBUG nova.network.neutron [-] [instance: a1dc9762-ad05-4cd4-b3c3-eb644ba4212b] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1237.455845] env[65726]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-233d9dca-e8a6-48a9-9603-f3d8eb9dca6b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1237.466566] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fb36a38-e128-42e8-bfba-fff0649ac349 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1237.499041] env[65726]: DEBUG nova.compute.manager [req-fb73edc7-2c45-46d3-aa00-586831974ff3 req-534b9851-a46e-4bab-908f-20dd951f5668 service nova] [instance: a1dc9762-ad05-4cd4-b3c3-eb644ba4212b] Detach interface failed, port_id=f2c7e927-2595-4d37-9f20-f76f432cc8f8, reason: Instance a1dc9762-ad05-4cd4-b3c3-eb644ba4212b could not be found. {{(pid=65726) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11669}} [ 1237.540546] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5117005, 'name': CreateVM_Task} progress is 99%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1237.684608] env[65726]: WARNING openstack [req-bf55f577-e3f8-4ff6-ba58-a699309176ed req-5335ae9a-3052-45b1-a098-ce00c0d3b287 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1237.684983] env[65726]: WARNING openstack [req-bf55f577-e3f8-4ff6-ba58-a699309176ed req-5335ae9a-3052-45b1-a098-ce00c0d3b287 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1237.797846] env[65726]: WARNING openstack [req-bf55f577-e3f8-4ff6-ba58-a699309176ed req-5335ae9a-3052-45b1-a098-ce00c0d3b287 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1237.798239] env[65726]: WARNING openstack [req-bf55f577-e3f8-4ff6-ba58-a699309176ed req-5335ae9a-3052-45b1-a098-ce00c0d3b287 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1237.863927] env[65726]: WARNING openstack [req-bf55f577-e3f8-4ff6-ba58-a699309176ed req-5335ae9a-3052-45b1-a098-ce00c0d3b287 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1237.864387] env[65726]: WARNING openstack [req-bf55f577-e3f8-4ff6-ba58-a699309176ed req-5335ae9a-3052-45b1-a098-ce00c0d3b287 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1237.916211] env[65726]: WARNING openstack [None req-a46b6a61-e631-41e2-b595-dffcedc65a71 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1237.916951] env[65726]: WARNING openstack [None req-a46b6a61-e631-41e2-b595-dffcedc65a71 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1237.921566] env[65726]: DEBUG nova.network.neutron [None req-a46b6a61-e631-41e2-b595-dffcedc65a71 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: e9e93672-d509-45b3-89fe-f1c6ff4d8e78] Instance cache missing network info. {{(pid=65726) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 1237.936834] env[65726]: INFO nova.compute.manager [-] [instance: a1dc9762-ad05-4cd4-b3c3-eb644ba4212b] Took 1.27 seconds to deallocate network for instance. [ 1237.947642] env[65726]: DEBUG nova.network.neutron [req-bf55f577-e3f8-4ff6-ba58-a699309176ed req-5335ae9a-3052-45b1-a098-ce00c0d3b287 service nova] [instance: a4c229bb-e5f9-41d3-86f9-ca4cfb087558] Updated VIF entry in instance network info cache for port 016b56b6-52e2-48c7-9a15-c91dda1dee97. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 1237.947997] env[65726]: DEBUG nova.network.neutron [req-bf55f577-e3f8-4ff6-ba58-a699309176ed req-5335ae9a-3052-45b1-a098-ce00c0d3b287 service nova] [instance: a4c229bb-e5f9-41d3-86f9-ca4cfb087558] Updating instance_info_cache with network_info: [{"id": "016b56b6-52e2-48c7-9a15-c91dda1dee97", "address": "fa:16:3e:7a:d4:c3", "network": {"id": "69642329-e986-48fa-accf-1f296c08116a", "bridge": "br-int", "label": "tempest-ServersTestJSON-1786538788-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f3f799aa94f64e8cb26d93f6124efd81", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "816c6e38-e200-4544-8c5b-9fc3e16c5761", "external-id": "nsx-vlan-transportzone-195", "segmentation_id": 195, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap016b56b6-52", "ovs_interfaceid": "016b56b6-52e2-48c7-9a15-c91dda1dee97", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1237.970733] env[65726]: WARNING openstack [None req-a46b6a61-e631-41e2-b595-dffcedc65a71 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1237.971455] env[65726]: WARNING openstack [None req-a46b6a61-e631-41e2-b595-dffcedc65a71 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1238.033218] env[65726]: WARNING openstack [None req-a46b6a61-e631-41e2-b595-dffcedc65a71 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1238.033629] env[65726]: WARNING openstack [None req-a46b6a61-e631-41e2-b595-dffcedc65a71 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1238.050640] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5117005, 'name': CreateVM_Task, 'duration_secs': 0.586256} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1238.050817] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a4c229bb-e5f9-41d3-86f9-ca4cfb087558] Created VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1238.051575] env[65726]: WARNING openstack [None req-e1322ace-4de0-4198-b5e5-47926fd137bd tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1238.052045] env[65726]: WARNING openstack [None req-e1322ace-4de0-4198-b5e5-47926fd137bd tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1238.057174] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e1322ace-4de0-4198-b5e5-47926fd137bd tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1238.057333] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e1322ace-4de0-4198-b5e5-47926fd137bd tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1238.057661] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e1322ace-4de0-4198-b5e5-47926fd137bd tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1238.057928] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8d1784e6-03e6-4c59-a08e-81abe465a991 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1238.063064] env[65726]: DEBUG oslo_vmware.api [None req-e1322ace-4de0-4198-b5e5-47926fd137bd tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Waiting for the task: (returnval){ [ 1238.063064] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]527580e6-de09-6798-6575-35bc1d620b3d" [ 1238.063064] env[65726]: _type = "Task" [ 1238.063064] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1238.071393] env[65726]: DEBUG oslo_vmware.api [None req-e1322ace-4de0-4198-b5e5-47926fd137bd tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]527580e6-de09-6798-6575-35bc1d620b3d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1238.116300] env[65726]: DEBUG nova.network.neutron [None req-a46b6a61-e631-41e2-b595-dffcedc65a71 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: e9e93672-d509-45b3-89fe-f1c6ff4d8e78] Updating instance_info_cache with network_info: [{"id": "bc612ce3-dca5-4716-9f5c-60a6247e4b42", "address": "fa:16:3e:86:90:ff", "network": {"id": "336fe51f-971a-447d-8e68-505640e0db1b", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1030314703-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6b56f93ba0e14521921484ac2c785c43", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8ee1c76d-1a61-4546-85cb-d4bd3c1b35ef", "external-id": "nsx-vlan-transportzone-161", "segmentation_id": 161, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbc612ce3-dc", "ovs_interfaceid": "bc612ce3-dca5-4716-9f5c-60a6247e4b42", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1238.446311] env[65726]: DEBUG oslo_concurrency.lockutils [None req-01b26dc0-172c-48e9-b562-30576ec9ec52 tempest-ServerPasswordTestJSON-84500404 tempest-ServerPasswordTestJSON-84500404-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1238.446663] env[65726]: DEBUG oslo_concurrency.lockutils [None req-01b26dc0-172c-48e9-b562-30576ec9ec52 tempest-ServerPasswordTestJSON-84500404 tempest-ServerPasswordTestJSON-84500404-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1238.446791] env[65726]: DEBUG nova.objects.instance [None req-01b26dc0-172c-48e9-b562-30576ec9ec52 tempest-ServerPasswordTestJSON-84500404 tempest-ServerPasswordTestJSON-84500404-project-member] Lazy-loading 'resources' on Instance uuid a1dc9762-ad05-4cd4-b3c3-eb644ba4212b {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1238.456057] env[65726]: DEBUG oslo_concurrency.lockutils [req-bf55f577-e3f8-4ff6-ba58-a699309176ed req-5335ae9a-3052-45b1-a098-ce00c0d3b287 service nova] Releasing lock "refresh_cache-a4c229bb-e5f9-41d3-86f9-ca4cfb087558" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1238.456057] env[65726]: DEBUG nova.compute.manager [req-bf55f577-e3f8-4ff6-ba58-a699309176ed req-5335ae9a-3052-45b1-a098-ce00c0d3b287 service nova] [instance: e9e93672-d509-45b3-89fe-f1c6ff4d8e78] Received event network-vif-plugged-bc612ce3-dca5-4716-9f5c-60a6247e4b42 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1238.456057] env[65726]: DEBUG oslo_concurrency.lockutils [req-bf55f577-e3f8-4ff6-ba58-a699309176ed req-5335ae9a-3052-45b1-a098-ce00c0d3b287 service nova] Acquiring lock "e9e93672-d509-45b3-89fe-f1c6ff4d8e78-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1238.456057] env[65726]: DEBUG oslo_concurrency.lockutils [req-bf55f577-e3f8-4ff6-ba58-a699309176ed req-5335ae9a-3052-45b1-a098-ce00c0d3b287 service nova] Lock "e9e93672-d509-45b3-89fe-f1c6ff4d8e78-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1238.456338] env[65726]: DEBUG oslo_concurrency.lockutils [req-bf55f577-e3f8-4ff6-ba58-a699309176ed req-5335ae9a-3052-45b1-a098-ce00c0d3b287 service nova] Lock "e9e93672-d509-45b3-89fe-f1c6ff4d8e78-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1238.456465] env[65726]: DEBUG nova.compute.manager [req-bf55f577-e3f8-4ff6-ba58-a699309176ed req-5335ae9a-3052-45b1-a098-ce00c0d3b287 service nova] [instance: e9e93672-d509-45b3-89fe-f1c6ff4d8e78] No waiting events found dispatching network-vif-plugged-bc612ce3-dca5-4716-9f5c-60a6247e4b42 {{(pid=65726) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1238.456739] env[65726]: WARNING nova.compute.manager [req-bf55f577-e3f8-4ff6-ba58-a699309176ed req-5335ae9a-3052-45b1-a098-ce00c0d3b287 service nova] [instance: e9e93672-d509-45b3-89fe-f1c6ff4d8e78] Received unexpected event network-vif-plugged-bc612ce3-dca5-4716-9f5c-60a6247e4b42 for instance with vm_state building and task_state spawning. [ 1238.456979] env[65726]: DEBUG nova.compute.manager [req-bf55f577-e3f8-4ff6-ba58-a699309176ed req-5335ae9a-3052-45b1-a098-ce00c0d3b287 service nova] [instance: e9e93672-d509-45b3-89fe-f1c6ff4d8e78] Received event network-changed-bc612ce3-dca5-4716-9f5c-60a6247e4b42 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1238.457220] env[65726]: DEBUG nova.compute.manager [req-bf55f577-e3f8-4ff6-ba58-a699309176ed req-5335ae9a-3052-45b1-a098-ce00c0d3b287 service nova] [instance: e9e93672-d509-45b3-89fe-f1c6ff4d8e78] Refreshing instance network info cache due to event network-changed-bc612ce3-dca5-4716-9f5c-60a6247e4b42. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 1238.457435] env[65726]: DEBUG oslo_concurrency.lockutils [req-bf55f577-e3f8-4ff6-ba58-a699309176ed req-5335ae9a-3052-45b1-a098-ce00c0d3b287 service nova] Acquiring lock "refresh_cache-e9e93672-d509-45b3-89fe-f1c6ff4d8e78" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1238.573735] env[65726]: DEBUG oslo_vmware.api [None req-e1322ace-4de0-4198-b5e5-47926fd137bd tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]527580e6-de09-6798-6575-35bc1d620b3d, 'name': SearchDatastore_Task, 'duration_secs': 0.010758} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1238.573968] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e1322ace-4de0-4198-b5e5-47926fd137bd tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1238.574217] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-e1322ace-4de0-4198-b5e5-47926fd137bd tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: a4c229bb-e5f9-41d3-86f9-ca4cfb087558] Processing image b52362a3-ee8a-4cbf-b06f-513b0cc8f95c {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1238.574482] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e1322ace-4de0-4198-b5e5-47926fd137bd tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1238.574624] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e1322ace-4de0-4198-b5e5-47926fd137bd tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1238.574802] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-e1322ace-4de0-4198-b5e5-47926fd137bd tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1238.575094] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-27de6d7b-affb-4147-8c52-f1928bbaebc4 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1238.583863] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-e1322ace-4de0-4198-b5e5-47926fd137bd tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1238.584064] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-e1322ace-4de0-4198-b5e5-47926fd137bd tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65726) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1238.584811] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-59f5baca-4bb0-406b-8bbc-9840966dfd2d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1238.590937] env[65726]: DEBUG oslo_vmware.api [None req-e1322ace-4de0-4198-b5e5-47926fd137bd tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Waiting for the task: (returnval){ [ 1238.590937] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]526375a7-e13a-9e43-70bc-dea8a9b0e9ae" [ 1238.590937] env[65726]: _type = "Task" [ 1238.590937] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1238.598798] env[65726]: DEBUG oslo_vmware.api [None req-e1322ace-4de0-4198-b5e5-47926fd137bd tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]526375a7-e13a-9e43-70bc-dea8a9b0e9ae, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1238.619598] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a46b6a61-e631-41e2-b595-dffcedc65a71 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Releasing lock "refresh_cache-e9e93672-d509-45b3-89fe-f1c6ff4d8e78" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1238.619977] env[65726]: DEBUG nova.compute.manager [None req-a46b6a61-e631-41e2-b595-dffcedc65a71 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: e9e93672-d509-45b3-89fe-f1c6ff4d8e78] Instance network_info: |[{"id": "bc612ce3-dca5-4716-9f5c-60a6247e4b42", "address": "fa:16:3e:86:90:ff", "network": {"id": "336fe51f-971a-447d-8e68-505640e0db1b", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1030314703-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6b56f93ba0e14521921484ac2c785c43", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8ee1c76d-1a61-4546-85cb-d4bd3c1b35ef", "external-id": "nsx-vlan-transportzone-161", "segmentation_id": 161, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbc612ce3-dc", "ovs_interfaceid": "bc612ce3-dca5-4716-9f5c-60a6247e4b42", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1238.620328] env[65726]: DEBUG oslo_concurrency.lockutils [req-bf55f577-e3f8-4ff6-ba58-a699309176ed req-5335ae9a-3052-45b1-a098-ce00c0d3b287 service nova] Acquired lock "refresh_cache-e9e93672-d509-45b3-89fe-f1c6ff4d8e78" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1238.620505] env[65726]: DEBUG nova.network.neutron [req-bf55f577-e3f8-4ff6-ba58-a699309176ed req-5335ae9a-3052-45b1-a098-ce00c0d3b287 service nova] [instance: e9e93672-d509-45b3-89fe-f1c6ff4d8e78] Refreshing network info cache for port bc612ce3-dca5-4716-9f5c-60a6247e4b42 {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 1238.621762] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-a46b6a61-e631-41e2-b595-dffcedc65a71 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: e9e93672-d509-45b3-89fe-f1c6ff4d8e78] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:86:90:ff', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '8ee1c76d-1a61-4546-85cb-d4bd3c1b35ef', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'bc612ce3-dca5-4716-9f5c-60a6247e4b42', 'vif_model': 'vmxnet3'}] {{(pid=65726) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1238.629780] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-a46b6a61-e631-41e2-b595-dffcedc65a71 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1238.631498] env[65726]: WARNING openstack [req-bf55f577-e3f8-4ff6-ba58-a699309176ed req-5335ae9a-3052-45b1-a098-ce00c0d3b287 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1238.631888] env[65726]: WARNING openstack [req-bf55f577-e3f8-4ff6-ba58-a699309176ed req-5335ae9a-3052-45b1-a098-ce00c0d3b287 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1238.638478] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e9e93672-d509-45b3-89fe-f1c6ff4d8e78] Creating VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1238.639314] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-864645c7-8f89-4005-a946-731593f746f7 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1238.662720] env[65726]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1238.662720] env[65726]: value = "task-5117006" [ 1238.662720] env[65726]: _type = "Task" [ 1238.662720] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1238.672090] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5117006, 'name': CreateVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1238.756039] env[65726]: WARNING openstack [req-bf55f577-e3f8-4ff6-ba58-a699309176ed req-5335ae9a-3052-45b1-a098-ce00c0d3b287 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1238.756397] env[65726]: WARNING openstack [req-bf55f577-e3f8-4ff6-ba58-a699309176ed req-5335ae9a-3052-45b1-a098-ce00c0d3b287 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1238.827384] env[65726]: WARNING openstack [req-bf55f577-e3f8-4ff6-ba58-a699309176ed req-5335ae9a-3052-45b1-a098-ce00c0d3b287 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1238.827851] env[65726]: WARNING openstack [req-bf55f577-e3f8-4ff6-ba58-a699309176ed req-5335ae9a-3052-45b1-a098-ce00c0d3b287 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1238.934655] env[65726]: DEBUG nova.network.neutron [req-bf55f577-e3f8-4ff6-ba58-a699309176ed req-5335ae9a-3052-45b1-a098-ce00c0d3b287 service nova] [instance: e9e93672-d509-45b3-89fe-f1c6ff4d8e78] Updated VIF entry in instance network info cache for port bc612ce3-dca5-4716-9f5c-60a6247e4b42. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 1238.935065] env[65726]: DEBUG nova.network.neutron [req-bf55f577-e3f8-4ff6-ba58-a699309176ed req-5335ae9a-3052-45b1-a098-ce00c0d3b287 service nova] [instance: e9e93672-d509-45b3-89fe-f1c6ff4d8e78] Updating instance_info_cache with network_info: [{"id": "bc612ce3-dca5-4716-9f5c-60a6247e4b42", "address": "fa:16:3e:86:90:ff", "network": {"id": "336fe51f-971a-447d-8e68-505640e0db1b", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1030314703-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6b56f93ba0e14521921484ac2c785c43", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8ee1c76d-1a61-4546-85cb-d4bd3c1b35ef", "external-id": "nsx-vlan-transportzone-161", "segmentation_id": 161, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbc612ce3-dc", "ovs_interfaceid": "bc612ce3-dca5-4716-9f5c-60a6247e4b42", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1239.054523] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-438310cd-8842-42a9-89e1-71e0f712647d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1239.063281] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59752913-01be-4014-9cd5-23942e81387b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1239.097079] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba7dcbe7-efc4-4543-bafc-fc773732ad9d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1239.108218] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6e68bbd-80f5-45ef-a828-bbed5dbed344 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1239.112070] env[65726]: DEBUG oslo_vmware.api [None req-e1322ace-4de0-4198-b5e5-47926fd137bd tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]526375a7-e13a-9e43-70bc-dea8a9b0e9ae, 'name': SearchDatastore_Task, 'duration_secs': 0.016691} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1239.113209] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a6a4c88e-afe0-4c54-945d-ca4204528e55 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1239.123343] env[65726]: DEBUG nova.compute.provider_tree [None req-01b26dc0-172c-48e9-b562-30576ec9ec52 tempest-ServerPasswordTestJSON-84500404 tempest-ServerPasswordTestJSON-84500404-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1239.128221] env[65726]: DEBUG oslo_vmware.api [None req-e1322ace-4de0-4198-b5e5-47926fd137bd tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Waiting for the task: (returnval){ [ 1239.128221] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52cb0286-62bd-cb6c-8251-492eda68018f" [ 1239.128221] env[65726]: _type = "Task" [ 1239.128221] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1239.137836] env[65726]: DEBUG oslo_vmware.api [None req-e1322ace-4de0-4198-b5e5-47926fd137bd tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52cb0286-62bd-cb6c-8251-492eda68018f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1239.173418] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5117006, 'name': CreateVM_Task, 'duration_secs': 0.318535} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1239.173614] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e9e93672-d509-45b3-89fe-f1c6ff4d8e78] Created VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1239.174365] env[65726]: WARNING openstack [None req-a46b6a61-e631-41e2-b595-dffcedc65a71 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1239.174712] env[65726]: WARNING openstack [None req-a46b6a61-e631-41e2-b595-dffcedc65a71 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1239.179745] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a46b6a61-e631-41e2-b595-dffcedc65a71 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1239.179901] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a46b6a61-e631-41e2-b595-dffcedc65a71 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1239.180231] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a46b6a61-e631-41e2-b595-dffcedc65a71 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1239.180499] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c3635e04-5c57-47c5-94fa-5f8e04f11752 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1239.185412] env[65726]: DEBUG oslo_vmware.api [None req-a46b6a61-e631-41e2-b595-dffcedc65a71 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Waiting for the task: (returnval){ [ 1239.185412] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5270818e-1382-7014-0c2a-663634403924" [ 1239.185412] env[65726]: _type = "Task" [ 1239.185412] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1239.194784] env[65726]: DEBUG oslo_vmware.api [None req-a46b6a61-e631-41e2-b595-dffcedc65a71 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5270818e-1382-7014-0c2a-663634403924, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1239.438311] env[65726]: DEBUG oslo_concurrency.lockutils [req-bf55f577-e3f8-4ff6-ba58-a699309176ed req-5335ae9a-3052-45b1-a098-ce00c0d3b287 service nova] Releasing lock "refresh_cache-e9e93672-d509-45b3-89fe-f1c6ff4d8e78" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1239.627157] env[65726]: DEBUG nova.scheduler.client.report [None req-01b26dc0-172c-48e9-b562-30576ec9ec52 tempest-ServerPasswordTestJSON-84500404 tempest-ServerPasswordTestJSON-84500404-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1239.640604] env[65726]: DEBUG oslo_vmware.api [None req-e1322ace-4de0-4198-b5e5-47926fd137bd tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52cb0286-62bd-cb6c-8251-492eda68018f, 'name': SearchDatastore_Task, 'duration_secs': 0.013152} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1239.641720] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e1322ace-4de0-4198-b5e5-47926fd137bd tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1239.641720] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-e1322ace-4de0-4198-b5e5-47926fd137bd tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] a4c229bb-e5f9-41d3-86f9-ca4cfb087558/a4c229bb-e5f9-41d3-86f9-ca4cfb087558.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1239.641720] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-fcaef1aa-8dcc-4b0b-bbfb-cf489cfeb71d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1239.649769] env[65726]: DEBUG oslo_vmware.api [None req-e1322ace-4de0-4198-b5e5-47926fd137bd tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Waiting for the task: (returnval){ [ 1239.649769] env[65726]: value = "task-5117007" [ 1239.649769] env[65726]: _type = "Task" [ 1239.649769] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1239.658221] env[65726]: DEBUG oslo_vmware.api [None req-e1322ace-4de0-4198-b5e5-47926fd137bd tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Task: {'id': task-5117007, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1239.696172] env[65726]: DEBUG oslo_vmware.api [None req-a46b6a61-e631-41e2-b595-dffcedc65a71 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5270818e-1382-7014-0c2a-663634403924, 'name': SearchDatastore_Task, 'duration_secs': 0.010094} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1239.696497] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a46b6a61-e631-41e2-b595-dffcedc65a71 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1239.696923] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-a46b6a61-e631-41e2-b595-dffcedc65a71 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: e9e93672-d509-45b3-89fe-f1c6ff4d8e78] Processing image b52362a3-ee8a-4cbf-b06f-513b0cc8f95c {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1239.697189] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a46b6a61-e631-41e2-b595-dffcedc65a71 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1239.697332] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a46b6a61-e631-41e2-b595-dffcedc65a71 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1239.697509] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-a46b6a61-e631-41e2-b595-dffcedc65a71 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1239.697774] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-df339d64-bfb2-4d0d-967e-4c6b59298644 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1239.707234] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-a46b6a61-e631-41e2-b595-dffcedc65a71 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1239.707445] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-a46b6a61-e631-41e2-b595-dffcedc65a71 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65726) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1239.708279] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-313eeabc-6660-4276-ab94-a52f48a4bbad {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1239.714560] env[65726]: DEBUG oslo_vmware.api [None req-a46b6a61-e631-41e2-b595-dffcedc65a71 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Waiting for the task: (returnval){ [ 1239.714560] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5216927e-910f-59da-2333-8f61718d93e7" [ 1239.714560] env[65726]: _type = "Task" [ 1239.714560] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1239.724138] env[65726]: DEBUG oslo_vmware.api [None req-a46b6a61-e631-41e2-b595-dffcedc65a71 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5216927e-910f-59da-2333-8f61718d93e7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1240.135520] env[65726]: DEBUG oslo_concurrency.lockutils [None req-01b26dc0-172c-48e9-b562-30576ec9ec52 tempest-ServerPasswordTestJSON-84500404 tempest-ServerPasswordTestJSON-84500404-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.689s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1240.160763] env[65726]: DEBUG oslo_vmware.api [None req-e1322ace-4de0-4198-b5e5-47926fd137bd tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Task: {'id': task-5117007, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1240.162541] env[65726]: INFO nova.scheduler.client.report [None req-01b26dc0-172c-48e9-b562-30576ec9ec52 tempest-ServerPasswordTestJSON-84500404 tempest-ServerPasswordTestJSON-84500404-project-member] Deleted allocations for instance a1dc9762-ad05-4cd4-b3c3-eb644ba4212b [ 1240.228111] env[65726]: DEBUG oslo_vmware.api [None req-a46b6a61-e631-41e2-b595-dffcedc65a71 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5216927e-910f-59da-2333-8f61718d93e7, 'name': SearchDatastore_Task, 'duration_secs': 0.018829} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1240.229480] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-11e20c1a-581e-47a9-97fd-de26c6688ac1 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1240.238285] env[65726]: DEBUG oslo_vmware.api [None req-a46b6a61-e631-41e2-b595-dffcedc65a71 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Waiting for the task: (returnval){ [ 1240.238285] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5239917e-6c2c-bcab-579e-fcb16e6440c8" [ 1240.238285] env[65726]: _type = "Task" [ 1240.238285] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1240.249837] env[65726]: DEBUG oslo_vmware.api [None req-a46b6a61-e631-41e2-b595-dffcedc65a71 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5239917e-6c2c-bcab-579e-fcb16e6440c8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1240.660840] env[65726]: DEBUG oslo_vmware.api [None req-e1322ace-4de0-4198-b5e5-47926fd137bd tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Task: {'id': task-5117007, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.893919} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1240.661216] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-e1322ace-4de0-4198-b5e5-47926fd137bd tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] a4c229bb-e5f9-41d3-86f9-ca4cfb087558/a4c229bb-e5f9-41d3-86f9-ca4cfb087558.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1240.661440] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-e1322ace-4de0-4198-b5e5-47926fd137bd tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: a4c229bb-e5f9-41d3-86f9-ca4cfb087558] Extending root virtual disk to 1048576 {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1240.661731] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-7ac32375-d2b9-407b-89d5-2d6547bae5c0 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1240.669538] env[65726]: DEBUG oslo_vmware.api [None req-e1322ace-4de0-4198-b5e5-47926fd137bd tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Waiting for the task: (returnval){ [ 1240.669538] env[65726]: value = "task-5117008" [ 1240.669538] env[65726]: _type = "Task" [ 1240.669538] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1240.680144] env[65726]: DEBUG oslo_vmware.api [None req-e1322ace-4de0-4198-b5e5-47926fd137bd tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Task: {'id': task-5117008, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1240.680423] env[65726]: DEBUG oslo_concurrency.lockutils [None req-01b26dc0-172c-48e9-b562-30576ec9ec52 tempest-ServerPasswordTestJSON-84500404 tempest-ServerPasswordTestJSON-84500404-project-member] Lock "a1dc9762-ad05-4cd4-b3c3-eb644ba4212b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.641s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1240.749396] env[65726]: DEBUG oslo_vmware.api [None req-a46b6a61-e631-41e2-b595-dffcedc65a71 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5239917e-6c2c-bcab-579e-fcb16e6440c8, 'name': SearchDatastore_Task, 'duration_secs': 0.075863} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1240.749727] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a46b6a61-e631-41e2-b595-dffcedc65a71 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1240.750073] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-a46b6a61-e631-41e2-b595-dffcedc65a71 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] e9e93672-d509-45b3-89fe-f1c6ff4d8e78/e9e93672-d509-45b3-89fe-f1c6ff4d8e78.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1240.750416] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-50b32db2-c5cf-495e-bf86-653a46fc9a7a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1240.757840] env[65726]: DEBUG oslo_vmware.api [None req-a46b6a61-e631-41e2-b595-dffcedc65a71 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Waiting for the task: (returnval){ [ 1240.757840] env[65726]: value = "task-5117009" [ 1240.757840] env[65726]: _type = "Task" [ 1240.757840] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1240.767225] env[65726]: DEBUG oslo_vmware.api [None req-a46b6a61-e631-41e2-b595-dffcedc65a71 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': task-5117009, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1241.183919] env[65726]: DEBUG oslo_vmware.api [None req-e1322ace-4de0-4198-b5e5-47926fd137bd tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Task: {'id': task-5117008, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067972} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1241.184132] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-e1322ace-4de0-4198-b5e5-47926fd137bd tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: a4c229bb-e5f9-41d3-86f9-ca4cfb087558] Extended root virtual disk {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1241.185125] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7457f2ba-a3a6-4985-98e3-ecbe7f790747 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1241.213804] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-e1322ace-4de0-4198-b5e5-47926fd137bd tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: a4c229bb-e5f9-41d3-86f9-ca4cfb087558] Reconfiguring VM instance instance-00000074 to attach disk [datastore1] a4c229bb-e5f9-41d3-86f9-ca4cfb087558/a4c229bb-e5f9-41d3-86f9-ca4cfb087558.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1241.214202] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0fc11c3f-4680-4394-9336-ea4e42b0975b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1241.240841] env[65726]: DEBUG oslo_vmware.api [None req-e1322ace-4de0-4198-b5e5-47926fd137bd tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Waiting for the task: (returnval){ [ 1241.240841] env[65726]: value = "task-5117010" [ 1241.240841] env[65726]: _type = "Task" [ 1241.240841] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1241.253602] env[65726]: DEBUG oslo_vmware.api [None req-e1322ace-4de0-4198-b5e5-47926fd137bd tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Task: {'id': task-5117010, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1241.271388] env[65726]: DEBUG oslo_vmware.api [None req-a46b6a61-e631-41e2-b595-dffcedc65a71 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': task-5117009, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1241.752442] env[65726]: DEBUG oslo_vmware.api [None req-e1322ace-4de0-4198-b5e5-47926fd137bd tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Task: {'id': task-5117010, 'name': ReconfigVM_Task, 'duration_secs': 0.344588} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1241.752808] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-e1322ace-4de0-4198-b5e5-47926fd137bd tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: a4c229bb-e5f9-41d3-86f9-ca4cfb087558] Reconfigured VM instance instance-00000074 to attach disk [datastore1] a4c229bb-e5f9-41d3-86f9-ca4cfb087558/a4c229bb-e5f9-41d3-86f9-ca4cfb087558.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1241.753491] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e178e013-3923-4b16-82db-a8954e0dff98 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1241.760563] env[65726]: DEBUG oslo_vmware.api [None req-e1322ace-4de0-4198-b5e5-47926fd137bd tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Waiting for the task: (returnval){ [ 1241.760563] env[65726]: value = "task-5117011" [ 1241.760563] env[65726]: _type = "Task" [ 1241.760563] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1241.774838] env[65726]: DEBUG oslo_vmware.api [None req-a46b6a61-e631-41e2-b595-dffcedc65a71 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': task-5117009, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.542526} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1241.775073] env[65726]: DEBUG oslo_vmware.api [None req-e1322ace-4de0-4198-b5e5-47926fd137bd tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Task: {'id': task-5117011, 'name': Rename_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1241.775339] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-a46b6a61-e631-41e2-b595-dffcedc65a71 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] e9e93672-d509-45b3-89fe-f1c6ff4d8e78/e9e93672-d509-45b3-89fe-f1c6ff4d8e78.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1241.775578] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-a46b6a61-e631-41e2-b595-dffcedc65a71 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: e9e93672-d509-45b3-89fe-f1c6ff4d8e78] Extending root virtual disk to 1048576 {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1241.775830] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a33ef215-e641-405e-ace7-0e99ee191132 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1241.782574] env[65726]: DEBUG oslo_vmware.api [None req-a46b6a61-e631-41e2-b595-dffcedc65a71 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Waiting for the task: (returnval){ [ 1241.782574] env[65726]: value = "task-5117012" [ 1241.782574] env[65726]: _type = "Task" [ 1241.782574] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1241.794130] env[65726]: DEBUG oslo_vmware.api [None req-a46b6a61-e631-41e2-b595-dffcedc65a71 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': task-5117012, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1242.274654] env[65726]: DEBUG oslo_vmware.api [None req-e1322ace-4de0-4198-b5e5-47926fd137bd tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Task: {'id': task-5117011, 'name': Rename_Task, 'duration_secs': 0.153149} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1242.274654] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-e1322ace-4de0-4198-b5e5-47926fd137bd tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: a4c229bb-e5f9-41d3-86f9-ca4cfb087558] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1242.274654] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-23cdff1b-4e98-43f4-abcf-bbc09723ec16 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1242.281070] env[65726]: DEBUG oslo_vmware.api [None req-e1322ace-4de0-4198-b5e5-47926fd137bd tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Waiting for the task: (returnval){ [ 1242.281070] env[65726]: value = "task-5117013" [ 1242.281070] env[65726]: _type = "Task" [ 1242.281070] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1242.292707] env[65726]: DEBUG oslo_vmware.api [None req-e1322ace-4de0-4198-b5e5-47926fd137bd tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Task: {'id': task-5117013, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1242.295911] env[65726]: DEBUG oslo_vmware.api [None req-a46b6a61-e631-41e2-b595-dffcedc65a71 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': task-5117012, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.076629} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1242.296261] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-a46b6a61-e631-41e2-b595-dffcedc65a71 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: e9e93672-d509-45b3-89fe-f1c6ff4d8e78] Extended root virtual disk {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1242.297512] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1caad301-761b-43c7-812c-974bb99335ad {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1242.323937] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-a46b6a61-e631-41e2-b595-dffcedc65a71 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: e9e93672-d509-45b3-89fe-f1c6ff4d8e78] Reconfiguring VM instance instance-00000075 to attach disk [datastore1] e9e93672-d509-45b3-89fe-f1c6ff4d8e78/e9e93672-d509-45b3-89fe-f1c6ff4d8e78.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1242.324376] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-08cc930c-7ae4-46bf-958a-3989747aec09 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1242.346975] env[65726]: DEBUG oslo_vmware.api [None req-a46b6a61-e631-41e2-b595-dffcedc65a71 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Waiting for the task: (returnval){ [ 1242.346975] env[65726]: value = "task-5117014" [ 1242.346975] env[65726]: _type = "Task" [ 1242.346975] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1242.360554] env[65726]: DEBUG oslo_vmware.api [None req-a46b6a61-e631-41e2-b595-dffcedc65a71 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': task-5117014, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1242.793181] env[65726]: DEBUG oslo_vmware.api [None req-e1322ace-4de0-4198-b5e5-47926fd137bd tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Task: {'id': task-5117013, 'name': PowerOnVM_Task, 'duration_secs': 0.510538} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1242.793509] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-e1322ace-4de0-4198-b5e5-47926fd137bd tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: a4c229bb-e5f9-41d3-86f9-ca4cfb087558] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1242.793687] env[65726]: INFO nova.compute.manager [None req-e1322ace-4de0-4198-b5e5-47926fd137bd tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: a4c229bb-e5f9-41d3-86f9-ca4cfb087558] Took 7.95 seconds to spawn the instance on the hypervisor. [ 1242.793874] env[65726]: DEBUG nova.compute.manager [None req-e1322ace-4de0-4198-b5e5-47926fd137bd tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: a4c229bb-e5f9-41d3-86f9-ca4cfb087558] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1242.794829] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2339567d-6ba7-4a00-9c47-d5b57de7af5a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1242.861075] env[65726]: DEBUG oslo_vmware.api [None req-a46b6a61-e631-41e2-b595-dffcedc65a71 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': task-5117014, 'name': ReconfigVM_Task, 'duration_secs': 0.346584} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1242.861379] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-a46b6a61-e631-41e2-b595-dffcedc65a71 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: e9e93672-d509-45b3-89fe-f1c6ff4d8e78] Reconfigured VM instance instance-00000075 to attach disk [datastore1] e9e93672-d509-45b3-89fe-f1c6ff4d8e78/e9e93672-d509-45b3-89fe-f1c6ff4d8e78.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1242.862176] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-fe6d405e-ab79-4c6b-91c4-363070c5842e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1242.868679] env[65726]: DEBUG oslo_vmware.api [None req-a46b6a61-e631-41e2-b595-dffcedc65a71 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Waiting for the task: (returnval){ [ 1242.868679] env[65726]: value = "task-5117015" [ 1242.868679] env[65726]: _type = "Task" [ 1242.868679] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1242.877832] env[65726]: DEBUG oslo_vmware.api [None req-a46b6a61-e631-41e2-b595-dffcedc65a71 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': task-5117015, 'name': Rename_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1243.315427] env[65726]: INFO nova.compute.manager [None req-e1322ace-4de0-4198-b5e5-47926fd137bd tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: a4c229bb-e5f9-41d3-86f9-ca4cfb087558] Took 12.74 seconds to build instance. [ 1243.379261] env[65726]: DEBUG oslo_vmware.api [None req-a46b6a61-e631-41e2-b595-dffcedc65a71 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': task-5117015, 'name': Rename_Task} progress is 99%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1243.818469] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e1322ace-4de0-4198-b5e5-47926fd137bd tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Lock "a4c229bb-e5f9-41d3-86f9-ca4cfb087558" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.254s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1243.879916] env[65726]: DEBUG oslo_vmware.api [None req-a46b6a61-e631-41e2-b595-dffcedc65a71 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': task-5117015, 'name': Rename_Task} progress is 99%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1244.379717] env[65726]: DEBUG oslo_vmware.api [None req-a46b6a61-e631-41e2-b595-dffcedc65a71 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': task-5117015, 'name': Rename_Task, 'duration_secs': 1.44711} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1244.380020] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-a46b6a61-e631-41e2-b595-dffcedc65a71 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: e9e93672-d509-45b3-89fe-f1c6ff4d8e78] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1244.380279] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-97439085-27ea-4082-b7f3-6a44b07cd0f1 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1244.386894] env[65726]: DEBUG oslo_vmware.api [None req-a46b6a61-e631-41e2-b595-dffcedc65a71 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Waiting for the task: (returnval){ [ 1244.386894] env[65726]: value = "task-5117016" [ 1244.386894] env[65726]: _type = "Task" [ 1244.386894] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1244.395191] env[65726]: DEBUG oslo_vmware.api [None req-a46b6a61-e631-41e2-b595-dffcedc65a71 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': task-5117016, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1244.650123] env[65726]: DEBUG oslo_concurrency.lockutils [None req-2033a75e-05f6-474d-9736-3bf92577a37b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Acquiring lock "1d141c86-6da9-41b9-a078-50ecf0236f85" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1244.650452] env[65726]: DEBUG oslo_concurrency.lockutils [None req-2033a75e-05f6-474d-9736-3bf92577a37b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Lock "1d141c86-6da9-41b9-a078-50ecf0236f85" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1244.900799] env[65726]: DEBUG oslo_vmware.api [None req-a46b6a61-e631-41e2-b595-dffcedc65a71 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': task-5117016, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1245.153099] env[65726]: DEBUG nova.compute.manager [None req-2033a75e-05f6-474d-9736-3bf92577a37b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 1d141c86-6da9-41b9-a078-50ecf0236f85] Starting instance... {{(pid=65726) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 1245.398413] env[65726]: DEBUG oslo_vmware.api [None req-a46b6a61-e631-41e2-b595-dffcedc65a71 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': task-5117016, 'name': PowerOnVM_Task, 'duration_secs': 0.560268} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1245.398686] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-a46b6a61-e631-41e2-b595-dffcedc65a71 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: e9e93672-d509-45b3-89fe-f1c6ff4d8e78] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1245.398888] env[65726]: INFO nova.compute.manager [None req-a46b6a61-e631-41e2-b595-dffcedc65a71 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: e9e93672-d509-45b3-89fe-f1c6ff4d8e78] Took 8.81 seconds to spawn the instance on the hypervisor. [ 1245.399079] env[65726]: DEBUG nova.compute.manager [None req-a46b6a61-e631-41e2-b595-dffcedc65a71 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: e9e93672-d509-45b3-89fe-f1c6ff4d8e78] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1245.399851] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da464655-f5fe-40f5-93c7-1e7ccc1dcc97 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1245.677216] env[65726]: DEBUG oslo_concurrency.lockutils [None req-2033a75e-05f6-474d-9736-3bf92577a37b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1245.677513] env[65726]: DEBUG oslo_concurrency.lockutils [None req-2033a75e-05f6-474d-9736-3bf92577a37b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1245.679624] env[65726]: INFO nova.compute.claims [None req-2033a75e-05f6-474d-9736-3bf92577a37b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 1d141c86-6da9-41b9-a078-50ecf0236f85] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1245.918260] env[65726]: INFO nova.compute.manager [None req-a46b6a61-e631-41e2-b595-dffcedc65a71 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: e9e93672-d509-45b3-89fe-f1c6ff4d8e78] Took 13.18 seconds to build instance. [ 1246.420901] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a46b6a61-e631-41e2-b595-dffcedc65a71 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Lock "e9e93672-d509-45b3-89fe-f1c6ff4d8e78" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.709s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1246.447702] env[65726]: DEBUG oslo_concurrency.lockutils [None req-f7bd014e-cae1-4564-bb38-0aa7f743015a tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Acquiring lock "e9e93672-d509-45b3-89fe-f1c6ff4d8e78" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1246.447877] env[65726]: DEBUG oslo_concurrency.lockutils [None req-f7bd014e-cae1-4564-bb38-0aa7f743015a tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Lock "e9e93672-d509-45b3-89fe-f1c6ff4d8e78" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1246.448027] env[65726]: DEBUG nova.compute.manager [None req-f7bd014e-cae1-4564-bb38-0aa7f743015a tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: e9e93672-d509-45b3-89fe-f1c6ff4d8e78] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1246.448963] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c2f1e66-d588-49c5-bcde-26e24196388a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1246.456895] env[65726]: DEBUG nova.compute.manager [None req-f7bd014e-cae1-4564-bb38-0aa7f743015a tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: e9e93672-d509-45b3-89fe-f1c6ff4d8e78] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=65726) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3459}} [ 1246.457472] env[65726]: DEBUG nova.objects.instance [None req-f7bd014e-cae1-4564-bb38-0aa7f743015a tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Lazy-loading 'flavor' on Instance uuid e9e93672-d509-45b3-89fe-f1c6ff4d8e78 {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1246.779660] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81e2b84a-91c9-4b51-a230-832ebce1e7eb {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1246.787476] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3217e1a-9f2c-4775-a078-be84534b65b0 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1246.816763] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02e5d432-8fc3-414e-9159-710f0de566f0 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1246.824209] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3cabedbd-f3ea-48f7-93a8-07cf7e6cf11d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1246.839091] env[65726]: DEBUG nova.compute.provider_tree [None req-2033a75e-05f6-474d-9736-3bf92577a37b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1247.342557] env[65726]: DEBUG nova.scheduler.client.report [None req-2033a75e-05f6-474d-9736-3bf92577a37b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1247.465220] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-f7bd014e-cae1-4564-bb38-0aa7f743015a tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: e9e93672-d509-45b3-89fe-f1c6ff4d8e78] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1247.465579] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-80c8bf8f-e41c-4ddd-931c-f3b226192127 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1247.473571] env[65726]: DEBUG oslo_vmware.api [None req-f7bd014e-cae1-4564-bb38-0aa7f743015a tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Waiting for the task: (returnval){ [ 1247.473571] env[65726]: value = "task-5117017" [ 1247.473571] env[65726]: _type = "Task" [ 1247.473571] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1247.482331] env[65726]: DEBUG oslo_vmware.api [None req-f7bd014e-cae1-4564-bb38-0aa7f743015a tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': task-5117017, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1247.848449] env[65726]: DEBUG oslo_concurrency.lockutils [None req-2033a75e-05f6-474d-9736-3bf92577a37b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.171s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1247.848995] env[65726]: DEBUG nova.compute.manager [None req-2033a75e-05f6-474d-9736-3bf92577a37b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 1d141c86-6da9-41b9-a078-50ecf0236f85] Start building networks asynchronously for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1247.985066] env[65726]: DEBUG oslo_vmware.api [None req-f7bd014e-cae1-4564-bb38-0aa7f743015a tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': task-5117017, 'name': PowerOffVM_Task, 'duration_secs': 0.174577} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1247.985066] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-f7bd014e-cae1-4564-bb38-0aa7f743015a tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: e9e93672-d509-45b3-89fe-f1c6ff4d8e78] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1247.985066] env[65726]: DEBUG nova.compute.manager [None req-f7bd014e-cae1-4564-bb38-0aa7f743015a tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: e9e93672-d509-45b3-89fe-f1c6ff4d8e78] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1247.985727] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-531b6b72-d784-4fd5-842d-b6072486cdd7 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1248.355035] env[65726]: DEBUG nova.compute.utils [None req-2033a75e-05f6-474d-9736-3bf92577a37b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Using /dev/sd instead of None {{(pid=65726) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1248.356858] env[65726]: DEBUG nova.compute.manager [None req-2033a75e-05f6-474d-9736-3bf92577a37b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 1d141c86-6da9-41b9-a078-50ecf0236f85] Allocating IP information in the background. {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1248.356993] env[65726]: DEBUG nova.network.neutron [None req-2033a75e-05f6-474d-9736-3bf92577a37b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 1d141c86-6da9-41b9-a078-50ecf0236f85] allocate_for_instance() {{(pid=65726) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 1248.357334] env[65726]: WARNING neutronclient.v2_0.client [None req-2033a75e-05f6-474d-9736-3bf92577a37b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1248.357647] env[65726]: WARNING neutronclient.v2_0.client [None req-2033a75e-05f6-474d-9736-3bf92577a37b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1248.358255] env[65726]: WARNING openstack [None req-2033a75e-05f6-474d-9736-3bf92577a37b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1248.358585] env[65726]: WARNING openstack [None req-2033a75e-05f6-474d-9736-3bf92577a37b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1248.421483] env[65726]: DEBUG nova.policy [None req-2033a75e-05f6-474d-9736-3bf92577a37b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ccb3513b5b4f453f842139ec1815c4af', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f3f799aa94f64e8cb26d93f6124efd81', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65726) authorize /opt/stack/nova/nova/policy.py:192}} [ 1248.498022] env[65726]: DEBUG oslo_concurrency.lockutils [None req-f7bd014e-cae1-4564-bb38-0aa7f743015a tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Lock "e9e93672-d509-45b3-89fe-f1c6ff4d8e78" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.050s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1248.694398] env[65726]: DEBUG nova.network.neutron [None req-2033a75e-05f6-474d-9736-3bf92577a37b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 1d141c86-6da9-41b9-a078-50ecf0236f85] Successfully created port: d7088b35-54c6-48ac-8244-46071889261e {{(pid=65726) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1248.867204] env[65726]: DEBUG nova.compute.manager [None req-2033a75e-05f6-474d-9736-3bf92577a37b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 1d141c86-6da9-41b9-a078-50ecf0236f85] Start building block device mappings for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1248.971805] env[65726]: INFO nova.compute.manager [None req-19eeb2b5-f166-426c-957f-e5a1c2480cc2 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: e9e93672-d509-45b3-89fe-f1c6ff4d8e78] Rebuilding instance [ 1249.021888] env[65726]: DEBUG nova.compute.manager [None req-19eeb2b5-f166-426c-957f-e5a1c2480cc2 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: e9e93672-d509-45b3-89fe-f1c6ff4d8e78] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1249.022912] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b20729c-4695-4f30-98a6-d6ecd54f0806 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1249.877777] env[65726]: DEBUG nova.compute.manager [None req-2033a75e-05f6-474d-9736-3bf92577a37b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 1d141c86-6da9-41b9-a078-50ecf0236f85] Start spawning the instance on the hypervisor. {{(pid=65726) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1249.905142] env[65726]: DEBUG nova.virt.hardware [None req-2033a75e-05f6-474d-9736-3bf92577a37b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T19:28:34Z,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1539f06fa3534e90acbc3a60c4b8bd3f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T19:28:35Z,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1249.905416] env[65726]: DEBUG nova.virt.hardware [None req-2033a75e-05f6-474d-9736-3bf92577a37b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1249.905638] env[65726]: DEBUG nova.virt.hardware [None req-2033a75e-05f6-474d-9736-3bf92577a37b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1249.905856] env[65726]: DEBUG nova.virt.hardware [None req-2033a75e-05f6-474d-9736-3bf92577a37b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1249.906015] env[65726]: DEBUG nova.virt.hardware [None req-2033a75e-05f6-474d-9736-3bf92577a37b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1249.906163] env[65726]: DEBUG nova.virt.hardware [None req-2033a75e-05f6-474d-9736-3bf92577a37b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1249.906372] env[65726]: DEBUG nova.virt.hardware [None req-2033a75e-05f6-474d-9736-3bf92577a37b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1249.906543] env[65726]: DEBUG nova.virt.hardware [None req-2033a75e-05f6-474d-9736-3bf92577a37b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1249.906711] env[65726]: DEBUG nova.virt.hardware [None req-2033a75e-05f6-474d-9736-3bf92577a37b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1249.906871] env[65726]: DEBUG nova.virt.hardware [None req-2033a75e-05f6-474d-9736-3bf92577a37b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1249.907051] env[65726]: DEBUG nova.virt.hardware [None req-2033a75e-05f6-474d-9736-3bf92577a37b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1249.907922] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c5e84e0-68b8-41ac-a6ae-a40294b2bd47 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1249.916381] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f1bb994-62cf-4aa0-8011-a5dc701c9def {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1250.039399] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-19eeb2b5-f166-426c-957f-e5a1c2480cc2 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: e9e93672-d509-45b3-89fe-f1c6ff4d8e78] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1250.039811] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-180c9a2d-a2f2-4ce3-9222-615afcef61b2 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1250.048170] env[65726]: DEBUG oslo_vmware.api [None req-19eeb2b5-f166-426c-957f-e5a1c2480cc2 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Waiting for the task: (returnval){ [ 1250.048170] env[65726]: value = "task-5117018" [ 1250.048170] env[65726]: _type = "Task" [ 1250.048170] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1250.058135] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-19eeb2b5-f166-426c-957f-e5a1c2480cc2 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: e9e93672-d509-45b3-89fe-f1c6ff4d8e78] VM already powered off {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1250.058412] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-19eeb2b5-f166-426c-957f-e5a1c2480cc2 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: e9e93672-d509-45b3-89fe-f1c6ff4d8e78] Destroying instance {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1250.059296] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b99cfec2-e176-41be-940f-75aa03578cfa {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1250.066828] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-19eeb2b5-f166-426c-957f-e5a1c2480cc2 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: e9e93672-d509-45b3-89fe-f1c6ff4d8e78] Unregistering the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1250.067145] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-61103463-65b9-4abc-8926-fac876e20429 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1250.075994] env[65726]: DEBUG nova.compute.manager [req-6c997680-b949-4cff-9ab9-dd0b9e4eb349 req-d388bb42-bdd5-49a4-a1e0-8ec384d609b2 service nova] [instance: 1d141c86-6da9-41b9-a078-50ecf0236f85] Received event network-vif-plugged-d7088b35-54c6-48ac-8244-46071889261e {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1250.076498] env[65726]: DEBUG oslo_concurrency.lockutils [req-6c997680-b949-4cff-9ab9-dd0b9e4eb349 req-d388bb42-bdd5-49a4-a1e0-8ec384d609b2 service nova] Acquiring lock "1d141c86-6da9-41b9-a078-50ecf0236f85-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1250.076697] env[65726]: DEBUG oslo_concurrency.lockutils [req-6c997680-b949-4cff-9ab9-dd0b9e4eb349 req-d388bb42-bdd5-49a4-a1e0-8ec384d609b2 service nova] Lock "1d141c86-6da9-41b9-a078-50ecf0236f85-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1250.076862] env[65726]: DEBUG oslo_concurrency.lockutils [req-6c997680-b949-4cff-9ab9-dd0b9e4eb349 req-d388bb42-bdd5-49a4-a1e0-8ec384d609b2 service nova] Lock "1d141c86-6da9-41b9-a078-50ecf0236f85-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1250.077040] env[65726]: DEBUG nova.compute.manager [req-6c997680-b949-4cff-9ab9-dd0b9e4eb349 req-d388bb42-bdd5-49a4-a1e0-8ec384d609b2 service nova] [instance: 1d141c86-6da9-41b9-a078-50ecf0236f85] No waiting events found dispatching network-vif-plugged-d7088b35-54c6-48ac-8244-46071889261e {{(pid=65726) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1250.077230] env[65726]: WARNING nova.compute.manager [req-6c997680-b949-4cff-9ab9-dd0b9e4eb349 req-d388bb42-bdd5-49a4-a1e0-8ec384d609b2 service nova] [instance: 1d141c86-6da9-41b9-a078-50ecf0236f85] Received unexpected event network-vif-plugged-d7088b35-54c6-48ac-8244-46071889261e for instance with vm_state building and task_state spawning. [ 1250.144254] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-19eeb2b5-f166-426c-957f-e5a1c2480cc2 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: e9e93672-d509-45b3-89fe-f1c6ff4d8e78] Unregistered the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1250.144800] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-19eeb2b5-f166-426c-957f-e5a1c2480cc2 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: e9e93672-d509-45b3-89fe-f1c6ff4d8e78] Deleting contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1250.144800] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-19eeb2b5-f166-426c-957f-e5a1c2480cc2 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Deleting the datastore file [datastore1] e9e93672-d509-45b3-89fe-f1c6ff4d8e78 {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1250.145013] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a83c899e-c204-4c9c-aadc-176fd9ddc504 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1250.153303] env[65726]: DEBUG oslo_vmware.api [None req-19eeb2b5-f166-426c-957f-e5a1c2480cc2 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Waiting for the task: (returnval){ [ 1250.153303] env[65726]: value = "task-5117020" [ 1250.153303] env[65726]: _type = "Task" [ 1250.153303] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1250.163319] env[65726]: DEBUG oslo_vmware.api [None req-19eeb2b5-f166-426c-957f-e5a1c2480cc2 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': task-5117020, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1250.169385] env[65726]: DEBUG nova.network.neutron [None req-2033a75e-05f6-474d-9736-3bf92577a37b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 1d141c86-6da9-41b9-a078-50ecf0236f85] Successfully updated port: d7088b35-54c6-48ac-8244-46071889261e {{(pid=65726) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 1250.665061] env[65726]: DEBUG oslo_vmware.api [None req-19eeb2b5-f166-426c-957f-e5a1c2480cc2 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': task-5117020, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.165744} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1250.665061] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-19eeb2b5-f166-426c-957f-e5a1c2480cc2 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Deleted the datastore file {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1250.665061] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-19eeb2b5-f166-426c-957f-e5a1c2480cc2 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: e9e93672-d509-45b3-89fe-f1c6ff4d8e78] Deleted contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1250.665452] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-19eeb2b5-f166-426c-957f-e5a1c2480cc2 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: e9e93672-d509-45b3-89fe-f1c6ff4d8e78] Instance destroyed {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1250.672726] env[65726]: DEBUG oslo_concurrency.lockutils [None req-2033a75e-05f6-474d-9736-3bf92577a37b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Acquiring lock "refresh_cache-1d141c86-6da9-41b9-a078-50ecf0236f85" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1250.673162] env[65726]: DEBUG oslo_concurrency.lockutils [None req-2033a75e-05f6-474d-9736-3bf92577a37b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Acquired lock "refresh_cache-1d141c86-6da9-41b9-a078-50ecf0236f85" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1250.673162] env[65726]: DEBUG nova.network.neutron [None req-2033a75e-05f6-474d-9736-3bf92577a37b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 1d141c86-6da9-41b9-a078-50ecf0236f85] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1251.175754] env[65726]: WARNING openstack [None req-2033a75e-05f6-474d-9736-3bf92577a37b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1251.176172] env[65726]: WARNING openstack [None req-2033a75e-05f6-474d-9736-3bf92577a37b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1251.183118] env[65726]: DEBUG nova.network.neutron [None req-2033a75e-05f6-474d-9736-3bf92577a37b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 1d141c86-6da9-41b9-a078-50ecf0236f85] Instance cache missing network info. {{(pid=65726) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 1251.237514] env[65726]: WARNING openstack [None req-2033a75e-05f6-474d-9736-3bf92577a37b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1251.237942] env[65726]: WARNING openstack [None req-2033a75e-05f6-474d-9736-3bf92577a37b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1251.303485] env[65726]: WARNING openstack [None req-2033a75e-05f6-474d-9736-3bf92577a37b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1251.303883] env[65726]: WARNING openstack [None req-2033a75e-05f6-474d-9736-3bf92577a37b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1251.389448] env[65726]: DEBUG nova.network.neutron [None req-2033a75e-05f6-474d-9736-3bf92577a37b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 1d141c86-6da9-41b9-a078-50ecf0236f85] Updating instance_info_cache with network_info: [{"id": "d7088b35-54c6-48ac-8244-46071889261e", "address": "fa:16:3e:da:f3:5a", "network": {"id": "69642329-e986-48fa-accf-1f296c08116a", "bridge": "br-int", "label": "tempest-ServersTestJSON-1786538788-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f3f799aa94f64e8cb26d93f6124efd81", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "816c6e38-e200-4544-8c5b-9fc3e16c5761", "external-id": "nsx-vlan-transportzone-195", "segmentation_id": 195, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd7088b35-54", "ovs_interfaceid": "d7088b35-54c6-48ac-8244-46071889261e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1251.701907] env[65726]: DEBUG nova.virt.hardware [None req-19eeb2b5-f166-426c-957f-e5a1c2480cc2 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T19:28:34Z,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1539f06fa3534e90acbc3a60c4b8bd3f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T19:28:35Z,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1251.702185] env[65726]: DEBUG nova.virt.hardware [None req-19eeb2b5-f166-426c-957f-e5a1c2480cc2 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1251.702346] env[65726]: DEBUG nova.virt.hardware [None req-19eeb2b5-f166-426c-957f-e5a1c2480cc2 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1251.702509] env[65726]: DEBUG nova.virt.hardware [None req-19eeb2b5-f166-426c-957f-e5a1c2480cc2 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1251.702649] env[65726]: DEBUG nova.virt.hardware [None req-19eeb2b5-f166-426c-957f-e5a1c2480cc2 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1251.702784] env[65726]: DEBUG nova.virt.hardware [None req-19eeb2b5-f166-426c-957f-e5a1c2480cc2 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1251.703059] env[65726]: DEBUG nova.virt.hardware [None req-19eeb2b5-f166-426c-957f-e5a1c2480cc2 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1251.703262] env[65726]: DEBUG nova.virt.hardware [None req-19eeb2b5-f166-426c-957f-e5a1c2480cc2 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1251.703453] env[65726]: DEBUG nova.virt.hardware [None req-19eeb2b5-f166-426c-957f-e5a1c2480cc2 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1251.703629] env[65726]: DEBUG nova.virt.hardware [None req-19eeb2b5-f166-426c-957f-e5a1c2480cc2 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1251.703801] env[65726]: DEBUG nova.virt.hardware [None req-19eeb2b5-f166-426c-957f-e5a1c2480cc2 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1251.704735] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3ec2d3c-cf54-43ac-8d3a-cb163f6677b4 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1251.713800] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4215c4e9-9d4c-45cc-875a-ad90bf005cf8 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1251.728373] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-19eeb2b5-f166-426c-957f-e5a1c2480cc2 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: e9e93672-d509-45b3-89fe-f1c6ff4d8e78] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:86:90:ff', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '8ee1c76d-1a61-4546-85cb-d4bd3c1b35ef', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'bc612ce3-dca5-4716-9f5c-60a6247e4b42', 'vif_model': 'vmxnet3'}] {{(pid=65726) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1251.736113] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-19eeb2b5-f166-426c-957f-e5a1c2480cc2 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1251.736409] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e9e93672-d509-45b3-89fe-f1c6ff4d8e78] Creating VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1251.736683] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-4bbaec7f-9b82-4f1a-9e49-cc918c1a7f2b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1251.757554] env[65726]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1251.757554] env[65726]: value = "task-5117021" [ 1251.757554] env[65726]: _type = "Task" [ 1251.757554] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1251.769319] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5117021, 'name': CreateVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1251.892087] env[65726]: DEBUG oslo_concurrency.lockutils [None req-2033a75e-05f6-474d-9736-3bf92577a37b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Releasing lock "refresh_cache-1d141c86-6da9-41b9-a078-50ecf0236f85" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1251.892657] env[65726]: DEBUG nova.compute.manager [None req-2033a75e-05f6-474d-9736-3bf92577a37b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 1d141c86-6da9-41b9-a078-50ecf0236f85] Instance network_info: |[{"id": "d7088b35-54c6-48ac-8244-46071889261e", "address": "fa:16:3e:da:f3:5a", "network": {"id": "69642329-e986-48fa-accf-1f296c08116a", "bridge": "br-int", "label": "tempest-ServersTestJSON-1786538788-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f3f799aa94f64e8cb26d93f6124efd81", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "816c6e38-e200-4544-8c5b-9fc3e16c5761", "external-id": "nsx-vlan-transportzone-195", "segmentation_id": 195, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd7088b35-54", "ovs_interfaceid": "d7088b35-54c6-48ac-8244-46071889261e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1251.893204] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-2033a75e-05f6-474d-9736-3bf92577a37b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 1d141c86-6da9-41b9-a078-50ecf0236f85] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:da:f3:5a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '816c6e38-e200-4544-8c5b-9fc3e16c5761', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd7088b35-54c6-48ac-8244-46071889261e', 'vif_model': 'vmxnet3'}] {{(pid=65726) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1251.901258] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-2033a75e-05f6-474d-9736-3bf92577a37b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1251.901537] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1d141c86-6da9-41b9-a078-50ecf0236f85] Creating VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1251.901779] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e0b3e726-6711-4ed7-a0e8-487f659421c3 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1251.922759] env[65726]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1251.922759] env[65726]: value = "task-5117022" [ 1251.922759] env[65726]: _type = "Task" [ 1251.922759] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1251.931161] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5117022, 'name': CreateVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1252.107251] env[65726]: DEBUG nova.compute.manager [req-7f2fac8c-930f-421f-bdb5-81a1372530d6 req-1caa8edb-6b95-4840-8fff-84b90cc5a126 service nova] [instance: 1d141c86-6da9-41b9-a078-50ecf0236f85] Received event network-changed-d7088b35-54c6-48ac-8244-46071889261e {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1252.107451] env[65726]: DEBUG nova.compute.manager [req-7f2fac8c-930f-421f-bdb5-81a1372530d6 req-1caa8edb-6b95-4840-8fff-84b90cc5a126 service nova] [instance: 1d141c86-6da9-41b9-a078-50ecf0236f85] Refreshing instance network info cache due to event network-changed-d7088b35-54c6-48ac-8244-46071889261e. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 1252.107721] env[65726]: DEBUG oslo_concurrency.lockutils [req-7f2fac8c-930f-421f-bdb5-81a1372530d6 req-1caa8edb-6b95-4840-8fff-84b90cc5a126 service nova] Acquiring lock "refresh_cache-1d141c86-6da9-41b9-a078-50ecf0236f85" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1252.107878] env[65726]: DEBUG oslo_concurrency.lockutils [req-7f2fac8c-930f-421f-bdb5-81a1372530d6 req-1caa8edb-6b95-4840-8fff-84b90cc5a126 service nova] Acquired lock "refresh_cache-1d141c86-6da9-41b9-a078-50ecf0236f85" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1252.108060] env[65726]: DEBUG nova.network.neutron [req-7f2fac8c-930f-421f-bdb5-81a1372530d6 req-1caa8edb-6b95-4840-8fff-84b90cc5a126 service nova] [instance: 1d141c86-6da9-41b9-a078-50ecf0236f85] Refreshing network info cache for port d7088b35-54c6-48ac-8244-46071889261e {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 1252.267756] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5117021, 'name': CreateVM_Task, 'duration_secs': 0.364199} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1252.268193] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e9e93672-d509-45b3-89fe-f1c6ff4d8e78] Created VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1252.268788] env[65726]: WARNING openstack [None req-19eeb2b5-f166-426c-957f-e5a1c2480cc2 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1252.269220] env[65726]: WARNING openstack [None req-19eeb2b5-f166-426c-957f-e5a1c2480cc2 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1252.274460] env[65726]: DEBUG oslo_concurrency.lockutils [None req-19eeb2b5-f166-426c-957f-e5a1c2480cc2 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1252.274610] env[65726]: DEBUG oslo_concurrency.lockutils [None req-19eeb2b5-f166-426c-957f-e5a1c2480cc2 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1252.274915] env[65726]: DEBUG oslo_concurrency.lockutils [None req-19eeb2b5-f166-426c-957f-e5a1c2480cc2 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1252.275209] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5c1acdca-8825-4426-87b3-584590581d30 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1252.280396] env[65726]: DEBUG oslo_vmware.api [None req-19eeb2b5-f166-426c-957f-e5a1c2480cc2 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Waiting for the task: (returnval){ [ 1252.280396] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52ef2a52-2cb1-3cd3-c9bd-a35102f2ead4" [ 1252.280396] env[65726]: _type = "Task" [ 1252.280396] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1252.289216] env[65726]: DEBUG oslo_vmware.api [None req-19eeb2b5-f166-426c-957f-e5a1c2480cc2 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52ef2a52-2cb1-3cd3-c9bd-a35102f2ead4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1252.433902] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5117022, 'name': CreateVM_Task, 'duration_secs': 0.324032} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1252.434108] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1d141c86-6da9-41b9-a078-50ecf0236f85] Created VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1252.434855] env[65726]: WARNING openstack [None req-2033a75e-05f6-474d-9736-3bf92577a37b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1252.435222] env[65726]: WARNING openstack [None req-2033a75e-05f6-474d-9736-3bf92577a37b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1252.440312] env[65726]: DEBUG oslo_concurrency.lockutils [None req-2033a75e-05f6-474d-9736-3bf92577a37b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1252.611178] env[65726]: WARNING openstack [req-7f2fac8c-930f-421f-bdb5-81a1372530d6 req-1caa8edb-6b95-4840-8fff-84b90cc5a126 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1252.611559] env[65726]: WARNING openstack [req-7f2fac8c-930f-421f-bdb5-81a1372530d6 req-1caa8edb-6b95-4840-8fff-84b90cc5a126 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1252.702045] env[65726]: WARNING openstack [req-7f2fac8c-930f-421f-bdb5-81a1372530d6 req-1caa8edb-6b95-4840-8fff-84b90cc5a126 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1252.702448] env[65726]: WARNING openstack [req-7f2fac8c-930f-421f-bdb5-81a1372530d6 req-1caa8edb-6b95-4840-8fff-84b90cc5a126 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1252.763019] env[65726]: WARNING openstack [req-7f2fac8c-930f-421f-bdb5-81a1372530d6 req-1caa8edb-6b95-4840-8fff-84b90cc5a126 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1252.763409] env[65726]: WARNING openstack [req-7f2fac8c-930f-421f-bdb5-81a1372530d6 req-1caa8edb-6b95-4840-8fff-84b90cc5a126 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1252.790840] env[65726]: DEBUG oslo_vmware.api [None req-19eeb2b5-f166-426c-957f-e5a1c2480cc2 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52ef2a52-2cb1-3cd3-c9bd-a35102f2ead4, 'name': SearchDatastore_Task, 'duration_secs': 0.010536} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1252.791162] env[65726]: DEBUG oslo_concurrency.lockutils [None req-19eeb2b5-f166-426c-957f-e5a1c2480cc2 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1252.791400] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-19eeb2b5-f166-426c-957f-e5a1c2480cc2 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: e9e93672-d509-45b3-89fe-f1c6ff4d8e78] Processing image b52362a3-ee8a-4cbf-b06f-513b0cc8f95c {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1252.791637] env[65726]: DEBUG oslo_concurrency.lockutils [None req-19eeb2b5-f166-426c-957f-e5a1c2480cc2 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1252.791779] env[65726]: DEBUG oslo_concurrency.lockutils [None req-19eeb2b5-f166-426c-957f-e5a1c2480cc2 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1252.791956] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-19eeb2b5-f166-426c-957f-e5a1c2480cc2 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1252.792249] env[65726]: DEBUG oslo_concurrency.lockutils [None req-2033a75e-05f6-474d-9736-3bf92577a37b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1252.792571] env[65726]: DEBUG oslo_concurrency.lockutils [None req-2033a75e-05f6-474d-9736-3bf92577a37b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1252.792807] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-162dd872-e223-45f6-b771-abfbf2f98ad6 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1252.794754] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-21435bdd-64ff-446a-b3d9-3583103cd5a7 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1252.802232] env[65726]: DEBUG oslo_vmware.api [None req-2033a75e-05f6-474d-9736-3bf92577a37b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Waiting for the task: (returnval){ [ 1252.802232] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52ada627-0a64-4d50-6a9d-dce1946e9915" [ 1252.802232] env[65726]: _type = "Task" [ 1252.802232] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1252.809776] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-19eeb2b5-f166-426c-957f-e5a1c2480cc2 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1252.809952] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-19eeb2b5-f166-426c-957f-e5a1c2480cc2 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65726) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1252.810762] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1cbba06e-974e-432d-9a0f-e3412511a4d0 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1252.816492] env[65726]: DEBUG oslo_vmware.api [None req-2033a75e-05f6-474d-9736-3bf92577a37b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52ada627-0a64-4d50-6a9d-dce1946e9915, 'name': SearchDatastore_Task, 'duration_secs': 0.011118} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1252.817097] env[65726]: DEBUG oslo_concurrency.lockutils [None req-2033a75e-05f6-474d-9736-3bf92577a37b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1252.817353] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-2033a75e-05f6-474d-9736-3bf92577a37b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 1d141c86-6da9-41b9-a078-50ecf0236f85] Processing image b52362a3-ee8a-4cbf-b06f-513b0cc8f95c {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1252.817543] env[65726]: DEBUG oslo_concurrency.lockutils [None req-2033a75e-05f6-474d-9736-3bf92577a37b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1252.820311] env[65726]: DEBUG oslo_vmware.api [None req-19eeb2b5-f166-426c-957f-e5a1c2480cc2 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Waiting for the task: (returnval){ [ 1252.820311] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]520f10bc-3be1-27c0-fa51-dd87717dbc78" [ 1252.820311] env[65726]: _type = "Task" [ 1252.820311] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1252.829182] env[65726]: DEBUG oslo_vmware.api [None req-19eeb2b5-f166-426c-957f-e5a1c2480cc2 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]520f10bc-3be1-27c0-fa51-dd87717dbc78, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1252.844732] env[65726]: DEBUG nova.network.neutron [req-7f2fac8c-930f-421f-bdb5-81a1372530d6 req-1caa8edb-6b95-4840-8fff-84b90cc5a126 service nova] [instance: 1d141c86-6da9-41b9-a078-50ecf0236f85] Updated VIF entry in instance network info cache for port d7088b35-54c6-48ac-8244-46071889261e. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 1252.845083] env[65726]: DEBUG nova.network.neutron [req-7f2fac8c-930f-421f-bdb5-81a1372530d6 req-1caa8edb-6b95-4840-8fff-84b90cc5a126 service nova] [instance: 1d141c86-6da9-41b9-a078-50ecf0236f85] Updating instance_info_cache with network_info: [{"id": "d7088b35-54c6-48ac-8244-46071889261e", "address": "fa:16:3e:da:f3:5a", "network": {"id": "69642329-e986-48fa-accf-1f296c08116a", "bridge": "br-int", "label": "tempest-ServersTestJSON-1786538788-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f3f799aa94f64e8cb26d93f6124efd81", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "816c6e38-e200-4544-8c5b-9fc3e16c5761", "external-id": "nsx-vlan-transportzone-195", "segmentation_id": 195, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd7088b35-54", "ovs_interfaceid": "d7088b35-54c6-48ac-8244-46071889261e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1253.331310] env[65726]: DEBUG oslo_vmware.api [None req-19eeb2b5-f166-426c-957f-e5a1c2480cc2 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]520f10bc-3be1-27c0-fa51-dd87717dbc78, 'name': SearchDatastore_Task, 'duration_secs': 0.011206} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1253.332100] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5902081a-bf08-453f-953d-93448db98ad0 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1253.338735] env[65726]: DEBUG oslo_vmware.api [None req-19eeb2b5-f166-426c-957f-e5a1c2480cc2 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Waiting for the task: (returnval){ [ 1253.338735] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]527efaf6-6868-949f-735b-a3035e9678e2" [ 1253.338735] env[65726]: _type = "Task" [ 1253.338735] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1253.348246] env[65726]: DEBUG oslo_concurrency.lockutils [req-7f2fac8c-930f-421f-bdb5-81a1372530d6 req-1caa8edb-6b95-4840-8fff-84b90cc5a126 service nova] Releasing lock "refresh_cache-1d141c86-6da9-41b9-a078-50ecf0236f85" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1253.348665] env[65726]: DEBUG oslo_vmware.api [None req-19eeb2b5-f166-426c-957f-e5a1c2480cc2 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]527efaf6-6868-949f-735b-a3035e9678e2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1253.849982] env[65726]: DEBUG oslo_vmware.api [None req-19eeb2b5-f166-426c-957f-e5a1c2480cc2 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]527efaf6-6868-949f-735b-a3035e9678e2, 'name': SearchDatastore_Task, 'duration_secs': 0.014445} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1253.850346] env[65726]: DEBUG oslo_concurrency.lockutils [None req-19eeb2b5-f166-426c-957f-e5a1c2480cc2 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1253.850618] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-19eeb2b5-f166-426c-957f-e5a1c2480cc2 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] e9e93672-d509-45b3-89fe-f1c6ff4d8e78/e9e93672-d509-45b3-89fe-f1c6ff4d8e78.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1253.850918] env[65726]: DEBUG oslo_concurrency.lockutils [None req-2033a75e-05f6-474d-9736-3bf92577a37b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1253.851120] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-2033a75e-05f6-474d-9736-3bf92577a37b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1253.851352] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2dfd26ac-f9f7-42b4-ad8c-f62c05ac6dab {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1253.853670] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-868c2421-5ac4-4a8c-8758-31b0dc9cc1f9 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1253.863136] env[65726]: DEBUG oslo_vmware.api [None req-19eeb2b5-f166-426c-957f-e5a1c2480cc2 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Waiting for the task: (returnval){ [ 1253.863136] env[65726]: value = "task-5117023" [ 1253.863136] env[65726]: _type = "Task" [ 1253.863136] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1253.864711] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-2033a75e-05f6-474d-9736-3bf92577a37b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1253.864910] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-2033a75e-05f6-474d-9736-3bf92577a37b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65726) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1253.868724] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4b92f7c4-1bea-462f-862d-30b7efafef2b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1253.877519] env[65726]: DEBUG oslo_vmware.api [None req-19eeb2b5-f166-426c-957f-e5a1c2480cc2 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': task-5117023, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1253.879446] env[65726]: DEBUG oslo_vmware.api [None req-2033a75e-05f6-474d-9736-3bf92577a37b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Waiting for the task: (returnval){ [ 1253.879446] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52d2218b-1887-0cb7-0db7-3dec3e0eb636" [ 1253.879446] env[65726]: _type = "Task" [ 1253.879446] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1253.890156] env[65726]: DEBUG oslo_vmware.api [None req-2033a75e-05f6-474d-9736-3bf92577a37b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52d2218b-1887-0cb7-0db7-3dec3e0eb636, 'name': SearchDatastore_Task, 'duration_secs': 0.011027} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1253.891057] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5d5ea7c4-d752-485f-ae7d-22171af0abb6 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1253.897744] env[65726]: DEBUG oslo_vmware.api [None req-2033a75e-05f6-474d-9736-3bf92577a37b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Waiting for the task: (returnval){ [ 1253.897744] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52030557-bfc3-3fd5-ddc4-3416447aeade" [ 1253.897744] env[65726]: _type = "Task" [ 1253.897744] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1253.907713] env[65726]: DEBUG oslo_vmware.api [None req-2033a75e-05f6-474d-9736-3bf92577a37b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52030557-bfc3-3fd5-ddc4-3416447aeade, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1254.376436] env[65726]: DEBUG oslo_vmware.api [None req-19eeb2b5-f166-426c-957f-e5a1c2480cc2 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': task-5117023, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1254.409894] env[65726]: DEBUG oslo_vmware.api [None req-2033a75e-05f6-474d-9736-3bf92577a37b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52030557-bfc3-3fd5-ddc4-3416447aeade, 'name': SearchDatastore_Task, 'duration_secs': 0.011819} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1254.410213] env[65726]: DEBUG oslo_concurrency.lockutils [None req-2033a75e-05f6-474d-9736-3bf92577a37b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1254.410500] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-2033a75e-05f6-474d-9736-3bf92577a37b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] 1d141c86-6da9-41b9-a078-50ecf0236f85/1d141c86-6da9-41b9-a078-50ecf0236f85.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1254.410989] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-992a7e69-4f5a-4b74-8bd4-7050141da55f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1254.421554] env[65726]: DEBUG oslo_vmware.api [None req-2033a75e-05f6-474d-9736-3bf92577a37b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Waiting for the task: (returnval){ [ 1254.421554] env[65726]: value = "task-5117024" [ 1254.421554] env[65726]: _type = "Task" [ 1254.421554] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1254.432809] env[65726]: DEBUG oslo_vmware.api [None req-2033a75e-05f6-474d-9736-3bf92577a37b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Task: {'id': task-5117024, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1254.877844] env[65726]: DEBUG oslo_vmware.api [None req-19eeb2b5-f166-426c-957f-e5a1c2480cc2 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': task-5117023, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.601485} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1254.878296] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-19eeb2b5-f166-426c-957f-e5a1c2480cc2 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] e9e93672-d509-45b3-89fe-f1c6ff4d8e78/e9e93672-d509-45b3-89fe-f1c6ff4d8e78.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1254.878561] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-19eeb2b5-f166-426c-957f-e5a1c2480cc2 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: e9e93672-d509-45b3-89fe-f1c6ff4d8e78] Extending root virtual disk to 1048576 {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1254.878940] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ba969561-e5e8-44cc-961b-d9c53071bf6d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1254.891906] env[65726]: DEBUG oslo_vmware.api [None req-19eeb2b5-f166-426c-957f-e5a1c2480cc2 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Waiting for the task: (returnval){ [ 1254.891906] env[65726]: value = "task-5117025" [ 1254.891906] env[65726]: _type = "Task" [ 1254.891906] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1254.901975] env[65726]: DEBUG oslo_vmware.api [None req-19eeb2b5-f166-426c-957f-e5a1c2480cc2 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': task-5117025, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1254.934285] env[65726]: DEBUG oslo_vmware.api [None req-2033a75e-05f6-474d-9736-3bf92577a37b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Task: {'id': task-5117024, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1254.966018] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a92da0c2-69e1-4c8b-a9b6-c6ef25fbbe48 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Acquiring lock "b353b5e9-500c-42d3-a87f-880a624febec" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1254.966424] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a92da0c2-69e1-4c8b-a9b6-c6ef25fbbe48 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Lock "b353b5e9-500c-42d3-a87f-880a624febec" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1254.966669] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a92da0c2-69e1-4c8b-a9b6-c6ef25fbbe48 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Acquiring lock "b353b5e9-500c-42d3-a87f-880a624febec-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1254.966819] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a92da0c2-69e1-4c8b-a9b6-c6ef25fbbe48 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Lock "b353b5e9-500c-42d3-a87f-880a624febec-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1254.967094] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a92da0c2-69e1-4c8b-a9b6-c6ef25fbbe48 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Lock "b353b5e9-500c-42d3-a87f-880a624febec-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1254.969794] env[65726]: INFO nova.compute.manager [None req-a92da0c2-69e1-4c8b-a9b6-c6ef25fbbe48 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: b353b5e9-500c-42d3-a87f-880a624febec] Terminating instance [ 1255.402474] env[65726]: DEBUG oslo_vmware.api [None req-19eeb2b5-f166-426c-957f-e5a1c2480cc2 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': task-5117025, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.151403} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1255.402864] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-19eeb2b5-f166-426c-957f-e5a1c2480cc2 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: e9e93672-d509-45b3-89fe-f1c6ff4d8e78] Extended root virtual disk {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1255.403715] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6be4be1-8eeb-439b-836b-cee87d5824dd {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1255.426577] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-19eeb2b5-f166-426c-957f-e5a1c2480cc2 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: e9e93672-d509-45b3-89fe-f1c6ff4d8e78] Reconfiguring VM instance instance-00000075 to attach disk [datastore1] e9e93672-d509-45b3-89fe-f1c6ff4d8e78/e9e93672-d509-45b3-89fe-f1c6ff4d8e78.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1255.427025] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7f2ff159-5dc7-49d4-b77a-010d7330a94f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1255.451432] env[65726]: DEBUG oslo_vmware.api [None req-2033a75e-05f6-474d-9736-3bf92577a37b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Task: {'id': task-5117024, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.794481} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1255.452872] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-2033a75e-05f6-474d-9736-3bf92577a37b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] 1d141c86-6da9-41b9-a078-50ecf0236f85/1d141c86-6da9-41b9-a078-50ecf0236f85.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1255.453158] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-2033a75e-05f6-474d-9736-3bf92577a37b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 1d141c86-6da9-41b9-a078-50ecf0236f85] Extending root virtual disk to 1048576 {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1255.453524] env[65726]: DEBUG oslo_vmware.api [None req-19eeb2b5-f166-426c-957f-e5a1c2480cc2 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Waiting for the task: (returnval){ [ 1255.453524] env[65726]: value = "task-5117026" [ 1255.453524] env[65726]: _type = "Task" [ 1255.453524] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1255.453854] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a1dc85ea-7617-48ca-bc6f-4c94fdac1a5e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1255.463855] env[65726]: DEBUG oslo_vmware.api [None req-19eeb2b5-f166-426c-957f-e5a1c2480cc2 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': task-5117026, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1255.465236] env[65726]: DEBUG oslo_vmware.api [None req-2033a75e-05f6-474d-9736-3bf92577a37b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Waiting for the task: (returnval){ [ 1255.465236] env[65726]: value = "task-5117027" [ 1255.465236] env[65726]: _type = "Task" [ 1255.465236] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1255.473282] env[65726]: DEBUG oslo_vmware.api [None req-2033a75e-05f6-474d-9736-3bf92577a37b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Task: {'id': task-5117027, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1255.473920] env[65726]: DEBUG nova.compute.manager [None req-a92da0c2-69e1-4c8b-a9b6-c6ef25fbbe48 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: b353b5e9-500c-42d3-a87f-880a624febec] Start destroying the instance on the hypervisor. {{(pid=65726) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1255.474225] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-a92da0c2-69e1-4c8b-a9b6-c6ef25fbbe48 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: b353b5e9-500c-42d3-a87f-880a624febec] Destroying instance {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1255.475043] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53ecec85-3e52-4887-ba12-d0f194e9696a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1255.482275] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-a92da0c2-69e1-4c8b-a9b6-c6ef25fbbe48 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: b353b5e9-500c-42d3-a87f-880a624febec] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1255.482567] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ea12924f-4985-4521-a310-58e182a2506d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1255.489998] env[65726]: DEBUG oslo_vmware.api [None req-a92da0c2-69e1-4c8b-a9b6-c6ef25fbbe48 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Waiting for the task: (returnval){ [ 1255.489998] env[65726]: value = "task-5117028" [ 1255.489998] env[65726]: _type = "Task" [ 1255.489998] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1255.498744] env[65726]: DEBUG oslo_vmware.api [None req-a92da0c2-69e1-4c8b-a9b6-c6ef25fbbe48 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Task: {'id': task-5117028, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1255.965727] env[65726]: DEBUG oslo_vmware.api [None req-19eeb2b5-f166-426c-957f-e5a1c2480cc2 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': task-5117026, 'name': ReconfigVM_Task, 'duration_secs': 0.439096} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1255.966187] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-19eeb2b5-f166-426c-957f-e5a1c2480cc2 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: e9e93672-d509-45b3-89fe-f1c6ff4d8e78] Reconfigured VM instance instance-00000075 to attach disk [datastore1] e9e93672-d509-45b3-89fe-f1c6ff4d8e78/e9e93672-d509-45b3-89fe-f1c6ff4d8e78.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1255.969791] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c8dbc024-f4cf-4332-b076-d9c1d1d598e6 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1255.976706] env[65726]: DEBUG oslo_vmware.api [None req-2033a75e-05f6-474d-9736-3bf92577a37b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Task: {'id': task-5117027, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1255.978029] env[65726]: DEBUG oslo_vmware.api [None req-19eeb2b5-f166-426c-957f-e5a1c2480cc2 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Waiting for the task: (returnval){ [ 1255.978029] env[65726]: value = "task-5117029" [ 1255.978029] env[65726]: _type = "Task" [ 1255.978029] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1255.987478] env[65726]: DEBUG oslo_vmware.api [None req-19eeb2b5-f166-426c-957f-e5a1c2480cc2 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': task-5117029, 'name': Rename_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1256.000886] env[65726]: DEBUG oslo_vmware.api [None req-a92da0c2-69e1-4c8b-a9b6-c6ef25fbbe48 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Task: {'id': task-5117028, 'name': PowerOffVM_Task, 'duration_secs': 0.230248} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1256.001230] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-a92da0c2-69e1-4c8b-a9b6-c6ef25fbbe48 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: b353b5e9-500c-42d3-a87f-880a624febec] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1256.001488] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-a92da0c2-69e1-4c8b-a9b6-c6ef25fbbe48 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: b353b5e9-500c-42d3-a87f-880a624febec] Unregistering the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1256.001940] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-116748d1-92bd-4090-953b-678f860d26eb {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1256.072690] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-a92da0c2-69e1-4c8b-a9b6-c6ef25fbbe48 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: b353b5e9-500c-42d3-a87f-880a624febec] Unregistered the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1256.072974] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-a92da0c2-69e1-4c8b-a9b6-c6ef25fbbe48 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: b353b5e9-500c-42d3-a87f-880a624febec] Deleting contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1256.073225] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-a92da0c2-69e1-4c8b-a9b6-c6ef25fbbe48 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Deleting the datastore file [datastore1] b353b5e9-500c-42d3-a87f-880a624febec {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1256.073547] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-cf1a44df-1f77-4b34-b63f-d2d94562eab6 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1256.081801] env[65726]: DEBUG oslo_vmware.api [None req-a92da0c2-69e1-4c8b-a9b6-c6ef25fbbe48 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Waiting for the task: (returnval){ [ 1256.081801] env[65726]: value = "task-5117031" [ 1256.081801] env[65726]: _type = "Task" [ 1256.081801] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1256.092278] env[65726]: DEBUG oslo_vmware.api [None req-a92da0c2-69e1-4c8b-a9b6-c6ef25fbbe48 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Task: {'id': task-5117031, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1256.475154] env[65726]: DEBUG oslo_vmware.api [None req-2033a75e-05f6-474d-9736-3bf92577a37b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Task: {'id': task-5117027, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.641215} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1256.475591] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-2033a75e-05f6-474d-9736-3bf92577a37b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 1d141c86-6da9-41b9-a078-50ecf0236f85] Extended root virtual disk {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1256.476216] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf295626-71ac-45be-9459-f4cee34b97f2 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1256.501456] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-2033a75e-05f6-474d-9736-3bf92577a37b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 1d141c86-6da9-41b9-a078-50ecf0236f85] Reconfiguring VM instance instance-00000076 to attach disk [datastore1] 1d141c86-6da9-41b9-a078-50ecf0236f85/1d141c86-6da9-41b9-a078-50ecf0236f85.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1256.504557] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c9a3e315-eeff-4056-8813-05b9173c7ade {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1256.519871] env[65726]: DEBUG oslo_vmware.api [None req-19eeb2b5-f166-426c-957f-e5a1c2480cc2 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': task-5117029, 'name': Rename_Task, 'duration_secs': 0.194924} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1256.520177] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-19eeb2b5-f166-426c-957f-e5a1c2480cc2 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: e9e93672-d509-45b3-89fe-f1c6ff4d8e78] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1256.520847] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-44eb5601-3c5c-46cf-bda4-60e6e8e77036 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1256.527353] env[65726]: DEBUG oslo_vmware.api [None req-2033a75e-05f6-474d-9736-3bf92577a37b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Waiting for the task: (returnval){ [ 1256.527353] env[65726]: value = "task-5117032" [ 1256.527353] env[65726]: _type = "Task" [ 1256.527353] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1256.528732] env[65726]: DEBUG oslo_vmware.api [None req-19eeb2b5-f166-426c-957f-e5a1c2480cc2 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Waiting for the task: (returnval){ [ 1256.528732] env[65726]: value = "task-5117033" [ 1256.528732] env[65726]: _type = "Task" [ 1256.528732] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1256.540668] env[65726]: DEBUG oslo_vmware.api [None req-19eeb2b5-f166-426c-957f-e5a1c2480cc2 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': task-5117033, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1256.543877] env[65726]: DEBUG oslo_vmware.api [None req-2033a75e-05f6-474d-9736-3bf92577a37b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Task: {'id': task-5117032, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1256.592700] env[65726]: DEBUG oslo_vmware.api [None req-a92da0c2-69e1-4c8b-a9b6-c6ef25fbbe48 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Task: {'id': task-5117031, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.234165} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1256.592992] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-a92da0c2-69e1-4c8b-a9b6-c6ef25fbbe48 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Deleted the datastore file {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1256.593222] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-a92da0c2-69e1-4c8b-a9b6-c6ef25fbbe48 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: b353b5e9-500c-42d3-a87f-880a624febec] Deleted contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1256.593411] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-a92da0c2-69e1-4c8b-a9b6-c6ef25fbbe48 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: b353b5e9-500c-42d3-a87f-880a624febec] Instance destroyed {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1256.593584] env[65726]: INFO nova.compute.manager [None req-a92da0c2-69e1-4c8b-a9b6-c6ef25fbbe48 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] [instance: b353b5e9-500c-42d3-a87f-880a624febec] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1256.593877] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-a92da0c2-69e1-4c8b-a9b6-c6ef25fbbe48 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1256.594104] env[65726]: DEBUG nova.compute.manager [-] [instance: b353b5e9-500c-42d3-a87f-880a624febec] Deallocating network for instance {{(pid=65726) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1256.594229] env[65726]: DEBUG nova.network.neutron [-] [instance: b353b5e9-500c-42d3-a87f-880a624febec] deallocate_for_instance() {{(pid=65726) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 1256.594936] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1256.595319] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1256.635363] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1256.635631] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1257.042261] env[65726]: DEBUG oslo_vmware.api [None req-2033a75e-05f6-474d-9736-3bf92577a37b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Task: {'id': task-5117032, 'name': ReconfigVM_Task, 'duration_secs': 0.28958} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1257.045401] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-2033a75e-05f6-474d-9736-3bf92577a37b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 1d141c86-6da9-41b9-a078-50ecf0236f85] Reconfigured VM instance instance-00000076 to attach disk [datastore1] 1d141c86-6da9-41b9-a078-50ecf0236f85/1d141c86-6da9-41b9-a078-50ecf0236f85.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1257.046202] env[65726]: DEBUG oslo_vmware.api [None req-19eeb2b5-f166-426c-957f-e5a1c2480cc2 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': task-5117033, 'name': PowerOnVM_Task, 'duration_secs': 0.507929} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1257.046417] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-af3961ac-d362-4578-9936-38a94f1d039e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1257.048509] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-19eeb2b5-f166-426c-957f-e5a1c2480cc2 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: e9e93672-d509-45b3-89fe-f1c6ff4d8e78] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1257.048509] env[65726]: DEBUG nova.compute.manager [None req-19eeb2b5-f166-426c-957f-e5a1c2480cc2 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: e9e93672-d509-45b3-89fe-f1c6ff4d8e78] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1257.049074] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f3531c6-cafe-4cdb-b690-b63739d41f6a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1257.062899] env[65726]: DEBUG oslo_vmware.api [None req-2033a75e-05f6-474d-9736-3bf92577a37b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Waiting for the task: (returnval){ [ 1257.062899] env[65726]: value = "task-5117034" [ 1257.062899] env[65726]: _type = "Task" [ 1257.062899] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1257.073258] env[65726]: DEBUG oslo_vmware.api [None req-2033a75e-05f6-474d-9736-3bf92577a37b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Task: {'id': task-5117034, 'name': Rename_Task} progress is 6%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1257.085774] env[65726]: DEBUG nova.compute.manager [req-e9846bbd-724a-49f0-b6f3-2110e87bed00 req-e05277fb-d52d-4a66-a543-e6c937ea9772 service nova] [instance: b353b5e9-500c-42d3-a87f-880a624febec] Received event network-vif-deleted-08b7016b-dc44-4206-a00c-5da943b82a38 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1257.085774] env[65726]: INFO nova.compute.manager [req-e9846bbd-724a-49f0-b6f3-2110e87bed00 req-e05277fb-d52d-4a66-a543-e6c937ea9772 service nova] [instance: b353b5e9-500c-42d3-a87f-880a624febec] Neutron deleted interface 08b7016b-dc44-4206-a00c-5da943b82a38; detaching it from the instance and deleting it from the info cache [ 1257.086039] env[65726]: DEBUG nova.network.neutron [req-e9846bbd-724a-49f0-b6f3-2110e87bed00 req-e05277fb-d52d-4a66-a543-e6c937ea9772 service nova] [instance: b353b5e9-500c-42d3-a87f-880a624febec] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1257.566067] env[65726]: INFO nova.compute.manager [None req-19eeb2b5-f166-426c-957f-e5a1c2480cc2 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: e9e93672-d509-45b3-89fe-f1c6ff4d8e78] bringing vm to original state: 'stopped' [ 1257.571020] env[65726]: DEBUG nova.network.neutron [-] [instance: b353b5e9-500c-42d3-a87f-880a624febec] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1257.581740] env[65726]: DEBUG oslo_vmware.api [None req-2033a75e-05f6-474d-9736-3bf92577a37b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Task: {'id': task-5117034, 'name': Rename_Task, 'duration_secs': 0.164651} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1257.582955] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-2033a75e-05f6-474d-9736-3bf92577a37b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 1d141c86-6da9-41b9-a078-50ecf0236f85] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1257.583327] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3718553d-1fab-438d-815a-2d3f5ff7ad44 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1257.592591] env[65726]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d1af634e-b600-4f32-a86e-429d5dc37efb {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1257.595046] env[65726]: DEBUG oslo_vmware.api [None req-2033a75e-05f6-474d-9736-3bf92577a37b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Waiting for the task: (returnval){ [ 1257.595046] env[65726]: value = "task-5117035" [ 1257.595046] env[65726]: _type = "Task" [ 1257.595046] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1257.604829] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8c77e02-5217-408f-88ae-0a0dfd995b23 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1257.620412] env[65726]: DEBUG oslo_vmware.api [None req-2033a75e-05f6-474d-9736-3bf92577a37b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Task: {'id': task-5117035, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1257.641615] env[65726]: DEBUG nova.compute.manager [req-e9846bbd-724a-49f0-b6f3-2110e87bed00 req-e05277fb-d52d-4a66-a543-e6c937ea9772 service nova] [instance: b353b5e9-500c-42d3-a87f-880a624febec] Detach interface failed, port_id=08b7016b-dc44-4206-a00c-5da943b82a38, reason: Instance b353b5e9-500c-42d3-a87f-880a624febec could not be found. {{(pid=65726) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11669}} [ 1258.074240] env[65726]: INFO nova.compute.manager [-] [instance: b353b5e9-500c-42d3-a87f-880a624febec] Took 1.48 seconds to deallocate network for instance. [ 1258.108391] env[65726]: DEBUG oslo_vmware.api [None req-2033a75e-05f6-474d-9736-3bf92577a37b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Task: {'id': task-5117035, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1258.573523] env[65726]: DEBUG oslo_concurrency.lockutils [None req-19eeb2b5-f166-426c-957f-e5a1c2480cc2 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Acquiring lock "e9e93672-d509-45b3-89fe-f1c6ff4d8e78" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1258.573888] env[65726]: DEBUG oslo_concurrency.lockutils [None req-19eeb2b5-f166-426c-957f-e5a1c2480cc2 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Lock "e9e93672-d509-45b3-89fe-f1c6ff4d8e78" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1258.574057] env[65726]: DEBUG nova.compute.manager [None req-19eeb2b5-f166-426c-957f-e5a1c2480cc2 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: e9e93672-d509-45b3-89fe-f1c6ff4d8e78] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1258.574973] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a2c6a19-f2cf-4774-adac-2ec195482994 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1258.580624] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a92da0c2-69e1-4c8b-a9b6-c6ef25fbbe48 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1258.580894] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a92da0c2-69e1-4c8b-a9b6-c6ef25fbbe48 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1258.581129] env[65726]: DEBUG nova.objects.instance [None req-a92da0c2-69e1-4c8b-a9b6-c6ef25fbbe48 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Lazy-loading 'resources' on Instance uuid b353b5e9-500c-42d3-a87f-880a624febec {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1258.584730] env[65726]: DEBUG nova.compute.manager [None req-19eeb2b5-f166-426c-957f-e5a1c2480cc2 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: e9e93672-d509-45b3-89fe-f1c6ff4d8e78] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=65726) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3459}} [ 1258.606713] env[65726]: DEBUG oslo_vmware.api [None req-2033a75e-05f6-474d-9736-3bf92577a37b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Task: {'id': task-5117035, 'name': PowerOnVM_Task, 'duration_secs': 0.5556} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1258.607888] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-2033a75e-05f6-474d-9736-3bf92577a37b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 1d141c86-6da9-41b9-a078-50ecf0236f85] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1258.607888] env[65726]: INFO nova.compute.manager [None req-2033a75e-05f6-474d-9736-3bf92577a37b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 1d141c86-6da9-41b9-a078-50ecf0236f85] Took 8.73 seconds to spawn the instance on the hypervisor. [ 1258.607888] env[65726]: DEBUG nova.compute.manager [None req-2033a75e-05f6-474d-9736-3bf92577a37b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 1d141c86-6da9-41b9-a078-50ecf0236f85] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1258.608268] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e6531ed-c1c5-47ee-862b-25321d53726c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1259.089685] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-19eeb2b5-f166-426c-957f-e5a1c2480cc2 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: e9e93672-d509-45b3-89fe-f1c6ff4d8e78] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1259.089685] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f36c6b04-34a0-4fbb-9207-50a7eb0d6d99 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1259.098040] env[65726]: DEBUG oslo_vmware.api [None req-19eeb2b5-f166-426c-957f-e5a1c2480cc2 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Waiting for the task: (returnval){ [ 1259.098040] env[65726]: value = "task-5117036" [ 1259.098040] env[65726]: _type = "Task" [ 1259.098040] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1259.107753] env[65726]: DEBUG oslo_vmware.api [None req-19eeb2b5-f166-426c-957f-e5a1c2480cc2 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': task-5117036, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1259.128073] env[65726]: INFO nova.compute.manager [None req-2033a75e-05f6-474d-9736-3bf92577a37b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 1d141c86-6da9-41b9-a078-50ecf0236f85] Took 13.47 seconds to build instance. [ 1259.204981] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e43ba65-5c7e-4e5a-8e3b-00b996f54882 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1259.215665] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8780d69d-51ba-4868-948a-f8bff09ee51e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1259.247808] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab29bc09-bd30-416a-85c4-26abf548cdac {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1259.258857] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d8c7ebb-2f4c-487c-9acc-256e8c87fe87 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1259.274726] env[65726]: DEBUG nova.compute.provider_tree [None req-a92da0c2-69e1-4c8b-a9b6-c6ef25fbbe48 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1259.608638] env[65726]: DEBUG oslo_vmware.api [None req-19eeb2b5-f166-426c-957f-e5a1c2480cc2 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': task-5117036, 'name': PowerOffVM_Task, 'duration_secs': 0.217282} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1259.608971] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-19eeb2b5-f166-426c-957f-e5a1c2480cc2 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: e9e93672-d509-45b3-89fe-f1c6ff4d8e78] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1259.609148] env[65726]: DEBUG nova.compute.manager [None req-19eeb2b5-f166-426c-957f-e5a1c2480cc2 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: e9e93672-d509-45b3-89fe-f1c6ff4d8e78] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1259.610092] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35b42999-cbd4-4da3-bcbf-592f7c1d77be {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1259.632810] env[65726]: DEBUG oslo_concurrency.lockutils [None req-2033a75e-05f6-474d-9736-3bf92577a37b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Lock "1d141c86-6da9-41b9-a078-50ecf0236f85" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.982s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1259.780025] env[65726]: DEBUG nova.scheduler.client.report [None req-a92da0c2-69e1-4c8b-a9b6-c6ef25fbbe48 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1260.124871] env[65726]: DEBUG oslo_concurrency.lockutils [None req-19eeb2b5-f166-426c-957f-e5a1c2480cc2 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Lock "e9e93672-d509-45b3-89fe-f1c6ff4d8e78" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.550s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1260.280376] env[65726]: DEBUG oslo_concurrency.lockutils [None req-f9267e03-8bcf-4f4e-acb8-f516abe74b2e tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Acquiring lock "1d141c86-6da9-41b9-a078-50ecf0236f85" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1260.281077] env[65726]: DEBUG oslo_concurrency.lockutils [None req-f9267e03-8bcf-4f4e-acb8-f516abe74b2e tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Lock "1d141c86-6da9-41b9-a078-50ecf0236f85" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1260.281226] env[65726]: DEBUG oslo_concurrency.lockutils [None req-f9267e03-8bcf-4f4e-acb8-f516abe74b2e tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Acquiring lock "1d141c86-6da9-41b9-a078-50ecf0236f85-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1260.281409] env[65726]: DEBUG oslo_concurrency.lockutils [None req-f9267e03-8bcf-4f4e-acb8-f516abe74b2e tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Lock "1d141c86-6da9-41b9-a078-50ecf0236f85-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1260.281568] env[65726]: DEBUG oslo_concurrency.lockutils [None req-f9267e03-8bcf-4f4e-acb8-f516abe74b2e tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Lock "1d141c86-6da9-41b9-a078-50ecf0236f85-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1260.284009] env[65726]: INFO nova.compute.manager [None req-f9267e03-8bcf-4f4e-acb8-f516abe74b2e tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 1d141c86-6da9-41b9-a078-50ecf0236f85] Terminating instance [ 1260.285847] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a92da0c2-69e1-4c8b-a9b6-c6ef25fbbe48 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.705s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1260.313064] env[65726]: INFO nova.scheduler.client.report [None req-a92da0c2-69e1-4c8b-a9b6-c6ef25fbbe48 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Deleted allocations for instance b353b5e9-500c-42d3-a87f-880a624febec [ 1260.633792] env[65726]: DEBUG oslo_concurrency.lockutils [None req-19eeb2b5-f166-426c-957f-e5a1c2480cc2 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1260.634139] env[65726]: DEBUG oslo_concurrency.lockutils [None req-19eeb2b5-f166-426c-957f-e5a1c2480cc2 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1260.634139] env[65726]: DEBUG nova.objects.instance [None req-19eeb2b5-f166-426c-957f-e5a1c2480cc2 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: e9e93672-d509-45b3-89fe-f1c6ff4d8e78] Trying to apply a migration context that does not seem to be set for this instance {{(pid=65726) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 1260.790726] env[65726]: DEBUG nova.compute.manager [None req-f9267e03-8bcf-4f4e-acb8-f516abe74b2e tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 1d141c86-6da9-41b9-a078-50ecf0236f85] Start destroying the instance on the hypervisor. {{(pid=65726) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1260.791043] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-f9267e03-8bcf-4f4e-acb8-f516abe74b2e tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 1d141c86-6da9-41b9-a078-50ecf0236f85] Destroying instance {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1260.792077] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ab1dd68-5227-4880-9269-539c893fff66 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1260.801632] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-f9267e03-8bcf-4f4e-acb8-f516abe74b2e tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 1d141c86-6da9-41b9-a078-50ecf0236f85] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1260.802312] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1e02fa39-7987-4832-bf39-3ac8a530fe40 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1260.810007] env[65726]: DEBUG oslo_vmware.api [None req-f9267e03-8bcf-4f4e-acb8-f516abe74b2e tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Waiting for the task: (returnval){ [ 1260.810007] env[65726]: value = "task-5117037" [ 1260.810007] env[65726]: _type = "Task" [ 1260.810007] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1260.823292] env[65726]: DEBUG oslo_vmware.api [None req-f9267e03-8bcf-4f4e-acb8-f516abe74b2e tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Task: {'id': task-5117037, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1260.824415] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a92da0c2-69e1-4c8b-a9b6-c6ef25fbbe48 tempest-AttachVolumeShelveTestJSON-679781638 tempest-AttachVolumeShelveTestJSON-679781638-project-member] Lock "b353b5e9-500c-42d3-a87f-880a624febec" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.858s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1261.320864] env[65726]: DEBUG oslo_vmware.api [None req-f9267e03-8bcf-4f4e-acb8-f516abe74b2e tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Task: {'id': task-5117037, 'name': PowerOffVM_Task, 'duration_secs': 0.204661} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1261.321246] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-f9267e03-8bcf-4f4e-acb8-f516abe74b2e tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 1d141c86-6da9-41b9-a078-50ecf0236f85] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1261.321457] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-f9267e03-8bcf-4f4e-acb8-f516abe74b2e tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 1d141c86-6da9-41b9-a078-50ecf0236f85] Unregistering the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1261.321752] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1b123b54-afc3-46a8-98dd-963c2e7fc9cb {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1261.387394] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-f9267e03-8bcf-4f4e-acb8-f516abe74b2e tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 1d141c86-6da9-41b9-a078-50ecf0236f85] Unregistered the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1261.387607] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-f9267e03-8bcf-4f4e-acb8-f516abe74b2e tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 1d141c86-6da9-41b9-a078-50ecf0236f85] Deleting contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1261.387794] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-f9267e03-8bcf-4f4e-acb8-f516abe74b2e tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Deleting the datastore file [datastore1] 1d141c86-6da9-41b9-a078-50ecf0236f85 {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1261.388096] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-72fc03bb-7814-4d4f-8391-1d6529823302 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1261.395572] env[65726]: DEBUG oslo_vmware.api [None req-f9267e03-8bcf-4f4e-acb8-f516abe74b2e tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Waiting for the task: (returnval){ [ 1261.395572] env[65726]: value = "task-5117039" [ 1261.395572] env[65726]: _type = "Task" [ 1261.395572] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1261.404061] env[65726]: DEBUG oslo_vmware.api [None req-f9267e03-8bcf-4f4e-acb8-f516abe74b2e tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Task: {'id': task-5117039, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1261.613567] env[65726]: DEBUG oslo_concurrency.lockutils [None req-258108f0-ee97-47d2-b9ce-1ad0b8ad8115 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Acquiring lock "e9e93672-d509-45b3-89fe-f1c6ff4d8e78" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1261.613900] env[65726]: DEBUG oslo_concurrency.lockutils [None req-258108f0-ee97-47d2-b9ce-1ad0b8ad8115 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Lock "e9e93672-d509-45b3-89fe-f1c6ff4d8e78" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1261.614156] env[65726]: DEBUG oslo_concurrency.lockutils [None req-258108f0-ee97-47d2-b9ce-1ad0b8ad8115 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Acquiring lock "e9e93672-d509-45b3-89fe-f1c6ff4d8e78-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1261.614346] env[65726]: DEBUG oslo_concurrency.lockutils [None req-258108f0-ee97-47d2-b9ce-1ad0b8ad8115 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Lock "e9e93672-d509-45b3-89fe-f1c6ff4d8e78-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1261.614575] env[65726]: DEBUG oslo_concurrency.lockutils [None req-258108f0-ee97-47d2-b9ce-1ad0b8ad8115 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Lock "e9e93672-d509-45b3-89fe-f1c6ff4d8e78-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1261.616939] env[65726]: INFO nova.compute.manager [None req-258108f0-ee97-47d2-b9ce-1ad0b8ad8115 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: e9e93672-d509-45b3-89fe-f1c6ff4d8e78] Terminating instance [ 1261.643691] env[65726]: DEBUG oslo_concurrency.lockutils [None req-19eeb2b5-f166-426c-957f-e5a1c2480cc2 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.010s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1261.906633] env[65726]: DEBUG oslo_vmware.api [None req-f9267e03-8bcf-4f4e-acb8-f516abe74b2e tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Task: {'id': task-5117039, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.178649} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1261.906633] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-f9267e03-8bcf-4f4e-acb8-f516abe74b2e tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Deleted the datastore file {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1261.906839] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-f9267e03-8bcf-4f4e-acb8-f516abe74b2e tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 1d141c86-6da9-41b9-a078-50ecf0236f85] Deleted contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1261.906969] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-f9267e03-8bcf-4f4e-acb8-f516abe74b2e tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 1d141c86-6da9-41b9-a078-50ecf0236f85] Instance destroyed {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1261.907164] env[65726]: INFO nova.compute.manager [None req-f9267e03-8bcf-4f4e-acb8-f516abe74b2e tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 1d141c86-6da9-41b9-a078-50ecf0236f85] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1261.907410] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-f9267e03-8bcf-4f4e-acb8-f516abe74b2e tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1261.907610] env[65726]: DEBUG nova.compute.manager [-] [instance: 1d141c86-6da9-41b9-a078-50ecf0236f85] Deallocating network for instance {{(pid=65726) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1261.907846] env[65726]: DEBUG nova.network.neutron [-] [instance: 1d141c86-6da9-41b9-a078-50ecf0236f85] deallocate_for_instance() {{(pid=65726) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 1261.908257] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1261.910078] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1261.948961] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1261.949269] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1262.121373] env[65726]: DEBUG nova.compute.manager [None req-258108f0-ee97-47d2-b9ce-1ad0b8ad8115 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: e9e93672-d509-45b3-89fe-f1c6ff4d8e78] Start destroying the instance on the hypervisor. {{(pid=65726) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1262.121642] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-258108f0-ee97-47d2-b9ce-1ad0b8ad8115 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: e9e93672-d509-45b3-89fe-f1c6ff4d8e78] Destroying instance {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1262.123097] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e1c8a9b-5857-4193-aee3-7f22f6569a0a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1262.131960] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-258108f0-ee97-47d2-b9ce-1ad0b8ad8115 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: e9e93672-d509-45b3-89fe-f1c6ff4d8e78] Unregistering the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1262.132313] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0d98031a-7ed1-476f-a798-e7c09b88c440 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1262.204647] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-258108f0-ee97-47d2-b9ce-1ad0b8ad8115 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: e9e93672-d509-45b3-89fe-f1c6ff4d8e78] Unregistered the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1262.204920] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-258108f0-ee97-47d2-b9ce-1ad0b8ad8115 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: e9e93672-d509-45b3-89fe-f1c6ff4d8e78] Deleting contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1262.205117] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-258108f0-ee97-47d2-b9ce-1ad0b8ad8115 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Deleting the datastore file [datastore1] e9e93672-d509-45b3-89fe-f1c6ff4d8e78 {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1262.205402] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-82a511c1-67af-4e10-b7ec-3a4915cb02d4 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1262.209506] env[65726]: DEBUG nova.compute.manager [req-ce058993-2797-4d61-8018-6abb0f317f4b req-30000063-3ef0-459e-a845-9cec0c8f80bb service nova] [instance: 1d141c86-6da9-41b9-a078-50ecf0236f85] Received event network-vif-deleted-d7088b35-54c6-48ac-8244-46071889261e {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1262.209666] env[65726]: INFO nova.compute.manager [req-ce058993-2797-4d61-8018-6abb0f317f4b req-30000063-3ef0-459e-a845-9cec0c8f80bb service nova] [instance: 1d141c86-6da9-41b9-a078-50ecf0236f85] Neutron deleted interface d7088b35-54c6-48ac-8244-46071889261e; detaching it from the instance and deleting it from the info cache [ 1262.209839] env[65726]: DEBUG nova.network.neutron [req-ce058993-2797-4d61-8018-6abb0f317f4b req-30000063-3ef0-459e-a845-9cec0c8f80bb service nova] [instance: 1d141c86-6da9-41b9-a078-50ecf0236f85] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1262.217775] env[65726]: DEBUG oslo_vmware.api [None req-258108f0-ee97-47d2-b9ce-1ad0b8ad8115 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Waiting for the task: (returnval){ [ 1262.217775] env[65726]: value = "task-5117042" [ 1262.217775] env[65726]: _type = "Task" [ 1262.217775] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1262.227382] env[65726]: DEBUG oslo_vmware.api [None req-258108f0-ee97-47d2-b9ce-1ad0b8ad8115 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': task-5117042, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1262.684923] env[65726]: DEBUG nova.network.neutron [-] [instance: 1d141c86-6da9-41b9-a078-50ecf0236f85] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1262.714015] env[65726]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a84fee64-426e-4a50-8abe-b76ca6a80181 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1262.725933] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11f9a23b-3b22-4887-9c64-16f07c885f57 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1262.742607] env[65726]: DEBUG oslo_vmware.api [None req-258108f0-ee97-47d2-b9ce-1ad0b8ad8115 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': task-5117042, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.185131} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1262.742919] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-258108f0-ee97-47d2-b9ce-1ad0b8ad8115 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Deleted the datastore file {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1262.743146] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-258108f0-ee97-47d2-b9ce-1ad0b8ad8115 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: e9e93672-d509-45b3-89fe-f1c6ff4d8e78] Deleted contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1262.743346] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-258108f0-ee97-47d2-b9ce-1ad0b8ad8115 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: e9e93672-d509-45b3-89fe-f1c6ff4d8e78] Instance destroyed {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1262.743517] env[65726]: INFO nova.compute.manager [None req-258108f0-ee97-47d2-b9ce-1ad0b8ad8115 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: e9e93672-d509-45b3-89fe-f1c6ff4d8e78] Took 0.62 seconds to destroy the instance on the hypervisor. [ 1262.744270] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-258108f0-ee97-47d2-b9ce-1ad0b8ad8115 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1262.744270] env[65726]: DEBUG nova.compute.manager [-] [instance: e9e93672-d509-45b3-89fe-f1c6ff4d8e78] Deallocating network for instance {{(pid=65726) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1262.744270] env[65726]: DEBUG nova.network.neutron [-] [instance: e9e93672-d509-45b3-89fe-f1c6ff4d8e78] deallocate_for_instance() {{(pid=65726) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 1262.744701] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1262.745050] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1262.760028] env[65726]: DEBUG nova.compute.manager [req-ce058993-2797-4d61-8018-6abb0f317f4b req-30000063-3ef0-459e-a845-9cec0c8f80bb service nova] [instance: 1d141c86-6da9-41b9-a078-50ecf0236f85] Detach interface failed, port_id=d7088b35-54c6-48ac-8244-46071889261e, reason: Instance 1d141c86-6da9-41b9-a078-50ecf0236f85 could not be found. {{(pid=65726) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11669}} [ 1262.780955] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1262.781206] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1263.187304] env[65726]: INFO nova.compute.manager [-] [instance: 1d141c86-6da9-41b9-a078-50ecf0236f85] Took 1.28 seconds to deallocate network for instance. [ 1263.487172] env[65726]: DEBUG nova.network.neutron [-] [instance: e9e93672-d509-45b3-89fe-f1c6ff4d8e78] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1263.695217] env[65726]: DEBUG oslo_concurrency.lockutils [None req-f9267e03-8bcf-4f4e-acb8-f516abe74b2e tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1263.695531] env[65726]: DEBUG oslo_concurrency.lockutils [None req-f9267e03-8bcf-4f4e-acb8-f516abe74b2e tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1263.695712] env[65726]: DEBUG nova.objects.instance [None req-f9267e03-8bcf-4f4e-acb8-f516abe74b2e tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Lazy-loading 'resources' on Instance uuid 1d141c86-6da9-41b9-a078-50ecf0236f85 {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1263.989830] env[65726]: INFO nova.compute.manager [-] [instance: e9e93672-d509-45b3-89fe-f1c6ff4d8e78] Took 1.25 seconds to deallocate network for instance. [ 1264.301370] env[65726]: DEBUG nova.compute.manager [req-cf69fdf9-fb66-440d-8125-56fbbd9a4b30 req-fa94d563-f9a8-463e-a2e4-058f6104d94d service nova] [instance: e9e93672-d509-45b3-89fe-f1c6ff4d8e78] Received event network-vif-deleted-bc612ce3-dca5-4716-9f5c-60a6247e4b42 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1264.364230] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5fd6ac53-02ad-4944-9fe4-6ddf3feae805 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1264.372148] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d67356ca-aded-4bf3-8770-989107d7a18f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1264.402816] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2df96edf-d925-4ee6-b720-df81cdcd6120 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1264.411425] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a28a0e1-0346-48fc-9cac-cfb6a2adaff1 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1264.427013] env[65726]: DEBUG nova.compute.provider_tree [None req-f9267e03-8bcf-4f4e-acb8-f516abe74b2e tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1264.496372] env[65726]: DEBUG oslo_concurrency.lockutils [None req-258108f0-ee97-47d2-b9ce-1ad0b8ad8115 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1264.930840] env[65726]: DEBUG nova.scheduler.client.report [None req-f9267e03-8bcf-4f4e-acb8-f516abe74b2e tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1265.436486] env[65726]: DEBUG oslo_concurrency.lockutils [None req-f9267e03-8bcf-4f4e-acb8-f516abe74b2e tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.741s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1265.439878] env[65726]: DEBUG oslo_concurrency.lockutils [None req-258108f0-ee97-47d2-b9ce-1ad0b8ad8115 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.943s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1265.440031] env[65726]: DEBUG nova.objects.instance [None req-258108f0-ee97-47d2-b9ce-1ad0b8ad8115 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Lazy-loading 'resources' on Instance uuid e9e93672-d509-45b3-89fe-f1c6ff4d8e78 {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1265.460196] env[65726]: INFO nova.scheduler.client.report [None req-f9267e03-8bcf-4f4e-acb8-f516abe74b2e tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Deleted allocations for instance 1d141c86-6da9-41b9-a078-50ecf0236f85 [ 1265.969788] env[65726]: DEBUG oslo_concurrency.lockutils [None req-f9267e03-8bcf-4f4e-acb8-f516abe74b2e tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Lock "1d141c86-6da9-41b9-a078-50ecf0236f85" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.689s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1266.038962] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a932aa47-41e0-450c-81f3-1ecd6f7204c4 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1266.047568] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-730d1c9a-71bd-4f0d-9402-fbc401e99113 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1266.078426] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41ba754b-b445-4402-b2e6-36a21faf4b03 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1266.088029] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f69da065-6f4f-43d4-be92-34e9c87743fe {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1266.102439] env[65726]: DEBUG nova.compute.provider_tree [None req-258108f0-ee97-47d2-b9ce-1ad0b8ad8115 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1266.421674] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e0c56071-ba98-4289-82f5-a0b33d02c9d6 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Acquiring lock "e3255f3b-028f-4a0b-b621-873730417936" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1266.421836] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e0c56071-ba98-4289-82f5-a0b33d02c9d6 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Lock "e3255f3b-028f-4a0b-b621-873730417936" acquired by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1266.422243] env[65726]: INFO nova.compute.manager [None req-e0c56071-ba98-4289-82f5-a0b33d02c9d6 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: e3255f3b-028f-4a0b-b621-873730417936] Rebooting instance [ 1266.497929] env[65726]: DEBUG oslo_service.periodic_task [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=65726) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1266.498634] env[65726]: DEBUG oslo_service.periodic_task [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=65726) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1266.498908] env[65726]: DEBUG oslo_service.periodic_task [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=65726) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1266.498971] env[65726]: DEBUG oslo_service.periodic_task [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=65726) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1266.499116] env[65726]: DEBUG oslo_service.periodic_task [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=65726) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1266.499256] env[65726]: DEBUG oslo_service.periodic_task [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=65726) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1266.499560] env[65726]: DEBUG oslo_service.periodic_task [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=65726) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1266.499624] env[65726]: DEBUG nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=65726) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11251}} [ 1266.606259] env[65726]: DEBUG nova.scheduler.client.report [None req-258108f0-ee97-47d2-b9ce-1ad0b8ad8115 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1266.939108] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e0c56071-ba98-4289-82f5-a0b33d02c9d6 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Acquiring lock "refresh_cache-e3255f3b-028f-4a0b-b621-873730417936" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1266.939314] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e0c56071-ba98-4289-82f5-a0b33d02c9d6 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Acquired lock "refresh_cache-e3255f3b-028f-4a0b-b621-873730417936" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1266.939549] env[65726]: DEBUG nova.network.neutron [None req-e0c56071-ba98-4289-82f5-a0b33d02c9d6 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: e3255f3b-028f-4a0b-b621-873730417936] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1267.062616] env[65726]: DEBUG oslo_concurrency.lockutils [None req-11154a5d-f93f-44ba-9a74-59c15507518c tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Acquiring lock "a4c229bb-e5f9-41d3-86f9-ca4cfb087558" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1267.062954] env[65726]: DEBUG oslo_concurrency.lockutils [None req-11154a5d-f93f-44ba-9a74-59c15507518c tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Lock "a4c229bb-e5f9-41d3-86f9-ca4cfb087558" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1267.063104] env[65726]: DEBUG oslo_concurrency.lockutils [None req-11154a5d-f93f-44ba-9a74-59c15507518c tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Acquiring lock "a4c229bb-e5f9-41d3-86f9-ca4cfb087558-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1267.063284] env[65726]: DEBUG oslo_concurrency.lockutils [None req-11154a5d-f93f-44ba-9a74-59c15507518c tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Lock "a4c229bb-e5f9-41d3-86f9-ca4cfb087558-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1267.063451] env[65726]: DEBUG oslo_concurrency.lockutils [None req-11154a5d-f93f-44ba-9a74-59c15507518c tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Lock "a4c229bb-e5f9-41d3-86f9-ca4cfb087558-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1267.065712] env[65726]: INFO nova.compute.manager [None req-11154a5d-f93f-44ba-9a74-59c15507518c tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: a4c229bb-e5f9-41d3-86f9-ca4cfb087558] Terminating instance [ 1267.117051] env[65726]: DEBUG oslo_concurrency.lockutils [None req-258108f0-ee97-47d2-b9ce-1ad0b8ad8115 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.677s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1267.139341] env[65726]: INFO nova.scheduler.client.report [None req-258108f0-ee97-47d2-b9ce-1ad0b8ad8115 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Deleted allocations for instance e9e93672-d509-45b3-89fe-f1c6ff4d8e78 [ 1267.442707] env[65726]: WARNING openstack [None req-e0c56071-ba98-4289-82f5-a0b33d02c9d6 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1267.443070] env[65726]: WARNING openstack [None req-e0c56071-ba98-4289-82f5-a0b33d02c9d6 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1267.569355] env[65726]: DEBUG nova.compute.manager [None req-11154a5d-f93f-44ba-9a74-59c15507518c tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: a4c229bb-e5f9-41d3-86f9-ca4cfb087558] Start destroying the instance on the hypervisor. {{(pid=65726) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1267.569581] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-11154a5d-f93f-44ba-9a74-59c15507518c tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: a4c229bb-e5f9-41d3-86f9-ca4cfb087558] Destroying instance {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1267.570529] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4cbd2699-1874-43c8-b2a4-36968c81090c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1267.578660] env[65726]: WARNING openstack [None req-e0c56071-ba98-4289-82f5-a0b33d02c9d6 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1267.579028] env[65726]: WARNING openstack [None req-e0c56071-ba98-4289-82f5-a0b33d02c9d6 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1267.588094] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-11154a5d-f93f-44ba-9a74-59c15507518c tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: a4c229bb-e5f9-41d3-86f9-ca4cfb087558] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1267.588371] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a86cd0bc-6de1-43d9-9afa-f37e81a127bf {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1267.596173] env[65726]: DEBUG oslo_vmware.api [None req-11154a5d-f93f-44ba-9a74-59c15507518c tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Waiting for the task: (returnval){ [ 1267.596173] env[65726]: value = "task-5117043" [ 1267.596173] env[65726]: _type = "Task" [ 1267.596173] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1267.607166] env[65726]: DEBUG oslo_vmware.api [None req-11154a5d-f93f-44ba-9a74-59c15507518c tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Task: {'id': task-5117043, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1267.649051] env[65726]: DEBUG oslo_concurrency.lockutils [None req-258108f0-ee97-47d2-b9ce-1ad0b8ad8115 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Lock "e9e93672-d509-45b3-89fe-f1c6ff4d8e78" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.035s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1267.660933] env[65726]: WARNING openstack [None req-e0c56071-ba98-4289-82f5-a0b33d02c9d6 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1267.661362] env[65726]: WARNING openstack [None req-e0c56071-ba98-4289-82f5-a0b33d02c9d6 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1267.750814] env[65726]: DEBUG nova.network.neutron [None req-e0c56071-ba98-4289-82f5-a0b33d02c9d6 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: e3255f3b-028f-4a0b-b621-873730417936] Updating instance_info_cache with network_info: [{"id": "03c9e6fb-2435-43be-bb55-8afecdcf38ff", "address": "fa:16:3e:d2:0f:96", "network": {"id": "a899453d-5f43-4650-9642-b33f54a4d2eb", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-614861139-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.231", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "acd800e0438940bc80ae3c6b672db4ee", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "205fb402-8eaf-4b61-8f57-8f216024179a", "external-id": "nsx-vlan-transportzone-78", "segmentation_id": 78, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap03c9e6fb-24", "ovs_interfaceid": "03c9e6fb-2435-43be-bb55-8afecdcf38ff", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1268.107252] env[65726]: DEBUG oslo_vmware.api [None req-11154a5d-f93f-44ba-9a74-59c15507518c tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Task: {'id': task-5117043, 'name': PowerOffVM_Task, 'duration_secs': 0.205503} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1268.107625] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-11154a5d-f93f-44ba-9a74-59c15507518c tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: a4c229bb-e5f9-41d3-86f9-ca4cfb087558] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1268.107700] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-11154a5d-f93f-44ba-9a74-59c15507518c tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: a4c229bb-e5f9-41d3-86f9-ca4cfb087558] Unregistering the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1268.107966] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-89023a58-e608-4238-9892-ada7491474f3 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1268.189156] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-11154a5d-f93f-44ba-9a74-59c15507518c tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: a4c229bb-e5f9-41d3-86f9-ca4cfb087558] Unregistered the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1268.189334] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-11154a5d-f93f-44ba-9a74-59c15507518c tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: a4c229bb-e5f9-41d3-86f9-ca4cfb087558] Deleting contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1268.189597] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-11154a5d-f93f-44ba-9a74-59c15507518c tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Deleting the datastore file [datastore1] a4c229bb-e5f9-41d3-86f9-ca4cfb087558 {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1268.189896] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-daa82481-d362-4c10-8c84-f9f38080d8f6 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1268.197093] env[65726]: DEBUG oslo_vmware.api [None req-11154a5d-f93f-44ba-9a74-59c15507518c tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Waiting for the task: (returnval){ [ 1268.197093] env[65726]: value = "task-5117045" [ 1268.197093] env[65726]: _type = "Task" [ 1268.197093] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1268.205742] env[65726]: DEBUG oslo_vmware.api [None req-11154a5d-f93f-44ba-9a74-59c15507518c tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Task: {'id': task-5117045, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1268.254666] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e0c56071-ba98-4289-82f5-a0b33d02c9d6 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Releasing lock "refresh_cache-e3255f3b-028f-4a0b-b621-873730417936" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1268.569268] env[65726]: DEBUG oslo_service.periodic_task [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Running periodic task ComputeManager.update_available_resource {{(pid=65726) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1268.712458] env[65726]: DEBUG oslo_vmware.api [None req-11154a5d-f93f-44ba-9a74-59c15507518c tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Task: {'id': task-5117045, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.226303} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1268.712802] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-11154a5d-f93f-44ba-9a74-59c15507518c tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Deleted the datastore file {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1268.712933] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-11154a5d-f93f-44ba-9a74-59c15507518c tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: a4c229bb-e5f9-41d3-86f9-ca4cfb087558] Deleted contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1268.713117] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-11154a5d-f93f-44ba-9a74-59c15507518c tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: a4c229bb-e5f9-41d3-86f9-ca4cfb087558] Instance destroyed {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1268.713288] env[65726]: INFO nova.compute.manager [None req-11154a5d-f93f-44ba-9a74-59c15507518c tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: a4c229bb-e5f9-41d3-86f9-ca4cfb087558] Took 1.14 seconds to destroy the instance on the hypervisor. [ 1268.713545] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-11154a5d-f93f-44ba-9a74-59c15507518c tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1268.713748] env[65726]: DEBUG nova.compute.manager [-] [instance: a4c229bb-e5f9-41d3-86f9-ca4cfb087558] Deallocating network for instance {{(pid=65726) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1268.713853] env[65726]: DEBUG nova.network.neutron [-] [instance: a4c229bb-e5f9-41d3-86f9-ca4cfb087558] deallocate_for_instance() {{(pid=65726) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 1268.714405] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1268.714705] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1268.753848] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1268.754171] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1268.764989] env[65726]: DEBUG nova.compute.manager [None req-e0c56071-ba98-4289-82f5-a0b33d02c9d6 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: e3255f3b-028f-4a0b-b621-873730417936] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1268.765936] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc352307-6b4b-45a9-a816-af50fd3c7924 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1269.011249] env[65726]: DEBUG nova.compute.manager [req-7f671e66-63e6-403a-9c37-396f3890386b req-763d312c-ed4f-4697-854c-3877e155d323 service nova] [instance: a4c229bb-e5f9-41d3-86f9-ca4cfb087558] Received event network-vif-deleted-016b56b6-52e2-48c7-9a15-c91dda1dee97 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1269.011924] env[65726]: INFO nova.compute.manager [req-7f671e66-63e6-403a-9c37-396f3890386b req-763d312c-ed4f-4697-854c-3877e155d323 service nova] [instance: a4c229bb-e5f9-41d3-86f9-ca4cfb087558] Neutron deleted interface 016b56b6-52e2-48c7-9a15-c91dda1dee97; detaching it from the instance and deleting it from the info cache [ 1269.012759] env[65726]: DEBUG nova.network.neutron [req-7f671e66-63e6-403a-9c37-396f3890386b req-763d312c-ed4f-4697-854c-3877e155d323 service nova] [instance: a4c229bb-e5f9-41d3-86f9-ca4cfb087558] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1269.075040] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1269.075040] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1269.075040] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1269.075040] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=65726) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:937}} [ 1269.075040] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d631e4f5-08a3-439c-b7ec-9e1d1734d0bc {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1269.085245] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44dcc01f-cb7b-4d51-b16b-e22a88f9f39c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1269.102336] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-602361c1-2db6-4d1d-a000-0233ae216b15 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1269.110551] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56389137-5f29-494f-906b-d710164d7009 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1269.141080] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=179620MB free_disk=95GB free_vcpus=48 pci_devices=None {{(pid=65726) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1136}} [ 1269.141234] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1269.141450] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1269.488669] env[65726]: DEBUG nova.network.neutron [-] [instance: a4c229bb-e5f9-41d3-86f9-ca4cfb087558] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1269.516622] env[65726]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d02fd686-b7b0-4017-883d-9ca16a49502e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1269.526527] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-998d7a62-2bf0-4b57-973e-08e4d3ac6355 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1269.556837] env[65726]: DEBUG nova.compute.manager [req-7f671e66-63e6-403a-9c37-396f3890386b req-763d312c-ed4f-4697-854c-3877e155d323 service nova] [instance: a4c229bb-e5f9-41d3-86f9-ca4cfb087558] Detach interface failed, port_id=016b56b6-52e2-48c7-9a15-c91dda1dee97, reason: Instance a4c229bb-e5f9-41d3-86f9-ca4cfb087558 could not be found. {{(pid=65726) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11669}} [ 1269.785811] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-744f64b6-65f1-4925-aff7-ab991ead41ba {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1269.794214] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-e0c56071-ba98-4289-82f5-a0b33d02c9d6 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: e3255f3b-028f-4a0b-b621-873730417936] Doing hard reboot of VM {{(pid=65726) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1063}} [ 1269.794484] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ResetVM_Task with opID=oslo.vmware-57a7b113-59cf-4459-9a94-565afd608456 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1269.801087] env[65726]: DEBUG oslo_vmware.api [None req-e0c56071-ba98-4289-82f5-a0b33d02c9d6 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Waiting for the task: (returnval){ [ 1269.801087] env[65726]: value = "task-5117046" [ 1269.801087] env[65726]: _type = "Task" [ 1269.801087] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1269.810228] env[65726]: DEBUG oslo_vmware.api [None req-e0c56071-ba98-4289-82f5-a0b33d02c9d6 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': task-5117046, 'name': ResetVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1269.886708] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e97dfc9c-52c4-40d1-b545-310ad12437a8 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Acquiring lock "58562386-ef51-44ef-869c-a3102a6aafad" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1269.886882] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e97dfc9c-52c4-40d1-b545-310ad12437a8 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Lock "58562386-ef51-44ef-869c-a3102a6aafad" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1269.990377] env[65726]: INFO nova.compute.manager [-] [instance: a4c229bb-e5f9-41d3-86f9-ca4cfb087558] Took 1.28 seconds to deallocate network for instance. [ 1270.166985] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Instance 9c0599e3-383f-46ab-809b-944cc3a4d206 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65726) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1270.167275] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Instance e3255f3b-028f-4a0b-b621-873730417936 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65726) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1270.167275] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Instance 55ee97c3-b014-48eb-b41b-fc5aa16bc09e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65726) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1270.167389] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Instance a4c229bb-e5f9-41d3-86f9-ca4cfb087558 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65726) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1270.312520] env[65726]: DEBUG oslo_vmware.api [None req-e0c56071-ba98-4289-82f5-a0b33d02c9d6 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': task-5117046, 'name': ResetVM_Task, 'duration_secs': 0.110095} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1270.312975] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-e0c56071-ba98-4289-82f5-a0b33d02c9d6 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: e3255f3b-028f-4a0b-b621-873730417936] Did hard reboot of VM {{(pid=65726) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1067}} [ 1270.312975] env[65726]: DEBUG nova.compute.manager [None req-e0c56071-ba98-4289-82f5-a0b33d02c9d6 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: e3255f3b-028f-4a0b-b621-873730417936] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1270.313726] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c556ea4-20bb-4c76-b129-2a39188a578c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1270.388893] env[65726]: DEBUG nova.compute.manager [None req-e97dfc9c-52c4-40d1-b545-310ad12437a8 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 58562386-ef51-44ef-869c-a3102a6aafad] Starting instance... {{(pid=65726) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 1270.499137] env[65726]: DEBUG oslo_concurrency.lockutils [None req-11154a5d-f93f-44ba-9a74-59c15507518c tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1270.670122] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Instance 58562386-ef51-44ef-869c-a3102a6aafad has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65726) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 1270.670334] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Total usable vcpus: 48, total allocated vcpus: 4 {{(pid=65726) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1159}} [ 1270.670448] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1280MB phys_disk=100GB used_disk=4GB total_vcpus=48 used_vcpus=4 pci_stats=[] stats={'failed_builds': '0', 'num_instances': '4', 'num_vm_active': '4', 'num_task_None': '2', 'num_os_type_None': '4', 'num_proj_f3f799aa94f64e8cb26d93f6124efd81': '2', 'io_workload': '0', 'num_task_reboot_pending_hard': '1', 'num_proj_acd800e0438940bc80ae3c6b672db4ee': '1', 'num_proj_6b56f93ba0e14521921484ac2c785c43': '1', 'num_task_deleting': '1'} {{(pid=65726) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1168}} [ 1270.745814] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbb05d5d-4241-4961-9c4c-d3aec6f652ad {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1270.754310] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8083a5c6-c3c8-49a1-8921-88125e39002c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1270.784827] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-932c5df4-e4f2-4f34-8d7d-4208b8992b88 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1270.792581] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc347507-1bcc-4bf1-9fb4-4188a3363a48 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1270.805865] env[65726]: DEBUG nova.compute.provider_tree [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1270.825600] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e0c56071-ba98-4289-82f5-a0b33d02c9d6 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Lock "e3255f3b-028f-4a0b-b621-873730417936" "released" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: held 4.404s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1270.909144] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e97dfc9c-52c4-40d1-b545-310ad12437a8 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1271.309278] env[65726]: DEBUG nova.scheduler.client.report [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1271.814175] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=65726) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1097}} [ 1271.814444] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.673s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1271.814791] env[65726]: DEBUG oslo_concurrency.lockutils [None req-11154a5d-f93f-44ba-9a74-59c15507518c tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.316s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1271.815060] env[65726]: DEBUG nova.objects.instance [None req-11154a5d-f93f-44ba-9a74-59c15507518c tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Lazy-loading 'resources' on Instance uuid a4c229bb-e5f9-41d3-86f9-ca4cfb087558 {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1272.387970] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2cedc66c-f7dc-46f6-810f-e4ab9af0498e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1272.395818] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17441aeb-3330-4de6-8785-f277631a0728 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1272.427984] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a02ee2a-0471-4e2d-ba2a-ee80df9951dd {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1272.437133] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0aa4a8b1-3074-4f90-8a59-0fe68f12abd6 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1272.453214] env[65726]: DEBUG nova.compute.provider_tree [None req-11154a5d-f93f-44ba-9a74-59c15507518c tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1272.957056] env[65726]: DEBUG nova.scheduler.client.report [None req-11154a5d-f93f-44ba-9a74-59c15507518c tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1273.462194] env[65726]: DEBUG oslo_concurrency.lockutils [None req-11154a5d-f93f-44ba-9a74-59c15507518c tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.647s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1273.464787] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e97dfc9c-52c4-40d1-b545-310ad12437a8 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.556s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1273.466494] env[65726]: INFO nova.compute.claims [None req-e97dfc9c-52c4-40d1-b545-310ad12437a8 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 58562386-ef51-44ef-869c-a3102a6aafad] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1273.482963] env[65726]: INFO nova.scheduler.client.report [None req-11154a5d-f93f-44ba-9a74-59c15507518c tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Deleted allocations for instance a4c229bb-e5f9-41d3-86f9-ca4cfb087558 [ 1273.991413] env[65726]: DEBUG oslo_concurrency.lockutils [None req-11154a5d-f93f-44ba-9a74-59c15507518c tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Lock "a4c229bb-e5f9-41d3-86f9-ca4cfb087558" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.928s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1274.494057] env[65726]: DEBUG nova.scheduler.client.report [None req-e97dfc9c-52c4-40d1-b545-310ad12437a8 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Refreshing inventories for resource provider 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 1274.508291] env[65726]: DEBUG nova.scheduler.client.report [None req-e97dfc9c-52c4-40d1-b545-310ad12437a8 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Updating ProviderTree inventory for provider 07c4692f-bdb4-4058-9173-ff9664830295 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 1274.508583] env[65726]: DEBUG nova.compute.provider_tree [None req-e97dfc9c-52c4-40d1-b545-310ad12437a8 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Updating inventory in ProviderTree for provider 07c4692f-bdb4-4058-9173-ff9664830295 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1274.519364] env[65726]: DEBUG nova.scheduler.client.report [None req-e97dfc9c-52c4-40d1-b545-310ad12437a8 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Refreshing aggregate associations for resource provider 07c4692f-bdb4-4058-9173-ff9664830295, aggregates: None {{(pid=65726) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 1274.536278] env[65726]: DEBUG nova.scheduler.client.report [None req-e97dfc9c-52c4-40d1-b545-310ad12437a8 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Refreshing trait associations for resource provider 07c4692f-bdb4-4058-9173-ff9664830295, traits: COMPUTE_IMAGE_TYPE_ISO,HW_ARCH_X86_64,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=65726) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 1274.595127] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ed24383-5632-41cb-80e2-003d1a74741a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1274.604827] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-789d1327-4df1-4ef0-9b01-44915b327288 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1274.636050] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59d275ea-d60a-4970-a001-c72b03f0a746 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1274.644120] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5edffbd0-0424-4645-9252-c5a313aa6caf {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1274.657877] env[65726]: DEBUG nova.compute.provider_tree [None req-e97dfc9c-52c4-40d1-b545-310ad12437a8 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1275.161839] env[65726]: DEBUG nova.scheduler.client.report [None req-e97dfc9c-52c4-40d1-b545-310ad12437a8 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1275.341574] env[65726]: DEBUG oslo_concurrency.lockutils [None req-db162dfe-c1d7-4167-bbfe-8ff6d4cf3921 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Acquiring lock "5eae4661-4212-4932-ad1c-cee6275b859c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1275.341795] env[65726]: DEBUG oslo_concurrency.lockutils [None req-db162dfe-c1d7-4167-bbfe-8ff6d4cf3921 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Lock "5eae4661-4212-4932-ad1c-cee6275b859c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1275.667156] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e97dfc9c-52c4-40d1-b545-310ad12437a8 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.202s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1275.667699] env[65726]: DEBUG nova.compute.manager [None req-e97dfc9c-52c4-40d1-b545-310ad12437a8 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 58562386-ef51-44ef-869c-a3102a6aafad] Start building networks asynchronously for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1275.811311] env[65726]: DEBUG oslo_service.periodic_task [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=65726) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1275.843598] env[65726]: DEBUG nova.compute.manager [None req-db162dfe-c1d7-4167-bbfe-8ff6d4cf3921 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 5eae4661-4212-4932-ad1c-cee6275b859c] Starting instance... {{(pid=65726) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 1276.173573] env[65726]: DEBUG nova.compute.utils [None req-e97dfc9c-52c4-40d1-b545-310ad12437a8 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Using /dev/sd instead of None {{(pid=65726) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1276.175768] env[65726]: DEBUG nova.compute.manager [None req-e97dfc9c-52c4-40d1-b545-310ad12437a8 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 58562386-ef51-44ef-869c-a3102a6aafad] Allocating IP information in the background. {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1276.175768] env[65726]: DEBUG nova.network.neutron [None req-e97dfc9c-52c4-40d1-b545-310ad12437a8 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 58562386-ef51-44ef-869c-a3102a6aafad] allocate_for_instance() {{(pid=65726) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 1276.175967] env[65726]: WARNING neutronclient.v2_0.client [None req-e97dfc9c-52c4-40d1-b545-310ad12437a8 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1276.176297] env[65726]: WARNING neutronclient.v2_0.client [None req-e97dfc9c-52c4-40d1-b545-310ad12437a8 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1276.176996] env[65726]: WARNING openstack [None req-e97dfc9c-52c4-40d1-b545-310ad12437a8 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1276.177419] env[65726]: WARNING openstack [None req-e97dfc9c-52c4-40d1-b545-310ad12437a8 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1276.185935] env[65726]: DEBUG nova.compute.manager [None req-e97dfc9c-52c4-40d1-b545-310ad12437a8 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 58562386-ef51-44ef-869c-a3102a6aafad] Start building block device mappings for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1276.233851] env[65726]: DEBUG nova.policy [None req-e97dfc9c-52c4-40d1-b545-310ad12437a8 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6ab5b309207a46bb9d95998ef0a7a46f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6b56f93ba0e14521921484ac2c785c43', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65726) authorize /opt/stack/nova/nova/policy.py:192}} [ 1276.366163] env[65726]: DEBUG oslo_concurrency.lockutils [None req-db162dfe-c1d7-4167-bbfe-8ff6d4cf3921 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1276.366446] env[65726]: DEBUG oslo_concurrency.lockutils [None req-db162dfe-c1d7-4167-bbfe-8ff6d4cf3921 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1276.368632] env[65726]: INFO nova.compute.claims [None req-db162dfe-c1d7-4167-bbfe-8ff6d4cf3921 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 5eae4661-4212-4932-ad1c-cee6275b859c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1276.535987] env[65726]: DEBUG nova.network.neutron [None req-e97dfc9c-52c4-40d1-b545-310ad12437a8 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 58562386-ef51-44ef-869c-a3102a6aafad] Successfully created port: dfc3c1e9-c8e6-41f5-aed1-e1c15c7f9ba6 {{(pid=65726) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1277.195653] env[65726]: DEBUG nova.compute.manager [None req-e97dfc9c-52c4-40d1-b545-310ad12437a8 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 58562386-ef51-44ef-869c-a3102a6aafad] Start spawning the instance on the hypervisor. {{(pid=65726) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1277.222282] env[65726]: DEBUG nova.virt.hardware [None req-e97dfc9c-52c4-40d1-b545-310ad12437a8 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T19:28:34Z,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1539f06fa3534e90acbc3a60c4b8bd3f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T19:28:35Z,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1277.222544] env[65726]: DEBUG nova.virt.hardware [None req-e97dfc9c-52c4-40d1-b545-310ad12437a8 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1277.222697] env[65726]: DEBUG nova.virt.hardware [None req-e97dfc9c-52c4-40d1-b545-310ad12437a8 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1277.222873] env[65726]: DEBUG nova.virt.hardware [None req-e97dfc9c-52c4-40d1-b545-310ad12437a8 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1277.223027] env[65726]: DEBUG nova.virt.hardware [None req-e97dfc9c-52c4-40d1-b545-310ad12437a8 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1277.223181] env[65726]: DEBUG nova.virt.hardware [None req-e97dfc9c-52c4-40d1-b545-310ad12437a8 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1277.223388] env[65726]: DEBUG nova.virt.hardware [None req-e97dfc9c-52c4-40d1-b545-310ad12437a8 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1277.223542] env[65726]: DEBUG nova.virt.hardware [None req-e97dfc9c-52c4-40d1-b545-310ad12437a8 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1277.223705] env[65726]: DEBUG nova.virt.hardware [None req-e97dfc9c-52c4-40d1-b545-310ad12437a8 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1277.223862] env[65726]: DEBUG nova.virt.hardware [None req-e97dfc9c-52c4-40d1-b545-310ad12437a8 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1277.224048] env[65726]: DEBUG nova.virt.hardware [None req-e97dfc9c-52c4-40d1-b545-310ad12437a8 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1277.224998] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-accfd4ea-a0a7-4816-881e-66b689a5aaa2 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1277.233168] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e08fc52-645b-4ee1-8e4e-4cbcc57e2df0 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1277.465647] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd9ec6b4-83d4-4c72-8803-5dc544e58666 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1277.474052] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af507ab8-a8b0-4355-99a0-e0ba69dc5a10 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1277.507787] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e412c88-0b70-4267-98d3-faa09224af11 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1277.516193] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2db9f7eb-72ee-4cc5-b1ea-74ca45c275c8 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1277.529931] env[65726]: DEBUG nova.compute.provider_tree [None req-db162dfe-c1d7-4167-bbfe-8ff6d4cf3921 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1277.954080] env[65726]: DEBUG nova.compute.manager [req-b19de506-22a9-4cd3-9724-d5c543b5e2b3 req-720e19d2-b222-4122-b28d-62e4f8e9767f service nova] [instance: 58562386-ef51-44ef-869c-a3102a6aafad] Received event network-vif-plugged-dfc3c1e9-c8e6-41f5-aed1-e1c15c7f9ba6 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1277.954350] env[65726]: DEBUG oslo_concurrency.lockutils [req-b19de506-22a9-4cd3-9724-d5c543b5e2b3 req-720e19d2-b222-4122-b28d-62e4f8e9767f service nova] Acquiring lock "58562386-ef51-44ef-869c-a3102a6aafad-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1277.954577] env[65726]: DEBUG oslo_concurrency.lockutils [req-b19de506-22a9-4cd3-9724-d5c543b5e2b3 req-720e19d2-b222-4122-b28d-62e4f8e9767f service nova] Lock "58562386-ef51-44ef-869c-a3102a6aafad-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1277.954793] env[65726]: DEBUG oslo_concurrency.lockutils [req-b19de506-22a9-4cd3-9724-d5c543b5e2b3 req-720e19d2-b222-4122-b28d-62e4f8e9767f service nova] Lock "58562386-ef51-44ef-869c-a3102a6aafad-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1277.954979] env[65726]: DEBUG nova.compute.manager [req-b19de506-22a9-4cd3-9724-d5c543b5e2b3 req-720e19d2-b222-4122-b28d-62e4f8e9767f service nova] [instance: 58562386-ef51-44ef-869c-a3102a6aafad] No waiting events found dispatching network-vif-plugged-dfc3c1e9-c8e6-41f5-aed1-e1c15c7f9ba6 {{(pid=65726) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1277.955142] env[65726]: WARNING nova.compute.manager [req-b19de506-22a9-4cd3-9724-d5c543b5e2b3 req-720e19d2-b222-4122-b28d-62e4f8e9767f service nova] [instance: 58562386-ef51-44ef-869c-a3102a6aafad] Received unexpected event network-vif-plugged-dfc3c1e9-c8e6-41f5-aed1-e1c15c7f9ba6 for instance with vm_state building and task_state spawning. [ 1278.033191] env[65726]: DEBUG nova.scheduler.client.report [None req-db162dfe-c1d7-4167-bbfe-8ff6d4cf3921 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1278.057968] env[65726]: DEBUG nova.network.neutron [None req-e97dfc9c-52c4-40d1-b545-310ad12437a8 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 58562386-ef51-44ef-869c-a3102a6aafad] Successfully updated port: dfc3c1e9-c8e6-41f5-aed1-e1c15c7f9ba6 {{(pid=65726) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 1278.538079] env[65726]: DEBUG oslo_concurrency.lockutils [None req-db162dfe-c1d7-4167-bbfe-8ff6d4cf3921 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.171s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1278.538578] env[65726]: DEBUG nova.compute.manager [None req-db162dfe-c1d7-4167-bbfe-8ff6d4cf3921 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 5eae4661-4212-4932-ad1c-cee6275b859c] Start building networks asynchronously for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1278.561238] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e97dfc9c-52c4-40d1-b545-310ad12437a8 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Acquiring lock "refresh_cache-58562386-ef51-44ef-869c-a3102a6aafad" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1278.561409] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e97dfc9c-52c4-40d1-b545-310ad12437a8 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Acquired lock "refresh_cache-58562386-ef51-44ef-869c-a3102a6aafad" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1278.561581] env[65726]: DEBUG nova.network.neutron [None req-e97dfc9c-52c4-40d1-b545-310ad12437a8 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 58562386-ef51-44ef-869c-a3102a6aafad] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1279.043552] env[65726]: DEBUG nova.compute.utils [None req-db162dfe-c1d7-4167-bbfe-8ff6d4cf3921 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Using /dev/sd instead of None {{(pid=65726) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1279.045365] env[65726]: DEBUG nova.compute.manager [None req-db162dfe-c1d7-4167-bbfe-8ff6d4cf3921 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 5eae4661-4212-4932-ad1c-cee6275b859c] Allocating IP information in the background. {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1279.045365] env[65726]: DEBUG nova.network.neutron [None req-db162dfe-c1d7-4167-bbfe-8ff6d4cf3921 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 5eae4661-4212-4932-ad1c-cee6275b859c] allocate_for_instance() {{(pid=65726) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 1279.045725] env[65726]: WARNING neutronclient.v2_0.client [None req-db162dfe-c1d7-4167-bbfe-8ff6d4cf3921 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1279.046104] env[65726]: WARNING neutronclient.v2_0.client [None req-db162dfe-c1d7-4167-bbfe-8ff6d4cf3921 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1279.046715] env[65726]: WARNING openstack [None req-db162dfe-c1d7-4167-bbfe-8ff6d4cf3921 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1279.047142] env[65726]: WARNING openstack [None req-db162dfe-c1d7-4167-bbfe-8ff6d4cf3921 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1279.065715] env[65726]: WARNING openstack [None req-e97dfc9c-52c4-40d1-b545-310ad12437a8 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1279.066104] env[65726]: WARNING openstack [None req-e97dfc9c-52c4-40d1-b545-310ad12437a8 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1279.071092] env[65726]: DEBUG nova.network.neutron [None req-e97dfc9c-52c4-40d1-b545-310ad12437a8 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 58562386-ef51-44ef-869c-a3102a6aafad] Instance cache missing network info. {{(pid=65726) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 1279.091678] env[65726]: DEBUG nova.policy [None req-db162dfe-c1d7-4167-bbfe-8ff6d4cf3921 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ccb3513b5b4f453f842139ec1815c4af', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f3f799aa94f64e8cb26d93f6124efd81', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65726) authorize /opt/stack/nova/nova/policy.py:192}} [ 1279.121584] env[65726]: WARNING openstack [None req-e97dfc9c-52c4-40d1-b545-310ad12437a8 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1279.121964] env[65726]: WARNING openstack [None req-e97dfc9c-52c4-40d1-b545-310ad12437a8 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1279.182260] env[65726]: WARNING openstack [None req-e97dfc9c-52c4-40d1-b545-310ad12437a8 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1279.182393] env[65726]: WARNING openstack [None req-e97dfc9c-52c4-40d1-b545-310ad12437a8 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1279.273234] env[65726]: DEBUG nova.network.neutron [None req-e97dfc9c-52c4-40d1-b545-310ad12437a8 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 58562386-ef51-44ef-869c-a3102a6aafad] Updating instance_info_cache with network_info: [{"id": "dfc3c1e9-c8e6-41f5-aed1-e1c15c7f9ba6", "address": "fa:16:3e:63:63:d0", "network": {"id": "336fe51f-971a-447d-8e68-505640e0db1b", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1030314703-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6b56f93ba0e14521921484ac2c785c43", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8ee1c76d-1a61-4546-85cb-d4bd3c1b35ef", "external-id": "nsx-vlan-transportzone-161", "segmentation_id": 161, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdfc3c1e9-c8", "ovs_interfaceid": "dfc3c1e9-c8e6-41f5-aed1-e1c15c7f9ba6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1279.339860] env[65726]: DEBUG nova.network.neutron [None req-db162dfe-c1d7-4167-bbfe-8ff6d4cf3921 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 5eae4661-4212-4932-ad1c-cee6275b859c] Successfully created port: c79ae8e4-e7f1-425a-a5b8-a203aaa7917a {{(pid=65726) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1279.555716] env[65726]: DEBUG nova.compute.manager [None req-db162dfe-c1d7-4167-bbfe-8ff6d4cf3921 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 5eae4661-4212-4932-ad1c-cee6275b859c] Start building block device mappings for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1279.776463] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e97dfc9c-52c4-40d1-b545-310ad12437a8 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Releasing lock "refresh_cache-58562386-ef51-44ef-869c-a3102a6aafad" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1279.776840] env[65726]: DEBUG nova.compute.manager [None req-e97dfc9c-52c4-40d1-b545-310ad12437a8 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 58562386-ef51-44ef-869c-a3102a6aafad] Instance network_info: |[{"id": "dfc3c1e9-c8e6-41f5-aed1-e1c15c7f9ba6", "address": "fa:16:3e:63:63:d0", "network": {"id": "336fe51f-971a-447d-8e68-505640e0db1b", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1030314703-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6b56f93ba0e14521921484ac2c785c43", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8ee1c76d-1a61-4546-85cb-d4bd3c1b35ef", "external-id": "nsx-vlan-transportzone-161", "segmentation_id": 161, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdfc3c1e9-c8", "ovs_interfaceid": "dfc3c1e9-c8e6-41f5-aed1-e1c15c7f9ba6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1279.777336] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-e97dfc9c-52c4-40d1-b545-310ad12437a8 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 58562386-ef51-44ef-869c-a3102a6aafad] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:63:63:d0', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '8ee1c76d-1a61-4546-85cb-d4bd3c1b35ef', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'dfc3c1e9-c8e6-41f5-aed1-e1c15c7f9ba6', 'vif_model': 'vmxnet3'}] {{(pid=65726) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1279.784831] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-e97dfc9c-52c4-40d1-b545-310ad12437a8 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1279.785064] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 58562386-ef51-44ef-869c-a3102a6aafad] Creating VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1279.785305] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-846228a3-2a79-404c-bcde-5e2f3194da27 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1279.806280] env[65726]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1279.806280] env[65726]: value = "task-5117047" [ 1279.806280] env[65726]: _type = "Task" [ 1279.806280] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1279.815770] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5117047, 'name': CreateVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1279.985678] env[65726]: DEBUG nova.compute.manager [req-88974c33-a696-4ec6-b395-a5f8b4805db8 req-c3f6ad76-2ea6-44fe-80ba-cf70495ea982 service nova] [instance: 58562386-ef51-44ef-869c-a3102a6aafad] Received event network-changed-dfc3c1e9-c8e6-41f5-aed1-e1c15c7f9ba6 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1279.985927] env[65726]: DEBUG nova.compute.manager [req-88974c33-a696-4ec6-b395-a5f8b4805db8 req-c3f6ad76-2ea6-44fe-80ba-cf70495ea982 service nova] [instance: 58562386-ef51-44ef-869c-a3102a6aafad] Refreshing instance network info cache due to event network-changed-dfc3c1e9-c8e6-41f5-aed1-e1c15c7f9ba6. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 1279.986181] env[65726]: DEBUG oslo_concurrency.lockutils [req-88974c33-a696-4ec6-b395-a5f8b4805db8 req-c3f6ad76-2ea6-44fe-80ba-cf70495ea982 service nova] Acquiring lock "refresh_cache-58562386-ef51-44ef-869c-a3102a6aafad" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1279.986301] env[65726]: DEBUG oslo_concurrency.lockutils [req-88974c33-a696-4ec6-b395-a5f8b4805db8 req-c3f6ad76-2ea6-44fe-80ba-cf70495ea982 service nova] Acquired lock "refresh_cache-58562386-ef51-44ef-869c-a3102a6aafad" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1279.986470] env[65726]: DEBUG nova.network.neutron [req-88974c33-a696-4ec6-b395-a5f8b4805db8 req-c3f6ad76-2ea6-44fe-80ba-cf70495ea982 service nova] [instance: 58562386-ef51-44ef-869c-a3102a6aafad] Refreshing network info cache for port dfc3c1e9-c8e6-41f5-aed1-e1c15c7f9ba6 {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 1280.319708] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5117047, 'name': CreateVM_Task, 'duration_secs': 0.304064} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1280.319708] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 58562386-ef51-44ef-869c-a3102a6aafad] Created VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1280.320388] env[65726]: WARNING openstack [None req-e97dfc9c-52c4-40d1-b545-310ad12437a8 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1280.320753] env[65726]: WARNING openstack [None req-e97dfc9c-52c4-40d1-b545-310ad12437a8 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1280.326160] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e97dfc9c-52c4-40d1-b545-310ad12437a8 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1280.326446] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e97dfc9c-52c4-40d1-b545-310ad12437a8 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1280.326639] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e97dfc9c-52c4-40d1-b545-310ad12437a8 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1280.326942] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-00133aef-1f38-4a68-86fc-72f009ac1cce {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1280.332725] env[65726]: DEBUG oslo_vmware.api [None req-e97dfc9c-52c4-40d1-b545-310ad12437a8 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Waiting for the task: (returnval){ [ 1280.332725] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5227f490-1c04-2923-f52e-90d0cf58d386" [ 1280.332725] env[65726]: _type = "Task" [ 1280.332725] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1280.341180] env[65726]: DEBUG oslo_vmware.api [None req-e97dfc9c-52c4-40d1-b545-310ad12437a8 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5227f490-1c04-2923-f52e-90d0cf58d386, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1280.489669] env[65726]: WARNING openstack [req-88974c33-a696-4ec6-b395-a5f8b4805db8 req-c3f6ad76-2ea6-44fe-80ba-cf70495ea982 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1280.490081] env[65726]: WARNING openstack [req-88974c33-a696-4ec6-b395-a5f8b4805db8 req-c3f6ad76-2ea6-44fe-80ba-cf70495ea982 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1280.567190] env[65726]: DEBUG nova.compute.manager [None req-db162dfe-c1d7-4167-bbfe-8ff6d4cf3921 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 5eae4661-4212-4932-ad1c-cee6275b859c] Start spawning the instance on the hypervisor. {{(pid=65726) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1280.591463] env[65726]: WARNING openstack [req-88974c33-a696-4ec6-b395-a5f8b4805db8 req-c3f6ad76-2ea6-44fe-80ba-cf70495ea982 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1280.591844] env[65726]: WARNING openstack [req-88974c33-a696-4ec6-b395-a5f8b4805db8 req-c3f6ad76-2ea6-44fe-80ba-cf70495ea982 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1280.600260] env[65726]: DEBUG nova.virt.hardware [None req-db162dfe-c1d7-4167-bbfe-8ff6d4cf3921 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T19:28:34Z,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1539f06fa3534e90acbc3a60c4b8bd3f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T19:28:35Z,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1280.600516] env[65726]: DEBUG nova.virt.hardware [None req-db162dfe-c1d7-4167-bbfe-8ff6d4cf3921 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1280.600675] env[65726]: DEBUG nova.virt.hardware [None req-db162dfe-c1d7-4167-bbfe-8ff6d4cf3921 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1280.600841] env[65726]: DEBUG nova.virt.hardware [None req-db162dfe-c1d7-4167-bbfe-8ff6d4cf3921 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1280.601026] env[65726]: DEBUG nova.virt.hardware [None req-db162dfe-c1d7-4167-bbfe-8ff6d4cf3921 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1280.601187] env[65726]: DEBUG nova.virt.hardware [None req-db162dfe-c1d7-4167-bbfe-8ff6d4cf3921 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1280.601396] env[65726]: DEBUG nova.virt.hardware [None req-db162dfe-c1d7-4167-bbfe-8ff6d4cf3921 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1280.601545] env[65726]: DEBUG nova.virt.hardware [None req-db162dfe-c1d7-4167-bbfe-8ff6d4cf3921 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1280.601709] env[65726]: DEBUG nova.virt.hardware [None req-db162dfe-c1d7-4167-bbfe-8ff6d4cf3921 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1280.601891] env[65726]: DEBUG nova.virt.hardware [None req-db162dfe-c1d7-4167-bbfe-8ff6d4cf3921 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1280.602081] env[65726]: DEBUG nova.virt.hardware [None req-db162dfe-c1d7-4167-bbfe-8ff6d4cf3921 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1280.603212] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21d6cf6f-3647-4fea-ae11-8b10702089f7 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1280.612981] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56514f46-c2a0-424b-b188-ba9421156eb8 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1280.662114] env[65726]: WARNING openstack [req-88974c33-a696-4ec6-b395-a5f8b4805db8 req-c3f6ad76-2ea6-44fe-80ba-cf70495ea982 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1280.662657] env[65726]: WARNING openstack [req-88974c33-a696-4ec6-b395-a5f8b4805db8 req-c3f6ad76-2ea6-44fe-80ba-cf70495ea982 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1280.738508] env[65726]: DEBUG nova.network.neutron [req-88974c33-a696-4ec6-b395-a5f8b4805db8 req-c3f6ad76-2ea6-44fe-80ba-cf70495ea982 service nova] [instance: 58562386-ef51-44ef-869c-a3102a6aafad] Updated VIF entry in instance network info cache for port dfc3c1e9-c8e6-41f5-aed1-e1c15c7f9ba6. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 1280.738856] env[65726]: DEBUG nova.network.neutron [req-88974c33-a696-4ec6-b395-a5f8b4805db8 req-c3f6ad76-2ea6-44fe-80ba-cf70495ea982 service nova] [instance: 58562386-ef51-44ef-869c-a3102a6aafad] Updating instance_info_cache with network_info: [{"id": "dfc3c1e9-c8e6-41f5-aed1-e1c15c7f9ba6", "address": "fa:16:3e:63:63:d0", "network": {"id": "336fe51f-971a-447d-8e68-505640e0db1b", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1030314703-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6b56f93ba0e14521921484ac2c785c43", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8ee1c76d-1a61-4546-85cb-d4bd3c1b35ef", "external-id": "nsx-vlan-transportzone-161", "segmentation_id": 161, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdfc3c1e9-c8", "ovs_interfaceid": "dfc3c1e9-c8e6-41f5-aed1-e1c15c7f9ba6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1280.804792] env[65726]: DEBUG nova.network.neutron [None req-db162dfe-c1d7-4167-bbfe-8ff6d4cf3921 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 5eae4661-4212-4932-ad1c-cee6275b859c] Successfully updated port: c79ae8e4-e7f1-425a-a5b8-a203aaa7917a {{(pid=65726) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 1280.844523] env[65726]: DEBUG oslo_vmware.api [None req-e97dfc9c-52c4-40d1-b545-310ad12437a8 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5227f490-1c04-2923-f52e-90d0cf58d386, 'name': SearchDatastore_Task, 'duration_secs': 0.015714} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1280.844843] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e97dfc9c-52c4-40d1-b545-310ad12437a8 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1280.845089] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-e97dfc9c-52c4-40d1-b545-310ad12437a8 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 58562386-ef51-44ef-869c-a3102a6aafad] Processing image b52362a3-ee8a-4cbf-b06f-513b0cc8f95c {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1280.845329] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e97dfc9c-52c4-40d1-b545-310ad12437a8 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1280.845470] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e97dfc9c-52c4-40d1-b545-310ad12437a8 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1280.845690] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-e97dfc9c-52c4-40d1-b545-310ad12437a8 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1280.846021] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4006377c-840e-4edd-917e-9c3ff5248ea3 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1280.857577] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-e97dfc9c-52c4-40d1-b545-310ad12437a8 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1280.857786] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-e97dfc9c-52c4-40d1-b545-310ad12437a8 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65726) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1280.858592] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-105f328b-86de-497b-96f5-8f35415ec163 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1280.864349] env[65726]: DEBUG oslo_vmware.api [None req-e97dfc9c-52c4-40d1-b545-310ad12437a8 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Waiting for the task: (returnval){ [ 1280.864349] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52915571-d43a-d6a5-708d-850a0347732d" [ 1280.864349] env[65726]: _type = "Task" [ 1280.864349] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1280.872724] env[65726]: DEBUG oslo_vmware.api [None req-e97dfc9c-52c4-40d1-b545-310ad12437a8 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52915571-d43a-d6a5-708d-850a0347732d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1281.241958] env[65726]: DEBUG oslo_concurrency.lockutils [req-88974c33-a696-4ec6-b395-a5f8b4805db8 req-c3f6ad76-2ea6-44fe-80ba-cf70495ea982 service nova] Releasing lock "refresh_cache-58562386-ef51-44ef-869c-a3102a6aafad" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1281.307805] env[65726]: DEBUG oslo_concurrency.lockutils [None req-db162dfe-c1d7-4167-bbfe-8ff6d4cf3921 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Acquiring lock "refresh_cache-5eae4661-4212-4932-ad1c-cee6275b859c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1281.308008] env[65726]: DEBUG oslo_concurrency.lockutils [None req-db162dfe-c1d7-4167-bbfe-8ff6d4cf3921 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Acquired lock "refresh_cache-5eae4661-4212-4932-ad1c-cee6275b859c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1281.308258] env[65726]: DEBUG nova.network.neutron [None req-db162dfe-c1d7-4167-bbfe-8ff6d4cf3921 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 5eae4661-4212-4932-ad1c-cee6275b859c] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1281.375773] env[65726]: DEBUG oslo_vmware.api [None req-e97dfc9c-52c4-40d1-b545-310ad12437a8 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52915571-d43a-d6a5-708d-850a0347732d, 'name': SearchDatastore_Task, 'duration_secs': 0.010047} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1281.376661] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-aaa9b971-99b9-48b5-bc0b-dccef4044898 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1281.382717] env[65726]: DEBUG oslo_vmware.api [None req-e97dfc9c-52c4-40d1-b545-310ad12437a8 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Waiting for the task: (returnval){ [ 1281.382717] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5252abf8-d470-86da-6136-4d33997dc496" [ 1281.382717] env[65726]: _type = "Task" [ 1281.382717] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1281.392443] env[65726]: DEBUG oslo_vmware.api [None req-e97dfc9c-52c4-40d1-b545-310ad12437a8 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5252abf8-d470-86da-6136-4d33997dc496, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1281.811175] env[65726]: WARNING openstack [None req-db162dfe-c1d7-4167-bbfe-8ff6d4cf3921 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1281.811691] env[65726]: WARNING openstack [None req-db162dfe-c1d7-4167-bbfe-8ff6d4cf3921 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1281.816803] env[65726]: DEBUG nova.network.neutron [None req-db162dfe-c1d7-4167-bbfe-8ff6d4cf3921 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 5eae4661-4212-4932-ad1c-cee6275b859c] Instance cache missing network info. {{(pid=65726) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 1281.864611] env[65726]: WARNING openstack [None req-db162dfe-c1d7-4167-bbfe-8ff6d4cf3921 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1281.865102] env[65726]: WARNING openstack [None req-db162dfe-c1d7-4167-bbfe-8ff6d4cf3921 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1281.893758] env[65726]: DEBUG oslo_vmware.api [None req-e97dfc9c-52c4-40d1-b545-310ad12437a8 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5252abf8-d470-86da-6136-4d33997dc496, 'name': SearchDatastore_Task, 'duration_secs': 0.050883} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1281.894041] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e97dfc9c-52c4-40d1-b545-310ad12437a8 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1281.894304] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-e97dfc9c-52c4-40d1-b545-310ad12437a8 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] 58562386-ef51-44ef-869c-a3102a6aafad/58562386-ef51-44ef-869c-a3102a6aafad.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1281.898301] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b25951a5-f86f-413d-981f-7363cfaac20a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1281.905567] env[65726]: DEBUG oslo_vmware.api [None req-e97dfc9c-52c4-40d1-b545-310ad12437a8 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Waiting for the task: (returnval){ [ 1281.905567] env[65726]: value = "task-5117048" [ 1281.905567] env[65726]: _type = "Task" [ 1281.905567] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1281.914279] env[65726]: DEBUG oslo_vmware.api [None req-e97dfc9c-52c4-40d1-b545-310ad12437a8 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': task-5117048, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1281.930965] env[65726]: WARNING openstack [None req-db162dfe-c1d7-4167-bbfe-8ff6d4cf3921 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1281.931413] env[65726]: WARNING openstack [None req-db162dfe-c1d7-4167-bbfe-8ff6d4cf3921 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1282.019549] env[65726]: DEBUG nova.compute.manager [req-e5852225-1f23-4e1c-8163-9095a1850fd1 req-8edc8098-f32c-422e-9f4b-0116ec24101b service nova] [instance: 5eae4661-4212-4932-ad1c-cee6275b859c] Received event network-vif-plugged-c79ae8e4-e7f1-425a-a5b8-a203aaa7917a {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1282.019549] env[65726]: DEBUG oslo_concurrency.lockutils [req-e5852225-1f23-4e1c-8163-9095a1850fd1 req-8edc8098-f32c-422e-9f4b-0116ec24101b service nova] Acquiring lock "5eae4661-4212-4932-ad1c-cee6275b859c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1282.019549] env[65726]: DEBUG oslo_concurrency.lockutils [req-e5852225-1f23-4e1c-8163-9095a1850fd1 req-8edc8098-f32c-422e-9f4b-0116ec24101b service nova] Lock "5eae4661-4212-4932-ad1c-cee6275b859c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1282.019765] env[65726]: DEBUG oslo_concurrency.lockutils [req-e5852225-1f23-4e1c-8163-9095a1850fd1 req-8edc8098-f32c-422e-9f4b-0116ec24101b service nova] Lock "5eae4661-4212-4932-ad1c-cee6275b859c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1282.019997] env[65726]: DEBUG nova.compute.manager [req-e5852225-1f23-4e1c-8163-9095a1850fd1 req-8edc8098-f32c-422e-9f4b-0116ec24101b service nova] [instance: 5eae4661-4212-4932-ad1c-cee6275b859c] No waiting events found dispatching network-vif-plugged-c79ae8e4-e7f1-425a-a5b8-a203aaa7917a {{(pid=65726) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1282.020141] env[65726]: WARNING nova.compute.manager [req-e5852225-1f23-4e1c-8163-9095a1850fd1 req-8edc8098-f32c-422e-9f4b-0116ec24101b service nova] [instance: 5eae4661-4212-4932-ad1c-cee6275b859c] Received unexpected event network-vif-plugged-c79ae8e4-e7f1-425a-a5b8-a203aaa7917a for instance with vm_state building and task_state spawning. [ 1282.020269] env[65726]: DEBUG nova.compute.manager [req-e5852225-1f23-4e1c-8163-9095a1850fd1 req-8edc8098-f32c-422e-9f4b-0116ec24101b service nova] [instance: 5eae4661-4212-4932-ad1c-cee6275b859c] Received event network-changed-c79ae8e4-e7f1-425a-a5b8-a203aaa7917a {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1282.020418] env[65726]: DEBUG nova.compute.manager [req-e5852225-1f23-4e1c-8163-9095a1850fd1 req-8edc8098-f32c-422e-9f4b-0116ec24101b service nova] [instance: 5eae4661-4212-4932-ad1c-cee6275b859c] Refreshing instance network info cache due to event network-changed-c79ae8e4-e7f1-425a-a5b8-a203aaa7917a. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 1282.020583] env[65726]: DEBUG oslo_concurrency.lockutils [req-e5852225-1f23-4e1c-8163-9095a1850fd1 req-8edc8098-f32c-422e-9f4b-0116ec24101b service nova] Acquiring lock "refresh_cache-5eae4661-4212-4932-ad1c-cee6275b859c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1282.023240] env[65726]: DEBUG nova.network.neutron [None req-db162dfe-c1d7-4167-bbfe-8ff6d4cf3921 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 5eae4661-4212-4932-ad1c-cee6275b859c] Updating instance_info_cache with network_info: [{"id": "c79ae8e4-e7f1-425a-a5b8-a203aaa7917a", "address": "fa:16:3e:23:28:63", "network": {"id": "69642329-e986-48fa-accf-1f296c08116a", "bridge": "br-int", "label": "tempest-ServersTestJSON-1786538788-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f3f799aa94f64e8cb26d93f6124efd81", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "816c6e38-e200-4544-8c5b-9fc3e16c5761", "external-id": "nsx-vlan-transportzone-195", "segmentation_id": 195, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc79ae8e4-e7", "ovs_interfaceid": "c79ae8e4-e7f1-425a-a5b8-a203aaa7917a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1282.417414] env[65726]: DEBUG oslo_vmware.api [None req-e97dfc9c-52c4-40d1-b545-310ad12437a8 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': task-5117048, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1282.526405] env[65726]: DEBUG oslo_concurrency.lockutils [None req-db162dfe-c1d7-4167-bbfe-8ff6d4cf3921 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Releasing lock "refresh_cache-5eae4661-4212-4932-ad1c-cee6275b859c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1282.526801] env[65726]: DEBUG nova.compute.manager [None req-db162dfe-c1d7-4167-bbfe-8ff6d4cf3921 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 5eae4661-4212-4932-ad1c-cee6275b859c] Instance network_info: |[{"id": "c79ae8e4-e7f1-425a-a5b8-a203aaa7917a", "address": "fa:16:3e:23:28:63", "network": {"id": "69642329-e986-48fa-accf-1f296c08116a", "bridge": "br-int", "label": "tempest-ServersTestJSON-1786538788-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f3f799aa94f64e8cb26d93f6124efd81", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "816c6e38-e200-4544-8c5b-9fc3e16c5761", "external-id": "nsx-vlan-transportzone-195", "segmentation_id": 195, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc79ae8e4-e7", "ovs_interfaceid": "c79ae8e4-e7f1-425a-a5b8-a203aaa7917a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1282.527221] env[65726]: DEBUG oslo_concurrency.lockutils [req-e5852225-1f23-4e1c-8163-9095a1850fd1 req-8edc8098-f32c-422e-9f4b-0116ec24101b service nova] Acquired lock "refresh_cache-5eae4661-4212-4932-ad1c-cee6275b859c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1282.527419] env[65726]: DEBUG nova.network.neutron [req-e5852225-1f23-4e1c-8163-9095a1850fd1 req-8edc8098-f32c-422e-9f4b-0116ec24101b service nova] [instance: 5eae4661-4212-4932-ad1c-cee6275b859c] Refreshing network info cache for port c79ae8e4-e7f1-425a-a5b8-a203aaa7917a {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 1282.528689] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-db162dfe-c1d7-4167-bbfe-8ff6d4cf3921 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 5eae4661-4212-4932-ad1c-cee6275b859c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:23:28:63', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '816c6e38-e200-4544-8c5b-9fc3e16c5761', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c79ae8e4-e7f1-425a-a5b8-a203aaa7917a', 'vif_model': 'vmxnet3'}] {{(pid=65726) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1282.536634] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-db162dfe-c1d7-4167-bbfe-8ff6d4cf3921 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1282.538232] env[65726]: WARNING openstack [req-e5852225-1f23-4e1c-8163-9095a1850fd1 req-8edc8098-f32c-422e-9f4b-0116ec24101b service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1282.538580] env[65726]: WARNING openstack [req-e5852225-1f23-4e1c-8163-9095a1850fd1 req-8edc8098-f32c-422e-9f4b-0116ec24101b service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1282.545047] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5eae4661-4212-4932-ad1c-cee6275b859c] Creating VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1282.546011] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d89d46df-8d59-4984-ad11-bcec41a3d202 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1282.568424] env[65726]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1282.568424] env[65726]: value = "task-5117049" [ 1282.568424] env[65726]: _type = "Task" [ 1282.568424] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1282.577791] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5117049, 'name': CreateVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1282.639065] env[65726]: WARNING openstack [req-e5852225-1f23-4e1c-8163-9095a1850fd1 req-8edc8098-f32c-422e-9f4b-0116ec24101b service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1282.639505] env[65726]: WARNING openstack [req-e5852225-1f23-4e1c-8163-9095a1850fd1 req-8edc8098-f32c-422e-9f4b-0116ec24101b service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1282.703188] env[65726]: WARNING openstack [req-e5852225-1f23-4e1c-8163-9095a1850fd1 req-8edc8098-f32c-422e-9f4b-0116ec24101b service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1282.703578] env[65726]: WARNING openstack [req-e5852225-1f23-4e1c-8163-9095a1850fd1 req-8edc8098-f32c-422e-9f4b-0116ec24101b service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1282.785083] env[65726]: DEBUG nova.network.neutron [req-e5852225-1f23-4e1c-8163-9095a1850fd1 req-8edc8098-f32c-422e-9f4b-0116ec24101b service nova] [instance: 5eae4661-4212-4932-ad1c-cee6275b859c] Updated VIF entry in instance network info cache for port c79ae8e4-e7f1-425a-a5b8-a203aaa7917a. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 1282.785474] env[65726]: DEBUG nova.network.neutron [req-e5852225-1f23-4e1c-8163-9095a1850fd1 req-8edc8098-f32c-422e-9f4b-0116ec24101b service nova] [instance: 5eae4661-4212-4932-ad1c-cee6275b859c] Updating instance_info_cache with network_info: [{"id": "c79ae8e4-e7f1-425a-a5b8-a203aaa7917a", "address": "fa:16:3e:23:28:63", "network": {"id": "69642329-e986-48fa-accf-1f296c08116a", "bridge": "br-int", "label": "tempest-ServersTestJSON-1786538788-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f3f799aa94f64e8cb26d93f6124efd81", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "816c6e38-e200-4544-8c5b-9fc3e16c5761", "external-id": "nsx-vlan-transportzone-195", "segmentation_id": 195, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc79ae8e4-e7", "ovs_interfaceid": "c79ae8e4-e7f1-425a-a5b8-a203aaa7917a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1282.917495] env[65726]: DEBUG oslo_vmware.api [None req-e97dfc9c-52c4-40d1-b545-310ad12437a8 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': task-5117048, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.533558} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1282.917848] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-e97dfc9c-52c4-40d1-b545-310ad12437a8 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] 58562386-ef51-44ef-869c-a3102a6aafad/58562386-ef51-44ef-869c-a3102a6aafad.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1282.917991] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-e97dfc9c-52c4-40d1-b545-310ad12437a8 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 58562386-ef51-44ef-869c-a3102a6aafad] Extending root virtual disk to 1048576 {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1282.918289] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5fb59b8b-d173-4014-89d1-240428423e9f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1282.927208] env[65726]: DEBUG oslo_vmware.api [None req-e97dfc9c-52c4-40d1-b545-310ad12437a8 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Waiting for the task: (returnval){ [ 1282.927208] env[65726]: value = "task-5117050" [ 1282.927208] env[65726]: _type = "Task" [ 1282.927208] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1282.935838] env[65726]: DEBUG oslo_vmware.api [None req-e97dfc9c-52c4-40d1-b545-310ad12437a8 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': task-5117050, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1283.079019] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5117049, 'name': CreateVM_Task, 'duration_secs': 0.438942} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1283.079226] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5eae4661-4212-4932-ad1c-cee6275b859c] Created VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1283.079934] env[65726]: WARNING openstack [None req-db162dfe-c1d7-4167-bbfe-8ff6d4cf3921 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1283.080301] env[65726]: WARNING openstack [None req-db162dfe-c1d7-4167-bbfe-8ff6d4cf3921 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1283.085307] env[65726]: DEBUG oslo_concurrency.lockutils [None req-db162dfe-c1d7-4167-bbfe-8ff6d4cf3921 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1283.085459] env[65726]: DEBUG oslo_concurrency.lockutils [None req-db162dfe-c1d7-4167-bbfe-8ff6d4cf3921 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1283.085852] env[65726]: DEBUG oslo_concurrency.lockutils [None req-db162dfe-c1d7-4167-bbfe-8ff6d4cf3921 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1283.086130] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-17e21054-1ef7-4fd1-9f56-7a1130e0ffbb {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1283.090856] env[65726]: DEBUG oslo_vmware.api [None req-db162dfe-c1d7-4167-bbfe-8ff6d4cf3921 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Waiting for the task: (returnval){ [ 1283.090856] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5220d582-097c-20e5-c616-2e524e375995" [ 1283.090856] env[65726]: _type = "Task" [ 1283.090856] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1283.098641] env[65726]: DEBUG oslo_vmware.api [None req-db162dfe-c1d7-4167-bbfe-8ff6d4cf3921 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5220d582-097c-20e5-c616-2e524e375995, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1283.288823] env[65726]: DEBUG oslo_concurrency.lockutils [req-e5852225-1f23-4e1c-8163-9095a1850fd1 req-8edc8098-f32c-422e-9f4b-0116ec24101b service nova] Releasing lock "refresh_cache-5eae4661-4212-4932-ad1c-cee6275b859c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1283.437951] env[65726]: DEBUG oslo_vmware.api [None req-e97dfc9c-52c4-40d1-b545-310ad12437a8 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': task-5117050, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.086091} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1283.437951] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-e97dfc9c-52c4-40d1-b545-310ad12437a8 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 58562386-ef51-44ef-869c-a3102a6aafad] Extended root virtual disk {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1283.438367] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5ff9ce2-1f6f-41ea-8b4e-169178572579 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1283.462792] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-e97dfc9c-52c4-40d1-b545-310ad12437a8 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 58562386-ef51-44ef-869c-a3102a6aafad] Reconfiguring VM instance instance-00000077 to attach disk [datastore1] 58562386-ef51-44ef-869c-a3102a6aafad/58562386-ef51-44ef-869c-a3102a6aafad.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1283.463160] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5748a178-b907-4654-8a69-54b5160f70cf {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1283.483650] env[65726]: DEBUG oslo_vmware.api [None req-e97dfc9c-52c4-40d1-b545-310ad12437a8 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Waiting for the task: (returnval){ [ 1283.483650] env[65726]: value = "task-5117051" [ 1283.483650] env[65726]: _type = "Task" [ 1283.483650] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1283.492605] env[65726]: DEBUG oslo_vmware.api [None req-e97dfc9c-52c4-40d1-b545-310ad12437a8 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': task-5117051, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1283.603204] env[65726]: DEBUG oslo_vmware.api [None req-db162dfe-c1d7-4167-bbfe-8ff6d4cf3921 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5220d582-097c-20e5-c616-2e524e375995, 'name': SearchDatastore_Task, 'duration_secs': 0.012518} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1283.603464] env[65726]: DEBUG oslo_concurrency.lockutils [None req-db162dfe-c1d7-4167-bbfe-8ff6d4cf3921 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1283.603695] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-db162dfe-c1d7-4167-bbfe-8ff6d4cf3921 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 5eae4661-4212-4932-ad1c-cee6275b859c] Processing image b52362a3-ee8a-4cbf-b06f-513b0cc8f95c {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1283.603951] env[65726]: DEBUG oslo_concurrency.lockutils [None req-db162dfe-c1d7-4167-bbfe-8ff6d4cf3921 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1283.604141] env[65726]: DEBUG oslo_concurrency.lockutils [None req-db162dfe-c1d7-4167-bbfe-8ff6d4cf3921 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1283.604330] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-db162dfe-c1d7-4167-bbfe-8ff6d4cf3921 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1283.604599] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-382cbc83-14f3-428b-85f4-119eb21d679c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1283.614747] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-db162dfe-c1d7-4167-bbfe-8ff6d4cf3921 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1283.614954] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-db162dfe-c1d7-4167-bbfe-8ff6d4cf3921 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65726) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1283.615788] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-33073951-b9d0-40b5-b7c5-1a64a58ca273 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1283.621264] env[65726]: DEBUG oslo_vmware.api [None req-db162dfe-c1d7-4167-bbfe-8ff6d4cf3921 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Waiting for the task: (returnval){ [ 1283.621264] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52d3416b-b306-d313-79f0-72f16a991fa1" [ 1283.621264] env[65726]: _type = "Task" [ 1283.621264] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1283.630085] env[65726]: DEBUG oslo_vmware.api [None req-db162dfe-c1d7-4167-bbfe-8ff6d4cf3921 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52d3416b-b306-d313-79f0-72f16a991fa1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1283.996016] env[65726]: DEBUG oslo_vmware.api [None req-e97dfc9c-52c4-40d1-b545-310ad12437a8 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': task-5117051, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1284.131697] env[65726]: DEBUG oslo_vmware.api [None req-db162dfe-c1d7-4167-bbfe-8ff6d4cf3921 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52d3416b-b306-d313-79f0-72f16a991fa1, 'name': SearchDatastore_Task, 'duration_secs': 0.011543} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1284.132510] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e5a942af-be83-459b-806d-777f95d12bca {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1284.137972] env[65726]: DEBUG oslo_vmware.api [None req-db162dfe-c1d7-4167-bbfe-8ff6d4cf3921 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Waiting for the task: (returnval){ [ 1284.137972] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52cf199b-d956-0246-0395-bc86dc5e4e44" [ 1284.137972] env[65726]: _type = "Task" [ 1284.137972] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1284.145852] env[65726]: DEBUG oslo_vmware.api [None req-db162dfe-c1d7-4167-bbfe-8ff6d4cf3921 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52cf199b-d956-0246-0395-bc86dc5e4e44, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1284.494463] env[65726]: DEBUG oslo_vmware.api [None req-e97dfc9c-52c4-40d1-b545-310ad12437a8 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': task-5117051, 'name': ReconfigVM_Task, 'duration_secs': 0.81646} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1284.494773] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-e97dfc9c-52c4-40d1-b545-310ad12437a8 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 58562386-ef51-44ef-869c-a3102a6aafad] Reconfigured VM instance instance-00000077 to attach disk [datastore1] 58562386-ef51-44ef-869c-a3102a6aafad/58562386-ef51-44ef-869c-a3102a6aafad.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1284.495445] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0c7585d3-c511-456a-a545-608a982b3d6c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1284.502898] env[65726]: DEBUG oslo_vmware.api [None req-e97dfc9c-52c4-40d1-b545-310ad12437a8 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Waiting for the task: (returnval){ [ 1284.502898] env[65726]: value = "task-5117052" [ 1284.502898] env[65726]: _type = "Task" [ 1284.502898] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1284.512195] env[65726]: DEBUG oslo_vmware.api [None req-e97dfc9c-52c4-40d1-b545-310ad12437a8 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': task-5117052, 'name': Rename_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1284.649548] env[65726]: DEBUG oslo_vmware.api [None req-db162dfe-c1d7-4167-bbfe-8ff6d4cf3921 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52cf199b-d956-0246-0395-bc86dc5e4e44, 'name': SearchDatastore_Task, 'duration_secs': 0.012171} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1284.649827] env[65726]: DEBUG oslo_concurrency.lockutils [None req-db162dfe-c1d7-4167-bbfe-8ff6d4cf3921 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1284.650107] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-db162dfe-c1d7-4167-bbfe-8ff6d4cf3921 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] 5eae4661-4212-4932-ad1c-cee6275b859c/5eae4661-4212-4932-ad1c-cee6275b859c.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1284.650383] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d9f27c2b-a04b-4283-bc79-05692f1bdf6f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1284.657616] env[65726]: DEBUG oslo_vmware.api [None req-db162dfe-c1d7-4167-bbfe-8ff6d4cf3921 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Waiting for the task: (returnval){ [ 1284.657616] env[65726]: value = "task-5117053" [ 1284.657616] env[65726]: _type = "Task" [ 1284.657616] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1284.665581] env[65726]: DEBUG oslo_vmware.api [None req-db162dfe-c1d7-4167-bbfe-8ff6d4cf3921 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Task: {'id': task-5117053, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1285.018222] env[65726]: DEBUG oslo_vmware.api [None req-e97dfc9c-52c4-40d1-b545-310ad12437a8 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': task-5117052, 'name': Rename_Task, 'duration_secs': 0.146161} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1285.018694] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-e97dfc9c-52c4-40d1-b545-310ad12437a8 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 58562386-ef51-44ef-869c-a3102a6aafad] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1285.018936] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5a8547c3-a8fa-4c2c-98a3-0d5d887d8a14 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1285.028121] env[65726]: DEBUG oslo_vmware.api [None req-e97dfc9c-52c4-40d1-b545-310ad12437a8 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Waiting for the task: (returnval){ [ 1285.028121] env[65726]: value = "task-5117054" [ 1285.028121] env[65726]: _type = "Task" [ 1285.028121] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1285.037936] env[65726]: DEBUG oslo_vmware.api [None req-e97dfc9c-52c4-40d1-b545-310ad12437a8 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': task-5117054, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1285.169180] env[65726]: DEBUG oslo_vmware.api [None req-db162dfe-c1d7-4167-bbfe-8ff6d4cf3921 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Task: {'id': task-5117053, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1285.539446] env[65726]: DEBUG oslo_vmware.api [None req-e97dfc9c-52c4-40d1-b545-310ad12437a8 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': task-5117054, 'name': PowerOnVM_Task} progress is 78%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1285.667927] env[65726]: DEBUG oslo_vmware.api [None req-db162dfe-c1d7-4167-bbfe-8ff6d4cf3921 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Task: {'id': task-5117053, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.544195} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1285.668230] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-db162dfe-c1d7-4167-bbfe-8ff6d4cf3921 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] 5eae4661-4212-4932-ad1c-cee6275b859c/5eae4661-4212-4932-ad1c-cee6275b859c.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1285.668475] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-db162dfe-c1d7-4167-bbfe-8ff6d4cf3921 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 5eae4661-4212-4932-ad1c-cee6275b859c] Extending root virtual disk to 1048576 {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1285.668710] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-adbacadc-7443-426d-9a38-40f7b8a2940e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1285.677335] env[65726]: DEBUG oslo_vmware.api [None req-db162dfe-c1d7-4167-bbfe-8ff6d4cf3921 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Waiting for the task: (returnval){ [ 1285.677335] env[65726]: value = "task-5117055" [ 1285.677335] env[65726]: _type = "Task" [ 1285.677335] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1285.687111] env[65726]: DEBUG oslo_vmware.api [None req-db162dfe-c1d7-4167-bbfe-8ff6d4cf3921 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Task: {'id': task-5117055, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1286.039602] env[65726]: DEBUG oslo_vmware.api [None req-e97dfc9c-52c4-40d1-b545-310ad12437a8 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': task-5117054, 'name': PowerOnVM_Task, 'duration_secs': 0.895485} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1286.040016] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-e97dfc9c-52c4-40d1-b545-310ad12437a8 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 58562386-ef51-44ef-869c-a3102a6aafad] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1286.040163] env[65726]: INFO nova.compute.manager [None req-e97dfc9c-52c4-40d1-b545-310ad12437a8 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 58562386-ef51-44ef-869c-a3102a6aafad] Took 8.84 seconds to spawn the instance on the hypervisor. [ 1286.040376] env[65726]: DEBUG nova.compute.manager [None req-e97dfc9c-52c4-40d1-b545-310ad12437a8 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 58562386-ef51-44ef-869c-a3102a6aafad] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1286.041380] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa809c04-38fb-4714-aac4-f72eb7ce480f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1286.187366] env[65726]: DEBUG oslo_vmware.api [None req-db162dfe-c1d7-4167-bbfe-8ff6d4cf3921 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Task: {'id': task-5117055, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.182984} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1286.187630] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-db162dfe-c1d7-4167-bbfe-8ff6d4cf3921 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 5eae4661-4212-4932-ad1c-cee6275b859c] Extended root virtual disk {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1286.188418] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-569a4fdf-1713-4502-b986-fe45d95bd6a6 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1286.210103] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-db162dfe-c1d7-4167-bbfe-8ff6d4cf3921 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 5eae4661-4212-4932-ad1c-cee6275b859c] Reconfiguring VM instance instance-00000078 to attach disk [datastore1] 5eae4661-4212-4932-ad1c-cee6275b859c/5eae4661-4212-4932-ad1c-cee6275b859c.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1286.210364] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0bd6d634-5cae-4b93-91ab-e778c47048b7 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1286.230218] env[65726]: DEBUG oslo_vmware.api [None req-db162dfe-c1d7-4167-bbfe-8ff6d4cf3921 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Waiting for the task: (returnval){ [ 1286.230218] env[65726]: value = "task-5117056" [ 1286.230218] env[65726]: _type = "Task" [ 1286.230218] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1286.238145] env[65726]: DEBUG oslo_vmware.api [None req-db162dfe-c1d7-4167-bbfe-8ff6d4cf3921 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Task: {'id': task-5117056, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1286.561911] env[65726]: INFO nova.compute.manager [None req-e97dfc9c-52c4-40d1-b545-310ad12437a8 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 58562386-ef51-44ef-869c-a3102a6aafad] Took 15.67 seconds to build instance. [ 1286.740867] env[65726]: DEBUG oslo_vmware.api [None req-db162dfe-c1d7-4167-bbfe-8ff6d4cf3921 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Task: {'id': task-5117056, 'name': ReconfigVM_Task, 'duration_secs': 0.270166} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1286.741349] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-db162dfe-c1d7-4167-bbfe-8ff6d4cf3921 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 5eae4661-4212-4932-ad1c-cee6275b859c] Reconfigured VM instance instance-00000078 to attach disk [datastore1] 5eae4661-4212-4932-ad1c-cee6275b859c/5eae4661-4212-4932-ad1c-cee6275b859c.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1286.741812] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-aa51a857-1a78-4f50-b03c-75b07c04990e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1286.751024] env[65726]: DEBUG oslo_vmware.api [None req-db162dfe-c1d7-4167-bbfe-8ff6d4cf3921 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Waiting for the task: (returnval){ [ 1286.751024] env[65726]: value = "task-5117057" [ 1286.751024] env[65726]: _type = "Task" [ 1286.751024] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1286.760126] env[65726]: DEBUG oslo_vmware.api [None req-db162dfe-c1d7-4167-bbfe-8ff6d4cf3921 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Task: {'id': task-5117057, 'name': Rename_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1287.065231] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e97dfc9c-52c4-40d1-b545-310ad12437a8 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Lock "58562386-ef51-44ef-869c-a3102a6aafad" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.178s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1287.261687] env[65726]: DEBUG oslo_vmware.api [None req-db162dfe-c1d7-4167-bbfe-8ff6d4cf3921 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Task: {'id': task-5117057, 'name': Rename_Task, 'duration_secs': 0.148593} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1287.261981] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-db162dfe-c1d7-4167-bbfe-8ff6d4cf3921 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 5eae4661-4212-4932-ad1c-cee6275b859c] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1287.262304] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-464c737b-00f3-4c30-998c-7e4d85530d51 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1287.269509] env[65726]: DEBUG oslo_vmware.api [None req-db162dfe-c1d7-4167-bbfe-8ff6d4cf3921 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Waiting for the task: (returnval){ [ 1287.269509] env[65726]: value = "task-5117058" [ 1287.269509] env[65726]: _type = "Task" [ 1287.269509] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1287.277872] env[65726]: DEBUG oslo_vmware.api [None req-db162dfe-c1d7-4167-bbfe-8ff6d4cf3921 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Task: {'id': task-5117058, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1287.382299] env[65726]: DEBUG nova.compute.manager [req-4cfd0059-0ed6-494c-85a8-9cb58934d5a1 req-ee223b32-d240-456a-a59f-137ce3b124ba service nova] [instance: 58562386-ef51-44ef-869c-a3102a6aafad] Received event network-changed-dfc3c1e9-c8e6-41f5-aed1-e1c15c7f9ba6 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1287.382299] env[65726]: DEBUG nova.compute.manager [req-4cfd0059-0ed6-494c-85a8-9cb58934d5a1 req-ee223b32-d240-456a-a59f-137ce3b124ba service nova] [instance: 58562386-ef51-44ef-869c-a3102a6aafad] Refreshing instance network info cache due to event network-changed-dfc3c1e9-c8e6-41f5-aed1-e1c15c7f9ba6. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 1287.382403] env[65726]: DEBUG oslo_concurrency.lockutils [req-4cfd0059-0ed6-494c-85a8-9cb58934d5a1 req-ee223b32-d240-456a-a59f-137ce3b124ba service nova] Acquiring lock "refresh_cache-58562386-ef51-44ef-869c-a3102a6aafad" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1287.382698] env[65726]: DEBUG oslo_concurrency.lockutils [req-4cfd0059-0ed6-494c-85a8-9cb58934d5a1 req-ee223b32-d240-456a-a59f-137ce3b124ba service nova] Acquired lock "refresh_cache-58562386-ef51-44ef-869c-a3102a6aafad" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1287.382885] env[65726]: DEBUG nova.network.neutron [req-4cfd0059-0ed6-494c-85a8-9cb58934d5a1 req-ee223b32-d240-456a-a59f-137ce3b124ba service nova] [instance: 58562386-ef51-44ef-869c-a3102a6aafad] Refreshing network info cache for port dfc3c1e9-c8e6-41f5-aed1-e1c15c7f9ba6 {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 1287.780815] env[65726]: DEBUG oslo_vmware.api [None req-db162dfe-c1d7-4167-bbfe-8ff6d4cf3921 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Task: {'id': task-5117058, 'name': PowerOnVM_Task, 'duration_secs': 0.493897} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1287.781178] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-db162dfe-c1d7-4167-bbfe-8ff6d4cf3921 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 5eae4661-4212-4932-ad1c-cee6275b859c] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1287.781440] env[65726]: INFO nova.compute.manager [None req-db162dfe-c1d7-4167-bbfe-8ff6d4cf3921 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 5eae4661-4212-4932-ad1c-cee6275b859c] Took 7.21 seconds to spawn the instance on the hypervisor. [ 1287.781663] env[65726]: DEBUG nova.compute.manager [None req-db162dfe-c1d7-4167-bbfe-8ff6d4cf3921 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 5eae4661-4212-4932-ad1c-cee6275b859c] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1287.782486] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ab495f0-328e-4614-a09f-1faa81d07702 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1287.885463] env[65726]: WARNING openstack [req-4cfd0059-0ed6-494c-85a8-9cb58934d5a1 req-ee223b32-d240-456a-a59f-137ce3b124ba service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1287.885874] env[65726]: WARNING openstack [req-4cfd0059-0ed6-494c-85a8-9cb58934d5a1 req-ee223b32-d240-456a-a59f-137ce3b124ba service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1288.009182] env[65726]: WARNING openstack [req-4cfd0059-0ed6-494c-85a8-9cb58934d5a1 req-ee223b32-d240-456a-a59f-137ce3b124ba service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1288.009323] env[65726]: WARNING openstack [req-4cfd0059-0ed6-494c-85a8-9cb58934d5a1 req-ee223b32-d240-456a-a59f-137ce3b124ba service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1288.075344] env[65726]: WARNING openstack [req-4cfd0059-0ed6-494c-85a8-9cb58934d5a1 req-ee223b32-d240-456a-a59f-137ce3b124ba service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1288.075778] env[65726]: WARNING openstack [req-4cfd0059-0ed6-494c-85a8-9cb58934d5a1 req-ee223b32-d240-456a-a59f-137ce3b124ba service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1288.156599] env[65726]: DEBUG nova.network.neutron [req-4cfd0059-0ed6-494c-85a8-9cb58934d5a1 req-ee223b32-d240-456a-a59f-137ce3b124ba service nova] [instance: 58562386-ef51-44ef-869c-a3102a6aafad] Updated VIF entry in instance network info cache for port dfc3c1e9-c8e6-41f5-aed1-e1c15c7f9ba6. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 1288.156960] env[65726]: DEBUG nova.network.neutron [req-4cfd0059-0ed6-494c-85a8-9cb58934d5a1 req-ee223b32-d240-456a-a59f-137ce3b124ba service nova] [instance: 58562386-ef51-44ef-869c-a3102a6aafad] Updating instance_info_cache with network_info: [{"id": "dfc3c1e9-c8e6-41f5-aed1-e1c15c7f9ba6", "address": "fa:16:3e:63:63:d0", "network": {"id": "336fe51f-971a-447d-8e68-505640e0db1b", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1030314703-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.183", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6b56f93ba0e14521921484ac2c785c43", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8ee1c76d-1a61-4546-85cb-d4bd3c1b35ef", "external-id": "nsx-vlan-transportzone-161", "segmentation_id": 161, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdfc3c1e9-c8", "ovs_interfaceid": "dfc3c1e9-c8e6-41f5-aed1-e1c15c7f9ba6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1288.299839] env[65726]: INFO nova.compute.manager [None req-db162dfe-c1d7-4167-bbfe-8ff6d4cf3921 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 5eae4661-4212-4932-ad1c-cee6275b859c] Took 11.95 seconds to build instance. [ 1288.659702] env[65726]: DEBUG oslo_concurrency.lockutils [req-4cfd0059-0ed6-494c-85a8-9cb58934d5a1 req-ee223b32-d240-456a-a59f-137ce3b124ba service nova] Releasing lock "refresh_cache-58562386-ef51-44ef-869c-a3102a6aafad" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1288.782942] env[65726]: DEBUG oslo_concurrency.lockutils [None req-909f68b4-6cd2-4d68-93b8-7226cd7e01ec tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Acquiring lock "5eae4661-4212-4932-ad1c-cee6275b859c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1288.802268] env[65726]: DEBUG oslo_concurrency.lockutils [None req-db162dfe-c1d7-4167-bbfe-8ff6d4cf3921 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Lock "5eae4661-4212-4932-ad1c-cee6275b859c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.460s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1288.802648] env[65726]: DEBUG oslo_concurrency.lockutils [None req-909f68b4-6cd2-4d68-93b8-7226cd7e01ec tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Lock "5eae4661-4212-4932-ad1c-cee6275b859c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.020s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1288.802866] env[65726]: DEBUG oslo_concurrency.lockutils [None req-909f68b4-6cd2-4d68-93b8-7226cd7e01ec tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Acquiring lock "5eae4661-4212-4932-ad1c-cee6275b859c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1288.803104] env[65726]: DEBUG oslo_concurrency.lockutils [None req-909f68b4-6cd2-4d68-93b8-7226cd7e01ec tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Lock "5eae4661-4212-4932-ad1c-cee6275b859c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1288.803370] env[65726]: DEBUG oslo_concurrency.lockutils [None req-909f68b4-6cd2-4d68-93b8-7226cd7e01ec tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Lock "5eae4661-4212-4932-ad1c-cee6275b859c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1288.805707] env[65726]: INFO nova.compute.manager [None req-909f68b4-6cd2-4d68-93b8-7226cd7e01ec tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 5eae4661-4212-4932-ad1c-cee6275b859c] Terminating instance [ 1289.310228] env[65726]: DEBUG nova.compute.manager [None req-909f68b4-6cd2-4d68-93b8-7226cd7e01ec tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 5eae4661-4212-4932-ad1c-cee6275b859c] Start destroying the instance on the hypervisor. {{(pid=65726) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1289.310651] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-909f68b4-6cd2-4d68-93b8-7226cd7e01ec tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 5eae4661-4212-4932-ad1c-cee6275b859c] Destroying instance {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1289.311783] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8269e0cd-a6b4-44b8-a564-640eba8771dc {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1289.320101] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-909f68b4-6cd2-4d68-93b8-7226cd7e01ec tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 5eae4661-4212-4932-ad1c-cee6275b859c] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1289.320350] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c913c47a-19be-4f97-9ee5-b7f104ba4d58 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1289.327207] env[65726]: DEBUG oslo_vmware.api [None req-909f68b4-6cd2-4d68-93b8-7226cd7e01ec tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Waiting for the task: (returnval){ [ 1289.327207] env[65726]: value = "task-5117059" [ 1289.327207] env[65726]: _type = "Task" [ 1289.327207] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1289.336956] env[65726]: DEBUG oslo_vmware.api [None req-909f68b4-6cd2-4d68-93b8-7226cd7e01ec tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Task: {'id': task-5117059, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1289.836765] env[65726]: DEBUG oslo_vmware.api [None req-909f68b4-6cd2-4d68-93b8-7226cd7e01ec tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Task: {'id': task-5117059, 'name': PowerOffVM_Task, 'duration_secs': 0.203956} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1289.837320] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-909f68b4-6cd2-4d68-93b8-7226cd7e01ec tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 5eae4661-4212-4932-ad1c-cee6275b859c] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1289.837320] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-909f68b4-6cd2-4d68-93b8-7226cd7e01ec tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 5eae4661-4212-4932-ad1c-cee6275b859c] Unregistering the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1289.837610] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-77301846-423a-401c-ad13-f8c94cf9a0d6 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1289.908981] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-909f68b4-6cd2-4d68-93b8-7226cd7e01ec tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 5eae4661-4212-4932-ad1c-cee6275b859c] Unregistered the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1289.909321] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-909f68b4-6cd2-4d68-93b8-7226cd7e01ec tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 5eae4661-4212-4932-ad1c-cee6275b859c] Deleting contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1289.909505] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-909f68b4-6cd2-4d68-93b8-7226cd7e01ec tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Deleting the datastore file [datastore1] 5eae4661-4212-4932-ad1c-cee6275b859c {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1289.909771] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5a659ecb-0f2e-4237-b61a-41177c59f2a4 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1289.916985] env[65726]: DEBUG oslo_vmware.api [None req-909f68b4-6cd2-4d68-93b8-7226cd7e01ec tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Waiting for the task: (returnval){ [ 1289.916985] env[65726]: value = "task-5117061" [ 1289.916985] env[65726]: _type = "Task" [ 1289.916985] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1289.925462] env[65726]: DEBUG oslo_vmware.api [None req-909f68b4-6cd2-4d68-93b8-7226cd7e01ec tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Task: {'id': task-5117061, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1290.427438] env[65726]: DEBUG oslo_vmware.api [None req-909f68b4-6cd2-4d68-93b8-7226cd7e01ec tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Task: {'id': task-5117061, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.157923} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1290.427848] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-909f68b4-6cd2-4d68-93b8-7226cd7e01ec tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Deleted the datastore file {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1290.427898] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-909f68b4-6cd2-4d68-93b8-7226cd7e01ec tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 5eae4661-4212-4932-ad1c-cee6275b859c] Deleted contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1290.428091] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-909f68b4-6cd2-4d68-93b8-7226cd7e01ec tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 5eae4661-4212-4932-ad1c-cee6275b859c] Instance destroyed {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1290.428274] env[65726]: INFO nova.compute.manager [None req-909f68b4-6cd2-4d68-93b8-7226cd7e01ec tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 5eae4661-4212-4932-ad1c-cee6275b859c] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1290.428517] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-909f68b4-6cd2-4d68-93b8-7226cd7e01ec tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1290.428711] env[65726]: DEBUG nova.compute.manager [-] [instance: 5eae4661-4212-4932-ad1c-cee6275b859c] Deallocating network for instance {{(pid=65726) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1290.428823] env[65726]: DEBUG nova.network.neutron [-] [instance: 5eae4661-4212-4932-ad1c-cee6275b859c] deallocate_for_instance() {{(pid=65726) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 1290.429355] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1290.429615] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1290.465174] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1290.465632] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1290.694170] env[65726]: DEBUG nova.compute.manager [req-9a6b6a27-c97f-4a1e-a437-2be1c11998d8 req-4cf7b70d-5cf5-46b8-ba41-69ac8549ae37 service nova] [instance: 5eae4661-4212-4932-ad1c-cee6275b859c] Received event network-vif-deleted-c79ae8e4-e7f1-425a-a5b8-a203aaa7917a {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1290.694382] env[65726]: INFO nova.compute.manager [req-9a6b6a27-c97f-4a1e-a437-2be1c11998d8 req-4cf7b70d-5cf5-46b8-ba41-69ac8549ae37 service nova] [instance: 5eae4661-4212-4932-ad1c-cee6275b859c] Neutron deleted interface c79ae8e4-e7f1-425a-a5b8-a203aaa7917a; detaching it from the instance and deleting it from the info cache [ 1290.694554] env[65726]: DEBUG nova.network.neutron [req-9a6b6a27-c97f-4a1e-a437-2be1c11998d8 req-4cf7b70d-5cf5-46b8-ba41-69ac8549ae37 service nova] [instance: 5eae4661-4212-4932-ad1c-cee6275b859c] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1291.169523] env[65726]: DEBUG nova.network.neutron [-] [instance: 5eae4661-4212-4932-ad1c-cee6275b859c] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1291.197491] env[65726]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f398a912-2521-4187-aced-04b3b831776b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1291.207257] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-926bf208-b71a-4668-9445-b641f69b7265 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1291.237367] env[65726]: DEBUG nova.compute.manager [req-9a6b6a27-c97f-4a1e-a437-2be1c11998d8 req-4cf7b70d-5cf5-46b8-ba41-69ac8549ae37 service nova] [instance: 5eae4661-4212-4932-ad1c-cee6275b859c] Detach interface failed, port_id=c79ae8e4-e7f1-425a-a5b8-a203aaa7917a, reason: Instance 5eae4661-4212-4932-ad1c-cee6275b859c could not be found. {{(pid=65726) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11669}} [ 1291.672058] env[65726]: INFO nova.compute.manager [-] [instance: 5eae4661-4212-4932-ad1c-cee6275b859c] Took 1.24 seconds to deallocate network for instance. [ 1292.179823] env[65726]: DEBUG oslo_concurrency.lockutils [None req-909f68b4-6cd2-4d68-93b8-7226cd7e01ec tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1292.180108] env[65726]: DEBUG oslo_concurrency.lockutils [None req-909f68b4-6cd2-4d68-93b8-7226cd7e01ec tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1292.180322] env[65726]: DEBUG nova.objects.instance [None req-909f68b4-6cd2-4d68-93b8-7226cd7e01ec tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Lazy-loading 'resources' on Instance uuid 5eae4661-4212-4932-ad1c-cee6275b859c {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1292.759590] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cafe5cb8-db61-473d-9051-fbb79477290f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1292.768395] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eddfa3d8-586a-4c67-8d45-768341c2818c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1292.808644] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9deaa266-d9e6-468e-b999-286b46c0fff3 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1292.817461] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-993e26f5-f0ff-4c99-aff4-9ac9525397c1 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1292.831585] env[65726]: DEBUG nova.compute.provider_tree [None req-909f68b4-6cd2-4d68-93b8-7226cd7e01ec tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1293.334711] env[65726]: DEBUG nova.scheduler.client.report [None req-909f68b4-6cd2-4d68-93b8-7226cd7e01ec tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1293.840873] env[65726]: DEBUG oslo_concurrency.lockutils [None req-909f68b4-6cd2-4d68-93b8-7226cd7e01ec tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.660s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1293.862172] env[65726]: INFO nova.scheduler.client.report [None req-909f68b4-6cd2-4d68-93b8-7226cd7e01ec tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Deleted allocations for instance 5eae4661-4212-4932-ad1c-cee6275b859c [ 1294.370855] env[65726]: DEBUG oslo_concurrency.lockutils [None req-909f68b4-6cd2-4d68-93b8-7226cd7e01ec tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Lock "5eae4661-4212-4932-ad1c-cee6275b859c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.568s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1296.003697] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e193c1fa-1e8f-4972-8bbd-a2663f5d008b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Acquiring lock "435b0908-fc57-4cd1-b008-48da19734835" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1296.004076] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e193c1fa-1e8f-4972-8bbd-a2663f5d008b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Lock "435b0908-fc57-4cd1-b008-48da19734835" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1296.507566] env[65726]: DEBUG nova.compute.manager [None req-e193c1fa-1e8f-4972-8bbd-a2663f5d008b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 435b0908-fc57-4cd1-b008-48da19734835] Starting instance... {{(pid=65726) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 1297.027317] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e193c1fa-1e8f-4972-8bbd-a2663f5d008b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1297.027605] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e193c1fa-1e8f-4972-8bbd-a2663f5d008b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1297.029487] env[65726]: INFO nova.compute.claims [None req-e193c1fa-1e8f-4972-8bbd-a2663f5d008b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 435b0908-fc57-4cd1-b008-48da19734835] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1298.116343] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd53c4af-3bea-42d3-b56d-143819227fe4 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1298.126703] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d477d9dc-faa1-49c2-ab68-3414b3e303a3 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1298.157555] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46e72a48-9fbd-4b84-82ae-943b14d03ac1 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1298.166195] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30247e60-4b06-455b-ad81-f0c91c603698 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1298.180594] env[65726]: DEBUG nova.compute.provider_tree [None req-e193c1fa-1e8f-4972-8bbd-a2663f5d008b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1298.684392] env[65726]: DEBUG nova.scheduler.client.report [None req-e193c1fa-1e8f-4972-8bbd-a2663f5d008b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1299.189851] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e193c1fa-1e8f-4972-8bbd-a2663f5d008b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.162s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1299.190333] env[65726]: DEBUG nova.compute.manager [None req-e193c1fa-1e8f-4972-8bbd-a2663f5d008b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 435b0908-fc57-4cd1-b008-48da19734835] Start building networks asynchronously for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1299.695058] env[65726]: DEBUG nova.compute.utils [None req-e193c1fa-1e8f-4972-8bbd-a2663f5d008b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Using /dev/sd instead of None {{(pid=65726) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1299.696594] env[65726]: DEBUG nova.compute.manager [None req-e193c1fa-1e8f-4972-8bbd-a2663f5d008b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 435b0908-fc57-4cd1-b008-48da19734835] Allocating IP information in the background. {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1299.696771] env[65726]: DEBUG nova.network.neutron [None req-e193c1fa-1e8f-4972-8bbd-a2663f5d008b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 435b0908-fc57-4cd1-b008-48da19734835] allocate_for_instance() {{(pid=65726) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 1299.697179] env[65726]: WARNING neutronclient.v2_0.client [None req-e193c1fa-1e8f-4972-8bbd-a2663f5d008b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1299.697408] env[65726]: WARNING neutronclient.v2_0.client [None req-e193c1fa-1e8f-4972-8bbd-a2663f5d008b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1299.699480] env[65726]: WARNING openstack [None req-e193c1fa-1e8f-4972-8bbd-a2663f5d008b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1299.699480] env[65726]: WARNING openstack [None req-e193c1fa-1e8f-4972-8bbd-a2663f5d008b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1299.770723] env[65726]: DEBUG nova.policy [None req-e193c1fa-1e8f-4972-8bbd-a2663f5d008b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ccb3513b5b4f453f842139ec1815c4af', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f3f799aa94f64e8cb26d93f6124efd81', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65726) authorize /opt/stack/nova/nova/policy.py:192}} [ 1300.021060] env[65726]: DEBUG nova.network.neutron [None req-e193c1fa-1e8f-4972-8bbd-a2663f5d008b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 435b0908-fc57-4cd1-b008-48da19734835] Successfully created port: 8a444f22-6684-414c-8a1c-c7f16f24d1c7 {{(pid=65726) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1300.210184] env[65726]: DEBUG nova.compute.manager [None req-e193c1fa-1e8f-4972-8bbd-a2663f5d008b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 435b0908-fc57-4cd1-b008-48da19734835] Start building block device mappings for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1301.223582] env[65726]: DEBUG nova.compute.manager [None req-e193c1fa-1e8f-4972-8bbd-a2663f5d008b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 435b0908-fc57-4cd1-b008-48da19734835] Start spawning the instance on the hypervisor. {{(pid=65726) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1301.251680] env[65726]: DEBUG nova.virt.hardware [None req-e193c1fa-1e8f-4972-8bbd-a2663f5d008b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T19:28:34Z,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1539f06fa3534e90acbc3a60c4b8bd3f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T19:28:35Z,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1301.251927] env[65726]: DEBUG nova.virt.hardware [None req-e193c1fa-1e8f-4972-8bbd-a2663f5d008b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1301.252095] env[65726]: DEBUG nova.virt.hardware [None req-e193c1fa-1e8f-4972-8bbd-a2663f5d008b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1301.252280] env[65726]: DEBUG nova.virt.hardware [None req-e193c1fa-1e8f-4972-8bbd-a2663f5d008b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1301.252420] env[65726]: DEBUG nova.virt.hardware [None req-e193c1fa-1e8f-4972-8bbd-a2663f5d008b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1301.252682] env[65726]: DEBUG nova.virt.hardware [None req-e193c1fa-1e8f-4972-8bbd-a2663f5d008b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1301.252915] env[65726]: DEBUG nova.virt.hardware [None req-e193c1fa-1e8f-4972-8bbd-a2663f5d008b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1301.253094] env[65726]: DEBUG nova.virt.hardware [None req-e193c1fa-1e8f-4972-8bbd-a2663f5d008b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1301.253268] env[65726]: DEBUG nova.virt.hardware [None req-e193c1fa-1e8f-4972-8bbd-a2663f5d008b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1301.253439] env[65726]: DEBUG nova.virt.hardware [None req-e193c1fa-1e8f-4972-8bbd-a2663f5d008b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1301.253612] env[65726]: DEBUG nova.virt.hardware [None req-e193c1fa-1e8f-4972-8bbd-a2663f5d008b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1301.254515] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6206226-cab1-4346-818e-1c18d3d014e5 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1301.263285] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39776cc2-954b-4d5a-8550-f58ce4020f09 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1301.416054] env[65726]: DEBUG nova.compute.manager [req-1a9ca796-37ab-4ae0-af7c-d45bf4a550a2 req-16a91117-e2f6-4d58-b44b-094a9f070d2b service nova] [instance: 435b0908-fc57-4cd1-b008-48da19734835] Received event network-vif-plugged-8a444f22-6684-414c-8a1c-c7f16f24d1c7 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1301.416301] env[65726]: DEBUG oslo_concurrency.lockutils [req-1a9ca796-37ab-4ae0-af7c-d45bf4a550a2 req-16a91117-e2f6-4d58-b44b-094a9f070d2b service nova] Acquiring lock "435b0908-fc57-4cd1-b008-48da19734835-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1301.416523] env[65726]: DEBUG oslo_concurrency.lockutils [req-1a9ca796-37ab-4ae0-af7c-d45bf4a550a2 req-16a91117-e2f6-4d58-b44b-094a9f070d2b service nova] Lock "435b0908-fc57-4cd1-b008-48da19734835-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1301.416689] env[65726]: DEBUG oslo_concurrency.lockutils [req-1a9ca796-37ab-4ae0-af7c-d45bf4a550a2 req-16a91117-e2f6-4d58-b44b-094a9f070d2b service nova] Lock "435b0908-fc57-4cd1-b008-48da19734835-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1301.416925] env[65726]: DEBUG nova.compute.manager [req-1a9ca796-37ab-4ae0-af7c-d45bf4a550a2 req-16a91117-e2f6-4d58-b44b-094a9f070d2b service nova] [instance: 435b0908-fc57-4cd1-b008-48da19734835] No waiting events found dispatching network-vif-plugged-8a444f22-6684-414c-8a1c-c7f16f24d1c7 {{(pid=65726) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1301.417170] env[65726]: WARNING nova.compute.manager [req-1a9ca796-37ab-4ae0-af7c-d45bf4a550a2 req-16a91117-e2f6-4d58-b44b-094a9f070d2b service nova] [instance: 435b0908-fc57-4cd1-b008-48da19734835] Received unexpected event network-vif-plugged-8a444f22-6684-414c-8a1c-c7f16f24d1c7 for instance with vm_state building and task_state spawning. [ 1301.507812] env[65726]: DEBUG nova.network.neutron [None req-e193c1fa-1e8f-4972-8bbd-a2663f5d008b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 435b0908-fc57-4cd1-b008-48da19734835] Successfully updated port: 8a444f22-6684-414c-8a1c-c7f16f24d1c7 {{(pid=65726) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 1302.011105] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e193c1fa-1e8f-4972-8bbd-a2663f5d008b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Acquiring lock "refresh_cache-435b0908-fc57-4cd1-b008-48da19734835" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1302.011322] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e193c1fa-1e8f-4972-8bbd-a2663f5d008b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Acquired lock "refresh_cache-435b0908-fc57-4cd1-b008-48da19734835" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1302.011503] env[65726]: DEBUG nova.network.neutron [None req-e193c1fa-1e8f-4972-8bbd-a2663f5d008b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 435b0908-fc57-4cd1-b008-48da19734835] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1302.514473] env[65726]: WARNING openstack [None req-e193c1fa-1e8f-4972-8bbd-a2663f5d008b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1302.514931] env[65726]: WARNING openstack [None req-e193c1fa-1e8f-4972-8bbd-a2663f5d008b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1302.519900] env[65726]: DEBUG nova.network.neutron [None req-e193c1fa-1e8f-4972-8bbd-a2663f5d008b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 435b0908-fc57-4cd1-b008-48da19734835] Instance cache missing network info. {{(pid=65726) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 1302.568916] env[65726]: WARNING openstack [None req-e193c1fa-1e8f-4972-8bbd-a2663f5d008b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1302.569304] env[65726]: WARNING openstack [None req-e193c1fa-1e8f-4972-8bbd-a2663f5d008b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1302.629467] env[65726]: WARNING openstack [None req-e193c1fa-1e8f-4972-8bbd-a2663f5d008b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1302.629853] env[65726]: WARNING openstack [None req-e193c1fa-1e8f-4972-8bbd-a2663f5d008b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1302.708017] env[65726]: DEBUG nova.network.neutron [None req-e193c1fa-1e8f-4972-8bbd-a2663f5d008b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 435b0908-fc57-4cd1-b008-48da19734835] Updating instance_info_cache with network_info: [{"id": "8a444f22-6684-414c-8a1c-c7f16f24d1c7", "address": "fa:16:3e:d3:a1:75", "network": {"id": "69642329-e986-48fa-accf-1f296c08116a", "bridge": "br-int", "label": "tempest-ServersTestJSON-1786538788-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f3f799aa94f64e8cb26d93f6124efd81", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "816c6e38-e200-4544-8c5b-9fc3e16c5761", "external-id": "nsx-vlan-transportzone-195", "segmentation_id": 195, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8a444f22-66", "ovs_interfaceid": "8a444f22-6684-414c-8a1c-c7f16f24d1c7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1303.210969] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e193c1fa-1e8f-4972-8bbd-a2663f5d008b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Releasing lock "refresh_cache-435b0908-fc57-4cd1-b008-48da19734835" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1303.211478] env[65726]: DEBUG nova.compute.manager [None req-e193c1fa-1e8f-4972-8bbd-a2663f5d008b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 435b0908-fc57-4cd1-b008-48da19734835] Instance network_info: |[{"id": "8a444f22-6684-414c-8a1c-c7f16f24d1c7", "address": "fa:16:3e:d3:a1:75", "network": {"id": "69642329-e986-48fa-accf-1f296c08116a", "bridge": "br-int", "label": "tempest-ServersTestJSON-1786538788-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f3f799aa94f64e8cb26d93f6124efd81", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "816c6e38-e200-4544-8c5b-9fc3e16c5761", "external-id": "nsx-vlan-transportzone-195", "segmentation_id": 195, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8a444f22-66", "ovs_interfaceid": "8a444f22-6684-414c-8a1c-c7f16f24d1c7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1303.212052] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-e193c1fa-1e8f-4972-8bbd-a2663f5d008b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 435b0908-fc57-4cd1-b008-48da19734835] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d3:a1:75', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '816c6e38-e200-4544-8c5b-9fc3e16c5761', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '8a444f22-6684-414c-8a1c-c7f16f24d1c7', 'vif_model': 'vmxnet3'}] {{(pid=65726) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1303.219998] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-e193c1fa-1e8f-4972-8bbd-a2663f5d008b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1303.220271] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 435b0908-fc57-4cd1-b008-48da19734835] Creating VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1303.220545] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a0a134c6-a545-42fc-a02c-24f79a3bd347 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1303.242859] env[65726]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1303.242859] env[65726]: value = "task-5117062" [ 1303.242859] env[65726]: _type = "Task" [ 1303.242859] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1303.251469] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5117062, 'name': CreateVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1303.446776] env[65726]: DEBUG nova.compute.manager [req-46f9e188-3574-4177-ba46-0da1e67ee19e req-0a595ce0-c29a-4e24-b346-ad90a4454b0c service nova] [instance: 435b0908-fc57-4cd1-b008-48da19734835] Received event network-changed-8a444f22-6684-414c-8a1c-c7f16f24d1c7 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1303.446991] env[65726]: DEBUG nova.compute.manager [req-46f9e188-3574-4177-ba46-0da1e67ee19e req-0a595ce0-c29a-4e24-b346-ad90a4454b0c service nova] [instance: 435b0908-fc57-4cd1-b008-48da19734835] Refreshing instance network info cache due to event network-changed-8a444f22-6684-414c-8a1c-c7f16f24d1c7. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 1303.447253] env[65726]: DEBUG oslo_concurrency.lockutils [req-46f9e188-3574-4177-ba46-0da1e67ee19e req-0a595ce0-c29a-4e24-b346-ad90a4454b0c service nova] Acquiring lock "refresh_cache-435b0908-fc57-4cd1-b008-48da19734835" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1303.447503] env[65726]: DEBUG oslo_concurrency.lockutils [req-46f9e188-3574-4177-ba46-0da1e67ee19e req-0a595ce0-c29a-4e24-b346-ad90a4454b0c service nova] Acquired lock "refresh_cache-435b0908-fc57-4cd1-b008-48da19734835" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1303.447687] env[65726]: DEBUG nova.network.neutron [req-46f9e188-3574-4177-ba46-0da1e67ee19e req-0a595ce0-c29a-4e24-b346-ad90a4454b0c service nova] [instance: 435b0908-fc57-4cd1-b008-48da19734835] Refreshing network info cache for port 8a444f22-6684-414c-8a1c-c7f16f24d1c7 {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 1303.752488] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5117062, 'name': CreateVM_Task, 'duration_secs': 0.327353} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1303.752939] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 435b0908-fc57-4cd1-b008-48da19734835] Created VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1303.753453] env[65726]: WARNING openstack [None req-e193c1fa-1e8f-4972-8bbd-a2663f5d008b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1303.753823] env[65726]: WARNING openstack [None req-e193c1fa-1e8f-4972-8bbd-a2663f5d008b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1303.758797] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e193c1fa-1e8f-4972-8bbd-a2663f5d008b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1303.758942] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e193c1fa-1e8f-4972-8bbd-a2663f5d008b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1303.759279] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e193c1fa-1e8f-4972-8bbd-a2663f5d008b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1303.759540] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d4849012-9a98-46e0-ac4b-4eedd27651d8 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1303.764369] env[65726]: DEBUG oslo_vmware.api [None req-e193c1fa-1e8f-4972-8bbd-a2663f5d008b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Waiting for the task: (returnval){ [ 1303.764369] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52a4ab6e-0682-f09d-f4a7-08e532258216" [ 1303.764369] env[65726]: _type = "Task" [ 1303.764369] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1303.774201] env[65726]: DEBUG oslo_vmware.api [None req-e193c1fa-1e8f-4972-8bbd-a2663f5d008b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52a4ab6e-0682-f09d-f4a7-08e532258216, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1303.950476] env[65726]: WARNING openstack [req-46f9e188-3574-4177-ba46-0da1e67ee19e req-0a595ce0-c29a-4e24-b346-ad90a4454b0c service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1303.950859] env[65726]: WARNING openstack [req-46f9e188-3574-4177-ba46-0da1e67ee19e req-0a595ce0-c29a-4e24-b346-ad90a4454b0c service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1304.042370] env[65726]: WARNING openstack [req-46f9e188-3574-4177-ba46-0da1e67ee19e req-0a595ce0-c29a-4e24-b346-ad90a4454b0c service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1304.042919] env[65726]: WARNING openstack [req-46f9e188-3574-4177-ba46-0da1e67ee19e req-0a595ce0-c29a-4e24-b346-ad90a4454b0c service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1304.101383] env[65726]: WARNING openstack [req-46f9e188-3574-4177-ba46-0da1e67ee19e req-0a595ce0-c29a-4e24-b346-ad90a4454b0c service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1304.101802] env[65726]: WARNING openstack [req-46f9e188-3574-4177-ba46-0da1e67ee19e req-0a595ce0-c29a-4e24-b346-ad90a4454b0c service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1304.177345] env[65726]: DEBUG nova.network.neutron [req-46f9e188-3574-4177-ba46-0da1e67ee19e req-0a595ce0-c29a-4e24-b346-ad90a4454b0c service nova] [instance: 435b0908-fc57-4cd1-b008-48da19734835] Updated VIF entry in instance network info cache for port 8a444f22-6684-414c-8a1c-c7f16f24d1c7. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 1304.177708] env[65726]: DEBUG nova.network.neutron [req-46f9e188-3574-4177-ba46-0da1e67ee19e req-0a595ce0-c29a-4e24-b346-ad90a4454b0c service nova] [instance: 435b0908-fc57-4cd1-b008-48da19734835] Updating instance_info_cache with network_info: [{"id": "8a444f22-6684-414c-8a1c-c7f16f24d1c7", "address": "fa:16:3e:d3:a1:75", "network": {"id": "69642329-e986-48fa-accf-1f296c08116a", "bridge": "br-int", "label": "tempest-ServersTestJSON-1786538788-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f3f799aa94f64e8cb26d93f6124efd81", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "816c6e38-e200-4544-8c5b-9fc3e16c5761", "external-id": "nsx-vlan-transportzone-195", "segmentation_id": 195, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8a444f22-66", "ovs_interfaceid": "8a444f22-6684-414c-8a1c-c7f16f24d1c7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1304.275402] env[65726]: DEBUG oslo_vmware.api [None req-e193c1fa-1e8f-4972-8bbd-a2663f5d008b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52a4ab6e-0682-f09d-f4a7-08e532258216, 'name': SearchDatastore_Task, 'duration_secs': 0.015994} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1304.275692] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e193c1fa-1e8f-4972-8bbd-a2663f5d008b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1304.275951] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-e193c1fa-1e8f-4972-8bbd-a2663f5d008b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 435b0908-fc57-4cd1-b008-48da19734835] Processing image b52362a3-ee8a-4cbf-b06f-513b0cc8f95c {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1304.276207] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e193c1fa-1e8f-4972-8bbd-a2663f5d008b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1304.276349] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e193c1fa-1e8f-4972-8bbd-a2663f5d008b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1304.276526] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-e193c1fa-1e8f-4972-8bbd-a2663f5d008b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1304.276785] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4dce01d7-d7e5-4518-ac97-091d626b2d8e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1304.285806] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-e193c1fa-1e8f-4972-8bbd-a2663f5d008b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1304.285995] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-e193c1fa-1e8f-4972-8bbd-a2663f5d008b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65726) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1304.286692] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0749dd61-8b5b-495e-b7e5-d6fb05f1f358 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1304.291532] env[65726]: DEBUG oslo_vmware.api [None req-e193c1fa-1e8f-4972-8bbd-a2663f5d008b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Waiting for the task: (returnval){ [ 1304.291532] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5299748c-d4c2-a318-09df-4e6c7be8a32d" [ 1304.291532] env[65726]: _type = "Task" [ 1304.291532] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1304.299180] env[65726]: DEBUG oslo_vmware.api [None req-e193c1fa-1e8f-4972-8bbd-a2663f5d008b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5299748c-d4c2-a318-09df-4e6c7be8a32d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1304.679982] env[65726]: DEBUG oslo_concurrency.lockutils [req-46f9e188-3574-4177-ba46-0da1e67ee19e req-0a595ce0-c29a-4e24-b346-ad90a4454b0c service nova] Releasing lock "refresh_cache-435b0908-fc57-4cd1-b008-48da19734835" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1304.802345] env[65726]: DEBUG oslo_vmware.api [None req-e193c1fa-1e8f-4972-8bbd-a2663f5d008b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5299748c-d4c2-a318-09df-4e6c7be8a32d, 'name': SearchDatastore_Task, 'duration_secs': 0.009365} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1304.803158] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3ac3e5b1-a165-4601-a0af-fe1ccc6b606c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1304.808670] env[65726]: DEBUG oslo_vmware.api [None req-e193c1fa-1e8f-4972-8bbd-a2663f5d008b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Waiting for the task: (returnval){ [ 1304.808670] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52b85fe7-2740-a5e4-bcb0-5bf544b20f76" [ 1304.808670] env[65726]: _type = "Task" [ 1304.808670] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1304.819056] env[65726]: DEBUG oslo_vmware.api [None req-e193c1fa-1e8f-4972-8bbd-a2663f5d008b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52b85fe7-2740-a5e4-bcb0-5bf544b20f76, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1305.319215] env[65726]: DEBUG oslo_vmware.api [None req-e193c1fa-1e8f-4972-8bbd-a2663f5d008b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52b85fe7-2740-a5e4-bcb0-5bf544b20f76, 'name': SearchDatastore_Task, 'duration_secs': 0.010622} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1305.319493] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e193c1fa-1e8f-4972-8bbd-a2663f5d008b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1305.319751] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-e193c1fa-1e8f-4972-8bbd-a2663f5d008b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] 435b0908-fc57-4cd1-b008-48da19734835/435b0908-fc57-4cd1-b008-48da19734835.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1305.320024] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-502ff5b9-6307-4ad1-894a-3ee78de876e1 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1305.327512] env[65726]: DEBUG oslo_vmware.api [None req-e193c1fa-1e8f-4972-8bbd-a2663f5d008b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Waiting for the task: (returnval){ [ 1305.327512] env[65726]: value = "task-5117063" [ 1305.327512] env[65726]: _type = "Task" [ 1305.327512] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1305.335541] env[65726]: DEBUG oslo_vmware.api [None req-e193c1fa-1e8f-4972-8bbd-a2663f5d008b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Task: {'id': task-5117063, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1305.837993] env[65726]: DEBUG oslo_vmware.api [None req-e193c1fa-1e8f-4972-8bbd-a2663f5d008b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Task: {'id': task-5117063, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.455146} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1305.838398] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-e193c1fa-1e8f-4972-8bbd-a2663f5d008b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] 435b0908-fc57-4cd1-b008-48da19734835/435b0908-fc57-4cd1-b008-48da19734835.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1305.838489] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-e193c1fa-1e8f-4972-8bbd-a2663f5d008b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 435b0908-fc57-4cd1-b008-48da19734835] Extending root virtual disk to 1048576 {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1305.838730] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c57785d5-9430-4026-8d5d-1c9ad0fb01cd {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1305.846098] env[65726]: DEBUG oslo_vmware.api [None req-e193c1fa-1e8f-4972-8bbd-a2663f5d008b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Waiting for the task: (returnval){ [ 1305.846098] env[65726]: value = "task-5117064" [ 1305.846098] env[65726]: _type = "Task" [ 1305.846098] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1305.854497] env[65726]: DEBUG oslo_vmware.api [None req-e193c1fa-1e8f-4972-8bbd-a2663f5d008b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Task: {'id': task-5117064, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1306.356368] env[65726]: DEBUG oslo_vmware.api [None req-e193c1fa-1e8f-4972-8bbd-a2663f5d008b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Task: {'id': task-5117064, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.063681} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1306.356681] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-e193c1fa-1e8f-4972-8bbd-a2663f5d008b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 435b0908-fc57-4cd1-b008-48da19734835] Extended root virtual disk {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1306.357427] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8d41421-54f1-4bdf-b912-cdf26d20a337 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1306.381646] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-e193c1fa-1e8f-4972-8bbd-a2663f5d008b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 435b0908-fc57-4cd1-b008-48da19734835] Reconfiguring VM instance instance-00000079 to attach disk [datastore1] 435b0908-fc57-4cd1-b008-48da19734835/435b0908-fc57-4cd1-b008-48da19734835.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1306.381980] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-50711caf-a534-4017-b326-2b07281e7001 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1306.403082] env[65726]: DEBUG oslo_vmware.api [None req-e193c1fa-1e8f-4972-8bbd-a2663f5d008b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Waiting for the task: (returnval){ [ 1306.403082] env[65726]: value = "task-5117065" [ 1306.403082] env[65726]: _type = "Task" [ 1306.403082] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1306.413067] env[65726]: DEBUG oslo_vmware.api [None req-e193c1fa-1e8f-4972-8bbd-a2663f5d008b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Task: {'id': task-5117065, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1306.913970] env[65726]: DEBUG oslo_vmware.api [None req-e193c1fa-1e8f-4972-8bbd-a2663f5d008b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Task: {'id': task-5117065, 'name': ReconfigVM_Task, 'duration_secs': 0.290821} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1306.915596] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-e193c1fa-1e8f-4972-8bbd-a2663f5d008b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 435b0908-fc57-4cd1-b008-48da19734835] Reconfigured VM instance instance-00000079 to attach disk [datastore1] 435b0908-fc57-4cd1-b008-48da19734835/435b0908-fc57-4cd1-b008-48da19734835.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1306.915757] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-acdd57ed-2ffa-4e65-83b8-fe9afa0ea49f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1306.923683] env[65726]: DEBUG oslo_vmware.api [None req-e193c1fa-1e8f-4972-8bbd-a2663f5d008b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Waiting for the task: (returnval){ [ 1306.923683] env[65726]: value = "task-5117066" [ 1306.923683] env[65726]: _type = "Task" [ 1306.923683] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1306.932940] env[65726]: DEBUG oslo_vmware.api [None req-e193c1fa-1e8f-4972-8bbd-a2663f5d008b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Task: {'id': task-5117066, 'name': Rename_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1307.433651] env[65726]: DEBUG oslo_vmware.api [None req-e193c1fa-1e8f-4972-8bbd-a2663f5d008b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Task: {'id': task-5117066, 'name': Rename_Task, 'duration_secs': 0.312135} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1307.433943] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-e193c1fa-1e8f-4972-8bbd-a2663f5d008b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 435b0908-fc57-4cd1-b008-48da19734835] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1307.434277] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-edd7e177-9b90-4836-9f9d-d5464c6c62ac {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1307.441297] env[65726]: DEBUG oslo_vmware.api [None req-e193c1fa-1e8f-4972-8bbd-a2663f5d008b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Waiting for the task: (returnval){ [ 1307.441297] env[65726]: value = "task-5117067" [ 1307.441297] env[65726]: _type = "Task" [ 1307.441297] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1307.450226] env[65726]: DEBUG oslo_vmware.api [None req-e193c1fa-1e8f-4972-8bbd-a2663f5d008b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Task: {'id': task-5117067, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1307.951362] env[65726]: DEBUG oslo_vmware.api [None req-e193c1fa-1e8f-4972-8bbd-a2663f5d008b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Task: {'id': task-5117067, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1308.129254] env[65726]: DEBUG oslo_concurrency.lockutils [None req-391ebcce-82a1-4069-8d93-7109cb45c105 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Acquiring lock "1ad4c4f9-1b02-420c-ac89-5e3390ddf773" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1308.129630] env[65726]: DEBUG oslo_concurrency.lockutils [None req-391ebcce-82a1-4069-8d93-7109cb45c105 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Lock "1ad4c4f9-1b02-420c-ac89-5e3390ddf773" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1308.455960] env[65726]: DEBUG oslo_vmware.api [None req-e193c1fa-1e8f-4972-8bbd-a2663f5d008b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Task: {'id': task-5117067, 'name': PowerOnVM_Task, 'duration_secs': 0.706164} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1308.456471] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-e193c1fa-1e8f-4972-8bbd-a2663f5d008b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 435b0908-fc57-4cd1-b008-48da19734835] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1308.456803] env[65726]: INFO nova.compute.manager [None req-e193c1fa-1e8f-4972-8bbd-a2663f5d008b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 435b0908-fc57-4cd1-b008-48da19734835] Took 7.23 seconds to spawn the instance on the hypervisor. [ 1308.457108] env[65726]: DEBUG nova.compute.manager [None req-e193c1fa-1e8f-4972-8bbd-a2663f5d008b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 435b0908-fc57-4cd1-b008-48da19734835] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1308.458672] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2cd2cdc7-6fd9-4030-8326-78d7438459cb {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1308.633061] env[65726]: DEBUG nova.compute.manager [None req-391ebcce-82a1-4069-8d93-7109cb45c105 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 1ad4c4f9-1b02-420c-ac89-5e3390ddf773] Starting instance... {{(pid=65726) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 1308.978562] env[65726]: INFO nova.compute.manager [None req-e193c1fa-1e8f-4972-8bbd-a2663f5d008b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 435b0908-fc57-4cd1-b008-48da19734835] Took 11.97 seconds to build instance. [ 1309.155162] env[65726]: DEBUG oslo_concurrency.lockutils [None req-391ebcce-82a1-4069-8d93-7109cb45c105 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1309.155447] env[65726]: DEBUG oslo_concurrency.lockutils [None req-391ebcce-82a1-4069-8d93-7109cb45c105 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1309.157517] env[65726]: INFO nova.compute.claims [None req-391ebcce-82a1-4069-8d93-7109cb45c105 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 1ad4c4f9-1b02-420c-ac89-5e3390ddf773] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1309.481029] env[65726]: DEBUG oslo_concurrency.lockutils [None req-e193c1fa-1e8f-4972-8bbd-a2663f5d008b tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Lock "435b0908-fc57-4cd1-b008-48da19734835" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.477s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1310.243260] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c733772-90c2-4a9d-bd07-e42afacfd6f6 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1310.251464] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9bd8f68-84c3-4c3b-93f7-810050ea8e36 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1310.282443] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-485c0476-a53c-455a-b829-ae25ee677176 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1310.289977] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd5abb13-4bd1-48f5-8464-ee3278b98e3d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1310.302755] env[65726]: DEBUG nova.compute.provider_tree [None req-391ebcce-82a1-4069-8d93-7109cb45c105 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1310.487973] env[65726]: DEBUG oslo_concurrency.lockutils [None req-f575f5b5-4d91-4cbe-9c54-c1839f3ecffa tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Acquiring lock "435b0908-fc57-4cd1-b008-48da19734835" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1310.488289] env[65726]: DEBUG oslo_concurrency.lockutils [None req-f575f5b5-4d91-4cbe-9c54-c1839f3ecffa tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Lock "435b0908-fc57-4cd1-b008-48da19734835" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1310.488432] env[65726]: DEBUG nova.compute.manager [None req-f575f5b5-4d91-4cbe-9c54-c1839f3ecffa tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 435b0908-fc57-4cd1-b008-48da19734835] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1310.489321] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-edde4ea0-5e7e-47c8-8bb3-011aed200928 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1310.496456] env[65726]: DEBUG nova.compute.manager [None req-f575f5b5-4d91-4cbe-9c54-c1839f3ecffa tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 435b0908-fc57-4cd1-b008-48da19734835] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=65726) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3459}} [ 1310.496986] env[65726]: DEBUG nova.objects.instance [None req-f575f5b5-4d91-4cbe-9c54-c1839f3ecffa tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Lazy-loading 'flavor' on Instance uuid 435b0908-fc57-4cd1-b008-48da19734835 {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1310.806205] env[65726]: DEBUG nova.scheduler.client.report [None req-391ebcce-82a1-4069-8d93-7109cb45c105 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1311.311509] env[65726]: DEBUG oslo_concurrency.lockutils [None req-391ebcce-82a1-4069-8d93-7109cb45c105 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.156s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1311.311975] env[65726]: DEBUG nova.compute.manager [None req-391ebcce-82a1-4069-8d93-7109cb45c105 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 1ad4c4f9-1b02-420c-ac89-5e3390ddf773] Start building networks asynchronously for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1311.508539] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-f575f5b5-4d91-4cbe-9c54-c1839f3ecffa tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 435b0908-fc57-4cd1-b008-48da19734835] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1311.508857] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d7d08ed9-142d-4759-9abb-555c8d328fec {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1311.517111] env[65726]: DEBUG oslo_vmware.api [None req-f575f5b5-4d91-4cbe-9c54-c1839f3ecffa tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Waiting for the task: (returnval){ [ 1311.517111] env[65726]: value = "task-5117068" [ 1311.517111] env[65726]: _type = "Task" [ 1311.517111] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1311.525391] env[65726]: DEBUG oslo_vmware.api [None req-f575f5b5-4d91-4cbe-9c54-c1839f3ecffa tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Task: {'id': task-5117068, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1311.819489] env[65726]: DEBUG nova.compute.utils [None req-391ebcce-82a1-4069-8d93-7109cb45c105 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Using /dev/sd instead of None {{(pid=65726) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1311.820945] env[65726]: DEBUG nova.compute.manager [None req-391ebcce-82a1-4069-8d93-7109cb45c105 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 1ad4c4f9-1b02-420c-ac89-5e3390ddf773] Allocating IP information in the background. {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1311.821157] env[65726]: DEBUG nova.network.neutron [None req-391ebcce-82a1-4069-8d93-7109cb45c105 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 1ad4c4f9-1b02-420c-ac89-5e3390ddf773] allocate_for_instance() {{(pid=65726) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 1311.821511] env[65726]: WARNING neutronclient.v2_0.client [None req-391ebcce-82a1-4069-8d93-7109cb45c105 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1311.821819] env[65726]: WARNING neutronclient.v2_0.client [None req-391ebcce-82a1-4069-8d93-7109cb45c105 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1311.822402] env[65726]: WARNING openstack [None req-391ebcce-82a1-4069-8d93-7109cb45c105 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1311.822741] env[65726]: WARNING openstack [None req-391ebcce-82a1-4069-8d93-7109cb45c105 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1312.028018] env[65726]: DEBUG oslo_vmware.api [None req-f575f5b5-4d91-4cbe-9c54-c1839f3ecffa tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Task: {'id': task-5117068, 'name': PowerOffVM_Task, 'duration_secs': 0.207389} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1312.028484] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-f575f5b5-4d91-4cbe-9c54-c1839f3ecffa tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 435b0908-fc57-4cd1-b008-48da19734835] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1312.028739] env[65726]: DEBUG nova.compute.manager [None req-f575f5b5-4d91-4cbe-9c54-c1839f3ecffa tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 435b0908-fc57-4cd1-b008-48da19734835] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1312.029479] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e45c6564-2a4d-4a8c-8b59-dd7f81af7402 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1312.263558] env[65726]: DEBUG nova.policy [None req-391ebcce-82a1-4069-8d93-7109cb45c105 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5143cc185c7645b48304c8ebc5c854de', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'acd800e0438940bc80ae3c6b672db4ee', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65726) authorize /opt/stack/nova/nova/policy.py:192}} [ 1312.331264] env[65726]: DEBUG nova.compute.manager [None req-391ebcce-82a1-4069-8d93-7109cb45c105 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 1ad4c4f9-1b02-420c-ac89-5e3390ddf773] Start building block device mappings for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1312.513574] env[65726]: DEBUG nova.network.neutron [None req-391ebcce-82a1-4069-8d93-7109cb45c105 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 1ad4c4f9-1b02-420c-ac89-5e3390ddf773] Successfully created port: 3aa1b8d0-3927-41ed-903b-212dc4ad95b6 {{(pid=65726) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1312.541485] env[65726]: DEBUG oslo_concurrency.lockutils [None req-f575f5b5-4d91-4cbe-9c54-c1839f3ecffa tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Lock "435b0908-fc57-4cd1-b008-48da19734835" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.053s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1313.171731] env[65726]: DEBUG oslo_concurrency.lockutils [None req-8d062369-1384-4677-a056-47218cc97eef tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Acquiring lock "435b0908-fc57-4cd1-b008-48da19734835" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1313.171997] env[65726]: DEBUG oslo_concurrency.lockutils [None req-8d062369-1384-4677-a056-47218cc97eef tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Lock "435b0908-fc57-4cd1-b008-48da19734835" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1313.172265] env[65726]: DEBUG oslo_concurrency.lockutils [None req-8d062369-1384-4677-a056-47218cc97eef tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Acquiring lock "435b0908-fc57-4cd1-b008-48da19734835-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1313.172460] env[65726]: DEBUG oslo_concurrency.lockutils [None req-8d062369-1384-4677-a056-47218cc97eef tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Lock "435b0908-fc57-4cd1-b008-48da19734835-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1313.172623] env[65726]: DEBUG oslo_concurrency.lockutils [None req-8d062369-1384-4677-a056-47218cc97eef tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Lock "435b0908-fc57-4cd1-b008-48da19734835-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1313.174907] env[65726]: INFO nova.compute.manager [None req-8d062369-1384-4677-a056-47218cc97eef tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 435b0908-fc57-4cd1-b008-48da19734835] Terminating instance [ 1313.341670] env[65726]: DEBUG nova.compute.manager [None req-391ebcce-82a1-4069-8d93-7109cb45c105 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 1ad4c4f9-1b02-420c-ac89-5e3390ddf773] Start spawning the instance on the hypervisor. {{(pid=65726) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1313.372157] env[65726]: DEBUG nova.virt.hardware [None req-391ebcce-82a1-4069-8d93-7109cb45c105 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T19:28:34Z,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1539f06fa3534e90acbc3a60c4b8bd3f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T19:28:35Z,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1313.372373] env[65726]: DEBUG nova.virt.hardware [None req-391ebcce-82a1-4069-8d93-7109cb45c105 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1313.372522] env[65726]: DEBUG nova.virt.hardware [None req-391ebcce-82a1-4069-8d93-7109cb45c105 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1313.372697] env[65726]: DEBUG nova.virt.hardware [None req-391ebcce-82a1-4069-8d93-7109cb45c105 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1313.372835] env[65726]: DEBUG nova.virt.hardware [None req-391ebcce-82a1-4069-8d93-7109cb45c105 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1313.372980] env[65726]: DEBUG nova.virt.hardware [None req-391ebcce-82a1-4069-8d93-7109cb45c105 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1313.373213] env[65726]: DEBUG nova.virt.hardware [None req-391ebcce-82a1-4069-8d93-7109cb45c105 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1313.373438] env[65726]: DEBUG nova.virt.hardware [None req-391ebcce-82a1-4069-8d93-7109cb45c105 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1313.373632] env[65726]: DEBUG nova.virt.hardware [None req-391ebcce-82a1-4069-8d93-7109cb45c105 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1313.373797] env[65726]: DEBUG nova.virt.hardware [None req-391ebcce-82a1-4069-8d93-7109cb45c105 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1313.373971] env[65726]: DEBUG nova.virt.hardware [None req-391ebcce-82a1-4069-8d93-7109cb45c105 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1313.374953] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b720e8b-1258-486b-8e43-2cbbf013bb99 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1313.384431] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2b31333-10d2-47f9-bf78-d4029ecad399 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1313.678771] env[65726]: DEBUG nova.compute.manager [None req-8d062369-1384-4677-a056-47218cc97eef tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 435b0908-fc57-4cd1-b008-48da19734835] Start destroying the instance on the hypervisor. {{(pid=65726) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1313.678994] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-8d062369-1384-4677-a056-47218cc97eef tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 435b0908-fc57-4cd1-b008-48da19734835] Destroying instance {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1313.679906] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58de85c4-0730-4ffb-8712-5ff84bdf3ab3 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1313.689356] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-8d062369-1384-4677-a056-47218cc97eef tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 435b0908-fc57-4cd1-b008-48da19734835] Unregistering the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1313.689673] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1f7bd810-699a-46c7-9e31-36c304dfc4fd {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1313.760642] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-8d062369-1384-4677-a056-47218cc97eef tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 435b0908-fc57-4cd1-b008-48da19734835] Unregistered the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1313.760911] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-8d062369-1384-4677-a056-47218cc97eef tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 435b0908-fc57-4cd1-b008-48da19734835] Deleting contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1313.761036] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-8d062369-1384-4677-a056-47218cc97eef tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Deleting the datastore file [datastore1] 435b0908-fc57-4cd1-b008-48da19734835 {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1313.761444] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a1562507-8fa0-41e9-bba1-0d1fdb2cdf53 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1313.768632] env[65726]: DEBUG oslo_vmware.api [None req-8d062369-1384-4677-a056-47218cc97eef tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Waiting for the task: (returnval){ [ 1313.768632] env[65726]: value = "task-5117070" [ 1313.768632] env[65726]: _type = "Task" [ 1313.768632] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1313.778288] env[65726]: DEBUG oslo_vmware.api [None req-8d062369-1384-4677-a056-47218cc97eef tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Task: {'id': task-5117070, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1313.960912] env[65726]: DEBUG nova.compute.manager [req-e36c9400-80fc-474f-bf4c-3bf2c357e403 req-d402c54c-8f50-481d-81c4-257d8f7ac43e service nova] [instance: 1ad4c4f9-1b02-420c-ac89-5e3390ddf773] Received event network-vif-plugged-3aa1b8d0-3927-41ed-903b-212dc4ad95b6 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1313.961486] env[65726]: DEBUG oslo_concurrency.lockutils [req-e36c9400-80fc-474f-bf4c-3bf2c357e403 req-d402c54c-8f50-481d-81c4-257d8f7ac43e service nova] Acquiring lock "1ad4c4f9-1b02-420c-ac89-5e3390ddf773-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1313.961789] env[65726]: DEBUG oslo_concurrency.lockutils [req-e36c9400-80fc-474f-bf4c-3bf2c357e403 req-d402c54c-8f50-481d-81c4-257d8f7ac43e service nova] Lock "1ad4c4f9-1b02-420c-ac89-5e3390ddf773-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1313.962090] env[65726]: DEBUG oslo_concurrency.lockutils [req-e36c9400-80fc-474f-bf4c-3bf2c357e403 req-d402c54c-8f50-481d-81c4-257d8f7ac43e service nova] Lock "1ad4c4f9-1b02-420c-ac89-5e3390ddf773-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1313.962299] env[65726]: DEBUG nova.compute.manager [req-e36c9400-80fc-474f-bf4c-3bf2c357e403 req-d402c54c-8f50-481d-81c4-257d8f7ac43e service nova] [instance: 1ad4c4f9-1b02-420c-ac89-5e3390ddf773] No waiting events found dispatching network-vif-plugged-3aa1b8d0-3927-41ed-903b-212dc4ad95b6 {{(pid=65726) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1313.963024] env[65726]: WARNING nova.compute.manager [req-e36c9400-80fc-474f-bf4c-3bf2c357e403 req-d402c54c-8f50-481d-81c4-257d8f7ac43e service nova] [instance: 1ad4c4f9-1b02-420c-ac89-5e3390ddf773] Received unexpected event network-vif-plugged-3aa1b8d0-3927-41ed-903b-212dc4ad95b6 for instance with vm_state building and task_state spawning. [ 1314.060323] env[65726]: DEBUG nova.network.neutron [None req-391ebcce-82a1-4069-8d93-7109cb45c105 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 1ad4c4f9-1b02-420c-ac89-5e3390ddf773] Successfully updated port: 3aa1b8d0-3927-41ed-903b-212dc4ad95b6 {{(pid=65726) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 1314.280067] env[65726]: DEBUG oslo_vmware.api [None req-8d062369-1384-4677-a056-47218cc97eef tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Task: {'id': task-5117070, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.193467} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1314.280067] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-8d062369-1384-4677-a056-47218cc97eef tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Deleted the datastore file {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1314.280067] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-8d062369-1384-4677-a056-47218cc97eef tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 435b0908-fc57-4cd1-b008-48da19734835] Deleted contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1314.280356] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-8d062369-1384-4677-a056-47218cc97eef tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 435b0908-fc57-4cd1-b008-48da19734835] Instance destroyed {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1314.280585] env[65726]: INFO nova.compute.manager [None req-8d062369-1384-4677-a056-47218cc97eef tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 435b0908-fc57-4cd1-b008-48da19734835] Took 0.60 seconds to destroy the instance on the hypervisor. [ 1314.280876] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-8d062369-1384-4677-a056-47218cc97eef tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1314.281115] env[65726]: DEBUG nova.compute.manager [-] [instance: 435b0908-fc57-4cd1-b008-48da19734835] Deallocating network for instance {{(pid=65726) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1314.281245] env[65726]: DEBUG nova.network.neutron [-] [instance: 435b0908-fc57-4cd1-b008-48da19734835] deallocate_for_instance() {{(pid=65726) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 1314.281783] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1314.282090] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1314.319872] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1314.320181] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1314.562962] env[65726]: DEBUG oslo_concurrency.lockutils [None req-391ebcce-82a1-4069-8d93-7109cb45c105 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Acquiring lock "refresh_cache-1ad4c4f9-1b02-420c-ac89-5e3390ddf773" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1314.563240] env[65726]: DEBUG oslo_concurrency.lockutils [None req-391ebcce-82a1-4069-8d93-7109cb45c105 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Acquired lock "refresh_cache-1ad4c4f9-1b02-420c-ac89-5e3390ddf773" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1314.563342] env[65726]: DEBUG nova.network.neutron [None req-391ebcce-82a1-4069-8d93-7109cb45c105 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 1ad4c4f9-1b02-420c-ac89-5e3390ddf773] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1315.047875] env[65726]: DEBUG nova.network.neutron [-] [instance: 435b0908-fc57-4cd1-b008-48da19734835] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1315.066348] env[65726]: WARNING openstack [None req-391ebcce-82a1-4069-8d93-7109cb45c105 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1315.066747] env[65726]: WARNING openstack [None req-391ebcce-82a1-4069-8d93-7109cb45c105 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1315.071718] env[65726]: DEBUG nova.network.neutron [None req-391ebcce-82a1-4069-8d93-7109cb45c105 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 1ad4c4f9-1b02-420c-ac89-5e3390ddf773] Instance cache missing network info. {{(pid=65726) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 1315.121381] env[65726]: WARNING openstack [None req-391ebcce-82a1-4069-8d93-7109cb45c105 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1315.121838] env[65726]: WARNING openstack [None req-391ebcce-82a1-4069-8d93-7109cb45c105 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1315.183606] env[65726]: WARNING openstack [None req-391ebcce-82a1-4069-8d93-7109cb45c105 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1315.183995] env[65726]: WARNING openstack [None req-391ebcce-82a1-4069-8d93-7109cb45c105 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1315.269444] env[65726]: DEBUG nova.network.neutron [None req-391ebcce-82a1-4069-8d93-7109cb45c105 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 1ad4c4f9-1b02-420c-ac89-5e3390ddf773] Updating instance_info_cache with network_info: [{"id": "3aa1b8d0-3927-41ed-903b-212dc4ad95b6", "address": "fa:16:3e:46:e9:71", "network": {"id": "a899453d-5f43-4650-9642-b33f54a4d2eb", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-614861139-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "acd800e0438940bc80ae3c6b672db4ee", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "205fb402-8eaf-4b61-8f57-8f216024179a", "external-id": "nsx-vlan-transportzone-78", "segmentation_id": 78, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3aa1b8d0-39", "ovs_interfaceid": "3aa1b8d0-3927-41ed-903b-212dc4ad95b6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1315.550947] env[65726]: INFO nova.compute.manager [-] [instance: 435b0908-fc57-4cd1-b008-48da19734835] Took 1.27 seconds to deallocate network for instance. [ 1315.772357] env[65726]: DEBUG oslo_concurrency.lockutils [None req-391ebcce-82a1-4069-8d93-7109cb45c105 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Releasing lock "refresh_cache-1ad4c4f9-1b02-420c-ac89-5e3390ddf773" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1315.772743] env[65726]: DEBUG nova.compute.manager [None req-391ebcce-82a1-4069-8d93-7109cb45c105 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 1ad4c4f9-1b02-420c-ac89-5e3390ddf773] Instance network_info: |[{"id": "3aa1b8d0-3927-41ed-903b-212dc4ad95b6", "address": "fa:16:3e:46:e9:71", "network": {"id": "a899453d-5f43-4650-9642-b33f54a4d2eb", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-614861139-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "acd800e0438940bc80ae3c6b672db4ee", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "205fb402-8eaf-4b61-8f57-8f216024179a", "external-id": "nsx-vlan-transportzone-78", "segmentation_id": 78, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3aa1b8d0-39", "ovs_interfaceid": "3aa1b8d0-3927-41ed-903b-212dc4ad95b6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1315.773251] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-391ebcce-82a1-4069-8d93-7109cb45c105 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 1ad4c4f9-1b02-420c-ac89-5e3390ddf773] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:46:e9:71', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '205fb402-8eaf-4b61-8f57-8f216024179a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3aa1b8d0-3927-41ed-903b-212dc4ad95b6', 'vif_model': 'vmxnet3'}] {{(pid=65726) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1315.780776] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-391ebcce-82a1-4069-8d93-7109cb45c105 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1315.780960] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1ad4c4f9-1b02-420c-ac89-5e3390ddf773] Creating VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1315.781200] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2f84c699-a854-4a53-abf0-661c5d45d7aa {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1315.802139] env[65726]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1315.802139] env[65726]: value = "task-5117071" [ 1315.802139] env[65726]: _type = "Task" [ 1315.802139] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1315.810035] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5117071, 'name': CreateVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1315.992597] env[65726]: DEBUG nova.compute.manager [req-dd5c9d6f-1005-4612-96e2-f26e7bc628a7 req-13edb659-3043-451a-b7e1-7000247f9984 service nova] [instance: 1ad4c4f9-1b02-420c-ac89-5e3390ddf773] Received event network-changed-3aa1b8d0-3927-41ed-903b-212dc4ad95b6 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1315.992816] env[65726]: DEBUG nova.compute.manager [req-dd5c9d6f-1005-4612-96e2-f26e7bc628a7 req-13edb659-3043-451a-b7e1-7000247f9984 service nova] [instance: 1ad4c4f9-1b02-420c-ac89-5e3390ddf773] Refreshing instance network info cache due to event network-changed-3aa1b8d0-3927-41ed-903b-212dc4ad95b6. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 1315.993290] env[65726]: DEBUG oslo_concurrency.lockutils [req-dd5c9d6f-1005-4612-96e2-f26e7bc628a7 req-13edb659-3043-451a-b7e1-7000247f9984 service nova] Acquiring lock "refresh_cache-1ad4c4f9-1b02-420c-ac89-5e3390ddf773" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1315.993603] env[65726]: DEBUG oslo_concurrency.lockutils [req-dd5c9d6f-1005-4612-96e2-f26e7bc628a7 req-13edb659-3043-451a-b7e1-7000247f9984 service nova] Acquired lock "refresh_cache-1ad4c4f9-1b02-420c-ac89-5e3390ddf773" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1315.993651] env[65726]: DEBUG nova.network.neutron [req-dd5c9d6f-1005-4612-96e2-f26e7bc628a7 req-13edb659-3043-451a-b7e1-7000247f9984 service nova] [instance: 1ad4c4f9-1b02-420c-ac89-5e3390ddf773] Refreshing network info cache for port 3aa1b8d0-3927-41ed-903b-212dc4ad95b6 {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 1316.057981] env[65726]: DEBUG oslo_concurrency.lockutils [None req-8d062369-1384-4677-a056-47218cc97eef tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1316.058305] env[65726]: DEBUG oslo_concurrency.lockutils [None req-8d062369-1384-4677-a056-47218cc97eef tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1316.058547] env[65726]: DEBUG nova.objects.instance [None req-8d062369-1384-4677-a056-47218cc97eef tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Lazy-loading 'resources' on Instance uuid 435b0908-fc57-4cd1-b008-48da19734835 {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1316.313064] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5117071, 'name': CreateVM_Task, 'duration_secs': 0.314854} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1316.315054] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1ad4c4f9-1b02-420c-ac89-5e3390ddf773] Created VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1316.315054] env[65726]: WARNING openstack [None req-391ebcce-82a1-4069-8d93-7109cb45c105 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1316.315054] env[65726]: WARNING openstack [None req-391ebcce-82a1-4069-8d93-7109cb45c105 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1316.319699] env[65726]: DEBUG oslo_concurrency.lockutils [None req-391ebcce-82a1-4069-8d93-7109cb45c105 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1316.319864] env[65726]: DEBUG oslo_concurrency.lockutils [None req-391ebcce-82a1-4069-8d93-7109cb45c105 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1316.320202] env[65726]: DEBUG oslo_concurrency.lockutils [None req-391ebcce-82a1-4069-8d93-7109cb45c105 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1316.320475] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-839183ad-e1a5-4c59-b1ec-ec208946a506 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1316.325631] env[65726]: DEBUG oslo_vmware.api [None req-391ebcce-82a1-4069-8d93-7109cb45c105 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Waiting for the task: (returnval){ [ 1316.325631] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52d9c9f8-b721-641e-2334-3bef60c64692" [ 1316.325631] env[65726]: _type = "Task" [ 1316.325631] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1316.334767] env[65726]: DEBUG oslo_vmware.api [None req-391ebcce-82a1-4069-8d93-7109cb45c105 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52d9c9f8-b721-641e-2334-3bef60c64692, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1316.498060] env[65726]: WARNING openstack [req-dd5c9d6f-1005-4612-96e2-f26e7bc628a7 req-13edb659-3043-451a-b7e1-7000247f9984 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1316.498060] env[65726]: WARNING openstack [req-dd5c9d6f-1005-4612-96e2-f26e7bc628a7 req-13edb659-3043-451a-b7e1-7000247f9984 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1316.598080] env[65726]: WARNING openstack [req-dd5c9d6f-1005-4612-96e2-f26e7bc628a7 req-13edb659-3043-451a-b7e1-7000247f9984 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1316.598080] env[65726]: WARNING openstack [req-dd5c9d6f-1005-4612-96e2-f26e7bc628a7 req-13edb659-3043-451a-b7e1-7000247f9984 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1316.648791] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fde03679-23ad-4c99-b712-0954690c0b23 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1316.656925] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e3069ad-4530-4fac-bbe3-eefe8e58d808 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1316.690157] env[65726]: WARNING openstack [req-dd5c9d6f-1005-4612-96e2-f26e7bc628a7 req-13edb659-3043-451a-b7e1-7000247f9984 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1316.690641] env[65726]: WARNING openstack [req-dd5c9d6f-1005-4612-96e2-f26e7bc628a7 req-13edb659-3043-451a-b7e1-7000247f9984 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1316.698728] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1851ed53-b53e-4c4c-8f5c-b74dd2ec9fd5 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1316.707370] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f40aacd1-9323-40ab-b831-1c05d37dcead {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1316.721840] env[65726]: DEBUG nova.compute.provider_tree [None req-8d062369-1384-4677-a056-47218cc97eef tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1316.778539] env[65726]: DEBUG nova.network.neutron [req-dd5c9d6f-1005-4612-96e2-f26e7bc628a7 req-13edb659-3043-451a-b7e1-7000247f9984 service nova] [instance: 1ad4c4f9-1b02-420c-ac89-5e3390ddf773] Updated VIF entry in instance network info cache for port 3aa1b8d0-3927-41ed-903b-212dc4ad95b6. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 1316.778908] env[65726]: DEBUG nova.network.neutron [req-dd5c9d6f-1005-4612-96e2-f26e7bc628a7 req-13edb659-3043-451a-b7e1-7000247f9984 service nova] [instance: 1ad4c4f9-1b02-420c-ac89-5e3390ddf773] Updating instance_info_cache with network_info: [{"id": "3aa1b8d0-3927-41ed-903b-212dc4ad95b6", "address": "fa:16:3e:46:e9:71", "network": {"id": "a899453d-5f43-4650-9642-b33f54a4d2eb", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-614861139-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "acd800e0438940bc80ae3c6b672db4ee", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "205fb402-8eaf-4b61-8f57-8f216024179a", "external-id": "nsx-vlan-transportzone-78", "segmentation_id": 78, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3aa1b8d0-39", "ovs_interfaceid": "3aa1b8d0-3927-41ed-903b-212dc4ad95b6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1316.836549] env[65726]: DEBUG oslo_vmware.api [None req-391ebcce-82a1-4069-8d93-7109cb45c105 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52d9c9f8-b721-641e-2334-3bef60c64692, 'name': SearchDatastore_Task, 'duration_secs': 0.013086} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1316.836803] env[65726]: DEBUG oslo_concurrency.lockutils [None req-391ebcce-82a1-4069-8d93-7109cb45c105 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1316.837044] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-391ebcce-82a1-4069-8d93-7109cb45c105 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 1ad4c4f9-1b02-420c-ac89-5e3390ddf773] Processing image b52362a3-ee8a-4cbf-b06f-513b0cc8f95c {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1316.837287] env[65726]: DEBUG oslo_concurrency.lockutils [None req-391ebcce-82a1-4069-8d93-7109cb45c105 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1316.837457] env[65726]: DEBUG oslo_concurrency.lockutils [None req-391ebcce-82a1-4069-8d93-7109cb45c105 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1316.837642] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-391ebcce-82a1-4069-8d93-7109cb45c105 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1316.837910] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-deebb6fd-1bba-49a5-96d4-fd68783a66da {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1316.849941] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-391ebcce-82a1-4069-8d93-7109cb45c105 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1316.850158] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-391ebcce-82a1-4069-8d93-7109cb45c105 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65726) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1316.850843] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e91a1449-ea83-49b7-ba19-fd5af4e222c0 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1316.858356] env[65726]: DEBUG oslo_vmware.api [None req-391ebcce-82a1-4069-8d93-7109cb45c105 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Waiting for the task: (returnval){ [ 1316.858356] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52008384-6aea-fcb7-a84f-f17d218ce971" [ 1316.858356] env[65726]: _type = "Task" [ 1316.858356] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1316.866687] env[65726]: DEBUG oslo_vmware.api [None req-391ebcce-82a1-4069-8d93-7109cb45c105 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52008384-6aea-fcb7-a84f-f17d218ce971, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1317.224797] env[65726]: DEBUG nova.scheduler.client.report [None req-8d062369-1384-4677-a056-47218cc97eef tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1317.282727] env[65726]: DEBUG oslo_concurrency.lockutils [req-dd5c9d6f-1005-4612-96e2-f26e7bc628a7 req-13edb659-3043-451a-b7e1-7000247f9984 service nova] Releasing lock "refresh_cache-1ad4c4f9-1b02-420c-ac89-5e3390ddf773" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1317.282727] env[65726]: DEBUG nova.compute.manager [req-dd5c9d6f-1005-4612-96e2-f26e7bc628a7 req-13edb659-3043-451a-b7e1-7000247f9984 service nova] [instance: 435b0908-fc57-4cd1-b008-48da19734835] Received event network-vif-deleted-8a444f22-6684-414c-8a1c-c7f16f24d1c7 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1317.370038] env[65726]: DEBUG oslo_vmware.api [None req-391ebcce-82a1-4069-8d93-7109cb45c105 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52008384-6aea-fcb7-a84f-f17d218ce971, 'name': SearchDatastore_Task, 'duration_secs': 0.010835} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1317.370465] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d3c9650d-d944-46cb-b07b-9ec307a4c7b8 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1317.375941] env[65726]: DEBUG oslo_vmware.api [None req-391ebcce-82a1-4069-8d93-7109cb45c105 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Waiting for the task: (returnval){ [ 1317.375941] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52eea6d8-d659-6914-ac7c-3c48726a57e5" [ 1317.375941] env[65726]: _type = "Task" [ 1317.375941] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1317.383785] env[65726]: DEBUG oslo_vmware.api [None req-391ebcce-82a1-4069-8d93-7109cb45c105 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52eea6d8-d659-6914-ac7c-3c48726a57e5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1317.729796] env[65726]: DEBUG oslo_concurrency.lockutils [None req-8d062369-1384-4677-a056-47218cc97eef tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.671s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1317.748504] env[65726]: INFO nova.scheduler.client.report [None req-8d062369-1384-4677-a056-47218cc97eef tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Deleted allocations for instance 435b0908-fc57-4cd1-b008-48da19734835 [ 1317.887058] env[65726]: DEBUG oslo_vmware.api [None req-391ebcce-82a1-4069-8d93-7109cb45c105 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52eea6d8-d659-6914-ac7c-3c48726a57e5, 'name': SearchDatastore_Task, 'duration_secs': 0.012276} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1317.887435] env[65726]: DEBUG oslo_concurrency.lockutils [None req-391ebcce-82a1-4069-8d93-7109cb45c105 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1317.887542] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-391ebcce-82a1-4069-8d93-7109cb45c105 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] 1ad4c4f9-1b02-420c-ac89-5e3390ddf773/1ad4c4f9-1b02-420c-ac89-5e3390ddf773.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1317.887918] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-dd9717f9-054e-4145-874f-8a6a2a7fa19e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1317.895549] env[65726]: DEBUG oslo_vmware.api [None req-391ebcce-82a1-4069-8d93-7109cb45c105 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Waiting for the task: (returnval){ [ 1317.895549] env[65726]: value = "task-5117072" [ 1317.895549] env[65726]: _type = "Task" [ 1317.895549] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1317.903577] env[65726]: DEBUG oslo_vmware.api [None req-391ebcce-82a1-4069-8d93-7109cb45c105 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': task-5117072, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1318.256662] env[65726]: DEBUG oslo_concurrency.lockutils [None req-8d062369-1384-4677-a056-47218cc97eef tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Lock "435b0908-fc57-4cd1-b008-48da19734835" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.084s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1318.407068] env[65726]: DEBUG oslo_vmware.api [None req-391ebcce-82a1-4069-8d93-7109cb45c105 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': task-5117072, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1318.907878] env[65726]: DEBUG oslo_vmware.api [None req-391ebcce-82a1-4069-8d93-7109cb45c105 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': task-5117072, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.577944} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1318.908373] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-391ebcce-82a1-4069-8d93-7109cb45c105 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] 1ad4c4f9-1b02-420c-ac89-5e3390ddf773/1ad4c4f9-1b02-420c-ac89-5e3390ddf773.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1318.908373] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-391ebcce-82a1-4069-8d93-7109cb45c105 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 1ad4c4f9-1b02-420c-ac89-5e3390ddf773] Extending root virtual disk to 1048576 {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1318.908613] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4fd2f730-52c4-43b8-8d9d-39109024d4d8 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1318.916147] env[65726]: DEBUG oslo_vmware.api [None req-391ebcce-82a1-4069-8d93-7109cb45c105 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Waiting for the task: (returnval){ [ 1318.916147] env[65726]: value = "task-5117073" [ 1318.916147] env[65726]: _type = "Task" [ 1318.916147] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1318.924039] env[65726]: DEBUG oslo_vmware.api [None req-391ebcce-82a1-4069-8d93-7109cb45c105 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': task-5117073, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1319.050839] env[65726]: DEBUG oslo_concurrency.lockutils [None req-ed09bf28-7012-492d-bc7b-871b04dcece6 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Acquiring lock "9c0599e3-383f-46ab-809b-944cc3a4d206" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1319.051126] env[65726]: DEBUG oslo_concurrency.lockutils [None req-ed09bf28-7012-492d-bc7b-871b04dcece6 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Lock "9c0599e3-383f-46ab-809b-944cc3a4d206" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1319.051344] env[65726]: DEBUG oslo_concurrency.lockutils [None req-ed09bf28-7012-492d-bc7b-871b04dcece6 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Acquiring lock "9c0599e3-383f-46ab-809b-944cc3a4d206-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1319.051521] env[65726]: DEBUG oslo_concurrency.lockutils [None req-ed09bf28-7012-492d-bc7b-871b04dcece6 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Lock "9c0599e3-383f-46ab-809b-944cc3a4d206-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1319.051687] env[65726]: DEBUG oslo_concurrency.lockutils [None req-ed09bf28-7012-492d-bc7b-871b04dcece6 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Lock "9c0599e3-383f-46ab-809b-944cc3a4d206-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1319.054230] env[65726]: INFO nova.compute.manager [None req-ed09bf28-7012-492d-bc7b-871b04dcece6 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 9c0599e3-383f-46ab-809b-944cc3a4d206] Terminating instance [ 1319.426828] env[65726]: DEBUG oslo_vmware.api [None req-391ebcce-82a1-4069-8d93-7109cb45c105 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': task-5117073, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.071216} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1319.427257] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-391ebcce-82a1-4069-8d93-7109cb45c105 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 1ad4c4f9-1b02-420c-ac89-5e3390ddf773] Extended root virtual disk {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1319.428152] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1bc5d751-c88b-4d18-b435-224574bec879 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1319.450911] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-391ebcce-82a1-4069-8d93-7109cb45c105 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 1ad4c4f9-1b02-420c-ac89-5e3390ddf773] Reconfiguring VM instance instance-0000007a to attach disk [datastore1] 1ad4c4f9-1b02-420c-ac89-5e3390ddf773/1ad4c4f9-1b02-420c-ac89-5e3390ddf773.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1319.451218] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d1950393-db47-4fbc-b736-ec12e5efb321 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1319.471658] env[65726]: DEBUG oslo_vmware.api [None req-391ebcce-82a1-4069-8d93-7109cb45c105 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Waiting for the task: (returnval){ [ 1319.471658] env[65726]: value = "task-5117074" [ 1319.471658] env[65726]: _type = "Task" [ 1319.471658] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1319.480381] env[65726]: DEBUG oslo_vmware.api [None req-391ebcce-82a1-4069-8d93-7109cb45c105 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': task-5117074, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1319.558348] env[65726]: DEBUG nova.compute.manager [None req-ed09bf28-7012-492d-bc7b-871b04dcece6 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 9c0599e3-383f-46ab-809b-944cc3a4d206] Start destroying the instance on the hypervisor. {{(pid=65726) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1319.558589] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-ed09bf28-7012-492d-bc7b-871b04dcece6 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 9c0599e3-383f-46ab-809b-944cc3a4d206] Destroying instance {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1319.559501] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58bb219b-ba96-4a1b-b3e4-feb4116ba309 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1319.567670] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-ed09bf28-7012-492d-bc7b-871b04dcece6 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 9c0599e3-383f-46ab-809b-944cc3a4d206] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1319.567947] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1ea2ae4e-7335-4ac5-8508-e16459445b11 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1319.574153] env[65726]: DEBUG oslo_vmware.api [None req-ed09bf28-7012-492d-bc7b-871b04dcece6 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Waiting for the task: (returnval){ [ 1319.574153] env[65726]: value = "task-5117075" [ 1319.574153] env[65726]: _type = "Task" [ 1319.574153] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1319.582405] env[65726]: DEBUG oslo_vmware.api [None req-ed09bf28-7012-492d-bc7b-871b04dcece6 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Task: {'id': task-5117075, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1319.982869] env[65726]: DEBUG oslo_vmware.api [None req-391ebcce-82a1-4069-8d93-7109cb45c105 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': task-5117074, 'name': ReconfigVM_Task, 'duration_secs': 0.320756} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1319.983295] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-391ebcce-82a1-4069-8d93-7109cb45c105 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 1ad4c4f9-1b02-420c-ac89-5e3390ddf773] Reconfigured VM instance instance-0000007a to attach disk [datastore1] 1ad4c4f9-1b02-420c-ac89-5e3390ddf773/1ad4c4f9-1b02-420c-ac89-5e3390ddf773.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1319.983926] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b879b4d2-2f80-46e9-b55c-a2807101ef97 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1319.991347] env[65726]: DEBUG oslo_vmware.api [None req-391ebcce-82a1-4069-8d93-7109cb45c105 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Waiting for the task: (returnval){ [ 1319.991347] env[65726]: value = "task-5117076" [ 1319.991347] env[65726]: _type = "Task" [ 1319.991347] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1319.999502] env[65726]: DEBUG oslo_vmware.api [None req-391ebcce-82a1-4069-8d93-7109cb45c105 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': task-5117076, 'name': Rename_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1320.083672] env[65726]: DEBUG oslo_vmware.api [None req-ed09bf28-7012-492d-bc7b-871b04dcece6 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Task: {'id': task-5117075, 'name': PowerOffVM_Task, 'duration_secs': 0.194931} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1320.083934] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-ed09bf28-7012-492d-bc7b-871b04dcece6 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 9c0599e3-383f-46ab-809b-944cc3a4d206] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1320.084136] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-ed09bf28-7012-492d-bc7b-871b04dcece6 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 9c0599e3-383f-46ab-809b-944cc3a4d206] Unregistering the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1320.084390] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b6ebf0f4-7dc4-468b-b8ae-3c6f0995ccfd {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1320.156791] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-ed09bf28-7012-492d-bc7b-871b04dcece6 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 9c0599e3-383f-46ab-809b-944cc3a4d206] Unregistered the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1320.157080] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-ed09bf28-7012-492d-bc7b-871b04dcece6 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 9c0599e3-383f-46ab-809b-944cc3a4d206] Deleting contents of the VM from datastore datastore2 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1320.157283] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-ed09bf28-7012-492d-bc7b-871b04dcece6 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Deleting the datastore file [datastore2] 9c0599e3-383f-46ab-809b-944cc3a4d206 {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1320.157561] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-72547fed-6001-41a7-a2f1-e48edd231180 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1320.164394] env[65726]: DEBUG oslo_vmware.api [None req-ed09bf28-7012-492d-bc7b-871b04dcece6 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Waiting for the task: (returnval){ [ 1320.164394] env[65726]: value = "task-5117078" [ 1320.164394] env[65726]: _type = "Task" [ 1320.164394] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1320.172522] env[65726]: DEBUG oslo_vmware.api [None req-ed09bf28-7012-492d-bc7b-871b04dcece6 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Task: {'id': task-5117078, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1320.502087] env[65726]: DEBUG oslo_vmware.api [None req-391ebcce-82a1-4069-8d93-7109cb45c105 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': task-5117076, 'name': Rename_Task, 'duration_secs': 0.141871} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1320.502387] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-391ebcce-82a1-4069-8d93-7109cb45c105 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 1ad4c4f9-1b02-420c-ac89-5e3390ddf773] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1320.502645] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8cfde2f2-f337-47f2-ac54-39b23257e8df {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1320.509469] env[65726]: DEBUG oslo_vmware.api [None req-391ebcce-82a1-4069-8d93-7109cb45c105 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Waiting for the task: (returnval){ [ 1320.509469] env[65726]: value = "task-5117079" [ 1320.509469] env[65726]: _type = "Task" [ 1320.509469] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1320.517483] env[65726]: DEBUG oslo_vmware.api [None req-391ebcce-82a1-4069-8d93-7109cb45c105 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': task-5117079, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1320.674268] env[65726]: DEBUG oslo_vmware.api [None req-ed09bf28-7012-492d-bc7b-871b04dcece6 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Task: {'id': task-5117078, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.138242} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1320.674590] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-ed09bf28-7012-492d-bc7b-871b04dcece6 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Deleted the datastore file {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1320.674896] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-ed09bf28-7012-492d-bc7b-871b04dcece6 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 9c0599e3-383f-46ab-809b-944cc3a4d206] Deleted contents of the VM from datastore datastore2 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1320.675174] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-ed09bf28-7012-492d-bc7b-871b04dcece6 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 9c0599e3-383f-46ab-809b-944cc3a4d206] Instance destroyed {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1320.675422] env[65726]: INFO nova.compute.manager [None req-ed09bf28-7012-492d-bc7b-871b04dcece6 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] [instance: 9c0599e3-383f-46ab-809b-944cc3a4d206] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1320.675711] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-ed09bf28-7012-492d-bc7b-871b04dcece6 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1320.675922] env[65726]: DEBUG nova.compute.manager [-] [instance: 9c0599e3-383f-46ab-809b-944cc3a4d206] Deallocating network for instance {{(pid=65726) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1320.676086] env[65726]: DEBUG nova.network.neutron [-] [instance: 9c0599e3-383f-46ab-809b-944cc3a4d206] deallocate_for_instance() {{(pid=65726) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 1320.676644] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1320.676909] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1320.714136] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1320.714434] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1320.966261] env[65726]: DEBUG nova.compute.manager [req-d1ad1eca-0e06-4e5b-929a-60304c16049e req-b595dd82-1fcd-449f-8e5b-3ed1c9d2c7f9 service nova] [instance: 9c0599e3-383f-46ab-809b-944cc3a4d206] Received event network-vif-deleted-c5730368-8477-43fb-ae16-4a4e35a403ab {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1320.966560] env[65726]: INFO nova.compute.manager [req-d1ad1eca-0e06-4e5b-929a-60304c16049e req-b595dd82-1fcd-449f-8e5b-3ed1c9d2c7f9 service nova] [instance: 9c0599e3-383f-46ab-809b-944cc3a4d206] Neutron deleted interface c5730368-8477-43fb-ae16-4a4e35a403ab; detaching it from the instance and deleting it from the info cache [ 1320.966655] env[65726]: DEBUG nova.network.neutron [req-d1ad1eca-0e06-4e5b-929a-60304c16049e req-b595dd82-1fcd-449f-8e5b-3ed1c9d2c7f9 service nova] [instance: 9c0599e3-383f-46ab-809b-944cc3a4d206] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1321.020780] env[65726]: DEBUG oslo_vmware.api [None req-391ebcce-82a1-4069-8d93-7109cb45c105 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': task-5117079, 'name': PowerOnVM_Task, 'duration_secs': 0.441061} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1321.021156] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-391ebcce-82a1-4069-8d93-7109cb45c105 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 1ad4c4f9-1b02-420c-ac89-5e3390ddf773] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1321.021206] env[65726]: INFO nova.compute.manager [None req-391ebcce-82a1-4069-8d93-7109cb45c105 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 1ad4c4f9-1b02-420c-ac89-5e3390ddf773] Took 7.68 seconds to spawn the instance on the hypervisor. [ 1321.021366] env[65726]: DEBUG nova.compute.manager [None req-391ebcce-82a1-4069-8d93-7109cb45c105 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 1ad4c4f9-1b02-420c-ac89-5e3390ddf773] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1321.022225] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-115e369d-56c9-472a-9875-3e18d2f93a22 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1321.449449] env[65726]: DEBUG nova.network.neutron [-] [instance: 9c0599e3-383f-46ab-809b-944cc3a4d206] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1321.469700] env[65726]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-eb6e61b1-461a-4a46-989a-9a2739e302c5 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1321.480081] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-525edf63-f10a-46a6-84df-3ab87d5b1f4c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1321.510443] env[65726]: DEBUG nova.compute.manager [req-d1ad1eca-0e06-4e5b-929a-60304c16049e req-b595dd82-1fcd-449f-8e5b-3ed1c9d2c7f9 service nova] [instance: 9c0599e3-383f-46ab-809b-944cc3a4d206] Detach interface failed, port_id=c5730368-8477-43fb-ae16-4a4e35a403ab, reason: Instance 9c0599e3-383f-46ab-809b-944cc3a4d206 could not be found. {{(pid=65726) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11669}} [ 1321.541419] env[65726]: INFO nova.compute.manager [None req-391ebcce-82a1-4069-8d93-7109cb45c105 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 1ad4c4f9-1b02-420c-ac89-5e3390ddf773] Took 12.40 seconds to build instance. [ 1321.952139] env[65726]: INFO nova.compute.manager [-] [instance: 9c0599e3-383f-46ab-809b-944cc3a4d206] Took 1.28 seconds to deallocate network for instance. [ 1322.043154] env[65726]: DEBUG oslo_concurrency.lockutils [None req-391ebcce-82a1-4069-8d93-7109cb45c105 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Lock "1ad4c4f9-1b02-420c-ac89-5e3390ddf773" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.913s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1322.459229] env[65726]: DEBUG oslo_concurrency.lockutils [None req-ed09bf28-7012-492d-bc7b-871b04dcece6 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1322.459543] env[65726]: DEBUG oslo_concurrency.lockutils [None req-ed09bf28-7012-492d-bc7b-871b04dcece6 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1322.459754] env[65726]: DEBUG nova.objects.instance [None req-ed09bf28-7012-492d-bc7b-871b04dcece6 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Lazy-loading 'resources' on Instance uuid 9c0599e3-383f-46ab-809b-944cc3a4d206 {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1322.846835] env[65726]: DEBUG oslo_concurrency.lockutils [None req-4a8d9522-c6c8-4b1b-88a5-25faf72db3f5 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Acquiring lock "58562386-ef51-44ef-869c-a3102a6aafad" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1322.847488] env[65726]: DEBUG oslo_concurrency.lockutils [None req-4a8d9522-c6c8-4b1b-88a5-25faf72db3f5 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Lock "58562386-ef51-44ef-869c-a3102a6aafad" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.001s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1323.039874] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ac9e2f5-809d-475d-9d31-e63c4c0a9ec2 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1323.049499] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6e150fb-6607-4b47-addf-8fde98330772 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1323.055513] env[65726]: DEBUG nova.compute.manager [req-8a8ad294-6d81-472c-b7e6-0c92ae7ff099 req-c348fa7e-3e0b-4526-8c7e-d57e2e3ee2ac service nova] [instance: 1ad4c4f9-1b02-420c-ac89-5e3390ddf773] Received event network-changed-3aa1b8d0-3927-41ed-903b-212dc4ad95b6 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1323.055737] env[65726]: DEBUG nova.compute.manager [req-8a8ad294-6d81-472c-b7e6-0c92ae7ff099 req-c348fa7e-3e0b-4526-8c7e-d57e2e3ee2ac service nova] [instance: 1ad4c4f9-1b02-420c-ac89-5e3390ddf773] Refreshing instance network info cache due to event network-changed-3aa1b8d0-3927-41ed-903b-212dc4ad95b6. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 1323.057524] env[65726]: DEBUG oslo_concurrency.lockutils [req-8a8ad294-6d81-472c-b7e6-0c92ae7ff099 req-c348fa7e-3e0b-4526-8c7e-d57e2e3ee2ac service nova] Acquiring lock "refresh_cache-1ad4c4f9-1b02-420c-ac89-5e3390ddf773" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1323.057674] env[65726]: DEBUG oslo_concurrency.lockutils [req-8a8ad294-6d81-472c-b7e6-0c92ae7ff099 req-c348fa7e-3e0b-4526-8c7e-d57e2e3ee2ac service nova] Acquired lock "refresh_cache-1ad4c4f9-1b02-420c-ac89-5e3390ddf773" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1323.057863] env[65726]: DEBUG nova.network.neutron [req-8a8ad294-6d81-472c-b7e6-0c92ae7ff099 req-c348fa7e-3e0b-4526-8c7e-d57e2e3ee2ac service nova] [instance: 1ad4c4f9-1b02-420c-ac89-5e3390ddf773] Refreshing network info cache for port 3aa1b8d0-3927-41ed-903b-212dc4ad95b6 {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 1323.086840] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1bff434c-325c-4516-a9c6-3f0a9dae1937 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1323.095634] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2773af41-e320-4b55-bd1a-512d60b72bb3 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1323.110385] env[65726]: DEBUG nova.compute.provider_tree [None req-ed09bf28-7012-492d-bc7b-871b04dcece6 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1323.350739] env[65726]: DEBUG nova.compute.utils [None req-4a8d9522-c6c8-4b1b-88a5-25faf72db3f5 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Using /dev/sd instead of None {{(pid=65726) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1323.560981] env[65726]: WARNING openstack [req-8a8ad294-6d81-472c-b7e6-0c92ae7ff099 req-c348fa7e-3e0b-4526-8c7e-d57e2e3ee2ac service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1323.561424] env[65726]: WARNING openstack [req-8a8ad294-6d81-472c-b7e6-0c92ae7ff099 req-c348fa7e-3e0b-4526-8c7e-d57e2e3ee2ac service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1323.569165] env[65726]: DEBUG oslo_service.periodic_task [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=65726) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1323.613577] env[65726]: DEBUG nova.scheduler.client.report [None req-ed09bf28-7012-492d-bc7b-871b04dcece6 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1323.686985] env[65726]: WARNING openstack [req-8a8ad294-6d81-472c-b7e6-0c92ae7ff099 req-c348fa7e-3e0b-4526-8c7e-d57e2e3ee2ac service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1323.687389] env[65726]: WARNING openstack [req-8a8ad294-6d81-472c-b7e6-0c92ae7ff099 req-c348fa7e-3e0b-4526-8c7e-d57e2e3ee2ac service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1323.750807] env[65726]: WARNING openstack [req-8a8ad294-6d81-472c-b7e6-0c92ae7ff099 req-c348fa7e-3e0b-4526-8c7e-d57e2e3ee2ac service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1323.751235] env[65726]: WARNING openstack [req-8a8ad294-6d81-472c-b7e6-0c92ae7ff099 req-c348fa7e-3e0b-4526-8c7e-d57e2e3ee2ac service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1323.833183] env[65726]: DEBUG nova.network.neutron [req-8a8ad294-6d81-472c-b7e6-0c92ae7ff099 req-c348fa7e-3e0b-4526-8c7e-d57e2e3ee2ac service nova] [instance: 1ad4c4f9-1b02-420c-ac89-5e3390ddf773] Updated VIF entry in instance network info cache for port 3aa1b8d0-3927-41ed-903b-212dc4ad95b6. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 1323.833552] env[65726]: DEBUG nova.network.neutron [req-8a8ad294-6d81-472c-b7e6-0c92ae7ff099 req-c348fa7e-3e0b-4526-8c7e-d57e2e3ee2ac service nova] [instance: 1ad4c4f9-1b02-420c-ac89-5e3390ddf773] Updating instance_info_cache with network_info: [{"id": "3aa1b8d0-3927-41ed-903b-212dc4ad95b6", "address": "fa:16:3e:46:e9:71", "network": {"id": "a899453d-5f43-4650-9642-b33f54a4d2eb", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-614861139-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.209", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "acd800e0438940bc80ae3c6b672db4ee", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "205fb402-8eaf-4b61-8f57-8f216024179a", "external-id": "nsx-vlan-transportzone-78", "segmentation_id": 78, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3aa1b8d0-39", "ovs_interfaceid": "3aa1b8d0-3927-41ed-903b-212dc4ad95b6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1323.853530] env[65726]: DEBUG oslo_concurrency.lockutils [None req-4a8d9522-c6c8-4b1b-88a5-25faf72db3f5 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Lock "58562386-ef51-44ef-869c-a3102a6aafad" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.006s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1324.119023] env[65726]: DEBUG oslo_concurrency.lockutils [None req-ed09bf28-7012-492d-bc7b-871b04dcece6 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.659s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1324.145561] env[65726]: INFO nova.scheduler.client.report [None req-ed09bf28-7012-492d-bc7b-871b04dcece6 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Deleted allocations for instance 9c0599e3-383f-46ab-809b-944cc3a4d206 [ 1324.335845] env[65726]: DEBUG oslo_concurrency.lockutils [req-8a8ad294-6d81-472c-b7e6-0c92ae7ff099 req-c348fa7e-3e0b-4526-8c7e-d57e2e3ee2ac service nova] Releasing lock "refresh_cache-1ad4c4f9-1b02-420c-ac89-5e3390ddf773" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1324.569340] env[65726]: DEBUG oslo_service.periodic_task [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=65726) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1324.569340] env[65726]: DEBUG oslo_service.periodic_task [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=65726) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1324.569666] env[65726]: DEBUG oslo_service.periodic_task [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=65726) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1324.653861] env[65726]: DEBUG oslo_concurrency.lockutils [None req-ed09bf28-7012-492d-bc7b-871b04dcece6 tempest-ServersTestJSON-1488720945 tempest-ServersTestJSON-1488720945-project-member] Lock "9c0599e3-383f-46ab-809b-944cc3a4d206" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.603s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1324.919855] env[65726]: DEBUG oslo_concurrency.lockutils [None req-4a8d9522-c6c8-4b1b-88a5-25faf72db3f5 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Acquiring lock "58562386-ef51-44ef-869c-a3102a6aafad" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1324.920164] env[65726]: DEBUG oslo_concurrency.lockutils [None req-4a8d9522-c6c8-4b1b-88a5-25faf72db3f5 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Lock "58562386-ef51-44ef-869c-a3102a6aafad" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1324.920410] env[65726]: INFO nova.compute.manager [None req-4a8d9522-c6c8-4b1b-88a5-25faf72db3f5 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 58562386-ef51-44ef-869c-a3102a6aafad] Attaching volume 98edee64-3e8c-4e8d-8eec-9f08f9e977f2 to /dev/sdb [ 1324.955502] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44707bac-9c23-4cf2-b19d-e23e081b21a3 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1324.963082] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6d13344-445e-4b33-947e-4fe44d367580 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1324.976277] env[65726]: DEBUG nova.virt.block_device [None req-4a8d9522-c6c8-4b1b-88a5-25faf72db3f5 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 58562386-ef51-44ef-869c-a3102a6aafad] Updating existing volume attachment record: 89bd308b-73d2-44a8-8e4d-9336ab6d8651 {{(pid=65726) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1325.565070] env[65726]: DEBUG oslo_service.periodic_task [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=65726) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1325.569498] env[65726]: DEBUG oslo_service.periodic_task [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=65726) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1325.569658] env[65726]: DEBUG nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=65726) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11251}} [ 1326.569309] env[65726]: DEBUG oslo_service.periodic_task [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=65726) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1328.569497] env[65726]: DEBUG oslo_service.periodic_task [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Running periodic task ComputeManager.update_available_resource {{(pid=65726) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1329.073590] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1329.073834] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1329.073997] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1329.074161] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=65726) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:937}} [ 1329.075054] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d650387c-ebd4-40d1-88ce-b6a5ae35a57d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1329.083565] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-266c391d-6973-4738-8c6c-5e494de27cc2 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1329.097628] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-053b95c0-b5d1-453c-beb9-351bba7a52d5 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1329.104704] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c7b00f9-37cb-46c5-8847-aa0b80bb54f2 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1329.135248] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=179504MB free_disk=95GB free_vcpus=48 pci_devices=None {{(pid=65726) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1136}} [ 1329.135439] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1329.135612] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1330.022424] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-4a8d9522-c6c8-4b1b-88a5-25faf72db3f5 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 58562386-ef51-44ef-869c-a3102a6aafad] Volume attach. Driver type: vmdk {{(pid=65726) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1330.022698] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-4a8d9522-c6c8-4b1b-88a5-25faf72db3f5 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 58562386-ef51-44ef-869c-a3102a6aafad] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-995347', 'volume_id': '98edee64-3e8c-4e8d-8eec-9f08f9e977f2', 'name': 'volume-98edee64-3e8c-4e8d-8eec-9f08f9e977f2', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '58562386-ef51-44ef-869c-a3102a6aafad', 'attached_at': '', 'detached_at': '', 'volume_id': '98edee64-3e8c-4e8d-8eec-9f08f9e977f2', 'serial': '98edee64-3e8c-4e8d-8eec-9f08f9e977f2'} {{(pid=65726) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1330.023590] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57761d99-bde7-4aa4-a63b-82dfcd4c836b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1330.041811] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-217417cd-eac9-43df-b799-49d485d4c8bc {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1330.067498] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-4a8d9522-c6c8-4b1b-88a5-25faf72db3f5 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 58562386-ef51-44ef-869c-a3102a6aafad] Reconfiguring VM instance instance-00000077 to attach disk [datastore2] volume-98edee64-3e8c-4e8d-8eec-9f08f9e977f2/volume-98edee64-3e8c-4e8d-8eec-9f08f9e977f2.vmdk or device None with type thin {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1330.067830] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d09bac5f-8f80-4693-ad74-021b6f5e1237 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1330.086163] env[65726]: DEBUG oslo_vmware.api [None req-4a8d9522-c6c8-4b1b-88a5-25faf72db3f5 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Waiting for the task: (returnval){ [ 1330.086163] env[65726]: value = "task-5117084" [ 1330.086163] env[65726]: _type = "Task" [ 1330.086163] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1330.094943] env[65726]: DEBUG oslo_vmware.api [None req-4a8d9522-c6c8-4b1b-88a5-25faf72db3f5 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': task-5117084, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1330.163771] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Instance e3255f3b-028f-4a0b-b621-873730417936 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65726) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1330.163975] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Instance 55ee97c3-b014-48eb-b41b-fc5aa16bc09e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65726) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1330.164124] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Instance 58562386-ef51-44ef-869c-a3102a6aafad actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65726) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1330.164243] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Instance 1ad4c4f9-1b02-420c-ac89-5e3390ddf773 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65726) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1330.164422] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Total usable vcpus: 48, total allocated vcpus: 4 {{(pid=65726) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1159}} [ 1330.164565] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1280MB phys_disk=100GB used_disk=4GB total_vcpus=48 used_vcpus=4 pci_stats=[] stats={'failed_builds': '0', 'num_instances': '4', 'num_vm_active': '4', 'num_task_None': '4', 'num_os_type_None': '4', 'num_proj_acd800e0438940bc80ae3c6b672db4ee': '2', 'io_workload': '0', 'num_proj_6b56f93ba0e14521921484ac2c785c43': '2'} {{(pid=65726) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1168}} [ 1330.231665] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad4548cf-8cb5-4cdb-acd2-4fbf644124bb {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1330.240522] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9852a97-fd17-4ede-afc9-b7ef386f3914 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1330.272797] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf2c119a-c98b-41e1-9580-e1f10afdb342 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1330.281115] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-088a6dd0-7575-4d46-bf95-6485438cc67a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1330.296626] env[65726]: DEBUG nova.compute.provider_tree [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1330.597145] env[65726]: DEBUG oslo_vmware.api [None req-4a8d9522-c6c8-4b1b-88a5-25faf72db3f5 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': task-5117084, 'name': ReconfigVM_Task, 'duration_secs': 0.361105} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1330.597480] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-4a8d9522-c6c8-4b1b-88a5-25faf72db3f5 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 58562386-ef51-44ef-869c-a3102a6aafad] Reconfigured VM instance instance-00000077 to attach disk [datastore2] volume-98edee64-3e8c-4e8d-8eec-9f08f9e977f2/volume-98edee64-3e8c-4e8d-8eec-9f08f9e977f2.vmdk or device None with type thin {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1330.602925] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7298daf0-1326-4dd3-9740-ff37e24edde1 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1330.618747] env[65726]: DEBUG oslo_vmware.api [None req-4a8d9522-c6c8-4b1b-88a5-25faf72db3f5 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Waiting for the task: (returnval){ [ 1330.618747] env[65726]: value = "task-5117085" [ 1330.618747] env[65726]: _type = "Task" [ 1330.618747] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1330.627417] env[65726]: DEBUG oslo_vmware.api [None req-4a8d9522-c6c8-4b1b-88a5-25faf72db3f5 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': task-5117085, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1330.800012] env[65726]: DEBUG nova.scheduler.client.report [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1331.129408] env[65726]: DEBUG oslo_vmware.api [None req-4a8d9522-c6c8-4b1b-88a5-25faf72db3f5 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': task-5117085, 'name': ReconfigVM_Task, 'duration_secs': 0.168228} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1331.129804] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-4a8d9522-c6c8-4b1b-88a5-25faf72db3f5 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 58562386-ef51-44ef-869c-a3102a6aafad] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-995347', 'volume_id': '98edee64-3e8c-4e8d-8eec-9f08f9e977f2', 'name': 'volume-98edee64-3e8c-4e8d-8eec-9f08f9e977f2', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '58562386-ef51-44ef-869c-a3102a6aafad', 'attached_at': '', 'detached_at': '', 'volume_id': '98edee64-3e8c-4e8d-8eec-9f08f9e977f2', 'serial': '98edee64-3e8c-4e8d-8eec-9f08f9e977f2'} {{(pid=65726) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1331.304788] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=65726) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1097}} [ 1331.305041] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.169s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1332.167776] env[65726]: DEBUG nova.objects.instance [None req-4a8d9522-c6c8-4b1b-88a5-25faf72db3f5 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Lazy-loading 'flavor' on Instance uuid 58562386-ef51-44ef-869c-a3102a6aafad {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1332.566663] env[65726]: INFO nova.compute.manager [None req-cb618542-8b34-4ac8-a12a-8db7c40cdc7f tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 58562386-ef51-44ef-869c-a3102a6aafad] Rebuilding instance [ 1332.612049] env[65726]: DEBUG nova.compute.manager [None req-cb618542-8b34-4ac8-a12a-8db7c40cdc7f tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 58562386-ef51-44ef-869c-a3102a6aafad] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1332.612907] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d702414-5d64-46f3-8bfb-5dada57e3c2c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1332.672588] env[65726]: DEBUG oslo_concurrency.lockutils [None req-4a8d9522-c6c8-4b1b-88a5-25faf72db3f5 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Lock "58562386-ef51-44ef-869c-a3102a6aafad" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.752s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1333.627696] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-cb618542-8b34-4ac8-a12a-8db7c40cdc7f tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 58562386-ef51-44ef-869c-a3102a6aafad] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1333.628131] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-48f54f39-5171-4ab0-865d-f2f28e8f2b31 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1333.636206] env[65726]: DEBUG oslo_vmware.api [None req-cb618542-8b34-4ac8-a12a-8db7c40cdc7f tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Waiting for the task: (returnval){ [ 1333.636206] env[65726]: value = "task-5117086" [ 1333.636206] env[65726]: _type = "Task" [ 1333.636206] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1333.644432] env[65726]: DEBUG oslo_vmware.api [None req-cb618542-8b34-4ac8-a12a-8db7c40cdc7f tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': task-5117086, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1334.146698] env[65726]: DEBUG oslo_vmware.api [None req-cb618542-8b34-4ac8-a12a-8db7c40cdc7f tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': task-5117086, 'name': PowerOffVM_Task, 'duration_secs': 0.215033} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1334.146965] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-cb618542-8b34-4ac8-a12a-8db7c40cdc7f tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 58562386-ef51-44ef-869c-a3102a6aafad] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1334.199629] env[65726]: INFO nova.compute.manager [None req-cb618542-8b34-4ac8-a12a-8db7c40cdc7f tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 58562386-ef51-44ef-869c-a3102a6aafad] Detaching volume 98edee64-3e8c-4e8d-8eec-9f08f9e977f2 [ 1334.231106] env[65726]: INFO nova.virt.block_device [None req-cb618542-8b34-4ac8-a12a-8db7c40cdc7f tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 58562386-ef51-44ef-869c-a3102a6aafad] Attempting to driver detach volume 98edee64-3e8c-4e8d-8eec-9f08f9e977f2 from mountpoint /dev/sdb [ 1334.231349] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-cb618542-8b34-4ac8-a12a-8db7c40cdc7f tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 58562386-ef51-44ef-869c-a3102a6aafad] Volume detach. Driver type: vmdk {{(pid=65726) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1334.231532] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-cb618542-8b34-4ac8-a12a-8db7c40cdc7f tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 58562386-ef51-44ef-869c-a3102a6aafad] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-995347', 'volume_id': '98edee64-3e8c-4e8d-8eec-9f08f9e977f2', 'name': 'volume-98edee64-3e8c-4e8d-8eec-9f08f9e977f2', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '58562386-ef51-44ef-869c-a3102a6aafad', 'attached_at': '', 'detached_at': '', 'volume_id': '98edee64-3e8c-4e8d-8eec-9f08f9e977f2', 'serial': '98edee64-3e8c-4e8d-8eec-9f08f9e977f2'} {{(pid=65726) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1334.232413] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-816ebd74-d394-49ec-991c-523d4ee552ee {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1334.254232] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-890093b3-4532-4daf-bfa1-588c5c74f267 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1334.261712] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c5f9a3c-da25-4a03-b984-786774371afe {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1334.282544] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41cf3a66-9acd-4b5a-88cb-0f3fc2e41211 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1334.297933] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-cb618542-8b34-4ac8-a12a-8db7c40cdc7f tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] The volume has not been displaced from its original location: [datastore2] volume-98edee64-3e8c-4e8d-8eec-9f08f9e977f2/volume-98edee64-3e8c-4e8d-8eec-9f08f9e977f2.vmdk. No consolidation needed. {{(pid=65726) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1334.303274] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-cb618542-8b34-4ac8-a12a-8db7c40cdc7f tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 58562386-ef51-44ef-869c-a3102a6aafad] Reconfiguring VM instance instance-00000077 to detach disk 2001 {{(pid=65726) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1334.304063] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8396fc47-852a-4629-a355-b20a5465e24b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1334.323218] env[65726]: DEBUG oslo_vmware.api [None req-cb618542-8b34-4ac8-a12a-8db7c40cdc7f tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Waiting for the task: (returnval){ [ 1334.323218] env[65726]: value = "task-5117087" [ 1334.323218] env[65726]: _type = "Task" [ 1334.323218] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1334.332576] env[65726]: DEBUG oslo_vmware.api [None req-cb618542-8b34-4ac8-a12a-8db7c40cdc7f tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': task-5117087, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1334.833628] env[65726]: DEBUG oslo_vmware.api [None req-cb618542-8b34-4ac8-a12a-8db7c40cdc7f tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': task-5117087, 'name': ReconfigVM_Task, 'duration_secs': 0.205502} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1334.834013] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-cb618542-8b34-4ac8-a12a-8db7c40cdc7f tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 58562386-ef51-44ef-869c-a3102a6aafad] Reconfigured VM instance instance-00000077 to detach disk 2001 {{(pid=65726) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1334.838643] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ba6e71d1-24f5-4afa-b5ad-4da996d65f64 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1334.854574] env[65726]: DEBUG oslo_vmware.api [None req-cb618542-8b34-4ac8-a12a-8db7c40cdc7f tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Waiting for the task: (returnval){ [ 1334.854574] env[65726]: value = "task-5117088" [ 1334.854574] env[65726]: _type = "Task" [ 1334.854574] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1334.863327] env[65726]: DEBUG oslo_vmware.api [None req-cb618542-8b34-4ac8-a12a-8db7c40cdc7f tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': task-5117088, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1335.367371] env[65726]: DEBUG oslo_vmware.api [None req-cb618542-8b34-4ac8-a12a-8db7c40cdc7f tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': task-5117088, 'name': ReconfigVM_Task, 'duration_secs': 0.150484} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1335.367665] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-cb618542-8b34-4ac8-a12a-8db7c40cdc7f tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 58562386-ef51-44ef-869c-a3102a6aafad] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-995347', 'volume_id': '98edee64-3e8c-4e8d-8eec-9f08f9e977f2', 'name': 'volume-98edee64-3e8c-4e8d-8eec-9f08f9e977f2', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '58562386-ef51-44ef-869c-a3102a6aafad', 'attached_at': '', 'detached_at': '', 'volume_id': '98edee64-3e8c-4e8d-8eec-9f08f9e977f2', 'serial': '98edee64-3e8c-4e8d-8eec-9f08f9e977f2'} {{(pid=65726) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1336.418062] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-cb618542-8b34-4ac8-a12a-8db7c40cdc7f tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 58562386-ef51-44ef-869c-a3102a6aafad] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1336.418507] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0f0e8128-1c12-4354-b6b0-1ab4dfb6e374 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1336.425494] env[65726]: DEBUG oslo_vmware.api [None req-cb618542-8b34-4ac8-a12a-8db7c40cdc7f tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Waiting for the task: (returnval){ [ 1336.425494] env[65726]: value = "task-5117089" [ 1336.425494] env[65726]: _type = "Task" [ 1336.425494] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1336.433475] env[65726]: DEBUG oslo_vmware.api [None req-cb618542-8b34-4ac8-a12a-8db7c40cdc7f tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': task-5117089, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1336.935907] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-cb618542-8b34-4ac8-a12a-8db7c40cdc7f tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 58562386-ef51-44ef-869c-a3102a6aafad] VM already powered off {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1336.936169] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-cb618542-8b34-4ac8-a12a-8db7c40cdc7f tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 58562386-ef51-44ef-869c-a3102a6aafad] Volume detach. Driver type: vmdk {{(pid=65726) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1336.936356] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-cb618542-8b34-4ac8-a12a-8db7c40cdc7f tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 58562386-ef51-44ef-869c-a3102a6aafad] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-995347', 'volume_id': '98edee64-3e8c-4e8d-8eec-9f08f9e977f2', 'name': 'volume-98edee64-3e8c-4e8d-8eec-9f08f9e977f2', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '58562386-ef51-44ef-869c-a3102a6aafad', 'attached_at': '', 'detached_at': '', 'volume_id': '98edee64-3e8c-4e8d-8eec-9f08f9e977f2', 'serial': '98edee64-3e8c-4e8d-8eec-9f08f9e977f2'} {{(pid=65726) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1336.937124] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13a91d77-a1e8-40e6-bb81-5b29e505aae2 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1336.955017] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9ba153f-aab9-487c-a8d4-da669233169a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1336.961957] env[65726]: WARNING nova.virt.vmwareapi.driver [None req-cb618542-8b34-4ac8-a12a-8db7c40cdc7f tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 58562386-ef51-44ef-869c-a3102a6aafad] The volume None does not exist!: nova.exception.DiskNotFound: Unable to find volume [ 1336.962261] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-cb618542-8b34-4ac8-a12a-8db7c40cdc7f tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 58562386-ef51-44ef-869c-a3102a6aafad] Destroying instance {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1336.962989] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-deb869ec-8110-47d5-aeeb-f4a60cff9c16 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1336.969460] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-cb618542-8b34-4ac8-a12a-8db7c40cdc7f tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 58562386-ef51-44ef-869c-a3102a6aafad] Unregistering the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1336.969674] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-30d21b70-c8d6-4fe9-b27a-dbc42803b396 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1337.033888] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-cb618542-8b34-4ac8-a12a-8db7c40cdc7f tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 58562386-ef51-44ef-869c-a3102a6aafad] Unregistered the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1337.034118] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-cb618542-8b34-4ac8-a12a-8db7c40cdc7f tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 58562386-ef51-44ef-869c-a3102a6aafad] Deleting contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1337.034235] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-cb618542-8b34-4ac8-a12a-8db7c40cdc7f tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Deleting the datastore file [datastore1] 58562386-ef51-44ef-869c-a3102a6aafad {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1337.034516] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8e4aa214-7788-4caf-b8bc-93f698255bb1 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1337.042080] env[65726]: DEBUG oslo_vmware.api [None req-cb618542-8b34-4ac8-a12a-8db7c40cdc7f tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Waiting for the task: (returnval){ [ 1337.042080] env[65726]: value = "task-5117091" [ 1337.042080] env[65726]: _type = "Task" [ 1337.042080] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1337.050742] env[65726]: DEBUG oslo_vmware.api [None req-cb618542-8b34-4ac8-a12a-8db7c40cdc7f tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': task-5117091, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1337.552496] env[65726]: DEBUG oslo_vmware.api [None req-cb618542-8b34-4ac8-a12a-8db7c40cdc7f tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': task-5117091, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.152798} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1337.552889] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-cb618542-8b34-4ac8-a12a-8db7c40cdc7f tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Deleted the datastore file {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1337.552937] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-cb618542-8b34-4ac8-a12a-8db7c40cdc7f tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 58562386-ef51-44ef-869c-a3102a6aafad] Deleted contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1337.553098] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-cb618542-8b34-4ac8-a12a-8db7c40cdc7f tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 58562386-ef51-44ef-869c-a3102a6aafad] Instance destroyed {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1338.058198] env[65726]: INFO nova.virt.block_device [None req-cb618542-8b34-4ac8-a12a-8db7c40cdc7f tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 58562386-ef51-44ef-869c-a3102a6aafad] Booting with volume 98edee64-3e8c-4e8d-8eec-9f08f9e977f2 at /dev/sdb [ 1338.092087] env[65726]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-cf5d694f-74af-4151-8ba2-eb339184e389 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1338.103105] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d4c61ee-769e-4239-ba53-24f99e715b45 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1338.132953] env[65726]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6db7dd16-b06c-47ba-9da9-cd4d42a01e13 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1338.142611] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6935b73-7f60-455c-b80b-44b67c64e27c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1338.172859] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-693416c1-e4bb-4e01-af63-b8ac465485e8 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1338.179185] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f259e2c-91d3-4f85-b2f5-3a514c2ecd47 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1338.192775] env[65726]: DEBUG nova.virt.block_device [None req-cb618542-8b34-4ac8-a12a-8db7c40cdc7f tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 58562386-ef51-44ef-869c-a3102a6aafad] Updating existing volume attachment record: 85b436a0-3337-4caa-af57-734f9dc10840 {{(pid=65726) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1340.305483] env[65726]: DEBUG nova.virt.hardware [None req-cb618542-8b34-4ac8-a12a-8db7c40cdc7f tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T19:28:34Z,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1539f06fa3534e90acbc3a60c4b8bd3f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T19:28:35Z,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1340.305806] env[65726]: DEBUG nova.virt.hardware [None req-cb618542-8b34-4ac8-a12a-8db7c40cdc7f tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1340.305926] env[65726]: DEBUG nova.virt.hardware [None req-cb618542-8b34-4ac8-a12a-8db7c40cdc7f tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1340.306117] env[65726]: DEBUG nova.virt.hardware [None req-cb618542-8b34-4ac8-a12a-8db7c40cdc7f tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1340.306267] env[65726]: DEBUG nova.virt.hardware [None req-cb618542-8b34-4ac8-a12a-8db7c40cdc7f tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1340.306407] env[65726]: DEBUG nova.virt.hardware [None req-cb618542-8b34-4ac8-a12a-8db7c40cdc7f tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1340.306621] env[65726]: DEBUG nova.virt.hardware [None req-cb618542-8b34-4ac8-a12a-8db7c40cdc7f tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1340.306808] env[65726]: DEBUG nova.virt.hardware [None req-cb618542-8b34-4ac8-a12a-8db7c40cdc7f tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1340.306982] env[65726]: DEBUG nova.virt.hardware [None req-cb618542-8b34-4ac8-a12a-8db7c40cdc7f tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1340.307158] env[65726]: DEBUG nova.virt.hardware [None req-cb618542-8b34-4ac8-a12a-8db7c40cdc7f tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1340.307329] env[65726]: DEBUG nova.virt.hardware [None req-cb618542-8b34-4ac8-a12a-8db7c40cdc7f tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1340.308240] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4c7a1c5-2b36-4ad8-84d9-d580eb98f11b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1340.316485] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9eb61da8-57ea-44a1-98c5-3f69e8a59db8 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1340.330607] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-cb618542-8b34-4ac8-a12a-8db7c40cdc7f tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 58562386-ef51-44ef-869c-a3102a6aafad] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:63:63:d0', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '8ee1c76d-1a61-4546-85cb-d4bd3c1b35ef', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'dfc3c1e9-c8e6-41f5-aed1-e1c15c7f9ba6', 'vif_model': 'vmxnet3'}] {{(pid=65726) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1340.338200] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-cb618542-8b34-4ac8-a12a-8db7c40cdc7f tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1340.338463] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 58562386-ef51-44ef-869c-a3102a6aafad] Creating VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1340.338681] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-63bf96b5-4444-4f98-933c-9ecd040a9ea1 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1340.358459] env[65726]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1340.358459] env[65726]: value = "task-5117092" [ 1340.358459] env[65726]: _type = "Task" [ 1340.358459] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1340.366046] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5117092, 'name': CreateVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1340.869110] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5117092, 'name': CreateVM_Task, 'duration_secs': 0.288911} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1340.869340] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 58562386-ef51-44ef-869c-a3102a6aafad] Created VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1340.870176] env[65726]: WARNING openstack [None req-cb618542-8b34-4ac8-a12a-8db7c40cdc7f tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1340.870635] env[65726]: WARNING openstack [None req-cb618542-8b34-4ac8-a12a-8db7c40cdc7f tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1340.875980] env[65726]: DEBUG oslo_concurrency.lockutils [None req-cb618542-8b34-4ac8-a12a-8db7c40cdc7f tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1340.876191] env[65726]: DEBUG oslo_concurrency.lockutils [None req-cb618542-8b34-4ac8-a12a-8db7c40cdc7f tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1340.876459] env[65726]: DEBUG oslo_concurrency.lockutils [None req-cb618542-8b34-4ac8-a12a-8db7c40cdc7f tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1340.877049] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-246f33d8-2c10-4395-872c-0a21757bc917 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1340.882015] env[65726]: DEBUG oslo_vmware.api [None req-cb618542-8b34-4ac8-a12a-8db7c40cdc7f tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Waiting for the task: (returnval){ [ 1340.882015] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52c327ed-76b9-ac3a-91ad-97ad7b816244" [ 1340.882015] env[65726]: _type = "Task" [ 1340.882015] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1340.890248] env[65726]: DEBUG oslo_vmware.api [None req-cb618542-8b34-4ac8-a12a-8db7c40cdc7f tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52c327ed-76b9-ac3a-91ad-97ad7b816244, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1341.392699] env[65726]: DEBUG oslo_vmware.api [None req-cb618542-8b34-4ac8-a12a-8db7c40cdc7f tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52c327ed-76b9-ac3a-91ad-97ad7b816244, 'name': SearchDatastore_Task, 'duration_secs': 0.013286} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1341.393086] env[65726]: DEBUG oslo_concurrency.lockutils [None req-cb618542-8b34-4ac8-a12a-8db7c40cdc7f tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1341.393252] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-cb618542-8b34-4ac8-a12a-8db7c40cdc7f tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 58562386-ef51-44ef-869c-a3102a6aafad] Processing image b52362a3-ee8a-4cbf-b06f-513b0cc8f95c {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1341.393484] env[65726]: DEBUG oslo_concurrency.lockutils [None req-cb618542-8b34-4ac8-a12a-8db7c40cdc7f tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1341.393621] env[65726]: DEBUG oslo_concurrency.lockutils [None req-cb618542-8b34-4ac8-a12a-8db7c40cdc7f tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1341.393795] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-cb618542-8b34-4ac8-a12a-8db7c40cdc7f tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1341.394073] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-89f87d56-b6cf-44ce-8af8-6517649a954e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1341.403175] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-cb618542-8b34-4ac8-a12a-8db7c40cdc7f tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1341.403404] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-cb618542-8b34-4ac8-a12a-8db7c40cdc7f tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65726) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1341.404059] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2f3f32e4-29d2-4cbf-84a5-2c67eceeb867 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1341.410929] env[65726]: DEBUG oslo_vmware.api [None req-cb618542-8b34-4ac8-a12a-8db7c40cdc7f tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Waiting for the task: (returnval){ [ 1341.410929] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]526dfc1d-0f84-55bc-7682-5ea490df8152" [ 1341.410929] env[65726]: _type = "Task" [ 1341.410929] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1341.418741] env[65726]: DEBUG oslo_vmware.api [None req-cb618542-8b34-4ac8-a12a-8db7c40cdc7f tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]526dfc1d-0f84-55bc-7682-5ea490df8152, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1341.921825] env[65726]: DEBUG oslo_vmware.api [None req-cb618542-8b34-4ac8-a12a-8db7c40cdc7f tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]526dfc1d-0f84-55bc-7682-5ea490df8152, 'name': SearchDatastore_Task, 'duration_secs': 0.008971} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1341.922646] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-656b8d5e-b2b7-43c1-8e58-49b9c0d3f4c4 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1341.928493] env[65726]: DEBUG oslo_vmware.api [None req-cb618542-8b34-4ac8-a12a-8db7c40cdc7f tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Waiting for the task: (returnval){ [ 1341.928493] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]527d9879-0894-3727-065c-e479f4eebe40" [ 1341.928493] env[65726]: _type = "Task" [ 1341.928493] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1341.937132] env[65726]: DEBUG oslo_vmware.api [None req-cb618542-8b34-4ac8-a12a-8db7c40cdc7f tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]527d9879-0894-3727-065c-e479f4eebe40, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1342.440030] env[65726]: DEBUG oslo_vmware.api [None req-cb618542-8b34-4ac8-a12a-8db7c40cdc7f tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]527d9879-0894-3727-065c-e479f4eebe40, 'name': SearchDatastore_Task, 'duration_secs': 0.010889} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1342.440685] env[65726]: DEBUG oslo_concurrency.lockutils [None req-cb618542-8b34-4ac8-a12a-8db7c40cdc7f tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1342.440983] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-cb618542-8b34-4ac8-a12a-8db7c40cdc7f tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] 58562386-ef51-44ef-869c-a3102a6aafad/58562386-ef51-44ef-869c-a3102a6aafad.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1342.441276] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-94f86b16-5759-4e5d-904a-902c7e1eab2f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1342.449262] env[65726]: DEBUG oslo_vmware.api [None req-cb618542-8b34-4ac8-a12a-8db7c40cdc7f tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Waiting for the task: (returnval){ [ 1342.449262] env[65726]: value = "task-5117093" [ 1342.449262] env[65726]: _type = "Task" [ 1342.449262] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1342.457445] env[65726]: DEBUG oslo_vmware.api [None req-cb618542-8b34-4ac8-a12a-8db7c40cdc7f tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': task-5117093, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1342.959601] env[65726]: DEBUG oslo_vmware.api [None req-cb618542-8b34-4ac8-a12a-8db7c40cdc7f tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': task-5117093, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1343.460919] env[65726]: DEBUG oslo_vmware.api [None req-cb618542-8b34-4ac8-a12a-8db7c40cdc7f tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': task-5117093, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.516792} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1343.461303] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-cb618542-8b34-4ac8-a12a-8db7c40cdc7f tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] 58562386-ef51-44ef-869c-a3102a6aafad/58562386-ef51-44ef-869c-a3102a6aafad.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1343.461303] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-cb618542-8b34-4ac8-a12a-8db7c40cdc7f tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 58562386-ef51-44ef-869c-a3102a6aafad] Extending root virtual disk to 1048576 {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1343.461564] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-df9fea0d-7259-4e36-b820-3f70cf98e1b8 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1343.468254] env[65726]: DEBUG oslo_vmware.api [None req-cb618542-8b34-4ac8-a12a-8db7c40cdc7f tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Waiting for the task: (returnval){ [ 1343.468254] env[65726]: value = "task-5117094" [ 1343.468254] env[65726]: _type = "Task" [ 1343.468254] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1343.478757] env[65726]: DEBUG oslo_vmware.api [None req-cb618542-8b34-4ac8-a12a-8db7c40cdc7f tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': task-5117094, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1343.978522] env[65726]: DEBUG oslo_vmware.api [None req-cb618542-8b34-4ac8-a12a-8db7c40cdc7f tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': task-5117094, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.063788} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1343.978812] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-cb618542-8b34-4ac8-a12a-8db7c40cdc7f tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 58562386-ef51-44ef-869c-a3102a6aafad] Extended root virtual disk {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1343.979619] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f4f4522-8e1f-414f-b151-7991b43d3ffe {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1344.001664] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-cb618542-8b34-4ac8-a12a-8db7c40cdc7f tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 58562386-ef51-44ef-869c-a3102a6aafad] Reconfiguring VM instance instance-00000077 to attach disk [datastore1] 58562386-ef51-44ef-869c-a3102a6aafad/58562386-ef51-44ef-869c-a3102a6aafad.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1344.001948] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-18c52a80-cf84-4b96-87ca-2a9dfee1b984 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1344.021935] env[65726]: DEBUG oslo_vmware.api [None req-cb618542-8b34-4ac8-a12a-8db7c40cdc7f tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Waiting for the task: (returnval){ [ 1344.021935] env[65726]: value = "task-5117095" [ 1344.021935] env[65726]: _type = "Task" [ 1344.021935] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1344.030457] env[65726]: DEBUG oslo_vmware.api [None req-cb618542-8b34-4ac8-a12a-8db7c40cdc7f tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': task-5117095, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1344.532313] env[65726]: DEBUG oslo_vmware.api [None req-cb618542-8b34-4ac8-a12a-8db7c40cdc7f tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': task-5117095, 'name': ReconfigVM_Task, 'duration_secs': 0.31608} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1344.532710] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-cb618542-8b34-4ac8-a12a-8db7c40cdc7f tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 58562386-ef51-44ef-869c-a3102a6aafad] Reconfigured VM instance instance-00000077 to attach disk [datastore1] 58562386-ef51-44ef-869c-a3102a6aafad/58562386-ef51-44ef-869c-a3102a6aafad.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1344.533815] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-cb618542-8b34-4ac8-a12a-8db7c40cdc7f tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 58562386-ef51-44ef-869c-a3102a6aafad] Block device information present: {'root_device_name': '/dev/sda', 'image': [{'disk_bus': None, 'encrypted': False, 'encryption_format': None, 'device_type': 'disk', 'encryption_secret_uuid': None, 'encryption_options': None, 'boot_index': 0, 'guest_format': None, 'size': 0, 'device_name': '/dev/sda', 'image_id': 'b52362a3-ee8a-4cbf-b06f-513b0cc8f95c'}], 'ephemerals': [], 'block_device_mapping': [{'disk_bus': None, 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-995347', 'volume_id': '98edee64-3e8c-4e8d-8eec-9f08f9e977f2', 'name': 'volume-98edee64-3e8c-4e8d-8eec-9f08f9e977f2', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '58562386-ef51-44ef-869c-a3102a6aafad', 'attached_at': '', 'detached_at': '', 'volume_id': '98edee64-3e8c-4e8d-8eec-9f08f9e977f2', 'serial': '98edee64-3e8c-4e8d-8eec-9f08f9e977f2'}, 'delete_on_termination': False, 'device_type': None, 'boot_index': None, 'mount_device': '/dev/sdb', 'attachment_id': '85b436a0-3337-4caa-af57-734f9dc10840', 'guest_format': None, 'volume_type': None}], 'swap': None} {{(pid=65726) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 1344.534088] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-cb618542-8b34-4ac8-a12a-8db7c40cdc7f tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 58562386-ef51-44ef-869c-a3102a6aafad] Volume attach. Driver type: vmdk {{(pid=65726) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1344.534339] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-cb618542-8b34-4ac8-a12a-8db7c40cdc7f tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 58562386-ef51-44ef-869c-a3102a6aafad] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-995347', 'volume_id': '98edee64-3e8c-4e8d-8eec-9f08f9e977f2', 'name': 'volume-98edee64-3e8c-4e8d-8eec-9f08f9e977f2', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '58562386-ef51-44ef-869c-a3102a6aafad', 'attached_at': '', 'detached_at': '', 'volume_id': '98edee64-3e8c-4e8d-8eec-9f08f9e977f2', 'serial': '98edee64-3e8c-4e8d-8eec-9f08f9e977f2'} {{(pid=65726) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1344.535203] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e35739b-bd3d-4289-8d0a-26b909c0b830 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1344.551178] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94bcdfcf-2bfb-402f-bd41-97b7b25e8d10 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1344.576896] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-cb618542-8b34-4ac8-a12a-8db7c40cdc7f tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 58562386-ef51-44ef-869c-a3102a6aafad] Reconfiguring VM instance instance-00000077 to attach disk [datastore2] volume-98edee64-3e8c-4e8d-8eec-9f08f9e977f2/volume-98edee64-3e8c-4e8d-8eec-9f08f9e977f2.vmdk or device None with type thin {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1344.577215] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-72068d73-bcd1-47df-b6fa-bc2641d36f33 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1344.596857] env[65726]: DEBUG oslo_vmware.api [None req-cb618542-8b34-4ac8-a12a-8db7c40cdc7f tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Waiting for the task: (returnval){ [ 1344.596857] env[65726]: value = "task-5117096" [ 1344.596857] env[65726]: _type = "Task" [ 1344.596857] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1344.605547] env[65726]: DEBUG oslo_vmware.api [None req-cb618542-8b34-4ac8-a12a-8db7c40cdc7f tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': task-5117096, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1345.108028] env[65726]: DEBUG oslo_vmware.api [None req-cb618542-8b34-4ac8-a12a-8db7c40cdc7f tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': task-5117096, 'name': ReconfigVM_Task, 'duration_secs': 0.285353} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1345.108341] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-cb618542-8b34-4ac8-a12a-8db7c40cdc7f tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 58562386-ef51-44ef-869c-a3102a6aafad] Reconfigured VM instance instance-00000077 to attach disk [datastore2] volume-98edee64-3e8c-4e8d-8eec-9f08f9e977f2/volume-98edee64-3e8c-4e8d-8eec-9f08f9e977f2.vmdk or device None with type thin {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1345.113105] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1299d43b-86ec-4ede-9dde-4b9f03b5c2b9 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1345.130419] env[65726]: DEBUG oslo_vmware.api [None req-cb618542-8b34-4ac8-a12a-8db7c40cdc7f tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Waiting for the task: (returnval){ [ 1345.130419] env[65726]: value = "task-5117097" [ 1345.130419] env[65726]: _type = "Task" [ 1345.130419] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1345.139115] env[65726]: DEBUG oslo_vmware.api [None req-cb618542-8b34-4ac8-a12a-8db7c40cdc7f tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': task-5117097, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1345.640960] env[65726]: DEBUG oslo_vmware.api [None req-cb618542-8b34-4ac8-a12a-8db7c40cdc7f tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': task-5117097, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1346.141797] env[65726]: DEBUG oslo_vmware.api [None req-cb618542-8b34-4ac8-a12a-8db7c40cdc7f tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': task-5117097, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1346.642580] env[65726]: DEBUG oslo_vmware.api [None req-cb618542-8b34-4ac8-a12a-8db7c40cdc7f tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': task-5117097, 'name': ReconfigVM_Task, 'duration_secs': 1.153671} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1346.642986] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-cb618542-8b34-4ac8-a12a-8db7c40cdc7f tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 58562386-ef51-44ef-869c-a3102a6aafad] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-995347', 'volume_id': '98edee64-3e8c-4e8d-8eec-9f08f9e977f2', 'name': 'volume-98edee64-3e8c-4e8d-8eec-9f08f9e977f2', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '58562386-ef51-44ef-869c-a3102a6aafad', 'attached_at': '', 'detached_at': '', 'volume_id': '98edee64-3e8c-4e8d-8eec-9f08f9e977f2', 'serial': '98edee64-3e8c-4e8d-8eec-9f08f9e977f2'} {{(pid=65726) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1346.643636] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a7e9d35f-ceb5-49c3-bf66-6d3a9d30430e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1346.650026] env[65726]: DEBUG oslo_vmware.api [None req-cb618542-8b34-4ac8-a12a-8db7c40cdc7f tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Waiting for the task: (returnval){ [ 1346.650026] env[65726]: value = "task-5117098" [ 1346.650026] env[65726]: _type = "Task" [ 1346.650026] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1346.657629] env[65726]: DEBUG oslo_vmware.api [None req-cb618542-8b34-4ac8-a12a-8db7c40cdc7f tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': task-5117098, 'name': Rename_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1347.160088] env[65726]: DEBUG oslo_vmware.api [None req-cb618542-8b34-4ac8-a12a-8db7c40cdc7f tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': task-5117098, 'name': Rename_Task, 'duration_secs': 0.14864} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1347.160410] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-cb618542-8b34-4ac8-a12a-8db7c40cdc7f tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 58562386-ef51-44ef-869c-a3102a6aafad] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1347.160637] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c7feccac-07c4-40cd-a6ac-06676a6ac857 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1347.167032] env[65726]: DEBUG oslo_vmware.api [None req-cb618542-8b34-4ac8-a12a-8db7c40cdc7f tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Waiting for the task: (returnval){ [ 1347.167032] env[65726]: value = "task-5117099" [ 1347.167032] env[65726]: _type = "Task" [ 1347.167032] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1347.174984] env[65726]: DEBUG oslo_vmware.api [None req-cb618542-8b34-4ac8-a12a-8db7c40cdc7f tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': task-5117099, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1347.677899] env[65726]: DEBUG oslo_vmware.api [None req-cb618542-8b34-4ac8-a12a-8db7c40cdc7f tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': task-5117099, 'name': PowerOnVM_Task, 'duration_secs': 0.453757} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1347.678234] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-cb618542-8b34-4ac8-a12a-8db7c40cdc7f tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 58562386-ef51-44ef-869c-a3102a6aafad] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1347.678425] env[65726]: DEBUG nova.compute.manager [None req-cb618542-8b34-4ac8-a12a-8db7c40cdc7f tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 58562386-ef51-44ef-869c-a3102a6aafad] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1347.679221] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1580b358-1ddc-4d94-86de-1a8600774815 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1348.197272] env[65726]: DEBUG oslo_concurrency.lockutils [None req-cb618542-8b34-4ac8-a12a-8db7c40cdc7f tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1348.197540] env[65726]: DEBUG oslo_concurrency.lockutils [None req-cb618542-8b34-4ac8-a12a-8db7c40cdc7f tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1348.197719] env[65726]: DEBUG nova.objects.instance [None req-cb618542-8b34-4ac8-a12a-8db7c40cdc7f tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 58562386-ef51-44ef-869c-a3102a6aafad] Trying to apply a migration context that does not seem to be set for this instance {{(pid=65726) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 1349.207743] env[65726]: DEBUG oslo_concurrency.lockutils [None req-cb618542-8b34-4ac8-a12a-8db7c40cdc7f tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.010s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1353.333331] env[65726]: INFO nova.compute.manager [None req-949e35bf-8a8c-45e4-89fa-ad5459f43fff tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 1ad4c4f9-1b02-420c-ac89-5e3390ddf773] Rebuilding instance [ 1353.372313] env[65726]: DEBUG nova.compute.manager [None req-949e35bf-8a8c-45e4-89fa-ad5459f43fff tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 1ad4c4f9-1b02-420c-ac89-5e3390ddf773] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1353.373240] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5054eb55-d5c1-4211-a8c8-8aec4df70897 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1354.386725] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-949e35bf-8a8c-45e4-89fa-ad5459f43fff tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 1ad4c4f9-1b02-420c-ac89-5e3390ddf773] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1354.387230] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0f7946b6-967c-479f-8f18-3f51ad152874 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1354.395991] env[65726]: DEBUG oslo_vmware.api [None req-949e35bf-8a8c-45e4-89fa-ad5459f43fff tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Waiting for the task: (returnval){ [ 1354.395991] env[65726]: value = "task-5117100" [ 1354.395991] env[65726]: _type = "Task" [ 1354.395991] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1354.404548] env[65726]: DEBUG oslo_vmware.api [None req-949e35bf-8a8c-45e4-89fa-ad5459f43fff tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': task-5117100, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1354.905587] env[65726]: DEBUG oslo_vmware.api [None req-949e35bf-8a8c-45e4-89fa-ad5459f43fff tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': task-5117100, 'name': PowerOffVM_Task, 'duration_secs': 0.177631} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1354.905847] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-949e35bf-8a8c-45e4-89fa-ad5459f43fff tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 1ad4c4f9-1b02-420c-ac89-5e3390ddf773] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1354.906088] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-949e35bf-8a8c-45e4-89fa-ad5459f43fff tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 1ad4c4f9-1b02-420c-ac89-5e3390ddf773] Destroying instance {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1354.906850] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34934a05-7912-41d3-9058-d3072f341d56 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1354.913501] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-949e35bf-8a8c-45e4-89fa-ad5459f43fff tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 1ad4c4f9-1b02-420c-ac89-5e3390ddf773] Unregistering the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1354.913738] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5bffcdb6-1cbd-4a30-bf2e-487c04b12f64 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1354.978782] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-949e35bf-8a8c-45e4-89fa-ad5459f43fff tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 1ad4c4f9-1b02-420c-ac89-5e3390ddf773] Unregistered the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1354.979015] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-949e35bf-8a8c-45e4-89fa-ad5459f43fff tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 1ad4c4f9-1b02-420c-ac89-5e3390ddf773] Deleting contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1354.979209] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-949e35bf-8a8c-45e4-89fa-ad5459f43fff tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Deleting the datastore file [datastore1] 1ad4c4f9-1b02-420c-ac89-5e3390ddf773 {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1354.979491] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f52718e8-ce69-4ed2-9984-232a8f3a23c3 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1354.986431] env[65726]: DEBUG oslo_vmware.api [None req-949e35bf-8a8c-45e4-89fa-ad5459f43fff tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Waiting for the task: (returnval){ [ 1354.986431] env[65726]: value = "task-5117102" [ 1354.986431] env[65726]: _type = "Task" [ 1354.986431] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1354.994538] env[65726]: DEBUG oslo_vmware.api [None req-949e35bf-8a8c-45e4-89fa-ad5459f43fff tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': task-5117102, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1355.497640] env[65726]: DEBUG oslo_vmware.api [None req-949e35bf-8a8c-45e4-89fa-ad5459f43fff tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': task-5117102, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.151056} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1355.498104] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-949e35bf-8a8c-45e4-89fa-ad5459f43fff tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Deleted the datastore file {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1355.498337] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-949e35bf-8a8c-45e4-89fa-ad5459f43fff tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 1ad4c4f9-1b02-420c-ac89-5e3390ddf773] Deleted contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1355.498618] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-949e35bf-8a8c-45e4-89fa-ad5459f43fff tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 1ad4c4f9-1b02-420c-ac89-5e3390ddf773] Instance destroyed {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1356.535378] env[65726]: DEBUG nova.virt.hardware [None req-949e35bf-8a8c-45e4-89fa-ad5459f43fff tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T19:28:34Z,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1539f06fa3534e90acbc3a60c4b8bd3f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T19:28:35Z,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1356.535733] env[65726]: DEBUG nova.virt.hardware [None req-949e35bf-8a8c-45e4-89fa-ad5459f43fff tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1356.535786] env[65726]: DEBUG nova.virt.hardware [None req-949e35bf-8a8c-45e4-89fa-ad5459f43fff tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1356.535964] env[65726]: DEBUG nova.virt.hardware [None req-949e35bf-8a8c-45e4-89fa-ad5459f43fff tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1356.536120] env[65726]: DEBUG nova.virt.hardware [None req-949e35bf-8a8c-45e4-89fa-ad5459f43fff tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1356.536264] env[65726]: DEBUG nova.virt.hardware [None req-949e35bf-8a8c-45e4-89fa-ad5459f43fff tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1356.536491] env[65726]: DEBUG nova.virt.hardware [None req-949e35bf-8a8c-45e4-89fa-ad5459f43fff tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1356.536656] env[65726]: DEBUG nova.virt.hardware [None req-949e35bf-8a8c-45e4-89fa-ad5459f43fff tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1356.536820] env[65726]: DEBUG nova.virt.hardware [None req-949e35bf-8a8c-45e4-89fa-ad5459f43fff tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1356.536978] env[65726]: DEBUG nova.virt.hardware [None req-949e35bf-8a8c-45e4-89fa-ad5459f43fff tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1356.537542] env[65726]: DEBUG nova.virt.hardware [None req-949e35bf-8a8c-45e4-89fa-ad5459f43fff tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1356.538103] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4dd31a59-187f-4793-bd33-ed28f384681b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1356.546682] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42af4cf1-68e6-49e8-bbed-ae2b87e8fe2e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1356.560982] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-949e35bf-8a8c-45e4-89fa-ad5459f43fff tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 1ad4c4f9-1b02-420c-ac89-5e3390ddf773] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:46:e9:71', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '205fb402-8eaf-4b61-8f57-8f216024179a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3aa1b8d0-3927-41ed-903b-212dc4ad95b6', 'vif_model': 'vmxnet3'}] {{(pid=65726) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1356.568535] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-949e35bf-8a8c-45e4-89fa-ad5459f43fff tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1356.568774] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1ad4c4f9-1b02-420c-ac89-5e3390ddf773] Creating VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1356.568985] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-19c07868-082d-472d-806d-abfce7a8b390 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1356.588097] env[65726]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1356.588097] env[65726]: value = "task-5117103" [ 1356.588097] env[65726]: _type = "Task" [ 1356.588097] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1356.595495] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5117103, 'name': CreateVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1357.098531] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5117103, 'name': CreateVM_Task, 'duration_secs': 0.337245} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1357.098773] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1ad4c4f9-1b02-420c-ac89-5e3390ddf773] Created VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1357.099523] env[65726]: WARNING openstack [None req-949e35bf-8a8c-45e4-89fa-ad5459f43fff tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1357.099928] env[65726]: WARNING openstack [None req-949e35bf-8a8c-45e4-89fa-ad5459f43fff tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1357.104958] env[65726]: DEBUG oslo_concurrency.lockutils [None req-949e35bf-8a8c-45e4-89fa-ad5459f43fff tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1357.105117] env[65726]: DEBUG oslo_concurrency.lockutils [None req-949e35bf-8a8c-45e4-89fa-ad5459f43fff tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1357.105451] env[65726]: DEBUG oslo_concurrency.lockutils [None req-949e35bf-8a8c-45e4-89fa-ad5459f43fff tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1357.105724] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b6deb545-9935-418b-af50-581190460fce {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1357.110569] env[65726]: DEBUG oslo_vmware.api [None req-949e35bf-8a8c-45e4-89fa-ad5459f43fff tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Waiting for the task: (returnval){ [ 1357.110569] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52d8db31-b41b-575e-c4f3-90b24659e5cb" [ 1357.110569] env[65726]: _type = "Task" [ 1357.110569] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1357.118218] env[65726]: DEBUG oslo_vmware.api [None req-949e35bf-8a8c-45e4-89fa-ad5459f43fff tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52d8db31-b41b-575e-c4f3-90b24659e5cb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1357.622637] env[65726]: DEBUG oslo_vmware.api [None req-949e35bf-8a8c-45e4-89fa-ad5459f43fff tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52d8db31-b41b-575e-c4f3-90b24659e5cb, 'name': SearchDatastore_Task, 'duration_secs': 0.012864} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1357.623063] env[65726]: DEBUG oslo_concurrency.lockutils [None req-949e35bf-8a8c-45e4-89fa-ad5459f43fff tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1357.623175] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-949e35bf-8a8c-45e4-89fa-ad5459f43fff tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 1ad4c4f9-1b02-420c-ac89-5e3390ddf773] Processing image b52362a3-ee8a-4cbf-b06f-513b0cc8f95c {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1357.623480] env[65726]: DEBUG oslo_concurrency.lockutils [None req-949e35bf-8a8c-45e4-89fa-ad5459f43fff tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1357.623650] env[65726]: DEBUG oslo_concurrency.lockutils [None req-949e35bf-8a8c-45e4-89fa-ad5459f43fff tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1357.623878] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-949e35bf-8a8c-45e4-89fa-ad5459f43fff tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1357.624312] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-152f56ea-2dca-426d-aa54-3f7118823fbe {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1357.635505] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-949e35bf-8a8c-45e4-89fa-ad5459f43fff tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1357.635750] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-949e35bf-8a8c-45e4-89fa-ad5459f43fff tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65726) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1357.636668] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-671e36a8-c9e0-4372-aebf-ef7adab87ef1 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1357.643245] env[65726]: DEBUG oslo_vmware.api [None req-949e35bf-8a8c-45e4-89fa-ad5459f43fff tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Waiting for the task: (returnval){ [ 1357.643245] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52f7aa5a-de98-f5bc-d895-a7402b6d5bd5" [ 1357.643245] env[65726]: _type = "Task" [ 1357.643245] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1357.651246] env[65726]: DEBUG oslo_vmware.api [None req-949e35bf-8a8c-45e4-89fa-ad5459f43fff tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52f7aa5a-de98-f5bc-d895-a7402b6d5bd5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1358.154545] env[65726]: DEBUG oslo_vmware.api [None req-949e35bf-8a8c-45e4-89fa-ad5459f43fff tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52f7aa5a-de98-f5bc-d895-a7402b6d5bd5, 'name': SearchDatastore_Task, 'duration_secs': 0.010676} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1358.155489] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-547d62b6-b461-4386-9fb4-bae85e0261d2 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1358.161089] env[65726]: DEBUG oslo_vmware.api [None req-949e35bf-8a8c-45e4-89fa-ad5459f43fff tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Waiting for the task: (returnval){ [ 1358.161089] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5229989b-26d4-e25c-49a1-de635cc384fc" [ 1358.161089] env[65726]: _type = "Task" [ 1358.161089] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1358.168929] env[65726]: DEBUG oslo_vmware.api [None req-949e35bf-8a8c-45e4-89fa-ad5459f43fff tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5229989b-26d4-e25c-49a1-de635cc384fc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1358.672798] env[65726]: DEBUG oslo_vmware.api [None req-949e35bf-8a8c-45e4-89fa-ad5459f43fff tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5229989b-26d4-e25c-49a1-de635cc384fc, 'name': SearchDatastore_Task, 'duration_secs': 0.011074} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1358.673189] env[65726]: DEBUG oslo_concurrency.lockutils [None req-949e35bf-8a8c-45e4-89fa-ad5459f43fff tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1358.673299] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-949e35bf-8a8c-45e4-89fa-ad5459f43fff tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] 1ad4c4f9-1b02-420c-ac89-5e3390ddf773/1ad4c4f9-1b02-420c-ac89-5e3390ddf773.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1358.673594] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-90b017f6-2a43-4615-8b93-df772d1cb1b6 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1358.680424] env[65726]: DEBUG oslo_vmware.api [None req-949e35bf-8a8c-45e4-89fa-ad5459f43fff tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Waiting for the task: (returnval){ [ 1358.680424] env[65726]: value = "task-5117104" [ 1358.680424] env[65726]: _type = "Task" [ 1358.680424] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1358.688016] env[65726]: DEBUG oslo_vmware.api [None req-949e35bf-8a8c-45e4-89fa-ad5459f43fff tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': task-5117104, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1359.191149] env[65726]: DEBUG oslo_vmware.api [None req-949e35bf-8a8c-45e4-89fa-ad5459f43fff tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': task-5117104, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1359.692279] env[65726]: DEBUG oslo_vmware.api [None req-949e35bf-8a8c-45e4-89fa-ad5459f43fff tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': task-5117104, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.520069} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1359.692644] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-949e35bf-8a8c-45e4-89fa-ad5459f43fff tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] 1ad4c4f9-1b02-420c-ac89-5e3390ddf773/1ad4c4f9-1b02-420c-ac89-5e3390ddf773.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1359.692775] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-949e35bf-8a8c-45e4-89fa-ad5459f43fff tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 1ad4c4f9-1b02-420c-ac89-5e3390ddf773] Extending root virtual disk to 1048576 {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1359.693031] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3900ec4f-4809-472a-a85c-99aecd446ebe {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1359.700577] env[65726]: DEBUG oslo_vmware.api [None req-949e35bf-8a8c-45e4-89fa-ad5459f43fff tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Waiting for the task: (returnval){ [ 1359.700577] env[65726]: value = "task-5117105" [ 1359.700577] env[65726]: _type = "Task" [ 1359.700577] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1359.710634] env[65726]: DEBUG oslo_vmware.api [None req-949e35bf-8a8c-45e4-89fa-ad5459f43fff tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': task-5117105, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1360.210993] env[65726]: DEBUG oslo_vmware.api [None req-949e35bf-8a8c-45e4-89fa-ad5459f43fff tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': task-5117105, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.076237} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1360.211282] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-949e35bf-8a8c-45e4-89fa-ad5459f43fff tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 1ad4c4f9-1b02-420c-ac89-5e3390ddf773] Extended root virtual disk {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1360.212055] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f880c84-be35-4002-954e-ad4478e2b57f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1360.234618] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-949e35bf-8a8c-45e4-89fa-ad5459f43fff tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 1ad4c4f9-1b02-420c-ac89-5e3390ddf773] Reconfiguring VM instance instance-0000007a to attach disk [datastore1] 1ad4c4f9-1b02-420c-ac89-5e3390ddf773/1ad4c4f9-1b02-420c-ac89-5e3390ddf773.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1360.234922] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-93adfeea-f8ea-48ef-91a5-13cc9227df98 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1360.254171] env[65726]: DEBUG oslo_vmware.api [None req-949e35bf-8a8c-45e4-89fa-ad5459f43fff tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Waiting for the task: (returnval){ [ 1360.254171] env[65726]: value = "task-5117106" [ 1360.254171] env[65726]: _type = "Task" [ 1360.254171] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1360.261903] env[65726]: DEBUG oslo_vmware.api [None req-949e35bf-8a8c-45e4-89fa-ad5459f43fff tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': task-5117106, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1360.765714] env[65726]: DEBUG oslo_vmware.api [None req-949e35bf-8a8c-45e4-89fa-ad5459f43fff tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': task-5117106, 'name': ReconfigVM_Task, 'duration_secs': 0.342238} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1360.766121] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-949e35bf-8a8c-45e4-89fa-ad5459f43fff tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 1ad4c4f9-1b02-420c-ac89-5e3390ddf773] Reconfigured VM instance instance-0000007a to attach disk [datastore1] 1ad4c4f9-1b02-420c-ac89-5e3390ddf773/1ad4c4f9-1b02-420c-ac89-5e3390ddf773.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1360.766664] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b8fd6046-a45f-4527-bc9e-05f9464eef48 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1360.774997] env[65726]: DEBUG oslo_vmware.api [None req-949e35bf-8a8c-45e4-89fa-ad5459f43fff tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Waiting for the task: (returnval){ [ 1360.774997] env[65726]: value = "task-5117107" [ 1360.774997] env[65726]: _type = "Task" [ 1360.774997] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1360.784240] env[65726]: DEBUG oslo_vmware.api [None req-949e35bf-8a8c-45e4-89fa-ad5459f43fff tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': task-5117107, 'name': Rename_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1361.285018] env[65726]: DEBUG oslo_vmware.api [None req-949e35bf-8a8c-45e4-89fa-ad5459f43fff tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': task-5117107, 'name': Rename_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1361.785934] env[65726]: DEBUG oslo_vmware.api [None req-949e35bf-8a8c-45e4-89fa-ad5459f43fff tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': task-5117107, 'name': Rename_Task, 'duration_secs': 0.921426} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1361.786244] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-949e35bf-8a8c-45e4-89fa-ad5459f43fff tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 1ad4c4f9-1b02-420c-ac89-5e3390ddf773] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1361.786469] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-83822e61-cf39-44c6-a1a8-18b9cfb71b94 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1361.794080] env[65726]: DEBUG oslo_vmware.api [None req-949e35bf-8a8c-45e4-89fa-ad5459f43fff tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Waiting for the task: (returnval){ [ 1361.794080] env[65726]: value = "task-5117108" [ 1361.794080] env[65726]: _type = "Task" [ 1361.794080] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1361.801890] env[65726]: DEBUG oslo_vmware.api [None req-949e35bf-8a8c-45e4-89fa-ad5459f43fff tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': task-5117108, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1362.303209] env[65726]: DEBUG oslo_vmware.api [None req-949e35bf-8a8c-45e4-89fa-ad5459f43fff tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': task-5117108, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1362.806171] env[65726]: DEBUG oslo_vmware.api [None req-949e35bf-8a8c-45e4-89fa-ad5459f43fff tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': task-5117108, 'name': PowerOnVM_Task, 'duration_secs': 0.511514} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1362.806452] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-949e35bf-8a8c-45e4-89fa-ad5459f43fff tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 1ad4c4f9-1b02-420c-ac89-5e3390ddf773] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1362.806670] env[65726]: DEBUG nova.compute.manager [None req-949e35bf-8a8c-45e4-89fa-ad5459f43fff tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 1ad4c4f9-1b02-420c-ac89-5e3390ddf773] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1362.807523] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a5f8f66-d778-4768-a6f6-cb742c3d56af {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1363.324291] env[65726]: DEBUG oslo_concurrency.lockutils [None req-949e35bf-8a8c-45e4-89fa-ad5459f43fff tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1363.325027] env[65726]: DEBUG oslo_concurrency.lockutils [None req-949e35bf-8a8c-45e4-89fa-ad5459f43fff tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1363.325027] env[65726]: DEBUG nova.objects.instance [None req-949e35bf-8a8c-45e4-89fa-ad5459f43fff tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 1ad4c4f9-1b02-420c-ac89-5e3390ddf773] Trying to apply a migration context that does not seem to be set for this instance {{(pid=65726) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 1364.333981] env[65726]: DEBUG oslo_concurrency.lockutils [None req-949e35bf-8a8c-45e4-89fa-ad5459f43fff tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.009s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1384.792235] env[65726]: DEBUG oslo_concurrency.lockutils [None req-83f9b618-4589-40f6-96d8-dbbed62c406b tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Acquiring lock "58562386-ef51-44ef-869c-a3102a6aafad" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1384.792748] env[65726]: DEBUG oslo_concurrency.lockutils [None req-83f9b618-4589-40f6-96d8-dbbed62c406b tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Lock "58562386-ef51-44ef-869c-a3102a6aafad" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1385.297030] env[65726]: INFO nova.compute.manager [None req-83f9b618-4589-40f6-96d8-dbbed62c406b tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 58562386-ef51-44ef-869c-a3102a6aafad] Detaching volume 98edee64-3e8c-4e8d-8eec-9f08f9e977f2 [ 1385.329381] env[65726]: INFO nova.virt.block_device [None req-83f9b618-4589-40f6-96d8-dbbed62c406b tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 58562386-ef51-44ef-869c-a3102a6aafad] Attempting to driver detach volume 98edee64-3e8c-4e8d-8eec-9f08f9e977f2 from mountpoint /dev/sdb [ 1385.329657] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-83f9b618-4589-40f6-96d8-dbbed62c406b tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 58562386-ef51-44ef-869c-a3102a6aafad] Volume detach. Driver type: vmdk {{(pid=65726) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1385.329881] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-83f9b618-4589-40f6-96d8-dbbed62c406b tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 58562386-ef51-44ef-869c-a3102a6aafad] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-995347', 'volume_id': '98edee64-3e8c-4e8d-8eec-9f08f9e977f2', 'name': 'volume-98edee64-3e8c-4e8d-8eec-9f08f9e977f2', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '58562386-ef51-44ef-869c-a3102a6aafad', 'attached_at': '', 'detached_at': '', 'volume_id': '98edee64-3e8c-4e8d-8eec-9f08f9e977f2', 'serial': '98edee64-3e8c-4e8d-8eec-9f08f9e977f2'} {{(pid=65726) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1385.330814] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3035c7fe-8b74-4185-9f98-4758aba690ab {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1385.352997] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71af84d5-e6db-48bf-ba5e-9a049449806d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1385.360796] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67544ba9-9825-4f9b-b8f8-4a61e2c39589 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1385.381224] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e2ac226-a9cc-43c0-8ae6-45aaa79f9005 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1385.395889] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-83f9b618-4589-40f6-96d8-dbbed62c406b tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] The volume has not been displaced from its original location: [datastore2] volume-98edee64-3e8c-4e8d-8eec-9f08f9e977f2/volume-98edee64-3e8c-4e8d-8eec-9f08f9e977f2.vmdk. No consolidation needed. {{(pid=65726) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1385.401206] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-83f9b618-4589-40f6-96d8-dbbed62c406b tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 58562386-ef51-44ef-869c-a3102a6aafad] Reconfiguring VM instance instance-00000077 to detach disk 2001 {{(pid=65726) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1385.401516] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-716074ed-aca7-4698-be9c-c23b92c9c550 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1385.423075] env[65726]: DEBUG oslo_vmware.api [None req-83f9b618-4589-40f6-96d8-dbbed62c406b tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Waiting for the task: (returnval){ [ 1385.423075] env[65726]: value = "task-5117109" [ 1385.423075] env[65726]: _type = "Task" [ 1385.423075] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1385.431728] env[65726]: DEBUG oslo_vmware.api [None req-83f9b618-4589-40f6-96d8-dbbed62c406b tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': task-5117109, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1385.932711] env[65726]: DEBUG oslo_vmware.api [None req-83f9b618-4589-40f6-96d8-dbbed62c406b tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': task-5117109, 'name': ReconfigVM_Task, 'duration_secs': 0.22853} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1385.933112] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-83f9b618-4589-40f6-96d8-dbbed62c406b tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 58562386-ef51-44ef-869c-a3102a6aafad] Reconfigured VM instance instance-00000077 to detach disk 2001 {{(pid=65726) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1385.937598] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8d5c5b14-0cf0-40aa-88f5-547b1a45d2fa {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1385.953047] env[65726]: DEBUG oslo_vmware.api [None req-83f9b618-4589-40f6-96d8-dbbed62c406b tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Waiting for the task: (returnval){ [ 1385.953047] env[65726]: value = "task-5117110" [ 1385.953047] env[65726]: _type = "Task" [ 1385.953047] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1385.961175] env[65726]: DEBUG oslo_vmware.api [None req-83f9b618-4589-40f6-96d8-dbbed62c406b tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': task-5117110, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1386.464359] env[65726]: DEBUG oslo_vmware.api [None req-83f9b618-4589-40f6-96d8-dbbed62c406b tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': task-5117110, 'name': ReconfigVM_Task, 'duration_secs': 0.140639} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1386.464649] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-83f9b618-4589-40f6-96d8-dbbed62c406b tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 58562386-ef51-44ef-869c-a3102a6aafad] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-995347', 'volume_id': '98edee64-3e8c-4e8d-8eec-9f08f9e977f2', 'name': 'volume-98edee64-3e8c-4e8d-8eec-9f08f9e977f2', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '58562386-ef51-44ef-869c-a3102a6aafad', 'attached_at': '', 'detached_at': '', 'volume_id': '98edee64-3e8c-4e8d-8eec-9f08f9e977f2', 'serial': '98edee64-3e8c-4e8d-8eec-9f08f9e977f2'} {{(pid=65726) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1387.004577] env[65726]: DEBUG nova.objects.instance [None req-83f9b618-4589-40f6-96d8-dbbed62c406b tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Lazy-loading 'flavor' on Instance uuid 58562386-ef51-44ef-869c-a3102a6aafad {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1387.305769] env[65726]: DEBUG oslo_service.periodic_task [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=65726) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1387.305967] env[65726]: DEBUG oslo_service.periodic_task [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=65726) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1387.306143] env[65726]: DEBUG oslo_service.periodic_task [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=65726) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1387.306297] env[65726]: DEBUG oslo_service.periodic_task [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=65726) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1387.306445] env[65726]: DEBUG oslo_service.periodic_task [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=65726) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1387.306593] env[65726]: DEBUG oslo_service.periodic_task [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=65726) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1387.306770] env[65726]: DEBUG nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=65726) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11251}} [ 1387.569892] env[65726]: DEBUG oslo_service.periodic_task [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=65726) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1388.013500] env[65726]: DEBUG oslo_concurrency.lockutils [None req-83f9b618-4589-40f6-96d8-dbbed62c406b tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Lock "58562386-ef51-44ef-869c-a3102a6aafad" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.221s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1389.038838] env[65726]: DEBUG oslo_concurrency.lockutils [None req-74541e4d-9cb3-4736-a648-4a0f94687769 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Acquiring lock "58562386-ef51-44ef-869c-a3102a6aafad" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1389.039250] env[65726]: DEBUG oslo_concurrency.lockutils [None req-74541e4d-9cb3-4736-a648-4a0f94687769 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Lock "58562386-ef51-44ef-869c-a3102a6aafad" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1389.039382] env[65726]: DEBUG oslo_concurrency.lockutils [None req-74541e4d-9cb3-4736-a648-4a0f94687769 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Acquiring lock "58562386-ef51-44ef-869c-a3102a6aafad-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1389.039558] env[65726]: DEBUG oslo_concurrency.lockutils [None req-74541e4d-9cb3-4736-a648-4a0f94687769 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Lock "58562386-ef51-44ef-869c-a3102a6aafad-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1389.039717] env[65726]: DEBUG oslo_concurrency.lockutils [None req-74541e4d-9cb3-4736-a648-4a0f94687769 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Lock "58562386-ef51-44ef-869c-a3102a6aafad-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1389.041988] env[65726]: INFO nova.compute.manager [None req-74541e4d-9cb3-4736-a648-4a0f94687769 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 58562386-ef51-44ef-869c-a3102a6aafad] Terminating instance [ 1389.545677] env[65726]: DEBUG nova.compute.manager [None req-74541e4d-9cb3-4736-a648-4a0f94687769 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 58562386-ef51-44ef-869c-a3102a6aafad] Start destroying the instance on the hypervisor. {{(pid=65726) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1389.546035] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-74541e4d-9cb3-4736-a648-4a0f94687769 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 58562386-ef51-44ef-869c-a3102a6aafad] Destroying instance {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1389.546996] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eed09e8b-a3fb-40d9-ae9c-b7ddb5048ece {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1389.555679] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-74541e4d-9cb3-4736-a648-4a0f94687769 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 58562386-ef51-44ef-869c-a3102a6aafad] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1389.555980] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7283ccaf-c110-45f2-93fe-b6057c77c9c4 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1389.563074] env[65726]: DEBUG oslo_vmware.api [None req-74541e4d-9cb3-4736-a648-4a0f94687769 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Waiting for the task: (returnval){ [ 1389.563074] env[65726]: value = "task-5117111" [ 1389.563074] env[65726]: _type = "Task" [ 1389.563074] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1389.571120] env[65726]: DEBUG oslo_service.periodic_task [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Running periodic task ComputeManager.update_available_resource {{(pid=65726) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1389.572145] env[65726]: DEBUG oslo_vmware.api [None req-74541e4d-9cb3-4736-a648-4a0f94687769 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': task-5117111, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1390.073886] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1390.074408] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1390.074408] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1390.074408] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=65726) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:937}} [ 1390.074654] env[65726]: DEBUG oslo_vmware.api [None req-74541e4d-9cb3-4736-a648-4a0f94687769 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': task-5117111, 'name': PowerOffVM_Task, 'duration_secs': 0.189194} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1390.075414] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc6ba87d-3c77-4489-b270-73f6488ccfd6 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1390.078319] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-74541e4d-9cb3-4736-a648-4a0f94687769 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 58562386-ef51-44ef-869c-a3102a6aafad] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1390.078483] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-74541e4d-9cb3-4736-a648-4a0f94687769 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 58562386-ef51-44ef-869c-a3102a6aafad] Unregistering the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1390.078723] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4996808b-ea70-4007-ac81-2b60d054c7a3 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1390.086188] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0398d71e-7e0c-4e2c-9a75-e144c1441a37 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1390.101011] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15e2f915-2d66-4cf5-9e0f-b655f9085fce {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1390.108531] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c9190a2-241d-459f-b6d7-34df19dfc2ef {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1390.140051] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=179543MB free_disk=95GB free_vcpus=48 pci_devices=None {{(pid=65726) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1136}} [ 1390.140307] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1390.140494] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1390.143397] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-74541e4d-9cb3-4736-a648-4a0f94687769 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 58562386-ef51-44ef-869c-a3102a6aafad] Unregistered the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1390.143596] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-74541e4d-9cb3-4736-a648-4a0f94687769 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 58562386-ef51-44ef-869c-a3102a6aafad] Deleting contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1390.143769] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-74541e4d-9cb3-4736-a648-4a0f94687769 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Deleting the datastore file [datastore1] 58562386-ef51-44ef-869c-a3102a6aafad {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1390.144318] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-885b4b68-8d37-4ff5-9665-fa83ef83e69b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1390.152534] env[65726]: DEBUG oslo_vmware.api [None req-74541e4d-9cb3-4736-a648-4a0f94687769 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Waiting for the task: (returnval){ [ 1390.152534] env[65726]: value = "task-5117113" [ 1390.152534] env[65726]: _type = "Task" [ 1390.152534] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1390.161488] env[65726]: DEBUG oslo_vmware.api [None req-74541e4d-9cb3-4736-a648-4a0f94687769 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': task-5117113, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1390.662716] env[65726]: DEBUG oslo_vmware.api [None req-74541e4d-9cb3-4736-a648-4a0f94687769 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': task-5117113, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.187057} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1390.662915] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-74541e4d-9cb3-4736-a648-4a0f94687769 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Deleted the datastore file {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1390.663086] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-74541e4d-9cb3-4736-a648-4a0f94687769 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 58562386-ef51-44ef-869c-a3102a6aafad] Deleted contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1390.663263] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-74541e4d-9cb3-4736-a648-4a0f94687769 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 58562386-ef51-44ef-869c-a3102a6aafad] Instance destroyed {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1390.663437] env[65726]: INFO nova.compute.manager [None req-74541e4d-9cb3-4736-a648-4a0f94687769 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 58562386-ef51-44ef-869c-a3102a6aafad] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1390.663672] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-74541e4d-9cb3-4736-a648-4a0f94687769 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1390.663858] env[65726]: DEBUG nova.compute.manager [-] [instance: 58562386-ef51-44ef-869c-a3102a6aafad] Deallocating network for instance {{(pid=65726) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1390.663953] env[65726]: DEBUG nova.network.neutron [-] [instance: 58562386-ef51-44ef-869c-a3102a6aafad] deallocate_for_instance() {{(pid=65726) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 1390.664466] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1390.664713] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1390.699840] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1390.700216] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1391.129858] env[65726]: DEBUG nova.compute.manager [req-95da5147-d7b8-4f1d-bda2-9e1298dd7fc8 req-d1a02e16-b1b8-44ed-8442-96358a523810 service nova] [instance: 58562386-ef51-44ef-869c-a3102a6aafad] Received event network-vif-deleted-dfc3c1e9-c8e6-41f5-aed1-e1c15c7f9ba6 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1391.130173] env[65726]: INFO nova.compute.manager [req-95da5147-d7b8-4f1d-bda2-9e1298dd7fc8 req-d1a02e16-b1b8-44ed-8442-96358a523810 service nova] [instance: 58562386-ef51-44ef-869c-a3102a6aafad] Neutron deleted interface dfc3c1e9-c8e6-41f5-aed1-e1c15c7f9ba6; detaching it from the instance and deleting it from the info cache [ 1391.130335] env[65726]: DEBUG nova.network.neutron [req-95da5147-d7b8-4f1d-bda2-9e1298dd7fc8 req-d1a02e16-b1b8-44ed-8442-96358a523810 service nova] [instance: 58562386-ef51-44ef-869c-a3102a6aafad] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1391.170660] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Instance e3255f3b-028f-4a0b-b621-873730417936 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65726) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1391.170804] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Instance 55ee97c3-b014-48eb-b41b-fc5aa16bc09e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65726) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1391.170921] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Instance 58562386-ef51-44ef-869c-a3102a6aafad actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65726) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1391.171047] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Instance 1ad4c4f9-1b02-420c-ac89-5e3390ddf773 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65726) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1391.171232] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Total usable vcpus: 48, total allocated vcpus: 4 {{(pid=65726) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1159}} [ 1391.171373] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1280MB phys_disk=100GB used_disk=4GB total_vcpus=48 used_vcpus=4 pci_stats=[] stats={'failed_builds': '0', 'num_instances': '4', 'num_vm_active': '4', 'num_task_None': '3', 'num_os_type_None': '4', 'num_proj_acd800e0438940bc80ae3c6b672db4ee': '2', 'io_workload': '0', 'num_proj_6b56f93ba0e14521921484ac2c785c43': '2', 'num_task_deleting': '1'} {{(pid=65726) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1168}} [ 1391.230763] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02fd7cd6-a4bf-4821-81da-37e509b55ab0 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1391.238762] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d780000a-a556-4a6a-9863-79fc32476f0b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1391.270489] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be0ff0c8-75f8-4590-bbc2-f6adfcadf57c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1391.278563] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e917923-2108-4668-a76e-783ae02413ff {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1391.291964] env[65726]: DEBUG nova.compute.provider_tree [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1391.613660] env[65726]: DEBUG nova.network.neutron [-] [instance: 58562386-ef51-44ef-869c-a3102a6aafad] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1391.632635] env[65726]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-3a072855-c074-4162-a218-c16c5f96e7b1 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1391.642355] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-569aa7bf-3342-466f-b73c-34f6da310a95 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1391.670787] env[65726]: DEBUG nova.compute.manager [req-95da5147-d7b8-4f1d-bda2-9e1298dd7fc8 req-d1a02e16-b1b8-44ed-8442-96358a523810 service nova] [instance: 58562386-ef51-44ef-869c-a3102a6aafad] Detach interface failed, port_id=dfc3c1e9-c8e6-41f5-aed1-e1c15c7f9ba6, reason: Instance 58562386-ef51-44ef-869c-a3102a6aafad could not be found. {{(pid=65726) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11669}} [ 1391.795151] env[65726]: DEBUG nova.scheduler.client.report [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1392.116757] env[65726]: INFO nova.compute.manager [-] [instance: 58562386-ef51-44ef-869c-a3102a6aafad] Took 1.45 seconds to deallocate network for instance. [ 1392.299547] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=65726) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1097}} [ 1392.299963] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.159s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1392.623656] env[65726]: DEBUG oslo_concurrency.lockutils [None req-74541e4d-9cb3-4736-a648-4a0f94687769 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1392.623943] env[65726]: DEBUG oslo_concurrency.lockutils [None req-74541e4d-9cb3-4736-a648-4a0f94687769 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1392.624157] env[65726]: DEBUG nova.objects.instance [None req-74541e4d-9cb3-4736-a648-4a0f94687769 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Lazy-loading 'resources' on Instance uuid 58562386-ef51-44ef-869c-a3102a6aafad {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1393.190566] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3c132f4-629f-4c99-a3fe-e708898f112d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1393.198792] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22d396d0-e35f-4d49-8363-59ecb4437499 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1393.228289] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8be1ffc1-1ca7-4d41-8961-044bf604531f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1393.235783] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71ebd783-6099-4bac-93c9-c38710cb4b66 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1393.249167] env[65726]: DEBUG nova.compute.provider_tree [None req-74541e4d-9cb3-4736-a648-4a0f94687769 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1393.752948] env[65726]: DEBUG nova.scheduler.client.report [None req-74541e4d-9cb3-4736-a648-4a0f94687769 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1394.258535] env[65726]: DEBUG oslo_concurrency.lockutils [None req-74541e4d-9cb3-4736-a648-4a0f94687769 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.634s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1394.280182] env[65726]: INFO nova.scheduler.client.report [None req-74541e4d-9cb3-4736-a648-4a0f94687769 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Deleted allocations for instance 58562386-ef51-44ef-869c-a3102a6aafad [ 1394.788250] env[65726]: DEBUG oslo_concurrency.lockutils [None req-74541e4d-9cb3-4736-a648-4a0f94687769 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Lock "58562386-ef51-44ef-869c-a3102a6aafad" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.749s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1396.293844] env[65726]: DEBUG oslo_service.periodic_task [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=65726) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1396.641131] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c7f89e7c-d9b3-4baf-aa24-d1596cc25228 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Acquiring lock "95617eac-9630-4a71-80ed-a9edd8e757c1" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1396.641358] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c7f89e7c-d9b3-4baf-aa24-d1596cc25228 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Lock "95617eac-9630-4a71-80ed-a9edd8e757c1" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1397.143861] env[65726]: DEBUG nova.compute.manager [None req-c7f89e7c-d9b3-4baf-aa24-d1596cc25228 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 95617eac-9630-4a71-80ed-a9edd8e757c1] Starting instance... {{(pid=65726) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 1397.667458] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c7f89e7c-d9b3-4baf-aa24-d1596cc25228 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1397.667745] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c7f89e7c-d9b3-4baf-aa24-d1596cc25228 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1397.669194] env[65726]: INFO nova.compute.claims [None req-c7f89e7c-d9b3-4baf-aa24-d1596cc25228 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 95617eac-9630-4a71-80ed-a9edd8e757c1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1398.736789] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6729265b-6ec4-4877-bcef-b12addac4425 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1398.744915] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3eb410c-7bbd-47dc-8ce8-9cb22137dce4 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1398.775630] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-270c0bd9-e716-4fde-8303-9d89d3d2d429 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1398.784759] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbeb38b3-becb-414f-9ab8-9b0155c32169 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1398.799277] env[65726]: DEBUG nova.compute.provider_tree [None req-c7f89e7c-d9b3-4baf-aa24-d1596cc25228 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1399.302854] env[65726]: DEBUG nova.scheduler.client.report [None req-c7f89e7c-d9b3-4baf-aa24-d1596cc25228 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1399.807743] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c7f89e7c-d9b3-4baf-aa24-d1596cc25228 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.140s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1399.808320] env[65726]: DEBUG nova.compute.manager [None req-c7f89e7c-d9b3-4baf-aa24-d1596cc25228 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 95617eac-9630-4a71-80ed-a9edd8e757c1] Start building networks asynchronously for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1400.313044] env[65726]: DEBUG nova.compute.utils [None req-c7f89e7c-d9b3-4baf-aa24-d1596cc25228 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Using /dev/sd instead of None {{(pid=65726) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1400.314371] env[65726]: DEBUG nova.compute.manager [None req-c7f89e7c-d9b3-4baf-aa24-d1596cc25228 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 95617eac-9630-4a71-80ed-a9edd8e757c1] Allocating IP information in the background. {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1400.314572] env[65726]: DEBUG nova.network.neutron [None req-c7f89e7c-d9b3-4baf-aa24-d1596cc25228 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 95617eac-9630-4a71-80ed-a9edd8e757c1] allocate_for_instance() {{(pid=65726) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 1400.314885] env[65726]: WARNING neutronclient.v2_0.client [None req-c7f89e7c-d9b3-4baf-aa24-d1596cc25228 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1400.315200] env[65726]: WARNING neutronclient.v2_0.client [None req-c7f89e7c-d9b3-4baf-aa24-d1596cc25228 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1400.315804] env[65726]: WARNING openstack [None req-c7f89e7c-d9b3-4baf-aa24-d1596cc25228 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1400.316162] env[65726]: WARNING openstack [None req-c7f89e7c-d9b3-4baf-aa24-d1596cc25228 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1400.367292] env[65726]: DEBUG nova.policy [None req-c7f89e7c-d9b3-4baf-aa24-d1596cc25228 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6ab5b309207a46bb9d95998ef0a7a46f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6b56f93ba0e14521921484ac2c785c43', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65726) authorize /opt/stack/nova/nova/policy.py:192}} [ 1400.468713] env[65726]: DEBUG oslo_concurrency.lockutils [None req-4f5d9381-2720-46cd-b56b-f135b25fee46 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Acquiring lock "1ad4c4f9-1b02-420c-ac89-5e3390ddf773" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1400.468985] env[65726]: DEBUG oslo_concurrency.lockutils [None req-4f5d9381-2720-46cd-b56b-f135b25fee46 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Lock "1ad4c4f9-1b02-420c-ac89-5e3390ddf773" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1400.469201] env[65726]: DEBUG oslo_concurrency.lockutils [None req-4f5d9381-2720-46cd-b56b-f135b25fee46 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Acquiring lock "1ad4c4f9-1b02-420c-ac89-5e3390ddf773-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1400.469382] env[65726]: DEBUG oslo_concurrency.lockutils [None req-4f5d9381-2720-46cd-b56b-f135b25fee46 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Lock "1ad4c4f9-1b02-420c-ac89-5e3390ddf773-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1400.469542] env[65726]: DEBUG oslo_concurrency.lockutils [None req-4f5d9381-2720-46cd-b56b-f135b25fee46 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Lock "1ad4c4f9-1b02-420c-ac89-5e3390ddf773-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1400.471694] env[65726]: INFO nova.compute.manager [None req-4f5d9381-2720-46cd-b56b-f135b25fee46 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 1ad4c4f9-1b02-420c-ac89-5e3390ddf773] Terminating instance [ 1400.608899] env[65726]: DEBUG nova.network.neutron [None req-c7f89e7c-d9b3-4baf-aa24-d1596cc25228 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 95617eac-9630-4a71-80ed-a9edd8e757c1] Successfully created port: b46e6107-621c-438e-9be6-fe1cd95a8a4e {{(pid=65726) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1400.823815] env[65726]: DEBUG nova.compute.manager [None req-c7f89e7c-d9b3-4baf-aa24-d1596cc25228 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 95617eac-9630-4a71-80ed-a9edd8e757c1] Start building block device mappings for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1400.975807] env[65726]: DEBUG nova.compute.manager [None req-4f5d9381-2720-46cd-b56b-f135b25fee46 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 1ad4c4f9-1b02-420c-ac89-5e3390ddf773] Start destroying the instance on the hypervisor. {{(pid=65726) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1400.976028] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-4f5d9381-2720-46cd-b56b-f135b25fee46 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 1ad4c4f9-1b02-420c-ac89-5e3390ddf773] Destroying instance {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1400.976940] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f58662c5-dc17-4e65-93ab-42f1462480cf {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1400.985155] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-4f5d9381-2720-46cd-b56b-f135b25fee46 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 1ad4c4f9-1b02-420c-ac89-5e3390ddf773] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1400.985396] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-96121a87-338b-4e26-82ba-74204ab87077 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1400.993085] env[65726]: DEBUG oslo_vmware.api [None req-4f5d9381-2720-46cd-b56b-f135b25fee46 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Waiting for the task: (returnval){ [ 1400.993085] env[65726]: value = "task-5117114" [ 1400.993085] env[65726]: _type = "Task" [ 1400.993085] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1401.002066] env[65726]: DEBUG oslo_vmware.api [None req-4f5d9381-2720-46cd-b56b-f135b25fee46 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': task-5117114, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1401.505191] env[65726]: DEBUG oslo_vmware.api [None req-4f5d9381-2720-46cd-b56b-f135b25fee46 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': task-5117114, 'name': PowerOffVM_Task, 'duration_secs': 0.185874} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1401.505468] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-4f5d9381-2720-46cd-b56b-f135b25fee46 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 1ad4c4f9-1b02-420c-ac89-5e3390ddf773] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1401.505633] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-4f5d9381-2720-46cd-b56b-f135b25fee46 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 1ad4c4f9-1b02-420c-ac89-5e3390ddf773] Unregistering the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1401.505932] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ed1710c2-1208-4464-8231-aac79e2aecb1 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1401.577841] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-4f5d9381-2720-46cd-b56b-f135b25fee46 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 1ad4c4f9-1b02-420c-ac89-5e3390ddf773] Unregistered the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1401.578150] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-4f5d9381-2720-46cd-b56b-f135b25fee46 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 1ad4c4f9-1b02-420c-ac89-5e3390ddf773] Deleting contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1401.578350] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-4f5d9381-2720-46cd-b56b-f135b25fee46 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Deleting the datastore file [datastore1] 1ad4c4f9-1b02-420c-ac89-5e3390ddf773 {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1401.578655] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a73350bd-ea09-4ca3-adf7-f80851d95a50 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1401.587206] env[65726]: DEBUG oslo_vmware.api [None req-4f5d9381-2720-46cd-b56b-f135b25fee46 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Waiting for the task: (returnval){ [ 1401.587206] env[65726]: value = "task-5117116" [ 1401.587206] env[65726]: _type = "Task" [ 1401.587206] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1401.595473] env[65726]: DEBUG oslo_vmware.api [None req-4f5d9381-2720-46cd-b56b-f135b25fee46 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': task-5117116, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1401.833688] env[65726]: DEBUG nova.compute.manager [None req-c7f89e7c-d9b3-4baf-aa24-d1596cc25228 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 95617eac-9630-4a71-80ed-a9edd8e757c1] Start spawning the instance on the hypervisor. {{(pid=65726) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1401.860369] env[65726]: DEBUG nova.virt.hardware [None req-c7f89e7c-d9b3-4baf-aa24-d1596cc25228 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T19:28:34Z,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1539f06fa3534e90acbc3a60c4b8bd3f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T19:28:35Z,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1401.860631] env[65726]: DEBUG nova.virt.hardware [None req-c7f89e7c-d9b3-4baf-aa24-d1596cc25228 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1401.860786] env[65726]: DEBUG nova.virt.hardware [None req-c7f89e7c-d9b3-4baf-aa24-d1596cc25228 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1401.860963] env[65726]: DEBUG nova.virt.hardware [None req-c7f89e7c-d9b3-4baf-aa24-d1596cc25228 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1401.861118] env[65726]: DEBUG nova.virt.hardware [None req-c7f89e7c-d9b3-4baf-aa24-d1596cc25228 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1401.861259] env[65726]: DEBUG nova.virt.hardware [None req-c7f89e7c-d9b3-4baf-aa24-d1596cc25228 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1401.861536] env[65726]: DEBUG nova.virt.hardware [None req-c7f89e7c-d9b3-4baf-aa24-d1596cc25228 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1401.861717] env[65726]: DEBUG nova.virt.hardware [None req-c7f89e7c-d9b3-4baf-aa24-d1596cc25228 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1401.861886] env[65726]: DEBUG nova.virt.hardware [None req-c7f89e7c-d9b3-4baf-aa24-d1596cc25228 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1401.862057] env[65726]: DEBUG nova.virt.hardware [None req-c7f89e7c-d9b3-4baf-aa24-d1596cc25228 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1401.862225] env[65726]: DEBUG nova.virt.hardware [None req-c7f89e7c-d9b3-4baf-aa24-d1596cc25228 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1401.863123] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28ef801c-f995-4658-860b-47a7299a57f2 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1401.871424] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e404e384-cf68-4d0f-a2cb-cdeba01f388a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1402.009623] env[65726]: DEBUG nova.compute.manager [req-fe8e8553-1ad0-4b16-8732-844c6b781c5f req-57993c58-3415-404f-8e0d-7a71a2854131 service nova] [instance: 95617eac-9630-4a71-80ed-a9edd8e757c1] Received event network-vif-plugged-b46e6107-621c-438e-9be6-fe1cd95a8a4e {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1402.009847] env[65726]: DEBUG oslo_concurrency.lockutils [req-fe8e8553-1ad0-4b16-8732-844c6b781c5f req-57993c58-3415-404f-8e0d-7a71a2854131 service nova] Acquiring lock "95617eac-9630-4a71-80ed-a9edd8e757c1-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1402.010065] env[65726]: DEBUG oslo_concurrency.lockutils [req-fe8e8553-1ad0-4b16-8732-844c6b781c5f req-57993c58-3415-404f-8e0d-7a71a2854131 service nova] Lock "95617eac-9630-4a71-80ed-a9edd8e757c1-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1402.010230] env[65726]: DEBUG oslo_concurrency.lockutils [req-fe8e8553-1ad0-4b16-8732-844c6b781c5f req-57993c58-3415-404f-8e0d-7a71a2854131 service nova] Lock "95617eac-9630-4a71-80ed-a9edd8e757c1-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1402.010389] env[65726]: DEBUG nova.compute.manager [req-fe8e8553-1ad0-4b16-8732-844c6b781c5f req-57993c58-3415-404f-8e0d-7a71a2854131 service nova] [instance: 95617eac-9630-4a71-80ed-a9edd8e757c1] No waiting events found dispatching network-vif-plugged-b46e6107-621c-438e-9be6-fe1cd95a8a4e {{(pid=65726) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1402.010546] env[65726]: WARNING nova.compute.manager [req-fe8e8553-1ad0-4b16-8732-844c6b781c5f req-57993c58-3415-404f-8e0d-7a71a2854131 service nova] [instance: 95617eac-9630-4a71-80ed-a9edd8e757c1] Received unexpected event network-vif-plugged-b46e6107-621c-438e-9be6-fe1cd95a8a4e for instance with vm_state building and task_state spawning. [ 1402.097172] env[65726]: DEBUG oslo_vmware.api [None req-4f5d9381-2720-46cd-b56b-f135b25fee46 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': task-5117116, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.168746} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1402.099603] env[65726]: DEBUG nova.network.neutron [None req-c7f89e7c-d9b3-4baf-aa24-d1596cc25228 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 95617eac-9630-4a71-80ed-a9edd8e757c1] Successfully updated port: b46e6107-621c-438e-9be6-fe1cd95a8a4e {{(pid=65726) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 1402.100856] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-4f5d9381-2720-46cd-b56b-f135b25fee46 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Deleted the datastore file {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1402.101046] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-4f5d9381-2720-46cd-b56b-f135b25fee46 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 1ad4c4f9-1b02-420c-ac89-5e3390ddf773] Deleted contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1402.101570] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-4f5d9381-2720-46cd-b56b-f135b25fee46 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 1ad4c4f9-1b02-420c-ac89-5e3390ddf773] Instance destroyed {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1402.101772] env[65726]: INFO nova.compute.manager [None req-4f5d9381-2720-46cd-b56b-f135b25fee46 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 1ad4c4f9-1b02-420c-ac89-5e3390ddf773] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1402.102059] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-4f5d9381-2720-46cd-b56b-f135b25fee46 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1402.102576] env[65726]: DEBUG nova.compute.manager [-] [instance: 1ad4c4f9-1b02-420c-ac89-5e3390ddf773] Deallocating network for instance {{(pid=65726) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1402.102680] env[65726]: DEBUG nova.network.neutron [-] [instance: 1ad4c4f9-1b02-420c-ac89-5e3390ddf773] deallocate_for_instance() {{(pid=65726) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 1402.103260] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1402.103631] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1402.140681] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1402.140946] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1402.603652] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c7f89e7c-d9b3-4baf-aa24-d1596cc25228 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Acquiring lock "refresh_cache-95617eac-9630-4a71-80ed-a9edd8e757c1" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1402.603881] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c7f89e7c-d9b3-4baf-aa24-d1596cc25228 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Acquired lock "refresh_cache-95617eac-9630-4a71-80ed-a9edd8e757c1" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1402.604079] env[65726]: DEBUG nova.network.neutron [None req-c7f89e7c-d9b3-4baf-aa24-d1596cc25228 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 95617eac-9630-4a71-80ed-a9edd8e757c1] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1403.034350] env[65726]: DEBUG nova.network.neutron [-] [instance: 1ad4c4f9-1b02-420c-ac89-5e3390ddf773] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1403.107031] env[65726]: WARNING openstack [None req-c7f89e7c-d9b3-4baf-aa24-d1596cc25228 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1403.108029] env[65726]: WARNING openstack [None req-c7f89e7c-d9b3-4baf-aa24-d1596cc25228 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1403.112809] env[65726]: DEBUG nova.network.neutron [None req-c7f89e7c-d9b3-4baf-aa24-d1596cc25228 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 95617eac-9630-4a71-80ed-a9edd8e757c1] Instance cache missing network info. {{(pid=65726) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 1403.168024] env[65726]: WARNING openstack [None req-c7f89e7c-d9b3-4baf-aa24-d1596cc25228 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1403.168486] env[65726]: WARNING openstack [None req-c7f89e7c-d9b3-4baf-aa24-d1596cc25228 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1403.230920] env[65726]: WARNING openstack [None req-c7f89e7c-d9b3-4baf-aa24-d1596cc25228 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1403.231336] env[65726]: WARNING openstack [None req-c7f89e7c-d9b3-4baf-aa24-d1596cc25228 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1403.312078] env[65726]: DEBUG nova.network.neutron [None req-c7f89e7c-d9b3-4baf-aa24-d1596cc25228 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 95617eac-9630-4a71-80ed-a9edd8e757c1] Updating instance_info_cache with network_info: [{"id": "b46e6107-621c-438e-9be6-fe1cd95a8a4e", "address": "fa:16:3e:38:e6:70", "network": {"id": "336fe51f-971a-447d-8e68-505640e0db1b", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1030314703-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6b56f93ba0e14521921484ac2c785c43", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8ee1c76d-1a61-4546-85cb-d4bd3c1b35ef", "external-id": "nsx-vlan-transportzone-161", "segmentation_id": 161, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb46e6107-62", "ovs_interfaceid": "b46e6107-621c-438e-9be6-fe1cd95a8a4e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1403.538061] env[65726]: INFO nova.compute.manager [-] [instance: 1ad4c4f9-1b02-420c-ac89-5e3390ddf773] Took 1.43 seconds to deallocate network for instance. [ 1403.815972] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c7f89e7c-d9b3-4baf-aa24-d1596cc25228 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Releasing lock "refresh_cache-95617eac-9630-4a71-80ed-a9edd8e757c1" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1403.815972] env[65726]: DEBUG nova.compute.manager [None req-c7f89e7c-d9b3-4baf-aa24-d1596cc25228 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 95617eac-9630-4a71-80ed-a9edd8e757c1] Instance network_info: |[{"id": "b46e6107-621c-438e-9be6-fe1cd95a8a4e", "address": "fa:16:3e:38:e6:70", "network": {"id": "336fe51f-971a-447d-8e68-505640e0db1b", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1030314703-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6b56f93ba0e14521921484ac2c785c43", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8ee1c76d-1a61-4546-85cb-d4bd3c1b35ef", "external-id": "nsx-vlan-transportzone-161", "segmentation_id": 161, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb46e6107-62", "ovs_interfaceid": "b46e6107-621c-438e-9be6-fe1cd95a8a4e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1403.816488] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-c7f89e7c-d9b3-4baf-aa24-d1596cc25228 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 95617eac-9630-4a71-80ed-a9edd8e757c1] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:38:e6:70', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '8ee1c76d-1a61-4546-85cb-d4bd3c1b35ef', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b46e6107-621c-438e-9be6-fe1cd95a8a4e', 'vif_model': 'vmxnet3'}] {{(pid=65726) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1403.824531] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-c7f89e7c-d9b3-4baf-aa24-d1596cc25228 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1403.824756] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 95617eac-9630-4a71-80ed-a9edd8e757c1] Creating VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1403.825495] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1c6788f6-2143-4c88-874b-937852f6e564 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1403.848950] env[65726]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1403.848950] env[65726]: value = "task-5117117" [ 1403.848950] env[65726]: _type = "Task" [ 1403.848950] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1403.859142] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5117117, 'name': CreateVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1404.044210] env[65726]: DEBUG oslo_concurrency.lockutils [None req-4f5d9381-2720-46cd-b56b-f135b25fee46 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1404.044549] env[65726]: DEBUG oslo_concurrency.lockutils [None req-4f5d9381-2720-46cd-b56b-f135b25fee46 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1404.044737] env[65726]: DEBUG nova.objects.instance [None req-4f5d9381-2720-46cd-b56b-f135b25fee46 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Lazy-loading 'resources' on Instance uuid 1ad4c4f9-1b02-420c-ac89-5e3390ddf773 {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1404.049942] env[65726]: DEBUG nova.compute.manager [req-918c4a43-7149-4b18-af50-2511f57ee18c req-91543a76-b049-45e1-b9e8-c48ce8e3af50 service nova] [instance: 95617eac-9630-4a71-80ed-a9edd8e757c1] Received event network-changed-b46e6107-621c-438e-9be6-fe1cd95a8a4e {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1404.049942] env[65726]: DEBUG nova.compute.manager [req-918c4a43-7149-4b18-af50-2511f57ee18c req-91543a76-b049-45e1-b9e8-c48ce8e3af50 service nova] [instance: 95617eac-9630-4a71-80ed-a9edd8e757c1] Refreshing instance network info cache due to event network-changed-b46e6107-621c-438e-9be6-fe1cd95a8a4e. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 1404.050068] env[65726]: DEBUG oslo_concurrency.lockutils [req-918c4a43-7149-4b18-af50-2511f57ee18c req-91543a76-b049-45e1-b9e8-c48ce8e3af50 service nova] Acquiring lock "refresh_cache-95617eac-9630-4a71-80ed-a9edd8e757c1" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1404.051012] env[65726]: DEBUG oslo_concurrency.lockutils [req-918c4a43-7149-4b18-af50-2511f57ee18c req-91543a76-b049-45e1-b9e8-c48ce8e3af50 service nova] Acquired lock "refresh_cache-95617eac-9630-4a71-80ed-a9edd8e757c1" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1404.051012] env[65726]: DEBUG nova.network.neutron [req-918c4a43-7149-4b18-af50-2511f57ee18c req-91543a76-b049-45e1-b9e8-c48ce8e3af50 service nova] [instance: 95617eac-9630-4a71-80ed-a9edd8e757c1] Refreshing network info cache for port b46e6107-621c-438e-9be6-fe1cd95a8a4e {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 1404.359168] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5117117, 'name': CreateVM_Task, 'duration_secs': 0.37544} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1404.359359] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 95617eac-9630-4a71-80ed-a9edd8e757c1] Created VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1404.360123] env[65726]: WARNING openstack [None req-c7f89e7c-d9b3-4baf-aa24-d1596cc25228 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1404.360518] env[65726]: WARNING openstack [None req-c7f89e7c-d9b3-4baf-aa24-d1596cc25228 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1404.365450] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c7f89e7c-d9b3-4baf-aa24-d1596cc25228 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1404.365609] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c7f89e7c-d9b3-4baf-aa24-d1596cc25228 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1404.365962] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c7f89e7c-d9b3-4baf-aa24-d1596cc25228 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1404.366245] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0fd8d0bb-5594-497a-9cfe-7ee8ed283949 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1404.371503] env[65726]: DEBUG oslo_vmware.api [None req-c7f89e7c-d9b3-4baf-aa24-d1596cc25228 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Waiting for the task: (returnval){ [ 1404.371503] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5213723f-c395-becb-927f-1558caec691c" [ 1404.371503] env[65726]: _type = "Task" [ 1404.371503] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1404.379528] env[65726]: DEBUG oslo_vmware.api [None req-c7f89e7c-d9b3-4baf-aa24-d1596cc25228 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5213723f-c395-becb-927f-1558caec691c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1404.552893] env[65726]: WARNING openstack [req-918c4a43-7149-4b18-af50-2511f57ee18c req-91543a76-b049-45e1-b9e8-c48ce8e3af50 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1404.553301] env[65726]: WARNING openstack [req-918c4a43-7149-4b18-af50-2511f57ee18c req-91543a76-b049-45e1-b9e8-c48ce8e3af50 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1404.613534] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7cd901f6-bdb9-46ab-b492-c368bcf60333 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1404.621842] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ad1967e-197b-460c-bcfd-3f6b9648af45 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1404.655700] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0dd64f96-6af6-4b28-b378-f440f1eeaea4 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1404.663547] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9c8f9d5-f44d-4a35-8f2e-549fdf4359ca {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1404.678159] env[65726]: DEBUG nova.compute.provider_tree [None req-4f5d9381-2720-46cd-b56b-f135b25fee46 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1404.680525] env[65726]: WARNING openstack [req-918c4a43-7149-4b18-af50-2511f57ee18c req-91543a76-b049-45e1-b9e8-c48ce8e3af50 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1404.680876] env[65726]: WARNING openstack [req-918c4a43-7149-4b18-af50-2511f57ee18c req-91543a76-b049-45e1-b9e8-c48ce8e3af50 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1404.740565] env[65726]: WARNING openstack [req-918c4a43-7149-4b18-af50-2511f57ee18c req-91543a76-b049-45e1-b9e8-c48ce8e3af50 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1404.740987] env[65726]: WARNING openstack [req-918c4a43-7149-4b18-af50-2511f57ee18c req-91543a76-b049-45e1-b9e8-c48ce8e3af50 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1404.818315] env[65726]: DEBUG nova.network.neutron [req-918c4a43-7149-4b18-af50-2511f57ee18c req-91543a76-b049-45e1-b9e8-c48ce8e3af50 service nova] [instance: 95617eac-9630-4a71-80ed-a9edd8e757c1] Updated VIF entry in instance network info cache for port b46e6107-621c-438e-9be6-fe1cd95a8a4e. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 1404.818671] env[65726]: DEBUG nova.network.neutron [req-918c4a43-7149-4b18-af50-2511f57ee18c req-91543a76-b049-45e1-b9e8-c48ce8e3af50 service nova] [instance: 95617eac-9630-4a71-80ed-a9edd8e757c1] Updating instance_info_cache with network_info: [{"id": "b46e6107-621c-438e-9be6-fe1cd95a8a4e", "address": "fa:16:3e:38:e6:70", "network": {"id": "336fe51f-971a-447d-8e68-505640e0db1b", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1030314703-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6b56f93ba0e14521921484ac2c785c43", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8ee1c76d-1a61-4546-85cb-d4bd3c1b35ef", "external-id": "nsx-vlan-transportzone-161", "segmentation_id": 161, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb46e6107-62", "ovs_interfaceid": "b46e6107-621c-438e-9be6-fe1cd95a8a4e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1404.882720] env[65726]: DEBUG oslo_vmware.api [None req-c7f89e7c-d9b3-4baf-aa24-d1596cc25228 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5213723f-c395-becb-927f-1558caec691c, 'name': SearchDatastore_Task, 'duration_secs': 0.011575} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1404.883120] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c7f89e7c-d9b3-4baf-aa24-d1596cc25228 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1404.883417] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-c7f89e7c-d9b3-4baf-aa24-d1596cc25228 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 95617eac-9630-4a71-80ed-a9edd8e757c1] Processing image b52362a3-ee8a-4cbf-b06f-513b0cc8f95c {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1404.883661] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c7f89e7c-d9b3-4baf-aa24-d1596cc25228 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1404.883829] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c7f89e7c-d9b3-4baf-aa24-d1596cc25228 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1404.884012] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-c7f89e7c-d9b3-4baf-aa24-d1596cc25228 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1404.884310] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-eac90e17-3fe4-4eb3-8bb8-100a8907a85f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1404.893701] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-c7f89e7c-d9b3-4baf-aa24-d1596cc25228 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1404.893856] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-c7f89e7c-d9b3-4baf-aa24-d1596cc25228 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65726) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1404.894601] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4d77b017-32c6-4361-8eb2-b18ecbe25211 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1404.899979] env[65726]: DEBUG oslo_vmware.api [None req-c7f89e7c-d9b3-4baf-aa24-d1596cc25228 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Waiting for the task: (returnval){ [ 1404.899979] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]528a5120-2704-2389-34e7-51971629089a" [ 1404.899979] env[65726]: _type = "Task" [ 1404.899979] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1404.907647] env[65726]: DEBUG oslo_vmware.api [None req-c7f89e7c-d9b3-4baf-aa24-d1596cc25228 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]528a5120-2704-2389-34e7-51971629089a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1405.188822] env[65726]: DEBUG nova.scheduler.client.report [None req-4f5d9381-2720-46cd-b56b-f135b25fee46 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1405.321567] env[65726]: DEBUG oslo_concurrency.lockutils [req-918c4a43-7149-4b18-af50-2511f57ee18c req-91543a76-b049-45e1-b9e8-c48ce8e3af50 service nova] Releasing lock "refresh_cache-95617eac-9630-4a71-80ed-a9edd8e757c1" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1405.321849] env[65726]: DEBUG nova.compute.manager [req-918c4a43-7149-4b18-af50-2511f57ee18c req-91543a76-b049-45e1-b9e8-c48ce8e3af50 service nova] [instance: 1ad4c4f9-1b02-420c-ac89-5e3390ddf773] Received event network-vif-deleted-3aa1b8d0-3927-41ed-903b-212dc4ad95b6 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1405.411213] env[65726]: DEBUG oslo_vmware.api [None req-c7f89e7c-d9b3-4baf-aa24-d1596cc25228 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]528a5120-2704-2389-34e7-51971629089a, 'name': SearchDatastore_Task, 'duration_secs': 0.012273} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1405.412016] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4cf6ba89-5406-4761-8be0-7a8516a9cfc1 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1405.417862] env[65726]: DEBUG oslo_vmware.api [None req-c7f89e7c-d9b3-4baf-aa24-d1596cc25228 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Waiting for the task: (returnval){ [ 1405.417862] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5277e958-63cf-d681-daea-def0fa53ff28" [ 1405.417862] env[65726]: _type = "Task" [ 1405.417862] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1405.426191] env[65726]: DEBUG oslo_vmware.api [None req-c7f89e7c-d9b3-4baf-aa24-d1596cc25228 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5277e958-63cf-d681-daea-def0fa53ff28, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1405.694054] env[65726]: DEBUG oslo_concurrency.lockutils [None req-4f5d9381-2720-46cd-b56b-f135b25fee46 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.649s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1405.712619] env[65726]: INFO nova.scheduler.client.report [None req-4f5d9381-2720-46cd-b56b-f135b25fee46 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Deleted allocations for instance 1ad4c4f9-1b02-420c-ac89-5e3390ddf773 [ 1405.929103] env[65726]: DEBUG oslo_vmware.api [None req-c7f89e7c-d9b3-4baf-aa24-d1596cc25228 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5277e958-63cf-d681-daea-def0fa53ff28, 'name': SearchDatastore_Task, 'duration_secs': 0.013943} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1405.929375] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c7f89e7c-d9b3-4baf-aa24-d1596cc25228 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1405.929630] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-c7f89e7c-d9b3-4baf-aa24-d1596cc25228 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] 95617eac-9630-4a71-80ed-a9edd8e757c1/95617eac-9630-4a71-80ed-a9edd8e757c1.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1405.929901] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6faf3067-0e7f-4eb5-99ca-15d18c66f17c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1405.936956] env[65726]: DEBUG oslo_vmware.api [None req-c7f89e7c-d9b3-4baf-aa24-d1596cc25228 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Waiting for the task: (returnval){ [ 1405.936956] env[65726]: value = "task-5117118" [ 1405.936956] env[65726]: _type = "Task" [ 1405.936956] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1405.945433] env[65726]: DEBUG oslo_vmware.api [None req-c7f89e7c-d9b3-4baf-aa24-d1596cc25228 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': task-5117118, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1406.221061] env[65726]: DEBUG oslo_concurrency.lockutils [None req-4f5d9381-2720-46cd-b56b-f135b25fee46 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Lock "1ad4c4f9-1b02-420c-ac89-5e3390ddf773" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.752s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1406.449412] env[65726]: DEBUG oslo_vmware.api [None req-c7f89e7c-d9b3-4baf-aa24-d1596cc25228 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': task-5117118, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1406.949976] env[65726]: DEBUG oslo_vmware.api [None req-c7f89e7c-d9b3-4baf-aa24-d1596cc25228 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': task-5117118, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.55418} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1406.950260] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-c7f89e7c-d9b3-4baf-aa24-d1596cc25228 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore1] 95617eac-9630-4a71-80ed-a9edd8e757c1/95617eac-9630-4a71-80ed-a9edd8e757c1.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1406.950695] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-c7f89e7c-d9b3-4baf-aa24-d1596cc25228 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 95617eac-9630-4a71-80ed-a9edd8e757c1] Extending root virtual disk to 1048576 {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1406.950805] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2ba38820-ac39-4ed6-b6ce-da8ed985ae65 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1406.958780] env[65726]: DEBUG oslo_vmware.api [None req-c7f89e7c-d9b3-4baf-aa24-d1596cc25228 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Waiting for the task: (returnval){ [ 1406.958780] env[65726]: value = "task-5117119" [ 1406.958780] env[65726]: _type = "Task" [ 1406.958780] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1406.969371] env[65726]: DEBUG oslo_vmware.api [None req-c7f89e7c-d9b3-4baf-aa24-d1596cc25228 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': task-5117119, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1407.468732] env[65726]: DEBUG oslo_vmware.api [None req-c7f89e7c-d9b3-4baf-aa24-d1596cc25228 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': task-5117119, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.077937} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1407.469116] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-c7f89e7c-d9b3-4baf-aa24-d1596cc25228 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 95617eac-9630-4a71-80ed-a9edd8e757c1] Extended root virtual disk {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1407.469955] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71b900cf-f17d-48ba-9a14-1895768e7ced {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1407.493331] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-c7f89e7c-d9b3-4baf-aa24-d1596cc25228 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 95617eac-9630-4a71-80ed-a9edd8e757c1] Reconfiguring VM instance instance-0000007b to attach disk [datastore1] 95617eac-9630-4a71-80ed-a9edd8e757c1/95617eac-9630-4a71-80ed-a9edd8e757c1.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1407.493703] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e4088393-f6cd-44b3-b921-df2a1831bd01 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1407.514478] env[65726]: DEBUG oslo_vmware.api [None req-c7f89e7c-d9b3-4baf-aa24-d1596cc25228 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Waiting for the task: (returnval){ [ 1407.514478] env[65726]: value = "task-5117120" [ 1407.514478] env[65726]: _type = "Task" [ 1407.514478] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1407.523135] env[65726]: DEBUG oslo_vmware.api [None req-c7f89e7c-d9b3-4baf-aa24-d1596cc25228 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': task-5117120, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1408.025394] env[65726]: DEBUG oslo_vmware.api [None req-c7f89e7c-d9b3-4baf-aa24-d1596cc25228 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': task-5117120, 'name': ReconfigVM_Task, 'duration_secs': 0.339104} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1408.025609] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-c7f89e7c-d9b3-4baf-aa24-d1596cc25228 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 95617eac-9630-4a71-80ed-a9edd8e757c1] Reconfigured VM instance instance-0000007b to attach disk [datastore1] 95617eac-9630-4a71-80ed-a9edd8e757c1/95617eac-9630-4a71-80ed-a9edd8e757c1.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1408.026431] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-5271da51-f4ef-4a6a-8bf1-283f0dad7ae5 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1408.035119] env[65726]: DEBUG oslo_vmware.api [None req-c7f89e7c-d9b3-4baf-aa24-d1596cc25228 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Waiting for the task: (returnval){ [ 1408.035119] env[65726]: value = "task-5117121" [ 1408.035119] env[65726]: _type = "Task" [ 1408.035119] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1408.043265] env[65726]: DEBUG oslo_vmware.api [None req-c7f89e7c-d9b3-4baf-aa24-d1596cc25228 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': task-5117121, 'name': Rename_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1408.547586] env[65726]: DEBUG oslo_vmware.api [None req-c7f89e7c-d9b3-4baf-aa24-d1596cc25228 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': task-5117121, 'name': Rename_Task, 'duration_secs': 0.161663} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1408.547966] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-c7f89e7c-d9b3-4baf-aa24-d1596cc25228 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 95617eac-9630-4a71-80ed-a9edd8e757c1] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1408.548303] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-942248aa-a9da-4f43-8ed8-9b5d5273431e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1408.556053] env[65726]: DEBUG oslo_vmware.api [None req-c7f89e7c-d9b3-4baf-aa24-d1596cc25228 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Waiting for the task: (returnval){ [ 1408.556053] env[65726]: value = "task-5117122" [ 1408.556053] env[65726]: _type = "Task" [ 1408.556053] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1408.565198] env[65726]: DEBUG oslo_vmware.api [None req-c7f89e7c-d9b3-4baf-aa24-d1596cc25228 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': task-5117122, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1408.613301] env[65726]: DEBUG nova.compute.manager [None req-76276b65-b10c-4377-a4b5-d8f16235a93a tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: e3255f3b-028f-4a0b-b621-873730417936] Stashing vm_state: active {{(pid=65726) _prep_resize /opt/stack/nova/nova/compute/manager.py:6193}} [ 1409.067065] env[65726]: DEBUG oslo_vmware.api [None req-c7f89e7c-d9b3-4baf-aa24-d1596cc25228 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': task-5117122, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1409.133862] env[65726]: DEBUG oslo_concurrency.lockutils [None req-76276b65-b10c-4377-a4b5-d8f16235a93a tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1409.134181] env[65726]: DEBUG oslo_concurrency.lockutils [None req-76276b65-b10c-4377-a4b5-d8f16235a93a tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1409.566317] env[65726]: DEBUG oslo_vmware.api [None req-c7f89e7c-d9b3-4baf-aa24-d1596cc25228 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': task-5117122, 'name': PowerOnVM_Task, 'duration_secs': 0.524027} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1409.566703] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-c7f89e7c-d9b3-4baf-aa24-d1596cc25228 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 95617eac-9630-4a71-80ed-a9edd8e757c1] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1409.566703] env[65726]: INFO nova.compute.manager [None req-c7f89e7c-d9b3-4baf-aa24-d1596cc25228 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 95617eac-9630-4a71-80ed-a9edd8e757c1] Took 7.73 seconds to spawn the instance on the hypervisor. [ 1409.566896] env[65726]: DEBUG nova.compute.manager [None req-c7f89e7c-d9b3-4baf-aa24-d1596cc25228 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 95617eac-9630-4a71-80ed-a9edd8e757c1] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1409.567734] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4113f08f-fae9-4070-930f-eec0f3a8affb {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1409.639423] env[65726]: INFO nova.compute.claims [None req-76276b65-b10c-4377-a4b5-d8f16235a93a tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: e3255f3b-028f-4a0b-b621-873730417936] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1410.086063] env[65726]: INFO nova.compute.manager [None req-c7f89e7c-d9b3-4baf-aa24-d1596cc25228 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 95617eac-9630-4a71-80ed-a9edd8e757c1] Took 12.44 seconds to build instance. [ 1410.146070] env[65726]: INFO nova.compute.resource_tracker [None req-76276b65-b10c-4377-a4b5-d8f16235a93a tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: e3255f3b-028f-4a0b-b621-873730417936] Updating resource usage from migration ac4cf9cd-eee7-48ff-8287-03052d30015d [ 1410.209374] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa10f2d8-85b6-4fdf-b06c-83a77412296f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1410.217902] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ce8d549-ae68-418e-ad6b-a0ce1ef916b9 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1410.248371] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b5cff2d-179f-454e-a069-4c2ace542bbd {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1410.256542] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9823410f-3dc9-4e72-b60b-80f0e3aea33c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1410.270236] env[65726]: DEBUG nova.compute.provider_tree [None req-76276b65-b10c-4377-a4b5-d8f16235a93a tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1410.588843] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c7f89e7c-d9b3-4baf-aa24-d1596cc25228 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Lock "95617eac-9630-4a71-80ed-a9edd8e757c1" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.947s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1410.773852] env[65726]: DEBUG nova.scheduler.client.report [None req-76276b65-b10c-4377-a4b5-d8f16235a93a tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1411.278736] env[65726]: DEBUG oslo_concurrency.lockutils [None req-76276b65-b10c-4377-a4b5-d8f16235a93a tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.144s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1411.278950] env[65726]: INFO nova.compute.manager [None req-76276b65-b10c-4377-a4b5-d8f16235a93a tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: e3255f3b-028f-4a0b-b621-873730417936] Migrating [ 1411.796068] env[65726]: DEBUG oslo_concurrency.lockutils [None req-76276b65-b10c-4377-a4b5-d8f16235a93a tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Acquiring lock "refresh_cache-e3255f3b-028f-4a0b-b621-873730417936" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1411.796068] env[65726]: DEBUG oslo_concurrency.lockutils [None req-76276b65-b10c-4377-a4b5-d8f16235a93a tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Acquired lock "refresh_cache-e3255f3b-028f-4a0b-b621-873730417936" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1411.796465] env[65726]: DEBUG nova.network.neutron [None req-76276b65-b10c-4377-a4b5-d8f16235a93a tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: e3255f3b-028f-4a0b-b621-873730417936] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1412.299531] env[65726]: WARNING openstack [None req-76276b65-b10c-4377-a4b5-d8f16235a93a tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1412.299872] env[65726]: WARNING openstack [None req-76276b65-b10c-4377-a4b5-d8f16235a93a tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1412.420684] env[65726]: WARNING openstack [None req-76276b65-b10c-4377-a4b5-d8f16235a93a tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1412.421099] env[65726]: WARNING openstack [None req-76276b65-b10c-4377-a4b5-d8f16235a93a tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1412.489296] env[65726]: WARNING openstack [None req-76276b65-b10c-4377-a4b5-d8f16235a93a tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1412.489673] env[65726]: WARNING openstack [None req-76276b65-b10c-4377-a4b5-d8f16235a93a tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1412.570423] env[65726]: DEBUG nova.network.neutron [None req-76276b65-b10c-4377-a4b5-d8f16235a93a tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: e3255f3b-028f-4a0b-b621-873730417936] Updating instance_info_cache with network_info: [{"id": "03c9e6fb-2435-43be-bb55-8afecdcf38ff", "address": "fa:16:3e:d2:0f:96", "network": {"id": "a899453d-5f43-4650-9642-b33f54a4d2eb", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-614861139-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.231", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "acd800e0438940bc80ae3c6b672db4ee", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "205fb402-8eaf-4b61-8f57-8f216024179a", "external-id": "nsx-vlan-transportzone-78", "segmentation_id": 78, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap03c9e6fb-24", "ovs_interfaceid": "03c9e6fb-2435-43be-bb55-8afecdcf38ff", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1413.073267] env[65726]: DEBUG oslo_concurrency.lockutils [None req-76276b65-b10c-4377-a4b5-d8f16235a93a tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Releasing lock "refresh_cache-e3255f3b-028f-4a0b-b621-873730417936" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1414.589123] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a01c4a58-de64-487c-9d1e-ad36983c141d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1414.607592] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-76276b65-b10c-4377-a4b5-d8f16235a93a tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: e3255f3b-028f-4a0b-b621-873730417936] Updating instance 'e3255f3b-028f-4a0b-b621-873730417936' progress to 0 {{(pid=65726) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1415.114607] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-76276b65-b10c-4377-a4b5-d8f16235a93a tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: e3255f3b-028f-4a0b-b621-873730417936] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1415.114955] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fc0d200a-5ade-430a-be36-6eacc8d8a3f6 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1415.124156] env[65726]: DEBUG oslo_vmware.api [None req-76276b65-b10c-4377-a4b5-d8f16235a93a tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Waiting for the task: (returnval){ [ 1415.124156] env[65726]: value = "task-5117125" [ 1415.124156] env[65726]: _type = "Task" [ 1415.124156] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1415.132897] env[65726]: DEBUG oslo_vmware.api [None req-76276b65-b10c-4377-a4b5-d8f16235a93a tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': task-5117125, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1415.635278] env[65726]: DEBUG oslo_vmware.api [None req-76276b65-b10c-4377-a4b5-d8f16235a93a tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': task-5117125, 'name': PowerOffVM_Task, 'duration_secs': 0.223287} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1415.635671] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-76276b65-b10c-4377-a4b5-d8f16235a93a tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: e3255f3b-028f-4a0b-b621-873730417936] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1415.635671] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-76276b65-b10c-4377-a4b5-d8f16235a93a tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: e3255f3b-028f-4a0b-b621-873730417936] Updating instance 'e3255f3b-028f-4a0b-b621-873730417936' progress to 17 {{(pid=65726) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1416.141945] env[65726]: DEBUG nova.virt.hardware [None req-76276b65-b10c-4377-a4b5-d8f16235a93a tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1416.142258] env[65726]: DEBUG nova.virt.hardware [None req-76276b65-b10c-4377-a4b5-d8f16235a93a tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1416.142427] env[65726]: DEBUG nova.virt.hardware [None req-76276b65-b10c-4377-a4b5-d8f16235a93a tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1416.142604] env[65726]: DEBUG nova.virt.hardware [None req-76276b65-b10c-4377-a4b5-d8f16235a93a tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1416.142743] env[65726]: DEBUG nova.virt.hardware [None req-76276b65-b10c-4377-a4b5-d8f16235a93a tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1416.142881] env[65726]: DEBUG nova.virt.hardware [None req-76276b65-b10c-4377-a4b5-d8f16235a93a tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1416.143163] env[65726]: DEBUG nova.virt.hardware [None req-76276b65-b10c-4377-a4b5-d8f16235a93a tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1416.143349] env[65726]: DEBUG nova.virt.hardware [None req-76276b65-b10c-4377-a4b5-d8f16235a93a tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1416.143617] env[65726]: DEBUG nova.virt.hardware [None req-76276b65-b10c-4377-a4b5-d8f16235a93a tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1416.143861] env[65726]: DEBUG nova.virt.hardware [None req-76276b65-b10c-4377-a4b5-d8f16235a93a tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1416.144180] env[65726]: DEBUG nova.virt.hardware [None req-76276b65-b10c-4377-a4b5-d8f16235a93a tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1416.149344] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d0062f4a-4876-48fc-bdc9-2f79f550ac96 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1416.167164] env[65726]: DEBUG oslo_vmware.api [None req-76276b65-b10c-4377-a4b5-d8f16235a93a tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Waiting for the task: (returnval){ [ 1416.167164] env[65726]: value = "task-5117126" [ 1416.167164] env[65726]: _type = "Task" [ 1416.167164] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1416.175753] env[65726]: DEBUG oslo_vmware.api [None req-76276b65-b10c-4377-a4b5-d8f16235a93a tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': task-5117126, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1416.680980] env[65726]: DEBUG oslo_vmware.api [None req-76276b65-b10c-4377-a4b5-d8f16235a93a tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': task-5117126, 'name': ReconfigVM_Task, 'duration_secs': 0.411562} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1416.681482] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-76276b65-b10c-4377-a4b5-d8f16235a93a tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: e3255f3b-028f-4a0b-b621-873730417936] Updating instance 'e3255f3b-028f-4a0b-b621-873730417936' progress to 33 {{(pid=65726) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1417.189866] env[65726]: DEBUG nova.virt.hardware [None req-76276b65-b10c-4377-a4b5-d8f16235a93a tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1417.190166] env[65726]: DEBUG nova.virt.hardware [None req-76276b65-b10c-4377-a4b5-d8f16235a93a tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1417.190385] env[65726]: DEBUG nova.virt.hardware [None req-76276b65-b10c-4377-a4b5-d8f16235a93a tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1417.190603] env[65726]: DEBUG nova.virt.hardware [None req-76276b65-b10c-4377-a4b5-d8f16235a93a tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1417.190778] env[65726]: DEBUG nova.virt.hardware [None req-76276b65-b10c-4377-a4b5-d8f16235a93a tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1417.190955] env[65726]: DEBUG nova.virt.hardware [None req-76276b65-b10c-4377-a4b5-d8f16235a93a tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1417.191214] env[65726]: DEBUG nova.virt.hardware [None req-76276b65-b10c-4377-a4b5-d8f16235a93a tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1417.191433] env[65726]: DEBUG nova.virt.hardware [None req-76276b65-b10c-4377-a4b5-d8f16235a93a tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1417.191643] env[65726]: DEBUG nova.virt.hardware [None req-76276b65-b10c-4377-a4b5-d8f16235a93a tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1417.191861] env[65726]: DEBUG nova.virt.hardware [None req-76276b65-b10c-4377-a4b5-d8f16235a93a tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1417.192087] env[65726]: DEBUG nova.virt.hardware [None req-76276b65-b10c-4377-a4b5-d8f16235a93a tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1417.197408] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-76276b65-b10c-4377-a4b5-d8f16235a93a tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: e3255f3b-028f-4a0b-b621-873730417936] Reconfiguring VM instance instance-0000006f to detach disk 2000 {{(pid=65726) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1417.197704] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8e8f818b-692c-426b-a58c-957675331a3b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1417.217471] env[65726]: DEBUG oslo_vmware.api [None req-76276b65-b10c-4377-a4b5-d8f16235a93a tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Waiting for the task: (returnval){ [ 1417.217471] env[65726]: value = "task-5117128" [ 1417.217471] env[65726]: _type = "Task" [ 1417.217471] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1417.225467] env[65726]: DEBUG oslo_vmware.api [None req-76276b65-b10c-4377-a4b5-d8f16235a93a tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': task-5117128, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1417.728172] env[65726]: DEBUG oslo_vmware.api [None req-76276b65-b10c-4377-a4b5-d8f16235a93a tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': task-5117128, 'name': ReconfigVM_Task, 'duration_secs': 0.185177} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1417.728642] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-76276b65-b10c-4377-a4b5-d8f16235a93a tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: e3255f3b-028f-4a0b-b621-873730417936] Reconfigured VM instance instance-0000006f to detach disk 2000 {{(pid=65726) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1417.729267] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef9be1b4-b2ac-4386-bc1d-b7dc02a4b574 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1417.751479] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-76276b65-b10c-4377-a4b5-d8f16235a93a tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: e3255f3b-028f-4a0b-b621-873730417936] Reconfiguring VM instance instance-0000006f to attach disk [datastore2] e3255f3b-028f-4a0b-b621-873730417936/e3255f3b-028f-4a0b-b621-873730417936.vmdk or device None with type thin {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1417.751713] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9b239987-6c3a-4356-b364-afcf61c0a3c9 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1417.770112] env[65726]: DEBUG oslo_vmware.api [None req-76276b65-b10c-4377-a4b5-d8f16235a93a tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Waiting for the task: (returnval){ [ 1417.770112] env[65726]: value = "task-5117129" [ 1417.770112] env[65726]: _type = "Task" [ 1417.770112] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1417.777849] env[65726]: DEBUG oslo_vmware.api [None req-76276b65-b10c-4377-a4b5-d8f16235a93a tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': task-5117129, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1418.280945] env[65726]: DEBUG oslo_vmware.api [None req-76276b65-b10c-4377-a4b5-d8f16235a93a tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': task-5117129, 'name': ReconfigVM_Task, 'duration_secs': 0.261207} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1418.281234] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-76276b65-b10c-4377-a4b5-d8f16235a93a tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: e3255f3b-028f-4a0b-b621-873730417936] Reconfigured VM instance instance-0000006f to attach disk [datastore2] e3255f3b-028f-4a0b-b621-873730417936/e3255f3b-028f-4a0b-b621-873730417936.vmdk or device None with type thin {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1418.281489] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-76276b65-b10c-4377-a4b5-d8f16235a93a tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: e3255f3b-028f-4a0b-b621-873730417936] Updating instance 'e3255f3b-028f-4a0b-b621-873730417936' progress to 50 {{(pid=65726) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1418.789373] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0a6d806-7a04-4343-ab50-51ba00b8e127 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1418.810055] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4fdece0-05d2-417c-ae7b-de89193d4845 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1418.828244] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-76276b65-b10c-4377-a4b5-d8f16235a93a tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: e3255f3b-028f-4a0b-b621-873730417936] Updating instance 'e3255f3b-028f-4a0b-b621-873730417936' progress to 67 {{(pid=65726) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1419.334612] env[65726]: WARNING neutronclient.v2_0.client [None req-76276b65-b10c-4377-a4b5-d8f16235a93a tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1419.370939] env[65726]: DEBUG nova.network.neutron [None req-76276b65-b10c-4377-a4b5-d8f16235a93a tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: e3255f3b-028f-4a0b-b621-873730417936] Port 03c9e6fb-2435-43be-bb55-8afecdcf38ff binding to destination host cpu-1 is already ACTIVE {{(pid=65726) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3236}} [ 1420.395456] env[65726]: DEBUG oslo_concurrency.lockutils [None req-76276b65-b10c-4377-a4b5-d8f16235a93a tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Acquiring lock "e3255f3b-028f-4a0b-b621-873730417936-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1420.395829] env[65726]: DEBUG oslo_concurrency.lockutils [None req-76276b65-b10c-4377-a4b5-d8f16235a93a tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Lock "e3255f3b-028f-4a0b-b621-873730417936-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1420.395829] env[65726]: DEBUG oslo_concurrency.lockutils [None req-76276b65-b10c-4377-a4b5-d8f16235a93a tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Lock "e3255f3b-028f-4a0b-b621-873730417936-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1421.400545] env[65726]: WARNING openstack [None req-76276b65-b10c-4377-a4b5-d8f16235a93a tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1421.400948] env[65726]: WARNING openstack [None req-76276b65-b10c-4377-a4b5-d8f16235a93a tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1421.436700] env[65726]: DEBUG oslo_concurrency.lockutils [None req-76276b65-b10c-4377-a4b5-d8f16235a93a tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Acquiring lock "refresh_cache-e3255f3b-028f-4a0b-b621-873730417936" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1421.436875] env[65726]: DEBUG oslo_concurrency.lockutils [None req-76276b65-b10c-4377-a4b5-d8f16235a93a tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Acquired lock "refresh_cache-e3255f3b-028f-4a0b-b621-873730417936" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1421.437079] env[65726]: DEBUG nova.network.neutron [None req-76276b65-b10c-4377-a4b5-d8f16235a93a tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: e3255f3b-028f-4a0b-b621-873730417936] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1421.940275] env[65726]: WARNING openstack [None req-76276b65-b10c-4377-a4b5-d8f16235a93a tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1421.940648] env[65726]: WARNING openstack [None req-76276b65-b10c-4377-a4b5-d8f16235a93a tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1422.071080] env[65726]: WARNING openstack [None req-76276b65-b10c-4377-a4b5-d8f16235a93a tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1422.071490] env[65726]: WARNING openstack [None req-76276b65-b10c-4377-a4b5-d8f16235a93a tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1422.144940] env[65726]: WARNING openstack [None req-76276b65-b10c-4377-a4b5-d8f16235a93a tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1422.145357] env[65726]: WARNING openstack [None req-76276b65-b10c-4377-a4b5-d8f16235a93a tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1422.228847] env[65726]: DEBUG nova.network.neutron [None req-76276b65-b10c-4377-a4b5-d8f16235a93a tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: e3255f3b-028f-4a0b-b621-873730417936] Updating instance_info_cache with network_info: [{"id": "03c9e6fb-2435-43be-bb55-8afecdcf38ff", "address": "fa:16:3e:d2:0f:96", "network": {"id": "a899453d-5f43-4650-9642-b33f54a4d2eb", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-614861139-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.231", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "acd800e0438940bc80ae3c6b672db4ee", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "205fb402-8eaf-4b61-8f57-8f216024179a", "external-id": "nsx-vlan-transportzone-78", "segmentation_id": 78, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap03c9e6fb-24", "ovs_interfaceid": "03c9e6fb-2435-43be-bb55-8afecdcf38ff", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1422.731722] env[65726]: DEBUG oslo_concurrency.lockutils [None req-76276b65-b10c-4377-a4b5-d8f16235a93a tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Releasing lock "refresh_cache-e3255f3b-028f-4a0b-b621-873730417936" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1423.255597] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0db479a-709b-4dc1-8bd7-4f82beee4ecb {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1423.275535] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a011ec4-b97e-4de5-94d4-1761b4c941f6 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1423.283162] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-76276b65-b10c-4377-a4b5-d8f16235a93a tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: e3255f3b-028f-4a0b-b621-873730417936] Updating instance 'e3255f3b-028f-4a0b-b621-873730417936' progress to 83 {{(pid=65726) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1423.789729] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-76276b65-b10c-4377-a4b5-d8f16235a93a tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: e3255f3b-028f-4a0b-b621-873730417936] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1423.790144] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b0ba8373-9152-401a-8867-57c3d9780955 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1423.798547] env[65726]: DEBUG oslo_vmware.api [None req-76276b65-b10c-4377-a4b5-d8f16235a93a tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Waiting for the task: (returnval){ [ 1423.798547] env[65726]: value = "task-5117133" [ 1423.798547] env[65726]: _type = "Task" [ 1423.798547] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1423.807041] env[65726]: DEBUG oslo_vmware.api [None req-76276b65-b10c-4377-a4b5-d8f16235a93a tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': task-5117133, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1424.309075] env[65726]: DEBUG oslo_vmware.api [None req-76276b65-b10c-4377-a4b5-d8f16235a93a tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': task-5117133, 'name': PowerOnVM_Task, 'duration_secs': 0.381174} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1424.309380] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-76276b65-b10c-4377-a4b5-d8f16235a93a tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: e3255f3b-028f-4a0b-b621-873730417936] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1424.309476] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-76276b65-b10c-4377-a4b5-d8f16235a93a tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: e3255f3b-028f-4a0b-b621-873730417936] Updating instance 'e3255f3b-028f-4a0b-b621-873730417936' progress to 100 {{(pid=65726) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1426.843730] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c04eeb3b-5e5d-4ea6-a41e-462912417015 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Acquiring lock "e3255f3b-028f-4a0b-b621-873730417936" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1426.844209] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c04eeb3b-5e5d-4ea6-a41e-462912417015 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Lock "e3255f3b-028f-4a0b-b621-873730417936" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1426.844270] env[65726]: DEBUG nova.compute.manager [None req-c04eeb3b-5e5d-4ea6-a41e-462912417015 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: e3255f3b-028f-4a0b-b621-873730417936] Going to confirm migration 8 {{(pid=65726) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:5307}} [ 1427.350303] env[65726]: WARNING openstack [None req-c04eeb3b-5e5d-4ea6-a41e-462912417015 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1427.350616] env[65726]: WARNING openstack [None req-c04eeb3b-5e5d-4ea6-a41e-462912417015 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1427.387304] env[65726]: WARNING openstack [None req-c04eeb3b-5e5d-4ea6-a41e-462912417015 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1427.387695] env[65726]: WARNING openstack [None req-c04eeb3b-5e5d-4ea6-a41e-462912417015 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1427.392461] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c04eeb3b-5e5d-4ea6-a41e-462912417015 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Acquiring lock "refresh_cache-e3255f3b-028f-4a0b-b621-873730417936" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1427.392620] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c04eeb3b-5e5d-4ea6-a41e-462912417015 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Acquired lock "refresh_cache-e3255f3b-028f-4a0b-b621-873730417936" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1427.392784] env[65726]: DEBUG nova.network.neutron [None req-c04eeb3b-5e5d-4ea6-a41e-462912417015 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: e3255f3b-028f-4a0b-b621-873730417936] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1427.392955] env[65726]: DEBUG nova.objects.instance [None req-c04eeb3b-5e5d-4ea6-a41e-462912417015 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Lazy-loading 'info_cache' on Instance uuid e3255f3b-028f-4a0b-b621-873730417936 {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1428.400247] env[65726]: WARNING openstack [None req-c04eeb3b-5e5d-4ea6-a41e-462912417015 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1428.400684] env[65726]: WARNING openstack [None req-c04eeb3b-5e5d-4ea6-a41e-462912417015 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1428.535634] env[65726]: WARNING openstack [None req-c04eeb3b-5e5d-4ea6-a41e-462912417015 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1428.536079] env[65726]: WARNING openstack [None req-c04eeb3b-5e5d-4ea6-a41e-462912417015 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1428.600554] env[65726]: WARNING openstack [None req-c04eeb3b-5e5d-4ea6-a41e-462912417015 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1428.600935] env[65726]: WARNING openstack [None req-c04eeb3b-5e5d-4ea6-a41e-462912417015 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1428.682052] env[65726]: DEBUG nova.network.neutron [None req-c04eeb3b-5e5d-4ea6-a41e-462912417015 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: e3255f3b-028f-4a0b-b621-873730417936] Updating instance_info_cache with network_info: [{"id": "03c9e6fb-2435-43be-bb55-8afecdcf38ff", "address": "fa:16:3e:d2:0f:96", "network": {"id": "a899453d-5f43-4650-9642-b33f54a4d2eb", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-614861139-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.231", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "acd800e0438940bc80ae3c6b672db4ee", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "205fb402-8eaf-4b61-8f57-8f216024179a", "external-id": "nsx-vlan-transportzone-78", "segmentation_id": 78, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap03c9e6fb-24", "ovs_interfaceid": "03c9e6fb-2435-43be-bb55-8afecdcf38ff", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1429.185031] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c04eeb3b-5e5d-4ea6-a41e-462912417015 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Releasing lock "refresh_cache-e3255f3b-028f-4a0b-b621-873730417936" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1429.185319] env[65726]: DEBUG nova.objects.instance [None req-c04eeb3b-5e5d-4ea6-a41e-462912417015 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Lazy-loading 'migration_context' on Instance uuid e3255f3b-028f-4a0b-b621-873730417936 {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1429.688934] env[65726]: DEBUG nova.objects.base [None req-c04eeb3b-5e5d-4ea6-a41e-462912417015 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Object Instance lazy-loaded attributes: info_cache,migration_context {{(pid=65726) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1429.689990] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48f8c488-a29d-44ae-9fec-14635e834cca {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1429.709981] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6e1480b1-f2f7-41b3-8758-8953f1596ae0 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1429.716157] env[65726]: DEBUG oslo_vmware.api [None req-c04eeb3b-5e5d-4ea6-a41e-462912417015 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Waiting for the task: (returnval){ [ 1429.716157] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]528eb1d9-1b23-1a67-f120-0494dba8227a" [ 1429.716157] env[65726]: _type = "Task" [ 1429.716157] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1429.724789] env[65726]: DEBUG oslo_vmware.api [None req-c04eeb3b-5e5d-4ea6-a41e-462912417015 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]528eb1d9-1b23-1a67-f120-0494dba8227a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1430.226512] env[65726]: DEBUG oslo_vmware.api [None req-c04eeb3b-5e5d-4ea6-a41e-462912417015 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]528eb1d9-1b23-1a67-f120-0494dba8227a, 'name': SearchDatastore_Task, 'duration_secs': 0.010238} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1430.226846] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c04eeb3b-5e5d-4ea6-a41e-462912417015 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1430.227080] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c04eeb3b-5e5d-4ea6-a41e-462912417015 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1430.618270] env[65726]: DEBUG oslo_concurrency.lockutils [None req-ece625a1-bea2-45d1-9407-04434761bc49 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Acquiring lock "a0119685-4f00-49f2-b97b-0e29175050cf" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1430.618475] env[65726]: DEBUG oslo_concurrency.lockutils [None req-ece625a1-bea2-45d1-9407-04434761bc49 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Lock "a0119685-4f00-49f2-b97b-0e29175050cf" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1430.808948] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65afc78e-ed57-4108-ba0e-81c299c54e16 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1430.818828] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a5fac64-7dfd-4455-9be9-76476a82babc {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1430.849943] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95cd7d89-0fa3-4736-965f-31d238fab70f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1430.857529] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-edd7390c-c2b4-4630-85f2-353c88872891 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1430.870644] env[65726]: DEBUG nova.compute.provider_tree [None req-c04eeb3b-5e5d-4ea6-a41e-462912417015 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1431.120884] env[65726]: DEBUG nova.compute.manager [None req-ece625a1-bea2-45d1-9407-04434761bc49 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: a0119685-4f00-49f2-b97b-0e29175050cf] Starting instance... {{(pid=65726) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 1431.374246] env[65726]: DEBUG nova.scheduler.client.report [None req-c04eeb3b-5e5d-4ea6-a41e-462912417015 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1431.643447] env[65726]: DEBUG oslo_concurrency.lockutils [None req-ece625a1-bea2-45d1-9407-04434761bc49 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1432.385296] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c04eeb3b-5e5d-4ea6-a41e-462912417015 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.158s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1432.388339] env[65726]: DEBUG oslo_concurrency.lockutils [None req-ece625a1-bea2-45d1-9407-04434761bc49 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.745s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1432.389842] env[65726]: INFO nova.compute.claims [None req-ece625a1-bea2-45d1-9407-04434761bc49 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: a0119685-4f00-49f2-b97b-0e29175050cf] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1432.952708] env[65726]: INFO nova.scheduler.client.report [None req-c04eeb3b-5e5d-4ea6-a41e-462912417015 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Deleted allocation for migration ac4cf9cd-eee7-48ff-8287-03052d30015d [ 1433.459011] env[65726]: DEBUG oslo_concurrency.lockutils [None req-c04eeb3b-5e5d-4ea6-a41e-462912417015 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Lock "e3255f3b-028f-4a0b-b621-873730417936" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 6.615s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1433.483740] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3059f6c2-51c4-4d5f-a2e6-484727c13cdc {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1433.493570] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e5d5ff9-97fa-488d-9683-482b5d3ee29d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1433.528575] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f02b749-b4fa-42ef-9a16-0beb4223b65f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1433.537904] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95a9cc62-5048-4659-8a86-689eac79e8e9 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1433.552525] env[65726]: DEBUG nova.compute.provider_tree [None req-ece625a1-bea2-45d1-9407-04434761bc49 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1433.903422] env[65726]: DEBUG oslo_concurrency.lockutils [None req-d1674e40-1087-4c64-897a-a89b6dcd052c tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Acquiring lock "e3255f3b-028f-4a0b-b621-873730417936" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1433.903775] env[65726]: DEBUG oslo_concurrency.lockutils [None req-d1674e40-1087-4c64-897a-a89b6dcd052c tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Lock "e3255f3b-028f-4a0b-b621-873730417936" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1433.904073] env[65726]: DEBUG oslo_concurrency.lockutils [None req-d1674e40-1087-4c64-897a-a89b6dcd052c tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Acquiring lock "e3255f3b-028f-4a0b-b621-873730417936-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1433.904347] env[65726]: DEBUG oslo_concurrency.lockutils [None req-d1674e40-1087-4c64-897a-a89b6dcd052c tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Lock "e3255f3b-028f-4a0b-b621-873730417936-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1433.904583] env[65726]: DEBUG oslo_concurrency.lockutils [None req-d1674e40-1087-4c64-897a-a89b6dcd052c tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Lock "e3255f3b-028f-4a0b-b621-873730417936-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1433.907393] env[65726]: INFO nova.compute.manager [None req-d1674e40-1087-4c64-897a-a89b6dcd052c tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: e3255f3b-028f-4a0b-b621-873730417936] Terminating instance [ 1434.055487] env[65726]: DEBUG nova.scheduler.client.report [None req-ece625a1-bea2-45d1-9407-04434761bc49 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1434.411874] env[65726]: DEBUG nova.compute.manager [None req-d1674e40-1087-4c64-897a-a89b6dcd052c tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: e3255f3b-028f-4a0b-b621-873730417936] Start destroying the instance on the hypervisor. {{(pid=65726) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1434.412111] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-d1674e40-1087-4c64-897a-a89b6dcd052c tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: e3255f3b-028f-4a0b-b621-873730417936] Destroying instance {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1434.414694] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26a3a801-84e2-4c32-8b7e-14e24593a819 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1434.423142] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-d1674e40-1087-4c64-897a-a89b6dcd052c tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: e3255f3b-028f-4a0b-b621-873730417936] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1434.423410] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a28b8388-9e74-49a4-9577-579ef04734df {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1434.430418] env[65726]: DEBUG oslo_vmware.api [None req-d1674e40-1087-4c64-897a-a89b6dcd052c tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Waiting for the task: (returnval){ [ 1434.430418] env[65726]: value = "task-5117136" [ 1434.430418] env[65726]: _type = "Task" [ 1434.430418] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1434.439436] env[65726]: DEBUG oslo_vmware.api [None req-d1674e40-1087-4c64-897a-a89b6dcd052c tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': task-5117136, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1434.561057] env[65726]: DEBUG oslo_concurrency.lockutils [None req-ece625a1-bea2-45d1-9407-04434761bc49 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.173s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1434.561646] env[65726]: DEBUG nova.compute.manager [None req-ece625a1-bea2-45d1-9407-04434761bc49 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: a0119685-4f00-49f2-b97b-0e29175050cf] Start building networks asynchronously for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1434.940933] env[65726]: DEBUG oslo_vmware.api [None req-d1674e40-1087-4c64-897a-a89b6dcd052c tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': task-5117136, 'name': PowerOffVM_Task, 'duration_secs': 0.25798} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1434.941213] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-d1674e40-1087-4c64-897a-a89b6dcd052c tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: e3255f3b-028f-4a0b-b621-873730417936] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1434.941379] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-d1674e40-1087-4c64-897a-a89b6dcd052c tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: e3255f3b-028f-4a0b-b621-873730417936] Unregistering the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1434.941667] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e220562a-8d46-4e66-9d9d-2bcb080b2c4b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1435.005896] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-d1674e40-1087-4c64-897a-a89b6dcd052c tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: e3255f3b-028f-4a0b-b621-873730417936] Unregistered the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1435.006309] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-d1674e40-1087-4c64-897a-a89b6dcd052c tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: e3255f3b-028f-4a0b-b621-873730417936] Deleting contents of the VM from datastore datastore2 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1435.006646] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-d1674e40-1087-4c64-897a-a89b6dcd052c tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Deleting the datastore file [datastore2] e3255f3b-028f-4a0b-b621-873730417936 {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1435.006878] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9eefcf06-31a8-480c-9447-cb161fb9e328 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1435.014930] env[65726]: DEBUG oslo_vmware.api [None req-d1674e40-1087-4c64-897a-a89b6dcd052c tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Waiting for the task: (returnval){ [ 1435.014930] env[65726]: value = "task-5117138" [ 1435.014930] env[65726]: _type = "Task" [ 1435.014930] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1435.024384] env[65726]: DEBUG oslo_vmware.api [None req-d1674e40-1087-4c64-897a-a89b6dcd052c tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': task-5117138, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1435.068212] env[65726]: DEBUG nova.compute.utils [None req-ece625a1-bea2-45d1-9407-04434761bc49 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Using /dev/sd instead of None {{(pid=65726) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1435.069635] env[65726]: DEBUG nova.compute.manager [None req-ece625a1-bea2-45d1-9407-04434761bc49 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: a0119685-4f00-49f2-b97b-0e29175050cf] Allocating IP information in the background. {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1435.069825] env[65726]: DEBUG nova.network.neutron [None req-ece625a1-bea2-45d1-9407-04434761bc49 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: a0119685-4f00-49f2-b97b-0e29175050cf] allocate_for_instance() {{(pid=65726) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 1435.070167] env[65726]: WARNING neutronclient.v2_0.client [None req-ece625a1-bea2-45d1-9407-04434761bc49 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1435.070471] env[65726]: WARNING neutronclient.v2_0.client [None req-ece625a1-bea2-45d1-9407-04434761bc49 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1435.071038] env[65726]: WARNING openstack [None req-ece625a1-bea2-45d1-9407-04434761bc49 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1435.071398] env[65726]: WARNING openstack [None req-ece625a1-bea2-45d1-9407-04434761bc49 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1435.126747] env[65726]: DEBUG nova.policy [None req-ece625a1-bea2-45d1-9407-04434761bc49 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6ab5b309207a46bb9d95998ef0a7a46f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6b56f93ba0e14521921484ac2c785c43', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65726) authorize /opt/stack/nova/nova/policy.py:192}} [ 1435.401079] env[65726]: DEBUG nova.network.neutron [None req-ece625a1-bea2-45d1-9407-04434761bc49 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: a0119685-4f00-49f2-b97b-0e29175050cf] Successfully created port: 5b57e69b-e348-46cb-932b-e1939c568791 {{(pid=65726) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1435.526999] env[65726]: DEBUG oslo_vmware.api [None req-d1674e40-1087-4c64-897a-a89b6dcd052c tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': task-5117138, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.14289} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1435.527454] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-d1674e40-1087-4c64-897a-a89b6dcd052c tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Deleted the datastore file {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1435.527795] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-d1674e40-1087-4c64-897a-a89b6dcd052c tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: e3255f3b-028f-4a0b-b621-873730417936] Deleted contents of the VM from datastore datastore2 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1435.528083] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-d1674e40-1087-4c64-897a-a89b6dcd052c tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: e3255f3b-028f-4a0b-b621-873730417936] Instance destroyed {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1435.528295] env[65726]: INFO nova.compute.manager [None req-d1674e40-1087-4c64-897a-a89b6dcd052c tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: e3255f3b-028f-4a0b-b621-873730417936] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1435.528610] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-d1674e40-1087-4c64-897a-a89b6dcd052c tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1435.528837] env[65726]: DEBUG nova.compute.manager [-] [instance: e3255f3b-028f-4a0b-b621-873730417936] Deallocating network for instance {{(pid=65726) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1435.528947] env[65726]: DEBUG nova.network.neutron [-] [instance: e3255f3b-028f-4a0b-b621-873730417936] deallocate_for_instance() {{(pid=65726) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 1435.529499] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1435.529811] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1435.568485] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1435.568781] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1435.578990] env[65726]: DEBUG nova.compute.manager [None req-ece625a1-bea2-45d1-9407-04434761bc49 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: a0119685-4f00-49f2-b97b-0e29175050cf] Start building block device mappings for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1435.985381] env[65726]: DEBUG nova.compute.manager [req-69d53cf2-3ed3-42db-af0d-133ada948f14 req-14ff8c1c-44cc-4a79-809a-42790de72dd9 service nova] [instance: e3255f3b-028f-4a0b-b621-873730417936] Received event network-vif-deleted-03c9e6fb-2435-43be-bb55-8afecdcf38ff {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1435.985572] env[65726]: INFO nova.compute.manager [req-69d53cf2-3ed3-42db-af0d-133ada948f14 req-14ff8c1c-44cc-4a79-809a-42790de72dd9 service nova] [instance: e3255f3b-028f-4a0b-b621-873730417936] Neutron deleted interface 03c9e6fb-2435-43be-bb55-8afecdcf38ff; detaching it from the instance and deleting it from the info cache [ 1435.985733] env[65726]: DEBUG nova.network.neutron [req-69d53cf2-3ed3-42db-af0d-133ada948f14 req-14ff8c1c-44cc-4a79-809a-42790de72dd9 service nova] [instance: e3255f3b-028f-4a0b-b621-873730417936] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1436.083788] env[65726]: INFO nova.virt.block_device [None req-ece625a1-bea2-45d1-9407-04434761bc49 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: a0119685-4f00-49f2-b97b-0e29175050cf] Booting with volume 790d262c-e3cc-459b-8c42-30915c88938a at /dev/sda [ 1436.124018] env[65726]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c7a9c1b6-ec6f-4279-ab00-b5f6a46d5f90 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1436.134443] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-290d38b4-66cb-4e27-8bfe-67106102fae1 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1436.168055] env[65726]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b506eaa0-9ae5-4081-a66a-a5b922f34c13 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1436.177527] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b20bcc4-4ff4-40fc-8fa1-307f9f2ff111 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1436.208766] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-411cccbe-b3a6-4f8a-84d8-64d04d6919d3 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1436.216605] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce249c2b-04df-43bf-a691-6dc16766bf8e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1436.230796] env[65726]: DEBUG nova.virt.block_device [None req-ece625a1-bea2-45d1-9407-04434761bc49 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: a0119685-4f00-49f2-b97b-0e29175050cf] Updating existing volume attachment record: 5c28c8cb-076e-4764-9ad3-1fbb4abb0323 {{(pid=65726) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1436.459666] env[65726]: DEBUG nova.network.neutron [-] [instance: e3255f3b-028f-4a0b-b621-873730417936] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1436.489369] env[65726]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-026428a7-00aa-4e5e-b094-e1263412fb7b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1436.500336] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6cfed1ed-0661-490e-a250-4e505bf7cb1a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1436.531959] env[65726]: DEBUG nova.compute.manager [req-69d53cf2-3ed3-42db-af0d-133ada948f14 req-14ff8c1c-44cc-4a79-809a-42790de72dd9 service nova] [instance: e3255f3b-028f-4a0b-b621-873730417936] Detach interface failed, port_id=03c9e6fb-2435-43be-bb55-8afecdcf38ff, reason: Instance e3255f3b-028f-4a0b-b621-873730417936 could not be found. {{(pid=65726) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11669}} [ 1436.826228] env[65726]: DEBUG nova.compute.manager [req-4c483acd-7746-422b-b4c5-222d9f877a33 req-fff806da-504a-45e9-b5aa-b3f299dbad6a service nova] [instance: a0119685-4f00-49f2-b97b-0e29175050cf] Received event network-vif-plugged-5b57e69b-e348-46cb-932b-e1939c568791 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1436.826499] env[65726]: DEBUG oslo_concurrency.lockutils [req-4c483acd-7746-422b-b4c5-222d9f877a33 req-fff806da-504a-45e9-b5aa-b3f299dbad6a service nova] Acquiring lock "a0119685-4f00-49f2-b97b-0e29175050cf-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1436.826655] env[65726]: DEBUG oslo_concurrency.lockutils [req-4c483acd-7746-422b-b4c5-222d9f877a33 req-fff806da-504a-45e9-b5aa-b3f299dbad6a service nova] Lock "a0119685-4f00-49f2-b97b-0e29175050cf-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1436.826831] env[65726]: DEBUG oslo_concurrency.lockutils [req-4c483acd-7746-422b-b4c5-222d9f877a33 req-fff806da-504a-45e9-b5aa-b3f299dbad6a service nova] Lock "a0119685-4f00-49f2-b97b-0e29175050cf-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1436.826993] env[65726]: DEBUG nova.compute.manager [req-4c483acd-7746-422b-b4c5-222d9f877a33 req-fff806da-504a-45e9-b5aa-b3f299dbad6a service nova] [instance: a0119685-4f00-49f2-b97b-0e29175050cf] No waiting events found dispatching network-vif-plugged-5b57e69b-e348-46cb-932b-e1939c568791 {{(pid=65726) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1436.827221] env[65726]: WARNING nova.compute.manager [req-4c483acd-7746-422b-b4c5-222d9f877a33 req-fff806da-504a-45e9-b5aa-b3f299dbad6a service nova] [instance: a0119685-4f00-49f2-b97b-0e29175050cf] Received unexpected event network-vif-plugged-5b57e69b-e348-46cb-932b-e1939c568791 for instance with vm_state building and task_state block_device_mapping. [ 1436.918303] env[65726]: DEBUG nova.network.neutron [None req-ece625a1-bea2-45d1-9407-04434761bc49 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: a0119685-4f00-49f2-b97b-0e29175050cf] Successfully updated port: 5b57e69b-e348-46cb-932b-e1939c568791 {{(pid=65726) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 1436.962166] env[65726]: INFO nova.compute.manager [-] [instance: e3255f3b-028f-4a0b-b621-873730417936] Took 1.43 seconds to deallocate network for instance. [ 1437.420938] env[65726]: DEBUG oslo_concurrency.lockutils [None req-ece625a1-bea2-45d1-9407-04434761bc49 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Acquiring lock "refresh_cache-a0119685-4f00-49f2-b97b-0e29175050cf" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1437.421218] env[65726]: DEBUG oslo_concurrency.lockutils [None req-ece625a1-bea2-45d1-9407-04434761bc49 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Acquired lock "refresh_cache-a0119685-4f00-49f2-b97b-0e29175050cf" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1437.421418] env[65726]: DEBUG nova.network.neutron [None req-ece625a1-bea2-45d1-9407-04434761bc49 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: a0119685-4f00-49f2-b97b-0e29175050cf] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1437.469199] env[65726]: DEBUG oslo_concurrency.lockutils [None req-d1674e40-1087-4c64-897a-a89b6dcd052c tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1437.469496] env[65726]: DEBUG oslo_concurrency.lockutils [None req-d1674e40-1087-4c64-897a-a89b6dcd052c tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1437.469678] env[65726]: DEBUG oslo_concurrency.lockutils [None req-d1674e40-1087-4c64-897a-a89b6dcd052c tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1437.492135] env[65726]: INFO nova.scheduler.client.report [None req-d1674e40-1087-4c64-897a-a89b6dcd052c tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Deleted allocations for instance e3255f3b-028f-4a0b-b621-873730417936 [ 1437.924783] env[65726]: WARNING openstack [None req-ece625a1-bea2-45d1-9407-04434761bc49 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1437.925138] env[65726]: WARNING openstack [None req-ece625a1-bea2-45d1-9407-04434761bc49 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1437.930171] env[65726]: DEBUG nova.network.neutron [None req-ece625a1-bea2-45d1-9407-04434761bc49 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: a0119685-4f00-49f2-b97b-0e29175050cf] Instance cache missing network info. {{(pid=65726) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 1437.980941] env[65726]: WARNING openstack [None req-ece625a1-bea2-45d1-9407-04434761bc49 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1437.981381] env[65726]: WARNING openstack [None req-ece625a1-bea2-45d1-9407-04434761bc49 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1437.999375] env[65726]: DEBUG oslo_concurrency.lockutils [None req-d1674e40-1087-4c64-897a-a89b6dcd052c tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Lock "e3255f3b-028f-4a0b-b621-873730417936" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 4.096s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1438.051897] env[65726]: WARNING openstack [None req-ece625a1-bea2-45d1-9407-04434761bc49 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1438.052370] env[65726]: WARNING openstack [None req-ece625a1-bea2-45d1-9407-04434761bc49 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1438.141579] env[65726]: DEBUG nova.network.neutron [None req-ece625a1-bea2-45d1-9407-04434761bc49 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: a0119685-4f00-49f2-b97b-0e29175050cf] Updating instance_info_cache with network_info: [{"id": "5b57e69b-e348-46cb-932b-e1939c568791", "address": "fa:16:3e:b5:4a:f0", "network": {"id": "336fe51f-971a-447d-8e68-505640e0db1b", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1030314703-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6b56f93ba0e14521921484ac2c785c43", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8ee1c76d-1a61-4546-85cb-d4bd3c1b35ef", "external-id": "nsx-vlan-transportzone-161", "segmentation_id": 161, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5b57e69b-e3", "ovs_interfaceid": "5b57e69b-e348-46cb-932b-e1939c568791", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1438.321623] env[65726]: DEBUG nova.compute.manager [None req-ece625a1-bea2-45d1-9407-04434761bc49 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: a0119685-4f00-49f2-b97b-0e29175050cf] Start spawning the instance on the hypervisor. {{(pid=65726) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1438.322298] env[65726]: DEBUG nova.virt.hardware [None req-ece625a1-bea2-45d1-9407-04434761bc49 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1438.322521] env[65726]: DEBUG nova.virt.hardware [None req-ece625a1-bea2-45d1-9407-04434761bc49 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1438.322673] env[65726]: DEBUG nova.virt.hardware [None req-ece625a1-bea2-45d1-9407-04434761bc49 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1438.322847] env[65726]: DEBUG nova.virt.hardware [None req-ece625a1-bea2-45d1-9407-04434761bc49 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1438.322983] env[65726]: DEBUG nova.virt.hardware [None req-ece625a1-bea2-45d1-9407-04434761bc49 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1438.323149] env[65726]: DEBUG nova.virt.hardware [None req-ece625a1-bea2-45d1-9407-04434761bc49 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1438.323367] env[65726]: DEBUG nova.virt.hardware [None req-ece625a1-bea2-45d1-9407-04434761bc49 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1438.323529] env[65726]: DEBUG nova.virt.hardware [None req-ece625a1-bea2-45d1-9407-04434761bc49 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1438.323696] env[65726]: DEBUG nova.virt.hardware [None req-ece625a1-bea2-45d1-9407-04434761bc49 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1438.323856] env[65726]: DEBUG nova.virt.hardware [None req-ece625a1-bea2-45d1-9407-04434761bc49 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1438.324071] env[65726]: DEBUG nova.virt.hardware [None req-ece625a1-bea2-45d1-9407-04434761bc49 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1438.325054] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fefa6155-dfc6-4f3d-b4b2-9693d7e56aa1 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1438.333610] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-293111e6-68d5-44b4-bed1-2cb54d992eaf {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1438.644539] env[65726]: DEBUG oslo_concurrency.lockutils [None req-ece625a1-bea2-45d1-9407-04434761bc49 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Releasing lock "refresh_cache-a0119685-4f00-49f2-b97b-0e29175050cf" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1438.644651] env[65726]: DEBUG nova.compute.manager [None req-ece625a1-bea2-45d1-9407-04434761bc49 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: a0119685-4f00-49f2-b97b-0e29175050cf] Instance network_info: |[{"id": "5b57e69b-e348-46cb-932b-e1939c568791", "address": "fa:16:3e:b5:4a:f0", "network": {"id": "336fe51f-971a-447d-8e68-505640e0db1b", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1030314703-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6b56f93ba0e14521921484ac2c785c43", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8ee1c76d-1a61-4546-85cb-d4bd3c1b35ef", "external-id": "nsx-vlan-transportzone-161", "segmentation_id": 161, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5b57e69b-e3", "ovs_interfaceid": "5b57e69b-e348-46cb-932b-e1939c568791", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1438.645165] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-ece625a1-bea2-45d1-9407-04434761bc49 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: a0119685-4f00-49f2-b97b-0e29175050cf] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b5:4a:f0', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '8ee1c76d-1a61-4546-85cb-d4bd3c1b35ef', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '5b57e69b-e348-46cb-932b-e1939c568791', 'vif_model': 'vmxnet3'}] {{(pid=65726) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1438.653167] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-ece625a1-bea2-45d1-9407-04434761bc49 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1438.653437] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a0119685-4f00-49f2-b97b-0e29175050cf] Creating VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1438.653680] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-23cf60a3-5394-4703-9748-eff785dcf2c1 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1438.673517] env[65726]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1438.673517] env[65726]: value = "task-5117139" [ 1438.673517] env[65726]: _type = "Task" [ 1438.673517] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1438.681665] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5117139, 'name': CreateVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1438.860687] env[65726]: DEBUG nova.compute.manager [req-6d2c4d86-6e00-4946-809d-e88f24bca66d req-bd2f8342-5152-432a-bf24-575d2c5e1c8e service nova] [instance: a0119685-4f00-49f2-b97b-0e29175050cf] Received event network-changed-5b57e69b-e348-46cb-932b-e1939c568791 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1438.860757] env[65726]: DEBUG nova.compute.manager [req-6d2c4d86-6e00-4946-809d-e88f24bca66d req-bd2f8342-5152-432a-bf24-575d2c5e1c8e service nova] [instance: a0119685-4f00-49f2-b97b-0e29175050cf] Refreshing instance network info cache due to event network-changed-5b57e69b-e348-46cb-932b-e1939c568791. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 1438.861137] env[65726]: DEBUG oslo_concurrency.lockutils [req-6d2c4d86-6e00-4946-809d-e88f24bca66d req-bd2f8342-5152-432a-bf24-575d2c5e1c8e service nova] Acquiring lock "refresh_cache-a0119685-4f00-49f2-b97b-0e29175050cf" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1438.861305] env[65726]: DEBUG oslo_concurrency.lockutils [req-6d2c4d86-6e00-4946-809d-e88f24bca66d req-bd2f8342-5152-432a-bf24-575d2c5e1c8e service nova] Acquired lock "refresh_cache-a0119685-4f00-49f2-b97b-0e29175050cf" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1438.861473] env[65726]: DEBUG nova.network.neutron [req-6d2c4d86-6e00-4946-809d-e88f24bca66d req-bd2f8342-5152-432a-bf24-575d2c5e1c8e service nova] [instance: a0119685-4f00-49f2-b97b-0e29175050cf] Refreshing network info cache for port 5b57e69b-e348-46cb-932b-e1939c568791 {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 1439.184379] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5117139, 'name': CreateVM_Task, 'duration_secs': 0.351785} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1439.184703] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a0119685-4f00-49f2-b97b-0e29175050cf] Created VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1439.185374] env[65726]: WARNING openstack [None req-ece625a1-bea2-45d1-9407-04434761bc49 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1439.185721] env[65726]: WARNING openstack [None req-ece625a1-bea2-45d1-9407-04434761bc49 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1439.190797] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-ece625a1-bea2-45d1-9407-04434761bc49 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: a0119685-4f00-49f2-b97b-0e29175050cf] Block device information present: {'root_device_name': '/dev/sda', 'image': [], 'ephemerals': [], 'block_device_mapping': [{'disk_bus': None, 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-995352', 'volume_id': '790d262c-e3cc-459b-8c42-30915c88938a', 'name': 'volume-790d262c-e3cc-459b-8c42-30915c88938a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'a0119685-4f00-49f2-b97b-0e29175050cf', 'attached_at': '', 'detached_at': '', 'volume_id': '790d262c-e3cc-459b-8c42-30915c88938a', 'serial': '790d262c-e3cc-459b-8c42-30915c88938a'}, 'delete_on_termination': True, 'device_type': None, 'boot_index': 0, 'mount_device': '/dev/sda', 'attachment_id': '5c28c8cb-076e-4764-9ad3-1fbb4abb0323', 'guest_format': None, 'volume_type': None}], 'swap': None} {{(pid=65726) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 1439.191009] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-ece625a1-bea2-45d1-9407-04434761bc49 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: a0119685-4f00-49f2-b97b-0e29175050cf] Root volume attach. Driver type: vmdk {{(pid=65726) attach_root_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:661}} [ 1439.193024] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-801fd4b7-c641-45cc-8484-623d9f34456e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1439.198631] env[65726]: DEBUG oslo_concurrency.lockutils [None req-24f45c06-45a5-4710-b612-50dc567da0ea tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Acquiring lock "08377673-6f69-4a5c-8135-7ef1683bbb55" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1439.198871] env[65726]: DEBUG oslo_concurrency.lockutils [None req-24f45c06-45a5-4710-b612-50dc567da0ea tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Lock "08377673-6f69-4a5c-8135-7ef1683bbb55" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1439.205547] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95a61334-ddad-4296-8757-5a13690f251f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1439.214609] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8eb8c999-704e-47da-919f-0e8b97360cb7 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1439.221702] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.RelocateVM_Task with opID=oslo.vmware-ac4ed619-9d70-4b48-96ad-ebf8c37abdac {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1439.228673] env[65726]: DEBUG oslo_vmware.api [None req-ece625a1-bea2-45d1-9407-04434761bc49 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Waiting for the task: (returnval){ [ 1439.228673] env[65726]: value = "task-5117140" [ 1439.228673] env[65726]: _type = "Task" [ 1439.228673] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1439.238411] env[65726]: DEBUG oslo_vmware.api [None req-ece625a1-bea2-45d1-9407-04434761bc49 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': task-5117140, 'name': RelocateVM_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1439.364476] env[65726]: WARNING openstack [req-6d2c4d86-6e00-4946-809d-e88f24bca66d req-bd2f8342-5152-432a-bf24-575d2c5e1c8e service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1439.364876] env[65726]: WARNING openstack [req-6d2c4d86-6e00-4946-809d-e88f24bca66d req-bd2f8342-5152-432a-bf24-575d2c5e1c8e service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1439.468564] env[65726]: WARNING openstack [req-6d2c4d86-6e00-4946-809d-e88f24bca66d req-bd2f8342-5152-432a-bf24-575d2c5e1c8e service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1439.468955] env[65726]: WARNING openstack [req-6d2c4d86-6e00-4946-809d-e88f24bca66d req-bd2f8342-5152-432a-bf24-575d2c5e1c8e service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1439.558643] env[65726]: WARNING openstack [req-6d2c4d86-6e00-4946-809d-e88f24bca66d req-bd2f8342-5152-432a-bf24-575d2c5e1c8e service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1439.559032] env[65726]: WARNING openstack [req-6d2c4d86-6e00-4946-809d-e88f24bca66d req-bd2f8342-5152-432a-bf24-575d2c5e1c8e service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1439.645367] env[65726]: DEBUG nova.network.neutron [req-6d2c4d86-6e00-4946-809d-e88f24bca66d req-bd2f8342-5152-432a-bf24-575d2c5e1c8e service nova] [instance: a0119685-4f00-49f2-b97b-0e29175050cf] Updated VIF entry in instance network info cache for port 5b57e69b-e348-46cb-932b-e1939c568791. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 1439.645727] env[65726]: DEBUG nova.network.neutron [req-6d2c4d86-6e00-4946-809d-e88f24bca66d req-bd2f8342-5152-432a-bf24-575d2c5e1c8e service nova] [instance: a0119685-4f00-49f2-b97b-0e29175050cf] Updating instance_info_cache with network_info: [{"id": "5b57e69b-e348-46cb-932b-e1939c568791", "address": "fa:16:3e:b5:4a:f0", "network": {"id": "336fe51f-971a-447d-8e68-505640e0db1b", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1030314703-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6b56f93ba0e14521921484ac2c785c43", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8ee1c76d-1a61-4546-85cb-d4bd3c1b35ef", "external-id": "nsx-vlan-transportzone-161", "segmentation_id": 161, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5b57e69b-e3", "ovs_interfaceid": "5b57e69b-e348-46cb-932b-e1939c568791", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1439.701782] env[65726]: DEBUG nova.compute.manager [None req-24f45c06-45a5-4710-b612-50dc567da0ea tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 08377673-6f69-4a5c-8135-7ef1683bbb55] Starting instance... {{(pid=65726) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 1439.739232] env[65726]: DEBUG oslo_vmware.api [None req-ece625a1-bea2-45d1-9407-04434761bc49 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': task-5117140, 'name': RelocateVM_Task, 'duration_secs': 0.027281} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1439.739560] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-ece625a1-bea2-45d1-9407-04434761bc49 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: a0119685-4f00-49f2-b97b-0e29175050cf] Volume attach. Driver type: vmdk {{(pid=65726) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1439.739782] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-ece625a1-bea2-45d1-9407-04434761bc49 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: a0119685-4f00-49f2-b97b-0e29175050cf] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-995352', 'volume_id': '790d262c-e3cc-459b-8c42-30915c88938a', 'name': 'volume-790d262c-e3cc-459b-8c42-30915c88938a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'a0119685-4f00-49f2-b97b-0e29175050cf', 'attached_at': '', 'detached_at': '', 'volume_id': '790d262c-e3cc-459b-8c42-30915c88938a', 'serial': '790d262c-e3cc-459b-8c42-30915c88938a'} {{(pid=65726) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1439.740537] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f59dcb9a-05ae-47d0-a15f-aca251d9f0eb {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1439.756969] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8e85ce4-6c92-48d1-bf37-f1009d7b4123 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1439.778592] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-ece625a1-bea2-45d1-9407-04434761bc49 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: a0119685-4f00-49f2-b97b-0e29175050cf] Reconfiguring VM instance instance-0000007c to attach disk [datastore2] volume-790d262c-e3cc-459b-8c42-30915c88938a/volume-790d262c-e3cc-459b-8c42-30915c88938a.vmdk or device None with type thin {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1439.778853] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4da89aac-7026-4730-baa9-8dbf38065412 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1439.799303] env[65726]: DEBUG oslo_vmware.api [None req-ece625a1-bea2-45d1-9407-04434761bc49 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Waiting for the task: (returnval){ [ 1439.799303] env[65726]: value = "task-5117141" [ 1439.799303] env[65726]: _type = "Task" [ 1439.799303] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1439.808046] env[65726]: DEBUG oslo_vmware.api [None req-ece625a1-bea2-45d1-9407-04434761bc49 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': task-5117141, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1440.149732] env[65726]: DEBUG oslo_concurrency.lockutils [req-6d2c4d86-6e00-4946-809d-e88f24bca66d req-bd2f8342-5152-432a-bf24-575d2c5e1c8e service nova] Releasing lock "refresh_cache-a0119685-4f00-49f2-b97b-0e29175050cf" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1440.228344] env[65726]: DEBUG oslo_concurrency.lockutils [None req-24f45c06-45a5-4710-b612-50dc567da0ea tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1440.228743] env[65726]: DEBUG oslo_concurrency.lockutils [None req-24f45c06-45a5-4710-b612-50dc567da0ea tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1440.230345] env[65726]: INFO nova.compute.claims [None req-24f45c06-45a5-4710-b612-50dc567da0ea tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 08377673-6f69-4a5c-8135-7ef1683bbb55] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1440.309563] env[65726]: DEBUG oslo_vmware.api [None req-ece625a1-bea2-45d1-9407-04434761bc49 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': task-5117141, 'name': ReconfigVM_Task, 'duration_secs': 0.262997} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1440.309842] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-ece625a1-bea2-45d1-9407-04434761bc49 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: a0119685-4f00-49f2-b97b-0e29175050cf] Reconfigured VM instance instance-0000007c to attach disk [datastore2] volume-790d262c-e3cc-459b-8c42-30915c88938a/volume-790d262c-e3cc-459b-8c42-30915c88938a.vmdk or device None with type thin {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1440.314434] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d3cb773e-6069-400c-9b93-f7c09c95bc1a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1440.330646] env[65726]: DEBUG oslo_vmware.api [None req-ece625a1-bea2-45d1-9407-04434761bc49 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Waiting for the task: (returnval){ [ 1440.330646] env[65726]: value = "task-5117142" [ 1440.330646] env[65726]: _type = "Task" [ 1440.330646] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1440.341935] env[65726]: DEBUG oslo_vmware.api [None req-ece625a1-bea2-45d1-9407-04434761bc49 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': task-5117142, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1440.569987] env[65726]: DEBUG oslo_service.periodic_task [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=65726) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1440.570215] env[65726]: DEBUG nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Cleaning up deleted instances with incomplete migration {{(pid=65726) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11970}} [ 1440.841204] env[65726]: DEBUG oslo_vmware.api [None req-ece625a1-bea2-45d1-9407-04434761bc49 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': task-5117142, 'name': ReconfigVM_Task, 'duration_secs': 0.124708} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1440.841515] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-ece625a1-bea2-45d1-9407-04434761bc49 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: a0119685-4f00-49f2-b97b-0e29175050cf] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-995352', 'volume_id': '790d262c-e3cc-459b-8c42-30915c88938a', 'name': 'volume-790d262c-e3cc-459b-8c42-30915c88938a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'a0119685-4f00-49f2-b97b-0e29175050cf', 'attached_at': '', 'detached_at': '', 'volume_id': '790d262c-e3cc-459b-8c42-30915c88938a', 'serial': '790d262c-e3cc-459b-8c42-30915c88938a'} {{(pid=65726) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1440.842072] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-dad05f63-ee17-435a-bb7e-f732960f57e6 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1440.849755] env[65726]: DEBUG oslo_vmware.api [None req-ece625a1-bea2-45d1-9407-04434761bc49 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Waiting for the task: (returnval){ [ 1440.849755] env[65726]: value = "task-5117143" [ 1440.849755] env[65726]: _type = "Task" [ 1440.849755] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1440.857858] env[65726]: DEBUG oslo_vmware.api [None req-ece625a1-bea2-45d1-9407-04434761bc49 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': task-5117143, 'name': Rename_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1441.307130] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a74b9a72-dde5-4094-9152-107d870c9f63 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1441.315245] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f85867b-2e91-4c7f-97c7-4dd9baf3e179 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1441.345109] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5854305b-6943-4878-a7bd-20cbf061ca40 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1441.356962] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f3fb67f-2723-48c2-be3c-a208cdff5193 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1441.364637] env[65726]: DEBUG oslo_vmware.api [None req-ece625a1-bea2-45d1-9407-04434761bc49 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': task-5117143, 'name': Rename_Task, 'duration_secs': 0.124072} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1441.365318] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-ece625a1-bea2-45d1-9407-04434761bc49 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: a0119685-4f00-49f2-b97b-0e29175050cf] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1441.365579] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3f4efb11-911a-4ef9-91a3-becdbf9e63b7 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1441.376140] env[65726]: DEBUG nova.compute.provider_tree [None req-24f45c06-45a5-4710-b612-50dc567da0ea tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1441.378614] env[65726]: DEBUG oslo_vmware.api [None req-ece625a1-bea2-45d1-9407-04434761bc49 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Waiting for the task: (returnval){ [ 1441.378614] env[65726]: value = "task-5117144" [ 1441.378614] env[65726]: _type = "Task" [ 1441.378614] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1441.388179] env[65726]: DEBUG oslo_vmware.api [None req-ece625a1-bea2-45d1-9407-04434761bc49 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': task-5117144, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1441.880378] env[65726]: DEBUG nova.scheduler.client.report [None req-24f45c06-45a5-4710-b612-50dc567da0ea tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1441.892799] env[65726]: DEBUG oslo_vmware.api [None req-ece625a1-bea2-45d1-9407-04434761bc49 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': task-5117144, 'name': PowerOnVM_Task, 'duration_secs': 0.435951} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1441.893253] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-ece625a1-bea2-45d1-9407-04434761bc49 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: a0119685-4f00-49f2-b97b-0e29175050cf] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1441.893253] env[65726]: INFO nova.compute.manager [None req-ece625a1-bea2-45d1-9407-04434761bc49 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: a0119685-4f00-49f2-b97b-0e29175050cf] Took 3.57 seconds to spawn the instance on the hypervisor. [ 1441.893411] env[65726]: DEBUG nova.compute.manager [None req-ece625a1-bea2-45d1-9407-04434761bc49 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: a0119685-4f00-49f2-b97b-0e29175050cf] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1441.894216] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab1cd3c7-b73f-40e6-9eaa-a056d88777f6 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1442.388278] env[65726]: DEBUG oslo_concurrency.lockutils [None req-24f45c06-45a5-4710-b612-50dc567da0ea tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.159s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1442.388752] env[65726]: DEBUG nova.compute.manager [None req-24f45c06-45a5-4710-b612-50dc567da0ea tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 08377673-6f69-4a5c-8135-7ef1683bbb55] Start building networks asynchronously for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1442.415854] env[65726]: INFO nova.compute.manager [None req-ece625a1-bea2-45d1-9407-04434761bc49 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: a0119685-4f00-49f2-b97b-0e29175050cf] Took 10.79 seconds to build instance. [ 1442.895040] env[65726]: DEBUG nova.compute.utils [None req-24f45c06-45a5-4710-b612-50dc567da0ea tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Using /dev/sd instead of None {{(pid=65726) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1442.896171] env[65726]: DEBUG nova.compute.manager [None req-24f45c06-45a5-4710-b612-50dc567da0ea tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 08377673-6f69-4a5c-8135-7ef1683bbb55] Allocating IP information in the background. {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1442.896391] env[65726]: DEBUG nova.network.neutron [None req-24f45c06-45a5-4710-b612-50dc567da0ea tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 08377673-6f69-4a5c-8135-7ef1683bbb55] allocate_for_instance() {{(pid=65726) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 1442.896705] env[65726]: WARNING neutronclient.v2_0.client [None req-24f45c06-45a5-4710-b612-50dc567da0ea tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1442.897012] env[65726]: WARNING neutronclient.v2_0.client [None req-24f45c06-45a5-4710-b612-50dc567da0ea tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1442.897615] env[65726]: WARNING openstack [None req-24f45c06-45a5-4710-b612-50dc567da0ea tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1442.898068] env[65726]: WARNING openstack [None req-24f45c06-45a5-4710-b612-50dc567da0ea tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1442.908858] env[65726]: DEBUG nova.compute.manager [req-bc3c213d-9974-4949-899d-a51ca1393cd8 req-af323de2-05d5-491e-8d8d-eb4b5a287675 service nova] [instance: 55ee97c3-b014-48eb-b41b-fc5aa16bc09e] Received event network-changed-65a8678e-a808-4eb1-9cb2-2f86b5fbd4b9 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1442.909035] env[65726]: DEBUG nova.compute.manager [req-bc3c213d-9974-4949-899d-a51ca1393cd8 req-af323de2-05d5-491e-8d8d-eb4b5a287675 service nova] [instance: 55ee97c3-b014-48eb-b41b-fc5aa16bc09e] Refreshing instance network info cache due to event network-changed-65a8678e-a808-4eb1-9cb2-2f86b5fbd4b9. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 1442.909248] env[65726]: DEBUG oslo_concurrency.lockutils [req-bc3c213d-9974-4949-899d-a51ca1393cd8 req-af323de2-05d5-491e-8d8d-eb4b5a287675 service nova] Acquiring lock "refresh_cache-55ee97c3-b014-48eb-b41b-fc5aa16bc09e" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1442.909391] env[65726]: DEBUG oslo_concurrency.lockutils [req-bc3c213d-9974-4949-899d-a51ca1393cd8 req-af323de2-05d5-491e-8d8d-eb4b5a287675 service nova] Acquired lock "refresh_cache-55ee97c3-b014-48eb-b41b-fc5aa16bc09e" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1442.909578] env[65726]: DEBUG nova.network.neutron [req-bc3c213d-9974-4949-899d-a51ca1393cd8 req-af323de2-05d5-491e-8d8d-eb4b5a287675 service nova] [instance: 55ee97c3-b014-48eb-b41b-fc5aa16bc09e] Refreshing network info cache for port 65a8678e-a808-4eb1-9cb2-2f86b5fbd4b9 {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 1442.917759] env[65726]: DEBUG oslo_concurrency.lockutils [None req-ece625a1-bea2-45d1-9407-04434761bc49 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Lock "a0119685-4f00-49f2-b97b-0e29175050cf" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 12.299s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1442.958598] env[65726]: DEBUG nova.policy [None req-24f45c06-45a5-4710-b612-50dc567da0ea tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5143cc185c7645b48304c8ebc5c854de', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'acd800e0438940bc80ae3c6b672db4ee', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65726) authorize /opt/stack/nova/nova/policy.py:192}} [ 1443.208017] env[65726]: DEBUG nova.network.neutron [None req-24f45c06-45a5-4710-b612-50dc567da0ea tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 08377673-6f69-4a5c-8135-7ef1683bbb55] Successfully created port: 5aea6240-edaa-4fb1-b696-877bc1ba685a {{(pid=65726) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1443.408227] env[65726]: DEBUG nova.compute.manager [None req-24f45c06-45a5-4710-b612-50dc567da0ea tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 08377673-6f69-4a5c-8135-7ef1683bbb55] Start building block device mappings for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1443.413020] env[65726]: WARNING openstack [req-bc3c213d-9974-4949-899d-a51ca1393cd8 req-af323de2-05d5-491e-8d8d-eb4b5a287675 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1443.413336] env[65726]: WARNING openstack [req-bc3c213d-9974-4949-899d-a51ca1393cd8 req-af323de2-05d5-491e-8d8d-eb4b5a287675 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1443.523875] env[65726]: WARNING openstack [req-bc3c213d-9974-4949-899d-a51ca1393cd8 req-af323de2-05d5-491e-8d8d-eb4b5a287675 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1443.523875] env[65726]: WARNING openstack [req-bc3c213d-9974-4949-899d-a51ca1393cd8 req-af323de2-05d5-491e-8d8d-eb4b5a287675 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1443.597267] env[65726]: WARNING openstack [req-bc3c213d-9974-4949-899d-a51ca1393cd8 req-af323de2-05d5-491e-8d8d-eb4b5a287675 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1443.597699] env[65726]: WARNING openstack [req-bc3c213d-9974-4949-899d-a51ca1393cd8 req-af323de2-05d5-491e-8d8d-eb4b5a287675 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1443.681633] env[65726]: DEBUG nova.network.neutron [req-bc3c213d-9974-4949-899d-a51ca1393cd8 req-af323de2-05d5-491e-8d8d-eb4b5a287675 service nova] [instance: 55ee97c3-b014-48eb-b41b-fc5aa16bc09e] Updated VIF entry in instance network info cache for port 65a8678e-a808-4eb1-9cb2-2f86b5fbd4b9. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 1443.681972] env[65726]: DEBUG nova.network.neutron [req-bc3c213d-9974-4949-899d-a51ca1393cd8 req-af323de2-05d5-491e-8d8d-eb4b5a287675 service nova] [instance: 55ee97c3-b014-48eb-b41b-fc5aa16bc09e] Updating instance_info_cache with network_info: [{"id": "65a8678e-a808-4eb1-9cb2-2f86b5fbd4b9", "address": "fa:16:3e:ba:7b:9c", "network": {"id": "336fe51f-971a-447d-8e68-505640e0db1b", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1030314703-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6b56f93ba0e14521921484ac2c785c43", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8ee1c76d-1a61-4546-85cb-d4bd3c1b35ef", "external-id": "nsx-vlan-transportzone-161", "segmentation_id": 161, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap65a8678e-a8", "ovs_interfaceid": "65a8678e-a808-4eb1-9cb2-2f86b5fbd4b9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1444.184607] env[65726]: DEBUG oslo_concurrency.lockutils [req-bc3c213d-9974-4949-899d-a51ca1393cd8 req-af323de2-05d5-491e-8d8d-eb4b5a287675 service nova] Releasing lock "refresh_cache-55ee97c3-b014-48eb-b41b-fc5aa16bc09e" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1444.406018] env[65726]: DEBUG nova.compute.manager [None req-a282bdd6-528c-4479-b8a1-8ae112917393 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: a0119685-4f00-49f2-b97b-0e29175050cf] Stashing vm_state: active {{(pid=65726) _prep_resize /opt/stack/nova/nova/compute/manager.py:6193}} [ 1444.418543] env[65726]: DEBUG nova.compute.manager [None req-24f45c06-45a5-4710-b612-50dc567da0ea tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 08377673-6f69-4a5c-8135-7ef1683bbb55] Start spawning the instance on the hypervisor. {{(pid=65726) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1444.447134] env[65726]: DEBUG nova.virt.hardware [None req-24f45c06-45a5-4710-b612-50dc567da0ea tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T19:28:34Z,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1539f06fa3534e90acbc3a60c4b8bd3f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T19:28:35Z,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1444.447503] env[65726]: DEBUG nova.virt.hardware [None req-24f45c06-45a5-4710-b612-50dc567da0ea tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1444.447741] env[65726]: DEBUG nova.virt.hardware [None req-24f45c06-45a5-4710-b612-50dc567da0ea tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1444.448018] env[65726]: DEBUG nova.virt.hardware [None req-24f45c06-45a5-4710-b612-50dc567da0ea tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1444.448203] env[65726]: DEBUG nova.virt.hardware [None req-24f45c06-45a5-4710-b612-50dc567da0ea tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1444.448354] env[65726]: DEBUG nova.virt.hardware [None req-24f45c06-45a5-4710-b612-50dc567da0ea tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1444.448598] env[65726]: DEBUG nova.virt.hardware [None req-24f45c06-45a5-4710-b612-50dc567da0ea tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1444.448752] env[65726]: DEBUG nova.virt.hardware [None req-24f45c06-45a5-4710-b612-50dc567da0ea tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1444.448913] env[65726]: DEBUG nova.virt.hardware [None req-24f45c06-45a5-4710-b612-50dc567da0ea tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1444.449109] env[65726]: DEBUG nova.virt.hardware [None req-24f45c06-45a5-4710-b612-50dc567da0ea tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1444.449299] env[65726]: DEBUG nova.virt.hardware [None req-24f45c06-45a5-4710-b612-50dc567da0ea tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1444.450234] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3a9cd54-a8e7-4b1c-831a-cb341181dd31 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1444.460256] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-545df639-c2a3-4d0c-b9ce-7c7d049979c4 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1444.764501] env[65726]: DEBUG nova.network.neutron [None req-24f45c06-45a5-4710-b612-50dc567da0ea tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 08377673-6f69-4a5c-8135-7ef1683bbb55] Successfully updated port: 5aea6240-edaa-4fb1-b696-877bc1ba685a {{(pid=65726) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 1444.926802] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a282bdd6-528c-4479-b8a1-8ae112917393 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1444.927109] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a282bdd6-528c-4479-b8a1-8ae112917393 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1444.939113] env[65726]: DEBUG nova.compute.manager [req-902e72cd-d1c4-4b48-a211-4dec05b46dcb req-231e4ab9-9c47-4973-a561-ed929b1322d7 service nova] [instance: a0119685-4f00-49f2-b97b-0e29175050cf] Received event network-changed-5b57e69b-e348-46cb-932b-e1939c568791 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1444.939113] env[65726]: DEBUG nova.compute.manager [req-902e72cd-d1c4-4b48-a211-4dec05b46dcb req-231e4ab9-9c47-4973-a561-ed929b1322d7 service nova] [instance: a0119685-4f00-49f2-b97b-0e29175050cf] Refreshing instance network info cache due to event network-changed-5b57e69b-e348-46cb-932b-e1939c568791. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 1444.939113] env[65726]: DEBUG oslo_concurrency.lockutils [req-902e72cd-d1c4-4b48-a211-4dec05b46dcb req-231e4ab9-9c47-4973-a561-ed929b1322d7 service nova] Acquiring lock "refresh_cache-a0119685-4f00-49f2-b97b-0e29175050cf" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1444.939113] env[65726]: DEBUG oslo_concurrency.lockutils [req-902e72cd-d1c4-4b48-a211-4dec05b46dcb req-231e4ab9-9c47-4973-a561-ed929b1322d7 service nova] Acquired lock "refresh_cache-a0119685-4f00-49f2-b97b-0e29175050cf" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1444.939113] env[65726]: DEBUG nova.network.neutron [req-902e72cd-d1c4-4b48-a211-4dec05b46dcb req-231e4ab9-9c47-4973-a561-ed929b1322d7 service nova] [instance: a0119685-4f00-49f2-b97b-0e29175050cf] Refreshing network info cache for port 5b57e69b-e348-46cb-932b-e1939c568791 {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 1445.267569] env[65726]: DEBUG oslo_concurrency.lockutils [None req-24f45c06-45a5-4710-b612-50dc567da0ea tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Acquiring lock "refresh_cache-08377673-6f69-4a5c-8135-7ef1683bbb55" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1445.267763] env[65726]: DEBUG oslo_concurrency.lockutils [None req-24f45c06-45a5-4710-b612-50dc567da0ea tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Acquired lock "refresh_cache-08377673-6f69-4a5c-8135-7ef1683bbb55" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1445.267947] env[65726]: DEBUG nova.network.neutron [None req-24f45c06-45a5-4710-b612-50dc567da0ea tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 08377673-6f69-4a5c-8135-7ef1683bbb55] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1445.433092] env[65726]: INFO nova.compute.claims [None req-a282bdd6-528c-4479-b8a1-8ae112917393 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: a0119685-4f00-49f2-b97b-0e29175050cf] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1445.442756] env[65726]: WARNING openstack [req-902e72cd-d1c4-4b48-a211-4dec05b46dcb req-231e4ab9-9c47-4973-a561-ed929b1322d7 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1445.443157] env[65726]: WARNING openstack [req-902e72cd-d1c4-4b48-a211-4dec05b46dcb req-231e4ab9-9c47-4973-a561-ed929b1322d7 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1445.557632] env[65726]: WARNING openstack [req-902e72cd-d1c4-4b48-a211-4dec05b46dcb req-231e4ab9-9c47-4973-a561-ed929b1322d7 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1445.558094] env[65726]: WARNING openstack [req-902e72cd-d1c4-4b48-a211-4dec05b46dcb req-231e4ab9-9c47-4973-a561-ed929b1322d7 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1445.621080] env[65726]: WARNING openstack [req-902e72cd-d1c4-4b48-a211-4dec05b46dcb req-231e4ab9-9c47-4973-a561-ed929b1322d7 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1445.621604] env[65726]: WARNING openstack [req-902e72cd-d1c4-4b48-a211-4dec05b46dcb req-231e4ab9-9c47-4973-a561-ed929b1322d7 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1445.703464] env[65726]: DEBUG nova.network.neutron [req-902e72cd-d1c4-4b48-a211-4dec05b46dcb req-231e4ab9-9c47-4973-a561-ed929b1322d7 service nova] [instance: a0119685-4f00-49f2-b97b-0e29175050cf] Updated VIF entry in instance network info cache for port 5b57e69b-e348-46cb-932b-e1939c568791. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 1445.703971] env[65726]: DEBUG nova.network.neutron [req-902e72cd-d1c4-4b48-a211-4dec05b46dcb req-231e4ab9-9c47-4973-a561-ed929b1322d7 service nova] [instance: a0119685-4f00-49f2-b97b-0e29175050cf] Updating instance_info_cache with network_info: [{"id": "5b57e69b-e348-46cb-932b-e1939c568791", "address": "fa:16:3e:b5:4a:f0", "network": {"id": "336fe51f-971a-447d-8e68-505640e0db1b", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1030314703-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.218", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6b56f93ba0e14521921484ac2c785c43", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8ee1c76d-1a61-4546-85cb-d4bd3c1b35ef", "external-id": "nsx-vlan-transportzone-161", "segmentation_id": 161, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5b57e69b-e3", "ovs_interfaceid": "5b57e69b-e348-46cb-932b-e1939c568791", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1445.771506] env[65726]: WARNING openstack [None req-24f45c06-45a5-4710-b612-50dc567da0ea tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1445.772017] env[65726]: WARNING openstack [None req-24f45c06-45a5-4710-b612-50dc567da0ea tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1445.777313] env[65726]: DEBUG nova.network.neutron [None req-24f45c06-45a5-4710-b612-50dc567da0ea tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 08377673-6f69-4a5c-8135-7ef1683bbb55] Instance cache missing network info. {{(pid=65726) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 1445.830229] env[65726]: WARNING openstack [None req-24f45c06-45a5-4710-b612-50dc567da0ea tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1445.830639] env[65726]: WARNING openstack [None req-24f45c06-45a5-4710-b612-50dc567da0ea tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1445.892961] env[65726]: WARNING openstack [None req-24f45c06-45a5-4710-b612-50dc567da0ea tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1445.893432] env[65726]: WARNING openstack [None req-24f45c06-45a5-4710-b612-50dc567da0ea tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1445.940989] env[65726]: INFO nova.compute.resource_tracker [None req-a282bdd6-528c-4479-b8a1-8ae112917393 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: a0119685-4f00-49f2-b97b-0e29175050cf] Updating resource usage from migration 530dfaf8-ce37-484f-80e7-196316df4994 [ 1445.979903] env[65726]: DEBUG nova.network.neutron [None req-24f45c06-45a5-4710-b612-50dc567da0ea tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 08377673-6f69-4a5c-8135-7ef1683bbb55] Updating instance_info_cache with network_info: [{"id": "5aea6240-edaa-4fb1-b696-877bc1ba685a", "address": "fa:16:3e:70:2b:bb", "network": {"id": "a899453d-5f43-4650-9642-b33f54a4d2eb", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-614861139-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "acd800e0438940bc80ae3c6b672db4ee", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "205fb402-8eaf-4b61-8f57-8f216024179a", "external-id": "nsx-vlan-transportzone-78", "segmentation_id": 78, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5aea6240-ed", "ovs_interfaceid": "5aea6240-edaa-4fb1-b696-877bc1ba685a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1446.021418] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b0767b1-bec8-494f-8aa2-60b104b6305a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1446.030807] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16f107f7-3696-432e-9e1c-3f63346a46d5 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1446.063361] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b52ddf24-261d-4d6b-ba99-8093a49c8d54 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1446.071468] env[65726]: DEBUG oslo_service.periodic_task [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=65726) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1446.071748] env[65726]: DEBUG oslo_service.periodic_task [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=65726) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1446.071888] env[65726]: DEBUG nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=65726) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11251}} [ 1446.073171] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-825c2911-c983-45e0-a1cc-ac7af212100f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1446.087742] env[65726]: DEBUG nova.compute.provider_tree [None req-a282bdd6-528c-4479-b8a1-8ae112917393 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1446.207586] env[65726]: DEBUG oslo_concurrency.lockutils [req-902e72cd-d1c4-4b48-a211-4dec05b46dcb req-231e4ab9-9c47-4973-a561-ed929b1322d7 service nova] Releasing lock "refresh_cache-a0119685-4f00-49f2-b97b-0e29175050cf" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1446.207853] env[65726]: DEBUG nova.compute.manager [req-902e72cd-d1c4-4b48-a211-4dec05b46dcb req-231e4ab9-9c47-4973-a561-ed929b1322d7 service nova] [instance: 08377673-6f69-4a5c-8135-7ef1683bbb55] Received event network-vif-plugged-5aea6240-edaa-4fb1-b696-877bc1ba685a {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1446.208039] env[65726]: DEBUG oslo_concurrency.lockutils [req-902e72cd-d1c4-4b48-a211-4dec05b46dcb req-231e4ab9-9c47-4973-a561-ed929b1322d7 service nova] Acquiring lock "08377673-6f69-4a5c-8135-7ef1683bbb55-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1446.208255] env[65726]: DEBUG oslo_concurrency.lockutils [req-902e72cd-d1c4-4b48-a211-4dec05b46dcb req-231e4ab9-9c47-4973-a561-ed929b1322d7 service nova] Lock "08377673-6f69-4a5c-8135-7ef1683bbb55-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1446.208417] env[65726]: DEBUG oslo_concurrency.lockutils [req-902e72cd-d1c4-4b48-a211-4dec05b46dcb req-231e4ab9-9c47-4973-a561-ed929b1322d7 service nova] Lock "08377673-6f69-4a5c-8135-7ef1683bbb55-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1446.208945] env[65726]: DEBUG nova.compute.manager [req-902e72cd-d1c4-4b48-a211-4dec05b46dcb req-231e4ab9-9c47-4973-a561-ed929b1322d7 service nova] [instance: 08377673-6f69-4a5c-8135-7ef1683bbb55] No waiting events found dispatching network-vif-plugged-5aea6240-edaa-4fb1-b696-877bc1ba685a {{(pid=65726) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1446.208945] env[65726]: WARNING nova.compute.manager [req-902e72cd-d1c4-4b48-a211-4dec05b46dcb req-231e4ab9-9c47-4973-a561-ed929b1322d7 service nova] [instance: 08377673-6f69-4a5c-8135-7ef1683bbb55] Received unexpected event network-vif-plugged-5aea6240-edaa-4fb1-b696-877bc1ba685a for instance with vm_state building and task_state spawning. [ 1446.208945] env[65726]: DEBUG nova.compute.manager [req-902e72cd-d1c4-4b48-a211-4dec05b46dcb req-231e4ab9-9c47-4973-a561-ed929b1322d7 service nova] [instance: 08377673-6f69-4a5c-8135-7ef1683bbb55] Received event network-changed-5aea6240-edaa-4fb1-b696-877bc1ba685a {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1446.209165] env[65726]: DEBUG nova.compute.manager [req-902e72cd-d1c4-4b48-a211-4dec05b46dcb req-231e4ab9-9c47-4973-a561-ed929b1322d7 service nova] [instance: 08377673-6f69-4a5c-8135-7ef1683bbb55] Refreshing instance network info cache due to event network-changed-5aea6240-edaa-4fb1-b696-877bc1ba685a. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 1446.209294] env[65726]: DEBUG oslo_concurrency.lockutils [req-902e72cd-d1c4-4b48-a211-4dec05b46dcb req-231e4ab9-9c47-4973-a561-ed929b1322d7 service nova] Acquiring lock "refresh_cache-08377673-6f69-4a5c-8135-7ef1683bbb55" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1446.485123] env[65726]: DEBUG oslo_concurrency.lockutils [None req-24f45c06-45a5-4710-b612-50dc567da0ea tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Releasing lock "refresh_cache-08377673-6f69-4a5c-8135-7ef1683bbb55" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1446.485523] env[65726]: DEBUG nova.compute.manager [None req-24f45c06-45a5-4710-b612-50dc567da0ea tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 08377673-6f69-4a5c-8135-7ef1683bbb55] Instance network_info: |[{"id": "5aea6240-edaa-4fb1-b696-877bc1ba685a", "address": "fa:16:3e:70:2b:bb", "network": {"id": "a899453d-5f43-4650-9642-b33f54a4d2eb", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-614861139-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "acd800e0438940bc80ae3c6b672db4ee", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "205fb402-8eaf-4b61-8f57-8f216024179a", "external-id": "nsx-vlan-transportzone-78", "segmentation_id": 78, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5aea6240-ed", "ovs_interfaceid": "5aea6240-edaa-4fb1-b696-877bc1ba685a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1446.485897] env[65726]: DEBUG oslo_concurrency.lockutils [req-902e72cd-d1c4-4b48-a211-4dec05b46dcb req-231e4ab9-9c47-4973-a561-ed929b1322d7 service nova] Acquired lock "refresh_cache-08377673-6f69-4a5c-8135-7ef1683bbb55" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1446.486084] env[65726]: DEBUG nova.network.neutron [req-902e72cd-d1c4-4b48-a211-4dec05b46dcb req-231e4ab9-9c47-4973-a561-ed929b1322d7 service nova] [instance: 08377673-6f69-4a5c-8135-7ef1683bbb55] Refreshing network info cache for port 5aea6240-edaa-4fb1-b696-877bc1ba685a {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 1446.487511] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-24f45c06-45a5-4710-b612-50dc567da0ea tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 08377673-6f69-4a5c-8135-7ef1683bbb55] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:70:2b:bb', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '205fb402-8eaf-4b61-8f57-8f216024179a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '5aea6240-edaa-4fb1-b696-877bc1ba685a', 'vif_model': 'vmxnet3'}] {{(pid=65726) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1446.495269] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-24f45c06-45a5-4710-b612-50dc567da0ea tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1446.497078] env[65726]: WARNING openstack [req-902e72cd-d1c4-4b48-a211-4dec05b46dcb req-231e4ab9-9c47-4973-a561-ed929b1322d7 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1446.497510] env[65726]: WARNING openstack [req-902e72cd-d1c4-4b48-a211-4dec05b46dcb req-231e4ab9-9c47-4973-a561-ed929b1322d7 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1446.504653] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 08377673-6f69-4a5c-8135-7ef1683bbb55] Creating VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1446.505643] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7bd1502f-a62f-4ef6-b33e-18a361902f83 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1446.531353] env[65726]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1446.531353] env[65726]: value = "task-5117145" [ 1446.531353] env[65726]: _type = "Task" [ 1446.531353] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1446.545129] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5117145, 'name': CreateVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1446.569396] env[65726]: DEBUG oslo_service.periodic_task [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=65726) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1446.569660] env[65726]: DEBUG oslo_service.periodic_task [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=65726) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1446.569896] env[65726]: DEBUG oslo_service.periodic_task [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=65726) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1446.570166] env[65726]: DEBUG nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Cleaning up deleted instances {{(pid=65726) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11932}} [ 1446.590741] env[65726]: DEBUG nova.scheduler.client.report [None req-a282bdd6-528c-4479-b8a1-8ae112917393 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1446.611815] env[65726]: WARNING openstack [req-902e72cd-d1c4-4b48-a211-4dec05b46dcb req-231e4ab9-9c47-4973-a561-ed929b1322d7 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1446.612259] env[65726]: WARNING openstack [req-902e72cd-d1c4-4b48-a211-4dec05b46dcb req-231e4ab9-9c47-4973-a561-ed929b1322d7 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1446.692179] env[65726]: WARNING openstack [req-902e72cd-d1c4-4b48-a211-4dec05b46dcb req-231e4ab9-9c47-4973-a561-ed929b1322d7 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1446.692560] env[65726]: WARNING openstack [req-902e72cd-d1c4-4b48-a211-4dec05b46dcb req-231e4ab9-9c47-4973-a561-ed929b1322d7 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1446.779613] env[65726]: DEBUG nova.network.neutron [req-902e72cd-d1c4-4b48-a211-4dec05b46dcb req-231e4ab9-9c47-4973-a561-ed929b1322d7 service nova] [instance: 08377673-6f69-4a5c-8135-7ef1683bbb55] Updated VIF entry in instance network info cache for port 5aea6240-edaa-4fb1-b696-877bc1ba685a. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 1446.779953] env[65726]: DEBUG nova.network.neutron [req-902e72cd-d1c4-4b48-a211-4dec05b46dcb req-231e4ab9-9c47-4973-a561-ed929b1322d7 service nova] [instance: 08377673-6f69-4a5c-8135-7ef1683bbb55] Updating instance_info_cache with network_info: [{"id": "5aea6240-edaa-4fb1-b696-877bc1ba685a", "address": "fa:16:3e:70:2b:bb", "network": {"id": "a899453d-5f43-4650-9642-b33f54a4d2eb", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-614861139-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "acd800e0438940bc80ae3c6b672db4ee", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "205fb402-8eaf-4b61-8f57-8f216024179a", "external-id": "nsx-vlan-transportzone-78", "segmentation_id": 78, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5aea6240-ed", "ovs_interfaceid": "5aea6240-edaa-4fb1-b696-877bc1ba685a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1447.050321] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5117145, 'name': CreateVM_Task, 'duration_secs': 0.350662} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1447.050514] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 08377673-6f69-4a5c-8135-7ef1683bbb55] Created VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1447.051332] env[65726]: WARNING openstack [None req-24f45c06-45a5-4710-b612-50dc567da0ea tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1447.051692] env[65726]: WARNING openstack [None req-24f45c06-45a5-4710-b612-50dc567da0ea tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1447.057377] env[65726]: DEBUG oslo_concurrency.lockutils [None req-24f45c06-45a5-4710-b612-50dc567da0ea tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1447.057585] env[65726]: DEBUG oslo_concurrency.lockutils [None req-24f45c06-45a5-4710-b612-50dc567da0ea tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1447.057936] env[65726]: DEBUG oslo_concurrency.lockutils [None req-24f45c06-45a5-4710-b612-50dc567da0ea tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1447.058614] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dfe7ac32-692f-45ee-aa3d-27ab4e25f406 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1447.064589] env[65726]: DEBUG oslo_vmware.api [None req-24f45c06-45a5-4710-b612-50dc567da0ea tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Waiting for the task: (returnval){ [ 1447.064589] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52b8cfc1-ceb9-eb60-d331-d778c74734e0" [ 1447.064589] env[65726]: _type = "Task" [ 1447.064589] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1447.079579] env[65726]: DEBUG nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] There are 22 instances to clean {{(pid=65726) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11941}} [ 1447.079834] env[65726]: DEBUG nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] [instance: 1ad4c4f9-1b02-420c-ac89-5e3390ddf773] Instance has had 0 of 5 cleanup attempts {{(pid=65726) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 1447.081535] env[65726]: DEBUG oslo_vmware.api [None req-24f45c06-45a5-4710-b612-50dc567da0ea tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52b8cfc1-ceb9-eb60-d331-d778c74734e0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1447.096411] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a282bdd6-528c-4479-b8a1-8ae112917393 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.169s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1447.096641] env[65726]: INFO nova.compute.manager [None req-a282bdd6-528c-4479-b8a1-8ae112917393 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: a0119685-4f00-49f2-b97b-0e29175050cf] Migrating [ 1447.283569] env[65726]: DEBUG oslo_concurrency.lockutils [req-902e72cd-d1c4-4b48-a211-4dec05b46dcb req-231e4ab9-9c47-4973-a561-ed929b1322d7 service nova] Releasing lock "refresh_cache-08377673-6f69-4a5c-8135-7ef1683bbb55" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1447.576593] env[65726]: DEBUG oslo_vmware.api [None req-24f45c06-45a5-4710-b612-50dc567da0ea tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52b8cfc1-ceb9-eb60-d331-d778c74734e0, 'name': SearchDatastore_Task, 'duration_secs': 0.013037} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1447.576593] env[65726]: DEBUG oslo_concurrency.lockutils [None req-24f45c06-45a5-4710-b612-50dc567da0ea tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1447.576845] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-24f45c06-45a5-4710-b612-50dc567da0ea tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 08377673-6f69-4a5c-8135-7ef1683bbb55] Processing image b52362a3-ee8a-4cbf-b06f-513b0cc8f95c {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1447.576947] env[65726]: DEBUG oslo_concurrency.lockutils [None req-24f45c06-45a5-4710-b612-50dc567da0ea tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1447.577040] env[65726]: DEBUG oslo_concurrency.lockutils [None req-24f45c06-45a5-4710-b612-50dc567da0ea tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1447.577233] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-24f45c06-45a5-4710-b612-50dc567da0ea tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1447.577499] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3d7d3024-489d-4765-91f6-f6086b6eb7db {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1447.583066] env[65726]: DEBUG nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] [instance: 435b0908-fc57-4cd1-b008-48da19734835] Instance has had 0 of 5 cleanup attempts {{(pid=65726) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 1447.587328] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-24f45c06-45a5-4710-b612-50dc567da0ea tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1447.587504] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-24f45c06-45a5-4710-b612-50dc567da0ea tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65726) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1447.588274] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-15938766-ed72-46c1-a6d9-7575b06d8e38 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1447.594512] env[65726]: DEBUG oslo_vmware.api [None req-24f45c06-45a5-4710-b612-50dc567da0ea tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Waiting for the task: (returnval){ [ 1447.594512] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5267510b-b26f-0140-e5c0-180d5fdb56a2" [ 1447.594512] env[65726]: _type = "Task" [ 1447.594512] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1447.603145] env[65726]: DEBUG oslo_vmware.api [None req-24f45c06-45a5-4710-b612-50dc567da0ea tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5267510b-b26f-0140-e5c0-180d5fdb56a2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1447.611804] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a282bdd6-528c-4479-b8a1-8ae112917393 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Acquiring lock "refresh_cache-a0119685-4f00-49f2-b97b-0e29175050cf" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1447.612033] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a282bdd6-528c-4479-b8a1-8ae112917393 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Acquired lock "refresh_cache-a0119685-4f00-49f2-b97b-0e29175050cf" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1447.612212] env[65726]: DEBUG nova.network.neutron [None req-a282bdd6-528c-4479-b8a1-8ae112917393 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: a0119685-4f00-49f2-b97b-0e29175050cf] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1448.086939] env[65726]: DEBUG nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] [instance: 5eae4661-4212-4932-ad1c-cee6275b859c] Instance has had 0 of 5 cleanup attempts {{(pid=65726) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 1448.105628] env[65726]: DEBUG oslo_vmware.api [None req-24f45c06-45a5-4710-b612-50dc567da0ea tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5267510b-b26f-0140-e5c0-180d5fdb56a2, 'name': SearchDatastore_Task, 'duration_secs': 0.009829} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1448.106471] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-150f066b-322e-48f1-97ab-67c25cd63f3b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1448.112412] env[65726]: DEBUG oslo_vmware.api [None req-24f45c06-45a5-4710-b612-50dc567da0ea tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Waiting for the task: (returnval){ [ 1448.112412] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]520ca517-dcb7-26c2-3e43-5959ffbc92bd" [ 1448.112412] env[65726]: _type = "Task" [ 1448.112412] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1448.116973] env[65726]: WARNING openstack [None req-a282bdd6-528c-4479-b8a1-8ae112917393 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1448.117345] env[65726]: WARNING openstack [None req-a282bdd6-528c-4479-b8a1-8ae112917393 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1448.130701] env[65726]: DEBUG oslo_vmware.api [None req-24f45c06-45a5-4710-b612-50dc567da0ea tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]520ca517-dcb7-26c2-3e43-5959ffbc92bd, 'name': SearchDatastore_Task, 'duration_secs': 0.010726} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1448.130956] env[65726]: DEBUG oslo_concurrency.lockutils [None req-24f45c06-45a5-4710-b612-50dc567da0ea tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1448.131288] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-24f45c06-45a5-4710-b612-50dc567da0ea tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore2] 08377673-6f69-4a5c-8135-7ef1683bbb55/08377673-6f69-4a5c-8135-7ef1683bbb55.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1448.131606] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c8e78ebd-050d-4fea-bae6-1a3feb63ae09 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1448.139362] env[65726]: DEBUG oslo_vmware.api [None req-24f45c06-45a5-4710-b612-50dc567da0ea tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Waiting for the task: (returnval){ [ 1448.139362] env[65726]: value = "task-5117146" [ 1448.139362] env[65726]: _type = "Task" [ 1448.139362] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1448.148598] env[65726]: DEBUG oslo_vmware.api [None req-24f45c06-45a5-4710-b612-50dc567da0ea tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': task-5117146, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1448.239793] env[65726]: WARNING openstack [None req-a282bdd6-528c-4479-b8a1-8ae112917393 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1448.240176] env[65726]: WARNING openstack [None req-a282bdd6-528c-4479-b8a1-8ae112917393 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1448.331826] env[65726]: WARNING openstack [None req-a282bdd6-528c-4479-b8a1-8ae112917393 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1448.332283] env[65726]: WARNING openstack [None req-a282bdd6-528c-4479-b8a1-8ae112917393 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1448.431747] env[65726]: DEBUG nova.network.neutron [None req-a282bdd6-528c-4479-b8a1-8ae112917393 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: a0119685-4f00-49f2-b97b-0e29175050cf] Updating instance_info_cache with network_info: [{"id": "5b57e69b-e348-46cb-932b-e1939c568791", "address": "fa:16:3e:b5:4a:f0", "network": {"id": "336fe51f-971a-447d-8e68-505640e0db1b", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1030314703-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.218", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6b56f93ba0e14521921484ac2c785c43", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8ee1c76d-1a61-4546-85cb-d4bd3c1b35ef", "external-id": "nsx-vlan-transportzone-161", "segmentation_id": 161, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5b57e69b-e3", "ovs_interfaceid": "5b57e69b-e348-46cb-932b-e1939c568791", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1448.590928] env[65726]: DEBUG nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] [instance: 58562386-ef51-44ef-869c-a3102a6aafad] Instance has had 0 of 5 cleanup attempts {{(pid=65726) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 1448.649741] env[65726]: DEBUG oslo_vmware.api [None req-24f45c06-45a5-4710-b612-50dc567da0ea tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': task-5117146, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.492591} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1448.649993] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-24f45c06-45a5-4710-b612-50dc567da0ea tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore2] 08377673-6f69-4a5c-8135-7ef1683bbb55/08377673-6f69-4a5c-8135-7ef1683bbb55.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1448.650221] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-24f45c06-45a5-4710-b612-50dc567da0ea tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 08377673-6f69-4a5c-8135-7ef1683bbb55] Extending root virtual disk to 1048576 {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1448.650490] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-bbd48491-cbcf-4f33-8c0b-f43c073c161b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1448.660255] env[65726]: DEBUG oslo_vmware.api [None req-24f45c06-45a5-4710-b612-50dc567da0ea tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Waiting for the task: (returnval){ [ 1448.660255] env[65726]: value = "task-5117147" [ 1448.660255] env[65726]: _type = "Task" [ 1448.660255] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1448.670908] env[65726]: DEBUG oslo_vmware.api [None req-24f45c06-45a5-4710-b612-50dc567da0ea tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': task-5117147, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1448.934814] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a282bdd6-528c-4479-b8a1-8ae112917393 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Releasing lock "refresh_cache-a0119685-4f00-49f2-b97b-0e29175050cf" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1449.094845] env[65726]: DEBUG nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] [instance: 1d141c86-6da9-41b9-a078-50ecf0236f85] Instance has had 0 of 5 cleanup attempts {{(pid=65726) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 1449.172109] env[65726]: DEBUG oslo_vmware.api [None req-24f45c06-45a5-4710-b612-50dc567da0ea tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': task-5117147, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.375052} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1449.172397] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-24f45c06-45a5-4710-b612-50dc567da0ea tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 08377673-6f69-4a5c-8135-7ef1683bbb55] Extended root virtual disk {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1449.173237] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e1f5b77-f30e-456a-9fdf-877c3e0bb90c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1449.196638] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-24f45c06-45a5-4710-b612-50dc567da0ea tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 08377673-6f69-4a5c-8135-7ef1683bbb55] Reconfiguring VM instance instance-0000007d to attach disk [datastore2] 08377673-6f69-4a5c-8135-7ef1683bbb55/08377673-6f69-4a5c-8135-7ef1683bbb55.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1449.196915] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a0c1d03c-7090-44b0-9ede-6cc42ccb6aef {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1449.218665] env[65726]: DEBUG oslo_vmware.api [None req-24f45c06-45a5-4710-b612-50dc567da0ea tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Waiting for the task: (returnval){ [ 1449.218665] env[65726]: value = "task-5117148" [ 1449.218665] env[65726]: _type = "Task" [ 1449.218665] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1449.228021] env[65726]: DEBUG oslo_vmware.api [None req-24f45c06-45a5-4710-b612-50dc567da0ea tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': task-5117148, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1449.598242] env[65726]: DEBUG nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] [instance: e9e93672-d509-45b3-89fe-f1c6ff4d8e78] Instance has had 0 of 5 cleanup attempts {{(pid=65726) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 1449.730135] env[65726]: DEBUG oslo_vmware.api [None req-24f45c06-45a5-4710-b612-50dc567da0ea tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': task-5117148, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1450.102019] env[65726]: DEBUG nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] [instance: a4c229bb-e5f9-41d3-86f9-ca4cfb087558] Instance has had 0 of 5 cleanup attempts {{(pid=65726) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 1450.230550] env[65726]: DEBUG oslo_vmware.api [None req-24f45c06-45a5-4710-b612-50dc567da0ea tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': task-5117148, 'name': ReconfigVM_Task, 'duration_secs': 0.694114} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1450.230846] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-24f45c06-45a5-4710-b612-50dc567da0ea tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 08377673-6f69-4a5c-8135-7ef1683bbb55] Reconfigured VM instance instance-0000007d to attach disk [datastore2] 08377673-6f69-4a5c-8135-7ef1683bbb55/08377673-6f69-4a5c-8135-7ef1683bbb55.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1450.231539] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-296aec14-5bae-4297-82e4-79f1a9985a01 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1450.238768] env[65726]: DEBUG oslo_vmware.api [None req-24f45c06-45a5-4710-b612-50dc567da0ea tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Waiting for the task: (returnval){ [ 1450.238768] env[65726]: value = "task-5117149" [ 1450.238768] env[65726]: _type = "Task" [ 1450.238768] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1450.248640] env[65726]: DEBUG oslo_vmware.api [None req-24f45c06-45a5-4710-b612-50dc567da0ea tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': task-5117149, 'name': Rename_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1450.450468] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90fad0c7-6036-4877-ae95-04744ce0edbd {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1450.470871] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-a282bdd6-528c-4479-b8a1-8ae112917393 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: a0119685-4f00-49f2-b97b-0e29175050cf] Updating instance 'a0119685-4f00-49f2-b97b-0e29175050cf' progress to 0 {{(pid=65726) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1450.604851] env[65726]: DEBUG nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] [instance: a1dc9762-ad05-4cd4-b3c3-eb644ba4212b] Instance has had 0 of 5 cleanup attempts {{(pid=65726) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 1450.751688] env[65726]: DEBUG oslo_vmware.api [None req-24f45c06-45a5-4710-b612-50dc567da0ea tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': task-5117149, 'name': Rename_Task, 'duration_secs': 0.286937} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1450.751977] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-24f45c06-45a5-4710-b612-50dc567da0ea tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 08377673-6f69-4a5c-8135-7ef1683bbb55] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1450.752255] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-87b38d64-412e-4107-8501-dc3a5e039dea {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1450.760678] env[65726]: DEBUG oslo_vmware.api [None req-24f45c06-45a5-4710-b612-50dc567da0ea tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Waiting for the task: (returnval){ [ 1450.760678] env[65726]: value = "task-5117150" [ 1450.760678] env[65726]: _type = "Task" [ 1450.760678] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1450.769962] env[65726]: DEBUG oslo_vmware.api [None req-24f45c06-45a5-4710-b612-50dc567da0ea tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': task-5117150, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1450.976742] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-a282bdd6-528c-4479-b8a1-8ae112917393 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: a0119685-4f00-49f2-b97b-0e29175050cf] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1450.977083] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d57d0a0d-49a3-4f6e-b32a-4ce65a5e4e7a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1450.984961] env[65726]: DEBUG oslo_vmware.api [None req-a282bdd6-528c-4479-b8a1-8ae112917393 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Waiting for the task: (returnval){ [ 1450.984961] env[65726]: value = "task-5117151" [ 1450.984961] env[65726]: _type = "Task" [ 1450.984961] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1450.994178] env[65726]: DEBUG oslo_vmware.api [None req-a282bdd6-528c-4479-b8a1-8ae112917393 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': task-5117151, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1451.108101] env[65726]: DEBUG nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] [instance: 86c23794-5d82-4a7a-aec6-de91601177dd] Instance has had 0 of 5 cleanup attempts {{(pid=65726) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 1451.271852] env[65726]: DEBUG oslo_vmware.api [None req-24f45c06-45a5-4710-b612-50dc567da0ea tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': task-5117150, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1451.495057] env[65726]: DEBUG oslo_vmware.api [None req-a282bdd6-528c-4479-b8a1-8ae112917393 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': task-5117151, 'name': PowerOffVM_Task, 'duration_secs': 0.182792} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1451.495394] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-a282bdd6-528c-4479-b8a1-8ae112917393 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: a0119685-4f00-49f2-b97b-0e29175050cf] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1451.495578] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-a282bdd6-528c-4479-b8a1-8ae112917393 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: a0119685-4f00-49f2-b97b-0e29175050cf] Updating instance 'a0119685-4f00-49f2-b97b-0e29175050cf' progress to 17 {{(pid=65726) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1451.611772] env[65726]: DEBUG nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] [instance: 6956bda7-5657-45d7-8f80-c6809bd836b7] Instance has had 0 of 5 cleanup attempts {{(pid=65726) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 1451.771333] env[65726]: DEBUG oslo_vmware.api [None req-24f45c06-45a5-4710-b612-50dc567da0ea tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': task-5117150, 'name': PowerOnVM_Task, 'duration_secs': 0.514012} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1451.771616] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-24f45c06-45a5-4710-b612-50dc567da0ea tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 08377673-6f69-4a5c-8135-7ef1683bbb55] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1451.771841] env[65726]: INFO nova.compute.manager [None req-24f45c06-45a5-4710-b612-50dc567da0ea tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 08377673-6f69-4a5c-8135-7ef1683bbb55] Took 7.35 seconds to spawn the instance on the hypervisor. [ 1451.772033] env[65726]: DEBUG nova.compute.manager [None req-24f45c06-45a5-4710-b612-50dc567da0ea tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 08377673-6f69-4a5c-8135-7ef1683bbb55] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1451.773092] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00aa70d7-470f-44ce-885c-af54b6d09f1b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1452.001773] env[65726]: DEBUG nova.virt.hardware [None req-a282bdd6-528c-4479-b8a1-8ae112917393 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1452.002047] env[65726]: DEBUG nova.virt.hardware [None req-a282bdd6-528c-4479-b8a1-8ae112917393 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1452.002188] env[65726]: DEBUG nova.virt.hardware [None req-a282bdd6-528c-4479-b8a1-8ae112917393 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1452.002365] env[65726]: DEBUG nova.virt.hardware [None req-a282bdd6-528c-4479-b8a1-8ae112917393 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1452.002504] env[65726]: DEBUG nova.virt.hardware [None req-a282bdd6-528c-4479-b8a1-8ae112917393 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1452.002643] env[65726]: DEBUG nova.virt.hardware [None req-a282bdd6-528c-4479-b8a1-8ae112917393 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1452.002847] env[65726]: DEBUG nova.virt.hardware [None req-a282bdd6-528c-4479-b8a1-8ae112917393 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1452.002999] env[65726]: DEBUG nova.virt.hardware [None req-a282bdd6-528c-4479-b8a1-8ae112917393 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1452.003184] env[65726]: DEBUG nova.virt.hardware [None req-a282bdd6-528c-4479-b8a1-8ae112917393 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1452.003344] env[65726]: DEBUG nova.virt.hardware [None req-a282bdd6-528c-4479-b8a1-8ae112917393 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1452.003556] env[65726]: DEBUG nova.virt.hardware [None req-a282bdd6-528c-4479-b8a1-8ae112917393 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1452.008877] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-721ac99d-31e9-4a86-8b05-6600c798733f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1452.025818] env[65726]: DEBUG oslo_vmware.api [None req-a282bdd6-528c-4479-b8a1-8ae112917393 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Waiting for the task: (returnval){ [ 1452.025818] env[65726]: value = "task-5117152" [ 1452.025818] env[65726]: _type = "Task" [ 1452.025818] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1452.037113] env[65726]: DEBUG oslo_vmware.api [None req-a282bdd6-528c-4479-b8a1-8ae112917393 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': task-5117152, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1452.115615] env[65726]: DEBUG nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] [instance: e3255f3b-028f-4a0b-b621-873730417936] Instance has had 0 of 5 cleanup attempts {{(pid=65726) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 1452.290739] env[65726]: INFO nova.compute.manager [None req-24f45c06-45a5-4710-b612-50dc567da0ea tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 08377673-6f69-4a5c-8135-7ef1683bbb55] Took 12.08 seconds to build instance. [ 1452.537137] env[65726]: DEBUG oslo_vmware.api [None req-a282bdd6-528c-4479-b8a1-8ae112917393 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': task-5117152, 'name': ReconfigVM_Task, 'duration_secs': 0.149185} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1452.537466] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-a282bdd6-528c-4479-b8a1-8ae112917393 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: a0119685-4f00-49f2-b97b-0e29175050cf] Updating instance 'a0119685-4f00-49f2-b97b-0e29175050cf' progress to 33 {{(pid=65726) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1452.618990] env[65726]: DEBUG nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] [instance: dc112e67-7c2d-4081-9a53-e4f43f61dcd6] Instance has had 0 of 5 cleanup attempts {{(pid=65726) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 1452.792941] env[65726]: DEBUG oslo_concurrency.lockutils [None req-24f45c06-45a5-4710-b612-50dc567da0ea tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Lock "08377673-6f69-4a5c-8135-7ef1683bbb55" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.594s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1453.044420] env[65726]: DEBUG nova.virt.hardware [None req-a282bdd6-528c-4479-b8a1-8ae112917393 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1453.044664] env[65726]: DEBUG nova.virt.hardware [None req-a282bdd6-528c-4479-b8a1-8ae112917393 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1453.044794] env[65726]: DEBUG nova.virt.hardware [None req-a282bdd6-528c-4479-b8a1-8ae112917393 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1453.044970] env[65726]: DEBUG nova.virt.hardware [None req-a282bdd6-528c-4479-b8a1-8ae112917393 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1453.045130] env[65726]: DEBUG nova.virt.hardware [None req-a282bdd6-528c-4479-b8a1-8ae112917393 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1453.045278] env[65726]: DEBUG nova.virt.hardware [None req-a282bdd6-528c-4479-b8a1-8ae112917393 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1453.045483] env[65726]: DEBUG nova.virt.hardware [None req-a282bdd6-528c-4479-b8a1-8ae112917393 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1453.045637] env[65726]: DEBUG nova.virt.hardware [None req-a282bdd6-528c-4479-b8a1-8ae112917393 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1453.045813] env[65726]: DEBUG nova.virt.hardware [None req-a282bdd6-528c-4479-b8a1-8ae112917393 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1453.045968] env[65726]: DEBUG nova.virt.hardware [None req-a282bdd6-528c-4479-b8a1-8ae112917393 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1453.046154] env[65726]: DEBUG nova.virt.hardware [None req-a282bdd6-528c-4479-b8a1-8ae112917393 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1453.051718] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-a282bdd6-528c-4479-b8a1-8ae112917393 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: a0119685-4f00-49f2-b97b-0e29175050cf] Reconfiguring VM instance instance-0000007c to detach disk 2000 {{(pid=65726) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1453.052068] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d9cf9ac5-bf25-4637-b131-0bf163658e4d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1453.073041] env[65726]: DEBUG oslo_vmware.api [None req-a282bdd6-528c-4479-b8a1-8ae112917393 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Waiting for the task: (returnval){ [ 1453.073041] env[65726]: value = "task-5117153" [ 1453.073041] env[65726]: _type = "Task" [ 1453.073041] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1453.083262] env[65726]: DEBUG oslo_vmware.api [None req-a282bdd6-528c-4479-b8a1-8ae112917393 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': task-5117153, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1453.122080] env[65726]: DEBUG nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] [instance: 9c0599e3-383f-46ab-809b-944cc3a4d206] Instance has had 0 of 5 cleanup attempts {{(pid=65726) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 1453.583211] env[65726]: DEBUG oslo_vmware.api [None req-a282bdd6-528c-4479-b8a1-8ae112917393 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': task-5117153, 'name': ReconfigVM_Task, 'duration_secs': 0.160436} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1453.583422] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-a282bdd6-528c-4479-b8a1-8ae112917393 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: a0119685-4f00-49f2-b97b-0e29175050cf] Reconfigured VM instance instance-0000007c to detach disk 2000 {{(pid=65726) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1453.584353] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed614d9f-9013-4e6f-935a-f678da4b6606 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1453.588914] env[65726]: DEBUG nova.compute.manager [req-9a6068d0-e030-4c0e-8a6e-6dad67efd0bd req-1f930961-321e-4910-8522-e54432f1ee01 service nova] [instance: 08377673-6f69-4a5c-8135-7ef1683bbb55] Received event network-changed-5aea6240-edaa-4fb1-b696-877bc1ba685a {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1453.589111] env[65726]: DEBUG nova.compute.manager [req-9a6068d0-e030-4c0e-8a6e-6dad67efd0bd req-1f930961-321e-4910-8522-e54432f1ee01 service nova] [instance: 08377673-6f69-4a5c-8135-7ef1683bbb55] Refreshing instance network info cache due to event network-changed-5aea6240-edaa-4fb1-b696-877bc1ba685a. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 1453.589320] env[65726]: DEBUG oslo_concurrency.lockutils [req-9a6068d0-e030-4c0e-8a6e-6dad67efd0bd req-1f930961-321e-4910-8522-e54432f1ee01 service nova] Acquiring lock "refresh_cache-08377673-6f69-4a5c-8135-7ef1683bbb55" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1453.589456] env[65726]: DEBUG oslo_concurrency.lockutils [req-9a6068d0-e030-4c0e-8a6e-6dad67efd0bd req-1f930961-321e-4910-8522-e54432f1ee01 service nova] Acquired lock "refresh_cache-08377673-6f69-4a5c-8135-7ef1683bbb55" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1453.589609] env[65726]: DEBUG nova.network.neutron [req-9a6068d0-e030-4c0e-8a6e-6dad67efd0bd req-1f930961-321e-4910-8522-e54432f1ee01 service nova] [instance: 08377673-6f69-4a5c-8135-7ef1683bbb55] Refreshing network info cache for port 5aea6240-edaa-4fb1-b696-877bc1ba685a {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 1453.612016] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-a282bdd6-528c-4479-b8a1-8ae112917393 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: a0119685-4f00-49f2-b97b-0e29175050cf] Reconfiguring VM instance instance-0000007c to attach disk [datastore2] volume-790d262c-e3cc-459b-8c42-30915c88938a/volume-790d262c-e3cc-459b-8c42-30915c88938a.vmdk or device None with type thin {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1453.613264] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-52878fed-1adf-4805-9d37-d777daa092cc {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1453.627362] env[65726]: DEBUG nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] [instance: 47e52bc1-c18e-4aa2-82a2-e4cb030a7a30] Instance has had 0 of 5 cleanup attempts {{(pid=65726) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 1453.636230] env[65726]: DEBUG oslo_vmware.api [None req-a282bdd6-528c-4479-b8a1-8ae112917393 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Waiting for the task: (returnval){ [ 1453.636230] env[65726]: value = "task-5117154" [ 1453.636230] env[65726]: _type = "Task" [ 1453.636230] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1453.650958] env[65726]: DEBUG oslo_vmware.api [None req-a282bdd6-528c-4479-b8a1-8ae112917393 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': task-5117154, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1454.093418] env[65726]: WARNING openstack [req-9a6068d0-e030-4c0e-8a6e-6dad67efd0bd req-1f930961-321e-4910-8522-e54432f1ee01 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1454.093765] env[65726]: WARNING openstack [req-9a6068d0-e030-4c0e-8a6e-6dad67efd0bd req-1f930961-321e-4910-8522-e54432f1ee01 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1454.133719] env[65726]: DEBUG nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] [instance: 5bfbaf25-5aed-46bc-97fc-f138d08ebbde] Instance has had 0 of 5 cleanup attempts {{(pid=65726) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 1454.147025] env[65726]: DEBUG oslo_vmware.api [None req-a282bdd6-528c-4479-b8a1-8ae112917393 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': task-5117154, 'name': ReconfigVM_Task, 'duration_secs': 0.263803} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1454.147198] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-a282bdd6-528c-4479-b8a1-8ae112917393 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: a0119685-4f00-49f2-b97b-0e29175050cf] Reconfigured VM instance instance-0000007c to attach disk [datastore2] volume-790d262c-e3cc-459b-8c42-30915c88938a/volume-790d262c-e3cc-459b-8c42-30915c88938a.vmdk or device None with type thin {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1454.147373] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-a282bdd6-528c-4479-b8a1-8ae112917393 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: a0119685-4f00-49f2-b97b-0e29175050cf] Updating instance 'a0119685-4f00-49f2-b97b-0e29175050cf' progress to 50 {{(pid=65726) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1454.227183] env[65726]: WARNING openstack [req-9a6068d0-e030-4c0e-8a6e-6dad67efd0bd req-1f930961-321e-4910-8522-e54432f1ee01 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1454.227642] env[65726]: WARNING openstack [req-9a6068d0-e030-4c0e-8a6e-6dad67efd0bd req-1f930961-321e-4910-8522-e54432f1ee01 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1454.298644] env[65726]: WARNING openstack [req-9a6068d0-e030-4c0e-8a6e-6dad67efd0bd req-1f930961-321e-4910-8522-e54432f1ee01 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1454.299088] env[65726]: WARNING openstack [req-9a6068d0-e030-4c0e-8a6e-6dad67efd0bd req-1f930961-321e-4910-8522-e54432f1ee01 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1454.386421] env[65726]: DEBUG nova.network.neutron [req-9a6068d0-e030-4c0e-8a6e-6dad67efd0bd req-1f930961-321e-4910-8522-e54432f1ee01 service nova] [instance: 08377673-6f69-4a5c-8135-7ef1683bbb55] Updated VIF entry in instance network info cache for port 5aea6240-edaa-4fb1-b696-877bc1ba685a. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 1454.386883] env[65726]: DEBUG nova.network.neutron [req-9a6068d0-e030-4c0e-8a6e-6dad67efd0bd req-1f930961-321e-4910-8522-e54432f1ee01 service nova] [instance: 08377673-6f69-4a5c-8135-7ef1683bbb55] Updating instance_info_cache with network_info: [{"id": "5aea6240-edaa-4fb1-b696-877bc1ba685a", "address": "fa:16:3e:70:2b:bb", "network": {"id": "a899453d-5f43-4650-9642-b33f54a4d2eb", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-614861139-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.231", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "acd800e0438940bc80ae3c6b672db4ee", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "205fb402-8eaf-4b61-8f57-8f216024179a", "external-id": "nsx-vlan-transportzone-78", "segmentation_id": 78, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5aea6240-ed", "ovs_interfaceid": "5aea6240-edaa-4fb1-b696-877bc1ba685a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1454.636912] env[65726]: DEBUG nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] [instance: 7c1188ce-8718-4719-8631-e59e7915b7aa] Instance has had 0 of 5 cleanup attempts {{(pid=65726) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 1454.655139] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0e015c2-1f63-49ac-be1c-ad32a52834b0 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1454.675946] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d489ffb7-084e-4617-9aec-e03775bf5484 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1454.695321] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-a282bdd6-528c-4479-b8a1-8ae112917393 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: a0119685-4f00-49f2-b97b-0e29175050cf] Updating instance 'a0119685-4f00-49f2-b97b-0e29175050cf' progress to 67 {{(pid=65726) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1454.892806] env[65726]: DEBUG oslo_concurrency.lockutils [req-9a6068d0-e030-4c0e-8a6e-6dad67efd0bd req-1f930961-321e-4910-8522-e54432f1ee01 service nova] Releasing lock "refresh_cache-08377673-6f69-4a5c-8135-7ef1683bbb55" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1455.140516] env[65726]: DEBUG nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] [instance: 3bc25953-21b1-4729-af09-e7211fd8b2c6] Instance has had 0 of 5 cleanup attempts {{(pid=65726) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 1455.644571] env[65726]: DEBUG nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] [instance: 378c1d18-4bb6-4245-80bf-370fb7af0575] Instance has had 0 of 5 cleanup attempts {{(pid=65726) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 1456.148417] env[65726]: DEBUG nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] [instance: b353b5e9-500c-42d3-a87f-880a624febec] Instance has had 0 of 5 cleanup attempts {{(pid=65726) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 1456.309812] env[65726]: WARNING neutronclient.v2_0.client [None req-a282bdd6-528c-4479-b8a1-8ae112917393 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1456.356938] env[65726]: DEBUG nova.network.neutron [None req-a282bdd6-528c-4479-b8a1-8ae112917393 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: a0119685-4f00-49f2-b97b-0e29175050cf] Port 5b57e69b-e348-46cb-932b-e1939c568791 binding to destination host cpu-1 is already ACTIVE {{(pid=65726) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3236}} [ 1456.651296] env[65726]: DEBUG nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] [instance: e7156a10-631a-4cc5-a544-88ce37763d02] Instance has had 0 of 5 cleanup attempts {{(pid=65726) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 1457.154672] env[65726]: DEBUG nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] [instance: 0d42fa75-20e9-4646-9b08-17015b7f068c] Instance has had 0 of 5 cleanup attempts {{(pid=65726) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 1457.380526] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a282bdd6-528c-4479-b8a1-8ae112917393 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Acquiring lock "a0119685-4f00-49f2-b97b-0e29175050cf-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1457.380717] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a282bdd6-528c-4479-b8a1-8ae112917393 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Lock "a0119685-4f00-49f2-b97b-0e29175050cf-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1457.380859] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a282bdd6-528c-4479-b8a1-8ae112917393 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Lock "a0119685-4f00-49f2-b97b-0e29175050cf-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1457.658084] env[65726]: DEBUG nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] [instance: 51f5337d-2e23-4b7c-b590-76be6bd82411] Instance has had 0 of 5 cleanup attempts {{(pid=65726) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 1458.161129] env[65726]: DEBUG oslo_service.periodic_task [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=65726) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1458.386960] env[65726]: WARNING openstack [None req-a282bdd6-528c-4479-b8a1-8ae112917393 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1458.387359] env[65726]: WARNING openstack [None req-a282bdd6-528c-4479-b8a1-8ae112917393 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1458.424241] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a282bdd6-528c-4479-b8a1-8ae112917393 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Acquiring lock "refresh_cache-a0119685-4f00-49f2-b97b-0e29175050cf" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1458.424419] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a282bdd6-528c-4479-b8a1-8ae112917393 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Acquired lock "refresh_cache-a0119685-4f00-49f2-b97b-0e29175050cf" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1458.424583] env[65726]: DEBUG nova.network.neutron [None req-a282bdd6-528c-4479-b8a1-8ae112917393 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: a0119685-4f00-49f2-b97b-0e29175050cf] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1458.928248] env[65726]: WARNING openstack [None req-a282bdd6-528c-4479-b8a1-8ae112917393 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1458.928798] env[65726]: WARNING openstack [None req-a282bdd6-528c-4479-b8a1-8ae112917393 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1459.083224] env[65726]: WARNING openstack [None req-a282bdd6-528c-4479-b8a1-8ae112917393 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1459.083604] env[65726]: WARNING openstack [None req-a282bdd6-528c-4479-b8a1-8ae112917393 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1459.150952] env[65726]: WARNING openstack [None req-a282bdd6-528c-4479-b8a1-8ae112917393 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1459.151355] env[65726]: WARNING openstack [None req-a282bdd6-528c-4479-b8a1-8ae112917393 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1459.232961] env[65726]: DEBUG nova.network.neutron [None req-a282bdd6-528c-4479-b8a1-8ae112917393 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: a0119685-4f00-49f2-b97b-0e29175050cf] Updating instance_info_cache with network_info: [{"id": "5b57e69b-e348-46cb-932b-e1939c568791", "address": "fa:16:3e:b5:4a:f0", "network": {"id": "336fe51f-971a-447d-8e68-505640e0db1b", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1030314703-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.218", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6b56f93ba0e14521921484ac2c785c43", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8ee1c76d-1a61-4546-85cb-d4bd3c1b35ef", "external-id": "nsx-vlan-transportzone-161", "segmentation_id": 161, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5b57e69b-e3", "ovs_interfaceid": "5b57e69b-e348-46cb-932b-e1939c568791", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1459.658629] env[65726]: DEBUG oslo_service.periodic_task [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=65726) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1459.658828] env[65726]: DEBUG oslo_service.periodic_task [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=65726) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1459.659028] env[65726]: DEBUG oslo_service.periodic_task [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=65726) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1459.659215] env[65726]: DEBUG oslo_service.periodic_task [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Running periodic task ComputeManager.update_available_resource {{(pid=65726) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1459.735844] env[65726]: DEBUG oslo_concurrency.lockutils [None req-a282bdd6-528c-4479-b8a1-8ae112917393 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Releasing lock "refresh_cache-a0119685-4f00-49f2-b97b-0e29175050cf" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1460.162504] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1460.162909] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1460.162984] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1460.163093] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=65726) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:937}} [ 1460.164070] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d8fe532-eb9c-4af1-b7e9-412db71559f3 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1460.173112] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-482c8afa-910c-4bb8-a0bf-bca3c4897252 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1460.188122] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1384136f-2ca2-46af-841e-6b0259e8019e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1460.195524] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-675c976d-d4b7-4688-8489-5dceb778c1d1 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1460.227891] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=179853MB free_disk=95GB free_vcpus=48 pci_devices=None {{(pid=65726) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1136}} [ 1460.228078] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1460.228303] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1460.244882] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74215173-765b-46c7-93de-fb7b9a90706a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1460.252993] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb82b17a-45bb-4ec3-b942-3d6136b6bd71 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1461.235955] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Applying migration context for instance a0119685-4f00-49f2-b97b-0e29175050cf as it has an incoming, in-progress migration 530dfaf8-ce37-484f-80e7-196316df4994. Migration status is post-migrating {{(pid=65726) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1046}} [ 1461.236864] env[65726]: INFO nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] [instance: a0119685-4f00-49f2-b97b-0e29175050cf] Updating resource usage from migration 530dfaf8-ce37-484f-80e7-196316df4994 [ 1461.288081] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Instance 55ee97c3-b014-48eb-b41b-fc5aa16bc09e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65726) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1461.288252] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Instance 95617eac-9630-4a71-80ed-a9edd8e757c1 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65726) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1461.288374] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Instance 08377673-6f69-4a5c-8135-7ef1683bbb55 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65726) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1461.288498] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Migration 530dfaf8-ce37-484f-80e7-196316df4994 is active on this compute host and has allocations in placement: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65726) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1745}} [ 1461.288701] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Instance a0119685-4f00-49f2-b97b-0e29175050cf actively managed on this compute host and has allocations in placement: {'resources': {'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=65726) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1461.288936] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Total usable vcpus: 48, total allocated vcpus: 5 {{(pid=65726) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1159}} [ 1461.289102] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1536MB phys_disk=100GB used_disk=3GB total_vcpus=48 used_vcpus=5 pci_stats=[] stats={'failed_builds': '0', 'num_instances': '4', 'num_vm_active': '4', 'num_task_None': '3', 'num_os_type_None': '4', 'num_proj_6b56f93ba0e14521921484ac2c785c43': '3', 'io_workload': '0', 'num_task_resize_finish': '1', 'num_proj_acd800e0438940bc80ae3c6b672db4ee': '1'} {{(pid=65726) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1168}} [ 1461.353159] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9203e90b-3eca-45dd-b3cb-e1e5fb92bddd {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1461.373921] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b5d0f4d-e932-4613-8eb0-b0db46cb1832 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1461.377165] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aae4f1e9-643a-4b84-86a1-bb373b5a7b6d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1461.384570] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-a282bdd6-528c-4479-b8a1-8ae112917393 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: a0119685-4f00-49f2-b97b-0e29175050cf] Updating instance 'a0119685-4f00-49f2-b97b-0e29175050cf' progress to 83 {{(pid=65726) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1461.390821] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47effc17-a041-4f91-8795-3f89d5d9647e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1461.422278] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1fd7fdaf-6ff3-4fb2-8296-0397fe9396f6 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1461.431376] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a993f0ee-66ca-4f26-8372-ba09920eb5dd {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1461.445803] env[65726]: DEBUG nova.compute.provider_tree [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1461.892856] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-a282bdd6-528c-4479-b8a1-8ae112917393 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: a0119685-4f00-49f2-b97b-0e29175050cf] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1461.893189] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f2de47be-9ca9-442a-b2e5-bcbae494c41f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1461.900694] env[65726]: DEBUG oslo_vmware.api [None req-a282bdd6-528c-4479-b8a1-8ae112917393 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Waiting for the task: (returnval){ [ 1461.900694] env[65726]: value = "task-5117155" [ 1461.900694] env[65726]: _type = "Task" [ 1461.900694] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1461.909043] env[65726]: DEBUG oslo_vmware.api [None req-a282bdd6-528c-4479-b8a1-8ae112917393 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': task-5117155, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1461.948817] env[65726]: DEBUG nova.scheduler.client.report [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1462.411137] env[65726]: DEBUG oslo_vmware.api [None req-a282bdd6-528c-4479-b8a1-8ae112917393 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': task-5117155, 'name': PowerOnVM_Task, 'duration_secs': 0.381305} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1462.411524] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-a282bdd6-528c-4479-b8a1-8ae112917393 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: a0119685-4f00-49f2-b97b-0e29175050cf] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1462.411654] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-a282bdd6-528c-4479-b8a1-8ae112917393 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: a0119685-4f00-49f2-b97b-0e29175050cf] Updating instance 'a0119685-4f00-49f2-b97b-0e29175050cf' progress to 100 {{(pid=65726) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1462.453538] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=65726) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1097}} [ 1462.453819] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.225s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1464.942345] env[65726]: DEBUG oslo_concurrency.lockutils [None req-d5a0bbfd-c62a-4d33-987c-b23e85f052b7 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Acquiring lock "a0119685-4f00-49f2-b97b-0e29175050cf" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1464.942793] env[65726]: DEBUG oslo_concurrency.lockutils [None req-d5a0bbfd-c62a-4d33-987c-b23e85f052b7 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Lock "a0119685-4f00-49f2-b97b-0e29175050cf" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.001s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1464.942793] env[65726]: DEBUG nova.compute.manager [None req-d5a0bbfd-c62a-4d33-987c-b23e85f052b7 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: a0119685-4f00-49f2-b97b-0e29175050cf] Going to confirm migration 9 {{(pid=65726) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:5307}} [ 1465.448872] env[65726]: WARNING openstack [None req-d5a0bbfd-c62a-4d33-987c-b23e85f052b7 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1465.449288] env[65726]: WARNING openstack [None req-d5a0bbfd-c62a-4d33-987c-b23e85f052b7 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1465.496244] env[65726]: WARNING openstack [None req-d5a0bbfd-c62a-4d33-987c-b23e85f052b7 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1465.496759] env[65726]: WARNING openstack [None req-d5a0bbfd-c62a-4d33-987c-b23e85f052b7 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1465.502488] env[65726]: DEBUG oslo_concurrency.lockutils [None req-d5a0bbfd-c62a-4d33-987c-b23e85f052b7 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Acquiring lock "refresh_cache-a0119685-4f00-49f2-b97b-0e29175050cf" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1465.502647] env[65726]: DEBUG oslo_concurrency.lockutils [None req-d5a0bbfd-c62a-4d33-987c-b23e85f052b7 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Acquired lock "refresh_cache-a0119685-4f00-49f2-b97b-0e29175050cf" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1465.502822] env[65726]: DEBUG nova.network.neutron [None req-d5a0bbfd-c62a-4d33-987c-b23e85f052b7 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: a0119685-4f00-49f2-b97b-0e29175050cf] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1465.503021] env[65726]: DEBUG nova.objects.instance [None req-d5a0bbfd-c62a-4d33-987c-b23e85f052b7 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Lazy-loading 'info_cache' on Instance uuid a0119685-4f00-49f2-b97b-0e29175050cf {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1466.509533] env[65726]: WARNING openstack [None req-d5a0bbfd-c62a-4d33-987c-b23e85f052b7 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1466.510039] env[65726]: WARNING openstack [None req-d5a0bbfd-c62a-4d33-987c-b23e85f052b7 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1466.629736] env[65726]: WARNING openstack [None req-d5a0bbfd-c62a-4d33-987c-b23e85f052b7 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1466.630146] env[65726]: WARNING openstack [None req-d5a0bbfd-c62a-4d33-987c-b23e85f052b7 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1466.690895] env[65726]: WARNING openstack [None req-d5a0bbfd-c62a-4d33-987c-b23e85f052b7 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1466.691282] env[65726]: WARNING openstack [None req-d5a0bbfd-c62a-4d33-987c-b23e85f052b7 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1466.770863] env[65726]: DEBUG nova.network.neutron [None req-d5a0bbfd-c62a-4d33-987c-b23e85f052b7 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: a0119685-4f00-49f2-b97b-0e29175050cf] Updating instance_info_cache with network_info: [{"id": "5b57e69b-e348-46cb-932b-e1939c568791", "address": "fa:16:3e:b5:4a:f0", "network": {"id": "336fe51f-971a-447d-8e68-505640e0db1b", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1030314703-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.218", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6b56f93ba0e14521921484ac2c785c43", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8ee1c76d-1a61-4546-85cb-d4bd3c1b35ef", "external-id": "nsx-vlan-transportzone-161", "segmentation_id": 161, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5b57e69b-e3", "ovs_interfaceid": "5b57e69b-e348-46cb-932b-e1939c568791", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1467.274696] env[65726]: DEBUG oslo_concurrency.lockutils [None req-d5a0bbfd-c62a-4d33-987c-b23e85f052b7 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Releasing lock "refresh_cache-a0119685-4f00-49f2-b97b-0e29175050cf" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1467.275099] env[65726]: DEBUG nova.objects.instance [None req-d5a0bbfd-c62a-4d33-987c-b23e85f052b7 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Lazy-loading 'migration_context' on Instance uuid a0119685-4f00-49f2-b97b-0e29175050cf {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1467.778920] env[65726]: DEBUG nova.objects.base [None req-d5a0bbfd-c62a-4d33-987c-b23e85f052b7 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Object Instance lazy-loaded attributes: info_cache,migration_context {{(pid=65726) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1467.779938] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d502f54-366d-453d-86ee-3c92c8639af8 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1467.801152] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-adc07555-ea8a-421a-9174-b1df6d491ce4 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1467.807541] env[65726]: DEBUG oslo_vmware.api [None req-d5a0bbfd-c62a-4d33-987c-b23e85f052b7 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Waiting for the task: (returnval){ [ 1467.807541] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5245c3ab-c90d-2da4-edc8-abb1f79fe721" [ 1467.807541] env[65726]: _type = "Task" [ 1467.807541] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1467.816314] env[65726]: DEBUG oslo_vmware.api [None req-d5a0bbfd-c62a-4d33-987c-b23e85f052b7 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5245c3ab-c90d-2da4-edc8-abb1f79fe721, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1468.317730] env[65726]: DEBUG oslo_vmware.api [None req-d5a0bbfd-c62a-4d33-987c-b23e85f052b7 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5245c3ab-c90d-2da4-edc8-abb1f79fe721, 'name': SearchDatastore_Task, 'duration_secs': 0.011025} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1468.318040] env[65726]: DEBUG oslo_concurrency.lockutils [None req-d5a0bbfd-c62a-4d33-987c-b23e85f052b7 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1468.318281] env[65726]: DEBUG oslo_concurrency.lockutils [None req-d5a0bbfd-c62a-4d33-987c-b23e85f052b7 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1468.896317] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78c3226f-f6d2-4bd7-8a05-1c1478ec993b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1468.904499] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1768b029-8de9-4f25-9143-e2f16646f0be {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1468.936675] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51eacc98-6eaa-4c5d-89a0-23b50cba4d39 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1468.944783] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a519459-cd22-449a-ab58-77198080b02a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1468.958493] env[65726]: DEBUG nova.compute.provider_tree [None req-d5a0bbfd-c62a-4d33-987c-b23e85f052b7 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1469.462363] env[65726]: DEBUG nova.scheduler.client.report [None req-d5a0bbfd-c62a-4d33-987c-b23e85f052b7 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1470.474673] env[65726]: DEBUG oslo_concurrency.lockutils [None req-d5a0bbfd-c62a-4d33-987c-b23e85f052b7 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.156s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1471.039354] env[65726]: INFO nova.scheduler.client.report [None req-d5a0bbfd-c62a-4d33-987c-b23e85f052b7 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Deleted allocation for migration 530dfaf8-ce37-484f-80e7-196316df4994 [ 1471.349982] env[65726]: INFO nova.compute.manager [None req-634d771d-de5a-41c6-9b94-fd91296e5a84 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: a0119685-4f00-49f2-b97b-0e29175050cf] Get console output [ 1471.350400] env[65726]: WARNING nova.virt.vmwareapi.driver [None req-634d771d-de5a-41c6-9b94-fd91296e5a84 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: a0119685-4f00-49f2-b97b-0e29175050cf] The console log is missing. Check your VSPC configuration [ 1471.545306] env[65726]: DEBUG oslo_concurrency.lockutils [None req-d5a0bbfd-c62a-4d33-987c-b23e85f052b7 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Lock "a0119685-4f00-49f2-b97b-0e29175050cf" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 6.603s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1489.678069] env[65726]: DEBUG nova.compute.manager [None req-9d36f804-83e3-49d8-bf78-46bd3e02dd5d tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 08377673-6f69-4a5c-8135-7ef1683bbb55] Stashing vm_state: active {{(pid=65726) _prep_resize /opt/stack/nova/nova/compute/manager.py:6193}} [ 1490.197355] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9d36f804-83e3-49d8-bf78-46bd3e02dd5d tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1490.197630] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9d36f804-83e3-49d8-bf78-46bd3e02dd5d tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1490.703699] env[65726]: INFO nova.compute.claims [None req-9d36f804-83e3-49d8-bf78-46bd3e02dd5d tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 08377673-6f69-4a5c-8135-7ef1683bbb55] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1491.210044] env[65726]: INFO nova.compute.resource_tracker [None req-9d36f804-83e3-49d8-bf78-46bd3e02dd5d tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 08377673-6f69-4a5c-8135-7ef1683bbb55] Updating resource usage from migration 4a3f7777-8d69-4a3c-a520-09cf29b571f3 [ 1491.282145] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa3b42a5-6265-443b-9482-9152cdfa0456 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1491.290372] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a3d9177-5a7a-496e-95ee-6763db7264c3 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1491.322237] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d932e8c9-8384-4337-b599-a9bf02a546ac {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1491.330238] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4e2472f-728a-4107-9dd5-5dd1d35227e9 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1491.343661] env[65726]: DEBUG nova.compute.provider_tree [None req-9d36f804-83e3-49d8-bf78-46bd3e02dd5d tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1491.847213] env[65726]: DEBUG nova.scheduler.client.report [None req-9d36f804-83e3-49d8-bf78-46bd3e02dd5d tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1492.352805] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9d36f804-83e3-49d8-bf78-46bd3e02dd5d tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.155s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1492.353041] env[65726]: INFO nova.compute.manager [None req-9d36f804-83e3-49d8-bf78-46bd3e02dd5d tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 08377673-6f69-4a5c-8135-7ef1683bbb55] Migrating [ 1492.868537] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9d36f804-83e3-49d8-bf78-46bd3e02dd5d tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Acquiring lock "refresh_cache-08377673-6f69-4a5c-8135-7ef1683bbb55" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1492.868932] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9d36f804-83e3-49d8-bf78-46bd3e02dd5d tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Acquired lock "refresh_cache-08377673-6f69-4a5c-8135-7ef1683bbb55" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1492.868932] env[65726]: DEBUG nova.network.neutron [None req-9d36f804-83e3-49d8-bf78-46bd3e02dd5d tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 08377673-6f69-4a5c-8135-7ef1683bbb55] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1493.372698] env[65726]: WARNING openstack [None req-9d36f804-83e3-49d8-bf78-46bd3e02dd5d tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1493.373113] env[65726]: WARNING openstack [None req-9d36f804-83e3-49d8-bf78-46bd3e02dd5d tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1493.526329] env[65726]: WARNING openstack [None req-9d36f804-83e3-49d8-bf78-46bd3e02dd5d tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1493.526746] env[65726]: WARNING openstack [None req-9d36f804-83e3-49d8-bf78-46bd3e02dd5d tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1493.600359] env[65726]: WARNING openstack [None req-9d36f804-83e3-49d8-bf78-46bd3e02dd5d tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1493.600745] env[65726]: WARNING openstack [None req-9d36f804-83e3-49d8-bf78-46bd3e02dd5d tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1493.687614] env[65726]: DEBUG nova.network.neutron [None req-9d36f804-83e3-49d8-bf78-46bd3e02dd5d tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 08377673-6f69-4a5c-8135-7ef1683bbb55] Updating instance_info_cache with network_info: [{"id": "5aea6240-edaa-4fb1-b696-877bc1ba685a", "address": "fa:16:3e:70:2b:bb", "network": {"id": "a899453d-5f43-4650-9642-b33f54a4d2eb", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-614861139-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.231", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "acd800e0438940bc80ae3c6b672db4ee", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "205fb402-8eaf-4b61-8f57-8f216024179a", "external-id": "nsx-vlan-transportzone-78", "segmentation_id": 78, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5aea6240-ed", "ovs_interfaceid": "5aea6240-edaa-4fb1-b696-877bc1ba685a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1494.190629] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9d36f804-83e3-49d8-bf78-46bd3e02dd5d tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Releasing lock "refresh_cache-08377673-6f69-4a5c-8135-7ef1683bbb55" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1495.706872] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7503147d-8b61-4856-8f63-c864c416207c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1495.725848] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-9d36f804-83e3-49d8-bf78-46bd3e02dd5d tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 08377673-6f69-4a5c-8135-7ef1683bbb55] Updating instance '08377673-6f69-4a5c-8135-7ef1683bbb55' progress to 0 {{(pid=65726) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1496.232180] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-9d36f804-83e3-49d8-bf78-46bd3e02dd5d tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 08377673-6f69-4a5c-8135-7ef1683bbb55] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1496.232610] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-bb1d1c54-eccc-4428-ac1e-c837f2345bde {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1496.240359] env[65726]: DEBUG oslo_vmware.api [None req-9d36f804-83e3-49d8-bf78-46bd3e02dd5d tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Waiting for the task: (returnval){ [ 1496.240359] env[65726]: value = "task-5117156" [ 1496.240359] env[65726]: _type = "Task" [ 1496.240359] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1496.250999] env[65726]: DEBUG oslo_vmware.api [None req-9d36f804-83e3-49d8-bf78-46bd3e02dd5d tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': task-5117156, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1496.750423] env[65726]: DEBUG oslo_vmware.api [None req-9d36f804-83e3-49d8-bf78-46bd3e02dd5d tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': task-5117156, 'name': PowerOffVM_Task, 'duration_secs': 0.258493} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1496.750857] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-9d36f804-83e3-49d8-bf78-46bd3e02dd5d tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 08377673-6f69-4a5c-8135-7ef1683bbb55] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1496.750857] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-9d36f804-83e3-49d8-bf78-46bd3e02dd5d tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 08377673-6f69-4a5c-8135-7ef1683bbb55] Updating instance '08377673-6f69-4a5c-8135-7ef1683bbb55' progress to 17 {{(pid=65726) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1497.257476] env[65726]: DEBUG nova.virt.hardware [None req-9d36f804-83e3-49d8-bf78-46bd3e02dd5d tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1497.257788] env[65726]: DEBUG nova.virt.hardware [None req-9d36f804-83e3-49d8-bf78-46bd3e02dd5d tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1497.257886] env[65726]: DEBUG nova.virt.hardware [None req-9d36f804-83e3-49d8-bf78-46bd3e02dd5d tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1497.258089] env[65726]: DEBUG nova.virt.hardware [None req-9d36f804-83e3-49d8-bf78-46bd3e02dd5d tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1497.258219] env[65726]: DEBUG nova.virt.hardware [None req-9d36f804-83e3-49d8-bf78-46bd3e02dd5d tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1497.258362] env[65726]: DEBUG nova.virt.hardware [None req-9d36f804-83e3-49d8-bf78-46bd3e02dd5d tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1497.258563] env[65726]: DEBUG nova.virt.hardware [None req-9d36f804-83e3-49d8-bf78-46bd3e02dd5d tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1497.258714] env[65726]: DEBUG nova.virt.hardware [None req-9d36f804-83e3-49d8-bf78-46bd3e02dd5d tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1497.258875] env[65726]: DEBUG nova.virt.hardware [None req-9d36f804-83e3-49d8-bf78-46bd3e02dd5d tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1497.259048] env[65726]: DEBUG nova.virt.hardware [None req-9d36f804-83e3-49d8-bf78-46bd3e02dd5d tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1497.259222] env[65726]: DEBUG nova.virt.hardware [None req-9d36f804-83e3-49d8-bf78-46bd3e02dd5d tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1497.264126] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-def53b7c-de79-4a85-bdf6-3f17316a8e95 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1497.280809] env[65726]: DEBUG oslo_vmware.api [None req-9d36f804-83e3-49d8-bf78-46bd3e02dd5d tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Waiting for the task: (returnval){ [ 1497.280809] env[65726]: value = "task-5117157" [ 1497.280809] env[65726]: _type = "Task" [ 1497.280809] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1497.289529] env[65726]: DEBUG oslo_vmware.api [None req-9d36f804-83e3-49d8-bf78-46bd3e02dd5d tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': task-5117157, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1497.792030] env[65726]: DEBUG oslo_vmware.api [None req-9d36f804-83e3-49d8-bf78-46bd3e02dd5d tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': task-5117157, 'name': ReconfigVM_Task, 'duration_secs': 0.181801} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1497.792820] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-9d36f804-83e3-49d8-bf78-46bd3e02dd5d tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 08377673-6f69-4a5c-8135-7ef1683bbb55] Updating instance '08377673-6f69-4a5c-8135-7ef1683bbb55' progress to 33 {{(pid=65726) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1498.235645] env[65726]: DEBUG oslo_concurrency.lockutils [None req-8bbfbdef-4cde-4887-bc25-41ef75e126a2 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Acquiring lock "a0119685-4f00-49f2-b97b-0e29175050cf" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1498.235946] env[65726]: DEBUG oslo_concurrency.lockutils [None req-8bbfbdef-4cde-4887-bc25-41ef75e126a2 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Lock "a0119685-4f00-49f2-b97b-0e29175050cf" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1498.236219] env[65726]: DEBUG oslo_concurrency.lockutils [None req-8bbfbdef-4cde-4887-bc25-41ef75e126a2 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Acquiring lock "a0119685-4f00-49f2-b97b-0e29175050cf-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1498.236415] env[65726]: DEBUG oslo_concurrency.lockutils [None req-8bbfbdef-4cde-4887-bc25-41ef75e126a2 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Lock "a0119685-4f00-49f2-b97b-0e29175050cf-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1498.236578] env[65726]: DEBUG oslo_concurrency.lockutils [None req-8bbfbdef-4cde-4887-bc25-41ef75e126a2 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Lock "a0119685-4f00-49f2-b97b-0e29175050cf-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1498.238987] env[65726]: INFO nova.compute.manager [None req-8bbfbdef-4cde-4887-bc25-41ef75e126a2 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: a0119685-4f00-49f2-b97b-0e29175050cf] Terminating instance [ 1498.298682] env[65726]: DEBUG nova.virt.hardware [None req-9d36f804-83e3-49d8-bf78-46bd3e02dd5d tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1498.298993] env[65726]: DEBUG nova.virt.hardware [None req-9d36f804-83e3-49d8-bf78-46bd3e02dd5d tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1498.299140] env[65726]: DEBUG nova.virt.hardware [None req-9d36f804-83e3-49d8-bf78-46bd3e02dd5d tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1498.299361] env[65726]: DEBUG nova.virt.hardware [None req-9d36f804-83e3-49d8-bf78-46bd3e02dd5d tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1498.299507] env[65726]: DEBUG nova.virt.hardware [None req-9d36f804-83e3-49d8-bf78-46bd3e02dd5d tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1498.299650] env[65726]: DEBUG nova.virt.hardware [None req-9d36f804-83e3-49d8-bf78-46bd3e02dd5d tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1498.299854] env[65726]: DEBUG nova.virt.hardware [None req-9d36f804-83e3-49d8-bf78-46bd3e02dd5d tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1498.300015] env[65726]: DEBUG nova.virt.hardware [None req-9d36f804-83e3-49d8-bf78-46bd3e02dd5d tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1498.300220] env[65726]: DEBUG nova.virt.hardware [None req-9d36f804-83e3-49d8-bf78-46bd3e02dd5d tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1498.300404] env[65726]: DEBUG nova.virt.hardware [None req-9d36f804-83e3-49d8-bf78-46bd3e02dd5d tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1498.300576] env[65726]: DEBUG nova.virt.hardware [None req-9d36f804-83e3-49d8-bf78-46bd3e02dd5d tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1498.305813] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-9d36f804-83e3-49d8-bf78-46bd3e02dd5d tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 08377673-6f69-4a5c-8135-7ef1683bbb55] Reconfiguring VM instance instance-0000007d to detach disk 2000 {{(pid=65726) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1498.306109] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-df0b9fbb-fe55-4925-9ce3-16864c6514de {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1498.326491] env[65726]: DEBUG oslo_vmware.api [None req-9d36f804-83e3-49d8-bf78-46bd3e02dd5d tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Waiting for the task: (returnval){ [ 1498.326491] env[65726]: value = "task-5117158" [ 1498.326491] env[65726]: _type = "Task" [ 1498.326491] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1498.334527] env[65726]: DEBUG oslo_vmware.api [None req-9d36f804-83e3-49d8-bf78-46bd3e02dd5d tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': task-5117158, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1498.743242] env[65726]: DEBUG nova.compute.manager [None req-8bbfbdef-4cde-4887-bc25-41ef75e126a2 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: a0119685-4f00-49f2-b97b-0e29175050cf] Start destroying the instance on the hypervisor. {{(pid=65726) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1498.743512] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-8bbfbdef-4cde-4887-bc25-41ef75e126a2 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: a0119685-4f00-49f2-b97b-0e29175050cf] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1498.743808] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a14a37fb-4e56-4aea-9f58-dfdcf9cca9e8 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1498.751876] env[65726]: DEBUG oslo_vmware.api [None req-8bbfbdef-4cde-4887-bc25-41ef75e126a2 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Waiting for the task: (returnval){ [ 1498.751876] env[65726]: value = "task-5117159" [ 1498.751876] env[65726]: _type = "Task" [ 1498.751876] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1498.759949] env[65726]: DEBUG oslo_vmware.api [None req-8bbfbdef-4cde-4887-bc25-41ef75e126a2 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': task-5117159, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1498.836970] env[65726]: DEBUG oslo_vmware.api [None req-9d36f804-83e3-49d8-bf78-46bd3e02dd5d tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': task-5117158, 'name': ReconfigVM_Task, 'duration_secs': 0.15365} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1498.837399] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-9d36f804-83e3-49d8-bf78-46bd3e02dd5d tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 08377673-6f69-4a5c-8135-7ef1683bbb55] Reconfigured VM instance instance-0000007d to detach disk 2000 {{(pid=65726) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1498.838193] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a30e2b9-98b7-4ec6-9237-d1297afa3c62 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1498.862592] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-9d36f804-83e3-49d8-bf78-46bd3e02dd5d tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 08377673-6f69-4a5c-8135-7ef1683bbb55] Reconfiguring VM instance instance-0000007d to attach disk [datastore2] 08377673-6f69-4a5c-8135-7ef1683bbb55/08377673-6f69-4a5c-8135-7ef1683bbb55.vmdk or device None with type thin {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1498.862944] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d3793f45-52c3-499c-a2a9-d5916efb5bce {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1498.883190] env[65726]: DEBUG oslo_vmware.api [None req-9d36f804-83e3-49d8-bf78-46bd3e02dd5d tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Waiting for the task: (returnval){ [ 1498.883190] env[65726]: value = "task-5117160" [ 1498.883190] env[65726]: _type = "Task" [ 1498.883190] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1498.892607] env[65726]: DEBUG oslo_vmware.api [None req-9d36f804-83e3-49d8-bf78-46bd3e02dd5d tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': task-5117160, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1499.262019] env[65726]: DEBUG oslo_vmware.api [None req-8bbfbdef-4cde-4887-bc25-41ef75e126a2 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': task-5117159, 'name': PowerOffVM_Task, 'duration_secs': 0.168159} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1499.262377] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-8bbfbdef-4cde-4887-bc25-41ef75e126a2 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: a0119685-4f00-49f2-b97b-0e29175050cf] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1499.262587] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-8bbfbdef-4cde-4887-bc25-41ef75e126a2 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: a0119685-4f00-49f2-b97b-0e29175050cf] Volume detach. Driver type: vmdk {{(pid=65726) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1499.262780] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-8bbfbdef-4cde-4887-bc25-41ef75e126a2 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: a0119685-4f00-49f2-b97b-0e29175050cf] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-995352', 'volume_id': '790d262c-e3cc-459b-8c42-30915c88938a', 'name': 'volume-790d262c-e3cc-459b-8c42-30915c88938a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'attaching', 'instance': 'a0119685-4f00-49f2-b97b-0e29175050cf', 'attached_at': '2025-12-12T19:45:24.000000', 'detached_at': '', 'volume_id': '790d262c-e3cc-459b-8c42-30915c88938a', 'serial': '790d262c-e3cc-459b-8c42-30915c88938a'} {{(pid=65726) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1499.263614] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78f7e3f2-c3b5-47bc-ba43-25b2f4a6cac7 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1499.284811] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f42e1d76-bd7a-443c-8b57-b21d93169539 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1499.292432] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eed6ccdf-b3bb-49c1-a6d3-2a66fd88addf {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1499.309615] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd4768df-f688-4a38-b215-11e0ff99c1bd {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1499.325075] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-8bbfbdef-4cde-4887-bc25-41ef75e126a2 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] The volume has not been displaced from its original location: [datastore2] volume-790d262c-e3cc-459b-8c42-30915c88938a/volume-790d262c-e3cc-459b-8c42-30915c88938a.vmdk. No consolidation needed. {{(pid=65726) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1499.330379] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-8bbfbdef-4cde-4887-bc25-41ef75e126a2 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: a0119685-4f00-49f2-b97b-0e29175050cf] Reconfiguring VM instance instance-0000007c to detach disk 2000 {{(pid=65726) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1499.330563] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b27b5cd8-c519-4b52-9863-2508c4326e6f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1499.350358] env[65726]: DEBUG oslo_vmware.api [None req-8bbfbdef-4cde-4887-bc25-41ef75e126a2 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Waiting for the task: (returnval){ [ 1499.350358] env[65726]: value = "task-5117161" [ 1499.350358] env[65726]: _type = "Task" [ 1499.350358] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1499.359919] env[65726]: DEBUG oslo_vmware.api [None req-8bbfbdef-4cde-4887-bc25-41ef75e126a2 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': task-5117161, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1499.393154] env[65726]: DEBUG oslo_vmware.api [None req-9d36f804-83e3-49d8-bf78-46bd3e02dd5d tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': task-5117160, 'name': ReconfigVM_Task, 'duration_secs': 0.262067} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1499.393458] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-9d36f804-83e3-49d8-bf78-46bd3e02dd5d tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 08377673-6f69-4a5c-8135-7ef1683bbb55] Reconfigured VM instance instance-0000007d to attach disk [datastore2] 08377673-6f69-4a5c-8135-7ef1683bbb55/08377673-6f69-4a5c-8135-7ef1683bbb55.vmdk or device None with type thin {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1499.393717] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-9d36f804-83e3-49d8-bf78-46bd3e02dd5d tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 08377673-6f69-4a5c-8135-7ef1683bbb55] Updating instance '08377673-6f69-4a5c-8135-7ef1683bbb55' progress to 50 {{(pid=65726) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1499.860668] env[65726]: DEBUG oslo_vmware.api [None req-8bbfbdef-4cde-4887-bc25-41ef75e126a2 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': task-5117161, 'name': ReconfigVM_Task, 'duration_secs': 0.161177} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1499.861090] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-8bbfbdef-4cde-4887-bc25-41ef75e126a2 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: a0119685-4f00-49f2-b97b-0e29175050cf] Reconfigured VM instance instance-0000007c to detach disk 2000 {{(pid=65726) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1499.865803] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2896d3fb-9cc9-4755-ab74-081844a98091 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1499.881635] env[65726]: DEBUG oslo_vmware.api [None req-8bbfbdef-4cde-4887-bc25-41ef75e126a2 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Waiting for the task: (returnval){ [ 1499.881635] env[65726]: value = "task-5117162" [ 1499.881635] env[65726]: _type = "Task" [ 1499.881635] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1499.890268] env[65726]: DEBUG oslo_vmware.api [None req-8bbfbdef-4cde-4887-bc25-41ef75e126a2 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': task-5117162, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1499.900804] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c77d2a50-475d-4bc9-b796-e28a3621d767 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1499.919431] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-801472a5-4960-4071-958f-2f3e31731c71 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1499.937632] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-9d36f804-83e3-49d8-bf78-46bd3e02dd5d tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 08377673-6f69-4a5c-8135-7ef1683bbb55] Updating instance '08377673-6f69-4a5c-8135-7ef1683bbb55' progress to 67 {{(pid=65726) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1500.395078] env[65726]: DEBUG oslo_vmware.api [None req-8bbfbdef-4cde-4887-bc25-41ef75e126a2 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': task-5117162, 'name': ReconfigVM_Task, 'duration_secs': 0.150567} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1500.395262] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-8bbfbdef-4cde-4887-bc25-41ef75e126a2 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: a0119685-4f00-49f2-b97b-0e29175050cf] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-995352', 'volume_id': '790d262c-e3cc-459b-8c42-30915c88938a', 'name': 'volume-790d262c-e3cc-459b-8c42-30915c88938a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'attaching', 'instance': 'a0119685-4f00-49f2-b97b-0e29175050cf', 'attached_at': '2025-12-12T19:45:24.000000', 'detached_at': '', 'volume_id': '790d262c-e3cc-459b-8c42-30915c88938a', 'serial': '790d262c-e3cc-459b-8c42-30915c88938a'} {{(pid=65726) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1500.395580] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-8bbfbdef-4cde-4887-bc25-41ef75e126a2 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: a0119685-4f00-49f2-b97b-0e29175050cf] Destroying instance {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1500.396793] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a3ae72c-9f31-44f2-bf32-6b6197824b38 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1500.405592] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-8bbfbdef-4cde-4887-bc25-41ef75e126a2 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: a0119685-4f00-49f2-b97b-0e29175050cf] Unregistering the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1500.405930] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-abd5e465-7582-4dcb-b107-837e491658ff {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1500.445596] env[65726]: WARNING neutronclient.v2_0.client [None req-9d36f804-83e3-49d8-bf78-46bd3e02dd5d tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1500.476026] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-8bbfbdef-4cde-4887-bc25-41ef75e126a2 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: a0119685-4f00-49f2-b97b-0e29175050cf] Unregistered the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1500.476026] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-8bbfbdef-4cde-4887-bc25-41ef75e126a2 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: a0119685-4f00-49f2-b97b-0e29175050cf] Deleting contents of the VM from datastore datastore2 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1500.476026] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-8bbfbdef-4cde-4887-bc25-41ef75e126a2 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Deleting the datastore file [datastore2] a0119685-4f00-49f2-b97b-0e29175050cf {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1500.476651] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-525b87e0-691c-4465-8996-db6bd5fd3802 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1500.484287] env[65726]: DEBUG oslo_vmware.api [None req-8bbfbdef-4cde-4887-bc25-41ef75e126a2 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Waiting for the task: (returnval){ [ 1500.484287] env[65726]: value = "task-5117164" [ 1500.484287] env[65726]: _type = "Task" [ 1500.484287] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1500.494980] env[65726]: DEBUG oslo_vmware.api [None req-8bbfbdef-4cde-4887-bc25-41ef75e126a2 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': task-5117164, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1500.546347] env[65726]: DEBUG nova.network.neutron [None req-9d36f804-83e3-49d8-bf78-46bd3e02dd5d tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 08377673-6f69-4a5c-8135-7ef1683bbb55] Port 5aea6240-edaa-4fb1-b696-877bc1ba685a binding to destination host cpu-1 is already ACTIVE {{(pid=65726) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3236}} [ 1500.994837] env[65726]: DEBUG oslo_vmware.api [None req-8bbfbdef-4cde-4887-bc25-41ef75e126a2 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': task-5117164, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.09089} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1500.995236] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-8bbfbdef-4cde-4887-bc25-41ef75e126a2 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Deleted the datastore file {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1500.995236] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-8bbfbdef-4cde-4887-bc25-41ef75e126a2 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: a0119685-4f00-49f2-b97b-0e29175050cf] Deleted contents of the VM from datastore datastore2 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1500.995386] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-8bbfbdef-4cde-4887-bc25-41ef75e126a2 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: a0119685-4f00-49f2-b97b-0e29175050cf] Instance destroyed {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1500.995550] env[65726]: INFO nova.compute.manager [None req-8bbfbdef-4cde-4887-bc25-41ef75e126a2 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: a0119685-4f00-49f2-b97b-0e29175050cf] Took 2.25 seconds to destroy the instance on the hypervisor. [ 1500.995789] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-8bbfbdef-4cde-4887-bc25-41ef75e126a2 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1500.995976] env[65726]: DEBUG nova.compute.manager [-] [instance: a0119685-4f00-49f2-b97b-0e29175050cf] Deallocating network for instance {{(pid=65726) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1500.996080] env[65726]: DEBUG nova.network.neutron [-] [instance: a0119685-4f00-49f2-b97b-0e29175050cf] deallocate_for_instance() {{(pid=65726) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 1500.996573] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1500.996873] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1501.033504] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1501.033801] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1501.453581] env[65726]: DEBUG nova.compute.manager [req-260e6b15-88e3-43f1-a9b5-096a45669a8f req-8a31ebac-b9b8-411d-8aca-3867817870b4 service nova] [instance: a0119685-4f00-49f2-b97b-0e29175050cf] Received event network-vif-deleted-5b57e69b-e348-46cb-932b-e1939c568791 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1501.453789] env[65726]: INFO nova.compute.manager [req-260e6b15-88e3-43f1-a9b5-096a45669a8f req-8a31ebac-b9b8-411d-8aca-3867817870b4 service nova] [instance: a0119685-4f00-49f2-b97b-0e29175050cf] Neutron deleted interface 5b57e69b-e348-46cb-932b-e1939c568791; detaching it from the instance and deleting it from the info cache [ 1501.454027] env[65726]: DEBUG nova.network.neutron [req-260e6b15-88e3-43f1-a9b5-096a45669a8f req-8a31ebac-b9b8-411d-8aca-3867817870b4 service nova] [instance: a0119685-4f00-49f2-b97b-0e29175050cf] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1501.568884] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9d36f804-83e3-49d8-bf78-46bd3e02dd5d tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Acquiring lock "08377673-6f69-4a5c-8135-7ef1683bbb55-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1501.569153] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9d36f804-83e3-49d8-bf78-46bd3e02dd5d tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Lock "08377673-6f69-4a5c-8135-7ef1683bbb55-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1501.569341] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9d36f804-83e3-49d8-bf78-46bd3e02dd5d tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Lock "08377673-6f69-4a5c-8135-7ef1683bbb55-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1501.939376] env[65726]: DEBUG nova.network.neutron [-] [instance: a0119685-4f00-49f2-b97b-0e29175050cf] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1501.956348] env[65726]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-787575a2-0742-42a4-9aed-77b765f1df8a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1501.967228] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23a3607c-2389-45e7-a819-c402171b78b5 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1501.996606] env[65726]: DEBUG nova.compute.manager [req-260e6b15-88e3-43f1-a9b5-096a45669a8f req-8a31ebac-b9b8-411d-8aca-3867817870b4 service nova] [instance: a0119685-4f00-49f2-b97b-0e29175050cf] Detach interface failed, port_id=5b57e69b-e348-46cb-932b-e1939c568791, reason: Instance a0119685-4f00-49f2-b97b-0e29175050cf could not be found. {{(pid=65726) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11669}} [ 1502.442521] env[65726]: INFO nova.compute.manager [-] [instance: a0119685-4f00-49f2-b97b-0e29175050cf] Took 1.45 seconds to deallocate network for instance. [ 1502.575931] env[65726]: WARNING openstack [None req-9d36f804-83e3-49d8-bf78-46bd3e02dd5d tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1502.575931] env[65726]: WARNING openstack [None req-9d36f804-83e3-49d8-bf78-46bd3e02dd5d tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1502.620085] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9d36f804-83e3-49d8-bf78-46bd3e02dd5d tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Acquiring lock "refresh_cache-08377673-6f69-4a5c-8135-7ef1683bbb55" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1502.620270] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9d36f804-83e3-49d8-bf78-46bd3e02dd5d tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Acquired lock "refresh_cache-08377673-6f69-4a5c-8135-7ef1683bbb55" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1502.620441] env[65726]: DEBUG nova.network.neutron [None req-9d36f804-83e3-49d8-bf78-46bd3e02dd5d tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 08377673-6f69-4a5c-8135-7ef1683bbb55] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1502.989967] env[65726]: INFO nova.compute.manager [None req-8bbfbdef-4cde-4887-bc25-41ef75e126a2 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: a0119685-4f00-49f2-b97b-0e29175050cf] Took 0.55 seconds to detach 1 volumes for instance. [ 1502.992502] env[65726]: DEBUG nova.compute.manager [None req-8bbfbdef-4cde-4887-bc25-41ef75e126a2 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: a0119685-4f00-49f2-b97b-0e29175050cf] Deleting volume: 790d262c-e3cc-459b-8c42-30915c88938a {{(pid=65726) _cleanup_volumes /opt/stack/nova/nova/compute/manager.py:3319}} [ 1503.123722] env[65726]: WARNING openstack [None req-9d36f804-83e3-49d8-bf78-46bd3e02dd5d tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1503.124165] env[65726]: WARNING openstack [None req-9d36f804-83e3-49d8-bf78-46bd3e02dd5d tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1503.248350] env[65726]: WARNING openstack [None req-9d36f804-83e3-49d8-bf78-46bd3e02dd5d tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1503.248754] env[65726]: WARNING openstack [None req-9d36f804-83e3-49d8-bf78-46bd3e02dd5d tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1503.313836] env[65726]: WARNING openstack [None req-9d36f804-83e3-49d8-bf78-46bd3e02dd5d tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1503.314229] env[65726]: WARNING openstack [None req-9d36f804-83e3-49d8-bf78-46bd3e02dd5d tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1503.396223] env[65726]: DEBUG nova.network.neutron [None req-9d36f804-83e3-49d8-bf78-46bd3e02dd5d tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 08377673-6f69-4a5c-8135-7ef1683bbb55] Updating instance_info_cache with network_info: [{"id": "5aea6240-edaa-4fb1-b696-877bc1ba685a", "address": "fa:16:3e:70:2b:bb", "network": {"id": "a899453d-5f43-4650-9642-b33f54a4d2eb", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-614861139-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.231", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "acd800e0438940bc80ae3c6b672db4ee", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "205fb402-8eaf-4b61-8f57-8f216024179a", "external-id": "nsx-vlan-transportzone-78", "segmentation_id": 78, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5aea6240-ed", "ovs_interfaceid": "5aea6240-edaa-4fb1-b696-877bc1ba685a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1503.534830] env[65726]: DEBUG oslo_concurrency.lockutils [None req-8bbfbdef-4cde-4887-bc25-41ef75e126a2 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1503.535139] env[65726]: DEBUG oslo_concurrency.lockutils [None req-8bbfbdef-4cde-4887-bc25-41ef75e126a2 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1503.535328] env[65726]: DEBUG oslo_concurrency.lockutils [None req-8bbfbdef-4cde-4887-bc25-41ef75e126a2 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1503.556320] env[65726]: INFO nova.scheduler.client.report [None req-8bbfbdef-4cde-4887-bc25-41ef75e126a2 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Deleted allocations for instance a0119685-4f00-49f2-b97b-0e29175050cf [ 1503.899447] env[65726]: DEBUG oslo_concurrency.lockutils [None req-9d36f804-83e3-49d8-bf78-46bd3e02dd5d tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Releasing lock "refresh_cache-08377673-6f69-4a5c-8135-7ef1683bbb55" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1504.064820] env[65726]: DEBUG oslo_concurrency.lockutils [None req-8bbfbdef-4cde-4887-bc25-41ef75e126a2 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Lock "a0119685-4f00-49f2-b97b-0e29175050cf" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.829s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1504.421602] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fa9ca78-4ce4-46e5-a808-d9b0a3edb214 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1504.440931] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d826689-608b-43d8-93b3-02eed2285443 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1504.448728] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-9d36f804-83e3-49d8-bf78-46bd3e02dd5d tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 08377673-6f69-4a5c-8135-7ef1683bbb55] Updating instance '08377673-6f69-4a5c-8135-7ef1683bbb55' progress to 83 {{(pid=65726) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1504.955742] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-9d36f804-83e3-49d8-bf78-46bd3e02dd5d tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 08377673-6f69-4a5c-8135-7ef1683bbb55] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1504.955742] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-115614d8-b201-4b4b-a7a0-5d64b82a64d6 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1504.965023] env[65726]: DEBUG oslo_vmware.api [None req-9d36f804-83e3-49d8-bf78-46bd3e02dd5d tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Waiting for the task: (returnval){ [ 1504.965023] env[65726]: value = "task-5117166" [ 1504.965023] env[65726]: _type = "Task" [ 1504.965023] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1504.973049] env[65726]: DEBUG oslo_vmware.api [None req-9d36f804-83e3-49d8-bf78-46bd3e02dd5d tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': task-5117166, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1504.990802] env[65726]: DEBUG oslo_concurrency.lockutils [None req-cda62c61-34a1-49ed-a6b7-5d6e24903605 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Acquiring lock "95617eac-9630-4a71-80ed-a9edd8e757c1" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1504.991149] env[65726]: DEBUG oslo_concurrency.lockutils [None req-cda62c61-34a1-49ed-a6b7-5d6e24903605 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Lock "95617eac-9630-4a71-80ed-a9edd8e757c1" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1504.991369] env[65726]: DEBUG oslo_concurrency.lockutils [None req-cda62c61-34a1-49ed-a6b7-5d6e24903605 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Acquiring lock "95617eac-9630-4a71-80ed-a9edd8e757c1-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1504.991550] env[65726]: DEBUG oslo_concurrency.lockutils [None req-cda62c61-34a1-49ed-a6b7-5d6e24903605 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Lock "95617eac-9630-4a71-80ed-a9edd8e757c1-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1504.991715] env[65726]: DEBUG oslo_concurrency.lockutils [None req-cda62c61-34a1-49ed-a6b7-5d6e24903605 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Lock "95617eac-9630-4a71-80ed-a9edd8e757c1-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1504.994055] env[65726]: INFO nova.compute.manager [None req-cda62c61-34a1-49ed-a6b7-5d6e24903605 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 95617eac-9630-4a71-80ed-a9edd8e757c1] Terminating instance [ 1505.476299] env[65726]: DEBUG oslo_vmware.api [None req-9d36f804-83e3-49d8-bf78-46bd3e02dd5d tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': task-5117166, 'name': PowerOnVM_Task, 'duration_secs': 0.406947} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1505.476643] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-9d36f804-83e3-49d8-bf78-46bd3e02dd5d tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 08377673-6f69-4a5c-8135-7ef1683bbb55] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1505.476781] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-9d36f804-83e3-49d8-bf78-46bd3e02dd5d tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 08377673-6f69-4a5c-8135-7ef1683bbb55] Updating instance '08377673-6f69-4a5c-8135-7ef1683bbb55' progress to 100 {{(pid=65726) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1505.499746] env[65726]: DEBUG nova.compute.manager [None req-cda62c61-34a1-49ed-a6b7-5d6e24903605 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 95617eac-9630-4a71-80ed-a9edd8e757c1] Start destroying the instance on the hypervisor. {{(pid=65726) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1505.500962] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-cda62c61-34a1-49ed-a6b7-5d6e24903605 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 95617eac-9630-4a71-80ed-a9edd8e757c1] Destroying instance {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1505.500962] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-896d7e88-9fd7-4889-a123-7c11aa074a20 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1505.508909] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-cda62c61-34a1-49ed-a6b7-5d6e24903605 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 95617eac-9630-4a71-80ed-a9edd8e757c1] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1505.509195] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c60b0d02-66f3-417d-b23c-20f536a56b5c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1505.516293] env[65726]: DEBUG oslo_vmware.api [None req-cda62c61-34a1-49ed-a6b7-5d6e24903605 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Waiting for the task: (returnval){ [ 1505.516293] env[65726]: value = "task-5117167" [ 1505.516293] env[65726]: _type = "Task" [ 1505.516293] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1505.526116] env[65726]: DEBUG oslo_vmware.api [None req-cda62c61-34a1-49ed-a6b7-5d6e24903605 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': task-5117167, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1506.028268] env[65726]: DEBUG oslo_vmware.api [None req-cda62c61-34a1-49ed-a6b7-5d6e24903605 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': task-5117167, 'name': PowerOffVM_Task, 'duration_secs': 0.208763} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1506.028758] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-cda62c61-34a1-49ed-a6b7-5d6e24903605 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 95617eac-9630-4a71-80ed-a9edd8e757c1] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1506.029056] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-cda62c61-34a1-49ed-a6b7-5d6e24903605 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 95617eac-9630-4a71-80ed-a9edd8e757c1] Unregistering the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1506.029440] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-da649492-e482-43e6-9909-2635b7eaba91 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1506.099970] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-cda62c61-34a1-49ed-a6b7-5d6e24903605 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 95617eac-9630-4a71-80ed-a9edd8e757c1] Unregistered the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1506.100229] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-cda62c61-34a1-49ed-a6b7-5d6e24903605 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 95617eac-9630-4a71-80ed-a9edd8e757c1] Deleting contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1506.100412] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-cda62c61-34a1-49ed-a6b7-5d6e24903605 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Deleting the datastore file [datastore1] 95617eac-9630-4a71-80ed-a9edd8e757c1 {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1506.100694] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-38d09211-03f9-49a7-b5d8-f06e5debb38c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1506.109239] env[65726]: DEBUG oslo_vmware.api [None req-cda62c61-34a1-49ed-a6b7-5d6e24903605 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Waiting for the task: (returnval){ [ 1506.109239] env[65726]: value = "task-5117169" [ 1506.109239] env[65726]: _type = "Task" [ 1506.109239] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1506.117860] env[65726]: DEBUG oslo_vmware.api [None req-cda62c61-34a1-49ed-a6b7-5d6e24903605 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': task-5117169, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1506.619989] env[65726]: DEBUG oslo_vmware.api [None req-cda62c61-34a1-49ed-a6b7-5d6e24903605 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': task-5117169, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.180842} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1506.620323] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-cda62c61-34a1-49ed-a6b7-5d6e24903605 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Deleted the datastore file {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1506.620439] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-cda62c61-34a1-49ed-a6b7-5d6e24903605 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 95617eac-9630-4a71-80ed-a9edd8e757c1] Deleted contents of the VM from datastore datastore1 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1506.620613] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-cda62c61-34a1-49ed-a6b7-5d6e24903605 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 95617eac-9630-4a71-80ed-a9edd8e757c1] Instance destroyed {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1506.620849] env[65726]: INFO nova.compute.manager [None req-cda62c61-34a1-49ed-a6b7-5d6e24903605 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 95617eac-9630-4a71-80ed-a9edd8e757c1] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1506.621049] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-cda62c61-34a1-49ed-a6b7-5d6e24903605 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1506.621265] env[65726]: DEBUG nova.compute.manager [-] [instance: 95617eac-9630-4a71-80ed-a9edd8e757c1] Deallocating network for instance {{(pid=65726) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1506.621363] env[65726]: DEBUG nova.network.neutron [-] [instance: 95617eac-9630-4a71-80ed-a9edd8e757c1] deallocate_for_instance() {{(pid=65726) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 1506.621867] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1506.622174] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1506.670009] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1506.670403] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1506.919418] env[65726]: DEBUG nova.compute.manager [req-77e62e88-a75f-41ed-b181-f8f56e848b92 req-399b3524-cd9a-4a65-81ea-0347c90035a9 service nova] [instance: 95617eac-9630-4a71-80ed-a9edd8e757c1] Received event network-vif-deleted-b46e6107-621c-438e-9be6-fe1cd95a8a4e {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1506.920028] env[65726]: INFO nova.compute.manager [req-77e62e88-a75f-41ed-b181-f8f56e848b92 req-399b3524-cd9a-4a65-81ea-0347c90035a9 service nova] [instance: 95617eac-9630-4a71-80ed-a9edd8e757c1] Neutron deleted interface b46e6107-621c-438e-9be6-fe1cd95a8a4e; detaching it from the instance and deleting it from the info cache [ 1506.920028] env[65726]: DEBUG nova.network.neutron [req-77e62e88-a75f-41ed-b181-f8f56e848b92 req-399b3524-cd9a-4a65-81ea-0347c90035a9 service nova] [instance: 95617eac-9630-4a71-80ed-a9edd8e757c1] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1507.391318] env[65726]: DEBUG nova.network.neutron [-] [instance: 95617eac-9630-4a71-80ed-a9edd8e757c1] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1507.422763] env[65726]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-95103ff5-243b-4bcc-835e-70b43aba60ac {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1507.435472] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbf85504-a363-4370-b9ff-81aa03cb422a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1507.468062] env[65726]: DEBUG nova.compute.manager [req-77e62e88-a75f-41ed-b181-f8f56e848b92 req-399b3524-cd9a-4a65-81ea-0347c90035a9 service nova] [instance: 95617eac-9630-4a71-80ed-a9edd8e757c1] Detach interface failed, port_id=b46e6107-621c-438e-9be6-fe1cd95a8a4e, reason: Instance 95617eac-9630-4a71-80ed-a9edd8e757c1 could not be found. {{(pid=65726) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11669}} [ 1507.494458] env[65726]: WARNING openstack [None req-96e312fd-0b71-43ed-baa9-b065b31062a1 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1507.494883] env[65726]: WARNING openstack [None req-96e312fd-0b71-43ed-baa9-b065b31062a1 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1507.549718] env[65726]: WARNING openstack [None req-96e312fd-0b71-43ed-baa9-b065b31062a1 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1507.550197] env[65726]: WARNING openstack [None req-96e312fd-0b71-43ed-baa9-b065b31062a1 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1507.555546] env[65726]: WARNING neutronclient.v2_0.client [None req-96e312fd-0b71-43ed-baa9-b065b31062a1 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1507.601586] env[65726]: DEBUG nova.network.neutron [None req-96e312fd-0b71-43ed-baa9-b065b31062a1 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 08377673-6f69-4a5c-8135-7ef1683bbb55] Port 5aea6240-edaa-4fb1-b696-877bc1ba685a binding to destination host cpu-1 is already ACTIVE {{(pid=65726) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3236}} [ 1507.601872] env[65726]: DEBUG oslo_concurrency.lockutils [None req-96e312fd-0b71-43ed-baa9-b065b31062a1 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Acquiring lock "refresh_cache-08377673-6f69-4a5c-8135-7ef1683bbb55" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1507.602048] env[65726]: DEBUG oslo_concurrency.lockutils [None req-96e312fd-0b71-43ed-baa9-b065b31062a1 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Acquired lock "refresh_cache-08377673-6f69-4a5c-8135-7ef1683bbb55" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1507.602213] env[65726]: DEBUG nova.network.neutron [None req-96e312fd-0b71-43ed-baa9-b065b31062a1 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 08377673-6f69-4a5c-8135-7ef1683bbb55] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1507.893831] env[65726]: INFO nova.compute.manager [-] [instance: 95617eac-9630-4a71-80ed-a9edd8e757c1] Took 1.27 seconds to deallocate network for instance. [ 1508.105920] env[65726]: WARNING openstack [None req-96e312fd-0b71-43ed-baa9-b065b31062a1 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1508.106324] env[65726]: WARNING openstack [None req-96e312fd-0b71-43ed-baa9-b065b31062a1 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1508.234495] env[65726]: WARNING openstack [None req-96e312fd-0b71-43ed-baa9-b065b31062a1 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1508.234887] env[65726]: WARNING openstack [None req-96e312fd-0b71-43ed-baa9-b065b31062a1 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1508.299921] env[65726]: WARNING openstack [None req-96e312fd-0b71-43ed-baa9-b065b31062a1 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1508.300353] env[65726]: WARNING openstack [None req-96e312fd-0b71-43ed-baa9-b065b31062a1 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1508.364480] env[65726]: DEBUG oslo_service.periodic_task [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=65726) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1508.364693] env[65726]: DEBUG oslo_service.periodic_task [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=65726) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1508.364850] env[65726]: DEBUG oslo_service.periodic_task [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=65726) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1508.364994] env[65726]: DEBUG nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=65726) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11251}} [ 1508.384333] env[65726]: DEBUG nova.network.neutron [None req-96e312fd-0b71-43ed-baa9-b065b31062a1 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 08377673-6f69-4a5c-8135-7ef1683bbb55] Updating instance_info_cache with network_info: [{"id": "5aea6240-edaa-4fb1-b696-877bc1ba685a", "address": "fa:16:3e:70:2b:bb", "network": {"id": "a899453d-5f43-4650-9642-b33f54a4d2eb", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-614861139-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.231", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "acd800e0438940bc80ae3c6b672db4ee", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "205fb402-8eaf-4b61-8f57-8f216024179a", "external-id": "nsx-vlan-transportzone-78", "segmentation_id": 78, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5aea6240-ed", "ovs_interfaceid": "5aea6240-edaa-4fb1-b696-877bc1ba685a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1508.400902] env[65726]: DEBUG oslo_concurrency.lockutils [None req-cda62c61-34a1-49ed-a6b7-5d6e24903605 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1508.401189] env[65726]: DEBUG oslo_concurrency.lockutils [None req-cda62c61-34a1-49ed-a6b7-5d6e24903605 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1508.401406] env[65726]: DEBUG nova.objects.instance [None req-cda62c61-34a1-49ed-a6b7-5d6e24903605 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Lazy-loading 'resources' on Instance uuid 95617eac-9630-4a71-80ed-a9edd8e757c1 {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1508.564708] env[65726]: DEBUG oslo_service.periodic_task [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=65726) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1508.569322] env[65726]: DEBUG oslo_service.periodic_task [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=65726) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1508.569490] env[65726]: DEBUG oslo_service.periodic_task [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=65726) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1508.887427] env[65726]: DEBUG oslo_concurrency.lockutils [None req-96e312fd-0b71-43ed-baa9-b065b31062a1 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Releasing lock "refresh_cache-08377673-6f69-4a5c-8135-7ef1683bbb55" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1508.968397] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6bd3ff3f-5822-475f-ab2f-ea2abf6c7904 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1508.977148] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0767adc-8d38-4aa2-9cc3-6cce6e2208c5 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1509.007177] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a424ef1b-5496-4c19-85d2-dc727e64d9b8 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1509.015824] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5074126-c726-4d2b-8522-d75a65597ed9 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1509.032046] env[65726]: DEBUG nova.compute.provider_tree [None req-cda62c61-34a1-49ed-a6b7-5d6e24903605 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1509.391468] env[65726]: DEBUG nova.compute.manager [None req-96e312fd-0b71-43ed-baa9-b065b31062a1 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 08377673-6f69-4a5c-8135-7ef1683bbb55] Hypervisor driver does not support instance shared storage check, assuming it's not on shared storage {{(pid=65726) _is_instance_storage_shared /opt/stack/nova/nova/compute/manager.py:924}} [ 1509.391717] env[65726]: DEBUG oslo_concurrency.lockutils [None req-96e312fd-0b71-43ed-baa9-b065b31062a1 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1509.535371] env[65726]: DEBUG nova.scheduler.client.report [None req-cda62c61-34a1-49ed-a6b7-5d6e24903605 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1510.041247] env[65726]: DEBUG oslo_concurrency.lockutils [None req-cda62c61-34a1-49ed-a6b7-5d6e24903605 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.640s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1510.044027] env[65726]: DEBUG oslo_concurrency.lockutils [None req-96e312fd-0b71-43ed-baa9-b065b31062a1 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: waited 0.652s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1510.063746] env[65726]: INFO nova.scheduler.client.report [None req-cda62c61-34a1-49ed-a6b7-5d6e24903605 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Deleted allocations for instance 95617eac-9630-4a71-80ed-a9edd8e757c1 [ 1510.546892] env[65726]: DEBUG nova.objects.instance [None req-96e312fd-0b71-43ed-baa9-b065b31062a1 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Lazy-loading 'migration_context' on Instance uuid 08377673-6f69-4a5c-8135-7ef1683bbb55 {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1510.570506] env[65726]: DEBUG oslo_concurrency.lockutils [None req-cda62c61-34a1-49ed-a6b7-5d6e24903605 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Lock "95617eac-9630-4a71-80ed-a9edd8e757c1" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.579s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1511.105176] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09ed0b1d-28ec-46e1-a5bc-92075524249d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1511.113017] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7644fed6-a190-4495-b14d-382ff26cf0a2 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1511.142801] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5130b4d-0c73-482b-8946-5593b9d35eaa {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1511.151114] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05785406-71c1-490a-868c-f40689811457 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1511.165656] env[65726]: DEBUG nova.compute.provider_tree [None req-96e312fd-0b71-43ed-baa9-b065b31062a1 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1511.568767] env[65726]: DEBUG oslo_service.periodic_task [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=65726) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1511.669355] env[65726]: DEBUG nova.scheduler.client.report [None req-96e312fd-0b71-43ed-baa9-b065b31062a1 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1512.681585] env[65726]: DEBUG oslo_concurrency.lockutils [None req-96e312fd-0b71-43ed-baa9-b065b31062a1 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: held 2.638s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1513.569243] env[65726]: DEBUG oslo_service.periodic_task [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Running periodic task ComputeManager.update_available_resource {{(pid=65726) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1514.072463] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1514.072793] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1514.072825] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1514.072982] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=65726) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:937}} [ 1514.073932] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa8cc24f-543e-4e80-b323-92162affcab9 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1514.082807] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a34becda-5484-4a42-be85-bb182670e319 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1514.096634] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-160de574-b28d-4632-a250-6dd037007023 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1514.103609] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d459a124-11bf-4a50-a1bf-e39cb9e4702f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1514.131915] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=179510MB free_disk=95GB free_vcpus=48 pci_devices=None {{(pid=65726) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1136}} [ 1514.132075] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1514.132265] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1514.218658] env[65726]: INFO nova.compute.manager [None req-96e312fd-0b71-43ed-baa9-b065b31062a1 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 08377673-6f69-4a5c-8135-7ef1683bbb55] Swapping old allocation on dict_keys(['07c4692f-bdb4-4058-9173-ff9664830295']) held by migration 4a3f7777-8d69-4a3c-a520-09cf29b571f3 for instance [ 1514.243183] env[65726]: DEBUG nova.scheduler.client.report [None req-96e312fd-0b71-43ed-baa9-b065b31062a1 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Overwriting current allocation {'allocations': {'07c4692f-bdb4-4058-9173-ff9664830295': {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}, 'generation': 186}}, 'project_id': 'acd800e0438940bc80ae3c6b672db4ee', 'user_id': '5143cc185c7645b48304c8ebc5c854de', 'consumer_generation': 1} on consumer 08377673-6f69-4a5c-8135-7ef1683bbb55 {{(pid=65726) move_allocations /opt/stack/nova/nova/scheduler/client/report.py:2036}} [ 1514.292523] env[65726]: WARNING openstack [None req-96e312fd-0b71-43ed-baa9-b065b31062a1 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1514.292952] env[65726]: WARNING openstack [None req-96e312fd-0b71-43ed-baa9-b065b31062a1 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1514.329629] env[65726]: DEBUG oslo_concurrency.lockutils [None req-96e312fd-0b71-43ed-baa9-b065b31062a1 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Acquiring lock "refresh_cache-08377673-6f69-4a5c-8135-7ef1683bbb55" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1514.329821] env[65726]: DEBUG oslo_concurrency.lockutils [None req-96e312fd-0b71-43ed-baa9-b065b31062a1 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Acquired lock "refresh_cache-08377673-6f69-4a5c-8135-7ef1683bbb55" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1514.330019] env[65726]: DEBUG nova.network.neutron [None req-96e312fd-0b71-43ed-baa9-b065b31062a1 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 08377673-6f69-4a5c-8135-7ef1683bbb55] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1514.833642] env[65726]: WARNING openstack [None req-96e312fd-0b71-43ed-baa9-b065b31062a1 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1514.833800] env[65726]: WARNING openstack [None req-96e312fd-0b71-43ed-baa9-b065b31062a1 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1514.882929] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b2286a8b-99dc-410d-b7aa-ad9c39bb9d31 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Acquiring lock "55ee97c3-b014-48eb-b41b-fc5aa16bc09e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1514.883188] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b2286a8b-99dc-410d-b7aa-ad9c39bb9d31 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Lock "55ee97c3-b014-48eb-b41b-fc5aa16bc09e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1514.883386] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b2286a8b-99dc-410d-b7aa-ad9c39bb9d31 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Acquiring lock "55ee97c3-b014-48eb-b41b-fc5aa16bc09e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1514.883580] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b2286a8b-99dc-410d-b7aa-ad9c39bb9d31 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Lock "55ee97c3-b014-48eb-b41b-fc5aa16bc09e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1514.883747] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b2286a8b-99dc-410d-b7aa-ad9c39bb9d31 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Lock "55ee97c3-b014-48eb-b41b-fc5aa16bc09e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1514.886650] env[65726]: INFO nova.compute.manager [None req-b2286a8b-99dc-410d-b7aa-ad9c39bb9d31 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 55ee97c3-b014-48eb-b41b-fc5aa16bc09e] Terminating instance [ 1514.964023] env[65726]: WARNING openstack [None req-96e312fd-0b71-43ed-baa9-b065b31062a1 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1514.964452] env[65726]: WARNING openstack [None req-96e312fd-0b71-43ed-baa9-b065b31062a1 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1515.024162] env[65726]: WARNING openstack [None req-96e312fd-0b71-43ed-baa9-b065b31062a1 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1515.024539] env[65726]: WARNING openstack [None req-96e312fd-0b71-43ed-baa9-b065b31062a1 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1515.103717] env[65726]: DEBUG nova.network.neutron [None req-96e312fd-0b71-43ed-baa9-b065b31062a1 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 08377673-6f69-4a5c-8135-7ef1683bbb55] Updating instance_info_cache with network_info: [{"id": "5aea6240-edaa-4fb1-b696-877bc1ba685a", "address": "fa:16:3e:70:2b:bb", "network": {"id": "a899453d-5f43-4650-9642-b33f54a4d2eb", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-614861139-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.231", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "acd800e0438940bc80ae3c6b672db4ee", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "205fb402-8eaf-4b61-8f57-8f216024179a", "external-id": "nsx-vlan-transportzone-78", "segmentation_id": 78, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5aea6240-ed", "ovs_interfaceid": "5aea6240-edaa-4fb1-b696-877bc1ba685a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1515.155269] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Instance 55ee97c3-b014-48eb-b41b-fc5aa16bc09e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65726) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1515.155424] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Instance 08377673-6f69-4a5c-8135-7ef1683bbb55 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65726) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1515.155603] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Total usable vcpus: 48, total allocated vcpus: 2 {{(pid=65726) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1159}} [ 1515.155752] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=896MB phys_disk=100GB used_disk=2GB total_vcpus=48 used_vcpus=2 pci_stats=[] stats={'failed_builds': '0', 'num_instances': '2', 'num_vm_active': '1', 'num_task_None': '1', 'num_os_type_None': '2', 'num_proj_6b56f93ba0e14521921484ac2c785c43': '1', 'io_workload': '0', 'num_vm_resized': '1', 'num_task_resize_reverting': '1', 'num_proj_acd800e0438940bc80ae3c6b672db4ee': '1'} {{(pid=65726) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1168}} [ 1515.190937] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a959238c-57d8-497d-ae46-47e7cdd5fd33 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1515.199235] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed6bd6f0-aff2-4cac-88e6-b8fa8969188e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1515.230338] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b58f90d7-75b8-499a-a6a1-5d2285dad404 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1515.238695] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4ace055-4954-46d0-9e73-4ad9b8a5d72a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1515.252952] env[65726]: DEBUG nova.compute.provider_tree [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1515.392116] env[65726]: DEBUG nova.compute.manager [None req-b2286a8b-99dc-410d-b7aa-ad9c39bb9d31 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 55ee97c3-b014-48eb-b41b-fc5aa16bc09e] Start destroying the instance on the hypervisor. {{(pid=65726) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1515.392293] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-b2286a8b-99dc-410d-b7aa-ad9c39bb9d31 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 55ee97c3-b014-48eb-b41b-fc5aa16bc09e] Destroying instance {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1515.393476] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d11b6677-1517-4e50-9172-48ddd19ec82f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1515.401954] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-b2286a8b-99dc-410d-b7aa-ad9c39bb9d31 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 55ee97c3-b014-48eb-b41b-fc5aa16bc09e] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1515.402219] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1f93f6d7-9f63-4a52-9ffe-fc6884d4230c {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1515.408811] env[65726]: DEBUG oslo_vmware.api [None req-b2286a8b-99dc-410d-b7aa-ad9c39bb9d31 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Waiting for the task: (returnval){ [ 1515.408811] env[65726]: value = "task-5117171" [ 1515.408811] env[65726]: _type = "Task" [ 1515.408811] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1515.416696] env[65726]: DEBUG oslo_vmware.api [None req-b2286a8b-99dc-410d-b7aa-ad9c39bb9d31 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': task-5117171, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1515.606776] env[65726]: DEBUG oslo_concurrency.lockutils [None req-96e312fd-0b71-43ed-baa9-b065b31062a1 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Releasing lock "refresh_cache-08377673-6f69-4a5c-8135-7ef1683bbb55" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1515.606936] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-96e312fd-0b71-43ed-baa9-b065b31062a1 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 08377673-6f69-4a5c-8135-7ef1683bbb55] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1515.607198] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c8a73018-9bf1-4627-8fa6-4efafb8de41b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1515.614546] env[65726]: DEBUG oslo_vmware.api [None req-96e312fd-0b71-43ed-baa9-b065b31062a1 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Waiting for the task: (returnval){ [ 1515.614546] env[65726]: value = "task-5117172" [ 1515.614546] env[65726]: _type = "Task" [ 1515.614546] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1515.623611] env[65726]: DEBUG oslo_vmware.api [None req-96e312fd-0b71-43ed-baa9-b065b31062a1 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': task-5117172, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1515.757107] env[65726]: DEBUG nova.scheduler.client.report [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1515.919293] env[65726]: DEBUG oslo_vmware.api [None req-b2286a8b-99dc-410d-b7aa-ad9c39bb9d31 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': task-5117171, 'name': PowerOffVM_Task, 'duration_secs': 0.218278} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1515.919577] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-b2286a8b-99dc-410d-b7aa-ad9c39bb9d31 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 55ee97c3-b014-48eb-b41b-fc5aa16bc09e] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1515.919741] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-b2286a8b-99dc-410d-b7aa-ad9c39bb9d31 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 55ee97c3-b014-48eb-b41b-fc5aa16bc09e] Unregistering the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1515.919994] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-012c8def-e011-464a-abc8-8e1768e70f86 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1515.982755] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-b2286a8b-99dc-410d-b7aa-ad9c39bb9d31 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 55ee97c3-b014-48eb-b41b-fc5aa16bc09e] Unregistered the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1515.982984] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-b2286a8b-99dc-410d-b7aa-ad9c39bb9d31 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 55ee97c3-b014-48eb-b41b-fc5aa16bc09e] Deleting contents of the VM from datastore datastore2 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1515.983188] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-b2286a8b-99dc-410d-b7aa-ad9c39bb9d31 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Deleting the datastore file [datastore2] 55ee97c3-b014-48eb-b41b-fc5aa16bc09e {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1515.983463] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9deb2d86-beac-4951-a14a-50a7a7554a93 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1515.990618] env[65726]: DEBUG oslo_vmware.api [None req-b2286a8b-99dc-410d-b7aa-ad9c39bb9d31 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Waiting for the task: (returnval){ [ 1515.990618] env[65726]: value = "task-5117174" [ 1515.990618] env[65726]: _type = "Task" [ 1515.990618] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1515.998980] env[65726]: DEBUG oslo_vmware.api [None req-b2286a8b-99dc-410d-b7aa-ad9c39bb9d31 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': task-5117174, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1516.125705] env[65726]: DEBUG oslo_vmware.api [None req-96e312fd-0b71-43ed-baa9-b065b31062a1 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': task-5117172, 'name': PowerOffVM_Task, 'duration_secs': 0.232902} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1516.126169] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-96e312fd-0b71-43ed-baa9-b065b31062a1 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 08377673-6f69-4a5c-8135-7ef1683bbb55] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1516.127030] env[65726]: DEBUG nova.virt.hardware [None req-96e312fd-0b71-43ed-baa9-b065b31062a1 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1516.127257] env[65726]: DEBUG nova.virt.hardware [None req-96e312fd-0b71-43ed-baa9-b065b31062a1 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1516.127412] env[65726]: DEBUG nova.virt.hardware [None req-96e312fd-0b71-43ed-baa9-b065b31062a1 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1516.127591] env[65726]: DEBUG nova.virt.hardware [None req-96e312fd-0b71-43ed-baa9-b065b31062a1 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1516.127743] env[65726]: DEBUG nova.virt.hardware [None req-96e312fd-0b71-43ed-baa9-b065b31062a1 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1516.127896] env[65726]: DEBUG nova.virt.hardware [None req-96e312fd-0b71-43ed-baa9-b065b31062a1 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1516.128107] env[65726]: DEBUG nova.virt.hardware [None req-96e312fd-0b71-43ed-baa9-b065b31062a1 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1516.128262] env[65726]: DEBUG nova.virt.hardware [None req-96e312fd-0b71-43ed-baa9-b065b31062a1 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1516.128425] env[65726]: DEBUG nova.virt.hardware [None req-96e312fd-0b71-43ed-baa9-b065b31062a1 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1516.128583] env[65726]: DEBUG nova.virt.hardware [None req-96e312fd-0b71-43ed-baa9-b065b31062a1 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1516.128749] env[65726]: DEBUG nova.virt.hardware [None req-96e312fd-0b71-43ed-baa9-b065b31062a1 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1516.134296] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0b3dae1d-629c-4461-8871-9900a3ea4a5f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1516.150537] env[65726]: DEBUG oslo_vmware.api [None req-96e312fd-0b71-43ed-baa9-b065b31062a1 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Waiting for the task: (returnval){ [ 1516.150537] env[65726]: value = "task-5117175" [ 1516.150537] env[65726]: _type = "Task" [ 1516.150537] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1516.162688] env[65726]: DEBUG oslo_vmware.api [None req-96e312fd-0b71-43ed-baa9-b065b31062a1 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': task-5117175, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1516.262864] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=65726) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1097}} [ 1516.263079] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.131s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1516.501646] env[65726]: DEBUG oslo_vmware.api [None req-b2286a8b-99dc-410d-b7aa-ad9c39bb9d31 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Task: {'id': task-5117174, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.139513} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1516.501886] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-b2286a8b-99dc-410d-b7aa-ad9c39bb9d31 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Deleted the datastore file {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1516.502085] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-b2286a8b-99dc-410d-b7aa-ad9c39bb9d31 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 55ee97c3-b014-48eb-b41b-fc5aa16bc09e] Deleted contents of the VM from datastore datastore2 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1516.502264] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-b2286a8b-99dc-410d-b7aa-ad9c39bb9d31 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 55ee97c3-b014-48eb-b41b-fc5aa16bc09e] Instance destroyed {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1516.502435] env[65726]: INFO nova.compute.manager [None req-b2286a8b-99dc-410d-b7aa-ad9c39bb9d31 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] [instance: 55ee97c3-b014-48eb-b41b-fc5aa16bc09e] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1516.502676] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-b2286a8b-99dc-410d-b7aa-ad9c39bb9d31 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1516.502872] env[65726]: DEBUG nova.compute.manager [-] [instance: 55ee97c3-b014-48eb-b41b-fc5aa16bc09e] Deallocating network for instance {{(pid=65726) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1516.502970] env[65726]: DEBUG nova.network.neutron [-] [instance: 55ee97c3-b014-48eb-b41b-fc5aa16bc09e] deallocate_for_instance() {{(pid=65726) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 1516.503497] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1516.503753] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1516.552888] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1516.553190] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1516.661550] env[65726]: DEBUG oslo_vmware.api [None req-96e312fd-0b71-43ed-baa9-b065b31062a1 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': task-5117175, 'name': ReconfigVM_Task, 'duration_secs': 0.146398} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1516.662393] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0998c09a-fdf0-49e4-b1c4-c38f0310d23f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1516.681967] env[65726]: DEBUG nova.virt.hardware [None req-96e312fd-0b71-43ed-baa9-b065b31062a1 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1516.682243] env[65726]: DEBUG nova.virt.hardware [None req-96e312fd-0b71-43ed-baa9-b065b31062a1 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1516.682395] env[65726]: DEBUG nova.virt.hardware [None req-96e312fd-0b71-43ed-baa9-b065b31062a1 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1516.682579] env[65726]: DEBUG nova.virt.hardware [None req-96e312fd-0b71-43ed-baa9-b065b31062a1 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1516.682716] env[65726]: DEBUG nova.virt.hardware [None req-96e312fd-0b71-43ed-baa9-b065b31062a1 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1516.682866] env[65726]: DEBUG nova.virt.hardware [None req-96e312fd-0b71-43ed-baa9-b065b31062a1 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1516.683200] env[65726]: DEBUG nova.virt.hardware [None req-96e312fd-0b71-43ed-baa9-b065b31062a1 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1516.683379] env[65726]: DEBUG nova.virt.hardware [None req-96e312fd-0b71-43ed-baa9-b065b31062a1 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1516.683571] env[65726]: DEBUG nova.virt.hardware [None req-96e312fd-0b71-43ed-baa9-b065b31062a1 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1516.683972] env[65726]: DEBUG nova.virt.hardware [None req-96e312fd-0b71-43ed-baa9-b065b31062a1 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1516.683972] env[65726]: DEBUG nova.virt.hardware [None req-96e312fd-0b71-43ed-baa9-b065b31062a1 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1516.684885] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d4f2d9d0-1075-4107-9b59-8a512df12a8f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1516.691315] env[65726]: DEBUG oslo_vmware.api [None req-96e312fd-0b71-43ed-baa9-b065b31062a1 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Waiting for the task: (returnval){ [ 1516.691315] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]520922b6-d09f-f7a7-71f8-330e933d1d15" [ 1516.691315] env[65726]: _type = "Task" [ 1516.691315] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1516.701273] env[65726]: DEBUG oslo_vmware.api [None req-96e312fd-0b71-43ed-baa9-b065b31062a1 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]520922b6-d09f-f7a7-71f8-330e933d1d15, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1516.811975] env[65726]: DEBUG nova.compute.manager [req-3d3b3965-3fef-4c20-9f96-1afd9e5be0f7 req-020cbb5e-c0ef-49a7-9134-4bc0263ea682 service nova] [instance: 55ee97c3-b014-48eb-b41b-fc5aa16bc09e] Received event network-vif-deleted-65a8678e-a808-4eb1-9cb2-2f86b5fbd4b9 {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1516.812234] env[65726]: INFO nova.compute.manager [req-3d3b3965-3fef-4c20-9f96-1afd9e5be0f7 req-020cbb5e-c0ef-49a7-9134-4bc0263ea682 service nova] [instance: 55ee97c3-b014-48eb-b41b-fc5aa16bc09e] Neutron deleted interface 65a8678e-a808-4eb1-9cb2-2f86b5fbd4b9; detaching it from the instance and deleting it from the info cache [ 1516.812413] env[65726]: DEBUG nova.network.neutron [req-3d3b3965-3fef-4c20-9f96-1afd9e5be0f7 req-020cbb5e-c0ef-49a7-9134-4bc0263ea682 service nova] [instance: 55ee97c3-b014-48eb-b41b-fc5aa16bc09e] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1517.202450] env[65726]: DEBUG oslo_vmware.api [None req-96e312fd-0b71-43ed-baa9-b065b31062a1 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]520922b6-d09f-f7a7-71f8-330e933d1d15, 'name': SearchDatastore_Task, 'duration_secs': 0.009441} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1517.208132] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-96e312fd-0b71-43ed-baa9-b065b31062a1 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 08377673-6f69-4a5c-8135-7ef1683bbb55] Reconfiguring VM instance instance-0000007d to detach disk 2000 {{(pid=65726) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1517.208463] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cf9ca57d-bd0a-4832-b986-515214fd00a7 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1517.229975] env[65726]: DEBUG oslo_vmware.api [None req-96e312fd-0b71-43ed-baa9-b065b31062a1 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Waiting for the task: (returnval){ [ 1517.229975] env[65726]: value = "task-5117176" [ 1517.229975] env[65726]: _type = "Task" [ 1517.229975] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1517.238179] env[65726]: DEBUG oslo_vmware.api [None req-96e312fd-0b71-43ed-baa9-b065b31062a1 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': task-5117176, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1517.291819] env[65726]: DEBUG nova.network.neutron [-] [instance: 55ee97c3-b014-48eb-b41b-fc5aa16bc09e] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1517.315274] env[65726]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ef71b56f-af55-4b37-8148-8a7892fdef18 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1517.328541] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-403d6be4-5ac3-4075-903e-e68c939147d7 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1517.357812] env[65726]: DEBUG nova.compute.manager [req-3d3b3965-3fef-4c20-9f96-1afd9e5be0f7 req-020cbb5e-c0ef-49a7-9134-4bc0263ea682 service nova] [instance: 55ee97c3-b014-48eb-b41b-fc5aa16bc09e] Detach interface failed, port_id=65a8678e-a808-4eb1-9cb2-2f86b5fbd4b9, reason: Instance 55ee97c3-b014-48eb-b41b-fc5aa16bc09e could not be found. {{(pid=65726) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11669}} [ 1517.740165] env[65726]: DEBUG oslo_vmware.api [None req-96e312fd-0b71-43ed-baa9-b065b31062a1 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': task-5117176, 'name': ReconfigVM_Task, 'duration_secs': 0.190264} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1517.740474] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-96e312fd-0b71-43ed-baa9-b065b31062a1 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 08377673-6f69-4a5c-8135-7ef1683bbb55] Reconfigured VM instance instance-0000007d to detach disk 2000 {{(pid=65726) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1517.741297] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c7ec457-aa8d-44dc-96f7-399109dc13a5 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1517.763351] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-96e312fd-0b71-43ed-baa9-b065b31062a1 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 08377673-6f69-4a5c-8135-7ef1683bbb55] Reconfiguring VM instance instance-0000007d to attach disk [datastore2] 08377673-6f69-4a5c-8135-7ef1683bbb55/08377673-6f69-4a5c-8135-7ef1683bbb55.vmdk or device None with type thin {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1517.763632] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-90139eb8-778f-4531-ac89-4c60751b44c9 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1517.782200] env[65726]: DEBUG oslo_vmware.api [None req-96e312fd-0b71-43ed-baa9-b065b31062a1 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Waiting for the task: (returnval){ [ 1517.782200] env[65726]: value = "task-5117177" [ 1517.782200] env[65726]: _type = "Task" [ 1517.782200] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1517.791531] env[65726]: DEBUG oslo_vmware.api [None req-96e312fd-0b71-43ed-baa9-b065b31062a1 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': task-5117177, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1517.794058] env[65726]: INFO nova.compute.manager [-] [instance: 55ee97c3-b014-48eb-b41b-fc5aa16bc09e] Took 1.29 seconds to deallocate network for instance. [ 1518.292359] env[65726]: DEBUG oslo_vmware.api [None req-96e312fd-0b71-43ed-baa9-b065b31062a1 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': task-5117177, 'name': ReconfigVM_Task, 'duration_secs': 0.259003} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1518.292693] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-96e312fd-0b71-43ed-baa9-b065b31062a1 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 08377673-6f69-4a5c-8135-7ef1683bbb55] Reconfigured VM instance instance-0000007d to attach disk [datastore2] 08377673-6f69-4a5c-8135-7ef1683bbb55/08377673-6f69-4a5c-8135-7ef1683bbb55.vmdk or device None with type thin {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1518.293534] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2efbb904-43e2-4bbb-a215-e690519b8266 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1518.312821] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b2286a8b-99dc-410d-b7aa-ad9c39bb9d31 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1518.313097] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b2286a8b-99dc-410d-b7aa-ad9c39bb9d31 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1518.313315] env[65726]: DEBUG nova.objects.instance [None req-b2286a8b-99dc-410d-b7aa-ad9c39bb9d31 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Lazy-loading 'resources' on Instance uuid 55ee97c3-b014-48eb-b41b-fc5aa16bc09e {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1518.315085] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d91451d2-095a-43e5-925c-8158fb416868 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1518.334417] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfe7450a-11eb-47bf-8764-4c06ba4c1f2b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1518.353611] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9fe7860-48c3-4b2e-a346-25832bfad2fd {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1518.361401] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-96e312fd-0b71-43ed-baa9-b065b31062a1 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 08377673-6f69-4a5c-8135-7ef1683bbb55] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1518.361733] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8d71fed1-1d8e-4706-9441-eb00f75b5122 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1518.368010] env[65726]: DEBUG oslo_vmware.api [None req-96e312fd-0b71-43ed-baa9-b065b31062a1 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Waiting for the task: (returnval){ [ 1518.368010] env[65726]: value = "task-5117178" [ 1518.368010] env[65726]: _type = "Task" [ 1518.368010] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1518.377597] env[65726]: DEBUG oslo_vmware.api [None req-96e312fd-0b71-43ed-baa9-b065b31062a1 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': task-5117178, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1518.861828] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27147006-8339-4eca-aaaf-c5d80e6b487a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1518.873275] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b66de267-f32c-4466-8650-040a2abf5507 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1518.882768] env[65726]: DEBUG oslo_vmware.api [None req-96e312fd-0b71-43ed-baa9-b065b31062a1 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': task-5117178, 'name': PowerOnVM_Task, 'duration_secs': 0.354673} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1518.906446] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-96e312fd-0b71-43ed-baa9-b065b31062a1 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 08377673-6f69-4a5c-8135-7ef1683bbb55] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1518.910770] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d07cd61f-f22f-4017-affd-5d2cc8f7368b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1518.918726] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-951de5f2-8321-4db6-9571-cd2fe4aafd18 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1518.933085] env[65726]: DEBUG nova.compute.provider_tree [None req-b2286a8b-99dc-410d-b7aa-ad9c39bb9d31 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1519.437055] env[65726]: DEBUG nova.scheduler.client.report [None req-b2286a8b-99dc-410d-b7aa-ad9c39bb9d31 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1519.920801] env[65726]: INFO nova.compute.manager [None req-96e312fd-0b71-43ed-baa9-b065b31062a1 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 08377673-6f69-4a5c-8135-7ef1683bbb55] Updating instance to original state: 'active' [ 1519.941945] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b2286a8b-99dc-410d-b7aa-ad9c39bb9d31 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.629s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1519.964121] env[65726]: INFO nova.scheduler.client.report [None req-b2286a8b-99dc-410d-b7aa-ad9c39bb9d31 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Deleted allocations for instance 55ee97c3-b014-48eb-b41b-fc5aa16bc09e [ 1520.474865] env[65726]: DEBUG oslo_concurrency.lockutils [None req-b2286a8b-99dc-410d-b7aa-ad9c39bb9d31 tempest-ServerActionsTestOtherA-1450592917 tempest-ServerActionsTestOtherA-1450592917-project-member] Lock "55ee97c3-b014-48eb-b41b-fc5aa16bc09e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.592s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1521.258543] env[65726]: DEBUG oslo_service.periodic_task [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=65726) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1521.356618] env[65726]: DEBUG oslo_concurrency.lockutils [None req-bcf3f172-6658-4a50-b1d3-bb672847eab6 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Acquiring lock "08377673-6f69-4a5c-8135-7ef1683bbb55" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1521.356928] env[65726]: DEBUG oslo_concurrency.lockutils [None req-bcf3f172-6658-4a50-b1d3-bb672847eab6 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Lock "08377673-6f69-4a5c-8135-7ef1683bbb55" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1521.357156] env[65726]: DEBUG oslo_concurrency.lockutils [None req-bcf3f172-6658-4a50-b1d3-bb672847eab6 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Acquiring lock "08377673-6f69-4a5c-8135-7ef1683bbb55-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1521.357337] env[65726]: DEBUG oslo_concurrency.lockutils [None req-bcf3f172-6658-4a50-b1d3-bb672847eab6 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Lock "08377673-6f69-4a5c-8135-7ef1683bbb55-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1521.357496] env[65726]: DEBUG oslo_concurrency.lockutils [None req-bcf3f172-6658-4a50-b1d3-bb672847eab6 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Lock "08377673-6f69-4a5c-8135-7ef1683bbb55-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1521.359608] env[65726]: INFO nova.compute.manager [None req-bcf3f172-6658-4a50-b1d3-bb672847eab6 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 08377673-6f69-4a5c-8135-7ef1683bbb55] Terminating instance [ 1521.863061] env[65726]: DEBUG nova.compute.manager [None req-bcf3f172-6658-4a50-b1d3-bb672847eab6 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 08377673-6f69-4a5c-8135-7ef1683bbb55] Start destroying the instance on the hypervisor. {{(pid=65726) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1521.863401] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-bcf3f172-6658-4a50-b1d3-bb672847eab6 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 08377673-6f69-4a5c-8135-7ef1683bbb55] Destroying instance {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1521.864196] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-638401da-b964-42d7-9e86-2c83580f944a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1521.871966] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-bcf3f172-6658-4a50-b1d3-bb672847eab6 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 08377673-6f69-4a5c-8135-7ef1683bbb55] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1521.872221] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-eb24d037-ecb3-4801-8a2d-293cd65307ac {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1521.879251] env[65726]: DEBUG oslo_vmware.api [None req-bcf3f172-6658-4a50-b1d3-bb672847eab6 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Waiting for the task: (returnval){ [ 1521.879251] env[65726]: value = "task-5117179" [ 1521.879251] env[65726]: _type = "Task" [ 1521.879251] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1521.888567] env[65726]: DEBUG oslo_vmware.api [None req-bcf3f172-6658-4a50-b1d3-bb672847eab6 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': task-5117179, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1522.389398] env[65726]: DEBUG oslo_vmware.api [None req-bcf3f172-6658-4a50-b1d3-bb672847eab6 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': task-5117179, 'name': PowerOffVM_Task, 'duration_secs': 0.212473} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1522.389699] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-bcf3f172-6658-4a50-b1d3-bb672847eab6 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 08377673-6f69-4a5c-8135-7ef1683bbb55] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1522.389861] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-bcf3f172-6658-4a50-b1d3-bb672847eab6 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 08377673-6f69-4a5c-8135-7ef1683bbb55] Unregistering the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1522.390218] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-77c97404-d961-4902-891e-6a15c63b8164 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1522.466313] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-bcf3f172-6658-4a50-b1d3-bb672847eab6 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 08377673-6f69-4a5c-8135-7ef1683bbb55] Unregistered the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1522.466564] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-bcf3f172-6658-4a50-b1d3-bb672847eab6 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 08377673-6f69-4a5c-8135-7ef1683bbb55] Deleting contents of the VM from datastore datastore2 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1522.466718] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-bcf3f172-6658-4a50-b1d3-bb672847eab6 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Deleting the datastore file [datastore2] 08377673-6f69-4a5c-8135-7ef1683bbb55 {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1522.467089] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-16c00b50-3e94-4cf1-89ca-b34991946bf0 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1522.474018] env[65726]: DEBUG oslo_vmware.api [None req-bcf3f172-6658-4a50-b1d3-bb672847eab6 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Waiting for the task: (returnval){ [ 1522.474018] env[65726]: value = "task-5117181" [ 1522.474018] env[65726]: _type = "Task" [ 1522.474018] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1522.485132] env[65726]: DEBUG oslo_vmware.api [None req-bcf3f172-6658-4a50-b1d3-bb672847eab6 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': task-5117181, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1522.984487] env[65726]: DEBUG oslo_vmware.api [None req-bcf3f172-6658-4a50-b1d3-bb672847eab6 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': task-5117181, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.170717} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1522.984809] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-bcf3f172-6658-4a50-b1d3-bb672847eab6 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Deleted the datastore file {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1522.985034] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-bcf3f172-6658-4a50-b1d3-bb672847eab6 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 08377673-6f69-4a5c-8135-7ef1683bbb55] Deleted contents of the VM from datastore datastore2 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1522.985259] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-bcf3f172-6658-4a50-b1d3-bb672847eab6 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 08377673-6f69-4a5c-8135-7ef1683bbb55] Instance destroyed {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1522.985431] env[65726]: INFO nova.compute.manager [None req-bcf3f172-6658-4a50-b1d3-bb672847eab6 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 08377673-6f69-4a5c-8135-7ef1683bbb55] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1522.985680] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-bcf3f172-6658-4a50-b1d3-bb672847eab6 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1522.985890] env[65726]: DEBUG nova.compute.manager [-] [instance: 08377673-6f69-4a5c-8135-7ef1683bbb55] Deallocating network for instance {{(pid=65726) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1522.986011] env[65726]: DEBUG nova.network.neutron [-] [instance: 08377673-6f69-4a5c-8135-7ef1683bbb55] deallocate_for_instance() {{(pid=65726) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 1522.986584] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1522.986851] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1523.030268] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1523.030619] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1523.559041] env[65726]: DEBUG nova.compute.manager [req-2cc5ea95-54f2-4d31-a8a9-f9818f46c4bb req-ec37c74b-45dc-4c82-b8e7-781ba8582fc4 service nova] [instance: 08377673-6f69-4a5c-8135-7ef1683bbb55] Received event network-vif-deleted-5aea6240-edaa-4fb1-b696-877bc1ba685a {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1523.559285] env[65726]: INFO nova.compute.manager [req-2cc5ea95-54f2-4d31-a8a9-f9818f46c4bb req-ec37c74b-45dc-4c82-b8e7-781ba8582fc4 service nova] [instance: 08377673-6f69-4a5c-8135-7ef1683bbb55] Neutron deleted interface 5aea6240-edaa-4fb1-b696-877bc1ba685a; detaching it from the instance and deleting it from the info cache [ 1523.559474] env[65726]: DEBUG nova.network.neutron [req-2cc5ea95-54f2-4d31-a8a9-f9818f46c4bb req-ec37c74b-45dc-4c82-b8e7-781ba8582fc4 service nova] [instance: 08377673-6f69-4a5c-8135-7ef1683bbb55] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1523.912288] env[65726]: DEBUG nova.network.neutron [-] [instance: 08377673-6f69-4a5c-8135-7ef1683bbb55] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1524.062280] env[65726]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d3158370-c94a-45bd-b2f2-c517c8e759c9 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1524.075265] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-133b330f-f9e6-4dfe-9b2d-8acf3446a8be {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1524.101685] env[65726]: DEBUG nova.compute.manager [req-2cc5ea95-54f2-4d31-a8a9-f9818f46c4bb req-ec37c74b-45dc-4c82-b8e7-781ba8582fc4 service nova] [instance: 08377673-6f69-4a5c-8135-7ef1683bbb55] Detach interface failed, port_id=5aea6240-edaa-4fb1-b696-877bc1ba685a, reason: Instance 08377673-6f69-4a5c-8135-7ef1683bbb55 could not be found. {{(pid=65726) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11669}} [ 1524.415443] env[65726]: INFO nova.compute.manager [-] [instance: 08377673-6f69-4a5c-8135-7ef1683bbb55] Took 1.43 seconds to deallocate network for instance. [ 1524.922111] env[65726]: DEBUG oslo_concurrency.lockutils [None req-bcf3f172-6658-4a50-b1d3-bb672847eab6 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1524.922330] env[65726]: DEBUG oslo_concurrency.lockutils [None req-bcf3f172-6658-4a50-b1d3-bb672847eab6 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1524.922553] env[65726]: DEBUG nova.objects.instance [None req-bcf3f172-6658-4a50-b1d3-bb672847eab6 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Lazy-loading 'resources' on Instance uuid 08377673-6f69-4a5c-8135-7ef1683bbb55 {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1525.462405] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ddf8c3c0-7330-49ae-8aa8-a36621f2369a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1525.470464] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32479325-36cb-4101-b5c5-78f01ef6f471 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1525.500052] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-835a0759-877e-4709-a243-f0add97a09e8 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1525.508072] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62431053-199d-41b3-9634-eed6f11fc243 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1525.522890] env[65726]: DEBUG nova.compute.provider_tree [None req-bcf3f172-6658-4a50-b1d3-bb672847eab6 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1526.026684] env[65726]: DEBUG nova.scheduler.client.report [None req-bcf3f172-6658-4a50-b1d3-bb672847eab6 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1526.531877] env[65726]: DEBUG oslo_concurrency.lockutils [None req-bcf3f172-6658-4a50-b1d3-bb672847eab6 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.609s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1526.551055] env[65726]: INFO nova.scheduler.client.report [None req-bcf3f172-6658-4a50-b1d3-bb672847eab6 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Deleted allocations for instance 08377673-6f69-4a5c-8135-7ef1683bbb55 [ 1527.059967] env[65726]: DEBUG oslo_concurrency.lockutils [None req-bcf3f172-6658-4a50-b1d3-bb672847eab6 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Lock "08377673-6f69-4a5c-8135-7ef1683bbb55" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.702s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1528.531340] env[65726]: DEBUG oslo_concurrency.lockutils [None req-7e3d3a43-2aff-493b-b2d8-31509401c7fd tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Acquiring lock "012a1edf-97c5-4ee8-a36c-222ca14eb1f6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1528.531707] env[65726]: DEBUG oslo_concurrency.lockutils [None req-7e3d3a43-2aff-493b-b2d8-31509401c7fd tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Lock "012a1edf-97c5-4ee8-a36c-222ca14eb1f6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1529.033895] env[65726]: DEBUG nova.compute.manager [None req-7e3d3a43-2aff-493b-b2d8-31509401c7fd tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 012a1edf-97c5-4ee8-a36c-222ca14eb1f6] Starting instance... {{(pid=65726) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 1529.554918] env[65726]: DEBUG oslo_concurrency.lockutils [None req-7e3d3a43-2aff-493b-b2d8-31509401c7fd tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1529.555244] env[65726]: DEBUG oslo_concurrency.lockutils [None req-7e3d3a43-2aff-493b-b2d8-31509401c7fd tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1529.556784] env[65726]: INFO nova.compute.claims [None req-7e3d3a43-2aff-493b-b2d8-31509401c7fd tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 012a1edf-97c5-4ee8-a36c-222ca14eb1f6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1530.591381] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c0aa3a9-e55e-4581-bda9-d402a2ebbd04 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1530.599698] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1939f7c6-37bc-41c7-9afc-387ab4fc6930 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1530.630224] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1066f6b6-fea5-4704-b195-c2e77a749e15 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1530.637970] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b78084e-d1d5-4ea4-8082-c0e995d3c5c5 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1530.651179] env[65726]: DEBUG nova.compute.provider_tree [None req-7e3d3a43-2aff-493b-b2d8-31509401c7fd tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1531.154038] env[65726]: DEBUG nova.scheduler.client.report [None req-7e3d3a43-2aff-493b-b2d8-31509401c7fd tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1531.658820] env[65726]: DEBUG oslo_concurrency.lockutils [None req-7e3d3a43-2aff-493b-b2d8-31509401c7fd tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.103s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1531.659378] env[65726]: DEBUG nova.compute.manager [None req-7e3d3a43-2aff-493b-b2d8-31509401c7fd tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 012a1edf-97c5-4ee8-a36c-222ca14eb1f6] Start building networks asynchronously for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1532.164676] env[65726]: DEBUG nova.compute.utils [None req-7e3d3a43-2aff-493b-b2d8-31509401c7fd tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Using /dev/sd instead of None {{(pid=65726) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1532.166225] env[65726]: DEBUG nova.compute.manager [None req-7e3d3a43-2aff-493b-b2d8-31509401c7fd tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 012a1edf-97c5-4ee8-a36c-222ca14eb1f6] Allocating IP information in the background. {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1532.166403] env[65726]: DEBUG nova.network.neutron [None req-7e3d3a43-2aff-493b-b2d8-31509401c7fd tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 012a1edf-97c5-4ee8-a36c-222ca14eb1f6] allocate_for_instance() {{(pid=65726) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 1532.166729] env[65726]: WARNING neutronclient.v2_0.client [None req-7e3d3a43-2aff-493b-b2d8-31509401c7fd tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1532.167486] env[65726]: WARNING neutronclient.v2_0.client [None req-7e3d3a43-2aff-493b-b2d8-31509401c7fd tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1532.167704] env[65726]: WARNING openstack [None req-7e3d3a43-2aff-493b-b2d8-31509401c7fd tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1532.168074] env[65726]: WARNING openstack [None req-7e3d3a43-2aff-493b-b2d8-31509401c7fd tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1532.176059] env[65726]: DEBUG nova.compute.manager [None req-7e3d3a43-2aff-493b-b2d8-31509401c7fd tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 012a1edf-97c5-4ee8-a36c-222ca14eb1f6] Start building block device mappings for instance. {{(pid=65726) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1532.237319] env[65726]: DEBUG nova.policy [None req-7e3d3a43-2aff-493b-b2d8-31509401c7fd tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5143cc185c7645b48304c8ebc5c854de', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'acd800e0438940bc80ae3c6b672db4ee', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65726) authorize /opt/stack/nova/nova/policy.py:192}} [ 1532.479784] env[65726]: DEBUG nova.network.neutron [None req-7e3d3a43-2aff-493b-b2d8-31509401c7fd tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 012a1edf-97c5-4ee8-a36c-222ca14eb1f6] Successfully created port: 2b4783b7-452c-486b-9cad-1ceed6a0d8dc {{(pid=65726) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1533.186669] env[65726]: DEBUG nova.compute.manager [None req-7e3d3a43-2aff-493b-b2d8-31509401c7fd tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 012a1edf-97c5-4ee8-a36c-222ca14eb1f6] Start spawning the instance on the hypervisor. {{(pid=65726) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1533.213968] env[65726]: DEBUG nova.virt.hardware [None req-7e3d3a43-2aff-493b-b2d8-31509401c7fd tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:28:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T19:28:34Z,direct_url=,disk_format='vmdk',id=b52362a3-ee8a-4cbf-b06f-513b0cc8f95c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1539f06fa3534e90acbc3a60c4b8bd3f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T19:28:35Z,virtual_size=,visibility=), allow threads: False {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1533.214243] env[65726]: DEBUG nova.virt.hardware [None req-7e3d3a43-2aff-493b-b2d8-31509401c7fd tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Flavor limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1533.214394] env[65726]: DEBUG nova.virt.hardware [None req-7e3d3a43-2aff-493b-b2d8-31509401c7fd tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Image limits 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1533.214567] env[65726]: DEBUG nova.virt.hardware [None req-7e3d3a43-2aff-493b-b2d8-31509401c7fd tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Flavor pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1533.214720] env[65726]: DEBUG nova.virt.hardware [None req-7e3d3a43-2aff-493b-b2d8-31509401c7fd tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Image pref 0:0:0 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1533.214866] env[65726]: DEBUG nova.virt.hardware [None req-7e3d3a43-2aff-493b-b2d8-31509401c7fd tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65726) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1533.215089] env[65726]: DEBUG nova.virt.hardware [None req-7e3d3a43-2aff-493b-b2d8-31509401c7fd tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1533.215249] env[65726]: DEBUG nova.virt.hardware [None req-7e3d3a43-2aff-493b-b2d8-31509401c7fd tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1533.215413] env[65726]: DEBUG nova.virt.hardware [None req-7e3d3a43-2aff-493b-b2d8-31509401c7fd tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Got 1 possible topologies {{(pid=65726) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1533.215567] env[65726]: DEBUG nova.virt.hardware [None req-7e3d3a43-2aff-493b-b2d8-31509401c7fd tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1533.215731] env[65726]: DEBUG nova.virt.hardware [None req-7e3d3a43-2aff-493b-b2d8-31509401c7fd tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65726) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1533.216629] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-196122b1-e2cd-4af1-9856-14c9404c8226 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1533.225224] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4039b16c-11c9-4f19-9198-d68f1ae4b7a1 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1533.949499] env[65726]: DEBUG nova.compute.manager [req-15f61e9b-2059-44b3-b5a4-b1832fd04e18 req-aa3b91ce-d6fc-48a7-a3e8-57cb58bcfabf service nova] [instance: 012a1edf-97c5-4ee8-a36c-222ca14eb1f6] Received event network-vif-plugged-2b4783b7-452c-486b-9cad-1ceed6a0d8dc {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1533.949918] env[65726]: DEBUG oslo_concurrency.lockutils [req-15f61e9b-2059-44b3-b5a4-b1832fd04e18 req-aa3b91ce-d6fc-48a7-a3e8-57cb58bcfabf service nova] Acquiring lock "012a1edf-97c5-4ee8-a36c-222ca14eb1f6-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1533.950068] env[65726]: DEBUG oslo_concurrency.lockutils [req-15f61e9b-2059-44b3-b5a4-b1832fd04e18 req-aa3b91ce-d6fc-48a7-a3e8-57cb58bcfabf service nova] Lock "012a1edf-97c5-4ee8-a36c-222ca14eb1f6-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1533.950213] env[65726]: DEBUG oslo_concurrency.lockutils [req-15f61e9b-2059-44b3-b5a4-b1832fd04e18 req-aa3b91ce-d6fc-48a7-a3e8-57cb58bcfabf service nova] Lock "012a1edf-97c5-4ee8-a36c-222ca14eb1f6-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1533.950383] env[65726]: DEBUG nova.compute.manager [req-15f61e9b-2059-44b3-b5a4-b1832fd04e18 req-aa3b91ce-d6fc-48a7-a3e8-57cb58bcfabf service nova] [instance: 012a1edf-97c5-4ee8-a36c-222ca14eb1f6] No waiting events found dispatching network-vif-plugged-2b4783b7-452c-486b-9cad-1ceed6a0d8dc {{(pid=65726) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1533.950590] env[65726]: WARNING nova.compute.manager [req-15f61e9b-2059-44b3-b5a4-b1832fd04e18 req-aa3b91ce-d6fc-48a7-a3e8-57cb58bcfabf service nova] [instance: 012a1edf-97c5-4ee8-a36c-222ca14eb1f6] Received unexpected event network-vif-plugged-2b4783b7-452c-486b-9cad-1ceed6a0d8dc for instance with vm_state building and task_state spawning. [ 1533.990343] env[65726]: DEBUG nova.network.neutron [None req-7e3d3a43-2aff-493b-b2d8-31509401c7fd tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 012a1edf-97c5-4ee8-a36c-222ca14eb1f6] Successfully updated port: 2b4783b7-452c-486b-9cad-1ceed6a0d8dc {{(pid=65726) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 1534.492829] env[65726]: DEBUG oslo_concurrency.lockutils [None req-7e3d3a43-2aff-493b-b2d8-31509401c7fd tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Acquiring lock "refresh_cache-012a1edf-97c5-4ee8-a36c-222ca14eb1f6" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1534.493100] env[65726]: DEBUG oslo_concurrency.lockutils [None req-7e3d3a43-2aff-493b-b2d8-31509401c7fd tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Acquired lock "refresh_cache-012a1edf-97c5-4ee8-a36c-222ca14eb1f6" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1534.493297] env[65726]: DEBUG nova.network.neutron [None req-7e3d3a43-2aff-493b-b2d8-31509401c7fd tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 012a1edf-97c5-4ee8-a36c-222ca14eb1f6] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1534.996885] env[65726]: WARNING openstack [None req-7e3d3a43-2aff-493b-b2d8-31509401c7fd tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1534.997347] env[65726]: WARNING openstack [None req-7e3d3a43-2aff-493b-b2d8-31509401c7fd tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1535.002830] env[65726]: DEBUG nova.network.neutron [None req-7e3d3a43-2aff-493b-b2d8-31509401c7fd tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 012a1edf-97c5-4ee8-a36c-222ca14eb1f6] Instance cache missing network info. {{(pid=65726) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 1535.053322] env[65726]: WARNING openstack [None req-7e3d3a43-2aff-493b-b2d8-31509401c7fd tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1535.053703] env[65726]: WARNING openstack [None req-7e3d3a43-2aff-493b-b2d8-31509401c7fd tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1535.117389] env[65726]: WARNING openstack [None req-7e3d3a43-2aff-493b-b2d8-31509401c7fd tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1535.117815] env[65726]: WARNING openstack [None req-7e3d3a43-2aff-493b-b2d8-31509401c7fd tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1535.198674] env[65726]: DEBUG nova.network.neutron [None req-7e3d3a43-2aff-493b-b2d8-31509401c7fd tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 012a1edf-97c5-4ee8-a36c-222ca14eb1f6] Updating instance_info_cache with network_info: [{"id": "2b4783b7-452c-486b-9cad-1ceed6a0d8dc", "address": "fa:16:3e:22:a9:49", "network": {"id": "a899453d-5f43-4650-9642-b33f54a4d2eb", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-614861139-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "acd800e0438940bc80ae3c6b672db4ee", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "205fb402-8eaf-4b61-8f57-8f216024179a", "external-id": "nsx-vlan-transportzone-78", "segmentation_id": 78, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2b4783b7-45", "ovs_interfaceid": "2b4783b7-452c-486b-9cad-1ceed6a0d8dc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1535.701302] env[65726]: DEBUG oslo_concurrency.lockutils [None req-7e3d3a43-2aff-493b-b2d8-31509401c7fd tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Releasing lock "refresh_cache-012a1edf-97c5-4ee8-a36c-222ca14eb1f6" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1535.701676] env[65726]: DEBUG nova.compute.manager [None req-7e3d3a43-2aff-493b-b2d8-31509401c7fd tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 012a1edf-97c5-4ee8-a36c-222ca14eb1f6] Instance network_info: |[{"id": "2b4783b7-452c-486b-9cad-1ceed6a0d8dc", "address": "fa:16:3e:22:a9:49", "network": {"id": "a899453d-5f43-4650-9642-b33f54a4d2eb", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-614861139-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "acd800e0438940bc80ae3c6b672db4ee", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "205fb402-8eaf-4b61-8f57-8f216024179a", "external-id": "nsx-vlan-transportzone-78", "segmentation_id": 78, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2b4783b7-45", "ovs_interfaceid": "2b4783b7-452c-486b-9cad-1ceed6a0d8dc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65726) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1535.702106] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-7e3d3a43-2aff-493b-b2d8-31509401c7fd tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 012a1edf-97c5-4ee8-a36c-222ca14eb1f6] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:22:a9:49', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '205fb402-8eaf-4b61-8f57-8f216024179a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2b4783b7-452c-486b-9cad-1ceed6a0d8dc', 'vif_model': 'vmxnet3'}] {{(pid=65726) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1535.709483] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-7e3d3a43-2aff-493b-b2d8-31509401c7fd tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1535.709692] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 012a1edf-97c5-4ee8-a36c-222ca14eb1f6] Creating VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1535.709916] env[65726]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-cfdcb23c-8d76-46d7-b3bc-40aea48e1a64 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1535.730230] env[65726]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1535.730230] env[65726]: value = "task-5117182" [ 1535.730230] env[65726]: _type = "Task" [ 1535.730230] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1535.738055] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5117182, 'name': CreateVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1535.987732] env[65726]: DEBUG nova.compute.manager [req-0958ce90-9000-48f1-a117-299ce01abc6f req-f32af7d2-4609-4f0d-b4f9-16c2738a6aae service nova] [instance: 012a1edf-97c5-4ee8-a36c-222ca14eb1f6] Received event network-changed-2b4783b7-452c-486b-9cad-1ceed6a0d8dc {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1535.987967] env[65726]: DEBUG nova.compute.manager [req-0958ce90-9000-48f1-a117-299ce01abc6f req-f32af7d2-4609-4f0d-b4f9-16c2738a6aae service nova] [instance: 012a1edf-97c5-4ee8-a36c-222ca14eb1f6] Refreshing instance network info cache due to event network-changed-2b4783b7-452c-486b-9cad-1ceed6a0d8dc. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 1535.988243] env[65726]: DEBUG oslo_concurrency.lockutils [req-0958ce90-9000-48f1-a117-299ce01abc6f req-f32af7d2-4609-4f0d-b4f9-16c2738a6aae service nova] Acquiring lock "refresh_cache-012a1edf-97c5-4ee8-a36c-222ca14eb1f6" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1535.988419] env[65726]: DEBUG oslo_concurrency.lockutils [req-0958ce90-9000-48f1-a117-299ce01abc6f req-f32af7d2-4609-4f0d-b4f9-16c2738a6aae service nova] Acquired lock "refresh_cache-012a1edf-97c5-4ee8-a36c-222ca14eb1f6" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1535.988637] env[65726]: DEBUG nova.network.neutron [req-0958ce90-9000-48f1-a117-299ce01abc6f req-f32af7d2-4609-4f0d-b4f9-16c2738a6aae service nova] [instance: 012a1edf-97c5-4ee8-a36c-222ca14eb1f6] Refreshing network info cache for port 2b4783b7-452c-486b-9cad-1ceed6a0d8dc {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 1536.240410] env[65726]: DEBUG oslo_vmware.api [-] Task: {'id': task-5117182, 'name': CreateVM_Task, 'duration_secs': 0.295824} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1536.240691] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 012a1edf-97c5-4ee8-a36c-222ca14eb1f6] Created VM on the ESX host {{(pid=65726) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1536.241413] env[65726]: WARNING openstack [None req-7e3d3a43-2aff-493b-b2d8-31509401c7fd tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1536.241767] env[65726]: WARNING openstack [None req-7e3d3a43-2aff-493b-b2d8-31509401c7fd tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1536.246947] env[65726]: DEBUG oslo_concurrency.lockutils [None req-7e3d3a43-2aff-493b-b2d8-31509401c7fd tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1536.247156] env[65726]: DEBUG oslo_concurrency.lockutils [None req-7e3d3a43-2aff-493b-b2d8-31509401c7fd tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1536.247491] env[65726]: DEBUG oslo_concurrency.lockutils [None req-7e3d3a43-2aff-493b-b2d8-31509401c7fd tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1536.247783] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0ffb16e4-7c60-4a28-a706-fc396bc21790 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1536.254727] env[65726]: DEBUG oslo_vmware.api [None req-7e3d3a43-2aff-493b-b2d8-31509401c7fd tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Waiting for the task: (returnval){ [ 1536.254727] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52cf0e75-64d9-d536-6500-0c5c8d1e3555" [ 1536.254727] env[65726]: _type = "Task" [ 1536.254727] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1536.263642] env[65726]: DEBUG oslo_vmware.api [None req-7e3d3a43-2aff-493b-b2d8-31509401c7fd tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52cf0e75-64d9-d536-6500-0c5c8d1e3555, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1536.492071] env[65726]: WARNING openstack [req-0958ce90-9000-48f1-a117-299ce01abc6f req-f32af7d2-4609-4f0d-b4f9-16c2738a6aae service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1536.492426] env[65726]: WARNING openstack [req-0958ce90-9000-48f1-a117-299ce01abc6f req-f32af7d2-4609-4f0d-b4f9-16c2738a6aae service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1536.595318] env[65726]: WARNING openstack [req-0958ce90-9000-48f1-a117-299ce01abc6f req-f32af7d2-4609-4f0d-b4f9-16c2738a6aae service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1536.596059] env[65726]: WARNING openstack [req-0958ce90-9000-48f1-a117-299ce01abc6f req-f32af7d2-4609-4f0d-b4f9-16c2738a6aae service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1536.654698] env[65726]: WARNING openstack [req-0958ce90-9000-48f1-a117-299ce01abc6f req-f32af7d2-4609-4f0d-b4f9-16c2738a6aae service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1536.655137] env[65726]: WARNING openstack [req-0958ce90-9000-48f1-a117-299ce01abc6f req-f32af7d2-4609-4f0d-b4f9-16c2738a6aae service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1536.733207] env[65726]: DEBUG nova.network.neutron [req-0958ce90-9000-48f1-a117-299ce01abc6f req-f32af7d2-4609-4f0d-b4f9-16c2738a6aae service nova] [instance: 012a1edf-97c5-4ee8-a36c-222ca14eb1f6] Updated VIF entry in instance network info cache for port 2b4783b7-452c-486b-9cad-1ceed6a0d8dc. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 1536.733570] env[65726]: DEBUG nova.network.neutron [req-0958ce90-9000-48f1-a117-299ce01abc6f req-f32af7d2-4609-4f0d-b4f9-16c2738a6aae service nova] [instance: 012a1edf-97c5-4ee8-a36c-222ca14eb1f6] Updating instance_info_cache with network_info: [{"id": "2b4783b7-452c-486b-9cad-1ceed6a0d8dc", "address": "fa:16:3e:22:a9:49", "network": {"id": "a899453d-5f43-4650-9642-b33f54a4d2eb", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-614861139-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "acd800e0438940bc80ae3c6b672db4ee", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "205fb402-8eaf-4b61-8f57-8f216024179a", "external-id": "nsx-vlan-transportzone-78", "segmentation_id": 78, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2b4783b7-45", "ovs_interfaceid": "2b4783b7-452c-486b-9cad-1ceed6a0d8dc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1536.766428] env[65726]: DEBUG oslo_vmware.api [None req-7e3d3a43-2aff-493b-b2d8-31509401c7fd tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]52cf0e75-64d9-d536-6500-0c5c8d1e3555, 'name': SearchDatastore_Task, 'duration_secs': 0.013304} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1536.766675] env[65726]: DEBUG oslo_concurrency.lockutils [None req-7e3d3a43-2aff-493b-b2d8-31509401c7fd tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1536.766894] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-7e3d3a43-2aff-493b-b2d8-31509401c7fd tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 012a1edf-97c5-4ee8-a36c-222ca14eb1f6] Processing image b52362a3-ee8a-4cbf-b06f-513b0cc8f95c {{(pid=65726) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1536.767216] env[65726]: DEBUG oslo_concurrency.lockutils [None req-7e3d3a43-2aff-493b-b2d8-31509401c7fd tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1536.767312] env[65726]: DEBUG oslo_concurrency.lockutils [None req-7e3d3a43-2aff-493b-b2d8-31509401c7fd tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1536.767464] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-7e3d3a43-2aff-493b-b2d8-31509401c7fd tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1536.767730] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ce75fa22-c875-40f9-8a4f-ebaf6e471ea8 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1536.776932] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-7e3d3a43-2aff-493b-b2d8-31509401c7fd tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65726) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1536.777142] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-7e3d3a43-2aff-493b-b2d8-31509401c7fd tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65726) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1536.777894] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7437c18d-b2a4-45ec-b4f1-cddf7b2b9570 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1536.782951] env[65726]: DEBUG oslo_vmware.api [None req-7e3d3a43-2aff-493b-b2d8-31509401c7fd tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Waiting for the task: (returnval){ [ 1536.782951] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5245400a-937c-6541-296a-4a3c4f70fc98" [ 1536.782951] env[65726]: _type = "Task" [ 1536.782951] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1536.790629] env[65726]: DEBUG oslo_vmware.api [None req-7e3d3a43-2aff-493b-b2d8-31509401c7fd tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5245400a-937c-6541-296a-4a3c4f70fc98, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1537.236805] env[65726]: DEBUG oslo_concurrency.lockutils [req-0958ce90-9000-48f1-a117-299ce01abc6f req-f32af7d2-4609-4f0d-b4f9-16c2738a6aae service nova] Releasing lock "refresh_cache-012a1edf-97c5-4ee8-a36c-222ca14eb1f6" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1537.294016] env[65726]: DEBUG oslo_vmware.api [None req-7e3d3a43-2aff-493b-b2d8-31509401c7fd tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5245400a-937c-6541-296a-4a3c4f70fc98, 'name': SearchDatastore_Task, 'duration_secs': 0.00899} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1537.294815] env[65726]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b906dc32-ca67-4c6c-b642-5b975e1b41fc {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1537.300173] env[65726]: DEBUG oslo_vmware.api [None req-7e3d3a43-2aff-493b-b2d8-31509401c7fd tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Waiting for the task: (returnval){ [ 1537.300173] env[65726]: value = "session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5201cc82-2564-fbe2-e95d-ae6d9e9a07c0" [ 1537.300173] env[65726]: _type = "Task" [ 1537.300173] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1537.308141] env[65726]: DEBUG oslo_vmware.api [None req-7e3d3a43-2aff-493b-b2d8-31509401c7fd tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5201cc82-2564-fbe2-e95d-ae6d9e9a07c0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1537.812077] env[65726]: DEBUG oslo_vmware.api [None req-7e3d3a43-2aff-493b-b2d8-31509401c7fd tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': session[52f663e2-317e-51dc-dc14-cf3a97cfedb8]5201cc82-2564-fbe2-e95d-ae6d9e9a07c0, 'name': SearchDatastore_Task, 'duration_secs': 0.010148} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1537.812530] env[65726]: DEBUG oslo_concurrency.lockutils [None req-7e3d3a43-2aff-493b-b2d8-31509401c7fd tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1537.812614] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-7e3d3a43-2aff-493b-b2d8-31509401c7fd tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore2] 012a1edf-97c5-4ee8-a36c-222ca14eb1f6/012a1edf-97c5-4ee8-a36c-222ca14eb1f6.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1537.812848] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-739c7cbf-700b-4c03-a9c3-93e59d7f6d61 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1537.820596] env[65726]: DEBUG oslo_vmware.api [None req-7e3d3a43-2aff-493b-b2d8-31509401c7fd tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Waiting for the task: (returnval){ [ 1537.820596] env[65726]: value = "task-5117183" [ 1537.820596] env[65726]: _type = "Task" [ 1537.820596] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1537.829291] env[65726]: DEBUG oslo_vmware.api [None req-7e3d3a43-2aff-493b-b2d8-31509401c7fd tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': task-5117183, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1538.330981] env[65726]: DEBUG oslo_vmware.api [None req-7e3d3a43-2aff-493b-b2d8-31509401c7fd tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': task-5117183, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.474519} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1538.331288] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-7e3d3a43-2aff-493b-b2d8-31509401c7fd tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c/b52362a3-ee8a-4cbf-b06f-513b0cc8f95c.vmdk to [datastore2] 012a1edf-97c5-4ee8-a36c-222ca14eb1f6/012a1edf-97c5-4ee8-a36c-222ca14eb1f6.vmdk {{(pid=65726) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1538.331500] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-7e3d3a43-2aff-493b-b2d8-31509401c7fd tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 012a1edf-97c5-4ee8-a36c-222ca14eb1f6] Extending root virtual disk to 1048576 {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1538.331769] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ae919510-811f-49f1-8394-e8bd78f926b4 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1538.338852] env[65726]: DEBUG oslo_vmware.api [None req-7e3d3a43-2aff-493b-b2d8-31509401c7fd tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Waiting for the task: (returnval){ [ 1538.338852] env[65726]: value = "task-5117184" [ 1538.338852] env[65726]: _type = "Task" [ 1538.338852] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1538.346703] env[65726]: DEBUG oslo_vmware.api [None req-7e3d3a43-2aff-493b-b2d8-31509401c7fd tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': task-5117184, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1538.849566] env[65726]: DEBUG oslo_vmware.api [None req-7e3d3a43-2aff-493b-b2d8-31509401c7fd tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': task-5117184, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066644} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1538.849915] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-7e3d3a43-2aff-493b-b2d8-31509401c7fd tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 012a1edf-97c5-4ee8-a36c-222ca14eb1f6] Extended root virtual disk {{(pid=65726) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1538.850658] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c832cb3-e520-4f9f-af82-0cfcecddfbd3 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1538.874040] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-7e3d3a43-2aff-493b-b2d8-31509401c7fd tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 012a1edf-97c5-4ee8-a36c-222ca14eb1f6] Reconfiguring VM instance instance-0000007e to attach disk [datastore2] 012a1edf-97c5-4ee8-a36c-222ca14eb1f6/012a1edf-97c5-4ee8-a36c-222ca14eb1f6.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1538.874358] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-95ded794-7c97-4d51-bc72-bc10a621ad26 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1538.894462] env[65726]: DEBUG oslo_vmware.api [None req-7e3d3a43-2aff-493b-b2d8-31509401c7fd tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Waiting for the task: (returnval){ [ 1538.894462] env[65726]: value = "task-5117185" [ 1538.894462] env[65726]: _type = "Task" [ 1538.894462] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1538.903167] env[65726]: DEBUG oslo_vmware.api [None req-7e3d3a43-2aff-493b-b2d8-31509401c7fd tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': task-5117185, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1539.405482] env[65726]: DEBUG oslo_vmware.api [None req-7e3d3a43-2aff-493b-b2d8-31509401c7fd tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': task-5117185, 'name': ReconfigVM_Task, 'duration_secs': 0.274578} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1539.405687] env[65726]: DEBUG nova.virt.vmwareapi.volumeops [None req-7e3d3a43-2aff-493b-b2d8-31509401c7fd tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 012a1edf-97c5-4ee8-a36c-222ca14eb1f6] Reconfigured VM instance instance-0000007e to attach disk [datastore2] 012a1edf-97c5-4ee8-a36c-222ca14eb1f6/012a1edf-97c5-4ee8-a36c-222ca14eb1f6.vmdk or device None with type sparse {{(pid=65726) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1539.406327] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f111b5d2-b7e1-4270-8782-afc849683fbb {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1539.413084] env[65726]: DEBUG oslo_vmware.api [None req-7e3d3a43-2aff-493b-b2d8-31509401c7fd tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Waiting for the task: (returnval){ [ 1539.413084] env[65726]: value = "task-5117186" [ 1539.413084] env[65726]: _type = "Task" [ 1539.413084] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1539.423294] env[65726]: DEBUG oslo_vmware.api [None req-7e3d3a43-2aff-493b-b2d8-31509401c7fd tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': task-5117186, 'name': Rename_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1539.923092] env[65726]: DEBUG oslo_vmware.api [None req-7e3d3a43-2aff-493b-b2d8-31509401c7fd tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': task-5117186, 'name': Rename_Task, 'duration_secs': 0.148593} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1539.923441] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-7e3d3a43-2aff-493b-b2d8-31509401c7fd tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 012a1edf-97c5-4ee8-a36c-222ca14eb1f6] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1539.923638] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-99f7d35d-79c4-45e3-b24c-82e0e9939eff {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1539.931090] env[65726]: DEBUG oslo_vmware.api [None req-7e3d3a43-2aff-493b-b2d8-31509401c7fd tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Waiting for the task: (returnval){ [ 1539.931090] env[65726]: value = "task-5117187" [ 1539.931090] env[65726]: _type = "Task" [ 1539.931090] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1539.938775] env[65726]: DEBUG oslo_vmware.api [None req-7e3d3a43-2aff-493b-b2d8-31509401c7fd tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': task-5117187, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1540.442417] env[65726]: DEBUG oslo_vmware.api [None req-7e3d3a43-2aff-493b-b2d8-31509401c7fd tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': task-5117187, 'name': PowerOnVM_Task, 'duration_secs': 0.453428} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1540.442706] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-7e3d3a43-2aff-493b-b2d8-31509401c7fd tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 012a1edf-97c5-4ee8-a36c-222ca14eb1f6] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1540.442946] env[65726]: INFO nova.compute.manager [None req-7e3d3a43-2aff-493b-b2d8-31509401c7fd tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 012a1edf-97c5-4ee8-a36c-222ca14eb1f6] Took 7.26 seconds to spawn the instance on the hypervisor. [ 1540.443171] env[65726]: DEBUG nova.compute.manager [None req-7e3d3a43-2aff-493b-b2d8-31509401c7fd tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 012a1edf-97c5-4ee8-a36c-222ca14eb1f6] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1540.444056] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e23465b1-07e9-438f-9826-1ecc39e039cc {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1540.962764] env[65726]: INFO nova.compute.manager [None req-7e3d3a43-2aff-493b-b2d8-31509401c7fd tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 012a1edf-97c5-4ee8-a36c-222ca14eb1f6] Took 11.42 seconds to build instance. [ 1541.465394] env[65726]: DEBUG oslo_concurrency.lockutils [None req-7e3d3a43-2aff-493b-b2d8-31509401c7fd tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Lock "012a1edf-97c5-4ee8-a36c-222ca14eb1f6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 12.934s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1541.592275] env[65726]: DEBUG nova.compute.manager [req-e6ebe01f-06e3-4019-8966-b67549979ee7 req-f2737e9f-c098-4691-bfbf-19d363e841a1 service nova] [instance: 012a1edf-97c5-4ee8-a36c-222ca14eb1f6] Received event network-changed-2b4783b7-452c-486b-9cad-1ceed6a0d8dc {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1541.593164] env[65726]: DEBUG nova.compute.manager [req-e6ebe01f-06e3-4019-8966-b67549979ee7 req-f2737e9f-c098-4691-bfbf-19d363e841a1 service nova] [instance: 012a1edf-97c5-4ee8-a36c-222ca14eb1f6] Refreshing instance network info cache due to event network-changed-2b4783b7-452c-486b-9cad-1ceed6a0d8dc. {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 1541.593512] env[65726]: DEBUG oslo_concurrency.lockutils [req-e6ebe01f-06e3-4019-8966-b67549979ee7 req-f2737e9f-c098-4691-bfbf-19d363e841a1 service nova] Acquiring lock "refresh_cache-012a1edf-97c5-4ee8-a36c-222ca14eb1f6" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1541.593569] env[65726]: DEBUG oslo_concurrency.lockutils [req-e6ebe01f-06e3-4019-8966-b67549979ee7 req-f2737e9f-c098-4691-bfbf-19d363e841a1 service nova] Acquired lock "refresh_cache-012a1edf-97c5-4ee8-a36c-222ca14eb1f6" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1541.593729] env[65726]: DEBUG nova.network.neutron [req-e6ebe01f-06e3-4019-8966-b67549979ee7 req-f2737e9f-c098-4691-bfbf-19d363e841a1 service nova] [instance: 012a1edf-97c5-4ee8-a36c-222ca14eb1f6] Refreshing network info cache for port 2b4783b7-452c-486b-9cad-1ceed6a0d8dc {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 1542.097517] env[65726]: WARNING openstack [req-e6ebe01f-06e3-4019-8966-b67549979ee7 req-f2737e9f-c098-4691-bfbf-19d363e841a1 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1542.097828] env[65726]: WARNING openstack [req-e6ebe01f-06e3-4019-8966-b67549979ee7 req-f2737e9f-c098-4691-bfbf-19d363e841a1 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1542.213211] env[65726]: WARNING openstack [req-e6ebe01f-06e3-4019-8966-b67549979ee7 req-f2737e9f-c098-4691-bfbf-19d363e841a1 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1542.213640] env[65726]: WARNING openstack [req-e6ebe01f-06e3-4019-8966-b67549979ee7 req-f2737e9f-c098-4691-bfbf-19d363e841a1 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1542.286755] env[65726]: WARNING openstack [req-e6ebe01f-06e3-4019-8966-b67549979ee7 req-f2737e9f-c098-4691-bfbf-19d363e841a1 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1542.287215] env[65726]: WARNING openstack [req-e6ebe01f-06e3-4019-8966-b67549979ee7 req-f2737e9f-c098-4691-bfbf-19d363e841a1 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1542.367546] env[65726]: DEBUG nova.network.neutron [req-e6ebe01f-06e3-4019-8966-b67549979ee7 req-f2737e9f-c098-4691-bfbf-19d363e841a1 service nova] [instance: 012a1edf-97c5-4ee8-a36c-222ca14eb1f6] Updated VIF entry in instance network info cache for port 2b4783b7-452c-486b-9cad-1ceed6a0d8dc. {{(pid=65726) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 1542.367911] env[65726]: DEBUG nova.network.neutron [req-e6ebe01f-06e3-4019-8966-b67549979ee7 req-f2737e9f-c098-4691-bfbf-19d363e841a1 service nova] [instance: 012a1edf-97c5-4ee8-a36c-222ca14eb1f6] Updating instance_info_cache with network_info: [{"id": "2b4783b7-452c-486b-9cad-1ceed6a0d8dc", "address": "fa:16:3e:22:a9:49", "network": {"id": "a899453d-5f43-4650-9642-b33f54a4d2eb", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-614861139-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.231", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "acd800e0438940bc80ae3c6b672db4ee", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "205fb402-8eaf-4b61-8f57-8f216024179a", "external-id": "nsx-vlan-transportzone-78", "segmentation_id": 78, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2b4783b7-45", "ovs_interfaceid": "2b4783b7-452c-486b-9cad-1ceed6a0d8dc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1542.870835] env[65726]: DEBUG oslo_concurrency.lockutils [req-e6ebe01f-06e3-4019-8966-b67549979ee7 req-f2737e9f-c098-4691-bfbf-19d363e841a1 service nova] Releasing lock "refresh_cache-012a1edf-97c5-4ee8-a36c-222ca14eb1f6" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1565.569534] env[65726]: DEBUG oslo_service.periodic_task [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=65726) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1566.570070] env[65726]: DEBUG oslo_service.periodic_task [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=65726) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1566.570070] env[65726]: DEBUG oslo_service.periodic_task [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=65726) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1566.570070] env[65726]: DEBUG nova.compute.manager [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=65726) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11251}} [ 1569.564967] env[65726]: DEBUG oslo_service.periodic_task [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=65726) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1569.568583] env[65726]: DEBUG oslo_service.periodic_task [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=65726) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1570.569303] env[65726]: DEBUG oslo_service.periodic_task [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=65726) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1572.569544] env[65726]: DEBUG oslo_service.periodic_task [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=65726) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1573.569731] env[65726]: DEBUG oslo_service.periodic_task [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Running periodic task ComputeManager.update_available_resource {{(pid=65726) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1574.072978] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1574.073256] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1574.073426] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1574.073581] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=65726) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:937}} [ 1574.074506] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39709df4-7bbc-4793-931a-9b3c7ba3ae98 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1574.083373] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e236e596-959d-43a4-b155-967a58be998e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1574.099205] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b3988ce-9444-492c-8897-ba927b00a6d6 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1574.105910] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91ac75f7-f38f-421f-ad71-d0de67a8fded {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1574.135589] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180254MB free_disk=95GB free_vcpus=48 pci_devices=None {{(pid=65726) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1136}} [ 1574.135767] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1574.135982] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1575.162599] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Instance 012a1edf-97c5-4ee8-a36c-222ca14eb1f6 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65726) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1575.162878] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Total usable vcpus: 48, total allocated vcpus: 1 {{(pid=65726) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1159}} [ 1575.162935] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=704MB phys_disk=100GB used_disk=1GB total_vcpus=48 used_vcpus=1 pci_stats=[] stats={'failed_builds': '0', 'num_instances': '1', 'num_vm_active': '1', 'num_task_None': '1', 'num_os_type_None': '1', 'num_proj_acd800e0438940bc80ae3c6b672db4ee': '1', 'io_workload': '0'} {{(pid=65726) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1168}} [ 1575.179597] env[65726]: DEBUG nova.scheduler.client.report [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Refreshing inventories for resource provider 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 1575.193341] env[65726]: DEBUG nova.scheduler.client.report [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Updating ProviderTree inventory for provider 07c4692f-bdb4-4058-9173-ff9664830295 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 1575.193526] env[65726]: DEBUG nova.compute.provider_tree [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Updating inventory in ProviderTree for provider 07c4692f-bdb4-4058-9173-ff9664830295 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1575.204543] env[65726]: DEBUG nova.scheduler.client.report [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Refreshing aggregate associations for resource provider 07c4692f-bdb4-4058-9173-ff9664830295, aggregates: None {{(pid=65726) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 1575.223177] env[65726]: DEBUG nova.scheduler.client.report [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Refreshing trait associations for resource provider 07c4692f-bdb4-4058-9173-ff9664830295, traits: COMPUTE_IMAGE_TYPE_ISO,HW_ARCH_X86_64,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=65726) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 1575.249278] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22b1526d-1112-4b11-af71-95f03210a17a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1575.257384] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7560526-8d64-4039-86ef-e0dde1890793 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1575.288510] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7628569-e7b0-4375-9eb3-e35022e41f64 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1575.296842] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3fa53db-0f6e-48c0-b8a5-53ce0ede488f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1575.310464] env[65726]: DEBUG nova.compute.provider_tree [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1575.813791] env[65726]: DEBUG nova.scheduler.client.report [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1576.319498] env[65726]: DEBUG nova.compute.resource_tracker [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=65726) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1097}} [ 1576.319872] env[65726]: DEBUG oslo_concurrency.lockutils [None req-3a590a19-0740-4fb6-a48b-e040cb27db2c None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.184s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1579.047985] env[65726]: DEBUG oslo_concurrency.lockutils [None req-d2473c5a-da5c-4cde-8779-4538492ab704 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Acquiring lock "012a1edf-97c5-4ee8-a36c-222ca14eb1f6" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1579.048365] env[65726]: DEBUG oslo_concurrency.lockutils [None req-d2473c5a-da5c-4cde-8779-4538492ab704 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Lock "012a1edf-97c5-4ee8-a36c-222ca14eb1f6" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1579.048975] env[65726]: DEBUG nova.compute.manager [None req-d2473c5a-da5c-4cde-8779-4538492ab704 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 012a1edf-97c5-4ee8-a36c-222ca14eb1f6] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1579.049908] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4388337-1087-427a-ab47-8f4486cfab0e {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1579.056869] env[65726]: DEBUG nova.compute.manager [None req-d2473c5a-da5c-4cde-8779-4538492ab704 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 012a1edf-97c5-4ee8-a36c-222ca14eb1f6] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=65726) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3459}} [ 1579.057500] env[65726]: DEBUG nova.objects.instance [None req-d2473c5a-da5c-4cde-8779-4538492ab704 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Lazy-loading 'flavor' on Instance uuid 012a1edf-97c5-4ee8-a36c-222ca14eb1f6 {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1580.065814] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-d2473c5a-da5c-4cde-8779-4538492ab704 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 012a1edf-97c5-4ee8-a36c-222ca14eb1f6] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1580.066231] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2f073c67-386f-47b2-93a0-999c69cd753a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1580.073477] env[65726]: DEBUG oslo_vmware.api [None req-d2473c5a-da5c-4cde-8779-4538492ab704 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Waiting for the task: (returnval){ [ 1580.073477] env[65726]: value = "task-5117188" [ 1580.073477] env[65726]: _type = "Task" [ 1580.073477] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1580.082115] env[65726]: DEBUG oslo_vmware.api [None req-d2473c5a-da5c-4cde-8779-4538492ab704 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': task-5117188, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1580.584467] env[65726]: DEBUG oslo_vmware.api [None req-d2473c5a-da5c-4cde-8779-4538492ab704 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': task-5117188, 'name': PowerOffVM_Task, 'duration_secs': 0.204345} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1580.584747] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-d2473c5a-da5c-4cde-8779-4538492ab704 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 012a1edf-97c5-4ee8-a36c-222ca14eb1f6] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1580.584940] env[65726]: DEBUG nova.compute.manager [None req-d2473c5a-da5c-4cde-8779-4538492ab704 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 012a1edf-97c5-4ee8-a36c-222ca14eb1f6] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1580.585767] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-108b479a-29f6-45cf-8b0e-532056e07d7a {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1581.098244] env[65726]: DEBUG oslo_concurrency.lockutils [None req-d2473c5a-da5c-4cde-8779-4538492ab704 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Lock "012a1edf-97c5-4ee8-a36c-222ca14eb1f6" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.050s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1581.446323] env[65726]: DEBUG nova.objects.instance [None req-bc646c21-bf73-4632-8992-3d65bac4af5b tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Lazy-loading 'flavor' on Instance uuid 012a1edf-97c5-4ee8-a36c-222ca14eb1f6 {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1581.951377] env[65726]: DEBUG oslo_concurrency.lockutils [None req-bc646c21-bf73-4632-8992-3d65bac4af5b tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Acquiring lock "refresh_cache-012a1edf-97c5-4ee8-a36c-222ca14eb1f6" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1581.951574] env[65726]: DEBUG oslo_concurrency.lockutils [None req-bc646c21-bf73-4632-8992-3d65bac4af5b tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Acquired lock "refresh_cache-012a1edf-97c5-4ee8-a36c-222ca14eb1f6" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1581.951707] env[65726]: DEBUG nova.network.neutron [None req-bc646c21-bf73-4632-8992-3d65bac4af5b tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 012a1edf-97c5-4ee8-a36c-222ca14eb1f6] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1581.951879] env[65726]: DEBUG nova.objects.instance [None req-bc646c21-bf73-4632-8992-3d65bac4af5b tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Lazy-loading 'info_cache' on Instance uuid 012a1edf-97c5-4ee8-a36c-222ca14eb1f6 {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1582.456061] env[65726]: DEBUG nova.objects.base [None req-bc646c21-bf73-4632-8992-3d65bac4af5b tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Object Instance<012a1edf-97c5-4ee8-a36c-222ca14eb1f6> lazy-loaded attributes: flavor,info_cache {{(pid=65726) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1582.958233] env[65726]: WARNING openstack [None req-bc646c21-bf73-4632-8992-3d65bac4af5b tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1582.958718] env[65726]: WARNING openstack [None req-bc646c21-bf73-4632-8992-3d65bac4af5b tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1583.101589] env[65726]: WARNING openstack [None req-bc646c21-bf73-4632-8992-3d65bac4af5b tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1583.102145] env[65726]: WARNING openstack [None req-bc646c21-bf73-4632-8992-3d65bac4af5b tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1583.166898] env[65726]: WARNING openstack [None req-bc646c21-bf73-4632-8992-3d65bac4af5b tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1583.166898] env[65726]: WARNING openstack [None req-bc646c21-bf73-4632-8992-3d65bac4af5b tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1583.246250] env[65726]: DEBUG nova.network.neutron [None req-bc646c21-bf73-4632-8992-3d65bac4af5b tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 012a1edf-97c5-4ee8-a36c-222ca14eb1f6] Updating instance_info_cache with network_info: [{"id": "2b4783b7-452c-486b-9cad-1ceed6a0d8dc", "address": "fa:16:3e:22:a9:49", "network": {"id": "a899453d-5f43-4650-9642-b33f54a4d2eb", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-614861139-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.231", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "acd800e0438940bc80ae3c6b672db4ee", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "205fb402-8eaf-4b61-8f57-8f216024179a", "external-id": "nsx-vlan-transportzone-78", "segmentation_id": 78, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2b4783b7-45", "ovs_interfaceid": "2b4783b7-452c-486b-9cad-1ceed6a0d8dc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1583.750794] env[65726]: DEBUG oslo_concurrency.lockutils [None req-bc646c21-bf73-4632-8992-3d65bac4af5b tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Releasing lock "refresh_cache-012a1edf-97c5-4ee8-a36c-222ca14eb1f6" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1584.757723] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-bc646c21-bf73-4632-8992-3d65bac4af5b tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 012a1edf-97c5-4ee8-a36c-222ca14eb1f6] Powering on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1584.758115] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1973802d-3f33-403a-97a4-efba0888b2fa {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1584.766101] env[65726]: DEBUG oslo_vmware.api [None req-bc646c21-bf73-4632-8992-3d65bac4af5b tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Waiting for the task: (returnval){ [ 1584.766101] env[65726]: value = "task-5117189" [ 1584.766101] env[65726]: _type = "Task" [ 1584.766101] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1584.774014] env[65726]: DEBUG oslo_vmware.api [None req-bc646c21-bf73-4632-8992-3d65bac4af5b tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': task-5117189, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1585.276902] env[65726]: DEBUG oslo_vmware.api [None req-bc646c21-bf73-4632-8992-3d65bac4af5b tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': task-5117189, 'name': PowerOnVM_Task, 'duration_secs': 0.454871} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1585.277116] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-bc646c21-bf73-4632-8992-3d65bac4af5b tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 012a1edf-97c5-4ee8-a36c-222ca14eb1f6] Powered on the VM {{(pid=65726) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1585.277318] env[65726]: DEBUG nova.compute.manager [None req-bc646c21-bf73-4632-8992-3d65bac4af5b tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 012a1edf-97c5-4ee8-a36c-222ca14eb1f6] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1585.278155] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31931792-80cf-48b9-b3c9-bc9743e18aa6 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1586.254573] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c21f26f6-e01f-44d7-9d51-d99bb90b6585 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1586.261698] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-c1700432-b49a-4415-aec3-d060be581e64 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 012a1edf-97c5-4ee8-a36c-222ca14eb1f6] Suspending the VM {{(pid=65726) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1162}} [ 1586.261947] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-9a3f1194-e0d0-4ca0-87f8-d91d5bd778aa {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1586.269249] env[65726]: DEBUG oslo_vmware.api [None req-c1700432-b49a-4415-aec3-d060be581e64 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Waiting for the task: (returnval){ [ 1586.269249] env[65726]: value = "task-5117190" [ 1586.269249] env[65726]: _type = "Task" [ 1586.269249] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1586.277779] env[65726]: DEBUG oslo_vmware.api [None req-c1700432-b49a-4415-aec3-d060be581e64 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': task-5117190, 'name': SuspendVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1586.780021] env[65726]: DEBUG oslo_vmware.api [None req-c1700432-b49a-4415-aec3-d060be581e64 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': task-5117190, 'name': SuspendVM_Task} progress is 66%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1587.280143] env[65726]: DEBUG oslo_vmware.api [None req-c1700432-b49a-4415-aec3-d060be581e64 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': task-5117190, 'name': SuspendVM_Task, 'duration_secs': 0.608251} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1587.280501] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-c1700432-b49a-4415-aec3-d060be581e64 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 012a1edf-97c5-4ee8-a36c-222ca14eb1f6] Suspended the VM {{(pid=65726) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1166}} [ 1587.280630] env[65726]: DEBUG nova.compute.manager [None req-c1700432-b49a-4415-aec3-d060be581e64 tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 012a1edf-97c5-4ee8-a36c-222ca14eb1f6] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1587.281359] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-903dcd3e-a226-4844-99ee-ff803b611617 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1588.637665] env[65726]: INFO nova.compute.manager [None req-4906c0e6-b557-44e4-83f9-2133a9cd636c tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 012a1edf-97c5-4ee8-a36c-222ca14eb1f6] Resuming [ 1588.638257] env[65726]: DEBUG nova.objects.instance [None req-4906c0e6-b557-44e4-83f9-2133a9cd636c tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Lazy-loading 'flavor' on Instance uuid 012a1edf-97c5-4ee8-a36c-222ca14eb1f6 {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1590.149154] env[65726]: DEBUG oslo_concurrency.lockutils [None req-4906c0e6-b557-44e4-83f9-2133a9cd636c tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Acquiring lock "refresh_cache-012a1edf-97c5-4ee8-a36c-222ca14eb1f6" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1590.149550] env[65726]: DEBUG oslo_concurrency.lockutils [None req-4906c0e6-b557-44e4-83f9-2133a9cd636c tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Acquired lock "refresh_cache-012a1edf-97c5-4ee8-a36c-222ca14eb1f6" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1590.149550] env[65726]: DEBUG nova.network.neutron [None req-4906c0e6-b557-44e4-83f9-2133a9cd636c tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 012a1edf-97c5-4ee8-a36c-222ca14eb1f6] Building network info cache for instance {{(pid=65726) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1590.654359] env[65726]: WARNING openstack [None req-4906c0e6-b557-44e4-83f9-2133a9cd636c tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1590.654701] env[65726]: WARNING openstack [None req-4906c0e6-b557-44e4-83f9-2133a9cd636c tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1590.791161] env[65726]: WARNING openstack [None req-4906c0e6-b557-44e4-83f9-2133a9cd636c tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1590.791587] env[65726]: WARNING openstack [None req-4906c0e6-b557-44e4-83f9-2133a9cd636c tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1590.858409] env[65726]: WARNING openstack [None req-4906c0e6-b557-44e4-83f9-2133a9cd636c tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1590.858793] env[65726]: WARNING openstack [None req-4906c0e6-b557-44e4-83f9-2133a9cd636c tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1590.938571] env[65726]: DEBUG nova.network.neutron [None req-4906c0e6-b557-44e4-83f9-2133a9cd636c tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 012a1edf-97c5-4ee8-a36c-222ca14eb1f6] Updating instance_info_cache with network_info: [{"id": "2b4783b7-452c-486b-9cad-1ceed6a0d8dc", "address": "fa:16:3e:22:a9:49", "network": {"id": "a899453d-5f43-4650-9642-b33f54a4d2eb", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-614861139-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.231", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "acd800e0438940bc80ae3c6b672db4ee", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "205fb402-8eaf-4b61-8f57-8f216024179a", "external-id": "nsx-vlan-transportzone-78", "segmentation_id": 78, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2b4783b7-45", "ovs_interfaceid": "2b4783b7-452c-486b-9cad-1ceed6a0d8dc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1591.441566] env[65726]: DEBUG oslo_concurrency.lockutils [None req-4906c0e6-b557-44e4-83f9-2133a9cd636c tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Releasing lock "refresh_cache-012a1edf-97c5-4ee8-a36c-222ca14eb1f6" {{(pid=65726) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1591.442562] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d802bd17-a79d-4dba-be0d-7ae9707b7e42 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1591.450769] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-4906c0e6-b557-44e4-83f9-2133a9cd636c tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 012a1edf-97c5-4ee8-a36c-222ca14eb1f6] Resuming the VM {{(pid=65726) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1183}} [ 1591.451034] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-891c20a7-0ae0-4b0a-8919-77067e638dd5 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1591.457731] env[65726]: DEBUG oslo_vmware.api [None req-4906c0e6-b557-44e4-83f9-2133a9cd636c tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Waiting for the task: (returnval){ [ 1591.457731] env[65726]: value = "task-5117191" [ 1591.457731] env[65726]: _type = "Task" [ 1591.457731] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1591.466365] env[65726]: DEBUG oslo_vmware.api [None req-4906c0e6-b557-44e4-83f9-2133a9cd636c tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': task-5117191, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1591.969447] env[65726]: DEBUG oslo_vmware.api [None req-4906c0e6-b557-44e4-83f9-2133a9cd636c tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': task-5117191, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1592.469207] env[65726]: DEBUG oslo_vmware.api [None req-4906c0e6-b557-44e4-83f9-2133a9cd636c tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': task-5117191, 'name': PowerOnVM_Task, 'duration_secs': 0.549481} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1592.469595] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-4906c0e6-b557-44e4-83f9-2133a9cd636c tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 012a1edf-97c5-4ee8-a36c-222ca14eb1f6] Resumed the VM {{(pid=65726) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1188}} [ 1592.469718] env[65726]: DEBUG nova.compute.manager [None req-4906c0e6-b557-44e4-83f9-2133a9cd636c tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 012a1edf-97c5-4ee8-a36c-222ca14eb1f6] Checking state {{(pid=65726) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1592.470485] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8edf7b2-9e18-4b78-8aa0-1a58ab827132 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1593.366320] env[65726]: DEBUG oslo_concurrency.lockutils [None req-f6e55432-94fd-49d5-b856-8043e49cd3fd tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Acquiring lock "012a1edf-97c5-4ee8-a36c-222ca14eb1f6" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1593.366594] env[65726]: DEBUG oslo_concurrency.lockutils [None req-f6e55432-94fd-49d5-b856-8043e49cd3fd tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Lock "012a1edf-97c5-4ee8-a36c-222ca14eb1f6" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1593.366822] env[65726]: DEBUG oslo_concurrency.lockutils [None req-f6e55432-94fd-49d5-b856-8043e49cd3fd tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Acquiring lock "012a1edf-97c5-4ee8-a36c-222ca14eb1f6-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1593.367021] env[65726]: DEBUG oslo_concurrency.lockutils [None req-f6e55432-94fd-49d5-b856-8043e49cd3fd tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Lock "012a1edf-97c5-4ee8-a36c-222ca14eb1f6-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1593.367191] env[65726]: DEBUG oslo_concurrency.lockutils [None req-f6e55432-94fd-49d5-b856-8043e49cd3fd tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Lock "012a1edf-97c5-4ee8-a36c-222ca14eb1f6-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1593.369897] env[65726]: INFO nova.compute.manager [None req-f6e55432-94fd-49d5-b856-8043e49cd3fd tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 012a1edf-97c5-4ee8-a36c-222ca14eb1f6] Terminating instance [ 1593.874316] env[65726]: DEBUG nova.compute.manager [None req-f6e55432-94fd-49d5-b856-8043e49cd3fd tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 012a1edf-97c5-4ee8-a36c-222ca14eb1f6] Start destroying the instance on the hypervisor. {{(pid=65726) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1593.874699] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-f6e55432-94fd-49d5-b856-8043e49cd3fd tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 012a1edf-97c5-4ee8-a36c-222ca14eb1f6] Destroying instance {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1593.875897] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f2a1d7c-8c21-4c5c-b7f7-f619504aa45b {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1593.883687] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-f6e55432-94fd-49d5-b856-8043e49cd3fd tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 012a1edf-97c5-4ee8-a36c-222ca14eb1f6] Powering off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1593.883919] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5c4caa41-98b1-46cc-87d2-faddeaa9d8fa {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1593.890149] env[65726]: DEBUG oslo_vmware.api [None req-f6e55432-94fd-49d5-b856-8043e49cd3fd tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Waiting for the task: (returnval){ [ 1593.890149] env[65726]: value = "task-5117192" [ 1593.890149] env[65726]: _type = "Task" [ 1593.890149] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1593.898986] env[65726]: DEBUG oslo_vmware.api [None req-f6e55432-94fd-49d5-b856-8043e49cd3fd tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': task-5117192, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1594.402305] env[65726]: DEBUG oslo_vmware.api [None req-f6e55432-94fd-49d5-b856-8043e49cd3fd tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': task-5117192, 'name': PowerOffVM_Task, 'duration_secs': 0.214459} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1594.402305] env[65726]: DEBUG nova.virt.vmwareapi.vm_util [None req-f6e55432-94fd-49d5-b856-8043e49cd3fd tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 012a1edf-97c5-4ee8-a36c-222ca14eb1f6] Powered off the VM {{(pid=65726) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1594.402305] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-f6e55432-94fd-49d5-b856-8043e49cd3fd tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 012a1edf-97c5-4ee8-a36c-222ca14eb1f6] Unregistering the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1594.402547] env[65726]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f8d15ab5-a4e1-4e11-b5c6-39c8ac8b9ec4 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1594.468119] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-f6e55432-94fd-49d5-b856-8043e49cd3fd tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 012a1edf-97c5-4ee8-a36c-222ca14eb1f6] Unregistered the VM {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1594.468385] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-f6e55432-94fd-49d5-b856-8043e49cd3fd tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 012a1edf-97c5-4ee8-a36c-222ca14eb1f6] Deleting contents of the VM from datastore datastore2 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1594.468537] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-f6e55432-94fd-49d5-b856-8043e49cd3fd tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Deleting the datastore file [datastore2] 012a1edf-97c5-4ee8-a36c-222ca14eb1f6 {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1594.468814] env[65726]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5bab0c75-7246-485c-ae01-aaf17d619f2f {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1594.476528] env[65726]: DEBUG oslo_vmware.api [None req-f6e55432-94fd-49d5-b856-8043e49cd3fd tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Waiting for the task: (returnval){ [ 1594.476528] env[65726]: value = "task-5117194" [ 1594.476528] env[65726]: _type = "Task" [ 1594.476528] env[65726]: } to complete. {{(pid=65726) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1594.485905] env[65726]: DEBUG oslo_vmware.api [None req-f6e55432-94fd-49d5-b856-8043e49cd3fd tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': task-5117194, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1594.986362] env[65726]: DEBUG oslo_vmware.api [None req-f6e55432-94fd-49d5-b856-8043e49cd3fd tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Task: {'id': task-5117194, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.207125} completed successfully. {{(pid=65726) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1594.986735] env[65726]: DEBUG nova.virt.vmwareapi.ds_util [None req-f6e55432-94fd-49d5-b856-8043e49cd3fd tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Deleted the datastore file {{(pid=65726) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1594.986782] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-f6e55432-94fd-49d5-b856-8043e49cd3fd tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 012a1edf-97c5-4ee8-a36c-222ca14eb1f6] Deleted contents of the VM from datastore datastore2 {{(pid=65726) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1594.986939] env[65726]: DEBUG nova.virt.vmwareapi.vmops [None req-f6e55432-94fd-49d5-b856-8043e49cd3fd tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 012a1edf-97c5-4ee8-a36c-222ca14eb1f6] Instance destroyed {{(pid=65726) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1594.987126] env[65726]: INFO nova.compute.manager [None req-f6e55432-94fd-49d5-b856-8043e49cd3fd tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] [instance: 012a1edf-97c5-4ee8-a36c-222ca14eb1f6] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1594.987367] env[65726]: DEBUG oslo.service.backend._common.loopingcall [None req-f6e55432-94fd-49d5-b856-8043e49cd3fd tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65726) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1594.987588] env[65726]: DEBUG nova.compute.manager [-] [instance: 012a1edf-97c5-4ee8-a36c-222ca14eb1f6] Deallocating network for instance {{(pid=65726) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1594.987688] env[65726]: DEBUG nova.network.neutron [-] [instance: 012a1edf-97c5-4ee8-a36c-222ca14eb1f6] deallocate_for_instance() {{(pid=65726) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 1594.988200] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1594.988455] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1595.024231] env[65726]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1595.024516] env[65726]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1595.459278] env[65726]: DEBUG nova.compute.manager [req-d6ebac53-dccd-416b-9138-7d3d56380477 req-6d622751-3f38-4cd0-a218-b7ba83ed0ad7 service nova] [instance: 012a1edf-97c5-4ee8-a36c-222ca14eb1f6] Received event network-vif-deleted-2b4783b7-452c-486b-9cad-1ceed6a0d8dc {{(pid=65726) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 1595.459480] env[65726]: INFO nova.compute.manager [req-d6ebac53-dccd-416b-9138-7d3d56380477 req-6d622751-3f38-4cd0-a218-b7ba83ed0ad7 service nova] [instance: 012a1edf-97c5-4ee8-a36c-222ca14eb1f6] Neutron deleted interface 2b4783b7-452c-486b-9cad-1ceed6a0d8dc; detaching it from the instance and deleting it from the info cache [ 1595.459720] env[65726]: DEBUG nova.network.neutron [req-d6ebac53-dccd-416b-9138-7d3d56380477 req-6d622751-3f38-4cd0-a218-b7ba83ed0ad7 service nova] [instance: 012a1edf-97c5-4ee8-a36c-222ca14eb1f6] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1595.934991] env[65726]: DEBUG nova.network.neutron [-] [instance: 012a1edf-97c5-4ee8-a36c-222ca14eb1f6] Updating instance_info_cache with network_info: [] {{(pid=65726) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1595.962289] env[65726]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ee405c89-f3cf-443f-9bdc-b75b871771d1 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1595.973079] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fda933a-3645-4613-869b-999420568277 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1596.000360] env[65726]: DEBUG nova.compute.manager [req-d6ebac53-dccd-416b-9138-7d3d56380477 req-6d622751-3f38-4cd0-a218-b7ba83ed0ad7 service nova] [instance: 012a1edf-97c5-4ee8-a36c-222ca14eb1f6] Detach interface failed, port_id=2b4783b7-452c-486b-9cad-1ceed6a0d8dc, reason: Instance 012a1edf-97c5-4ee8-a36c-222ca14eb1f6 could not be found. {{(pid=65726) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11669}} [ 1596.438241] env[65726]: INFO nova.compute.manager [-] [instance: 012a1edf-97c5-4ee8-a36c-222ca14eb1f6] Took 1.45 seconds to deallocate network for instance. [ 1596.945126] env[65726]: DEBUG oslo_concurrency.lockutils [None req-f6e55432-94fd-49d5-b856-8043e49cd3fd tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1596.945391] env[65726]: DEBUG oslo_concurrency.lockutils [None req-f6e55432-94fd-49d5-b856-8043e49cd3fd tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1596.945648] env[65726]: DEBUG nova.objects.instance [None req-f6e55432-94fd-49d5-b856-8043e49cd3fd tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Lazy-loading 'resources' on Instance uuid 012a1edf-97c5-4ee8-a36c-222ca14eb1f6 {{(pid=65726) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1597.486336] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85cce9f2-9c52-43fa-8d24-ec03d8f8c4c1 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1597.494612] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0972ef68-b71d-4d27-89a1-e5ed6eee4eca {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1597.526269] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a00a97f6-d619-45fb-998f-94eacb6cd64d {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1597.534058] env[65726]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a4f9224-cac9-400e-869d-feb24c085462 {{(pid=65726) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1597.548033] env[65726]: DEBUG nova.compute.provider_tree [None req-f6e55432-94fd-49d5-b856-8043e49cd3fd tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Inventory has not changed in ProviderTree for provider: 07c4692f-bdb4-4058-9173-ff9664830295 {{(pid=65726) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1598.051792] env[65726]: DEBUG nova.scheduler.client.report [None req-f6e55432-94fd-49d5-b856-8043e49cd3fd tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Inventory has not changed for provider 07c4692f-bdb4-4058-9173-ff9664830295 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65726) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1598.558280] env[65726]: DEBUG oslo_concurrency.lockutils [None req-f6e55432-94fd-49d5-b856-8043e49cd3fd tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.613s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1598.580632] env[65726]: INFO nova.scheduler.client.report [None req-f6e55432-94fd-49d5-b856-8043e49cd3fd tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Deleted allocations for instance 012a1edf-97c5-4ee8-a36c-222ca14eb1f6 [ 1599.091996] env[65726]: DEBUG oslo_concurrency.lockutils [None req-f6e55432-94fd-49d5-b856-8043e49cd3fd tempest-ServerActionsTestJSON-1690106724 tempest-ServerActionsTestJSON-1690106724-project-member] Lock "012a1edf-97c5-4ee8-a36c-222ca14eb1f6" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.725s {{(pid=65726) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}}